source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Design%20closure
Design Closure is a part of the digital electronic design automation workflow by which an integrated circuit (i.e. VLSI) design is modified from its initial description to meet a growing list of design constraints and objectives. Every step in the IC design (such as static timing analysis, placement, routing, and so on) is already complex and often forms its own field of study. This article, however, looks at the overall design closure process, which takes a chip from its initial design state to the final form in which all of its design constraints are met. Introduction Every chip starts off as someone’s idea of a good thing: "If we can make a part that performs function X, we will all be rich!" Once the concept is established, someone from marketing says "To make this chip profitably, it must cost $C and run at frequency F." Someone from manufacturing says "To meet this chip’s targets, it must have a yield of Y%." Someone from packaging says “It must fit in the P package and dissipate no more than W watts.” Eventually, the team generates an extensive list of all the constraints and objectives they must meet to manufacture a product that can be sold profitably. The management then forms a design team, which consists of chip architects, logic designers, functional verification engineers, physical designers, and timing engineers, and assigns them to create a chip to the specifications. Constraints vs Objectives The distinction between constraints and objectives is straightforward: a constraint is a design target that must be met for the design to be successful. For example, a chip may be required to run at a specific frequency so it can interface with other components in a system. In contrast, an objective is a design target where more (or less) is better. For example, yield is generally an objective, which is maximized to lower manufacturing cost. For the purposes of design closure, the distinction between constraints and objectives is not important; this artic
https://en.wikipedia.org/wiki/Cameron%20graph
The Cameron graph is a strongly regular graph of parameters . External links A.E. Brouwer's website: the Cameron graph Individual graphs Regular graphs Strongly regular graphs
https://en.wikipedia.org/wiki/Gamow%20factor
The Gamow factor, Sommerfeld factor or Gamow–Sommerfeld factor, named after its discoverer George Gamow or after Arnold Sommerfeld, is a probability factor for two nuclear particles' chance of overcoming the Coulomb barrier in order to undergo nuclear reactions, for example in nuclear fusion. By classical physics, there is almost no possibility for protons to fuse by crossing each other's Coulomb barrier at temperatures commonly observed to cause fusion, such as those found in the sun. When George Gamow instead applied quantum mechanics to the problem, he found that there was a significant chance for the fusion due to tunneling. The probability of two nuclear particles overcoming their electrostatic barriers is given by the following equation: where is the Gamow energy, Here, is the reduced mass of the two particles. The constant is the fine structure constant, is the speed of light, and and are the respective atomic numbers of each particle. While the probability of overcoming the Coulomb barrier increases rapidly with increasing particle energy, for a given temperature, the probability of a particle having such an energy falls off very fast, as described by the Maxwell–Boltzmann distribution. Gamow found that, taken together, these effects mean that for any given temperature, the particles that fuse are mostly in a temperature-dependent narrow range of energies known as the Gamow window. Derivation Gamow first solved the one-dimensional case of quantum tunneling using the WKB approximation. Considering a wave function of a particle of mass m, we take area 1 to be where a wave is emitted, area 2 the potential barrier which has height V and width l (at ), and area 3 its other side, where the wave is arriving, partly transmitted and partly reflected. For a wave number k and energy E we get: where and . This is solved for given A and α by taking the boundary conditions at the both barrier edges, at and , where both and its derivative must be equal
https://en.wikipedia.org/wiki/Optic%20canal
The optic foramen is the opening to the optic canal. The canal is located in the sphenoid bone; it is bounded medially by the body of the sphenoid and laterally by the lesser wing of the sphenoid. The superior surface of the sphenoid bone is bounded behind by a ridge, which forms the anterior border of a narrow, transverse groove, the chiasmatic groove (optic groove), above and behind which lies the optic chiasma; the groove ends on either side in the optic foramen, which transmits the optic nerve and ophthalmic artery (with accompanying sympathetic nerve fibres) into the orbital cavity. Compared to the optic nerve, the ophthalmic artery is located inferolaterally within the canal. The left and right optic canals are 25mm apart posteriorly and 30mm apart anteriorly. The canals themselves are funnel-shaped (narrowest anteriorly). Additional images See also Foramina of skull Ophryon
https://en.wikipedia.org/wiki/Chromosome%20microdissection
Chromosome microdissection is a technique that physically removes a large section of DNA from a complete chromosome. The smallest portion of DNA that can be isolated using this method comprises 10 million base pairs - hundreds or thousands of individual genes. Scientists who study chromosomes are known as cytogeneticists. They are able to identify each chromosome based on its unique pattern of dark and light bands. Certain abnormalities, however, cause chromosomes to have unusual banding patterns. For example, one chromosome may have a piece of another chromosome inserted within it, creating extra bands. Or, a portion of a chromosome may be repeated over and over again, resulting in an unusually wide, dark band (known as a homogeneously staining region). Some chromosomal aberrations have been linked to cancer and inherited genetic disorders, and the chromosomes of many tumor cells exhibit irregular bands. To understand more about what causes these conditions, scientists hope to determine which genes and DNA sequences are located near these unusual bands. Chromosome microdissection is a specialized way of isolating these regions by removing the DNA from the band and making that DNA available for further study. To prepare cells for chromosome microdissection, a scientist first treats them with a chemical that forces them into metaphase: a phase of the cell's life-cycle where the chromosomes are tightly coiled and highly visible. Next, the cells are dropped onto a microscope slide so that the nucleus, which holds all of the genetic material together, breaks apart and releases the chromosomes onto the slide. Then, under a microscope, the scientist locates the specific band of interest, and, using a very fine needle, tears that band away from the rest of the chromosome. The researcher next produces multiple copies of the isolated DNA using a procedure called PCR (polymerase chain reaction). The scientist uses these copies to study the DNA from the unusual region of the
https://en.wikipedia.org/wiki/Atmospheric%20window
An atmospheric window is a range of wavelengths of the electromagnetic spectrum that can pass through the atmosphere of Earth. The optical, infrared and radio windows comprise the three main atmospheric windows. The windows provide direct channels for Earth's surface to receive electromagnetic energy from the Sun, and for thermal radiation from the surface to leave to space. Atmospheric windows are useful for astronomy, remote sensing, telecommunications and other science & technology applications. In the study of the greenhouse effect, the term atmospheric window may be limited to mean the infrared window, which is the primary escape route for a fraction of the thermal radiation emitted near the surface. In other fields of science and technology, such as radio astronomy and remote sensing, the term is used as a hypernym, covering the whole electromagnetic spectrum as in the present article. Role in Earth's energy budget Atmospheric windows, especially the optical and infrared, affect the distribution of energy flows and temperatures within Earth's energy balance. The windows are themselves dependent upon clouds, water vapor, trace greenhouse gases, and other components of the atmosphere. Out of an average 340 watts per square meter (W/m2) of solar irradiance at the top of the atmosphere, about 200 W/m2 reaches the surface via windows, mostly the optical and infrared. Also, out of about 340 W/m2 of reflected shortwave (105 W/m2) plus outgoing longwave radiation (235 W/m2), 80-100 W/m2 exits to space through the infrared window depending on cloudiness. About 40 W/m2 of this transmitted amount is emitted by the surface, while most of the remainder comes from lower regions of the atmosphere. In a complimentary manner, the infrared window also transmits to the surface a portion of down-welling thermal radiation that is emitted within colder upper regions of the atmosphere. The "window" concept is useful to provide qualitative insight into some important features o
https://en.wikipedia.org/wiki/24-Isopropylcholestane
24-isopropyl cholestane is an organic molecule produced by specific sponges, protists and marine algae. The identification of this molecule at high abundances in Neoproterozoic rocks has been interpreted to reflect the presence of multicellular life prior to the rapid diversification and radiation of life during the Cambrian explosion. In this transitional period at the start of the Phanerozoic, single-celled organisms evolved to produce many of the evolutionary lineages present on Earth today. Interpreting 24-isopropyl cholestane in ancient rocks as indicating the presence of sponges before this rapid diversification event alters the traditional understanding of the evolution of multicellular life and the coupling of biology to changes in end-Neoproterozoic climate. However, there are several arguments against causally linking 24-isopropyl cholestane to sponges based on considerations of marine algae and the potential alteration of organic molecules over geologic time. In particular the discovery of 24-isopropyl cholestane in rhizarian protists implies that this biomarker cannot be used on its own to trace sponges. Interpreting the presence of 24-isopropyl cholestane in the context of changingglobal biogeochemical cycles at the Proterozoic-Phanerozoic transition remains an area of active research. 24-isopropyl cholestane Chemical argument for Precambrian sponges 24-isopropyl cholestane (figure 1, left) is a C30 sterane with chemical formula C30H54 and molecular mass 414.76 g/mol. The molecule has a cholestane skeleton with an isopropyl moiety at C24 and is the geologically stable form of 24-isopropyl cholesterol. A related and important molecule is 24-n-propyl cholestane (figure 1, right), also with the cholestane skeleton, but with an n-propyl moiety at C24. 24-isopropyl cholestane is produced copiously by a particular group of sponges in the class Demospongiae within the phylum Porifera. Like other molecular fossils, the presence of 24-isopropyl cholestane in
https://en.wikipedia.org/wiki/107th%20meridian%20east
The meridian 107° east of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Asia, the Indian Ocean, the Southern Ocean, and Antarctica to the South Pole. The 107th meridian east forms a great circle with the 73rd meridian west. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 107th meridian east passes through: {| class="wikitable" ! Width="130" | Co-ordinates ! Country, territory or sea ! Notes |- | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | Laptev Sea | style="background:#b0e0e6;" | |- | | | Krasnoyarsk Krai — Maly Taymyr Island, Severnaya Zemlya |-valign="top" | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | Laptev Sea | style="background:#b0e0e6;" | Passing between the Komsomolskaya Pravda Islands, Krasnoyarsk Krai, (at ) |- | | | Krasnoyarsk Krai — Taymyr Peninsula |- | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | Faddey Bay | style="background:#b0e0e6;" | |- | | | Krasnoyarsk Krai — Taymyr Peninsula |- | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | Khatanga Gulf | style="background:#b0e0e6;" | |-valign="top" | | | Krasnoyarsk Krai Sakha Republic — from Krasnoyarsk Krai — from Irkutsk Oblast — from Republic of Buryatia — from (border is in Lake Baikal) |- | | | Passing just east of Ulan Bator (at ) |-valign="top" | | | Inner Mongolia Ningxia – from Gansu – from Shaanxi – from Sichuan – from Chongqing – from Guizhou – from Guangxi – from |- | | | Mainland and Cat Ba Island |- | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | South China Sea | style="background:#b0e0e6;" | Gulf of Tonkin |- | | | |- | | | |- | | | |- | | | |-valign="top" | style="background:#b0e0e6;" | | style="background:#b0e0e6;" | South China Sea | style="background:#b0e0e6;" | Pas
https://en.wikipedia.org/wiki/Superspace
Superspace is the coordinate space of a theory exhibiting supersymmetry. In such a formulation, along with ordinary space dimensions x, y, z, ..., there are also "anticommuting" dimensions whose coordinates are labeled in Grassmann numbers rather than real numbers. The ordinary space dimensions correspond to bosonic degrees of freedom, the anticommuting dimensions to fermionic degrees of freedom. The word "superspace" was first used by John Wheeler in an unrelated sense to describe the configuration space of general relativity; for example, this usage may be seen in his 1973 textbook Gravitation. Informal discussion There are several similar, but not equivalent, definitions of superspace that have been used, and continue to be used in the mathematical and physics literature. One such usage is as a synonym for super Minkowski space. In this case, one takes ordinary Minkowski space, and extends it with anti-commuting fermionic degrees of freedom, taken to be anti-commuting Weyl spinors from the Clifford algebra associated to the Lorentz group. Equivalently, the super Minkowski space can be understood as the quotient of the super Poincaré algebra modulo the algebra of the Lorentz group. A typical notation for the coordinates on such a space is with the overline being the give-away that super Minkowski space is the intended space. Superspace is also commonly used as a synonym for the super vector space. This is taken to be an ordinary vector space, together with additional coordinates taken from the Grassmann algebra, i.e. coordinate directions that are Grassmann numbers. There are several conventions for constructing a super vector space in use; two of these are described by Rogers and DeWitt. A third usage of the term "superspace" is as a synonym for a supermanifold: a supersymmetric generalization of a manifold. Note that both super Minkowski spaces and super vector spaces can be taken as special cases of supermanifolds. A fourth, and completely unrelated mean
https://en.wikipedia.org/wiki/Kundu%20equation
The Kundu equation is a general form of integrable system that is gauge-equivalent to the mixed nonlinear Schrödinger equation. It was proposed by Anjan Kundu as with arbitrary function and the subscripts denoting partial derivatives. Equation (1) is shown to be reducible for the choice of to an integrable class of mixed nonlinear Schrödinger equation with cubic–quintic nonlinearity, given in a representative form Here are independent parameters, while Equation , more specifically equation is known as the Kundu equation. Properties and applications The Kundu equation is a completely integrable system, allowing Lax pair representation, exact solutions, and higher conserved quantity. Along with its different particular cases, this equation has been investigated for finding its exact travelling wave solutions, exact solitary wave solutions via bilinearization, and Darboux transformation together with the orbital stability for such solitary wave solutions. The Kundu equation has been applied to various physical processes such as fluid dynamics, plasma physics, and nonlinear optics. It is linked to the mixed nonlinear Schrödinger equation through a gauge transformation and is reducible to a variety of known integrable equations such as the nonlinear Schrödinger equation (NLSE), derivative NLSE, higher nonlinear derivative NLSE, Chen–Lee–Liu, Gerjikov-Vanov, and Kundu–Eckhaus equations, for different choices of the parameters. Kundu-Eckhaus equation A generalization of the nonlinear Schrödinger equation with additional quintic nonlinearity and a nonlinear dispersive term was proposed in the form which may be obtained from the Kundu Equation , when restricted to . The same equation, limited further to the particular case was introduced later as the Eckhaus equation, following which equation is presently known as the Kundu-Ekchaus equation. The Kundu-Ekchaus equation can be reduced to the nonlinear Schrödinger equation through a nonlinear transformation of th
https://en.wikipedia.org/wiki/Mr.%20Do%27s%20Wild%20Ride
Mr. Do's Wild Ride is a platform game released in 1984 as the third in Universal's Mr. Do! arcade video game series. An MSX version was published in 1985. Gameplay Mr. Do!'s scenario is a roller coaster, and the object is to reach the top. As the cars (and eventually other objects) speed around the track, the player must escape by using a super speed button, or by climbing up small ladders scattered about the track to dodge the hazards. Two icons at the end of the level range from cakes to EXTRA letters or diamonds change upon collecting cherries at the top of each letter. The game is timed, and the timer ticks faster when the super speed button is held down. Collision with a roller coaster car or another object is fatal, knocking Mr. Do! off the coaster and costing a life. After the sixth level is completed, the game cycles back to the first with various obstacles and/or more roller coaster cars to avoid. Reception In Japan, Game Machine listed Mr. Do's Wild Ride on their April 15, 1984 issue as being the sixth most-successful table arcade unit of the month. Legacy Ocean Software published a clone for the Commodore 64, Amstrad CPC, and ZX Spectrum as Kong Strikes Back! It incorporates cosmetic aspects of Donkey Kong.
https://en.wikipedia.org/wiki/Caloribacterium
Caldanaerobius is a moderately thermophilic and anaerobic genus of bacteria from the family of Thermoanaerobacteraceae with one known species (Caloribacterium cisternae).
https://en.wikipedia.org/wiki/Redback%20Networks
Redback Networks provided hardware and software used by Internet service providers to manage broadband services. The company's products included the SMS (Subscriber Management System), SmartEdge, and SmartMetro product lines. In January 2007, the company was acquired by Ericsson. History Redback Networks was founded in August 1996 by Gaurav Garg, Asher Waldfogel, and William M. Salkewicz. The company received seed money from Sequoia Capital. In May 1999, during the dot-com bubble, the company became a public company via an initial public offering. After pricing at $23 each, shares soared 266% on the first day of trading. In November 1999, the company acquired Siara Systems, which at the time only had products in the prototype stage, for $4.3 billion in stock. In 2000, its share price peaked at $198 but fell to $0.27 in October 2002, after the burst of the dot-com bubble. In August 2000, the company acquired Abatis Systems. In October 2000, the company opened a regional headquarters in Hong Kong. In January 2007, the company was acquired by Ericsson for $1.9 billion, or $25 per share.
https://en.wikipedia.org/wiki/Versatile%20Toroidal%20Facility
The Versatile Toroidal Facility (VTF) is a research group within the Physics Research Division of the MIT Plasma Science and Fusion Center at the Massachusetts Institute of Technology. The VTF is a laboratory focused on studying the phenomenon of magnetic reconnection. For this purpose the group has a small tokamak designed to observe rarefied plasmas with probes. These probes measure electric and magnetic field behavior as well as various plasma characteristics in order to better understand the poorly understood processes involved in magnetic reconnection. The VTF is a fundamental physics research group, and its research has wide-ranging and immediate impact on our understanding of such plasma-related subjects as solar flares, the aurora borealis, magnetic confinement fusion, and magnetohydrodynamic theory in general. The VTF is was built and originally led by Dr. Marcel Gaudreau, and prior to its retirement, was led by Dr. Miklos Porkolab and Dr. Jan Egedal, all MIT faculty at the time. External links PSFC homepage VTF homepage Massachusetts Institute of Technology Plasma physics facilities Tokamaks
https://en.wikipedia.org/wiki/Softmax%20function
The softmax function, also known as softargmax or normalized exponential function, converts a vector of real numbers into a probability distribution of possible outcomes. It is a generalization of the logistic function to multiple dimensions, and used in multinomial logistic regression. The softmax function is often used as the last activation function of a neural network to normalize the output of a network to a probability distribution over predicted output classes, based on Luce's choice axiom. Definition The softmax function takes as input a vector of real numbers, and normalizes it into a probability distribution consisting of probabilities proportional to the exponentials of the input numbers. That is, prior to applying softmax, some vector components could be negative, or greater than one; and might not sum to 1; but after applying softmax, each component will be in the interval , and the components will add up to 1, so that they can be interpreted as probabilities. Furthermore, the larger input components will correspond to larger probabilities. The standard (unit) softmax function where is defined by the formula In words, it applies the standard exponential function to each element of the input vector and normalizes these values by dividing by the sum of all these exponentials. The normalization ensures that the sum of the components of the output vector is 1. The term "softmax" derives from the amplifying effects of the exponential on any maxima in the input vector. For example, the standard softmax of is approximately , which amounts to assigning almost all of the total unit weight in the result to the position of the vector's maximal element (of 8). In general, instead of a different base can be used. If , smaller input components will result in larger output probabilities, and decreasing the value of will create probability distributions that are more concentrated around the positions of the smallest input values. Conversely, as ab
https://en.wikipedia.org/wiki/Xg%20antigen%20system
The XG antigen is a red blood cell surface antigen discovered in 1962. by researchers at the MRC Blood Group Unit. The PBDX gene that encodes the antigen is located on the short arm of the X chromosome. Since males normally have one X chromosome they are considered hemizygotes. Since women have two copies of the gene and could be heterozygotic for the presence or absence of the functioning gene they could (through the process of lyonisation) express the functioning protein on just some of their red blood cells. Frequency
https://en.wikipedia.org/wiki/Cometabolism
Cometabolism is defined as the simultaneous degradation of two compounds, in which the degradation of the second compound (the secondary substrate) depends on the presence of the first compound (the primary substrate). This is in contrast to simultaneous catabolism, where each substrate is catabolized concomitantly by different enzymes. Cometabolism occurs when an enzyme produced by an organism to catalyze the degradation of its growth-substrate to derive energy and carbon from it is also capable of degrading additional compounds. The fortuitous degradation of these additional compounds does not support the growth of the bacteria, and some of these compounds can even be toxic in certain concentrations to the bacteria. The first report of this phenomenon was the degradation of ethane by the species Pseudomonas methanica. These bacteria degrade their growth-substrate methane with the enzyme methane monooxygenase (MMO). MMO was discovered to be capable of degrading ethane and propane, although the bacteria were unable to use these compounds as energy and carbon sources to grow. Another example is Mycobacterium vaccae, which uses an alkane monooxygenase enzyme to oxidize propane. Accidentally, this enzyme also oxidizes, at no additional cost for M. vaccae, cyclohexane into cyclohexanol. Thus, cyclohexane is co-metabolized in the presence of propane. This allows for the commensal growth of Pseudomonas on cyclohexane. The latter can metabolize cyclohexanol, but not cyclohexane. Cometabolism in Bioremediation Some of the molecules that are cometabolically degraded by bacteria are xenobiotic, persistent compounds, such as PCE, TCE, and MTBE, that have harmful effects on several types of environments. Co-metabolism is thus used as an approach to biologically degrade hazardous solvents. Cometabolism can be used for the biodegradation of methyl-tert-butyl ether (MTBE): an aquatic environment pollutant. Some Pseudomonas members were found to be able to fully degrade MTBE c
https://en.wikipedia.org/wiki/History%20of%20Norsk%20Data
Norsk Data (ND) was a Norwegian manufacturer of minicomputers which operated between 1967 and 1992. The company was established as A/S Nordata – Norsk Data-Elektronikk on 7 July 1967 and took into use the Norsk Data brand in 1975. The company was founded by Lars Monrad-Krohn, Rolf Skår and Per Bjørge, three computer engineers working at the Norwegian Defence Research Establishment which had just built the minicomputer SAM 2. ND's first contract was the delivery of a Nord-1 computer to Norcontrol. Initially in competition with Kongsberg, ND started delivering computers to Norwegian institutions. By 1972 the company had developed Sintran operating system, the 32-bit Nord-5 and a time sharing system. The international break-through came with the 1973 delivery of computers to CERN and the company soon had half their sales abroad. Two years later the database program Sibas (SIBAS is (tm) of SRS that has full rights to the code developed at the Central Institute for Industrial Research by Olli, Salter, Aschim and Hoffmann) had been completely ported and made available, and the following year a 150-terminal system connected via X.21/X.25 based XMSG and a flight simulator backbone for the F-16 were delivered. In 1978 Norsk Data both bought Tandberg and launched its office suite Notis, although Tandberg was sold again in 1980. ND became the first foreign-listed Norwegian company in 1981, which also saw the launch of the 32-bit ND-500. Throughout the 1980s ND acquired a series of domestic and foreign hardware and software companies, many loosely oriented at increased hardware sales. At the peak in 1986 and 1987, Norsk Data had 4,500 employees, 2.5 billion Norwegian krone (NOK) in revenue and was Norway's second-largest company by market capitalization—having increased fifty-fold between 1977 and 1985. Despite late attempts to develop Ndix, ND never succeeded at entering the Unix market which started to dominate in the late 1980s. The company's share value halved on 19 Octob
https://en.wikipedia.org/wiki/Danny%20Hillis
William Daniel Hillis (born September 25, 1956) is an American inventor, entrepreneur, and computer scientist, who pioneered parallel computers and their use in artificial intelligence. He founded Thinking Machines Corporation, a parallel supercomputer manufacturer, and subsequently was Vice President of Research and Disney Fellow at Walt Disney Imagineering. Hillis was elected a member of the National Academy of Engineering in 2001 for advances in parallel computers, parallel software, and parallel storage. More recently, Hillis co-founded Applied Minds, and Applied Invention, an interdisciplinary group of engineers, scientists, and artists. He is a visiting professor at the MIT Media Lab. Biography Early life and academic work Born September 25, 1956, in Baltimore, Maryland, Danny Hillis spent much of his childhood living overseas, in Europe, Africa, and Asia. He attended the Massachusetts Institute of Technology (MIT) and received his bachelor of science in mathematics in 1978. As an undergraduate, he worked at the MIT Logo Laboratory under the tutelage of Seymour Papert, developing computer hardware and software for children. During this time, he also designed computer-oriented toys and games for the Milton Bradley Company. While still in college, he co-founded Terrapin Inc., a producer of computer software, including Logo, for elementary schools. As a graduate student at the MIT Computer Science and Artificial Intelligence Laboratory, Hillis designed tendon-controlled robot arms and a touch-sensitive robot "skin". During his college years, Hillis was part of the team that built a computer composed entirely of Tinkertoys, currently at the Computer History Museum in Mountain View, California. At MIT, Hillis began to study Artificial Intelligence under Marvin Minsky. In 1981, he proposed building a massively parallel computer for Artificial Intelligence, consisting of a million processors, each similar to a modern Graphics Processing Unit. This work culmi
https://en.wikipedia.org/wiki/Journal%20of%20Mathematical%20Analysis%20and%20Applications
The Journal of Mathematical Analysis and Applications is an academic journal in mathematics, specializing in mathematical analysis and related topics in applied mathematics. It was founded in 1960, as part of a series of new journals on areas of mathematics published by Academic Press, and is now published by Elsevier. For most years since 1997 it has been ranked by SCImago Journal Rank as among the top 50% of journals in its topic areas.
https://en.wikipedia.org/wiki/Bankruptcy%20problem
A bankruptcy problem, also called a claims problem, is a problem of distributing a homogeneous divisible good (such as money) among people with different claims. The focus is on the case where the amount is insufficient to satisfy all the claims. The canonical application is a bankrupt firm that is to be liquidated. The firm owes different amounts of money to different creditors, but the total worth of the company's assets is smaller than its total debt. The problem is how to divide the scarce existing money among the creditors. Another application would be the division of an estate amongst several heirs, particularly when the estate cannot meet all the deceased's commitments. A third application is tax assessment. One can consider the claimants as taxpayers, the claims as the incomes, and the endowment as the total after-tax income. Determining the allocation of total after-tax income is equivalent to determining the allocation of tax payments. Definitions The amount available to divide is denoted by (=Estate or Endowment). There are n claimants. Each claimant i has a claim denoted by . It is assumed that , that is, the total claims are (weakly) larger than the estate. A division rule is a function that maps a problem instance to a vector such that and for all i. That is: each claimant receives at most its claim, and the sum of allocations is exactly the estate E. Generalizations There are generalized variants in which the total claims might be smaller than the estate. In these generalized variants, is not assumed and is not required. Another generalization, inspired by realistic bankruptcy problems, is to add an exogeneous priority ordering among the claimants, that may be different even for claimants with identical claims. This problem is called a claims problem with priorities. Another variant is called a claims problem with weights. Rules There are various rules for solving bankruptcy problems in practice. The proportional rule divide
https://en.wikipedia.org/wiki/Macrophage%20migration%20inhibitory%20factor%20domain
Macrophage migration inhibitory factor domain is an evolutionary conserved protein domain. Macrophage migration inhibitory factor (MIF) is a key regulatory cytokine within innate and adaptive immune responses, capable of promoting and modulating the magnitude of the response. MIF is released from T-cells and macrophages, and acts within the neuroendocrine system. MIF is capable of tautomerase activity, although its biological function has not been fully characterised. It is induced by glucocorticoid and is capable of overriding the anti-inflammatory actions of glucocorticoid. MIF regulates cytokine secretion and the expression of receptors involved in the immune response. It can be taken up into target cells in order to interact with intracellular signalling molecules, inhibiting p53 function, and/or activating components of the mitogen-activated protein kinase and Jun-activation domain-binding protein-1 (Jab-1). MIF has been linked to various inflammatory diseases, such as rheumatoid arthritis and atherosclerosis. The MIF homologue D-dopachrome tautomerase (EC 4.1.1.84) is involved in detoxification through the conversion of dopaminechrome (and possibly norepinephrinechrome), the toxic quinine product of the neurotransmitter dopamine (and norepinephrine), to an indole derivative that can serve as a precursor to neuromelanin. Examples Human genes encoding proteins that contain this domain include DDT and MIF.
https://en.wikipedia.org/wiki/KQCA
KQCA (channel 58) is a television station licensed to Stockton, California, United States, serving the Sacramento area as a dual affiliate of The CW and MyNetworkTV. It is owned by Hearst Television alongside NBC affiliate KCRA-TV (channel 3). Both stations share studios on Television Circle off D Street in downtown Sacramento, while KQCA's transmitter is located in Walnut Grove, California. History The station first signed on the air on April 13, 1986, as KSCH. The first program to air on the station was a "preview" show hosted by Jim Finnerty and Lori Sequest. It was 51 percent owned by Schuyler Communications, Inc., and 49 percent by the SFN Companies. It originally operated as an independent station and aired classic television series from the 1950s, 1960s, and 1970s, as well as some daytime programs that were preempted by KCRA-TV and KXTV (channel 10). The station originally operated from studios located on West Weber Avenue in Stockton. KSCH was also the first station in the Sacramento market to provide stereo sound from its sign-on. On August 9 of that year, SFN sold the station to Pegasus Broadcasting, which consisted of SFN management and outside investors; channel 58 along with three other television stations and three radio stations sold for $154 million. In 1988, the station moved its studios to a new building located on Gold Canal Drive in Rancho Cordova. In 1990, GE Capital, which had been one of the investors that formed Pegasus, purchased the company outright. In 1993, GE Capital began shopping KSCH-TV for sale; in one potential proposal, both KSCH and Koplar Communications-owned KRBK (channel 31, now KMAX-TV) would have been sold to one buyer, who would have been able to sell off one of the stations to a noncompetitive entity. In 1994, Sacramento restaurant owner Wing Fat and Barbara Scurfield purchased KSCH-TV from GE Capital for $8 million. The new owners entered into a local marketing agreement with Kelly Broadcasting, then-owner of KCRA. KCRA
https://en.wikipedia.org/wiki/Kennel%20cough
Kennel cough (also "canine infectious respiratory disease" (CIRD), "canine infectious respiratory disease complex" (CIRDC) or "canine infectious tracheobronchitis" (CIT)) is an upper respiratory infection affecting dogs. There are multiple causative agents, the most common being the bacterium Bordetella bronchiseptica (found in 78.7% of cases in Southern Germany), followed by canine parainfluenza virus (CPIV; 37.7% of cases), and to a lesser extent canine coronavirus (9.8% of cases). It is highly contagious; however, adult dogs may display immunity to reinfection even under constant exposure. Kennel cough is so named because the infection can spread quickly among dogs in the close quarters of a kennel or animal shelter. Viral and bacterial causes of canine cough are spread through airborne droplets produced by sneezing and coughing. These agents also spread through contact with contaminated surfaces. Symptoms begin after a several day incubation period post-exposure, and in most cases will clear up on their own. However, in young puppies or immunocompromised animals, mixed or secondary infections can progress to lower respiratory infections such as pneumonia. Symptoms The incubation period is 5–7 days (with a range of 3–10). Symptoms can include a harsh, dry cough, retching, sneezing, snorting, gagging or vomiting in response to light pressing of the trachea or after excitement or exercise. The presence of a fever varies from case to case. Types Although kennel cough is considered to be a multifactorial infection, there are two main forms. The first is more mild and is caused by B. bronchiseptica and canine parainfluenza infections, without complications from canine distemper virus (CDV) or canine mastadenovirus A (formerly canine adenovirus-1). This form occurs most regularly in autumn, and can be distinguished by symptoms such as a retching cough and vomiting. The second form has a more complex combination of causative organisms, including CDV and CAV. It typi
https://en.wikipedia.org/wiki/Purple%20Rain%20%28song%29
"Purple Rain" is a song by American musician Prince and his backing band the Revolution. It is the title track from the 1984 album of the same name, which in turn is the soundtrack album for the 1984 film of the same name starring Prince, and was released as the third single from the album. The song is a power ballad that combines rock, R&B, gospel, and orchestral music. "Purple Rain" reached number two on the US Billboard Hot 100 and stayed there for two weeks, being kept off the top spot by "Wake Me Up Before You Go-Go" by Wham!. It reached the summit in Belgium and the Netherlands. It is certified gold by the Recording Industry Association of America (RIAA) and is considered to be one of Prince's signature songs. Following Prince's death in 2016, "Purple Rain" re-entered the Billboard Hot 100, where it reached number four. It also re-entered the UK Singles Chart at number six, placing two spaces higher than its original peak. In France, where it originally peaked at number 12, "Purple Rain" reached number one around a week after Prince's death. "Purple Rain" is ranked at number 18 on Rolling Stone list of the 500 Greatest Songs of All Time and is included in the Rock and Roll Hall of Fame's 500 Songs that Shaped Rock and Roll. During the Super Bowl XLI halftime show in 2007, for which Prince was the featured performer, "Purple Rain" was the last song of his set; the event became especially notable when actual rain fell during the performance while the stage and stadium were lit up with purple lights, and the show continues to top lists of the best Super Bowl halftime shows of all time. Prince performed the song as the opening of a medley of his hits with Beyoncé at the 2004 Grammy Awards. It was also the final song he performed live, taking place at the end of his final performance in Atlanta on April 14, 2016, one week before he died. Composition Origins "Purple Rain" was originally written as a country song and intended to be a collaboration with Stevie Nic
https://en.wikipedia.org/wiki/Cultrex%20BME
Cultrex Basement Membrane Extract (BME) is the trade name for a extracellular protein mixture secreted by Engelbreth-Holm-Swarm (EHS) mouse sarcoma cells and manufactured into a hydrogel by R&D Systems, a brand of Bio-Techne. Similar to Matrigel, this hydrogel is a natural extracellular matrix that mimics the complex extracellular environment within complex tissues. It is used as a general cell culture substrate across a wide variety of research applications. Composition The major components of Cultrex BME include laminin, collagen, entactin/nidogen, and heparin sulfate proteoglycans. These components are found in the basement membrane extracellular matrix that forms an interface between stromal tissue and adjacent endothelial, epithelial, muscle, or neuronal cells. Also present in Cultrex BME are growth factors, however these can be reduced during production of the hydrogel. Reduced growth factor (RGF) Cultrex BME is commonly used for the culture of pluripotent stem cells and organoids. Cell culture Cultrex BME is gelatinous at 4 °C. The matrix proteins polymerize and solidify at temperatures above 18 °C. Ice-cold basement membrane extract can be dispensed directly onto plastic cell culture labware or it can be diluted in ice-cold phosphate buffered saline or cell culture media prior to dispensing. Due to its heterogenous extracellular matrix protein composition, cells cultured using basement membrane extract show complex cellular behaviors that are difficult to reproduce under laboratory conditions. These include cell adhesion, migration, invasion, proliferation, differentiation, and endothelial cell tube formation. EHS tumor-derived matrices, such as Cultrex BME, are used for a variety of cell culture applications, including angiogenesis, spheroid formation, organoid culture, pluripotent stem cell culture, xenograft, and in vivo and in vitro tumor modeling Pluripotent stem cells Cultrex BME is used as an attachment substrate to promote the expansion and ma
https://en.wikipedia.org/wiki/QA%20%26%20UX%20Manager
A QA & UX Manager work with both Quality Assurance (QA) and User Experience (UX) in relation to video game and software development. QA & UX Manager can work independently or in co-operation with other QA & UX Managers, like in larger QA & UX teams with other QA & UX Managers and game testers. In the larger QA & UX teams, there is usually a lead QA & UX Group Manager that works as the daily leader of this team of QA & UX Managers and game testers. QA & UX Manager also usually work very close with the project managers and the QA Programmer as part of video game development. QA & UX Managers work and tasks In terms of work assignments that QA & UX Managers do, this can include planning and management, technical testing, User Experience, metrics and communication. In terms of planning and management done by the QA & UX Manager, this may include making test plans, test cycles and test cases for a video game as well as being responsible for the overall testing and recruiting of game testers in a video game development. In terms of technical testing that the QA & UX Managers do, this can be very alternating and can be everything from testing correctness and regression to check an optimization in a video game. In terms of User Experience that the QA & UX Manager work with, this is particularly centered around making interviews and doing monitoring in relation to User Experience of a video game by the player. In terms of Metrics that the QA & UX Manager work with, this task mainly is about collecting and analyze data an about game and gamer activity in relation to a particular game. In terms of the communication characterized parts of the work the QA & UX Manager do, this is often done in relation with meetings with the development team and where the QA & UX Manager pass on his or her knowledge and data to the development team.
https://en.wikipedia.org/wiki/66th%20meridian%20west
The meridian 66° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, North America, the Atlantic Ocean, the Caribbean Sea, South America, the Southern Ocean, and Antarctica to the South Pole. The 66th meridian west forms a great circle with the 114th meridian east. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 66th meridian west passes through: {| class="wikitable plainrowheaders" ! scope="col" width="120" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Lincoln Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Ellesmere Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Nares Strait | style="background:#b0e0e6;" | |- | ! scope="row" | |Washington Land |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Nares Strait | style="background:#b0e0e6;" | Kane Basin |- | ! scope="row" | |Inglefield Land |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Baffin Bay | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Davis Strait | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Cumberland Peninsula, Baffin Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Davis Strait | style="background:#b0e0e6;" | Cumberland Sound |- | ! scope="row" | | Nunavut — Hall Peninsula, Baffin Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Frobisher Bay | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Meta Incognita Peninsula, Baffin Island |- | style="background:#b0e0e6;" | ! scope="row" st
https://en.wikipedia.org/wiki/Endoscopic%20group
In mathematics, endoscopic groups of reductive algebraic groups were introduced by in his work on the stable trace formula. Roughly speaking, an endoscopic group H of G is a quasi-split group whose L-group is the connected component of the centralizer of a semisimple element of the L-group of G. In the stable trace formula, unstable orbital integrals on a group G correspond to stable orbital integrals on its endoscopic groups H. The relation between them is given by the fundamental lemma.
https://en.wikipedia.org/wiki/Hassler%20Whitney
Hassler Whitney (March 23, 1907 – May 10, 1989) was an American mathematician. He was one of the founders of singularity theory, and did foundational work in manifolds, embeddings, immersions, characteristic classes, and geometric integration theory. Biography Life Hassler Whitney was born on March 23, 1907, in New York City, where his father, Edward Baldwin Whitney, was the First District New York Supreme Court judge. His mother, A. Josepha Newcomb Whitney, was an artist and political activist. He was the paternal nephew of Connecticut Governor and Chief Justice Simeon Eben Baldwin, his paternal grandfather was William Dwight Whitney, professor of Ancient Languages at Yale University, linguist and Sanskrit scholar. Whitney was the great-grandson of Connecticut Governor and US Senator Roger Sherman Baldwin, and the great-great-grandson of American founding father Roger Sherman. His maternal grandparents were astronomer and mathematician Simon Newcomb (1835-1909), a Steeves descendant, and Mary Hassler Newcomb, granddaughter of the first superintendent of the Coast Survey Ferdinand Rudolph Hassler. His great uncle Josiah Whitney was the first to survey Mount Whitney. He married three times: his first wife was Margaret R. Howell, married on the 30 May 1930. They had three children, James Newcomb, Carol and Marian. After his first divorce, on January 16, 1955 he married Mary Barnett Garfield. He and Mary had two daughters, Sarah Newcomb and Emily Baldwin. Finally, Whitney divorced his second wife and married Barbara Floyd Osterman on 8 February 1986. Whitney and his first wife Margaret made an innovative decision in 1939 that influenced the history of modern architecture in New England, when they commissioned the architect Edwin B. Goodell, Jr. to design a new residence for their family in Weston, Massachusetts. They purchased a rocky hillside site on a historic road, next door to another International Style house by Goodell from several years earlier, designed for
https://en.wikipedia.org/wiki/Answer%20to%20reset
An Answer To Reset (ATR) is a message output by a contact Smart Card conforming to ISO/IEC 7816 standards, following electrical reset of the card's chip by a card reader. The ATR conveys information about the communication parameters proposed by the card, and the card's nature and state. By extension, ATR often refers to a message obtained from a Smart Card in an early communication stage; or from the card reader used to access that card, which may transform the card's message into an ATR-like format (this occurs e.g. for some PC/SC card readers when accessing an ISO/IEC 14443 Smart Card). The presence of an ATR is often used as a first indication that a Smart Card appears operative, and its content examined as a first test that it is of the appropriate kind for a given usage. Contact Smart Cards communicate over a signal named Input/Output (I/O) either synchronously (data bits are sent and received at the rhythm of one per period of the clock supplied to the card on its CLK signal) or asynchronously (data bits are exchanged over I/O with another mechanism for bit delimitation, similar to traditional asynchronous serial communication). The two modes are exclusive in a given communication session, and most cards are built with support for a single mode. Microprocessor-based contact Smart Cards are mostly of the asynchronous variety, used for all Subscriber Identity Modules (SIM) for mobile phones, those bank cards with contacts that conform to EMV specifications, all contact Java Cards, and Smart Cards for pay television. Memory-only cards are generally of the synchronous variety. ATR under asynchronous and synchronous transmission have entirely different form and content. The ATR in asynchronous transmission is precisely normalized (in order to allow interoperability between cards and readers of different origin), and relatively complex to parse. Some Smart Cards (mostly of the asynchronous variety) send different ATR depending on if the reset is the first sinc
https://en.wikipedia.org/wiki/Envy%20ratio
Envy ratio, in finance, is the ratio of the price paid by investors to that paid by the management team for their respective shares of the equity. It is used to consider an opportunity for a management buyout. Managers are often allowed to invest at a lower valuation to make their ownership possible and to create a personal financial incentive for them to approve the buyout and to work diligently towards the success of the investment. The envy ratio is somewhat similar to the concept of financial leverage; managers can increase returns on their investments by using other investors' money. Basic formula Source Example If private equity investors paid $500M for 80% of a company's equity, and a management team paid $60M for 20%, then ER=(500/0,8)/(60/0,2)=2.08x. This means that the investors paid for a share 2.08 times more than did the managers. The ratio demonstrates how generous institutional investors are to a management team—the higher the ratio is, the better is the deal for management. As a rule of thumb, management should be expected to invest anywhere from six months to one year’s gross salary to demonstrate commitment and have some personal financial risk. In any transaction, the envy ratio is affected by how keen the investors are to do the deal; the competition they are facing; and economic factors. See also Management buy-in LBO Takeover Financial ratio
https://en.wikipedia.org/wiki/Sensor%20fish
A sensor fish is a small, plastic tubular device containing sensors. It is designed to record information such as the physical stresses that a fish experiences while navigating currents from dam turbines. Description Created by the US Department of Energy's Pacific Northwest National Laboratory (PNNL), the tubular device is long, in diameter, and weighs . It is roughly the same size as a juvenile salmon. The sensor fish has neutral buoyancy allowing it to remain underwater. Inside are sensors and a lithium-ion battery. Taking 2,048 measurements each second, it is able to record five minutes of turbulence, pressure, and acceleration, saving the data to flash memory. It records a maximum of 1.2 MPa (174 pounds per square inch) of external pressure, up to 200 gs of acceleration, temperatures ranging from -40 to 127 °C (-40 and +260 degrees F), and rotational velocity of up to 2,000 degrees per second. Construction The sensor fish is built manually at the Bio-Acoustics & Flow Laboratory within the PNNL. It receives funding from the Electric Power Research Institute and the US Department of Energy's Office of Energy Efficiency and Renewable Energy. Specifications The following are the specifications as stated by the Pacific Northwest National Laboratory: Power: rechargeable 3.7-volt lithium-ion battery Length: ~90 mm Diameter: ~25 mm Mass: ~42.1g Cost: $1,200 each Gyroscope: Model ITG-3200, InvenSense Inc. Orientation: Model LSM303DLHC eCompass module made by STMicroelectronics, Geneva, Switzerland Usage: Kaplan turbine; Francis turbine; small hydropower structures; pumped storage hydroelectric facilities Memory: ~5 minutes of data with flash memory Speed: 2,048 measurements per second Pressure: Model MS5412-BM, Measurement Specialties, Inc., made in Hampton, Virginia capable of 174 pounds per square inch of pressure Acceleration: Model ADXL377 accelerometer, Analog Devices, Inc. Rotational velocity: 2,000 degrees per second Temperature range: -40 to +125 degrees
https://en.wikipedia.org/wiki/CUT%26RUN%20sequencing
CUT&RUN sequencing, also known as cleavage under targets and release using nuclease, is a method used to analyze protein interactions with DNA. CUT&RUN sequencing combines antibody-targeted controlled cleavage by micrococcal nuclease with massively parallel DNA sequencing to identify the binding sites of DNA-associated proteins. It can be used to map global DNA binding sites precisely for any protein of interest. Currently, ChIP-Seq is the most common technique utilized to study protein–DNA relations, however, it suffers from a number of practical and economical limitations that CUT&RUN sequencing does not. Uses CUT&RUN sequencing can be used to examine gene regulation or to analyze transcription factor and other chromatin-associated protein binding. Protein-DNA interactions regulate gene expression and are responsible for many biological processes and disease states. This epigenetic information is complementary to genotype and expression analysis. CUT&RUN is an alternative to the current standard of ChIP-seq. ChIP-Seq suffers from limitations due to the cross linking step in ChIP-Seq protocols that can promote epitope masking and generate false-positive binding sites. As well, ChIP-seq suffers from suboptimal signal-to-noise ratios and poor resolution. CUT&RUN sequencing has the advantage of being a simpler technique with lower costs due to the high signal-to-noise ratio, requiring less depth in sequencing. Specific DNA sites in direct physical interaction with proteins such as transcription factors can be isolated by Protein-A (pA) conjugated micrococcal nuclease (MNase) bound to a protein of interest. MNase mediated cleavage produces a library of target DNA sites bound to a protein of interest in situ. Sequencing of prepared DNA libraries and comparison to whole-genome sequence databases allows researchers to analyze the interactions between target proteins and DNA, as well as differences in epigenetic chromatin modifications. Therefore, the CUT&RUN method may
https://en.wikipedia.org/wiki/Electric%20discharge
In electromagnetism, an electric discharge is the release and transmission of electricity in an applied electric field through a medium such as a gas (ie., an outgoing flow of electric current through a non-metal medium). Applications The properties and effects of electric discharges are useful over a wide range of magnitudes. Tiny pulses of current are used to detect ionizing radiation in a Geiger–Müller tube. A low steady current can be used to illustrate the spectrum of gases in a gas-filled tube. A neon lamp is an example of a gas-discharge lamp, useful both for illumination and as a voltage regulator. A flashtube generates a short pulse of intense light useful for photography by sending a heavy current through a gas arc discharge. Corona discharges are used in photocopiers. Electric discharges can convey substantial energy to the electrodes at the ends of the discharge. A spark gap is used in internal combustion engines to ignite the fuel/air mixture on every power stroke. Spark gaps are also used to switch heavy currents in a Marx generator and to protect electrical apparatus. In electric discharge machining, multiple tiny electric arcs are used to erode a conductive workpiece to a finished shape. Arc welding is used to assemble heavy steel structures, where the base metal is heated to melting by the heat of the arc. An electric arc furnace sustains arc currents of tens of thousands of amperes and is used for steelmaking and production of alloys and other products. Examples Examples of electric discharge phenomena include: Brush discharge Dielectric barrier discharge Corona discharge Electric glow discharge Electric arc Electrostatic discharge Electric discharge in gases Leader (spark) Partial discharge Streamer discharge Vacuum arc Townsend discharge St. Elmo's fire Lightning Electric organ See also Electrical breakdown Electric discharge in gases Lichtenberg figure Space charge Debye sheath
https://en.wikipedia.org/wiki/Satellite%20%28biology%29
A satellite is a subviral agent that depends on the coinfection of a host cell with a helper virus for its replication. Satellites can be divided into two major classes: satellite viruses and satellite nucleic acids. Satellite viruses, which are most commonly associated with plants, are also found in mammals, arthropods, and bacteria. They encode structural proteins to enclose their genetic material, which are therefore distinct from the structural proteins of their helper viruses. Satellite nucleic acids, in contrast, do not encode their own structural proteins, but instead are encapsulated by proteins encoded by their helper viruses. The genomes of satellites range upward from 359 nucleotides in length for satellite tobacco ringspot virus RNA (STobRV). Most viruses have the capability to use host enzymes or their own replication machinery to independently replicate their own viral RNA. Satellites, in contrast, are completely dependent on a helper virus for replication. The symbiotic relationship between a satellite and a helper virus to catalyze the replication of a satellite genome is also dependent on the host to provide components like replicases to carry out replication. A satellite virus of mamavirus that inhibits the replication of its host has been termed a virophage. However, the usage of this term remains controversial due to the lack of fundamental differences between virophages and classical satellite viruses. History and discovery The tobacco necrosis virus was the virus that led to the discovery of the first satellite virus in 1962. Scientists discovered that the first satellite had the components to make its own protein shell. A few years later in 1969, scientists discovered another symbiotic relationship with the tobacco ringspot nepovirus (TobRV) and another satellite virus. The emergence of satellite RNA is said to have come from either the genome of the host or its co-infecting agents, and any vectors leading to transmission. A satellite vi
https://en.wikipedia.org/wiki/Stolarsky%20mean
In mathematics, the Stolarsky mean is a generalization of the logarithmic mean. It was introduced by Kenneth B. Stolarsky in 1975. Definition For two positive real numbers x, y the Stolarsky Mean is defined as: Derivation It is derived from the mean value theorem, which states that a secant line, cutting the graph of a differentiable function at and , has the same slope as a line tangent to the graph at some point in the interval . The Stolarsky mean is obtained by when choosing . Special cases is the minimum. is the geometric mean. is the logarithmic mean. It can be obtained from the mean value theorem by choosing . is the power mean with exponent . is the identric mean. It can be obtained from the mean value theorem by choosing . is the arithmetic mean. is a connection to the quadratic mean and the geometric mean. is the maximum. Generalizations One can generalize the mean to n + 1 variables by considering the mean value theorem for divided differences for the nth derivative. One obtains for . See also Mean
https://en.wikipedia.org/wiki/Lysis
Lysis ( ) is the breaking down of the membrane of a cell, often by viral, enzymic, or osmotic (that is, "lytic" ) mechanisms that compromise its integrity. A fluid containing the contents of lysed cells is called a lysate. In molecular biology, biochemistry, and cell biology laboratories, cell cultures may be subjected to lysis in the process of purifying their components, as in protein purification, DNA extraction, RNA extraction, or in purifying organelles. Many species of bacteria are subject to lysis by the enzyme lysozyme, found in animal saliva, egg white, and other secretions. Phage lytic enzymes (lysins) produced during bacteriophage infection are responsible for the ability of these viruses to lyse bacterial cells. Penicillin and related β-lactam antibiotics cause the death of bacteria through enzyme-mediated lysis that occurs after the drug causes the bacterium to form a defective cell wall. If the cell wall is completely lost and the penicillin was used on gram-positive bacteria, then the bacterium is referred to as a protoplast, but if penicillin was used on gram-negative bacteria, then it is called a spheroplast. Cytolysis Cytolysis occurs when a cell bursts due to an osmotic imbalance that has caused excess water to move into the cell. Cytolysis can be prevented by several different mechanisms, including the contractile vacuole that exists in some paramecia, which rapidly pump water out of the cell. Cytolysis does not occur under normal conditions in plant cells because plant cells have a strong cell wall that contains the osmotic pressure, or turgor pressure, that would otherwise cause cytolysis to occur. Oncolysis Oncolysis is the destruction of neoplastic cells or of a tumour. The term is also used to refer to the reduction of any swelling. Plasmolysis Plasmolysis is the contraction of cells within plants due to the loss of water through osmosis. In a hypertonic environment, the cell membrane peels off of the cell wall and the vacuole co
https://en.wikipedia.org/wiki/Federation%20of%20American%20Societies%20for%20Experimental%20Biology
The Federation of American Societies for Experimental Biology (FASEB), based in Rockville, Maryland, is a non-profit organization of scientific societies in the United States. With a focus on the biological and biomedical sciences, the federation represents scientists in such fields as anatomy, physiology, immunology, biochemistry, molecular biology, toxicology, genetics, and nutrition. Description Founded in 1912, FASEB was established to provide a forum for biological and biomedical researchers enabling them to hold educational meetings, develop publications, and disseminate biological research results. FASEB currently comprises 27 scientific societies. As a federation, FASEB's collective mission is to advance health and well-being by promoting research and education in the biological and biomedical sciences. To advance its mission, FASEB provides a variety of programs and services to its member societies to support their individual members, including advocating for federal funding for their research, hosting scientific conferences and meetings to share their research findings, publishing scientific journals to promote their research, working to create a diverse and representative workforce in the biological and biomedical sciences, and celebrating researchers’ efforts to advance the biological and biomedical sciences through an awards program. Members Membership in FASEB is limited to scientific societies. Societies are either Full or Associate members. Each member society shares a common vision for the advancement of research and education in the biological and biomedical sciences. Full Members American Aging Association American Association for Anatomy American Association of Immunologists American College of Sports Medicine American Federation for Medical Research American Physiological Society American Society for Biochemistry and Molecular Biology American Society for Bone and Mineral Research American Society for Clinical Investigation American Soci
https://en.wikipedia.org/wiki/Peer-to-peer%20SIP
Peer-to-peer SIP (P2P-SIP) is an implementation of a distributed voice over Internet Protocol (VoIP) or instant messaging communications application using a peer-to-peer (P2P) architecture in which session control between communication end points is facilitated with the Session Initiation Protocol (SIP). SIP in a P2P architecture In a pure peer-to-peer application architecture no central servers are required, whereas traditional SIP telephony networks have relied on using centrally deployed and managed SIP servers, in analogy to the centralized switching architecture of the public switched telephone network (PSTN). P2P application design can improve scalability and survivability in the event of central network outages. The Session Initiation Protocol is in principle a client-server protocol, however it has been described in analogy to the P2P relationship, called a dialog, which is defined by a unique combination of SIP protocol parameters (To-tag, From-tag, Call-ID). Both endpoints of a communication session implement a user-agent server and a user-agent client, which enables any two user agents to communicate directly with one another without the mediation of another a central switching system. SIP also provides facilities of registering (REGISTER request) the network location of a user agent with other SIP elements and subscription (SUBSCRIBE request) and notification (NOTIFY request) features for event tracking between user agents. Based on these inherent SIP features it is possible to construct a peer-to-peer network of SIP nodes. In another approach, SIP over P2P, the SIP location service is replaced by conventional P2P overlay networking approach, such as the OverSim framework. In this model, the overlay network is used for service or node discovery and rendezvous. The search key in such a mechanism is the Uniform Resource Identifier (URI) of a user agent. This URI requires resolution to a particular device or Uniform Resource Locator (URL) that must be
https://en.wikipedia.org/wiki/Nicod%27s%20axiom
In logic, Nicod's axiom (named after the French logician and philosopher Jean Nicod) is a formula that can be used as the sole axiom of a semantically complete system of propositional calculus. The only connective used in the formulation of Nicod's axiom is the Sheffer's stroke. The axiom has the following form: ((φ | (χ | ψ)) | ((τ | (τ | τ)) | ((θ | χ) | ((φ | θ) | (φ | θ))))) Nicod showed that the whole propositional logic of Principia Mathematica could be derived from this axiom alone by using one inference rule, called "Nicod's modus ponens": 1. φ 2. (φ | (χ | ψ)) ∴ ψ In 1931, the Polish logician Mordechaj Wajsberg discovered an equally powerful and easier-to-work-with alternative: ((φ | (ψ | χ)) | (((τ | χ) | ((φ | τ) | (φ | τ))) | (φ | (φ | ψ))))
https://en.wikipedia.org/wiki/A%20History%20of%20the%20Kerala%20School%20of%20Hindu%20Astronomy
A History of the Kerala School of Hindu Astronomy (in perspective) is the first definitive book giving a comprehensive description of the contribution of Kerala to astronomy and mathematics. The book was authored by K. V. Sarma who was a Reader in Sanskrit at Vishveshvaranand Institute of Sanskrit and Indological Studies, Panjab University, Hoshiarpur, at the time of publication of the book (1972). The book, among other things, contains details of the lives and works of about 80 astronomers and mathematicians belonging to the Kerala School. It has also identified 752 works belonging to the Kerala school. Even though C. M. Whish, an officer of East India Company, had presented a paper on the achievements of the mathematicians of Kerala School as early as 1842, western scholars had hardly taken note of these contributions. Much later in the 1940s, C. T. Rajagopal and his associates made some efforts to study and popularize the discoveries of Whish. Their work was lying scattered in several journals and as parts of books. Even after these efforts by C. T. Rajagopal and others, the view that Bhaskara II was the last significant mathematician pre-modern India had produced had prevailed among scholars, and surprisingly, even among Indian scholars. It was in this context K. V. Sarma published his book as an attempt to present in a succinct form the results of the investigations of C. T. Rajagopal and others and also the findings of his own investigations into the history of the Kerala school of astronomy and mathematics. Summary of the book The book is divided into six chapters. Chapter 1 gives an outline of the salient features of Kerala astronomy. Sarma emphasizes the spirit of inquiry, stress on observation and experimentation, concern for accuracy, and continuity of tradition as the important features of Kerala astronomy. Adherence to the Aryabhatan system, use of the katapayadi system for expressing numbers, the use of the Parahita and Drik systems for astronomica
https://en.wikipedia.org/wiki/Modern%20synthesis%20%2820th%20century%29
The modern synthesis was the early 20th-century synthesis of Charles Darwin's theory of evolution and Gregor Mendel's ideas on heredity into a joint mathematical framework. Julian Huxley coined the term in his 1942 book, Evolution: The Modern Synthesis. The synthesis combined the ideas of natural selection, Mendelian genetics, and population genetics. It also related the broad-scale macroevolution seen by palaeontologists to the small-scale microevolution of local populations. The synthesis was defined differently by its founders, with Ernst Mayr in 1959, G. Ledyard Stebbins in 1966, and Theodosius Dobzhansky in 1974 offering differing basic postulates, though they all include natural selection, working on heritable variation supplied by mutation. Other major figures in the synthesis included E. B. Ford, Bernhard Rensch, Ivan Schmalhausen, and George Gaylord Simpson. An early event in the modern synthesis was R. A. Fisher's 1918 paper on mathematical population genetics, though William Bateson, and separately Udny Yule, had already started to show how Mendelian genetics could work in evolution in 1902. Different syntheses followed, including with social behaviour in E. O. Wilson's sociobiology in 1975, evolutionary developmental biology's integration of embryology with genetics and evolution, starting in 1977, and Massimo Pigliucci's and Gerd B. Müller's proposed extended evolutionary synthesis of 2007. In the view of evolutionary biologist Eugene Koonin in 2009, the modern synthesis will be replaced by a 'post-modern' synthesis that will include revolutionary changes in molecular biology, the study of prokaryotes and the resulting tree of life, and genomics. Developments leading up to the synthesis Darwin's evolution by natural selection, 1859 Charles Darwin's 1859 book, On the Origin of Species, convinced most biologists that evolution had occurred, but not that natural selection was its primary mechanism. In the 19th and early 20th centuries, variations of L
https://en.wikipedia.org/wiki/Virt-manager
virt-manager is a desktop virtual machine monitor primarily developed by Red Hat. Features Virtual Machine Manager allows users to: create, edit, start and stop VMs view and control each VM's console see performance and utilization statistics for each VM view all running VMs and hosts, and their live performance or resource utilization statistics. use KVM, Xen or QEMU virtual machines, running either locally or remotely. use LXC containers Support for FreeBSD's bhyve hypervisor has been included since 2014, though it remains disabled by default. Distributions including Virtual Machine Manager Virtual Machine Manager comes as the package in: Arch Linux CentOS Debian (since lenny) Fedora (since version 6) FreeBSD (via Ports collection) Frugalware Gentoo Mandriva Linux (since release 2007.1) MXLinux NetBSD (via pkgsrc) NixOS OpenBSD (via Ports collection) openSUSE (since release 10.3) Red Hat Enterprise Linux (versions 5 through 7 only) Scientific Linux Trisquel TrueOS Ubuntu (version 8.04 and above) Void Linux See also libvirt, the API used by Virtual Machine Manager to create and manage virtual machines
https://en.wikipedia.org/wiki/List%20of%20unusual%20units%20of%20measurement
An unusual unit of measurement is a unit of measurement that does not form part of a coherent system of measurement, especially because its exact quantity may not be well known or because it may be an inconvenient multiple or fraction of a base unit. Many of the unusual units of measurements listed here are colloquial measurements, units devised to compare a measurement to common and familiar objects. Length Hammer unit Valve's Source game engine uses the Hammer unit as its base unit of length. This unit refers to Source's official map creation software, Hammer. The exact definition varies from game to game, but a Hammer unit is usually defined as a sixteenth of a foot (16 Hammer units = 1 foot). This means that 1 Hammer unit is equal to exactly . Rack unit One rack unit (U) is and is used to measure rack-mountable audiovisual, computing and industrial equipment. Rack units are typically denoted without a space between the number of units and the 'U'. Thus, a 4U server enclosure (case) is high, or more practically, built to occupy a vertical space seven inches high, with sufficient clearance to allow movement of adjacent hardware. Hand The hand is a non-SI unit of length equal to exactly . It is normally used to measure the height of horses in some English-speaking countries, including Australia, Canada, Ireland, the United Kingdom, and the United States. It is customary when measuring in hands to use a point to indicate inches (quarter-hands) and not tenths of a hand. For example, 15.1 hands normally means 15 hands, 1 inch (5 ft 1 in), rather than 15 hands. Light-nanosecond The light-nanosecond is defined as exactly 29.9792458 cm. It was popularized in information technology as a unit of distance by Grace Hopper as the distance which a photon could travel in one billionth of a second (roughly 30 cm or one foot): "The speed of light is one foot per nanosecond." Metric feet A metric foot, defined as ), has been used occasionally in the UK but has never b
https://en.wikipedia.org/wiki/Locality-sensitive%20hashing
In computer science, locality-sensitive hashing (LSH) is a fuzzy hashing technique that hashes similar input items into the same "buckets" with high probability. (The number of buckets is much smaller than the universe of possible input items.) Since similar items end up in the same buckets, this technique can be used for data clustering and nearest neighbor search. It differs from conventional hashing techniques in that hash collisions are maximized, not minimized. Alternatively, the technique can be seen as a way to reduce the dimensionality of high-dimensional data; high-dimensional input items can be reduced to low-dimensional versions while preserving relative distances between items. Hashing-based approximate nearest-neighbor search algorithms generally use one of two main categories of hashing methods: either data-independent methods, such as locality-sensitive hashing (LSH); or data-dependent methods, such as locality-preserving hashing (LPH). Locality-preserving hashing was initially devised as a way to facilitate data pipelining in implementations of massively parallel algorithms that use randomized routing and universal hashing to reduce memory contention and network congestion. Definitions An LSH family is defined for a metric space , a threshold , an approximation factor , and probabilities and . This family is a set of functions that map elements of the metric space to buckets . An LSH family must satisfy the following conditions for any two points and any hash function chosen uniformly at random from : if , then (i.e., and collide) with probability at least , if , then with probability at most . A family is interesting when . Such a family is called -sensitive. Alternatively it is defined with respect to a universe of items that have a similarity function . An LSH scheme is a family of hash functions coupled with a probability distribution over the functions such that a function chosen according to satisfies the prope
https://en.wikipedia.org/wiki/Potassium%20bromide
Potassium bromide (KBr) is a salt, widely used as an anticonvulsant and a sedative in the late 19th and early 20th centuries, with over-the-counter use extending to 1975 in the US. Its action is due to the bromide ion (sodium bromide is equally effective). Potassium bromide is used as a veterinary drug, as an antiepileptic medication for dogs. Under standard conditions, potassium bromide is a white crystalline powder. It is freely soluble in water; it is not soluble in acetonitrile. In a dilute aqueous solution, potassium bromide tastes sweet, at higher concentrations it tastes bitter, and tastes salty when the concentration is even higher. These effects are mainly due to the properties of the potassium ion—sodium bromide tastes salty at any concentration. In high concentration, potassium bromide strongly irritates the gastric mucous membrane, causing nausea and sometimes vomiting (a typical effect of all soluble potassium salts). Chemical properties Potassium bromide, a typical ionic salt, is fully dissociated and near pH 7 in aqueous solution. It serves as a source of bromide ions. This reaction is important for the manufacture of silver bromide for photographic film: KBr_{(aq)}{} + AgNO3_{(aq)} -> AgBr_{(s)}{} + KNO3_{(aq)} Aqueous bromide also forms complexes when reacted with some metal halides such as copper(II) bromide: 2 KBr_{(aq)}{} + CuBr2_{(aq)} -> K2[CuBr3]_{(aq)} Preparation A traditional method for the manufacture of KBr is the reaction of potassium carbonate with an iron(III, II) bromide, , made by treating scrap iron under water with excess bromine: 4 K2CO3 + Fe3Br8 -> 8 KBr + Fe3O4 + 4 CO2 Applications Medical and veterinary The anticonvulsant properties of potassium bromide were first noted by Sir Charles Locock at a meeting of the Royal Medical and Chirurgical Society in 1857. Bromide can be regarded as the first effective medication for epilepsy. At the time, it was commonly thought that epilepsy was caused by masturbation. Locock note
https://en.wikipedia.org/wiki/List%20of%20the%20most%20common%20passwords
This is a list of the most common passwords, discovered in various data breaches. Common passwords generally are not recommended on account of low password strength. List NordPass NordPass conducted the most breached passwords research in 2021. The company gathered top 200 worst passwords this year from a database of 275,699,516 passwords. SplashData The Worst Passwords List is an annual list of the 25 most common passwords from each year as produced by internet security firm SplashData. Since 2011, the firm has published the list based on data examined from millions of passwords leaked in data breaches, mostly in North America and Western Europe, over each year. In the 2016 edition, the 25 most common passwords made up more than 10% of the surveyed passwords, with the most common password of 2016, "123456", making up 4%. Keeper Password manager Keeper compiled its own list of the 25 most common passwords in 2016, from 25 million passwords leaked in data breaches that year. National Cyber Security Centre The National Cyber Security Centre (NCSC) compiled its own list of the 20 most common passwords in 2019, from 100 million passwords leaked in data breaches that year. See also Password cracking 10,000 most common passwords Notes
https://en.wikipedia.org/wiki/Quantasome
Quantasomes are particles found in the thylakoid membrane of chloroplasts in which photosynthesis takes place. They are embedded in a paracrystalline array on the surface of thylakoid discs in chloroplasts. They are composed of lipids and proteins that include various photosynthetic pigments and redox carriers. For this reason they are considered to be photosynthetic units. They occur in 2 sizes: the smaller quantasome is thought to represent the site of photosystem I, the larger to represent the site of photosystem II. See also Light-dependent reactions Photophosphorylation Photosynthetic reaction centre Photosystem II Thylakoid
https://en.wikipedia.org/wiki/Sigmoid%20sulcus
The inner surface of the mastoid portion of the temporal bone presents a deep, curved groove, the sigmoid sulcus, which lodges part of the transverse sinus; in it may be seen the opening of the mastoid foramen.
https://en.wikipedia.org/wiki/Magnetomotive%20force
In physics, the magnetomotive force (abbreviated mmf or MMF, symbol ) is a quantity appearing in the equation for the magnetic flux in a magnetic circuit, Hopkinson's law. It is the property of certain substances or phenomena that give rise to magnetic fields: where is the magnetic flux and is the reluctance of the circuit. It can be seen that the magnetomotive force plays a role in this equation analogous to the voltage in Ohm's law, , since it is the cause of magnetic flux in a magnetic circuit: where is the number of turns in the coil and is the electric current through the circuit. where is the magnetic flux and is the magnetic reluctance where is the magnetizing force (the strength of the magnetizing field) and is the mean length of a solenoid or the circumference of a toroid. Units The SI unit of mmf is the ampere, the same as the unit of current (analogously the units of emf and voltage are both the volt). Informally, and frequently, this unit is stated as the ampere-turn to avoid confusion with current. This was the unit name in the MKS system. Occasionally, the cgs system unit of the gilbert may also be encountered. History The term magnetomotive force was coined by Henry Augustus Rowland in 1880. Rowland intended this to indicate a direct analogy with electromotive force. The idea of a magnetic analogy to electromotive force can be found much earlier in the work of Michael Faraday (1791–1867) and it is hinted at by James Clerk Maxwell (1831–1879). However, Rowland coined the term and was the first to make explicit an Ohm's law for magnetic circuits in 1873. Ohm's law for magnetic circuits is sometimes referred to as Hopkinson's law rather than Rowland's law as some authors attribute the law to John Hopkinson instead of Rowland. According to a review of magnetic circuit analysis methods this is an incorrect attribution originating from an 1885 paper by Hopkinson. Furthermore, Hopkinson actually cites Rowland's 1873 paper in th
https://en.wikipedia.org/wiki/Fragilariforma
Fragilariforma is a genus of diatoms belonging to the family Fragilariaceae. The genus was first described by D. M. Williams and Round in 1988. The genus has cosmopolitan distribution. Species: Fragilariforma bicapitata Fragilariforma neoproducta Fragilariforma virescens Fragilariforma virescens
https://en.wikipedia.org/wiki/Communicating%20finite-state%20machine
In computer science, a communicating finite-state machine is a finite state machine labeled with "receive" and "send" operations over some alphabet of channels. They were introduced by Brand and Zafiropulo, and can be used as a model of concurrent processes like Petri nets. Communicating finite state machines are used frequently for modeling a communication protocol since they make it possible to detect major protocol design errors, including boundedness, deadlocks, and unspecified receptions. The advantage of communicating finite state machines is that they make it possible to decide many properties in communication protocols, beyond the level of just detecting such properties. This advantage rules out the need for human assistance or restriction in generality. Communicating finite state machines can be more powerful than finite state machines in situations where the propagation delay is not negligible (so that several messages can be in transit at one time) and in situations where it is natural to describe the protocol parties and the communication medium as separate entities. Communicating Hierarchical State Machine Hierarchical state machines are finite state machines whose states themselves can be other machines. Since a communicating finite state machine is characterized by concurrency, the most notable trait in a communicating hierarchical state machine is the coexistence of hierarchy and concurrency. This has been considered highly suitable as it signifies stronger interaction inside the machine. However, it was proved that the coexistence of hierarchy and concurrency intrinsically costs language inclusion, language equivalence, and all of universality. Definition Protocol For an arbitrary positive integer , a protocol with process(es) is a quadruple with: , a sequence of disjoint finite sets. Each set is used to represent a process, and each element of represents a possible state of the -th process. (with ), a sequence representing the initi
https://en.wikipedia.org/wiki/Femoral%20canal
The femoral canal is the medial (and smallest) compartment of the three compartments of the femoral sheath. It is conical in shape. The femoral canal contains lymphatic vessels, and adipose and loose connective tissue, as well as - sometimes - a deep inguinal lymph node. The function of the femoral canal is to accommodate the distension of the femoral vein when venous return from the leg is increased or temporarily restricted (e.g. during a Valsalva maneuver). The proximal, abdominal end of the femoral canal forms the femoral ring. The femoral canal should not be confused with the nearby adductor canal. Anatomy The femoral canal is bordered: anterosuperiorly by the inguinal ligament posteriorly by the pectineal ligament lying anterior to the superior pubic ramus Medially by the lacunar ligament Laterally by the femoral vein Physiological significance The position of the femoral canal medially to the femoral vein is of physiologic importance. The space of the canal allows for the expansion of the femoral vein when venous return from the lower limbs is increased or when increased intra-abdominal pressure (valsalva maneuver) causes a temporary stasis in the venous flow. Clinical significance The entrance to the femoral canal is the femoral ring, through which bowel can sometimes enter, causing a femoral hernia. Though femoral hernias are rare, their passage through the inflexible femoral ring puts them at particular risk of strangulation, giving them surgical priority. See also Femoral ring
https://en.wikipedia.org/wiki/Gecko%20%28software%29
Gecko is a browser engine developed by Mozilla. It is used in the Firefox browser, the Thunderbird email client, and many other projects. Gecko is designed to support open Internet standards, and is used by different applications to display web pages and, in some cases, an application's user interface itself (by rendering XUL). Gecko offers a rich programming API that makes it suitable for a wide variety of roles in Internet-enabled applications, such as web browsers, content presentation, and client/server. Gecko is written in C++ and JavaScript, and, since 2016, additionally in Rust. It is free and open-source software subject to the terms of the Mozilla Public License version 2. Mozilla officially supports its use on Android, Linux, macOS, and Windows. History Development of the layout engine now known as Gecko began at Netscape in 1997, following the company's purchase of DigitalStyle. The existing Netscape rendering engine, originally written for Netscape Navigator 1.0 and upgraded through the years, was slow, did not comply well with W3C standards, had limited support for dynamic HTML and lacked features such as incremental reflow (when the layout engine rearranges elements on the screen as new data is downloaded and added to the page). The new layout engine was developed in parallel with the old, with the intention being to integrate it into Netscape Communicator when it was mature and stable. At least one more major revision of Netscape was expected to be released with the old layout engine before the switch. After the launch of the Mozilla project in early 1998, the new layout engine code was released under an open-source license. Originally unveiled as Raptor, the name had to be changed to NGLayout (next generation layout) due to trademark problems. Netscape later rebranded NGLayout as Gecko. While Mozilla Organization (the forerunner of the Mozilla Foundation) initially continued to use the NGLayout name (Gecko was a Netscape trademark), eventually th
https://en.wikipedia.org/wiki/Rapid%20control%20prototyping
Rapid Control Prototyping (RCP) is a type of simulation methodology that allows for the rapid evaluation of control systems, especially for large machinery. It can test and evaluate algorithms as well as associated components such as sensors, actuators, pumps etc. The system requires some type of mock up, usually a scaled down version of the system to be tested, plus high powered computer simulation software. Rapid Control Prototyping has gained popularity thanks to its ability to accelerate product development and reduce their time-to-market. The approach also helps mitigate design risks, thanks to their early identification. Methodology The methodology seeks to rapidly address experimental activities in order to quickly identify and correct potential issues. If necessary, design iterations can be made using computer-assisted simulations. RCP mainly focuses on control system development (by opposition to the plant itself) and may be complementary with other techniques, such as HIL, PIL, or PHIL. Using Rapid Control Prototyping for product development requires implementing some sort of mockup (often downscaled) of the system under study as well a control system. The control system hardware often differs from the final hardware, as the latter may not yet be available (or clearly defined) at the time of prototyping. In fact, generic control hardware with superior flexibility and performance is often preferred over the final hardware, which is often cost-optimized for a specific application and use. This use of generic control hardware during the early development phase may even be considered as a definition of Rapid Control Prototyping, distinguishing it from design methodologies that exclusively involve the final hardware, possibly with several iterations. Fields of application Rapid Control Prototyping is involved in various fields of engineering, for instance: Mechatronics, robotics Power electronics, motor drives, motion control Automotive, aerospace Milit
https://en.wikipedia.org/wiki/Hellas%20%28personification%29
Hellas or Ellada is the personification of the nation of Greece, dating back to Ancient Greece. There was a desire for unification in Greece, and Hellas is the only national personification known from that period. She is mentioned frequently in literature but only appears once in the arts of late classical Athens. Description Hellas is usually depicted as a woman who wears simple clothes similar to ancient Greek clothes. On her head she wears a crown or an olive wreath.
https://en.wikipedia.org/wiki/GIS%20and%20public%20health
Geographic information systems (GISs) and geographic information science (GIScience) combine computer-mapping capabilities with additional database management and data analysis tools. Commercial GIS systems are very powerful and have touched many applications and industries, including environmental science, urban planning, agricultural applications, and others. Public health is another focus area that has made increasing use of GIS techniques. A strict definition of public health is difficult to pin down, as it is used in different ways by different groups. In general, public health differs from personal health in that it is (1) focused on the health of populations rather than of individuals, (2) focused more on prevention than on treatment, and (3) operates in a mainly governmental (rather than private) context. These efforts fall naturally within the domain of problems requiring use of spatial analysis as part of the solution, and GIS and other spatial analysis tools are therefore recognized as providing potentially transformational capabilities for public health efforts. This article presents some history of use of geographic information and geographic information systems in public health application areas, provides some examples showing the utilization of GIS techniques in solving specific public health problems, and finally addresses several potential issues arising from increased use of these GIS techniques in the public health arena. History Public health efforts have been based on analysis and use of spatial data for many years. Dr. John Snow (physician), often credited as the father of epidemiology, is arguably the most famous of those examples. Dr. Snow used a hand-drawn map to analyze the geographic locations of deaths related to cholera in London in the mid-1850s. His map, which superimposed the locations of cholera deaths with those of public water supplies, pinpointed the Broad Street pump as the most likely source of the cholera outbreak.
https://en.wikipedia.org/wiki/Vaccine%20efficacy
Vaccine efficacy or vaccine effectiveness is the percentage reduction of disease cases in a vaccinated group of people compared to an unvaccinated group. For example, a vaccine efficacy or effectiveness of 80% indicates an 80% decrease in the number of disease cases among a group of vaccinated people compared to a group in which nobody was vaccinated. When a study is carried out using the most favorable, ideal or perfectly controlled conditions, such as those in a clinical trial, the term vaccine efficacy is used. On the other hand, when a study is carried out to show how well a vaccine works when they are used in a bigger, typical population under less-than-perfectly controlled conditions, the term vaccine effectiveness is used. Vaccine efficacy was designed and calculated by Greenwood and Yule in 1915 for the cholera and typhoid vaccines. It is best measured using double-blind, randomized, clinical controlled trials, such that it is studied under "best case scenarios." Vaccine efficacy studies are used to measure several important and critical outcomes of interest such as disease attack rates, hospitalizations due to the disease, deaths due to the disease, asymptomatic infection, serious adverse events due to vaccination, vaccine reactogenicity, and cost effectiveness of the vaccine. Vaccine efficacy is calculated on a set population (and therefore is not a constant value when counting in other populations), and may be misappropriated to be how efficacious a vaccine is in all populations. Formula The outcome data (vaccine efficacy) generally are expressed as a proportionate reduction in disease attack rate (AR) between the unvaccinated (ARU) and vaccinated (ARV), or can be calculated from the relative risk (RR) of disease among the vaccinated group. The basic formula is written as:with = Vaccine efficacy, = Attack rate of unvaccinated people, = Attack rate of vaccinated people. An alternative, equivalent formulation of vaccine efficacy is: where is th
https://en.wikipedia.org/wiki/Gabriel%20Andrew%20Dirac
Gabriel Andrew Dirac (13 March 1925 – 20 July 1984) was a Hungarian-British mathematician who mainly worked in graph theory. He served as Erasmus Smith's Professor of Mathematics at Trinity College Dublin from 1964 to 1966. In 1952, he gave a sufficient condition for a graph to contain a Hamiltonian circuit. The previous year, he conjectured that n points in the plane, not all collinear, must span at least two-point lines, where is the largest integer not exceeding . This conjecture was proven true when n is sufficiently large by Green and Tao in 2012. Education Dirac started his studies at St John's College, Cambridge in 1942, but in that same year the war saw him serving in the aircraft industry. He received his MA in 1949, and moved to the University of London, getting his Ph.D. "On the Colouring of Graphs: Combinatorial topology of Linear Complexes" there under Richard Rado. Career Dirac's main academic positions were at the King's College London (1948-1954), University of Toronto (1952-1953), University of Vienna (1954-1958), University of Hamburg (1958-1963), Trinity College Dublin (Erasmus Smith's Professor of Mathematics, 1964-1966), University of Wales at Swansea (1967-1970), and Aarhus University (1970-1984). Family He was born Balázs Gábor in Budapest, to Richárd Balázs, a military officer and businessman, and Margit "Manci" Wigner (sister of Eugene Wigner). When his mother married Paul Dirac in 1937, he and his sister resettled in England and were formally adopted, changing their family name to Dirac. He married Rosemari Dirac and they had four children together: Meike, Barbara, Holger and Annette. See also Dirac's theorem on Hamiltonian cycles Dirac's theorem on chordal graphs Dirac's theorem on cycles in -connected graphs Notes
https://en.wikipedia.org/wiki/Australasian%20Ornithological%20Conference
Australasian Ornithological Conference is a biennial meeting of ornithologists that focuses on the Australasian region and Antarctica. Preceded by the short-lived series of two Southern Hemisphere Ornithological Congresses, they were initiated by the Royal Australasian Ornithologists Union (RAOU), also known as Birds Australia, with the inaugural meeting held at Bathurst, New South Wales in 2001. They have subsequently been jointly sponsored by the BirdLife Australia and the Ornithological Society of New Zealand (OSNZ). Conferences 2001 - Bathurst, New South Wales 2003 - Canberra, Australian Capital Territory 2005 - Blenheim, New Zealand 2007 - Perth, Western Australia 2009 - Armidale, New South Wales 2011 - Cairns, Queensland 2013 - Auckland, New Zealand 2015 - Adelaide, South Australia 2017 - Geelong, Victoria 2019 - Darwin, Northern Territory
https://en.wikipedia.org/wiki/Spectral%20geometry
Spectral geometry is a field in mathematics which concerns relationships between geometric structures of manifolds and spectra of canonically defined differential operators. The case of the Laplace–Beltrami operator on a closed Riemannian manifold has been most intensively studied, although other Laplace operators in differential geometry have also been examined. The field concerns itself with two kinds of questions: direct problems and inverse problems. Inverse problems seek to identify features of the geometry from information about the eigenvalues of the Laplacian. One of the earliest results of this kind was due to Hermann Weyl who used David Hilbert's theory of integral equation in 1911 to show that the volume of a bounded domain in Euclidean space can be determined from the asymptotic behavior of the eigenvalues for the Dirichlet boundary value problem of the Laplace operator. This question is usually expressed as "Can one hear the shape of a drum?", the popular phrase due to Mark Kac. A refinement of Weyl's asymptotic formula obtained by Pleijel and Minakshisundaram produces a series of local spectral invariants involving covariant differentiations of the curvature tensor, which can be used to establish spectral rigidity for a special class of manifolds. However as the example given by John Milnor tells us, the information of eigenvalues is not enough to determine the isometry class of a manifold (see isospectral). A general and systematic method due to Toshikazu Sunada gave rise to a veritable cottage industry of such examples which clarifies the phenomenon of isospectral manifolds. Direct problems attempt to infer the behavior of the eigenvalues of a Riemannian manifold from knowledge of the geometry. The solutions to direct problems are typified by the Cheeger inequality which gives a relation between the first positive eigenvalue and an isoperimetric constant (the Cheeger constant). Many versions of the inequality have been established since Cheeger's
https://en.wikipedia.org/wiki/DNA%20microarray
A DNA microarray (also commonly known as DNA chip or biochip) is a collection of microscopic DNA spots attached to a solid surface. Scientists use DNA microarrays to measure the expression levels of large numbers of genes simultaneously or to genotype multiple regions of a genome. Each DNA spot contains picomoles (10−12 moles) of a specific DNA sequence, known as probes (or reporters or oligos). These can be a short section of a gene or other DNA element that are used to hybridize a cDNA or cRNA (also called anti-sense RNA) sample (called target) under high-stringency conditions. Probe-target hybridization is usually detected and quantified by detection of fluorophore-, silver-, or chemiluminescence-labeled targets to determine relative abundance of nucleic acid sequences in the target. The original nucleic acid arrays were macro arrays approximately 9 cm × 12 cm and the first computerized image based analysis was published in 1981. It was invented by Patrick O. Brown. An example of its application is in SNPs arrays for polymorphisms in cardiovascular diseases, cancer, pathogens and GWAS analysis. It is also used for the identification of structural variations and the measurement of gene expression. Principle The core principle behind microarrays is hybridization between two DNA strands, the property of complementary nucleic acid sequences to specifically pair with each other by forming hydrogen bonds between complementary nucleotide base pairs. A high number of complementary base pairs in a nucleotide sequence means tighter non-covalent bonding between the two strands. After washing off non-specific bonding sequences, only strongly paired strands will remain hybridized. Fluorescently labeled target sequences that bind to a probe sequence generate a signal that depends on the hybridization conditions (such as temperature), and washing after hybridization. Total strength of the signal, from a spot (feature), depends upon the amount of target sample binding to the
https://en.wikipedia.org/wiki/V-Key
V-Key is a software-based digital security provider. Headquartered in Singapore, it provides products to financial institutions, mobile payment providers and governments to implement cloud-based payments, authentication for mobile banking, and secured mobile applications for user access and data protection. Background & founders V-Key was founded in 2011 by entrepreneurs Eddie Chau, Benjamin Mah and Joseph Gan. Eddie Chau, who are the founders of digital agency Brandtology, acquired by iSentia in 2014, started V-Key primarily to secure mobile devices and applications with patented technology. Benjamin Mah is the co-founder and chief executive officer of V-Key. He was general manager at e-Cop (acquired by a wholly owned subsidiary of Temasek Holdings) and regional director at Encentuate (acquired by IBM), before he co-founded V-Key. He is a concurrently venture partner of Venture Craft, chairman of Jump Start Asia and a mentor at UOB Finlabs. Joseph Gan is the third co-founder of V-Key. Before joining V-Key, he was at the Center for Strategic Info comm Technologies (CSIT) as the head of the Cryptography Lab, where he oversaw research and development into cryptographic software for the Ministry of Defence (Singapore). Companies that have funded V-Key are IPV Capital and Ant Financial Services, which runs the Alipay mobile wallet app. Technology V-Key provides security to businesses to support cloud-based payments, digital identity and authentication for mobile banking as well as other secured mobile applications via its core technology—V-OS. V-Key's partners are financial institutions, governments and mobile payment providers in various markets. Its technology has been used by: ChinaPnR - financial payment provider: V-Key integrates a virtual secure element into "ChinaPnR POS Acquirer" (Point-of-Sale payment acquisition query platform) to protect mobile applications' runtime environment, program logic and important data. See also Application Security Encr
https://en.wikipedia.org/wiki/Euler%20product
In number theory, an Euler product is an expansion of a Dirichlet series into an infinite product indexed by prime numbers. The original such product was given for the sum of all positive integers raised to a certain power as proven by Leonhard Euler. This series and its continuation to the entire complex plane would later become known as the Riemann zeta function. Definition In general, if is a bounded multiplicative function, then the Dirichlet series is equal to where the product is taken over prime numbers , and is the sum In fact, if we consider these as formal generating functions, the existence of such a formal Euler product expansion is a necessary and sufficient condition that be multiplicative: this says exactly that is the product of the whenever factors as the product of the powers of distinct primes . An important special case is that in which is totally multiplicative, so that is a geometric series. Then as is the case for the Riemann zeta function, where , and more generally for Dirichlet characters. Convergence In practice all the important cases are such that the infinite series and infinite product expansions are absolutely convergent in some region that is, in some right half-plane in the complex numbers. This already gives some information, since the infinite product, to converge, must give a non-zero value; hence the function given by the infinite series is not zero in such a half-plane. In the theory of modular forms it is typical to have Euler products with quadratic polynomials in the denominator here. The general Langlands philosophy includes a comparable explanation of the connection of polynomials of degree , and the representation theory for . Examples The following examples will use the notation for the set of all primes, that is: The Euler product attached to the Riemann zeta function , also using the sum of the geometric series, is while for the Liouville function , it is Using their reciprocals, two Euler produc
https://en.wikipedia.org/wiki/Gas%20in%20scattering%20media%20absorption%20spectroscopy
Gas in scattering media absorption spectroscopy (GASMAS) is an optical technique for sensing and analysis of gas located within porous and highly scattering solids, e.g. powders, ceramics, wood, fruit, translucent packages, pharmaceutical tablets, foams, human paranasal sinuses etc. It was introduced in 2001 by Prof. Sune Svanberg and co-workers at Lund University (Sweden). The technique is related to conventional high-resolution laser spectroscopy for sensing and spectroscopy of gas (e.g. tunable diode laser absorption spectroscopy, TDLAS), but the fact that the gas here is "hidden" inside solid materials give rise to important differences. Basic Principles Free gases exhibit very sharp spectral features, and different gas species have their own unique spectral fingerprints. At atmospheric pressure, absorption linewidths are typically on the order of 0.1 cm−1 (i.e. ~3 GHz in optical frequency or 0.006 nm in wavelength), while solid media have dull spectral behavior with absorption features thousand times wider. By looking for the sharp absorption imprints in light emerging from porous samples, it is thus possible to detect gases confined in solids – even though the solid often attenuates light much stronger than the gas itself. The basic principle of GASMAS is shown in figure 1. Laser light is sent into a sample with gas cavities, which could either be small pores (left) or larger gas-filled chambers. The heterogeneous nature of the porous material often give rise to strong light scattering, and pathlengths are often surprisingly long (10 or 100 times the sample dimension are not uncommon). In addition, light will experience absorption related to the solid material. When travelling through the material, light will travel partly through the pores, and will thus experience the spectrally sharp gas absorption. Light leaving the material will carry this information, and can be collected by a detector either in a transmission mode (left) or in a reflection mode (rig
https://en.wikipedia.org/wiki/Piano%20key%20frequencies
This is a list of the fundamental frequencies in Hertz (cycles per second) of the keys of a modern 88-key standard or 108-key extended piano in twelve-tone equal temperament, with the 49th key, the fifth A (called A4), tuned to 440 Hz (referred to as A440). Every octave is made of twelve steps called semitones. A jump from the lowest semitone to the highest semitone in one octave doubles the frequency (for example, the fifth A is 440 Hz and the sixth A is 880 Hz). The frequency of a pitch is derived by multiplying (ascending) or dividing (descending) the frequency of the previous pitch by the twelfth root of two (approximately 1.059463). For example, to get the frequency one semitone up from A4 (A4), multiply 440 Hz by the twelfth root of two. To go from A4 up two semitones (one whole tone) to B4, multiply 440 twice by the twelfth root of two (or once by the sixth root of two, approximately 1.122462). To go from A4 up three semitones to C5 (a minor third), multiply 440 Hz three times by the twelfth root of two (or once by the fourth root of two, approximately 1.189207). For other tuning schemes, refer to musical tuning. This list of frequencies is for a theoretically ideal piano. On an actual piano, the ratio between semitones is slightly larger, especially at the high and low ends, where string stiffness causes inharmonicity, i.e., the tendency for the harmonic makeup of each note to run sharp. To compensate for this, octaves are tuned slightly wide, stretched according to the inharmonic characteristics of each instrument. This deviation from equal temperament is called the Railsback curve. The following equation gives the frequency (Hz) of the th key on the idealized standard piano with the 49th key tuned to A4 at 440 Hz: where is shown in the table below. Conversely, the key number of a pitch with a frequency (Hz) on the idealized standard piano is: List Values in bold are exact on an idealized standard piano. Keys shaded gray are rare and only appear
https://en.wikipedia.org/wiki/Omega2%20%28computer%29
The Onion Omega series of personal single-board computer created by a startup company called Onion that is based in Boston, Toronto and Shenzhen. It is advertised as "the world's smallest Linux Server". The system combines a tiny form factor and power-efficiency with the power of a general purpose Operating System. They ship with a Linux kernel based lightweight operating system for embedded systems called OpenWRT, but is capable of running other lightweight Unix-based operating systems. The first shipments of the Onion Omega went out in October, 2015. History Omega2 is the next generation of the old product Onion makes, Omega. The original Omega was based on the Qualcomm Atheros AR9331 (MIPS architecture) SoC which runs a full Linux operating system designed for embedded system and sold for $19.99. The company has discontinued development of the Omega, and replaced it with the successor, Omega2, using another SoC chip - Mediatek MT7688 which also has a metal cover over the chip. They have also drastically cut the price to $5 (but later increased it to $7.5). As of the beginning of 2017, Onion has already attracted crowdfunding of more than $850,000 for the Omega2, which has greatly exceeded their initial goal of $440,000. Hardware Features Omega2 comes in two versions, the basic Omega2 and Omega2 Plus. Omega2 CPU is based on MIPS architecture running at 580 MHz clock speed, equipped with 64 MB of RAM and 16 MB of flash memory. Omega2 Plus is similar to Omega2, except it has 128 MB RAM and 32 MB memory and a MicroSD slot, sold for $9 USD. The system comes in a small PCB footprint with dual-in-line 16x2mm pins. The board runs at 3.3 volts with an average power consumption of 0.6W. The devices are intended for as headless computers with no graphical interfaces in Embedded systems.
https://en.wikipedia.org/wiki/24-Norcholestane
24-Norcholestane, a steroid derivative, is used as a biomarker to constrain the source age of sediments and petroleum through the ratio between 24-norcholestane and 27-norcholestane (24-norcholestane ratio, NCR), especially when used with other age diagnostic biomarkers, like oleanane. While the origins of this compound are still unknown, it is thought that they are derived from diatoms due to their identification in diatom rich sediments and environments. In addition, it was found that 24-norcholestane levels increased in correlation with diatom evolution. Another possible source of 24-norcholestane is from dinoflagellates, albeit to a much lower extent. Structure 24-Norcholestane is a tetracyclic compound, with 20R,5α(H),14α(H),17α(H) stereochemistry, derived from steroids or sterols. It consists of three 6-membered rings and one 5-membered ring, with carbon 24 removed from the side chain off of C17. Background 24-Norcholestane is a 26-carbon (C26) sterane created from the removal of carbon 24 from cholestane. It has been found that 24-norcholestane is relatively high in abundance, up to 10% of sterols, in Thalassiosira aff. antarctica, a diatom. It has also been found in the dinoflagellate Gymnodinium simplex, albeit at much lower levels (around 0.2% of sterols). Origins Since 24-norcholestane origins are still unknown, the synthesis of it is also unknown as well. However, some pathways have been proposed. Possible sources of 24-norcholestane include 24-norcholesterol, which is present in many marine invertebrates and some algae in addition to diatoms and dinoflagellates. Measurement techniques Samples are collected from rocks or crude oils. Asphaltenes are first extracted before the sample is fractionated by passing through a silica column and eluting with solvents of increasing polarity. Traditional gas chromatography-mass spectrometry (GC/MS) techniques are not used, as C26 steranes are present in samples in much lower quantities, generally a magnitud
https://en.wikipedia.org/wiki/Hypsometric%20equation
The hypsometric equation, also known as the thickness equation, relates an atmospheric pressure ratio to the equivalent thickness of an atmospheric layer considering the layer mean of virtual temperature, gravity, and occasionally wind. It is derived from the hydrostatic equation and the ideal gas law. Formulation The hypsometric equation is expressed as: where: = thickness of the layer [m], = geometric height [m], = specific gas constant for dry air, = mean virtual temperature in Kelvin [K], = gravitational acceleration [m/s2], = pressure [Pa]. In meteorology, and are isobaric surfaces. In radiosonde observation, the hypsometric equation can be used to compute the height of a pressure level given the height of a reference pressure level and the mean virtual temperature in between. Then, the newly computed height can be used as a new reference level to compute the height of the next level given the mean virtual temperature in between, and so on. Derivation The hydrostatic equation: where is the density [kg/m3], is used to generate the equation for hydrostatic equilibrium, written in differential form: This is combined with the ideal gas law: to eliminate : This is integrated from to : R and g are constant with z, so they can be brought outside the integral. If temperature varies linearly with z (e.g., given a small change in z), it can also be brought outside the integral when replaced with , the average virtual temperature between and . Integration gives simplifying to Rearranging: or, eliminating the natural log: Correction The Eötvös effect can be taken into account as a correction to the hypsometric equation. Physically, using a frame of reference that rotates with Earth, an air mass moving eastward effectively weighs less, which corresponds to an increase in thickness between pressure levels, and vice versa. The corrected hypsometric equation follows: where the correction due to the Eötvös effect, A, can be expressed as follows
https://en.wikipedia.org/wiki/Invader%20potential
Ecologically, invader potential is the qualitative and quantitative measures of a given invasive species probability to invade a given ecosystem. This is often seen through climate matching. There are many reasons why a species may invade a new area. The term invader potential may also be interchangeable with invasiveness. Invader potential is a large threat to global biodiversity. It has been shown that there is an ecosystem function loss due to the introduction of species in areas they are not native to. Invaders are species that, through biomass, abundance, and strong interactions with natives, have significantly altered the structure and composition of the established community. This differs greatly from the term "introduced", which merely refers to species that have been introduced to an environment, disregarding whether or not they have created a successful establishment.1 They are simply organisms that have been accidentally, or deliberately, placed into an unfamiliar area .2 Many times, in fact, species do not have a strong impact on the introduced habitat. This can be for a variety of reasons; either the newcomers are not abundant or because they are small and unobtrusive.1 Understanding the mechanisms of invader potential is important to understanding why species relocate and to predict future invasions. There are three predicted reasons as to why species invade an area. They are as follows: adaptation to physical environment, resource competition and/or utilization, and enemy release. Some of these reasons as to why species move seem relatively simple to understand. For example, species may adapt to the new physical environment through having great phenotypic plasticity and environmental tolerance. Species with high rates of these find it easier to adapt to new environments. In terms of resources, those with low resource requirements thrive in unknown areas more than those with complex resource needs. This is shown directly through Tilman's R* rule. Tho
https://en.wikipedia.org/wiki/IBM%20303X
The IBM 303X is a discontinued line of mainframe computers, the first model of which, the IBM 3033 Processor, nicknamed "The Big One", was introduced March 25, 1977. Two additional processors, the 3031 and the 3032, were announced on October 6, 1977. All three 303X systems were withdrawn on February 5, 1985. Features The CPUs feature instruction pipelining, "several instructions can be pre-fetched while one is being executed". "Processor storage ... is four-way interleaved" resulting in "a significantly faster data rate than... non-interleaved". Each of the three systems includes, as a standard feature, a Dual-display console, the newly announced IBM 3036. The systems consume less than half the floor space of a System/370 with an equal amount of computer memory and an identical number of channels because "the channels are physically integrated within the processor mainframe." Each group of six channels has its own microprogrammed channel processor, called a Channel Director. The Director is internally very similar to a System/370 Model 158, running special channel microcode. The initial announcement of the 3033 also introduced new operating system versions "MVS/System Extensions (MVS/SE) and VM/System Extensions (VM/SE)." 303X as successor to System 370 Beginning in 1977, IBM began to introduce new systems, using the following descriptions: "A compatible member of the System/370 family." "System/370 Compatible - 303(1/2/3)" "the System/370 3033 Processor." Competing compatibles At this time, other companies, known as plug compatible manufacturers (PCMs), were competing with IBM by producing IBM-compatible systems. By the late 1970s and early 1980s, patented technology allowed Amdahl IBM-compatible mainframes of this era to be completely air-cooled, unlike IBM systems that required chilled water and its supporting infrastructure— the 8 largest of the 18 models of the ES/9000 systems introduced in 1990 were water-cooled; the other ten were air-cooled. "The
https://en.wikipedia.org/wiki/Online%20poker
Online poker is the game of poker played over the Internet. It has been partly responsible for a huge increase in the number of poker players worldwide. Christiansen Capital Advisors stated online poker revenues grew from $82.7 million in 2001 to $2.4 billion in 2005, while a survey carried out by DrKW and Global Betting and Gaming Consultants asserted online poker revenues in 2004 were at $1.4 billion. In a testimony before the United States Senate regarding Internet Gaming, Grant Eve, a Certified Public Accountant representing the US Accounting Firm Joseph Eve, Certified Public Accountants, estimated that one in every four dollars gambled is gambled online. Traditional (or "brick and mortar", B&M, live, land-based) venues for playing poker, such as casinos and poker rooms, may be intimidating for novice players and are often located in geographically disparate locations. Also, brick and mortar casinos are reluctant to promote poker because it is difficult for them to profit from it. Though the rake, or time charge, of traditional casinos is often high, the opportunity costs of running a poker room are even higher. Brick and mortar casinos often make much more money by removing poker rooms and adding more slot machines. For example, figures from the Gaming Accounting Firm Joseph Eve estimate that poker accounts for 1% of brick and mortar casino revenues. Online venues, by contrast, are dramatically cheaper because they have much smaller overhead costs. For example, adding another table does not take up valuable space like it would for a brick and mortar casino. Online poker rooms also allow the players to play for low stakes (as low as 1¢/2¢) and often offer poker freeroll tournaments (where there is no entry fee), attracting beginners and/or less wealthy clientele. Online venues may be more vulnerable to certain types of fraud, especially collusion between players. However, they have collusion detection abilities that do not exist in brick and mortar casinos. F
https://en.wikipedia.org/wiki/Bioceramic
Bioceramics and bioglasses are ceramic materials that are biocompatible. Bioceramics are an important subset of biomaterials. Bioceramics range in biocompatibility from the ceramic oxides, which are inert in the body, to the other extreme of resorbable materials, which are eventually replaced by the body after they have assisted repair. Bioceramics are used in many types of medical procedures. Bioceramics are typically used as rigid materials in surgical implants, though some bioceramics are flexible. The ceramic materials used are not the same as porcelain type ceramic materials. Rather, bioceramics are closely related to either the body's own materials or are extremely durable metal oxides. History Prior to 1925, the materials used in implant surgery were primarily relatively pure metals. The success of these materials was surprising considering the relatively primitive surgical techniques. The 1930s marked the beginning of the era of better surgical techniques as well as the first use of alloys such as vitallium. In 1969, L. L. Hench and others discovered that various kinds of glasses and ceramics could bond to living bone. Hench was inspired by the idea on his way to a conference on materials. He was seated next to a colonel who had just returned from the Vietnam War. The colonel shared that after an injury the bodies of soldiers would often reject the implant. Hench was intrigued and began to investigate materials that would be biocompatible. The final product was a new material which he called bioglass. This work inspired a new field called bioceramics. With the discovery of bioglass, interest in bioceramics grew rapidly. On April 26, 1988, the first international symposium on bioceramics was held in Kyoto, Japan. Applications Ceramics are now commonly used in the medical fields as dental and bone implants. Surgical cermets are used regularly. Joint replacements are commonly coated with bioceramic materials to reduce wear and inflammatory response. Other
https://en.wikipedia.org/wiki/Distance%20from%20a%20point%20to%20a%20line
In Euclidean geometry, the distance from a point to a line is the shortest distance from a given point to any point on an infinite straight line. It is the perpendicular distance of the point to the line, the length of the line segment which joins the point to nearest point on the line. The algebraic expression for calculating it can be derived and expressed in several ways. Knowing the distance from a point to a line can be useful in various situations—for example, finding the shortest distance to reach a road, quantifying the scatter on a graph, etc. In Deming regression, a type of linear curve fitting, if the dependent and independent variables have equal variance this results in orthogonal regression in which the degree of imperfection of the fit is measured for each data point as the perpendicular distance of the point from the regression line. Line defined by an equation In the case of a line in the plane given by the equation , where , and are real constants with and not both zero, the distance from the line to a point is The point on this line which is closest to has coordinates: Horizontal and vertical lines In the general equation of a line, , and cannot both be zero unless is also zero, in which case the equation does not define a line. If and , the line is horizontal and has equation . The distance from to this line is measured along a vertical line segment of length in accordance with the formula. Similarly, for vertical lines (b = 0) the distance between the same point and the line is , as measured along a horizontal line segment. Line defined by two points If the line passes through two points and then the distance of from the line is: The denominator of this expression is the distance between and . The numerator is twice the area of the triangle with its vertices at the three points, , and . See: . The expression is equivalent to , which can be obtained by rearranging the standard formula for the area of a triangle: , where
https://en.wikipedia.org/wiki/Database%20of%20Molecular%20Motions
The Database of Macromolecular Motions is a bioinformatics database and software-as-a-service tool that attempts to categorize macromolecular motions, sometimes also known as conformational change. It was originally developed by Mark B. Gerstein, Werner Krebs, and Nat Echols in the Molecular Biophysics & Biochemistry Department at Yale University. Discussion Since its introduction in the late 1990s, peer-reviewed papers on the database have received thousands of citations. The database has been mentioned in news articles in major scientific journals, book chapters, and elsewhere. Users can search the database for a particular motion by either protein name or Protein Data Bank ID number. Typically, however, users will enter the database via the Protein Data Bank, which often provides a hyperlink to the molmovdb entry for proteins found in both databases. The database includes a web-based tool (the Morph server) which allows non-experts to animate and visualize certain types of protein conformational change through the generation of short movies. This system uses molecular modelling techniques to interpolate the structural changes between two different protein conformers and to generate a set of intermediate structures. A hyperlink pointing to the morph results is then emailed to the user. The Morph Server was originally primarily a research tool rather than general molecular animation tool, and thus offered only limited user control over rendering, animation parameters, color, and point of view, and the original methods sometimes required a fair amount of CPU time to completion. Since their initial introduction in 1996, the database and associated morph server have undergone development to try to address some of these shortcomings as well as add new features, such as Normal Mode Analysis. Other research grounds have subsequently developed alternative systems, such as MovieMaker from the University of Alberta. Commercialization Bioinformatics vendor DNASTAR ha
https://en.wikipedia.org/wiki/Bamberger%20Ranch%20Preserve
Bamberger Ranch, also called Selah, is an ecological restoration and conservation project near Johnson City, in Blanco County, Texas in the Texas Hill Country. History The preserve was created in 1969 by restaurant chain executive J. David Bamberger. Bamberger specifically sought out a worn-out ranch with poor economic prospects and when he found Selah, “The ranch was mostly bare ground or infested with cedar (Juniperus ashei). There was no grass and absolutely no live creeks, springs, or ponds.” After removing the ”cedar” and planting native grasses, rainwater infiltration improved and the aquifer underlying the property refilled. The site now has several running springs that did not exist when Bamberger bought the place. Bird biodiversity on the land has increased from around 50 species to over 220 species. The ranch also works to promulgate endangered species. It has a herd of scimitar-horned oryx (native to Africa), and Bamberger also markets the Texas snowbell tree to other ranchers. Bat cave The site has its own human-built bat cave, completed in 1998 and called the Chiroptorium. The Chiroptorium is now home to a colony of 200,000 Mexican free-tailed bats.
https://en.wikipedia.org/wiki/Fideicommissum%20%28sculpture%29
Fideicommissum is a public artwork by the Swedish artist Ann-Sofi Sidén, originally created in 2000 for the Wanås Sculpture Park in Skåne, Sweden. The work is an edition of six. History The first edition of Fideicommissum was placed in the Wanås Sculpture Park, which is situated at Wanås Castle in Skåne. Description Name The title Fideicommissum refers to the former legal institution of the entail, under which estates and other family properties were passed down from one generation to the next, often benefitting the eldest son. In Sweden, a law abolishing entailed estates was implemented in 1964. Sidén's sculpture is located on the shore of a pond, with a view of the castle, which was entailed until 1978. The title Fideicommissum and the placement of the sculpture emphasize that the work refers to a staking out of territory. Composition Fideicommissum is a bronze cast of the artist herself squatting down to urinate. The artwork functions as a fountain and makes reference to the fountain motifs of young males urinating. Sidén conceived the idea for the sculpture in 1999 during a visit to the Wanås Sculpture Park, when she was working on a film and had to retire to the bushes to relieve herself. According to the artist, the urinating narrative echoes to a primitive way to mark a territory. The composition of the artwork has historical references. The sculpture's hair, sandals and folds of clothing are modelled on Roman, Greek and Etruscan sculptures. Editions The first version of Fideicommissum was made for the Wanås Sculpture Park. Another is at Ekebergparken Sculpture Park in Oslo, Norway. Sidén also documented the installation of the first edition of the sculpture at the sculpture park in Wanås. That project resulted in a video work called Head Lake Piss Down (2000).
https://en.wikipedia.org/wiki/Interferon%20gamma
Interferon gamma (IFN-γ) is a dimerized soluble cytokine that is the only member of the type II class of interferons. The existence of this interferon, which early in its history was known as immune interferon, was described by E. F. Wheelock as a product of human leukocytes stimulated with phytohemagglutinin, and by others as a product of antigen-stimulated lymphocytes. It was also shown to be produced in human lymphocytes. or tuberculin-sensitized mouse peritoneal lymphocytes challenged with Mantoux test (PPD); the resulting supernatants were shown to inhibit growth of vesicular stomatitis virus. Those reports also contained the basic observation underlying the now widely employed IFN-γ release assay used to test for tuberculosis. In humans, the IFN-γ protein is encoded by the IFNG gene. Through cell signaling, IFN-γ plays a role in regulating the immune response of its target cell. A key signaling pathway that is activated by type II IFN is the JAK-STAT signaling pathway. IFN-γ plays an important role in both innate and adaptive immunity. Type II IFN is primarily secreted by adaptive immune cells, more specifically CD4+ T helper 1 (Th1) cells, natural killer (NK) cells, and CD8+ cytotoxic T cells. The expression of type II IFN is upregulated and downregulated by cytokines. By activating signaling pathways in cells such as macrophages, B cells, and CD8+ cytotoxic T cells, it is able to promote inflammation, antiviral or antibacterial activity, and cell proliferation and differentiation. Type II IFN is serologically different from interferon type 1, binds to different receptors, and is encoded by a separate chromosomal locus. Type II IFN has played a role in the development of cancer immunotherapy treatments due to its ability to prevent tumor growth. Function IFN-γ, or type II interferon, is a cytokine that is critical for innate and adaptive immunity against viral, some bacterial and protozoan infections. IFN-γ is an important activator of macrophages and in
https://en.wikipedia.org/wiki/Repligen%20Corporation%20Award%20in%20Chemistry%20of%20Biological%20Processes
The Repligen Award in Chemistry of Biological Processes was established in 1985 and consists of a silver medal and honorarium. Its purpose is to acknowledge and encourage outstanding contributions to the understanding of the chemistry of biological processes, with particular emphasis on structure, function, and mechanism. The Award is administered by the Division of Biological Chemistry of the American Chemical Society. The award was suspended in 2018 until a patron can be found. Recipients Source: ACS - Division of Biological Chemistry 1986 – Gregorio Weber 1987 – Thomas C. Bruice 1988 – Robert H. Abeles 1989 – Stephen J. Benkovic 1990 – Harold A. Scheraga 1991 – William W. Parson 1992 – Frank H. Westheimer 1993 – Jeremy R. Knowles 1994 – Judith P. Klinman 1995 – W. Wallace Cleland 1996 – William P. Jencks 1997 – James A. Spudich 1998 – David S. Eisenberg 1999 – Christopher T. Walsh 2000 – Perry A. Frey 2001 – Rowena G. Matthews 2002 – C. Dale Poulter 2003 – John A. Gerlt 2004 – JoAnne Stubbe 2005 – David E. Cane 2006 – Vern L. Schramm 2007 – Michael Marletta 2008 – Hung-Wen (Ben) Liu 2009 – Frank Raushel 2010 – Ronald T. Raines 2011 – Richard Armstrong 2012 – Carol Fierke 2013 – David W. Christianson 2014 – John Lipscomb 2015 – John S. Blanchard 2016 – Tadhg Begley 2017 – Wilfred A. van der Donk 2018 – Michael H. Gelb award suspended - 2018 See also List of biochemistry awards
https://en.wikipedia.org/wiki/Nordic%20Mathematical%20Contest
The Nordic Mathematical Contest (NMC) is a mathematics competition for secondary school students from the five Nordic countries: Denmark, Finland, Iceland, Norway and Sweden. It takes place every year in March or April and serves the double purpose of being a regional secondary school level mathematics competition for the Nordic region and a step in the process of selection of the teams of the participating countries for the International Mathematical Olympiad (IMO) and the regional Baltic Way competition. Participation At most twenty participants from each country are appointed by the organisers of the national secondary school level mathematics competitions. They must either be eligible to the IMO or attend a secondary school. (The foreword of ref. renders the eligibility requirements unlike the past and present regulations.) Problems The exam consists of four problems to be answered in four hours. Only writing and drawing tools are permitted. For each problem the contestant can get from zero to seven points. The problems are of the IMO type and harder than those of the national secondary school level competitions in mathematics of the Nordic countries but not as hard as those of the IMO. They are chosen by the organising committee of the host country of the year from proposals submitted by the national organising committees. The official web site of the NMC provides a complete collection in English with solutions of the problems from all the years. It is compiled by Matti Lehtinen. Selected versions of the problems in other Nordic languages are also available at the site Organisation The NMC is run in a decentralised manner involving no travel of the contestants nor any other personnel. The contestants write the exam in their own schools on the same day. Thence the papers are sent to a committee in the contestants' country who mark them preliminarily. They are then forwarded with the preliminary marking to a committee in the host country of the year, who coo
https://en.wikipedia.org/wiki/Oncotic%20pressure
Oncotic pressure, or colloid osmotic-pressure, is a type of osmotic pressure induced by the plasma proteins, notably albumin, in a blood vessel's plasma (or any other body fluid such as blood and lymph) that causes a pull on fluid back into the capillary. Participating colloids displace water molecules, thus creating a relative water molecule deficit with water molecules moving back into the circulatory system within the lower venous pressure end of capillaries. It has the opposing effect of both hydrostatic blood pressure pushing water and small molecules out of the blood into the interstitial spaces within the arterial end of capillaries and interstitial colloidal osmotic pressure. These interacting factors determine the partition balancing of extracellular water between the blood plasma and outside the blood stream. Oncotic pressure strongly affects the physiological function of the circulatory system. It is suspected to have a major effect on the pressure across the glomerular filter. However, this concept has been strongly criticised and attention has been shifted to the impact of the intravascular glycocalyx layer as the major player. Etymology The word 'Oncotic' by definition is termed as 'pertaining to swelling', indicating the effect of oncotic imbalance on the swelling of tissues. The word itself is derived from onco- and -ic; 'onco-' meaning 'pertaining to mass or tumors' and '-ic', which forms an adjective. Description Throughout the body, dissolved compounds have an osmotic pressure. Because large plasma proteins cannot easily cross through the capillary walls, their effect on the osmotic pressure of the capillary interiors will, to some extent, balance out the tendency for fluid to leak out of the capillaries. In other words, the oncotic pressure tends to pull fluid into the capillaries. In conditions where plasma proteins are reduced, e.g. from being lost in the urine (proteinuria), there will be a reduction in oncotic pressure and an increase
https://en.wikipedia.org/wiki/Priest%20of%20Nature
Priest of Nature: The Religious Worlds of Isaac Newton is a 2017 book by science historian Rob Iliffe on the religious views of Isaac Newton. Background Rob Iliffe is Professor of the History of Science, Linacre College, University of Oxford. For nearly twenty years leading up to the book's publication, he directed The Newton Project – an online repository of Isaac Newton's manuscripts with editorial commentary. Previously, Iliffe wrote another book on Newton, Newton: A Very Short Introduction, that was published by Oxford University Press in 2007. Themes Newton had a lifelong interest in theology, especially prophecies in the Book of Revelation. The book shows that in one of the Philosophiæ Naturalis Principia Mathematicas appendices, General Scholium, Newton argued that the "divine mode of being" was unknown, an argument that threatened the traditional theological concept of incarnation. Newton never believed in the Trinity, a heretical view that played a large role in his religious writings being only recently published after nearly 300 years. Reception The book was met with critical acclaim. The book was reviewed by David Brion Davis, Graham Farmelo, Peter Harrison, Eamon Duffy, William Gibson, Floris Cohen, and Guy Stroumsa, along with many others, including reviews in newspapers and magazines, literary publications, academic journals, and theological journals. The book's review in Publishers Weekly stated that "Iliffe skillfully chronicles Newton's life" and that he "adroitly illustrates that, from the beginning, Newton displayed deep interests in scriptural interpretation, the history of the early Church, and the idea of prophecy". The review closes by stating: "Iliffe's fascinating study provides an absorbing glimpse into Newton's work and early modern culture". In his review of the book, Peter Harrison stated that the book "presents what is surely the definitive account of Newton's religious world view" and went on to write: "It is a remarkable work o
https://en.wikipedia.org/wiki/Cramer%27s%20theorem%20%28algebraic%20curves%29
In algebraic geometry, Cramer's theorem on algebraic curves gives the necessary and sufficient number of points in the real plane falling on an algebraic curve to uniquely determine the curve in non-degenerate cases. This number is where is the degree of the curve. The theorem is due to Gabriel Cramer, who published it in 1750. For example, a line (of degree 1) is determined by 2 distinct points on it: one and only one line goes through those two points. Likewise, a non-degenerate conic (polynomial equation in and with the sum of their powers in any term not exceeding 2, hence with degree 2) is uniquely determined by 5 points in general position (no three of which are on a straight line). The intuition of the conic case is this: Suppose the given points fall on, specifically, an ellipse. Then five pieces of information are necessary and sufficient to identify the ellipse—the horizontal location of the ellipse's center, the vertical location of the center, the major axis (the length of the longest chord), the minor axis (the length of the shortest chord through the center, perpendicular to the major axis), and the ellipse's rotational orientation (the extent to which the major axis departs from the horizontal). Five points in general position suffice to provide these five pieces of information, while four points do not. Derivation of the formula The number of distinct terms (including those with a zero coefficient) in an n-th degree equation in two variables is (n + 1)(n + 2) / 2. This is because the n-th degree terms are numbering n + 1 in total; the (n − 1) degree terms are numbering n in total; and so on through the first degree terms and numbering 2 in total, and the single zero degree term (the constant). The sum of these is (n + 1) + n + (n – 1) + ... + 2 + 1 = (n + 1)(n + 2) / 2 terms, each with its own coefficient. However, one of these coefficients is redundant in determining the curve, because we can always divide through the polynomial equatio
https://en.wikipedia.org/wiki/Multiresolution%20Fourier%20transform
Multiresolution Fourier Transform is an integral fourier transform that represents a specific wavelet-like transform with a fully scalable modulated window, but not all possible translations. Comparison of Fourier transform and wavelet transform The Fourier transform is one of the most common approaches when it comes to digital signal processing and signal analysis. It represents a signal through sine and cosine functions thus transforming the time-domain into frequency-domain. A disadvantage of the Fourier transform is that both sine and cosine function are defined in the whole time plane, meaning that there is no time resolution. Certain variants of Fourier transform, such as Short Time Fourier Transform (STFT) utilize a window for sampling, but the window length is fixed meaning that the results will be satisfactory only for either low or high frequency components. Fast fourier transform (FFT) is used often because of its computational speed, but shows better results for stationary signals. On the other hand, the wavelet transform can improve all the aforementioned downsides. It preserves both time and frequency information and it uses a window of variable length, meaning that both low and high frequency components will be derived with higher accuracy than the Fourier transform. The wavelet transform also shows better results in transient states. Multiresolution Fourier Transform leverages the advantageous properties of the wavelet transform and uses them for Fourier transform. Definition Let be a function that has its Fourier transform defined as    The time line can be split by intervals of length π/ω with centers at integer multiples of π/ω    Then, new transforms of function can be introduced       and       where , when n is an integer. Functions and can be used in order to define the complex Fourier transform    Then, set of points in the frequency-time plane is defined for the computation of the introduced transforms   
https://en.wikipedia.org/wiki/List%20of%20C%2B%2B%20multi-threading%20libraries
List of cross-platform multi-threading libraries for the C++ programming language. Apache Portable Runtime Boost.Thread C++ Standard Library Thread Concurrencpp Dlib HPX IPP OpenMP OpenThreads Parallel Patterns Library POCO C++ Libraries Threading POSIX Threads Qt QThread Rogue Wave SourcePro Threads Module Stapl Taskflow TBB C++ libraries C++ Multi-threading libraries Multi-threading libraries
https://en.wikipedia.org/wiki/Tosyl%20phenylalanyl%20chloromethyl%20ketone
Tosyl phenylalanyl chloromethyl ketone (TPCK) is a protease inhibitor. Its structural formula is 1-chloro-3-tosylamido-4-phenyl-2-butanone. Uses TPCK is an irreversible inhibitor of chymotrypsin. Also inhibits some cysteine proteases such as caspase, papain, bromelain or ficin. It does not inhibit trypsin or zymogens. TPCK is observed covalently bound in the active site of Caspase 3 in the crystal structure of the complex solved in 2010. The chloromethyl group reacts with the active site cysteine to form a covalent bond with the loss of the chlorine. TPCK is chosen for the chemical labelling of active histidine in enzyme analysis. The phenylalanine moiety is bound to the enzyme because of specificity for aromatic amino acid residues at the active site (as in chymotrypsin, in which it binds to the Histidine-57 residue in the active site). Virology TPCK-treated trypsin is used to improve infection yield in laboratory tissue culture of some wild virus isolates that are not well-adapted to growth in vitro, such as some low-pathogenic avian influenza strains or fresh clinical isolates of SARS-CoV-2. The trypsin performs the maturation cleavage of the viral envelope proteins efficiently.
https://en.wikipedia.org/wiki/E-selectin
E-selectin, also known as CD62 antigen-like family member E (CD62E), endothelial-leukocyte adhesion molecule 1 (ELAM-1), or leukocyte-endothelial cell adhesion molecule 2 (LECAM2), is a selectin cell adhesion molecule expressed only on endothelial cells activated by cytokines. Like other selectins, it plays an important part in inflammation. In humans, E-selectin is encoded by the SELE gene. Structure E selectin has a cassette structure: an N-terminal, C-type lectin domain, an EGF (epidermal-growth-factor)-like domain, 6 Sushi domain (SCR repeat) units, a transmembrane domain (TM) and an intracellular cytoplasmic tail (cyto). The three-dimensional structure of the ligand-binding region of human E-selectin has been determined at 2.0 Å resolution in 1994. The structure reveals limited contact between the two domains and a coordination of Ca2+ not predicted from other C-type lectins. Structure/function analysis indicates a defined region and specific amino-acid side chains that may be involved in ligand binding. The E-selectin bound to sialyl-LewisX (SLeX; NeuNAcα2,3Galβ1,4[Fucα1,3]GlcNAc) tetrasaccharide was solved in 2000. Gene and regulation In humans, E-selectin is encoded by the SELE gene. Its C-type lectin domain, EGF-like, SCR repeats, and transmembrane domains are each encoded by separate exons, whereas the E-selectin cytosolic domain derives from two exons. The E-selectin locus flanks the L-selectin locus on chromosome 1. Different from P-selectin, which is stored in vesicles called Weibel-Palade bodies, E-selectin is not stored in the cell and has to be transcribed, translated, and transported to the cell surface. The production of E-selectin is stimulated by the expression of P-selectin which in turn, is stimulated by tumor necrosis factor α (TNFα), interleukin-1 (IL-1) and lipopolysaccharide (LPS). It takes about two hours, after cytokine recognition, for E-selectin to be expressed on the endothelial cell's surface. Maximal expression of E-selectin
https://en.wikipedia.org/wiki/Secure%20Computing%20Corporation
Secure Computing Corporation (SCC) was a public company that developed and sold computer security appliances and hosted services to protect users and data. McAfee acquired the company in 2008. The company also developed filtering systems used by governments such as Iran and Saudi Arabia that blocks their citizens from accessing information on the Internet. Company history In 1984, a research group called the Secure Computing Technology Center (SCTC) was formed at Honeywell in Minneapolis, Minnesota. The centerpiece of SCTC was its work on security-evaluated operating systems for the NSA. This work included the Secure Ada Target (SAT) and the Logical Coprocessing Kernel (LOCK), both designed to meet the stringent A1 level of the Trusted Computer Systems Evaluation Criteria (TCSEC). Over the next several years, Secure Computing morphed from a small defense contractor into a commercial product vendor, largely because the investment community was much less interested in purchasing security goods from defense contractors than from commercial product vendors, especially vendors in the growing Internet space. Secure Computing became a publicly traded company in 1995. Following the pattern of other Internet-related startups, the stock price tripled its first day: it opened at $16 a share and closed at $48. The price peaked around $64 in the next several weeks and then collapsed over the following year or so. It ranged between roughly $3 and $20 afterward until the company was purchased by McAfee. The company headquarters were moved to San Jose, California, in 1998, though the bulk of the workforce remained in the Twin Cities. The Roseville employees completed a move to St. Paul, Minnesota, in February 2006. Several other sites now exist, largely the result of mergers. Mergers and acquisitions Secure Computing consisted of several merged units, one of the oldest being Enigma Logic, Inc., which was started around 1982. Bob Bosen, the founder, claims to have created t
https://en.wikipedia.org/wiki/List%20of%20symbolic%20stars
This is a list of symbolic uses of "star" ideograms. Star (classification), a scoring system for hotels, restaurants and movies Star (football badge), representing trophies won by a football team Barnstar, a decorative painted object or image often used to adorn a barn Brunswick star, an eight- or sixteen-pointed star surrounding the British Royal Cypher, used on police badges Hex sign, a form of Pennsylvania Dutch folk art Mullet (heraldry), unconventional shapes of stars on coats-of-arms Nautical star, a popular tattoo design Red star, a political symbol of communism and socialism Star of Life, representing emergency medical services units and personnel Geometry Star polygon, a star drawn with a number of lines equal to the number of points Pentagram, a five-pointed star polygon Five-pointed star, a pentagram with internal line segments removed Lute of Pythagoras, a pentagram-based fractal pattern Hexagram, a six-pointed star polygon Heptagram, a seven-pointed star polygon Octagram, an eight-pointed star polygon Enneagram, a nine-pointed star polygon Decagram, a ten-pointed star polygon Hendecagram, an eleven-pointed star polygon Dodecagram, a twelve-pointed star polygon Magic star, a star polygon in which numbers can be placed at each of the vertices and intersections, such that the four numbers on each line sum to the same "magic" constant Typography Star (glyph), any of a number of star-shaped glyphs in typography Asterisk, a typographical symbol (*) Arabic star, a typographical symbol developed to be distinct from the asterisk Medals and awards 1-, 2-, 3-, 4-, or 5-star rank, officer ranks used in many armed services, as well as the rare 6-star rank. Africa Star, awarded by the British Commonwealth for service in World War II. Award star, issued by the United States military for meritorious action in combat. Bronze Star Medal, a United States Armed Forces individual military decoration. Gold star, the highest state decoration in the Soviet Union and sever
https://en.wikipedia.org/wiki/Ecotin
In molecular biology, ecotin is a protease inhibitor which belongs to MEROPS inhibitor family I11, clan IN. Ecotins are dimeric periplasmic proteins from Escherichia coli and related Gram-negative bacteria that have been shown to be potent inhibitors of many trypsin-fold serine proteases of widely varying substrate specificity, which belong to MEROPS peptidase family S1. Phylogenetic analysis suggested that ecotin has an exogenous target, possibly neutrophil elastase. Ecotin from E. coli, Yersinia pestis, and Pseudomonas aeruginosa, all species that encounter the mammalian immune system, inhibit neutrophil elastase strongly while ecotin from the plant pathogen Pantoea citrea inhibits neutrophil elastase 1000-fold less potently. Ecotins all potently inhibit pancreatic digestive peptidases trypsin and chymotrypsin, while showing more variable inhibition of the blood peptidases Factor Xa, thrombin, and urokinase-type plasminogen activator. External links MEROPS entry for ecotin MEROPS peptidase family S1
https://en.wikipedia.org/wiki/Geodetic%20control%20network
A geodetic control network (also geodetic network, reference network, control point network, or control network) is a network, often of triangles, which are measured precisely by techniques of control surveying, such as terrestrial surveying or satellite geodesy. A geodetic control network consists of stable, identifiable points with published datum values derived from observations that tie the points together. Classically, a control is divided into horizontal (X-Y) and vertical (Z) controls (components of the control), however with the advent of satellite navigation systems, GPS in particular, this division is becoming obsolete. Many organizations contribute information to the geodetic control network. The higher-order (high precision, usually millimeter-to-decimeter on a scale of continents) control points are normally defined in both space and time using global or space techniques, and are used for "lower-order" points to be tied into. The lower-order control points are normally used for engineering, construction and navigation. The scientific discipline that deals with the establishing of coordinates of points in a control network is called geodesy. Cartography applications After a cartographer registers key points in a digital map to the real world coordinates of those points on the ground, the map is then said to be "in control". Having a base map and other data in geodetic control means that they will overlay correctly. When map layers are not in control, it requires extra work to adjust them to line up, which introduces additional error. Those real world coordinates are generally in some particular map projection, unit, and geodetic datum. Measurement techniques Terrestrial techniques Triangulation In "classical geodesy" (up to the sixties) control networks were established by triangulation using measurements of angles and of some spare distances. The precise orientation to the geographic north is achieved through methods of geodetic astronomy.
https://en.wikipedia.org/wiki/Ionic%20strength
The ionic strength of a solution is a measure of the concentration of ions in that solution. Ionic compounds, when dissolved in water, dissociate into ions. The total electrolyte concentration in solution will affect important properties such as the dissociation constant or the solubility of different salts. One of the main characteristics of a solution with dissolved ions is the ionic strength. Ionic strength can be molar (mol/L solution) or molal (mol/kg solvent) and to avoid confusion the units should be stated explicitly. The concept of ionic strength was first introduced by Lewis and Randall in 1921 while describing the activity coefficients of strong electrolytes. Quantifying ionic strength The molar ionic strength, I, of a solution is a function of the concentration of all ions present in that solution. where one half is because we are including both cations and anions, ci is the molar concentration of ion i (M, mol/L), zi is the charge number of that ion, and the sum is taken over all ions in the solution. For a 1:1 electrolyte such as sodium chloride, where each ion is singly-charged, the ionic strength is equal to the concentration. For the electrolyte MgSO4, however, each ion is doubly-charged, leading to an ionic strength that is four times higher than an equivalent concentration of sodium chloride: Generally multivalent ions contribute strongly to the ionic strength. Calculation example As a more complex example, the ionic strength of a mixed solution 0.050 M in Na2SO4 and 0.020 M in KCl is: Non-ideal solutions Because in non-ideal solutions volumes are no longer strictly additive it is often preferable to work with molality b (mol/kg of H2O) rather than molarity c (mol/L). In that case, molal ionic strength is defined as: in which i = ion identification number z = charge of ion b = molality (mol solute per Kg solvent) Importance The ionic strength plays a central role in the Debye–Hückel theory that describes the strong deviations from id
https://en.wikipedia.org/wiki/V%C3%A1mos%20matroid
In mathematics, the Vámos matroid or Vámos cube is a matroid over a set of eight elements that cannot be represented as a matrix over any field. It is named after English mathematician Peter Vámos, who first described it in an unpublished manuscript in 1968. Definition The Vámos matroid has eight elements, which may be thought of as the eight vertices of a cube or cuboid. The matroid has rank 4: all sets of three or fewer elements are independent, and 65 of the 70 possible sets of four elements are also independent. The five exceptions are four-element circuits in the matroid. Four of these five circuits are formed by faces of the cuboid (omitting two opposite faces). The fifth circuit connects two opposite edges of the cuboid, each of which is shared by two of the chosen four faces. Another way of describing the same structure is that it has two elements for each vertex of the diamond graph, and a four-element circuit for each edge of the diamond graph. Properties The Vámos matroid is a paving matroid, meaning that all of its circuits have size at least equal to its rank. The Vámos matroid is isomorphic to its dual matroid, but it is not identically self-dual (the isomorphism requires a nontrivial permutation of the matroid elements). The Vámos matroid cannot be represented over any field. That is, it is not possible to find a vector space, and a system of eight vectors within that space, such that the matroid of linear independence of these vectors is isomorphic to the Vámos matroid. Indeed, it is one of the smallest non-representable matroids, and served as a counterexample to a conjecture of Ingleton that the matroids on eight or fewer elements were all representable. The Vámos matroid is a forbidden minor for the matroids representable over a field , whenever has five or more elements. However, it is not possible to test in polynomial time whether it is a minor of a given matroid , given access to through an independence oracle. The Vámos matroid is not al
https://en.wikipedia.org/wiki/Charles%20Maitland%20%28physician%29
Charles Maitland (–1748) was a Scottish surgeon who inoculated people against smallpox. Career In March 1718, Lady Mary Wortley Montagu had Maitland, who was then serving in the British embassy in Constantinople, Turkey, inoculate her five-year-old son Edward. The process was done by an elderly Greek woman from Pera under Maitland's direction. Montague did not tell her husband until a week after when it proved to be successful. They returned to London in April 1721, when Montagu requested that her daughter Mary, who was four, be inoculated. Maitland reluctantly agreed if there were other witnesses present, so three physicians from the Royal College of Physicians were there for the procedure, the first professional inoculation in England. One of the witnesses, James Keith, was so pleased by the success that he had Maitland inoculate his six-year-old son; Keith's other children had all died of smallpox. The Montague family promoted inoculation in England, calling it a "useful invention". On 9 August 1721, Maitland received a Royal Licence that allowed him to test variolation on six prisoners from Newgate Prison. The experiment took place in August 1722, under the direction of Sir Hans Sloane. All prisoners survived, and they were pardoned later that year. One prisoner who was exposed to the disease proved to be immune. In late 1722, Caroline of Ansbach ordered the inoculation of five orphans of St. James's Parish in London. Following their success, Caroline had Maitland inoculate her eldest son, Frederick and one other child. The surgeon died in Aberdeen, Scotland in 1748. See also List of surgeons
https://en.wikipedia.org/wiki/Fujitsu%20FR
The Fujitsu FR (Fujitsu RISC) is a 32-bit RISC processor family. New variants include a floating point unit and partly video input analog-to-digital converter and digital signal processor. It is supported by Softune, GNU Compiler Collection and other integrated development environments. Applications Fujitsu FR are used to control previous versions of Milbeaut signal processors specialized for image processing. Although variants of the 6th generation in 2011 and later generations changed to dual-core ARM architecture, ASSP/ASIC variants with FR controller are continued. They are also used as processor cores inside versions 1 to 3 of the Nikon Expeed image processors (versions 3A and 4 have moved to ARM CPUs). See also SPARClite FR-V (microprocessor)
https://en.wikipedia.org/wiki/Comparison%20of%20Unicode%20encodings
This article compares Unicode encodings. Two situations are considered: 8-bit-clean environments (which can be assumed), and environments that forbid use of byte values that have the high bit set. Originally such prohibitions were to allow for links that used only seven data bits, but they remain in some standards and so some standard-conforming software must generate messages that comply with the restrictions. Standard Compression Scheme for Unicode and Binary Ordered Compression for Unicode are excluded from the comparison tables because it is difficult to simply quantify their size. Compatibility issues A UTF-8 file that contains only ASCII characters is identical to an ASCII file. Legacy programs can generally handle UTF-8 encoded files, even if they contain non-ASCII characters. For instance, the C printf function can print a UTF-8 string, as it only looks for the ASCII '%' character to define a formatting string, and prints all other bytes unchanged, thus non-ASCII characters will be output unchanged. UTF-16 and UTF-32 are incompatible with ASCII files, and thus require Unicode-aware programs to display, print and manipulate them, even if the file is known to contain only characters in the ASCII subset. Because they contain many zero bytes, the strings cannot be manipulated by normal null-terminated string handling for even simple operations such as copy. Therefore, even on most UTF-16 systems such as Windows and Java, UTF-16 text files are not common; older 8-bit encodings such as ASCII or ISO-8859-1 are still used, forgoing Unicode support; or UTF-8 is used for Unicode. One rare counter-example is the "strings" file used by macOS (Mac OS X 10.3 Panther and later) applications for lookup of internationalized versions of messages which defaults to UTF-16, with "files encoded using UTF-8 ... not guaranteed to work." XML is, by convention, encoded as UTF-8, and all XML processors must at least support UTF-8 (including US-ASCII by definition) and UTF-16. Ef
https://en.wikipedia.org/wiki/Baked%20beans
Baked beans is a dish traditionally containing white beans that are parboiled and then, in the US, baked in sauce at low temperature for a lengthy period. In the United Kingdom, the dish is sometimes baked, but usually stewed in sauce. Canned baked beans are not baked, but are cooked through a steam process. Baked beans occurred in Native American cuisine, and are made from beans indigenous to the Americas. It is thought that the dish was adopted and adapted by English colonists in New England in the 17th century and, through cookbooks published in the 19th century, spread to other regions of the United States and into Canada. However, the connection to Native American cuisine may be apocryphal, as legumes such as broad beans and lentils prepared in various sauces had been established in European cuisine long before the Middle Ages. Today, in the New England region of the United States, a variety of indigenous legumes are used in restaurants or in the home, such as Jacob's cattle, soldier beans, yellow-eyed beans, and navy beans (also known as native beans). Originally, Native Americans sweetened baked beans with maple syrup, a tradition some recipes still follow, but some English colonists used brown sugar beginning in the 17th century. In the 18th century, the convention of using American-made molasses as a sweetening agent became increasingly popular to avoid British taxes on sugar. Boston baked beans use a sauce prepared with molasses and salt pork, a dish whose popularity has given Boston the nickname "Beantown". Today, baked beans are served throughout the United States alongside barbecue foods and at picnics. Beans in a brown sugar, sugar, or corn syrup sauce (with or without tomatoes) are widely available throughout the United States. Bush Brothers are the largest producer. After the American Revolutionary War, Independence Day celebrations often included baked beans. Canned baked beans are used as a convenience food; most are made from haricot beans in
https://en.wikipedia.org/wiki/Room%20for%20PlayStation%20Portable
Room for PlayStation Portable (officially styled R∞M — that is, with the initial and final letters in capital case, and an intervening pair of (lowercase) letters O suggested by an intervening infinity symbol. also marketed as PlayStation Room) was a community-based social networking service that halted development on April 15, 2010. It was being developed by Sony Computer Entertainment Japan and beta tested in Japan, from October 2009 to April 2010 for the PlayStation Portable on the PlayStation Network. The service was similar to the PlayStation 3's service, PlayStation Home. Room was first announced at TGS 2009. It could be launched directly from the PlayStation Network section of the PSP's XrossMediaBar after it had been downloaded and installed onto the PSP system. Just as with Home, PSP owners would have been able to invite other PSP owners into their Simulated rooms (called "my rooms") to "enjoy real time communication." The service would have been free, with additional premium content available at launch. It was to be available for all models of the PSP; Sony announced that it would allow users to create 3-D avatars, room spaces, blogs, albums, and chat support. A closed beta test was conducted from October 2009 up to April 15, 2010, with selected PSP owners receiving an invitation to participate, and further details. PSP owners in Japan were able to sign up from September 24 to October 5 of 2009, on the Japanese website, requiring access to the internet from their PSP, age of 18 years or more, and a PlayStation Network master account. Development of the service was halted on April 15, 2010, in light of the feedback from the participants, with SCE Japan thanking them for their feedback. See also Avatars Xbox 360 Avatars | Wii Miis Avatar worlds PlayStation Home | Second Life | Free Realms | IMVU | OLIVE | Active Worlds
https://en.wikipedia.org/wiki/Jungian%20neuroscience
The theories of Carl Jung are grounded in his evolutionary conception of human brain evolution. This had led to a resurgence of research into his work, beginning in the early 2000s, from the perspective of contemporary neuroscience. Much of this work looks at Jung's theories of a genetically inherited 'collective unconscious' common to all of humankind. This hypothesis was postulated by Jung in his efforts to account for similar patterns of behaviour and symbolic expression in myth, dream imagery and religion in various cultures around the world. Jung believed that the 'collective unconscious' was structured by archetypes - that is species typical patterns of behaviour and cognition common to all humans. Contemporary researchers have postulated such recurrent archetypes reside in 'environmentally closed' subcortical brain systems that evolved in the human lineage prior to the emergence of self-consciousness and the uniquely human self-reflective ego. Jung's theories of brain evolution In a lecture given in 1957 at the Second International Conference for Psychiatry in Zurich, Jung wrote: Throughout his career Jung developed this notion and applied it frequently in his analysis mythic and cultural symbolism as well as clinical treatments. One of his main arguments was that this ancient genetically inherited aspect of the brain was dissociated from more recently evolved regions associated with the human ego complex. For him the goal of therapy was to integrate this more ancient emotionally based subcortical brain system with the more recently evolved ego complex and its capacities for self-conscious awareness and rational insight. It was such integration that formed the basis of his theory of individuation