source
stringlengths 31
227
| text
stringlengths 9
2k
|
|---|---|
https://en.wikipedia.org/wiki/McCutcheon%20index
|
The McCutcheon index or chemotactic ratio is a numerical metric that quantifies the efficiency of movement. It is calculated as the ratio of the net displacement of a moving entity to the total length of the path it has traveled.
The index acts as an evaluative measure of the directness of movement. A value close to 1 indicates that a moving entity performed its movement in a very direct manner, minimizing detours. On the other hand, a lower value indicates that the entity has achieved only a marginal net displacement, despite traveling a considerable distance. The index is used to evaluate movements of, for example, leukocytes, bacteria, or amoebae.
It is named after Morton McCutcheon who introduced it to describe chemotaxis in leukocytes.
|
https://en.wikipedia.org/wiki/Prescription%20drug
|
A prescription drug (also prescription medication, prescription medicine or prescription-only medication) is a pharmaceutical drug that is permitted to be dispensed only to those with a medical prescription. In contrast, over-the-counter drugs can be obtained without a prescription. The reason for this difference in substance control is the potential scope of misuse, from drug abuse to practicing medicine without a license and without sufficient education. Different jurisdictions have different definitions of what constitutes a prescription drug.
In North America, , usually printed as "Rx", is used as an abbreviation of the word "prescription". It is a contraction of the Latin word "recipe" (an imperative form of "recipere") meaning "take". Prescription drugs are often dispensed together with a monograph (in Europe, a Patient Information Leaflet or PIL) that gives detailed information about the drug.
The use of prescription drugs has been increasing since the 1960s.
Regulation
Australia
In Australia, the Standard for the Uniform Scheduling of Medicines and Poisons (SUSMP) governs the manufacture and supply of drugs with several categories:
Schedule 1 – Defunct Drug.
Schedule 2 – Pharmacy Medicine
Schedule 3 – Pharmacist-Only Medicine
Schedule 4 – Prescription-Only Medicine/Prescription Animal Remedy
Schedule 5 – Caution/Poison.
Schedule 6 – Poison
Schedule 7 – Dangerous Poison
Schedule 8 – Controlled Drug (Possession without authority illegal)
Schedule 9 – Prohibited Substance (Possession illegal without a license legal only for research purposes)
Schedule 10 – Controlled Poison.
Unscheduled Substances.
As in other developed countries, the person requiring a prescription drug attends the clinic of a qualified health practitioner, such as a physician, who may write the prescription for the required drug.
Many prescriptions issued by health practitioners in Australia are covered by the Pharmaceutical Benefits Scheme, a scheme that provides subsidised
|
https://en.wikipedia.org/wiki/Mach%203%20%281987%20video%20game%29
|
Mach 3 is a 1987 3D shooter video game by Loriciels for Amiga, Amstrad CPC, Atari ST, MSX, Thomson TO7, ZX Spectrum and DOS. The DOS (PC) version uses CGA 320x200 video mode.
Gameplay
The player controls a spacecraft and shoots various enemy crafts while avoiding mines and obstacles. During the intro screen, a sampled is phrase is spoken: "Get ready for Mach 3."
Reviews
http://www.atarimania.com/game-atari-st-mach-iii_9879.html
https://worldofspectrum.org/archive/magazines/your-sinclair/64/0/1991/4/0#10
http://amr.abime.net/review_12918
Jeux & Stratégie #49
|
https://en.wikipedia.org/wiki/Image%20%28mathematics%29
|
In mathematics, the image of a function is the set of all output values it may produce.
More generally, evaluating a given function at each element of a given subset of its domain produces a set, called the "image of under (or through) ". Similarly, the inverse image (or preimage) of a given subset of the codomain of is the set of all elements of the domain that map to the members of
Image and inverse image may also be defined for general binary relations, not just functions.
Definition
The word "image" is used in three related ways. In these definitions, is a function from the set to the set
Image of an element
If is a member of then the image of under denoted is the value of when applied to is alternatively known as the output of for argument
Given the function is said to "" or "" if there exists some in the function's domain such that
Similarly, given a set is said to "" if there exists in the function's domain such that
However, "" and "" means that for point in 's domain.
Image of a subset
Throughout, let be a function.
The under of a subset of is the set of all for It is denoted by or by when there is no risk of confusion. Using set-builder notation, this definition can be written as
This induces a function where denotes the power set of a set that is the set of all subsets of See below for more.
Image of a function
The image of a function is the image of its entire domain, also known as the range of the function. This last usage should be avoided because the word "range" is also commonly used to mean the codomain of
Generalization to binary relations
If is an arbitrary binary relation on then the set is called the image, or the range, of Dually, the set is called the domain of
Inverse image
Let be a function from to The preimage or inverse image of a set under denoted by is the subset of defined by
Other notations include and
The inverse image of a singleton set, denoted by or by
|
https://en.wikipedia.org/wiki/Colorimeter%20%28chemistry%29
|
A colorimeter is a device used in colorimetry that measures the absorbance of particular wavelengths of light by a specific solution. It is commonly used to determine the concentration of a known solute in a given solution by the application of the Beer–Lambert law, which states that the concentration of a solute is proportional to the absorbance.
Construction
The essential parts of a colorimeter are:
a light source (often an ordinary low-voltage filament lamp);
an adjustable aperture;
a set of colored filters;
a cuvette to hold the working solution;
a detector (usually a photoresistor) to measure the transmitted light;
a meter to display the output from the detector.
In addition, there may be:
a voltage regulator, to protect the instrument from fluctuations in mains voltage;
a second light path, cuvette and detector. This enables comparison between the working solution and a "blank", consisting of pure solvent, to improve accuracy.
There are many commercialized colorimeters as well as open source versions with construction documentation for education and for research.
Filters
Changeable optics filters are used in the colorimeter to select the wavelength which the solute absorbs the most, in order to maximize accuracy. The usual wavelength range is from 400 to 700 nm. If it is necessary to operate in the ultraviolet range then some modifications to the colorimeter are needed. In modern colorimeters the filament lamp and filters may be replaced by several (light-emitting diode) of different colors.
Cuvettes
In a manual colorimeter the cuvettes are inserted and removed by hand. An automated colorimeter (as used in an AutoAnalyzer) is fitted with a flowcell through which solution flows continuously.
Output
The output from a colorimeter may be displayed by an analogue or digital meter and may be shown as transmittance (a linear scale from 0 to 100%) or as absorbance (a logarithmic scale from zero to infinity). The useful range of the absorbance scale is
|
https://en.wikipedia.org/wiki/Regular%20chain
|
In mathematics, and more specifically in computer algebra and elimination theory, a regular chain is a particular kind of triangular set of multivariate polynomials over a field, where a triangular set is a finite sequence of polynomials such that each one contains at least one more indeterminate than the preceding one. The condition that a triangular set must satisfy to be a regular chain is that, for every , every common zero (in an algebraically closed field) of the first polynomials may be prolongated to a common zero of the th polynomial. In other words, regular chains allow solving systems of polynomial equations by solving successive univariate equations without considering different cases.
Regular chains enhance the notion of Wu's characteristic sets in the sense that they provide a better result with a similar method of computation.
Introduction
Given a linear system, one can convert it to a triangular system via Gaussian elimination. For the non-linear case, given a polynomial system F over a field, one can convert (decompose or triangularize) it to a finite set of triangular sets, in the sense that the algebraic variety V(F) is described by these triangular sets.
A triangular set may merely describe the empty set. To fix this degenerated case, the notion of regular chain was introduced, independently by Kalkbrener (1993), Yang and Zhang (1994). Regular chains also appear in Chou and Gao (1992). Regular chains are special triangular sets which are used in different algorithms for computing unmixed-dimensional decompositions of algebraic varieties. Without using factorization, these decompositions have better properties that the ones produced by Wu's algorithm. Kalkbrener's original definition was based on the following observation: every irreducible variety is uniquely determined by one of its generic points and varieties can be represented by describing the generic points of their irreducible components. These generic points are given by regular cha
|
https://en.wikipedia.org/wiki/Quantum%20bus
|
A quantum bus is a device which can be used to store or transfer information between independent qubits in a quantum computer, or combine two qubits into a superposition. It is the quantum analog of a classical bus.
There are several physical systems that can be used to realize a quantum bus, including trapped ions, photons, and superconducting qubits. Trapped ions, for example, can use the quantized motion of ions (phonons) as a quantum bus, while photons can act as a carrier of quantum information by utilizing the increased interaction strength provided by cavity quantum electrodynamics. Circuit quantum electrodynamics, which uses superconducting qubits coupled to a microwave cavity on a chip, is another example of a quantum bus that has been successfully demonstrated in experiments.
History
The concept was first demonstrated by researchers at Yale University and the National Institute of Standards and Technology (NIST) in 2007. Prior to this experimental demonstration, the quantum bus had been described by scientists at NIST as one of the possible cornerstone building blocks in quantum computing architectures.
Mathematical description
A quantum bus for superconducting qubits can be built with a resonance cavity. The hamiltonian for a system with qubit A, qubit B, and the resonance cavity or quantum bus connecting the two is where is the single qubit hamiltonian, is the raising or lowering operator for creating or destroying excitations in the th qubit, and is controlled by the amplitude of the D.C. and radio frequency flux bias.
|
https://en.wikipedia.org/wiki/Cecropin
|
Cecropins are antimicrobial peptides. They were first isolated from the hemolymph of Hyalophora cecropia, whence the term cecropin was derived. Cecropins lyse bacterial cell membranes; they also inhibit proline uptake and cause leaky membranes.
Cecropins constitute a main part of the innate immune system of insects. Cecropins are small proteins anywhere from 31 - 37 amino acids long and are active against both gram-positive and gram-negative bacteria. Cecropins isolated from insects other than Hyalophora cecropia (Cecropia moth) have been given various names, such as bactericidin, , and sarcotoxin. All of these peptides are structurally related.
Members
Members include:
Cecropin A Peptide Sequence (KWKLFKKIEKVGQNIRDGIIKAGPAVAVVGQATQIAK). Secondary structure includes two α helices. At low peptide to lipid ratios ion channels are formed, at high peptide to lipid ratios pores are formed.
Cecropin B Peptide Sequence (KWKVFKKIEKMGRNIRNGIVKAGPAIAVLGEAKAL). Secondary structure includes two α helices.
CECD from Aedes aegypti (Yellowfever mosquito).
Papiliocin from Papilio xuthus (a butterfly)
Cecropin P1 Peptide Sequence (SWLSKTAKKLENSAKKRISEGIAIAIQGGPR). An antibacterial peptide from Ascaris suum, a parasitic nematode that resides in the pig intestine, also belongs to this family.
Derivatives
A derivative of Cecropin B is an anticancer polypeptide(L). Structure consists of mainly alpha helixes, determined by solution NMR. Protein molecular weight = 4203.4g/mol.
Some of the cecropins (e.g. cecropin A, and cecropin B) have anticancer properties and are called anticancer peptides (ACPs).
Hybrid ACPs based on Cecropin A have been studied for anticancer properties.
Anticancer properties
Anticancer activities of cecropin B, cecropin P1, and Shiva-1 were first demonstrated with in vitro studies of mammalian leukemia and lymphoma cell lines, where cells were sensitive to peptide concentrations on the order of 10−6 M. Two multidrug-resistant breast and ovarian cancer c
|
https://en.wikipedia.org/wiki/Glass%20cockpit
|
A glass cockpit is an aircraft cockpit that features electronic (digital) flight instrument displays, typically large LCD screens, rather than the traditional style of analog dials and gauges. While a traditional cockpit relies on numerous mechanical gauges (nicknamed "steam gauges") to display information, a glass cockpit uses several multi-function displays driven by flight management systems, that can be adjusted to display flight information as needed. This simplifies aircraft operation and navigation and allows pilots to focus only on the most pertinent information. They are also popular with airline companies as they usually eliminate the need for a flight engineer, saving costs. In recent years the technology has also become widely available in small aircraft.
As aircraft displays have modernized, the sensors that feed them have modernized as well. Traditional gyroscopic flight instruments have been replaced by electronic attitude and heading reference systems (AHRS) and air data computers (ADCs), improving reliability and reducing cost and maintenance. GPS receivers are usually integrated into glass cockpits.
Early glass cockpits, found in the McDonnell Douglas MD-80, Boeing 737 Classic, ATR 42, ATR 72 and in the Airbus A300-600 and A310, used electronic flight instrument systems (EFIS) to display attitude and navigational information only, with traditional mechanical gauges retained for airspeed, altitude, vertical speed, and engine performance. The Boeing 757 and 767-200/-300 introduced an electronic engine-indicating and crew-alerting system (EICAS) for monitoring engine performance while retaining mechanical gauges for airspeed, altitude and vertical speed.
Later glass cockpits, found in the Boeing 737NG, 747-400, 767-400, 777, Airbus A320, later Airbuses, Ilyushin Il-96 and Tupolev Tu-204 have completely replaced the mechanical gauges and warning lights in previous generations of aircraft. While glass cockpit-equipped aircraft throughout the late 20t
|
https://en.wikipedia.org/wiki/Computational%20Infrastructure%20for%20Geodynamics
|
The Computational Infrastructure for Geodynamics (CIG) is a community-driven organization that advances Earth science by developing and disseminating software for geophysics and related fields. It is a National Science Foundation-sponsored collaborative effort to improve geodynamic modelling and develop, support, and disseminate open-source software for the geodynamics research and higher education communities.
CIG is located at the University of California, Davis, and is a member-governed consortium with 62 US institutional members and 15 international affiliates.
History
CIG was established in 2005 in response to the need for coordinated development and dissemination of software for geodynamics applications. Founded with an NSF cooperative agreement to Caltech, in 2010, CIG moved to UC Davis under a new cooperative agreement from NSF.
Software
CIG hosts open source software in a wide range of disciplines and topic areas, such as geodynamics, computational science, seismology, mantle convection, long-term tectonics, and short-term crustal dynamics.
Software Attribution for Geoscience Applications (SAGA)
CIG started the SAGA project with an NSF EAGER award from the SBE Office of Multidisciplinary Activities for "Development of Software Citation Methodology for Open Source Computational Science".
|
https://en.wikipedia.org/wiki/Taylor%E2%80%93Green%20vortex
|
In fluid dynamics, the Taylor–Green vortex is an unsteady flow of a decaying vortex, which has an exact closed form solution of the incompressible Navier–Stokes equations in Cartesian coordinates. It is named after the British physicist and mathematician Geoffrey Ingram Taylor and his collaborator A. E. Green.
Original work
In the original work of Taylor and Green, a particular flow is analyzed in three spatial dimensions, with the three velocity components at time specified by
The continuity equation determines that . The small time behavior of the flow is then found through simplification of the incompressible Navier–Stokes equations using the initial flow to give a step-by-step solution as time progresses.
An exact solution in two spatial dimensions is known, and is presented below.
Incompressible Navier–Stokes equations
The incompressible Navier–Stokes equations in the absence of body force, and in two spatial dimensions, are given by
The first of the above equation represents the continuity equation and the other two represent the momentum equations.
Taylor–Green vortex solution
In the domain , the solution is given by
where , being the kinematic viscosity of the fluid. Following the analysis of Taylor and Green for the two-dimensional situation, and for , gives agreement with this exact solution, if the exponential is expanded as a Taylor series, i.e. .
The pressure field can be obtained by substituting the velocity solution in the momentum equations and is given by
The stream function of the Taylor–Green vortex solution, i.e. which satisfies for flow velocity , is
Similarly, the vorticity, which satisfies , is given by
The Taylor–Green vortex solution may be used for testing and validation of temporal accuracy of Navier–Stokes algorithms.
A generalization of the Taylor–Green vortex solution in three dimensions in described in.
|
https://en.wikipedia.org/wiki/Outer%20nuclear%20layer
|
The outer nuclear layer (or layer of outer granules or external nuclear layer), is one of the layers of the vertebrate retina, the light-detecting portion of the eye. Like the inner nuclear layer, the outer nuclear layer contains several strata of oval nuclear bodies; they are of two kinds, viz.: rod and cone granules, so named on account of their being respectively connected with the rods and cones of the next layer.
Rod granules
The spherical rod granules are much more numerous, and are placed at different levels throughout the layer.
Their nuclei present a peculiar cross-striped appearance, and prolonged from either extremity of each cell is a fine process; the outer process is continuous with a single rod of the layer of rods and cones; the inner ends in the outer plexiform layer in an enlarged extremity, and is imbedded in the tuft into which the outer processes of the rod bipolar cells break up.
In its course it presents numerous varicosities.
Cone granules
The stem-like cone granules, fewer in number than the rod granules, are placed close to the membrana limitans externa, through which they are continuous with the cones of the layer of rods and cones.
They do not present any cross-striation, but contain a pyriform nucleus, which almost completely fills the cell.
From the inner extremity of the granule a thick process passes into the outer plexiform layer, and there expands into a pyramidal enlargement or foot plate, from which are given off numerous fine fibrils, that come in contact with the outer processes of the cone bipolars.
|
https://en.wikipedia.org/wiki/Capability%20curve
|
Capability curve of an electrical generator describes the limits of the active (MW) and reactive power (MVAr) that the generator can provide. The curve represents a boundary of all operating points in the MW/MVAr plane; it is typically drawn with the real power on the horizontal axis, and, for the synchronous generator, resembles a letter D in shape, thus another name for the same curve, D-curve. In some sources the axes are switched, and the curve gets a dome-shaped appearance.
Synchronous generators
For a traditional synchronous generator the curve consists of multiple segments, each due to some physical constraint:
at the right part of the curve (close to the rated voltage), the generator is constrained by the heat dissipation in the armature (stator for large generators). The heating is proportional to the sum of squares of active and reactive currents, at the near-constant voltage it is closely proportional to the sum of squares of MW and MVAr, therefore this part of the curve (armature heating limit) resembles a section of a semicircle with the center at (0,0);
at the upper part of the curve (generator produces a lot of reactive power) operation requires higher voltage on the output of the generator and thus higher excitation field. The rotating excitation winding has its own field heating limit;
at the bottom of the curve (generator absorbs a lot of reactive power) the magnetic flux constraints in the stator cause heating of the magnetic core at the stator end (core end heating limit).
The corners between the sections of the curve define the limits of the power factor (PF) that the generator can sustain at its nameplate capacity (the illustration has the PF ticks placed at 0.85 lagging and 0.95 leading angles). In practice, the prime mover (a power source that drives the generator) is designed for less active power than the generator is capable of (due to the fact that in real life generator always has to deliver some reactive power), so a prime move
|
https://en.wikipedia.org/wiki/Bovista
|
Bovista is a genus of fungi commonly known as the true puffballs. It was formerly classified within the now-obsolete order Lycoperdales, which, following a restructuring of fungal taxonomy brought about by molecular phylogeny, has been split; the species of Bovista are now placed in the family Agaricaceae of the order Agaricales. Bovista species have a collectively widespread distribution, and are found largely in temperate regions of the world. Various species have historically been used in homeopathic preparations.
Description
Fruit bodies are oval to spherical to pear-shaped, and typically in diameter with a white or light-colored thin and fragile exoperidium (outer layer of the peridium). Depending on the species, the exoperidium in a young specimen may be smooth, granular, or finely echinulate. This exoperidium sloughs off at maturity to expose a smooth endoperidium with a single apical pore (ostiole). The fruit bodies may be attached to the ground by fine rhizomorphs that may appear like a small cord. Some species develop a subgleba—a sterile base that is typically not well developed. The fruit bodies of mature specimens can develop surface alterations such as scales, plates, areolae, or verrucae. At the microscopic level, these features are made of hyphae, sphaerocysts (rounded cells), claviform (club-shaped) cells. Bovista sclerocystis is the only species in the genus with mycosclereids (setoid elements) in the peridium.
Spores are brown to purple-brown, roughly spherical or ellipsoid in shape, and 3.5–7 μm in diameter. A short or long pedicel (stalk) may be present. At maturity, the entire fruit body may become detached from the ground, and the spores spread as the puffball is blown around like a tumbleweed.
In Bovista, the capillitium (a network of thread-like cells in which the spores are embedded) is not connected directly to the interior wall of the peridium. Instead, it is made of separate, irregularly branched units that end in tapered points. Thi
|
https://en.wikipedia.org/wiki/Sagittaria%20sagittifolia
|
Sagittaria sagittifolia (also called arrowhead because of the shape of its leaves) is a flowering plant in the family Alismataceae, native to wetlands in most of Europe from Ireland and Portugal to Finland and Bulgaria, and in Russia, Ukraine, Siberia, Japan, Turkey, China, India, Australia, Vietnam and the Caucasus. It is also cultivated as a food crop in some other countries. In Britain it is the only native Sagittaria.
Sagittaria sagittifolia is a herbaceous perennial plant, growing in water from 10–50 cm deep. The leaves above water are arrowhead-shaped, the leaf blade 15–25 cm long and 10–22 cm broad, on a long petiole holding the leaf up to 45 cm above water level. The plant also has narrow linear submerged leaves, up to 80 cm long and 2 cm broad. The flowers are 2-2.5 cm broad, with three small sepals and three white petals, and numerous purple stamens.
Cultivation and uses
The round tuber is edible. It tastes bland, with a starchy texture, similar to a potato but somewhat crunchier, even when cooked. In Japan, it is known as () and its tuber is eaten particularly during the New Year. In China, it is known as and often used in winter hot pots. In Vietnam, the plant's young petiole leaves and rhizomes are used for soups.
Remnants of Sagittaria sagittifolia have been found in the Paleolithic/Mesolithic site of Całowanie in Poland.
Sagittaria sagittifolia is used in Chinese medicine, and in 2006 seven new ent-rosane diterpenoids and a new labdane diterpene were purified from the plant. Four of these compounds (Sagittine A–D) exhibited antibacterial activity against Streptococcus mutans and Actinomyces naeslundii while another (Sagittine E) was only active against A. naeslundii (MIC = 62.5 μg ml–1). Recently, the same group identified five new diterpenoids from Sagittaria pygmaea. None displayed activity against A. actinomycetemcomitans, while four of the others were active against A. viscosus and three against S. mutans, of which 18-ß-D-3',6'-diacetoxyg
|
https://en.wikipedia.org/wiki/Nielsen%20theory
|
Nielsen theory is a branch of mathematical research with its origins in topological fixed-point theory. Its central ideas were developed by Danish mathematician Jakob Nielsen, and bear his name.
The theory developed in the study of the so-called minimal number of a map f from a compact space to itself, denoted MF[f]. This is defined as:
where ~ indicates homotopy of mappings, and #Fix(g) indicates the number of fixed points of g. The minimal number was very difficult to compute in Nielsen's time, and remains so today. Nielsen's approach is to group the fixed-point set into classes, which are judged "essential" or "nonessential" according to whether or not they can be "removed" by a homotopy.
Nielsen's original formulation is equivalent to the following:
We define an equivalence relation on the set of fixed points of a self-map f on a space X. We say that x is equivalent to y if and only if there exists a path c from x to y with f(c) homotopic to c as paths. The equivalence classes with respect to this relation are called the Nielsen classes of f, and the Nielsen number N(f) is defined as the number of Nielsen classes having non-zero fixed-point index sum.
Nielsen proved that
making his invariant a good tool for estimating the much more difficult MF[f]. This leads immediately to what is now known as the Nielsen fixed-point theorem: Any map f has at least N(f) fixed points.
Because of its definition in terms of the fixed-point index, the Nielsen number is closely related to the Lefschetz number. Indeed, shortly after Nielsen's initial work, the two invariants were combined into a single "generalized Lefschetz number" (more recently called the Reidemeister trace) by Wecken and Reidemeister.
Bibliography
External links
Survey article on Nielsen theory by Robert F. Brown at Topology Atlas
Fixed-point theorems
Fixed points (mathematics)
Topology
|
https://en.wikipedia.org/wiki/Pseudescherichia
|
Pseudescherichia is a Gram-negative genus of non-spore-forming, facultatively anaerobic rod-shaped bacteria from the family Enterobacteriaceae. Based on conserved signature indels (CSIs) differentiating it from other members of this family, this genus and its sole species P. vulneris were divided from Escherichia, the genus of E. coli, in 2017.
A November 2022 preprint article has identified another member species from a Oryza sativa rice seedling growing in Arkansas, proposing the name P. oryzae.
|
https://en.wikipedia.org/wiki/Wi-Fi%20positioning%20system
|
Wi-Fi positioning system (WPS, also abbreviated as WiPS or WFPS) is a geolocation system that uses the characteristics of nearby Wi-Fi hotspots and other wireless access points to discover where a device is located.
It is used where satellite navigation such as GPS is inadequate due to various causes including multipath and signal blockage indoors, or where acquiring a satellite fix would take too long. Such systems include assisted GPS, urban positioning services through hotspot databases, and indoor positioning systems. Wi-Fi positioning takes advantage of the rapid growth in the early 21st century of wireless access points in urban areas.
The most common and widespread localization technique used for positioning with wireless access points is based on measuring the intensity of the received signal (received signal strength indication or RSSI) and the method of "fingerprinting". Typical parameters useful to geolocate the wireless access point include its SSID and MAC address. The accuracy depends on the number of nearby access points whose positions have been entered into the database. The Wi-Fi hotspot database gets filled by correlating mobile device GPS location data with Wi-Fi hotspot MAC addresses. The possible signal fluctuations that may occur can increase errors and inaccuracies in the path of the user. To minimize fluctuations in the received signal, there are certain techniques that can be applied to filter the noise.
In the case of low precision, some techniques have been proposed to merge the Wi-Fi traces with other data sources such as geographical information and time constraints (i.e., time geography).
Motivation and applications
Accurate indoor localization is becoming more important for Wi-Fi based devices due to the increased use of augmented reality, social networking, health care monitoring, personal tracking, inventory control and other indoor location-aware applications.
In wireless security, it is an important task used to locate and
|
https://en.wikipedia.org/wiki/Maekawa%27s%20theorem
|
Maekawa's theorem is a theorem in the mathematics of paper folding named after Jun Maekawa. It relates to flat-foldable origami crease patterns and states that at every vertex, the numbers of valley and mountain folds always differ by two in either direction. The same result was also discovered by Jacques Justin and, even earlier, by S. Murata.
Parity and coloring
One consequence of Maekawa's theorem is that the total number of folds at each vertex must be an even number. This implies (via a form of planar graph duality between Eulerian graphs and bipartite graphs) that, for any flat-foldable crease pattern, it is always possible to color the regions between the creases with two colors, such that each crease separates regions of differing colors. The same result can also be seen by considering which side of the sheet of paper is uppermost in each region of the folded shape.
Related results
Maekawa's theorem does not completely characterize the flat-foldable vertices, because it only takes into account the numbers of folds of each type, and not their angles.
Kawasaki's theorem gives a complementary condition on the angles between the folds at a vertex (regardless of which folds are mountain folds and which are valley folds) that is also necessary for a vertex to be flat-foldable.
|
https://en.wikipedia.org/wiki/Operation%20Bayonet%20%28darknet%29
|
Operation Bayonet was a multinational law enforcement operation culminating in 2017 targeting the AlphaBay and Hansa darknet markets. Many other darknet markets were also shut down.
|
https://en.wikipedia.org/wiki/Braarudosphaera%20bigelowii
|
Braarudosphaera bigelowii is a coastal coccolithophore in the fossil record going back 100 million years. The family Braarudosphaeraceae are single-celled coastal phytoplanktonic algae with calcareous scales with five-fold symmetry, called pentaliths. With 12 sides, it has a regular dodecahedral structure, approximately 10 micrometers across.
The genus name of Braarudosphaera is in honour of Trygve Braarud (1903–1985), who was a Norwegian botanist. He specialized in marine biology, and was affiliated with the University of Oslo.
|
https://en.wikipedia.org/wiki/Hao%20Huang%20%28mathematician%29
|
Hao Huang is a mathematician known for solving the sensitivity conjecture. Huang is currently an associate professor in the mathematics department at National University of Singapore.
Huang was an assistant professor from 2015 to 2021 in the Department of Mathematics at Emory University. He obtained his Ph.D in mathematics from UCLA in 2012 advised by Benny Sudakov. His postdoctoral research was done at the Institute for Advanced Study in Princeton and DIMACS at Rutgers University in 2012-2014, followed by a year at the Institute for Mathematics and its Applications at University of Minnesota.
In July 2019, Huang announced a breakthrough, which gave a proof of the sensitivity conjecture. At that point the conjecture had been open for nearly 30 years, having been posed by Noam Nisan and Mario Szegedy in 1992.
Theoretical computer scientist Scott Aaronson said of Huang's ingenious two-page proof, "I find it hard to imagine that even God knows how to prove the Sensitivity Conjecture in any simpler way than this."
Huang received an NSF Career Award in 2019 and a Sloan Research Fellowship in 2020.
|
https://en.wikipedia.org/wiki/Years%20of%20potential%20life%20lost
|
Years of potential life lost (YPLL) or potential years of life lost (PYLL) is an estimate of the average years a person would have lived if they had not died prematurely. It is, therefore, a measure of premature mortality. As an alternative to death rates, it is a method that gives more weight to deaths that occur among younger people. An alternative is to consider the effects of both disability and premature death using disability adjusted life years.
Calculation
To calculate the years of potential life lost, the analyst has to set an upper reference age. The reference age should correspond roughly to the life expectancy of the population under study. In the developed world, this is commonly set at age 75, but it is essentially arbitrary. Thus, PYLL should be written with respect to the reference age used in the calculation: e.g., PYLL[75].
PYLL can be calculated using individual level data or using age grouped data.
Briefly, for the individual method, each person's PYLL is calculated by subtracting the person's age at death from the reference age. If a person is older than the reference age when they die, that person's PYLL is set to zero (i.e., there are no "negative" PYLLs). In effect, only those who die before the reference age are included in the calculation. Some examples:
Reference age = 75; Age at death = 60; PYLL[75] = 75 − 60 = 15
Reference age = 75; Age at death = 6 months; PYLL[75] = 75 − 0.5 = 74.5
Reference age = 75; Age at death = 80; PYLL[75] = 0 (age at death greater than reference age)
To calculate the PYLL for a particular population in a particular year, the analyst sums the individual PYLLs for all individuals in that population who died in that year. This can be done for all-cause mortality or for cause-specific mortality.
Significance
In the developed world, mortality counts and rates tend to emphasise the most common causes of death in older people because the risk of death increases with age. Because YPLL gives more weight to d
|
https://en.wikipedia.org/wiki/Beta%20decay%20transition
|
In nuclear physics, a beta decay transition is the change in state of an atomic nucleus undergoing beta decay. (β-decay) When undergoing beta decay, a nucleus emits a beta particle and a corresponding neutrino, transforming the original nuclide into one with the same mass, but differing charge. (an isobar)
There are several types of beta decay transition. In a Fermi transition, the spins of the two emitted particles are anti-parallel, for a combined spin . As a result, the total angular momentum of the nucleus is unchanged by the transition. By contrast, in a Gamow-Teller transition, the spins of the two emitted particles are parallel, with total spin , leading to a change in angular momentum between the initial and final states of the nucleus.
The theoretical work in describing these transitions was done between 1934 and 1936 by George Gamow and Edward Teller at George Washington University.
The weak interaction and beta decay
β decay had been first described theoretically by Fermi's original ansatz which was Lorentz-invariant and involved a 4-point fermion vector current. However, this did not incorporate parity violation within the matrix element in Fermi's Golden Rule seen in weak interactions. The Gamow–Teller theory was necessary for the inclusion of parity violation by modifying the matrix element to include vector and axial-vector couplings of fermions.
This formed the matrix element that completed the Fermi theory of β decay and described parity violation, neutrino helicity, muon decay properties along with the concept of lepton universality. Before the Standard Model of Particle Physics was developed, George Sudarshan and Robert Marshak, and also independently Richard Feynman and Murray Gell-Mann, determined the correct tensor structure (vector minus axial vector, ) of the four-fermion interaction.
From there modern electroweak theory was developed, which described the weak interaction in terms of massive gauge bosons which was required for describi
|
https://en.wikipedia.org/wiki/Web%20development%20tools
|
Web development tools (often called devtools or inspect element) allow web developers to test and debug their source code. They are different from website builders and integrated development environments (IDEs) in that they do not assist in the direct creation of a webpage, rather they are tools used for testing the user interface of a website or web application.
Web development tools come as browser add-ons or built-in features in modern web browsers. Browsers such as Google Chrome, Firefox, Internet Explorer, Safari, Microsoft Edge and Opera, have built-in tools to help web developers, and many additional add-ons can be found in their respective plugin download centers.
Web development tools allow developers to work with a variety of web technologies, including HTML, CSS, the DOM, JavaScript, and other components that are handled by the web browser. Due to increasing demand from web browsers to do more, popular web browsers have included more features geared for developers.
Web developer tools support
Several notable web browsers have support for web developer tools that allow web designers and developers to look at the make-up of their pages. These are all tools that are built into the browser and do not require additional modules or configuration.
Firefox – opens the Web Console / Browser Console (since Firefox 4). The Web Console applies to a single content tab; the Browser Console applies to the whole browser. Many add-ons also exist, including Firebug.
Google Chrome – Chrome Developer Tools (DevTools)
Internet Explorer and Microsoft Edge – opens Web Developer Tools (as of version 8)
Opera – Opera Dragonfly
Safari – Safari Web Development Tools (as of version 3)
Most used features
The built-in web developer tools in the browser are commonly accessed by hovering over an item on a webpage and selecting the "Inspect Element" or similar option from the context menu. Alternatively the key tends to be another common shortcut.
HTML and the DOM
HTML and
|
https://en.wikipedia.org/wiki/Opte%20Project
|
The Opte Project, created in 2003 by Barrett Lyon, seeks to generate an accurate representation of the breadth of the Internet using visual graphics. Lyon believes that his network mapping can help teach students more about the Internet while also acting as a gauge illustrating both overall Internet growth and the specific areas where that growth occurs. It was not the first such project; others predated it, such as the Bell Labs Internet Mapping Project.
Lyon has been generating image maps using traceroute, and later switched to mapping using BGP routes. The generated images were published on the Opte Project website. In 2021, Lyon created different video animations, using his mapping technique: shedding light on internet growth between 1997 and 2021, the Iranian internet shutdown of 2019, the United States Department of Defense's place on the internet as well as the few entry points into the Chinese internet.
The project has gathered notice worldwide having been featured by Time, Cornell University, New Scientist, and Kaspersky Lab. In addition, Opte Project maps have found homes in at least two art galleries and exhibits such as The Museum of Modern Art and the Museum of Science's Mapping the World Around Us permanent exhibit.
Opte images are licensed under a Creative Commons license and while use of The Opte Image is free for all non-commercial applications, a license fee is required for all others.
|
https://en.wikipedia.org/wiki/Plural%20quantification
|
In mathematics and logic, plural quantification is the theory that an individual variable x may take on plural, as well as singular, values. As well as substituting individual objects such as Alice, the number 1, the tallest building in London etc. for x, we may substitute both Alice and Bob, or all the numbers between 0 and 10, or all the buildings in London over 20 stories.
The point of the theory is to give first-order logic the power of set theory, but without any "existential commitment" to such objects as sets. The classic expositions are Boolos 1984 and Lewis 1991.
History
The view is commonly associated with George Boolos, though it is older (see notably Simons 1982), and is related to the view of classes defended by John Stuart Mill and other nominalist philosophers. Mill argued that universals or "classes" are not a peculiar kind of thing, having an objective existence distinct from the individual objects that fall under them, but "is neither more nor less than the individual things in the class". (Mill 1904, II. ii. 2,also I. iv. 3).
A similar position was also discussed by Bertrand Russell in chapter VI of Russell (1903), but later dropped in favour of a "no-classes" theory. See also Gottlob Frege 1895 for a critique of an earlier view defended by Ernst Schroeder.
The general idea can be traced back to Leibniz. (Levey 2011, pp. 129–133)
Interest revived in plurals with work in linguistics in the 1970s by Remko Scha, Godehard Link, Fred Landman, Friederike Moltmann, Roger Schwarzschild, Peter Lasersohn and others, who developed ideas for a semantics of plurals.
Background and motivation
Multigrade (variably polyadic) predicates and relations
Sentences like
Alice and Bob cooperate.
Alice, Bob and Carol cooperate.
are said to involve a multigrade (also known as variably polyadic, also anadic) predicate or relation ("cooperate" in this example), meaning that they stand for the same concept even though they don't have a fixed arity (cf. Linnebo
|
https://en.wikipedia.org/wiki/Organic%20farming%20and%20biodiversity
|
The effect of organic farming has been a subject of interest for researchers. Theory suggests that organic farming practices, which exclude the use of most synthetic pesticides and fertilizers, may be beneficial for biodiversity. This is generally shown to be true for soils scaled to the area of cultivated land, where species abundance is, on average, 30% richer than that of conventional farms. However, for crop yield-scaled land the effect of organic farming on biodiversity is highly debated due to the significantly lower yields compared to conventional farms.
In ancient farming practices, farmers did not possess the technology or manpower to have a significant impact on the destruction of biodiversity even as mass-production agriculture was rising. Nowadays, common farming methods generally rely on pesticides to maintain high yields. With such, most agricultural landscapes favor mono-culture crops with very little flora or fauna co-existence (van Elsen 2000). Modern organic farm practices such as the removal of pesticides and the inclusion of animal manure, crop rotation, and multi-cultural crops provides the chance for biodiversity to thrive.
Benefits of organic farming to biodiversity
Nearly all non-crop, naturally occurring species observed in comparative farm land practice studies show a preference in organic farming both by population and richness. Spanning all associated species, there is an average of 30% more on organic farms versus conventional farming methods, however this does not account for possible loss of biodiversity due to decreased yields. Birds, butterflies, soil microbes, beetles, earthworms, spiders, vegetation, and mammals are particularly affected. Some organic farms may use less pesticides and thus biodiversity fitness and population density may benefit. Larger farms however tend to use pesticides more liberally and in some cases to larger extent than conventional farms. Many weed species attract beneficial insects that improve soil qual
|
https://en.wikipedia.org/wiki/Electronic%20circuit%20simulation
|
Electronic circuit simulation uses mathematical models to replicate the behavior of an actual electronic device or circuit.
Simulation software allows for modeling of circuit operation and is an invaluable analysis tool. Due to its highly accurate modeling capability, many colleges and universities use this type of software for the teaching of electronics technician and electronics engineering programs. Electronics simulation software engages its users by integrating them into the learning experience. These kinds of interactions actively engage learners to analyze, synthesize, organize, and evaluate content and result in learners constructing their own knowledge.
Simulating a circuit’s behavior before actually building it can greatly improve design efficiency by making faulty designs known as such, and providing insight into the behavior of electronics circuit designs. In particular, for integrated circuits, the tooling (photomasks) is expensive, breadboards are impractical, and probing the behavior of internal signals is extremely difficult. Therefore, almost all IC design relies heavily on simulation. The most well known analog simulator is SPICE. Probably the best known digital simulators are those based on Verilog and VHDL.
Some electronics simulators integrate a schematic editor, a simulation engine, and on-screen waveform display (see Figure 1), allowing designers to rapidly modify a simulated circuit and see what effect the changes have on the output. They also typically contain extensive model and device libraries. These models typically include IC specific transistor models such as BSIM, generic components such as resistors, capacitors, inductors and transformers, user defined models (such as controlled current and voltage sources, or models in Verilog-A or VHDL-AMS). Printed circuit board (PCB) design requires specific models as well, such as transmission lines for the traces and IBIS models for driving and receiving electronics.
Types
While
|
https://en.wikipedia.org/wiki/Michele%20Mosca
|
Michele Mosca is co-founder and deputy director of the Institute for Quantum Computing at the University of Waterloo, researcher and founding member of the Perimeter Institute for Theoretical Physics, and professor of mathematics in the department of Combinatorics & Optimization at the University of Waterloo. He has held a Tier 2 Canada Research Chair in Quantum Computation since January 2002, and has been a scholar for the Canadian Institute for Advanced Research since September 2003. Mosca's principal research interests concern the design of quantum algorithms, but he is also known for his early work on NMR quantum computation together with Jonathan A. Jones.
Graduate and post-graduate education
Mosca received a B.Math degree from the University of Waterloo in 1995. In 1996 he received a Commonwealth Scholarship to attend Wolfson College, Oxford University, where he received his M.Sc. degree in mathematics and foundations of computer science. On another scholarship (and while holding a fellowship), Mosca received his D.Phil degree on the topic of quantum computer algorithms, also at the University of Oxford.
Awards and honors
2010 Canada's Top 40 under 40 from The Globe and Mail.
Fellow of the Canadian Institute for Advanced Research program in Quantum Information since 2010. Scholar since 2003.
2010 Waterloo Region 40 under 40: "Honouring those making a difference in our region".
Invited Speaker, AAAS Science and Technology Workshop "Plug into Canada", organized by the Canadian Embassy, January 2005 (with National Science Advisor, the NSERC President, and 2 other Canadian researchers).
One of fifteen PAGSE Symposium "Leaders of Tomorrow", Ottawa, Canada, 2004. The Partnership Group for Science and Engineering was formed in June 1995 at the invitation of the Academy of Science of the Royal Society of Canada to represent the Canadian science and engineering community to the Government of Canada.
Invited article in inno'va-tion and inno'v@-tion2: Essays
|
https://en.wikipedia.org/wiki/Herbal%20medicine
|
Herbal medicine (also called herbalism, phytomedicine or phytotherapy) is the study of pharmacognosy and the use of medicinal plants, which are a basis of traditional medicine. With worldwide research into pharmacology, some herbal medicines have been translated into modern remedies, such as the anti-malarial group of drugs called artemisinin isolated from Artemisia annua, a herb that was known in Chinese medicine to treat fever. There is limited scientific evidence for the safety and efficacy of many plants used in 21st century herbalism, which generally does not provide standards for purity or dosage. The scope of herbal medicine sometimes include fungal and bee products, as well as minerals, shells and certain animal parts.
Paraherbalism describes alternative and pseudoscientific practices of using unrefined plant or animal extracts as unproven medicines or health-promoting agents. Paraherbalism relies on the belief that preserving various substances from a given source with less processing is safer or more effective than manufactured products, a concept for which there is no evidence.
History
Archaeological evidence indicates that the use of medicinal plants dates back to the Paleolithic age, approximately 60,000 years ago. Written evidence of herbal remedies dates back over 5,000 years to the Sumerians, who compiled lists of plants. Some ancient cultures wrote about plants and their medical uses in books called herbals. In ancient Egypt, herbs are mentioned in Egyptian medical papyri, depicted in tomb illustrations, or on rare occasions found in medical jars containing trace amounts of herbs. In ancient Egypt, the Ebers papyrus dates from about 1550 BC, and covers more than 700 compounds, mainly of plant origin. The earliest known Greek herbals came from Theophrastus of Eresos who, in the 4th century BC, wrote in Greek Historia Plantarum, from Diocles of Carystus who wrote during the 3rd century BC, and from Krateuas who wrote in the 1st century BC. Only a f
|
https://en.wikipedia.org/wiki/Vegetation%20%28pathology%29
|
In medicine, a vegetation is an abnormal growth named for its similarity to natural vegetation. Vegetations are often associated with endocarditis. They can be made of fibrin and platelets.
Types
Certain conditions are associated with specific vegetation patterns:
|
https://en.wikipedia.org/wiki/Integrated%20Visual%20Augmentation%20System
|
The Integrated Visual Augmentation System (IVAS) is an augmented reality headset being developed by Microsoft for the United States Army, providing a wide variety of new capabilities to soldiers. Originally developed for infantry, it is also being adapted for use by mounted soldiers and aircrew. Its development was initiated in 2018 and is currently undergoing testing, although numerous problems have repeatedly delayed its deployment in the field.
Design
IVAS is an augmented reality system based on the Microsoft Hololens 2 headset. It intends to provide soldiers with "improved situational awareness, target engagement, and informed decision-making".
The system consists of a display, a computer known as a "puck", a networked data radio, and three conformal batteries. The display can augment the soldier's vision with imagery from thermal imaging and low-light imaging sensors. It also contains a digital magnetic compass for navigation and can display imagery from the Family of Weapons Sights-Individual mounted on the soldier's weapon. The radio allows data from the soldiers' individual IVAS headsets to be passed among members of the company.
The latest version of IVAS, version 1.2, weighs , although developers are working to reduce this to the target weight of . The computer is attached to the back of the helmet in order to distribute weight and move the center of mass. The display has a field of view of 60 degrees, using a flat display that can be flipped upwards.
History
Although it is unknown when the idea for IVAS was first conceived, the Army Acquisition Executive first approved IVAS's development effort on September 25, 2018, and an Other Transaction Agreement for the development of IVAS was issued to Microsoft in November 2018. Initial testing began in March 2019.
2020
As of October 2020, IVAS was on its third iteration. Previous tests used commercial Microsoft HoloLens 2 headsets which were not resistant to inclement weather. A ruggedized version of the
|
https://en.wikipedia.org/wiki/Adventures%20Among%20the%20Toroids
|
Adventures Among the Toroids: A study of orientable polyhedra with regular faces is a book on toroidal polyhedra that have regular polygons as their faces. It was written, hand-lettered, and illustrated by mathematician Bonnie Stewart, and self-published under the imprint "Number One Tall Search Book" in 1970. Stewart put out a second edition, again hand-lettered and self-published, in 1980. Although out of print, the Basic Library List Committee of the Mathematical Association of America has recommended its inclusion in undergraduate mathematics libraries.
Topics
The Platonic solids, known to antiquity, have all faces regular polygons, all symmetric to each other (each face can be taken to each other face by a symmetry of the polyhedron). However, if less symmetry is required, a greater number of polyhedra can be formed while having all faces regular. The convex polyhedra with all faces regular were catalogued in 1966 by Norman Johnson (after earlier study e.g. by Martyn Cundy and A. P. Rollett), and have come to be known as the Johnson solids. Adventures Among the Toroids extends the investigation of polyhedra with regular faces to non-convex polyhedra, and in particular to polyhedra of higher genus than the sphere. Many of these polyhedra can be formed by gluing together smaller polyhedral pieces, carving polyhedral tunnels through them, or piling them into elaborate towers. The toroidal polyhedra described in this book, formed from regular polygons with no self-intersections or flat angles, have come to be called Stewart toroids.
The second edition is rewritten in a different page format, letter sized in landscape mode compared to the tall and narrow by page size of the first edition, with two columns per page. It includes new material on knotted polyhedra and on rings of regular octahedra and regular dodecahedra; as the ring of dodecahedra forms the outline of a golden rhombus, it can be extended to make skeletal pentagon-faced versions of the convex polyh
|
https://en.wikipedia.org/wiki/Broadcast%20transmitter
|
A broadcast transmitter is an electronic device which radiates radio waves modulated with information content intended to be received by the general public. Examples are a radio broadcasting transmitter which transmits audio (sound) to broadcast radio receivers (radios) owned by the public, or a television transmitter, which transmits moving images (video) to television receivers (televisions). The term often includes the antenna which radiates the radio waves, and the building and facilities associated with the transmitter. A broadcasting station (radio station or television station) consists of a broadcast transmitter along with the production studio which originates the broadcasts. Broadcast transmitters must be licensed by governments, and are restricted to specific frequencies and power levels. Each transmitter is assigned a unique identifier consisting of a string of letters and numbers called a callsign, which must be used in all broadcasts.
Exciter
In broadcasting and telecommunication, the part which contains the oscillator, modulator, and sometimes audio processor, is called the "exciter". Most transmitters use the heterodyne principle, so they also have frequency conversion units. Confusingly, the high-power amplifier which the exciter then feeds into is often called the "transmitter" by broadcast engineers. The final output is given as transmitter power output (TPO), although this is not what most stations are rated by.
Effective radiated power (ERP) is used when calculating station coverage, even for most non-broadcast stations. It is the TPO, minus any attenuation or radiated loss in the line to the antenna, multiplied by the gain (magnification) which the antenna provides toward the horizon. This antenna gain is important, because achieving a desired signal strength without it would result in an enormous electric utility bill for the transmitter, and a prohibitively expensive transmitter. For most large stations in the VHF- and UHF-range, the t
|
https://en.wikipedia.org/wiki/Perception%20of%20infrasound
|
Infrasound is sound at frequencies lower than the low frequency end of human hearing threshold at 20 Hz. It is known, however, that humans can perceive sounds below this frequency at very high pressure levels. Infrasound can come from many natural as well as man-made sources, including weather patterns, topographic features, ocean wave activity, thunderstorms, geomagnetic storms, earthquakes, jet streams, mountain ranges, and rocket launchings. Infrasounds are also present in the vocalizations of some animals. Low frequency sounds can travel for long distances with very little attenuation and can be detected hundreds of miles away from their sources.
Mammals
The production and perception of infrasound has been observed in multiple mammals, including whale, elephant, giraffe, hippopotamus, and rhinoceros. For most of these animals, observations are preliminary and their sensitivity to infrasound has not been quantified. If an animal produces a low frequency sound, and uses it in communication, it suggests the animal might also be sensitive to infrasound.
Elephants
Elephants are the terrestrial animal in which the production of infrasonic calls was first noted by M. Krishnan, later discovered by Katy Payne. The use of low frequency sounds to communicate over long distances may explain certain elephant behaviors that have previously puzzled observers. Elephant groups that are separated by several kilometers have been observed to travel in parallel or to change the direction simultaneously and move directly towards each other in order to meet. The time of estrus for females is asynchronous, lasts only for a few days, and occurs only every several years. Nevertheless, males, which usually wander apart from female groups, rapidly gather from many directions to compete for a receptive female. Since infrasound can travel for very long distances, it has been suggested that calls in the infrasonic range might be important for long distance communication for such coordin
|
https://en.wikipedia.org/wiki/Key%20clustering
|
Key or hash function should avoid clustering, the mapping of two or more keys to consecutive slots. Such clustering may cause the lookup cost to skyrocket, even if the load factor is low and collisions are infrequent. The popular multiplicative hash is claimed to have particularly poor clustering behaviour.
|
https://en.wikipedia.org/wiki/152%20%28number%29
|
152 (one hundred [and] fifty-two) is the natural number following 151 and preceding 153.
In mathematics
152 is the sum of four consecutive primes (31 + 37 + 41 + 43). It is a nontotient since there is no integer with 152 coprimes below it.
152 is a refactorable number since it is divisible by the total number of divisors it has, and in base 10 it is divisible by the sum of its digits, making it a Harshad number.
Recently, the smallest repunit probable prime in base 152 was found, it has 589570 digits.
The number of surface points on a 6*6*6 cube is 152.
In the military
Focke-Wulf Ta 152 was a Luftwaffe high-altitude interceptor fighter aircraft during World War II
was a United States Navy during World War II
was a United States Navy during World War II
was a United States Navy supply ship during World War II
was a United States Navy during World War II
was a United States Navy ship during World War II
was a United States Navy during World War II
was a United States Navy during World War II
152.3 (5.9"), common medium artillery (and historically heavy tank destroyer) caliber utilized by Russia, China and former members of the Soviet Union, akin to the 155 mm standard caliber of NATO nations.
In transportation
The Baade 152, the first German jet passenger airliner in 1958
The Cessna 152 airplane
Garuda Indonesia Flight 152 was an Indonesian flight from Jakarta to Medan that crashed on September 26, 1997
London Buses route 152
In TV, radio, games and cinema
The aviation-frequency radio exchange (pronounced one-fifty-two), as 152 is associated with the Cessna 152
"NY152" AOL e-mail account use by Joe in the movie You've Got Mail
In other fields
152 is also:
The year AD 152 or 152 BC
152 AH is a year in the Islamic calendar that corresponds to 759 – 760 CE
152 Atala is a dark type D main belt asteroid
The atomic number of an element temporarily called Unpentbium
Sonnet 152
The Garmin GPS 152, produced in 2001
The Xerox DocuMate
|
https://en.wikipedia.org/wiki/Gravity%20laser
|
A gravity laser, also sometimes referred to as a Gaser, Graser, or Glaser, is a hypothetical device for stimulated emission of coherent gravitational radiation or gravitons, much in the same way that a standard laser produces coherent electromagnetic radiation.
Principle of function
While photons exist as excitations of a vector potential and so contain an oscillating dipole term, gravitons are a spin-2 field and so have an oscillating quadrupole term. For efficient lasing to occur, there are several conditions that must be met:
There must be particles in an excited state capable of emitting radiation at the desired frequency. In a normal laser, these would be valence electrons in an excited state. For a gaser, the more straightforward analog would be a binary system of massive bodies.
These particles must couple to supplied radiation, in order to provide stimulated emission. This could be possible in a gaser by a stimulated analog of the Penrose process.
The particles must be in an inverted population, where more are in the excited state than the ground state. This typically requires some type of pumping, such as optical pumping.
The lasing medium must be long enough for the radiation to persist and excite more of the same. In optical systems this can typically be created by mirrors, effectively making a larger optical path length. For a gaser, a large-scale, slowly spatially varying gravitational potential could act as a mirror (by the WKB approximation). Alternately, a hypothetical gaser could simply be built with sufficient length to begin with.
Alternate design proposals involve free undulators akin to a free-electron laser. Several proposals involve exploiting the momentum transport properties of superconductors, where s-waves and d-waves couple distinctly to gravitational radiation.
As of 2019, there are no plans to construct a gravity laser.
Use in science fiction
The idea of gravity lasers has been popularized by science fiction works such as Davi
|
https://en.wikipedia.org/wiki/Energy%20condition
|
In relativistic classical field theories of gravitation, particularly general relativity, an energy condition is a generalization of the statement "the energy density of a region of space cannot be negative" in a relativistically-phrased mathematical formulation. There are multiple possible alternative ways to express such a condition such that can be applied to the matter content of the theory. The hope is then that any reasonable matter theory will satisfy this condition or at least will preserve the condition if it is satisfied by the starting conditions.
Energy conditions are not physical constraints , but are rather mathematically imposed boundary conditions that attempt to capture a belief that "energy should be positive". Many energy conditions are known to not correspond to physical reality—for example, the observable effects of dark energy are well-known to violate the strong energy condition.
In general relativity, energy conditions are often used (and required) in proofs of various important theorems about black holes, such as the no hair theorem or the laws of black hole thermodynamics.
Motivation
In general relativity and allied theories, the distribution of the mass, momentum, and stress due to matter and to any non-gravitational fields is described by the energy–momentum tensor (or matter tensor) . However, the Einstein field equation in itself does not specify what kinds of states of matter or non-gravitational fields are admissible in a spacetime model. This is both a strength, since a good general theory of gravitation should be maximally independent of any assumptions concerning non-gravitational physics, and a weakness, because without some further criterion the Einstein field equation admits putative solutions with properties most physicists regard as unphysical, i.e. too weird to resemble anything in the real universe even approximately.
The energy conditions represent such criteria. Roughly speaking, they crudely describe properties common
|
https://en.wikipedia.org/wiki/Hydrogen%20turboexpander-generator
|
A hydrogen turboexpander-generator or generator-loaded expander for hydrogen gas is an axial flow turbine or radial expander for energy recovery through which a high pressure hydrogen gas is expanded to produce work used to drive an electrical generator. It replaces the control valve or regulator where the pressure drops to the appropriate pressure for the low-pressure network. A turboexpander generator can help recover energy losses and offset electrical requirements and emissions.
Description
Per stage, 200 bar is handled with up to 15,000 kW power and a maximum expansion ratio of 14, the generator loaded expander for hydrogen gas is fitted with an automatic thrust balance, a dry gas seal, and a programmable logic control with remote monitoring and diagnostics.
Application
The hydrogen turboexpander-generators are used for hydrogen pipeline transport in combination with hydrogen compressors and energy recovery in underground hydrogen storage. A variation is the compressor loaded turboexpanders which are used in the liquefaction of gases such as liquid hydrogen
See also
compressed hydrogen
letdown station
hydrogen infrastructure
turboexpander
|
https://en.wikipedia.org/wiki/Boolean%20conjunctive%20query
|
In the theory of relational databases, a Boolean conjunctive query is a conjunctive query without distinguished predicates, i.e., a query in the form , where each is a relation symbol and each is a tuple of variables and constants; the number of elements in is equal to the arity of . Such a query evaluates to either true or false depending on whether the relations in the database contain the appropriate tuples of values, i.e. the conjunction is valid according to the facts in the database.
As an example, if a database schema contains the relation symbols (binary, who's the father of whom) and (unary, who is employed), a conjunctive query could be . This query evaluates to true if there exists an individual who is a child of Mark and employed. In other words, this query expresses the question: "does Mark have an employed child?"
See also
Logical conjunction
Conjunctive query
|
https://en.wikipedia.org/wiki/Sapphire%20Technology
|
Sapphire Technology () is a Hong Kong-based technology company, founded in 2001, which produces graphics cards for personal computers and workstations, motherboards, TV tuner cards, digital audio players and LCDTVs
Sapphire's products are based on AMD graphics processing units, and both AMD (ATI) and Intel motherboard chipset technology. The company is the largest supplier of AMD-based video cards in the world.
Sapphire was the first company to release a video card with a high definition multimedia interface (HDMI) connector.
Sapphire was the first company to release a video card having clock speed of 1000 MHz (1 GHz) with the release of the Sapphire Atomic Edition HD 4890.
Manufacturing facilities
As of 2007, Sapphire has two ISO 9001 and ISO 14001-certified manufacturing facilities in Dongguan, China, which have a monthly production capacity of 1.8 million video cards.
The manufacturing facility had an area of about 250,000 m2 used by 16 independent production lines as of May 2005.
Manufacturing process
Sapphire buys printed circuit boards (PCB) from an external contractor, but they place components on the PCB and reflow them in their own factories. AMD GPUs have historically been used in their products.
Gallery
|
https://en.wikipedia.org/wiki/Am486
|
The Am486 is a 80486-class family of computer processors that was produced by AMD in the 1990s. Intel beat AMD to market by nearly four years, but AMD priced its 40 MHz 486 at or below Intel's price for a 33 MHz chip, offering about 20% better performance for the same price.
While competing 486 chips, such as those from Cyrix, benchmarked lower than the equivalent Intel chip, AMD's 486 matched Intel's performance on a clock-for-clock basis.
While the Am386 was primarily used by small computer manufacturers, the Am486DX, DX2, and SX2 chips gained acceptance among larger computer manufacturers, especially Acer and Compaq, in the 1994 time frame.
AMD's higher clocked 486 chips provided superior performance to many of the early Pentium chips, especially the 60 and 66 MHz launch products. While equivalent Intel 80486DX4 chips were priced high and required a minor socket modification, AMD priced low. Intel's DX4 chips initially had twice the cache of the AMD chips, giving them a slight performance edge, but AMD's DX4-100 usually cost less than Intel's DX2-66.
The enhanced Am486 series supported new features like extended power-saving modes and an 8 KiB Write-Back L1-Cache, later versions even got an upgrade to 16 KiB Write-Back L1-Cache.
The 133 MHz AMD Am5x86 was a higher clocked enhanced Am486.
One derivative of the Am486 family is the core used in the AMD Élan SC4xx family of microcontrollers marketed by AMD.
Features
Am486 models
WT = Write-Through cache strategy, WB = Write-Back cache strategy
|
https://en.wikipedia.org/wiki/Micromatabilin
|
Micromatabilin, the green pigment of the spider species Micrommata virescens, is characterized as a mixture of biliverdin conjugates. The two isolated fractions have identical absorption bands (free base: 620–630 μm, hydrochloride: 690 μm, zinc complex: 685–690 μm). Chromic acid degradation yields imides I, II, IIIa, and IIIb. Differences in the non-hydrolytic degradation and in polarity lead to the conclusion that fraction 1 is a monoconjugate and fraction 2a diconjugate of biliverdin.
|
https://en.wikipedia.org/wiki/Ernest%20Julius%20Wilczynski
|
Ernest Julius Wilczynski (November 13, 1876 – September 14, 1932) was an American mathematician considered the founder of projective differential geometry.
Born in Hamburg, Germany, Wilczynski's family emigrated to America and settled in Chicago, Illinois when he was very young. He attended public school in the US but went to college in Germany and received his PhD from the University of Berlin in 1897. He taught at the University of California until 1907, the University of Illinois from 1907 to 1910, and the University of Chicago from 1910 until illness forced his absence from the classroom in 1923. His doctoral students include Archibald Henderson, Ernest Preston Lane, Pauline Sperry, Ellis Stouffer, and Charles Thompson Sullivan.
Selected publications
Projective differential geometry of curves and ruled surfaces, Leipzig, Teubner 1906
Projective differential geometry of curved surfaces, Parts I–V, Transactions American Mathematical Society
vol. 8, 1907, Part I, pp. 223–260
vol. 9, 1908, Part II, pp. 79–120 ; Part III, pp. 293–315
vol. 10, 1909, Part IV, pp. 176–200 ; Part V, pp. 279–296
|
https://en.wikipedia.org/wiki/Wolf%20summation
|
The Wolf summation is a method for computing the electrostatic interactions of systems (e.g. crystals). This method is generally more computationally efficient than the Ewald summation. It was proposed by Dieter Wolf.
|
https://en.wikipedia.org/wiki/Lipoblast
|
A lipoblast is a precursor cell for an adipocyte.
Alternate terms include adipoblast and preadipocyte.
Early stages are almost indistinguishable from fibroblasts.
Liposarcoma
Lipoblasts are seen in liposarcoma and characteristically have abundant multivacuolated clear cytoplasm and a dark staining (hyperchromatic), indented nucleus.
See also
Adipogenesis
Adipose differentiation-related protein
Lipoblastoma
List of human cell types derived from the germ layers
|
https://en.wikipedia.org/wiki/Who%27ll%20Stop%20the%20Rain%20%28song%29
|
"Who'll Stop the Rain" is a song written by John Fogerty and originally recorded by Creedence Clearwater Revival for their 1970 album Cosmo's Factory. Backed with "Travelin' Band", it was one of three double-sided singles from that album to reach the top five on the Billboard Pop Singles chart and the first of two to reach the No. 2 spot on the American charts, alongside "Lookin' Out My Back Door"/"Long As I Can See the Light". In 2004, Rolling Stone ranked it No. 188 on its "500 Greatest Songs of All Time" list.
History
Lyrically, "Who'll Stop the Rain" breaks into three verses, with a historical, recent past, and present tense approach. All three verses allude to a sense of unending malaise, pondered by "good men through the ages", "Five Year Plans and New Deals/wrapped in golden chains", and the Woodstock generation.
Musically, in contrast to the 1950s-Rock-inspired "Travelin' Band", "Who'll Stop the Rain" has more of an acoustic, folk-rock feel to it. Like many folk-rock songs, it starts off with a ringing acoustic guitar riff, though the backing throughout has more of a roots rock sound than that heard on more standard folk-rock recordings. Interpreting the song in its time period (1970), and the resigned but somewhat angry feeling of the song, many see "Who'll Stop the Rain" as a thinly veiled protest against the Vietnam War, with the final verse lyrics and its references to music, large crowds, rain, and crowds trying to keep warm being about the band's experience at the Woodstock Festival in August 1969. For his part, when asked by Rolling Stone about the meaning of the song's lyrics, John Fogerty was quoted as saying,
In 2007 during a concert in Shelburne, Vermont, he said the following about the song:
Reception
Billboard called it a "blockbuster side" that "has the beat and feel of [Creedence's] hits." Cash Box said that it spotlights Creedence's "originated river-rock style." Record World called the single a "two-sided smash."
Ultimate Classic Rock
|
https://en.wikipedia.org/wiki/Shimura%27s%20reciprocity%20law
|
In mathematics, Shimura's reciprocity law, introduced by , describes the action of ideles of imaginary quadratic fields on the values of modular functions at singular moduli. It forms a part of the Kronecker Jugendtraum, explicit class field theory for such fields. There are also higher-dimensional generalizations.
|
https://en.wikipedia.org/wiki/Polynomial%20conjoint%20measurement
|
Polynomial conjoint measurement is an extension of the theory of conjoint measurement to three or more attributes. It was initially developed by the mathematical psychologists David Krantz (1968) and Amos Tversky (1967). The theory was given a comprehensive mathematical exposition in the first volume of Foundations of Measurement (Krantz, Luce, Suppes & Tversky, 1971), which Krantz and Tversky wrote in collaboration with the mathematical psychologist R. Duncan Luce and philosopher Patrick Suppes. Krantz & Tversky (1971) also published a non-technical paper on polynomial conjoint measurement for behavioural scientists in the journal Psychological Review.
As with the theory of conjoint measurement, the significance of polynomial conjoint measurement lies in the quantification of natural attributes in the absence of concatenation operations. Polynomial conjoint measurement differs from the two attribute case discovered by Luce & Tukey (1964) in that more complex composition rules are involved.
Polynomial conjoint measurement
Krantz's (1968) schema
Most scientific theories involve more than just two attributes; and thus the two variable case of conjoint measurement has rather limited scope. Moreover, contrary to the theory of n – component conjoint measurement, many attributes are non-additive compositions of other attributes (Krantz, et al., 1971). Krantz (1968) proposed a general schema to ascertain the sufficient set of cancellation axioms for a class of polynomial combination rules he called simple polynomials. The formal definition of this schema given by Krantz, et al., (1971, p. 328) is as follows.
Let . The set is the smallest set of simple polynomials such that:
;
such that and , then and are in .
Informally, the schema argues:
a) single attributes are simple polynomials;
b) if G1 and G2 are simple polynomials that are disjoint (i.e. have no attributes in common), then G1 + G2 and G1 G2 are simple polynomials; and
c) no polynomials are simple exc
|
https://en.wikipedia.org/wiki/Cheese%20ripening
|
Cheese ripening, alternatively cheese maturation or affinage, is a process in cheesemaking. It is responsible for the distinct flavour of cheese, and through the modification of "ripening agents", determines the features that define many different varieties of cheeses, such as taste, texture, and body. The process is "characterized by a series of complex physical, chemical and microbiological changes" that incorporates the agents of "bacteria and enzymes of the milk, lactic culture, rennet, lipases, added moulds or yeasts, and environmental contaminants". The majority of cheese is ripened, except for fresh cheese.
History
Cheese ripening was not always the highly industrialised process it is today; in the past, cellars and caves were used to ripen cheeses instead of the current highly regulated process involving machinery and biochemistry. Some cheeses still are made using more historical methods, such as the blue cheese Roquefort, which is required to be ripened in designated caves in south-western France. However, with the invention of refrigeration in the 20th century, the process evolved considerably, and is much more efficient at producing a consistent quality of cheese, at a faster pace, and a lower cost (depending on the type of cheese).
Process
After the initial manufacturing process of the cheese is done, the cheese ripening process occurs. This process is especially important, since it defines the flavour and texture of the cheese, which differentiates the many varieties. Duration is dependent on the type of cheese and the desired quality, and typically ranges from "three weeks to two or more years".
Ripening is influenced by a variety of factors, ranging from the microflora to the curd, and others. The enzymatic process is the most crucial process for all cheeses, although bacteria play a role in many varieties. The most important agents in this process include the following elements:
Rennet, or a substitute for rennet
starter bacteria an
|
https://en.wikipedia.org/wiki/Nagell%E2%80%93Lutz%20theorem
|
In mathematics, the Nagell–Lutz theorem is a result in the diophantine geometry of elliptic curves, which describes rational torsion points on elliptic curves over the integers.
It is named for Trygve Nagell and Élisabeth Lutz.
Definition of the terms
Suppose that the equation
defines a non-singular cubic curve with integer coefficients a, b, c, and let D be the discriminant of the cubic polynomial on the right side:
Statement of the theorem
If P = (x,y) is a rational point of finite order on C, for the elliptic curve group law, then:
1) x and y are integers
2) either y = 0, in which case P has order two, or else y divides D, which immediately implies that y2 divides D.
Generalizations
The Nagell–Lutz theorem generalizes to arbitrary number fields and more
general cubic equations.
For curves over the rationals, the
generalization says that, for a nonsingular cubic curve
whose Weierstrass form
has integer coefficients, any rational point P=(x,y) of finite
order must have integer coordinates, or else have order 2 and
coordinates of the form x=m/4, y=n/8, for m and n integers.
History
The result is named for its two independent discoverers, the Norwegian Trygve Nagell (1895–1988) who published it in 1935, and Élisabeth Lutz (1937).
See also
Mordell–Weil theorem
|
https://en.wikipedia.org/wiki/Lozenge%20%28shape%29
|
A lozenge ( ; symbol: ), often referred to as a diamond, is a form of rhombus. The definition of lozenge is not strictly fixed, and the word is sometimes used simply as a synonym () for rhombus. Most often, though, lozenge refers to a thin rhombus—a rhombus with two acute and two obtuse angles, especially one with acute angles of 45°. The lozenge shape is often used in parquetry (with acute angles that are 360°/n with n being an integer higher than 4, because they can be used to form a set of tiles of the same shape and size, reusable to cover the plane in various geometric patterns as the result of a tiling process called tessellation in mathematics) and as decoration on ceramics, silverware and textiles. It also features in heraldry and playing cards.
Symbolism
The lozenge motif dates from the Neolithic and Paleolithic period in Eastern Europe and represents a sown field and female fertility. The ancient lozenge pattern often shows up in Diamond vault architecture, in traditional dress patterns of Slavic peoples, and in traditional Ukrainian embroidery. The lozenge pattern also appears extensively in Celtic art, art from the Ottoman Empire, and ancient Phrygian art.
The lozenge symbolism is one of the main symbols for women in Berber carpets.
Common Berber jewelry from the Aurès Mountains or Kabylie in Algeria also uses this pattern as a female fertility sign.
In 1658, the English philosopher Sir Thomas Browne published The Garden of Cyrus, subtitled The Quincunciall Lozenge, or Network Plantations of the Ancients, in which he outlined the mystical interconnection of art, nature and the universe via the quincunx pattern. He also suggested therein that ancient plantations were laid out in a lozenge pattern.
Lozenges appear as symbols in ancient classic element systems, in amulets, and in religious symbolism. In playing cards, the symbol for the suit of diamonds is a lozenge.
Encodings
In Unicode, the lozenge is encoded in multiple variants:
|
https://en.wikipedia.org/wiki/Horologium%20Oscillatorium
|
(English: The Pendulum Clock: or Geometrical Demonstrations Concerning the Motion of Pendula as Applied to Clocks) is a book published by Dutch mathematician and physicist Christiaan Huygens in 1673 and his major work on pendula and horology. It is regarded as one of the three most important works on mechanics in the 17th century, the other two being Galileo’s Discourses and Mathematical Demonstrations Relating to Two New Sciences (1638) and Newton’s (1687).
Much more than a mere description of clocks, Huygens's is the first modern treatise in which a physical problem (the accelerated motion of a falling body) is idealized by a set of parameters then analyzed mathematically and constitutes one of the seminal works of applied mathematics. The book is also known for its strangely worded dedication to Louis XIV. The appearance of the book in 1673 was a political issue, since at that time the Dutch Republic was at war with France; Huygens was anxious to show his allegiance to his patron, which can be seen in the obsequious dedication to Louis XIV.
Overview
The motivation behind Horologium Oscillatorium (1673) goes back to the idea of using a pendulum to keep time, which had already been proposed by people engaged in astronomical observations such as Galileo. Mechanical clocks at the time were instead regulated by balances that were often very unreliable. Moreover, without reliable clocks, there was no good way to measure longitude at sea, which was particularly problematic for a country dependent on sea trade like the Dutch Republic.
Huygens interest in using a freely suspended pendulum to regulate clocks began in earnest in December 1656. He had a working model by the next year which he patented and then communicated to others such as Frans van Schooten and Claude Mylon. Although Huygens’s design, published in a short tract entitled Horologium (1658), was a combination of existing ideas, it nonetheless became widely popular and many pendulum clocks by Salomon Co
|
https://en.wikipedia.org/wiki/Dysdiadochokinesia
|
Dysdiadochokinesia (DDK) is the medical term for an impaired ability to perform rapid, alternating movements (i.e., diadochokinesia). Complete inability is called adiadochokinesia. The term is from Greek δυς dys "bad", διάδοχος diadochos "succeeding", κίνησις kinesis "movement".
Signs and symptoms
Abnormalities in diadochokinesia can be seen in the upper extremity, lower extremity and in speech. The deficits become visible in the rate of alternation, the completeness of the sequence, and in the variation in amplitude involving both motor coordination and sequencing. Average rate can be used as a measure of performance when testing for dysdiadochokinesia.
Dysdiadochokinesia is demonstrated clinically by asking the patient to tap the palm of one hand with the fingers of the other, then rapidly turn over the fingers and tap the palm with the back of them, repeatedly. This movement is known as a pronation/supination test of the upper extremity. A simpler method using this same concept is to ask the patient to demonstrate the movement of trying a doorknob or screwing in a light bulb. When testing for this condition in legs, ask the patient to tap your hand as quickly as possible with the ball of each foot in turn. Movements tend to be slow or awkward. The feet normally perform less well than the hands. When testing for dysdiadochokinesia with speech the patient is asked to repeat syllables such as /pə/, /tə/, and /kə/; variation, excess loudness, and irregular articular breakdown are signs of dysdiadochokinesia.
Causes
Dysdiadochokinesia is a feature of cerebellar ataxia and may be the result of lesions to either the cerebellar hemispheres or the frontal lobe (of the cerebrum), it can also be a combination of both. It is thought to be caused by the inability to switch on and switch off antagonising muscle groups in a coordinated fashion due to hypotonia, secondary to the central lesion.
Dysdiadochokinesia is also seen in Friedreich's ataxia and multiple sclerosis, as
|
https://en.wikipedia.org/wiki/Sodium%20in%20biology
|
Sodium ions () are necessary in small amounts for some types of plants, but sodium as a nutrient is more generally needed in larger amounts by animals, due to their use of it for generation of nerve impulses and for maintenance of electrolyte balance and fluid balance. In animals, sodium ions are necessary for the aforementioned functions and for heart activity and certain metabolic functions. The health effects of salt reflect what happens when the body has too much or too little sodium.
Characteristic concentrations of sodium in model organisms are: 10 mM in E. coli, 30 mM in budding yeast, 10 mM in mammalian cell and 100 mM in blood plasma.
Sodium distribution in species
Humans
The minimum physiological requirement for sodium is between 115 and 500 mg per day depending on sweating due to physical activity, and whether the person is adapted to the climate. Sodium chloride is the principal source of sodium in the diet, and is used as seasoning and preservative, such as for pickling and jerky; most of it comes from processed foods. The Adequate Intake for sodium is 1.2 to 1.5 g per day, but on average people in the United States consume 3.4 g per day, the minimum amount that promotes hypertension. Note that salt contains about 39.3% sodium by massthe rest being chlorine and other trace chemicals; thus the Tolerable Upper Intake Level of 2.3 g sodium would be about 5.9 g of salt—about 1 teaspoon. The average daily excretion of sodium is between 40 and 220 mEq.
Normal serum sodium levels are between approximately 135 and 145 mEq/L (135 to 145 mmol/L). A serum sodium level of less than 135 mEq/L qualifies as hyponatremia, which is considered severe when the serum sodium level is below 125 mEq/L.
The renin–angiotensin system and the atrial natriuretic peptide indirectly regulate the amount of signal transduction in the human central nervous system, which depends on sodium ion motion across the nerve cell membrane, in all nerves. Sodium is thus important in neuron
|
https://en.wikipedia.org/wiki/Gas%20laws
|
The laws describing the behaviour of gases under fixed pressure, volume and absolute temperature conditions are called Gas Laws. The basic gas laws were discovered by the end of the 18th century when scientists found out that relationships between pressure, volume and temperature of a sample of gas could be obtained which would hold to approximation for all gases. These macroscopic gas laws were found to be consistent with atomic and kinetic theory.
History
Following the invention of the Torricelli mercury barometer in mid 17th century, the pressure-volume gas law was soon revealed by Robert Boyle while keeping temperature constant. Marriott, however, did notice small temperature dependence. It took another century and a half to develop thermometry and recognise the absolute zero temperature scale before the discovery of temperature-dependent gas laws.
Boyle's law
In 1662, Robert Boyle systematically studied the relationship between the volume and pressure of a fixed amount of gas at a constant temperature. He observed that the volume of a given mass of a gas is inversely proportional to its pressure at a constant temperature.
Boyle's law, published in 1662, states that, at a constant temperature, the product of the pressure and volume of a given mass of an ideal gas in a closed system is always constant. It can be verified experimentally using a pressure gauge and a variable volume container. It can also be derived from the kinetic theory of gases: if a container, with a fixed number of molecules inside, is reduced in volume, more molecules will strike a given area of the sides of the container per unit time, causing a greater pressure.
Statement
Boyle's law states that:
The concept can be represented with these formulae:
, meaning "Volume is inversely proportional to Pressure", or
, meaning "Pressure is inversely proportional to Volume", or
, or
where is the pressure, is the volume of a gas, and is the constant in this equation (and is not the same as
|
https://en.wikipedia.org/wiki/Shift%20rule
|
The shift rule is a mathematical rule for sequences and series.
Here and are natural numbers.
For sequences, the rule states that if is a sequence, then it converges if and only if also converges, and in this case both sequences always converge to the same number.
For series, the rule states that the series converges to a number if and only if converges.
|
https://en.wikipedia.org/wiki/I%20Love%20a%20Rainy%20Night
|
"I Love a Rainy Night" is a song co-written and recorded by American country music artist Eddie Rabbitt. It was released in November 1980 as the second single from his album Horizon. It reached number one on the Hot Country Singles, Billboard Hot 100, and Adult Contemporary Singles charts in early 1981. It was written by Rabbitt, Even Stevens, and David Malloy.
Song history
According to music historian Fred Bronson, "I Love a Rainy Night" was 12 years in the making. Rabbitt had a collection of old tapes he kept in the basement of his home. While rummaging through the tapes one day in 1980, he heard a fragment of a song he had recorded one rainy night in the late 1960s.
"It brought back the memory of sitting in a small apartment, staring out the window at one o'clock in the morning, watching the rain come down," wrote Bronson in The Billboard Book of Number One Hits. "He sang into his tape recorder, 'I love a rainy night, I love a rainy night.'"
Upon rediscovery of the old lyrics, Rabbitt completed the song (with help from frequent songwriting partners Even Stevens and David Malloy) and recorded it.
The result included vivid descriptions of a man's fondness for thunderstorms and the peace it brings him ("I love to hear the thunder/watch the lightnin' when it lights up the sky/you know it makes me feel good") and a renewed sense of hope the storms bring ("Showers wash all my cares away/I wake up to a sunny day").
The song's other distinctive feature is its rhythmic pattern of alternating finger snaps and hand claps, which was included with the help of percussionist Farrell Morris, who, according to The Billboard Book of Number One Country Hits, mixed two tracks of each to complete the record.
Chart performance
On February 28, the song succeeded Dolly Parton's hit film theme song "9 to 5" in the number-one position on the Billboard Hot 100 pop singles chart. On March 14, Parton's song returned to the top spot – the last time that the pop chart featured back-to-ba
|
https://en.wikipedia.org/wiki/Smallest%20grammar%20problem
|
In data compression and the theory of formal languages, the smallest grammar problem is the problem of finding the smallest context-free grammar that generates a given string of characters (but no other string). The size of a grammar is defined by some authors as the number of symbols on the right side of the production rules.
Others also add the number of rules to that. The (decision version of the) problem is NP-complete.
The smallest context-free grammar that generates a given string is always a straight-line grammar without useless rules.
See also
Grammar-based code
Kolmogorov Complexity
Lossless data compression
|
https://en.wikipedia.org/wiki/Certified%20interconnect%20designer
|
Certified Interconnect Designer (CID) is a certification from the IPC Designer's Council for experienced PCB design professionals. CID+ is the advanced version of this certification.
External links
IPC Designers Council, Designer Certification – WAGO PCB Interconnect connector 2059 series Click
Printed circuit board manufacturing
Professional titles and certifications
|
https://en.wikipedia.org/wiki/Emo%20Welzl
|
Emmerich (Emo) Welzl (born 4 August 1958 in Linz, Austria) is a computer scientist known for his research in computational geometry. He is a professor in the Institute for Theoretical Computer Science at ETH Zurich in Switzerland.
Biography
Welzl was born on 4 August 1958 in Linz, Austria. He studied at the Graz University of Technology receiving a Diplom in Applied Mathematics in 1981 and a doctorate in 1983 under the supervision of Hermann Maurer. Following postdoctoral studies at Leiden University, he became a professor at the Free University of Berlin in 1987 at age 28 and was the youngest professor in Germany. Since 1996 he has been professor of Computer Science at the ETH Zurich.
Welzl is a member of multiple journal editorial boards, and has been program chair for the Symposium on Computational Geometry in 1995, one of the tracks of the International Colloquium on Automata, Languages and Programming in 2000, and one of the tracks of the European Symposium on Algorithms in 2007.
Research
Much of Welzl's research has been in computational geometry. With David Haussler, he showed that machinery from computational learning theory including ε-nets and VC dimension could be useful in geometric problems such as the development of space-efficient range searching data structures. He devised linear time randomized algorithms for the smallest circle problem and for low-dimensional linear programming, and developed the combinatorial framework of LP-type problems that generalizes both of these problems. Other highly cited research publications by Welzl and his co-authors describe algorithms for constructing visibility graphs and using them to find shortest paths among obstacles in the plane, test whether two point sets can be mapped to each other by a combination of a geometric transformation and a small perturbation, and pioneer the use of space-filling curves for range query data structures.
Awards and honors
Welzl won the Gottfried Wilhelm Leibniz Prize in 1995. He
|
https://en.wikipedia.org/wiki/Clinical%20clerkship
|
Clinical clerkships encompass a period of medical education in which students – medical, nursing, dental, or otherwise – practice medicine under the supervision of a health practitioner.
Medical clerkships
In medical education, a clerkship, or rotation, refers to the practice of medicine by medical students (M.D., D.O., D.P.M) during their final year(s) of study. Traditionally, the first half of medical school trains students in the classroom setting, and the second half takes place in a teaching hospital. Clerkships give students experience in all parts of the hospital setting, including the operating room, emergency department, and various other departments that allow learning by viewing and doing.
Students are required to undergo a pre-clerkship course, which include introduction to clinical medicine, clinical skills, and clinical reasoning. A performance assessment such as the Objective Structured Clinical Examination (OSCE) is conducted at the end of this period. During the clerkship training, students are required to rotate through different medical specialties and treat patients under the supervision of physicians. Students elicit patient histories, complete physical examinations, write progress notes, and assist in surgeries and medical procedures. They are also actively involved in the diagnoses and treatment of patients under the supervision of a resident or faculty.
Students undergoing two-year clerkships spend their first year in patient care environment in month-long rotations with limited patient workloads. In their final year, when they are sometimes referred to as sub-interns or externs, they are given more patient care responsibilities in a variety or elective rotations.
The work hours are that of a full-time job, generally similar to that of residents. Students may also be required to work on weekends and to be on call.
For medical students, clerkships occur after the basic science curriculum, and are supervised by medical specialists at a te
|
https://en.wikipedia.org/wiki/Tissue%20cytometry
|
Tissue image cytometry or tissue cytometry is a method of digital histopathology and combines classical digital pathology (glass slides scanning and virtual slide generation) and computational pathology (digital analysis) into one integrated approach with solutions for all kinds of diseases, tissue and cell types as well as molecular markers and corresponding staining methods to visualize these markers. Tissue cytometry uses virtual slides as they can be generated by multiple, commercially available slide scanners, as well as dedicated image analysis software – preferentially including machine and deep learning algorithms. Tissue cytometry enables cellular analysis within thick tissues, retaining morphological and contextual information, including spatial information on defined cellular subpopulations. In this process, a tissue sample, either formalin-fixed paraffin-embedded (FFPE) or frozen tissue section, also referred to as “cryocut”, is labelled with either immunohistochemistry(IHC) or immunofluorescent markers, scanned with high-throughput slide scanners and the data gathered from virtual slides is processed and analyzed using software that is able to identify individual cells in tissue context automatically and distinguish between nucleus and cytoplasm for each cell. Additional algorithms can identify cellular membranes, subcellular structures (like cytoskeletal fibers, vacuoles, nucleoli) and/or multicellular tissue structures (glands, glomeruli, epidermis, or tumor foci).
Fluorescence Activated Cell Sorting (FACS) is a method of analysis that measures fluorescence signals on single cells, where the signal comes from antibody-mediated staining techniques and phenotypes detected by flow cytometry. The major limitation of flow cytometry is that it can only be applied – as the name suggest – to cells in solution. Although methods of “solubilizing” solid tissue exist, any such processing irrevocably destroys the tissue architecture and any spatial context. Hen
|
https://en.wikipedia.org/wiki/Chlorothalonil
|
Chlorothalonil (2,4,5,6-tetrachloroisophthalonitrile) is an organic compound mainly used as a broad spectrum, nonsystemic fungicide, with other uses as a wood protectant, pesticide, acaricide, and to control mold, mildew, bacteria, algae. Chlorothalonil-containing products are sold under the names Bravo, Echo, and Daconil. It was first registered for use in the US in 1966. In 1997, the most recent year for which data are available, it was the third most used fungicide in the US, behind only sulfur and copper, with used in agriculture that year. Including nonagricultural uses, the United States Environmental Protection Agency (EPA) estimates, on average, almost were used annually from 1990 to 1996.
Uses
In the US, chlorothalonil is used predominantly on peanuts (about 34% of usage), potatoes (about 12%), and tomatoes (about 7%), although the EPA recognizes its use on many other crops. It is also used on golf courses and lawns (about 10%) and as a preservative additive in some paints (about 13%), resins, emulsions, and coatings.
Chlorothalonil is commercially available in many different formulations and delivery methods. It is applied as a dust, dry or water-soluble grains, a wettable powder, a liquid spray, a fog, and a dip. It may be applied by hand, by ground sprayer, or by aircraft.
Mechanism of action
Chlorothalonil reacts with glutathione giving an glutathione adduct with elimination of HCl. Its mechanism of action is similar to that of trichloromethyl sulfenyl fungicides such as captan and folpet.
Toxicity
Acute
According to the EPS, chlorothalonil is a toxicity category I eye irritant, producing severe eye irritation. It is in toxicity category II, "moderately toxic", if inhaled (inhaled 0.094 mg/L in rats.) For skin contact and ingestion, chlorothalonil is rated toxicity category IV, "practically nontoxic", meaning the oral and dermal is greater than 10,000 mg/kg.
Chronic
Long-term exposure to chlorothalonil resulted in kidney damage and tumors in
|
https://en.wikipedia.org/wiki/Luis%20Caffarelli
|
Luis Ángel Caffarelli (; born December 8, 1948) is an Argentine–American mathematician. He studies partial differential equations and their applications.
Career
Caffarelli was born and grew up in Buenos Aires. He obtained his Masters of Science (1968) and Ph.D. (1972) at the University of Buenos Aires. His Ph.D. advisor was Calixto Calderón. He currently holds the Sid Richardson Chair at the University of Texas at Austin. He also has been a professor at the University of Minnesota, the University of Chicago, and the Courant Institute of Mathematical Sciences at New York University. From 1986 to 1996 he was a professor at the Institute for Advanced Study in Princeton.
Research
Caffarelli received recognition with "The regularity of free boundaries in higher dimensions" published in 1977 in Acta Mathematica. He is considered an expert in free boundary problems and nonlinear partial differential equations. He proved several regularity results for fully nonlinear elliptic equations including the Monge-Ampere equation, and also contributed to homogenization. He is also interested in integro-differential equations.
One of his most cited results regards the Partial regularity of suitable weak solutions of the Navier–Stokes equations; it was obtained in 1982 in collaboration with Louis Nirenberg and Robert V. Kohn.
Awards and recognition
In 1991 he was elected to the U.S. National Academy of Sciences. He was awarded honorary doctorates by the École Normale Supérieure, Paris, the University of Notre Dame, the Universidad Autónoma de Madrid, and the Universidad de La Plata, Argentina. He received the Bôcher Memorial Prize in 1984. He is listed as an ISI highly cited researcher.
In 2003 Konex Foundation from Argentina granted him the Diamond Konex Award, one of the most prestigious awards in Argentina, as the most important Scientist of his country in the last decade. In 2005, he received the prestigious Rolf Schock Prize of the Royal Swedish Academy of Sciences "for his
|
https://en.wikipedia.org/wiki/Melittin
|
Melittin is the main component (40–60% of the dry weight) and the major pain-producing substance of honeybee (Apis mellifera) venom. Melittin is a basic peptide consisting of 26 amino acids.
Function
The principal function of melittin as a component of bee venom is to cause pain and destruction of tissue of intruders that threaten a beehive. However, in honey bees, melittin is not only expressed in the venom gland, but also in other tissues when infected with pathogens. The two venom molecules, melittin and secapin, that are over-expressed in honey bees infected with various pathogens, possibly indicate a role for melittin in the immune response of bees to infectious diseases.
Structure
Melittin is a small peptide with no disulfide bridge; the N-terminal part of the molecule is predominantly hydrophobic and the C-terminal part is hydrophilic and strongly basic. In water, it forms a tetramer but it also can spontaneously integrate itself into cell membranes.
Mechanism of action
Injection of melittin into animals and humans causes pain sensation. It has strong surface effects on cell membranes causing pore-formation in epithelial cells and the destruction of red blood cells. Melittin also activates nociceptor (pain receptor) cells through a variety of mechanisms.
Melittin can open thermal nociceptor TRPV1 channels via cyclooxygenase metabolites resulting in depolarization of nociceptor cells. The pore forming effects in cells causes the release of pro-inflammatory cytokines. It also activates G-protein-coupled receptor-mediated opening of transient receptor potential channels. Finally melittin up-regulates the expression of Nav1.8 and Nav1.9 sodium channels in nociceptor cell causing long term action potential firing and pain sensation.
Melittin inhibits protein kinase C, Ca2+/calmodulin-dependent protein kinase II, myosin light chain kinase, and Na+/K+-ATPase (synaptosomal membrane). Melittin blocks transport pumps such as the Na+-K+-ATPase and the H+
|
https://en.wikipedia.org/wiki/Onion%20routing
|
Onion routing is a technique for anonymous communication over a computer network. In an onion network, messages are encapsulated in layers of encryption, analogous to the layers of an onion. The encrypted data is transmitted through a series of network nodes called "onion routers," each of which "peels" away a single layer, revealing the data's next destination. When the final layer is decrypted, the message arrives at its destination. The sender remains anonymous because each intermediary knows only the location of the immediately preceding and following nodes. While onion routing provides a high level of security and anonymity, there are methods to break the anonymity of this technique, such as timing analysis.
History
Onion routing was developed in the mid-1990s at the U.S. Naval Research Laboratory by employees Paul Syverson, Michael G. Reed, and David Goldschlag to protect U.S. intelligence communications online. It was then refined by the Defense Advanced Research Projects Agency (DARPA) and patented by the Navy in 1998.
This method was publicly released by the same employees through publishing an article in the IEEE Journal on Selected Areas in Communications the same year. It depicted the use of the method to protect the user from the network and outside observers who eavesdrop and conduct traffic analysis attacks. The most important part of this research is the configurations and applications of onion routing on the existing e-services, such as Virtual private network, Web-browsing, Email, Remote login, and Electronic cash.
Based on the existing onion routing technology, computer scientists Roger Dingledine and Nick Mathewson joined Paul Syverson in 2002 to develop what has become the largest and best-known implementation of onion routing, then called The Onion Routing project (Tor project).
After the Naval Research Laboratory released the code for Tor under a free license, Dingledine, Mathewson and five others founded The Tor Project as a non-profit o
|
https://en.wikipedia.org/wiki/Biohydrometallurgy
|
Biohydrometallurgy is a technique in the world of metallurgy that utilizes biological agents (bacteria) to recover and treat metals such as copper. Modern biohydrometallurgy advances started with the bioleaching of copper more efficiently in the 1950's
Important Definitions
Bio: Shortened form of Biology; refers to usage of bacteria.
Hydro: Term referring to the usage of water; process occurs in aqueous environments
Metallurgy: A process involving the separating and refining of metals from other substances;
Bioleaching: Using biological agents (bacteria) to extract metals or soils; general term used to encompass all forms biotechnological forms of extraction (hydrometallurgy, biohydrometallurgy, biomining, etc)
General Information
Interdisciplinary field involving processes that
make use of microbes, usually bacteria and archaea
mainly take place in aqueous environment
deal with metal production and treatment of metal containing materials and solutions
"Biohydrometallurgy may generally referred to as the branch of biotechnology dealing with the study and application of the economic potential of the interactions between microbes and minerals. It concerns, thus, all those engaged, directly or indirectly, in the exploitation of mineral resources and in environmental protection: geologists, economic geologists, mining engineers, metallurgists, hydrometallurgists, chemists and chemical engineers. In addition to these specialists, there are the microbiologists whose work is indispensable in the design, implementation and running of biohydrometallurgical processes."
Biohydrometallurgy was first used more than 300 years ago to recover copper. The uses have evolved to extracting gold, uranium, and other metals.
Hydrometallurgy
Hydrometallurgy refers to a specific process involving the chemical properties of water to create an aqueous solution for metal extraction through a series of chemical reactions
Biohydrometallurgy as a Science
Biohydrometallurgy repre
|
https://en.wikipedia.org/wiki/Corn%20starch
|
Corn starch, maize starch, or cornflour (British English) is the starch derived from corn (maize) grain. The starch is obtained from the endosperm of the kernel. Corn starch is a common food ingredient, often used to thicken sauces or soups, and to make corn syrup and other sugars. Corn starch is versatile, easily modified, and finds many uses in industry such as adhesives, in paper products, as an anti-sticking agent, and textile manufacturing. It has medical uses as well, such as to supply glucose for people with glycogen storage disease.
Like many products in dust form, it can be hazardous in large quantities due to its flammability—see dust explosion. When mixed with a fluid, corn starch can rearrange itself into a non-Newtonian fluid. For example, adding water transforms corn starch into a material commonly known as oobleck while adding oil transforms corn starch into an electrorheological (ER) fluid. The concept can be explained through the mixture termed "cornflour slime".
History
Until 1851, corn starch was used primarily for starching laundry and for other industrial uses.
A method to produce pure culinary starch from maize was patented by John Polson of Brown & Polson, in Paisley, Scotland in 1854. This was sold as "Patented Corn Flour". Brown & Polson were muslin manufacturers who had been producing laundry starch for the Paisley shawl industry and would become the largest starch producers in the UK.
Uses
Although mostly used for cooking and as a household item, corn starch is used for many purposes in several industries, ranging from its use as a chemical additive for certain products, to medical therapy for certain illnesses.
Culinary
Corn starch is used as a thickening agent in liquid-based foods (e.g., soup, sauces, gravies, custard), usually by mixing it with a lower-temperature liquid to form a paste or slurry. It is sometimes preferred over flour alone because it forms a translucent, rather than opaque mixture. As the starch is heated ov
|
https://en.wikipedia.org/wiki/Gamma-glutamyltransferase%206
|
Gamma-glutamyltransferase 6 is a protein that in humans is encoded by the GGT6 gene.
Function
GGT6 belongs to the gamma-glutamyltransferase (GGT; EC 2.3.2.2) gene family. GGT is a membrane-bound extracellular enzyme that cleaves gamma-glutamyl peptide bonds in glutathione and other peptides and transfers the gamma-glutamyl moiety to acceptors. GGT is also key to glutathione homeostasis because it provides substrates for glutathione synthesis (Heisterkamp et al., 2008 [PubMed 18357469]).
|
https://en.wikipedia.org/wiki/Aj%C3%AD%20%28sauce%29
|
Ají is a spicy sauce that contains ají peppers, oil, tomatoes, cilantro (coriander), garlic, onions, and water. It is served as a condiment to complement main dishes, most oftentimes in Latin American cuisines, and prepared by blending its ingredients using a food processor or blender. Although ají sauce recipes can vary from person to person, there are generally country-specific and region-specific varieties.
Description
Ají is a spicy sauce made from ají peppers that is usually served to accompany other dishes in a variety of Latin American cuisines. Its most basic ingredients include ají peppers, water, oil, garlic, cilantro, and salt. Ingredients are usually blended together using a blender or food processor.
Ají has been prepared in Andean countries such as Bolivia, Colombia, and Peru since at least the time of the Incas, who called it uchu. It is usually added to other foods such as anticuchos, chugchucaras, soup, chorizo, or empanadas.
In Colombia and Ecuador, food is traditionally milder, so ají can be added to almost any dish to add flavor and spice. Recipes vary dramatically from person to person and from region to region, depending on preference.
The core ingredient of ají sauce, ají peppers (Capsicum baccatum), was originally grown in South America. While these peppers have a Scoville Heat Unit of 30,000 - 50,000, depending on the variety of pepper and preparation technique, the spice level of ají sauce is variable.
Varieties
Chile
In Chile there is a related variety of the condiment known as ají chileno, which contains the additional ingredient of lemon juice.
Ecuador
In Ecuador, ají sauce is prepared using one of the over 30 ají pepper varieties available in the country. These ají peppers vary in spice level and this, combined with the amount of water used to dilute the sauce, can create variation in the level of spice between sauces. Some regions are also known for their addition of fruits, in addition to the basic ingredients, which lead
|
https://en.wikipedia.org/wiki/Promiscuous%20mode
|
In computer networking, promiscuous mode is a mode for a wired network interface controller (NIC) or wireless network interface controller (WNIC) that causes the controller to pass all traffic it receives to the central processing unit (CPU) rather than passing only the frames that the controller is specifically programmed to receive. This mode is normally used for packet sniffing that takes place on a router or on a computer connected to a wired network or one being part of a wireless LAN. Interfaces are placed into promiscuous mode by software bridges often used with hardware virtualization.
In IEEE 802 networks such as Ethernet or IEEE 802.11, each frame includes a destination MAC address. In non-promiscuous mode, when a NIC receives a frame, it drops it unless the frame is addressed to that NIC's MAC address or is a broadcast or multicast addressed frame. In promiscuous mode, however, the NIC allows all frames through, thus allowing the computer to read frames intended for other machines or network devices.
Many operating systems require superuser privileges to enable promiscuous mode. A non-routing node in promiscuous mode can generally only monitor traffic to and from other nodes within the same broadcast domain (for Ethernet and IEEE 802.11) or ring (for Token Ring). Computers attached to the same Ethernet hub satisfy this requirement, which is why network switches are used to combat malicious use of promiscuous mode. A router may monitor all traffic that it routes.
Promiscuous mode is often used to diagnose network connectivity issues. There are programs that make use of this feature to show the user all the data being transferred over the network. Some protocols like FTP and Telnet transfer data and passwords in clear text, without encryption, and network scanners can see this data. Therefore, computer users are encouraged to stay away from insecure protocols like telnet and use more secure ones such as SSH.
Detection
As promiscuous mode can be used
|
https://en.wikipedia.org/wiki/Sudan%20I
|
Sudan I (also commonly known as CI Solvent Yellow 14 and Solvent Orange R) is an organic compound, typically classified as an azo dye. It is an intensely orange-red solid that is added to colourise waxes, oils, petrol, solvents, and polishes. Sudan I has also been adopted for colouring various foodstuffs, especially curry powder and chili powder, although the use of Sudan I in foods is now banned in many countries, because Sudan I, Sudan III, and Sudan IV have been classified as category 3 carcinogens (not classifiable as to its carcinogenicity to humans) by the International Agency for Research on Cancer. Sudan I is still used in some orange-coloured smoke formulations and as a colouring for cotton refuse used in chemistry experiments.
Application
The Sudan dyes are a group of azo compounds which have been used to color hydrocarbon solvents, oils, fats, waxes, shoes, and floor polishes. As recently as 1974, about of Sudan I, of Sudan II, of Sudan III, and of Sudan IV were produced in the United States.
Sudan I and Sudan III (1-(4-(phenyldiazenyl)phenyl) azonaphthalen-2-ol) are used for mostly the same application. Sudan III melts at a 68 °C higher temperature than Sudan I.
Synthesis
The synthesis of Sudan I involves the reaction of phenyldiazonium salts with 2-naphthol.
Sudan I suffers from oxidative photo-degradation by two mechanisms, singlet oxygen degradation and free radical degradation, decreasing its fastness on materials.
Degradation and metabolism
The metabolism of Sudan I, as characterized in rabbits, involves both oxidative or reductive reactions.
Azo-reduction of Sudan I produces aniline and 1-amino-2-naphthol, and this reaction seems to be responsible for the detoxification. In vivo, after oxidation of Sudan I, C-hydroxylated metabolites are formed as major oxidation products and are excreted in urine. These metabolites are also found after oxidation with rat hepatic microsomes in vitro.
The C-hydroxylated metabolites may be considered as th
|
https://en.wikipedia.org/wiki/NDUFS2
|
NADH dehydrogenase [ubiquinone] iron-sulfur protein 2, mitochondrial (NDUFS2) also known as NADH-ubiquinone oxidoreductase 49 kDa subunit is an enzyme that in humans is encoded by the NDUFS2 gene. The protein encoded by this gene is a core subunit of the mitochondrial membrane respiratory chain NADH dehydrogenase (complex I). Mutations in this gene are associated with mitochondrial complex I deficiency.
Structure
NDUFS2 is located on the q arm of chromosome 1 in position 23.3 and has 15 exons. The NDUFS2 gene produces a 52.5 kDa protein composed of 463 amino acids. NDUFS2, the protein encoded by this gene, is a member of the complex I 49 kDa subunit family. It is a peripheral membrane protein on the matrix side of the inner mitochondrial membrane. It contains a cofactor binding site for a [4Fe-4S] cluster, a transit peptide, 5 turns, 11 beta strands, and 18 alpha helixes. Alternatively spliced transcript variants encoding different isoforms have been found for this gene.
Function
Mitochondrial complex I is the first multimeric complex of the respiratory chain that catalyzes the NADH oxidation with concomitant ubiquinone reduction and proton ejection out of the mitochondria. Mammalian mitochondrial complex I is an assembly of at least 43 different subunits. Seven of the subunits are encoded by the mitochondrial genome; the remainder are the products of nuclear genes. The iron-sulfur protein (IP) fraction of complex I is made up of 7 subunits, including NDUFS2. Dimethylation at Arg-118 by NDUFAF7 takes place after NDUFS2 assembles into the complex I, leading to the stabilization of the early intermediate complex.
Clinical significance
Mutations in the NDUFS2 gene are associated with Mitochondrial Complex I Deficiency, which is autosomal recessive. This deficiency is the most common enzymatic defect of the oxidative phosphorylation disorders. Mitochondrial complex I deficiency shows extreme genetic heterogeneity and can be caused by mutation in nuclear-encoded
|
https://en.wikipedia.org/wiki/Guar%20gum
|
Guar gum, also called guaran, is a galactomannan polysaccharide extracted from guar beans that has thickening and stabilizing properties useful in food, feed, and industrial applications. The guar seeds are mechanically dehusked, hydrated, milled and screened according to application. It is typically produced as a free-flowing, off-white powder.
Production and trade
The guar bean is principally grown in India, Pakistan, the United States, Australia and Africa. India is the largest producer, accounting for nearly 80% of the world production. In India, Rajasthan, Gujarat, and Haryana are the main producing regions. The US has produced 4,600 to 14,000 tonnes of guar over the last 5 years. Texas acreage since 1999 has fluctuated from about 7,000 to 50,000 acres. The world production for guar gum and its derivatives is about 1.0 million tonnes. Non-food guar gum accounts for about 40% of the total demand.
Properties
Chemical composition
Chemically, guar gum is an exo-polysaccharide composed of the sugars galactose and mannose. The backbone is a linear chain of β 1,4-linked mannose residues to which galactose residues are 1,6-linked at every second mannose, forming short side-branches. Guar gum has the ability to withstand temperatures of 80 °C (176 °F) for five minutes.
Solubility and viscosity
Guar gum is more soluble than locust bean gum due to its extra galactose branch points. Unlike locust bean gum, it is not self-gelling. Either borax or calcium can cross-link guar gum, causing it to gel. In water, it is nonionic and hydrocolloidal. It is not affected by ionic strength or pH, but will degrade at extreme pH and temperature (e.g. pH 3 at 50 °C). It remains stable in solution over pH range 5–7. Strong acids cause hydrolysis and loss of viscosity and alkalies in strong concentration also tend to reduce viscosity. It is insoluble in most hydrocarbon solvents. The viscosity attained is dependent on time, temperature, concentration, pH, rate of agitation and particl
|
https://en.wikipedia.org/wiki/Row%20equivalence
|
In linear algebra, two matrices are row equivalent if one can be changed to the other by a sequence of elementary row operations. Alternatively, two m × n matrices are row equivalent if and only if they have the same row space. The concept is most commonly applied to matrices that represent systems of linear equations, in which case two matrices of the same size are row equivalent if and only if the corresponding homogeneous systems have the same set of solutions, or equivalently the matrices have the same null space.
Because elementary row operations are reversible, row equivalence is an equivalence relation. It is commonly denoted by a tilde (~).
There is a similar notion of column equivalence, defined by elementary column operations; two matrices are column equivalent if and only if their transpose matrices are row equivalent. Two rectangular matrices that can be converted into one another allowing both elementary row and column operations are called simply equivalent.
Elementary row operations
An elementary row operation is any one of the following moves:
Swap: Swap two rows of a matrix.
Scale: Multiply a row of a matrix by a nonzero constant.
Pivot: Add a multiple of one row of a matrix to another row.
Two matrices A and B are row equivalent if it is possible to transform A into B by a sequence of elementary row operations.
Row space
The row space of a matrix is the set of all possible linear combinations of its row vectors. If the rows of the matrix represent a system of linear equations, then the row space consists of all linear equations that can be deduced algebraically from those in the system. Two m × n matrices are row equivalent if and only if they have the same row space.
For example, the matrices
are row equivalent, the row space being all vectors of the form . The corresponding systems of homogeneous equations convey the same information:
In particular, both of these systems imply every equation of the form
Equivalence of the defin
|
https://en.wikipedia.org/wiki/Normal%20mode
|
A normal mode of a dynamical system is a pattern of motion in which all parts of the system move sinusoidally with the same frequency and with a fixed phase relation. The free motion described by the normal modes takes place at fixed frequencies. These fixed frequencies of the normal modes of a system are known as its natural frequencies or resonant frequencies. A physical object, such as a building, bridge, or molecule, has a set of normal modes and their natural frequencies that depend on its structure, materials and boundary conditions.
The most general motion of a linear system is a superposition of its normal modes. The modes are normal in the sense that they can move independently, that is to say that an excitation of one mode will never cause motion of a different mode. In mathematical terms, normal modes are orthogonal to each other.
General definitions
Mode
In the wave theory of physics and engineering, a mode in a dynamical system is a standing wave state of excitation, in which all the components of the system will be affected sinusoidally at a fixed frequency associated with that mode.
Because no real system can perfectly fit under the standing wave framework, the mode concept is taken as a general characterization of specific states of oscillation, thus treating the dynamic system in a linear fashion, in which linear superposition of states can be performed.
Classical examples include
In a mechanical dynamical system, a vibrating rope is the most clear example of a mode, in which the rope is the medium, the stress on the rope is the excitation, and the displacement of the rope with respect to its static state is the modal variable.
In an acoustic dynamical system, a single sound pitch is a mode, in which the air is the medium, the sound pressure in the air is the excitation, and the displacement of the air molecules is the modal variable.
In a structural dynamical system, a high tall building oscillating under its most flexural axis is a mo
|
https://en.wikipedia.org/wiki/DigiCipher%202
|
DigiCipher 2, or simply DCII, is a proprietary standard format of digital signal transmission and it doubles as an encryption standard with MPEG-2/MPEG-4 signal video compression used on many communications satellite television and audio signals. The DCII standard was originally developed in 1997 by General Instrument, which then became the Home and Network Mobility division of Motorola, then bought by Google in Aug 2011, and lastly became the Home portion of the division to Arris.
The original attempt for a North American digital signal encryption and compression standard was DigiCipher 1, which was used most notably in the now-defunct PrimeStar medium-power direct broadcast satellite (DBS) system during the early 1990s. The DCII standard predates wide acceptance of DVB-based digital terrestrial television compression (although not cable or satellite DVB) and therefore is incompatible with the DVB standard.
Approximately 70% of newer first-generation digital cable networks in North America use the 4DTV/DigiCipher 2 format. The use of DCII is most prevalent in North American digital cable television set-top boxes. DCII is also used on Motorola's 4DTV digital satellite television tuner and Shaw Direct's DBS receiver.
The DigiCipher 2 encryption standard was reverse engineered in 2016.
Technical specifications
DigiCipher II uses QPSK and BPSK at the same time. The primary difference between DigiCipher 2 and DVB lies in how each standard handles SI metadata, or System Information, where DVB reserves packet identifiers from 16 to 31 for metadata, DigiCipher reserves only packet identifier 8187 for its master guide table which acts as a look-up table for all other metadata tables. DigiCipher 2 also extends the MPEG program number that is assigned for each service in a transport stream with the concept of a virtual channel number, whereas the DVB system never defined this type of remapping preferring to use a registry of network identifiers to further differentiate p
|
https://en.wikipedia.org/wiki/VBites
|
VBites is a plant-based meat alternative and vegan cheese company that develops 140 products and retails in 24 countries worldwide. All its foods are manufactured in the UK at VBites' factory in Corby. The company was bought by Heather Mills in 2009.
The manufacturing company was founded in 1993 and mainly traded as the Redwood Wholefood Company before its name was changed to VBites in 2013 to match the restaurant she owned.
Brands include Cheezly - a selection of Cheese equivalents, since 2017 Domino's Pizza have been using it for their vegan options.
See also
List of vegetarian restaurants
External links
Official website
|
https://en.wikipedia.org/wiki/Glycoinformatics
|
Glycoinformatics is a field of bioinformatics that pertains to the study of carbohydrates involved in protein post-translational modification. It broadly includes (but is not restricted to) database, software, and algorithm development for the study of carbohydrate structures, glycoconjugates, enzymatic carbohydrate synthesis and degradation, as well as carbohydrate interactions. Conventional usage of the term does not currently include the treatment of carbohydrates from the better-known nutritive aspect.
Issues to consider
Even though glycosylation is the most common form of protein modification, with highly complex carbohydrate structures, the bioinformatics on glycome is still very poor.
Unlike proteins and nucleic acids which are linear, carbohydrates are often branched and extremely complex. For instance, just four sugars can be strung together to form more than 5 million different types of carbohydrates or nine different sugars may be assembled into 15 million possible four-sugar-chains.
Also, the number of simple sugars that make up glycans is more than the number of nucleotides that make up DNA or RNA. Therefore, it is more computationally expensive to evaluate their structures.
One of the main constrains in the glycoinformatics is the difficulty of representing sugars in the sequence form especially due to their branching nature. Owing to the lack of a genetic blue print, carbohydrates do not have a "fixed" sequence. Instead, the sequence is largely determined by the presence of a variety of enzymes, their kinetic differences and variations in the biosynthetic micro-environment of the cells. This increases the complexity of analysis and experimental reproducibility of the carbohydrate structure of interest. It is for this reason that carbohydrates are often considered as the "information poor" molecules.
Databases
Table of major glyco-databases.
|
https://en.wikipedia.org/wiki/Quotition%20and%20partition
|
In arithmetic, quotition and partition are two ways of viewing fractions and division.
In quotition division one asks, "how many parts are there?"; while in partition division one asks, "what is the size of each part?".
For example, the expression can be constructed of either of two ways:
"How many parts of the size of 2 must be added to get the amount of 6?" (Quotition division)
One can write
Since it takes 3 parts, the conclusion is that
"What is the size of 2 equal parts whose sum is that of 6?". (Partition division)
One can write
Since the size of each part is 3, the conclusion is that
It is a fact of elementary theoretical mathematics that the numerical answer is always the same no matter which way you put it, 6 ÷ 2 = 3. This is essentially equivalent to the commutativity of multiplication in multiplication arithmetic.
Division involves thinking about a whole in terms of its parts. One frequent division special case, is that of a natural number (positive integers) of equal parts, is known to teachers as a partition or sharing: the whole entity becomes an integer number with equal parts.
What quotition focuses on, is explained by removing the word integer in the last sentence. Allow the number to be any fraction and you may have a quotition instead of a partition.
See also
List of partition topics
|
https://en.wikipedia.org/wiki/Allelomimetic%20behavior
|
Allelomimetic behavior or allomimetic behavior is a range of activities in which the performance of a behavior increases the probability of that behavior being performed by other nearby animals. Allelomimetic behavior is sometimes called contagious behavior and has strong components of social facilitation, imitation, and group coordination. It is usually considered to occur between members of the same species. An alternate definition is that allelomimetic behavior is a more intense or more frequent response or the initiation of an already known response, when others around the individual are engaged in the same behavior. It is often referred to as synchronous behavior, mimetic behavior, imitative behavior, and social facilitation.
Allelomimetic behavior is displayed in all animals and can occur in any stage of life, but usually starts at a young age. This behavior will continue throughout life, especially when an individual is living in a large group that emphasizes group cohesion. Cohesion is seen as a prerequisite for group living, with synchronous activity being crucial for social cohesion. However, animals in large cohesive groups face trade-offs when allelomimetic behavior is adopted. If the behavior is adopted then the risk of predation or capture decreases significantly but the inter-individual competition for immediate resources, such as food, mates, and space, will increase when cohesion is still stressed. Many collective group decisions in animals are the result of allelomimetism and can be explained by allelomimetic behaviors. Some examples are the cockroaches choosing a single aggregation site, schooling behaviors in fishes, and pheromone-based path selection in ants that allows all the workers to go down the same path to a specific food source. Allelomimetic behavior can also be seen as an animal welfare indicator. For example, if cattle do not have enough room to all lie down simultaneously then it indicates that there are not enough resources present
|
https://en.wikipedia.org/wiki/Automatic%20Transmitter%20Identification%20System%20%28television%29
|
The Automatic Transmitter Identification System (ATIS) is a communications protocol used for the station identification of television channels carried on satellite television.
ATIS is only required for analog television transmission and only via satellites or earth stations under United States jurisdiction. It is continuously repeated whilst an earth station is using a transponder on a satellite. The Federal Communications Commission (FCC) rules for ATIS are set forth in 47 CFR §25.281.
The system was developed in response to the "Captain Midnight" satellite jamming incident. In 2009, HBO and Elmer Musser were awarded a Technology & Engineering Emmy Award for ATIS.
ATIS is a Morse code transmission sent on a subcarrier of 7.1 MHz and must activate automatically any time the station is transmitting. The center frequency must be no more than 25 kHz from this nominal value and the frequency deviation must be no more than 25 kHz peak when being modulated. Injection must be at a minimum of −26dB referenced to the unmodulated carrier. The tone used to modulate the subcarrier is nominally 1200 Hz, but may vary by as much as 800 Hz (400 to 2000 Hz).
The ATIS message is sent at a transmission rate of 15 to 25 words per minute and must not exceed 30 seconds in its entire length. The message includes the FCC-assigned callsign of the earth station, its telephone number, a ten-digit serial number, and sometimes other information which is voluntary. The telephone number must immediately connect to personnel who can resolve radio interference and other frequency coordination issues. The serial number cannot be easily changed.
ATIS encoders must be included in the uplink airchain of all transmitters as of March 1, 1991 "in a method that cannot easily be defeated".
Other subcarriers piggybacked on the video carrier are also included in transmissions, often at 6.2 MHz and 6.8 MHz, just above the bandpass of the video signal (even if scrambled). These carry left and right stereo
|
https://en.wikipedia.org/wiki/Truchet%20tiles
|
In information visualization and graphic design, Truchet tiles are square tiles decorated with patterns that are not rotationally symmetric. When placed in a square tiling of the plane, they can form varied patterns, and the orientation of each tile can be used to visualize information associated with the tile's position within the tiling.
Truchet tiles were first described in a 1704 memoir by Sébastien Truchet entitled "Mémoire sur les combinaisons", and were popularized in 1987 by Cyril Stanley Smith.
Variations
Contrasting triangles
The tile originally studied by Truchet is split along the diagonal into two triangles of contrasting colors. The tile has four possible orientations.
Some examples of surface filling made tiling such a pattern.
With a scheme:
With random placement:
Quarter-circles
A second common form of the Truchet tiles, due to , decorates each tile with two quarter-circles connecting the midpoints of adjacent sides. Each such tile has two possible orientations.
We have such a tiling:
This type of tile has also been used in abstract strategy games Trax and the Black Path Game, prior to Smith's work.
Diagonal
A labyrinth can be generated by tiles in the form of a white square with a black diagonal. As with the quarter-circle tiles, each such tile has two orientations.
The connectivity of the resulting labyrinth can be analyzed mathematically using percolation theory as bond percolation at the critical point of a diagonally-oriented grid.
Nick Montfort considers the single line of Commodore 64 BASIC required to generate such patterns - 10 PRINT CHR$(205.5+RND(1)); : GOTO 10 - to be "a concrete poem, a found poem".
See also
Girih tiles
Wallpaper group
Wang tiles
|
https://en.wikipedia.org/wiki/Stem%20cell%20laws%20and%20policy%20in%20China
|
The laws and policies regarding stem cell research in the People's Republic of China are relatively relaxed in comparison to that of other nations. The reason for this is due to different traditional and cultural views in relation to that of the West.
Laws and regulations
China has one of the most unrestrictive embryonic stem cell research policies in the world. In recent years, seeing the research opportunities that China's lax regulations provide, many expatriate Chinese scientists from the West are returning to China to establish stem cell research centers and laboratories there.
As a result of the increased interest in this field of research, in 2003, the People's Republic of China Ministry of Science and Technology and Ministry of Health issued official ethical guidelines for human embryonic stem cell research in its territories. The guidelines strictly forbid any research aimed at human reproductive cloning and require that the embryos used for stem cell research come only from:
Spared gamete or blastocyst after in vitro fertilization (IVF) procedures;
Fetal cells from accidental spontaneous or voluntarily selected abortions;
Blastocyst or parthenogenetic split blastocyst obtained by somatic cell nuclear transfer technology; or
Germ cells voluntarily donated.
American scientific journals Science and Nature have both reported in recent years that China's stem cell programs hold potential, and in 2004 a delegation from Britain's Department of Trade and Industry concluded more emphatically that Chinese research in the field was already world-class. Funding for stem cell research by the Chinese government is extremely limited compared to Western nations, with the Chinese Ministry of Science and Technology planning to devote between US$33 million and US$132 million on stem cell research during the next 5 years. By contrast, the state of California alone has earmarked US$3 billion to fund stem cell research at California institutions during the next decade.
|
https://en.wikipedia.org/wiki/Calcium%20in%20biology
|
Calcium ions (Ca2+) contribute to the physiology and biochemistry of organisms' cells. They play an important role in signal transduction pathways, where they act as a second messenger, in neurotransmitter release from neurons, in contraction of all muscle cell types, and in fertilization. Many enzymes require calcium ions as a cofactor, including several of the coagulation factors. Extracellular calcium is also important for maintaining the potential difference across excitable cell membranes, as well as proper bone formation.
Plasma calcium levels in mammals are tightly regulated, with bone acting as the major mineral storage site. Calcium ions, Ca2+, are released from bone into the bloodstream under controlled conditions. Calcium is transported through the bloodstream as dissolved ions or bound to proteins such as serum albumin. Parathyroid hormone secreted by the parathyroid gland regulates the resorption of Ca2+ from bone, reabsorption in the kidney back into circulation, and increases in the activation of vitamin D3 to calcitriol. Calcitriol, the active form of vitamin D3, promotes absorption of calcium from the intestines and bones. Calcitonin secreted from the parafollicular cells of the thyroid gland also affects calcium levels by opposing parathyroid hormone; however, its physiological significance in humans is dubious.
Intracellular calcium is stored in organelles which repetitively release and then reaccumulate Ca2+ ions in response to specific cellular events: storage sites include mitochondria and the endoplasmic reticulum.
Characteristic concentrations of calcium in model organisms are: in E. coli 3mM (bound), 100nM (free), in budding yeast 2mM (bound), in mammalian cell 10-100nM (free) and in blood plasma 2mM.
Humans
In 2020, calcium was the 204th most commonly prescribed medication in the United States, with more than 2million prescriptions.
Dietary recommendations
The U.S. Institute of Medicine (IOM) established Recommended Dietary Allowanc
|
https://en.wikipedia.org/wiki/Starent%20Networks
|
Starent Networks was an information technology and computer networking products company. It was headquartered in Tewksbury, Massachusetts, USA, with major engineering facilities in Pune, India and Bangalore, India. Starent Networks was bought by Cisco in 2009.
Starent was a provider of infrastructure products that enabled mobile operators to deliver multimedia services such as video, Internet access, voice-over-IP, e-mail, mobile TV, photo sharing, and gaming to their subscribers. The products act as gateways that connect the radio access network to the IP network. The gateways allow the service provider to have fine control over the end user experience of the high bandwidth/low latency applications, transaction accounting details and access control.
On October 13, 2009, Cisco announced a plan to purchase Starent for US$2.9 billion.
On December 18, 2009, Cisco completed acquisition of Starent Networks.
|
https://en.wikipedia.org/wiki/GNF6702
|
GNF6702 is the name for a broad-spectrum antiprotozoal drug invented by researchers working at the Genomics Institute of the Novartis Research Foundation in 2013, with activity against leishmaniasis, Chagas disease and sleeping sickness. These three diseases are caused by related kinetoplastid parasites, which share similar biology. GNF6702 acts as allosteric proteasome inhibitor which was effective against infection with any of the three protozoal diseases in mice, while having little evident toxicity to mammalian cells.
See also
Proteasome inhibitor
|
https://en.wikipedia.org/wiki/Fetus
|
A fetus or foetus (; : fetuses, feti, foetuses, or foeti) is the unborn offspring that develops from an animal embryo. Following embryonic development the fetal stage of development takes place. In human prenatal development, fetal development begins from the ninth week after fertilization (or eleventh week gestational age) and continues until birth. Prenatal development is a continuum, with no clear defining feature distinguishing an embryo from a fetus. However, a fetus is characterized by the presence of all the major body organs, though they will not yet be fully developed and functional and some not yet situated in their final anatomical location.
Etymology
The word fetus (plural fetuses or feti) is related to the Latin fētus ("offspring", "bringing forth", "hatching of young") and the Greek "φυτώ" to plant. The word "fetus" was used by Ovid in Metamorphoses, book 1, line 104.
The predominant British, Irish, and Commonwealth spelling is foetus, which has been in use since at least 1594. The spelling with -oe- arose in Late Latin, in which the distinction between the vowel sounds -oe- and -e- had been lost. This spelling is the most common in most Commonwealth nations, except in the medical literature, where the fetus is used. The more classical spelling fetus is used in Canada and the United States. In addition, fetus is now the standard English spelling throughout the world in medical journals. The spelling faetus was also used historically.
Development in humans
Weeks 9 to 16 (2 to 3.6 months)
In humans, the fetal stage starts nine weeks after fertilization. At the start of the fetal stage, the fetus is typically about in length from crown-rump, and weighs about 8 grams. The head makes up nearly half of the size of the fetus. Breathing-like movements of the fetus are necessary for the stimulation of lung development, rather than for obtaining oxygen. The heart, hands, feet, brain, and other organs are present, but are only at the beginning of developme
|
https://en.wikipedia.org/wiki/Plate%20%28anatomy%29
|
A plate in animal anatomy may refer to several things:
Flat bones (examples: bony plates, dermal plates) of vertebrates
an appendage of the Stegosauria group of dinosaurs
articulated armoured plates covering the head of thorax of Placodermi (literally "plate-skinned"), an extinct class of prehistoric fish (including skull, thoracic and tooth plates)
bony shields of the Ostracoderms (armored jawless fishes) such as the dermal head armour of members of the class Pteraspidomorphi that include dorsal, ventral, rostral and pineal plates
plates of a carapace, such as the dermal plates of the shell of a turtle
dermal plates partly or completely covering the body of the fish in the order Gasterosteiformes that includes the sticklebacks and relatives
plates of dermal bones of the armadillo
Zygomatic plate, a bony plate derived from the flattened front part of the zygomatic arch (cheekbone) in rodent anatomy
Other flat structures
hairy plate-like keratin scales of the pangolin
Basal plate (disambiguation), several anatomy-related meanings
Other meanings in human anatomy
Alar plate, a neural structure in the embryonic nervous system
Cribriform plate, of the ethmoid bone (horizontal lamina) received into the ethmoidal notch of the frontal bone and roofs in the nasal cavities
Epiphyseal plate, a hyaline cartilage plate in the metaphysis at each end of a long bone
Lateral pterygoid plate of the sphenoid, a broad, thin and everted bone that forms the lateral part of a horseshoe like process that extends from the inferior aspect of the sphenoid bone
Nail plate, the hard and translucent portion of the nail
Perpendicular plate of the ethmoid bone (vertical plate), a thin, flattened lamina, polygonal in form, which descends from the under surface of the cribriform plate, and assists in forming the septum of the nose
Related structures
Scute, a bony external plate or scale overlaid with horn, as on the shell of a turtle, the skin of crocodilians and the feet of
|
https://en.wikipedia.org/wiki/Zolotarev%20polynomials
|
In mathematics, Zolotarev polynomials are polynomials used in approximation theory. They are sometimes used as an alternative to the Chebyshev polynomials where accuracy of approximation near the origin is of less importance. Zolotarev polynomials differ from the Chebyshev polynomials in that two of the coefficients are fixed in advance rather than allowed to take on any value. The Chebyshev polynomials of the first kind are a special case of Zolotarev polynomials. These polynomials were introduced by Russian mathematician Yegor Ivanovich Zolotarev in 1868.
Definition and properties
Zolotarev polynomials of degree in are of the form
where is a prescribed value for and the are otherwise chosen such that the deviation of from zero is minimum in the interval .
A subset of Zolotarev polynomials can be expressed in terms of Chebyshev polynomials of the first kind, . For
then
For values of greater than the maximum of this range, Zolotarev polynomials can be expressed in terms of elliptic functions. For , the Zolotarev polynomial is identical to the equivalent Chebyshev polynomial. For negative values of , the polynomial can be found from the polynomial of the positive value,
The Zolotarev polynomial can be expanded into a sum of Chebyshev polynomials using the relationship
In terms of Jacobi elliptic functions
The original solution to the approximation problem given by Zolotarev was in terms of Jacobi elliptic functions. Zolotarev gave the general solution where the number of zeroes to the left of the peak value () in the interval is not equal to the number of zeroes to the right of this peak (). The degree of the polynomial is . For many applications, is used and then only need be considered. The general Zolotarev polynomials are defined as
where
is the Jacobi eta function
is the incomplete elliptic integral of the first kind
is the quarter-wave complete elliptic integral of the first kind. That is,
is the Jacobi elliptic
|
https://en.wikipedia.org/wiki/Transrectal%20biopsy
|
Transrectal biopsy is a biopsy procedure in which a sample of tissue is removed from the prostate using a thin needle that is inserted through the rectum and into the prostate. Transrectal ultrasound (TRUS) is usually used to guide the needle. The sample is examined under a microscope to see if it contains cancer.
|
https://en.wikipedia.org/wiki/RNA%20modification%20database
|
This RNA modification databases are a compilation of databases and web portals and servers used for RNA modification. RNA modification occurs in all living organisms, and is one of the most evolutionarily conserved properties of RNAs. More than 100 different types of RNA modifications have been characterized across all living organisms. It can affect the activity, localization as well as stability of RNAs, and has been linked with human cancer and diseases.
RNA Modification Databases
|
https://en.wikipedia.org/wiki/Imaging%20genetics
|
Imaging genetics refers to the use of anatomical or physiological imaging technologies as phenotypic assays to evaluate genetic variation. Scientists that first used the term imaging genetics were interested in how genes influence psychopathology and used functional neuroimaging to investigate genes that are expressed in the brain (neuroimaging genetics).
Imaging genetics uses research approaches in which genetic information and fMRI data in the same subjects are combined to define neuro-mechanisms linked to genetic variation. With the images and genetic information, it can be determined how individual differences in single nucleotide polymorphisms, or SNPs, lead to differences in brain wiring structure, and intellectual function. Imaging genetics allows the direct observation of the link between genes and brain activity in which the overall idea is that common variants in SNPs lead to common diseases. A neuroimaging phenotype is attractive because it is closer to the biology of genetic function than illnesses or cognitive phenotypes.
Alzheimer's disease
By combining the outputs of the polygenic and neuro-imaging within a linear model, it has been shown that genetic information provides additive value in the task of predicting Alzheimer's disease (AD). AD traditionally has been considered a disease marked by neuronal cell loss and widespread gray matter atrophy and the apolipoprotein E allele (APOE4) is a widely confirmed genetic risk factor for late-onset AD.
Another gene risk variant is associated with Alzheimer's, which is known as the CLU gene risk variant. The CLU gene risk variant showed a distinct profile of lower white matter integrity that may increase vulnerability to developing AD later in life. Each CLU-C allele was associated with lower FA in frontal, temporal, parietal, occipital, and subcortical white matter. Brain regions with lower FA included corticocortical pathways previously demonstrated to have lower FA in AD patients and APOE4 carriers
|
https://en.wikipedia.org/wiki/Agrammatism
|
Agrammatism is a characteristic of non-fluent aphasia. Individuals with agrammatism present with speech that is characterized by containing mainly content words, with a lack of function words. For example, when asked to describe a picture of children playing in the park, the affected individual responds with, "trees..children..run." People with agrammatism may have telegraphic speech, a unique speech pattern with simplified formation of sentences (in which many or all function words are omitted), akin to that found in telegraph messages. Deficits in agrammaticism are often language-specific, however—in other words, "agrammaticism" in speakers of one language may present differently from in speakers of another.
Errors made in agrammatism depend on the severity of aphasia. In severe forms language production is severely telegraphic and in more mild to moderate cases necessary elements for sentence construction are missing. Common errors include errors in tense, number, and gender. Patients also find it very hard to produce sentences involving "movement" of elements, such as passive sentences, wh-questions or complex sentences.
Agrammatism is seen in many brain disease syndromes, including expressive aphasia and traumatic brain injury.
History
Agrammatism was first coined by Adolf Kussmaul in 1887 to explain the inability to form words grammatically and to syntactically order them into a sentence. Later on, Harold Goodglass defined the term as the omission of connective words, auxiliaries and inflectional morphemes, all of these generating a speech production with extremely rudimentary grammar. Agrammatism, today seen as a symptom of the Broca's syndrome (Tesak & Code, 2008), has been also referred as 'motor aphasia' (Goldstein, 1948), 'syntactic aphasia' (Wepman & Jones, 1964), 'efferent motor aphasia' (Luria, 1970), and 'non-fluent aphasia' (Goodglass et al., 1964).
The early accounts of agrammatism involved cases of German and French participants. The gre
|
https://en.wikipedia.org/wiki/Bayesian%20learning%20mechanisms
|
Bayesian learning mechanisms are probabilistic causal models used in computer science to research the fundamental underpinnings of machine learning, and in cognitive neuroscience, to model conceptual development.
Bayesian learning mechanisms have also been used in economics and cognitive psychology to study social learning in theoretical models of herd behavior.
See also
Active learning
Bayesian learning
Cognitive acceleration
Cognitivism (learning theory)
Constructivist epistemology
Developmental psychology
Fluid and crystallized intelligence
Inquiry-based learning
Kohlberg's stages of moral development
Theory theory
|
https://en.wikipedia.org/wiki/Adenylyl-sulfate%20kinase
|
In enzymology, an adenylyl-sulfate kinase () is an enzyme that catalyzes the chemical reaction
ATP + adenylyl sulfate ADP + 3'-phosphoadenylyl sulfate
Thus, the two substrates of this enzyme are ATP and adenylyl sulfate, whereas its two products are ADP and 3'-phosphoadenylyl sulfate.
This enzyme belongs to the family of transferases, specifically those transferring phosphorus-containing groups (phosphotransferases) with an alcohol group as acceptor. The systematic name of this enzyme class is ATP:adenylyl-sulfate 3'-phosphotransferase. Other names in common use include adenylylsulfate kinase (phosphorylating), 5'-phosphoadenosine sulfate kinase, adenosine 5'-phosphosulfate kinase, adenosine phosphosulfate kinase, adenosine phosphosulfokinase, adenosine-5'-phosphosulfate-3'-phosphokinase, and APS kinase. This enzyme participates in 3 metabolic pathways: purine metabolism, selenoamino acid metabolism, and sulfur metabolism.
This enzyme contains an ATP binding P-loop motif.
Structural studies
As of late 2007, 11 structures have been solved for this class of enzymes, with PDB accession codes , , , , , , , , , , and .
|
https://en.wikipedia.org/wiki/Octeract%20Engine
|
Octeract Engine is a proprietary massively parallel deterministic global optimization solver for general Mixed-Integer Nonlinear Programs (MINLP) and the current world record holder in MINLP performance.
It uses MPI as a means of accelerating solution times. It is known for breaking four consecutive performance world records in the Mittelmann MINLPLIB benchmark, as well as its native parallelism and high degree of configurability. The latest world records were set in April 2023, when it became the first optimization solver to ever solve all problems in the benchmark, with a world record setting unscaled shifted geometric mean of 36.8.
History
Octeract Engine was developed by Nikos Kazazakis and Gabriel Lau. The first public beta version of Octeract Engine was released in August 2019, and it came out of beta in August 2020.
Performance
Octeract Engine exhibits world-leading performance on a single thread, and also has the ability to speed up these single thread solution times by several fold through supercomputing. In 23 July 2022 it ranked first on the single thread Mittelmann MINLPLIB benchmark. This lead has been maintained for all releases of Octeract Engine hence.
As of August 2022 it is the first and only solver to solve the largest open transmission switching problems in the industry standard MINLPLIB library, namely transswitch2736spp
and transswitch2736spr.
World records
Octeract Engine currently holds two world records in MINLP benchmarks. The first world record, set in 20 April 2023, is in number of problems solved (100%). The second world record, also set in 20 April 2023, is the smallest unscaled shifted geometric mean ever achieved for this test set, which was 36.8. The runner-ups achieved means of 138.2 (BARON) and 380.5 (SCIP), with Couenne ranking last with a mean of 3304.4.
World record history
In 27 October 2022 Octeract Engine set its first world record by solving more 91% of problems in the benchmark, which was more than any solver had ever
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.