source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Oak%E2%80%93heath%20forest | An oak–heath forest is a plant community association and type of forest ecology. It is a deciduous forest type of well-drained, acidic soils, characterized by oaks (Quercus) and plants of the heath family (Ericaceae). It is commonly found in the high elevations of the eastern United States. Such forest areas typically have a dense fibrous root layer at the surface of the soil, and in many areas predominate on south-facing or southwest-facing slopes. Many of the existing oak–heath forests once featured American chestnut as an important canopy species.
Oaks
Oaks (Quercus) characteristic of oak–heath associations include white oak (Quercus alba), black oak (Quercus velutina), scarlet oak (Quercus coccinea), chestnut oak (Quercus prinus), and red oak (Quercus rubra).
Heaths
Heath plants common to this ecology include mountain-laurel (Kalmia latifolia), various blueberries (genus Vaccinium), huckleberries (genus Gaylussacia), sourwood or sorrel-tree (Oxydendrum arboreum), and azaleas and rhododendrons (genus Rhododendron). These are all usually shrubs, except for Oxydendrum, which is usually a small tree. There are also heaths that are sub-shrubs, usually trailing on the ground, including wintergreens (genus Chimaphila and Gaultheria) and trailing arbutus (Epigaea repens). |
https://en.wikipedia.org/wiki/Ebonite | Ebonite is a brand name for a material generically known as hard rubber, obtained via vulcanizing natural rubber for prolonged periods. Ebonite may contain from 25% to 80% sulfur and linseed oil. Its name comes from its intended use as an artificial substitute for ebony wood. The material has also been called vulcanite, although that name formally refers to the mineral vulcanite.
Charles Goodyear's brother, Nelson Goodyear, experimented with the chemistry of ebonite composites. In 1851, he used zinc oxide as a filler. Hugh Silver was responsible for giving it its name.
Properties
The sulfur percentage and the applied temperatures and duration of vulcanizing are the main variables that determine the technical properties of the hard rubber polysulfide elastomer. The occurring reaction is basically addition of sulfur at the double bonds, forming intramolecular ring structures, so a large portion of the sulfur is highly cross-linked in the form of intramolecular addition. As a result of having a maximum sulfur content up to 40%, it may be used to resist swelling and minimize dielectric loss. The strongest mechanical properties and greatest heat resistance is obtained with sulfur contents around 35% while the highest impact strength can be obtained with a lower sulfur content of 30%. The rigidity of hard rubber at room temperature is attributed to the van der Waals forces between the intramolecular sulfur atoms. Raising the temperature gradually increases the molecular vibrations that overcome the van der Waals forces making it elastic. Hard rubber has a content mixture dependent density around 1.1 to 1.2. When reheated hard rubber exhibits shape-memory effect and can be fairly easily reshaped within certain limits. Depending on the sulfur percentage hard rubber has a thermoplastic transition or softening temperature of .
The material is brittle, which produces problems in its use in battery cases for example, where the integrity of the case is vital to prevent leaka |
https://en.wikipedia.org/wiki/Zhores%20Alferov | Zhores Ivanovich Alferov (; ; 15 March 19301 March 2019) was a Soviet and Russian physicist and academic who contributed significantly to the creation of modern heterostructure physics and electronics. He shared the 2000 Nobel Prize in Physics for the development of the semiconductor heterojunction for optoelectronics. He also became a politician in his later life, serving in the lower house of the Russian parliament, the State Duma, as a member of the Communist Party from 1995.
Early life and education
Alferov was born in Vitebsk, Byelorussian SSR, Soviet Union, to a Russian father, Ivan Karpovich Alferov, a factory manager, and a Jewish mother, Anna Vladimirovna Rosenblum. He was named after French socialist Jean Jaurès while his older brother was named Marx after Karl Marx. Alferov graduated from secondary school in Minsk in 1947 and started Belarusian Polytechnic Academy. In 1952, he received his B.S. from the V. I. Ulyanov (Lenin) Electrotechnical Institute (LETI) in Leningrad. Starting in 1953, Alferov worked in the Ioffe Physico-Technical Institute of the Academy of Sciences of the Soviet Union. From the institute, he earned several scientific degrees: a Candidate of Sciences in Technology in 1961 and a Doctor of Sciences in Physics and Mathematics in 1970.
Alferov then served as the director of the Ioffe Institute from 1987 to 2003. He was elected a corresponding member of the Academy of Sciences of the Soviet Union in 1972, and a full member in 1979. From 1989, he was Vice-President of the USSR Academy of Sciences and President of its Saint Petersburg Scientific Center.
Research
Starting at Ioffe Institute in 1953, Alferov worked with a group led by Vladimir Tuchkevich, who became director of the Ioffe Institute in 1967, on planar semiconductor amplifiers for use in radio receivers. These planar semiconductor amplifiers would be referred to as transistors in the present day. Alferov's contribution included work on germanium diodes for use as a rectifier. |
https://en.wikipedia.org/wiki/Kaibun | Kaibun () is a Japanese equivalent of the palindrome, or in other words, a sentence that reads the same from the beginning to the end or from the end to the beginning. The unit of kaibun is mora, since the Japanese language uses syllabaries, hiragana and katakana.
Single word palindromes are not uncommon in Japanese. For example, Ku-ku (九九, multiplication table), Shi-n-bu-n-shi (新聞紙, newspaper), to-ma-to (トマト, tomato), etc. So kaibun usually refers to a palindromic sentence, but a passage can be a kaibun too.
The topic marker "wa" (は) can be treated as "ha" and small kana ゃ,ゅ and ょ are usually allowed to be interpreted as big kana や, ゆ and よ. In classics, diacritic marks are often ignored.
Rather than saying "read the same forwards and backwards", because Japanese is traditionally written vertically, Japanese people describe the word as being the same when read from the top (ue kara yomu) as when read from the bottom (shita kara yomu).
Famous kaibun
Ta-ke-ya-bu ya-ke-ta (竹薮焼けた) - A bamboo grove has been burned.
Wa-ta-shi ma-ke-ma-shi-ta-wa (私負けましたわ) - I have lost.
Na-ru-to wo to-ru-na (なるとを取るな) - Do not take my naruto (spiral-shaped fishcake).
Dan-su-ga sun-da (ダンスが済んだ) - The dance is over.
Shi-na-mo-n pa-n mo re-mo-n pa-n mo na-shi (シナモンパンもレモンパンも無し) - There is neither cinnamon bread nor lemon bread.
Na-ga-ki yo-no to-ho-no ne-bu-ri-no mi-na me-za-me na-mi-no-ri-bu-ne-no o-to-no-yo-ki-ka-na (長き世の 遠の眠りの 皆目覚め 波乗り船の 音の良きかな) Tanka - Everybody gets awakened from a long sleep and enjoys the sound of waves on which the boat is gliding along.
Yo-no-na-ka, ho-ka-ho-ka na-no-yo (世の中、ホカホカなのよ) - The world is a warm place.
Yo-no-na-ka-ne, ka-o-ka o-ka-ne-ka na-no-yo (世の中ね、顔かお金かなのよ) - What matters in the world is your appearance or money.
External links
Japanese word games
Palindromes
ja:回文 |
https://en.wikipedia.org/wiki/Otto%20Toeplitz | Otto Toeplitz (1 August 1881 – 15 February 1940) was a German mathematician working in functional analysis.
Life and work
Toeplitz was born to a Jewish family of mathematicians. Both his father and grandfather were Gymnasium mathematics teachers and published papers in mathematics. Toeplitz grew up in Breslau and graduated from the Gymnasium there. He then studied mathematics at the University of Breslau and was awarded a doctorate in algebraic geometry in 1905. In 1906 Toeplitz arrived at Göttingen University, which was then the world's leading mathematical center, and he remained there for seven years. The mathematics faculty included David Hilbert, Felix Klein, and Hermann Minkowski. Toeplitz joined a group of young people working with Hilbert: Max Born, Richard Courant and Ernst Hellinger, with whom he collaborated for many years afterward. At that time Toeplitz began to rework the theory of linear functionals and quadratic forms on n-dimensional spaces for infinite dimensional spaces. He wrote five papers directly related to spectral theory of operators which Hilbert was developing. During this period he also published a paper on summation processes and discovered the basic ideas of what are now called the Toeplitz operators. In 1913 Toeplitz became an extraordinary professor at the University of Kiel. He was promoted to a professor in 1920.
In 1911, Toeplitz proposed the inscribed square problem:
Does every Jordan curve contain an inscribed square?
This has been established for convex curves and smooth curves, but the question remains open in general (2007).
Together with Hans Rademacher, he wrote a classic of popular mathematics Von Zahlen und Figuren, which was first published in 1930 and later translated into English as Enjoyment of Mathematics.
Toeplitz was deeply interested in the history of mathematics. In 1929, he cofounded "Quellen und Studien zur Geschichte der Mathematik" with Otto Neugebauer and Julius Stenzel. Beginning in the 1920s, Toepl |
https://en.wikipedia.org/wiki/History%20of%20thermodynamics | The history of thermodynamics is a fundamental strand in, the history of physics, the history of chemistry, and the history of science in general. Owing in the relevance of thermodynamics in much of science and technology, its history is finely woven with the developments of classical mechanics, quantum mechanics, magnetism, and chemical kinetics, to more distant applied fields such as meteorology, information theory, and biology (physiology), and to technological developments such as the steam engine, internal combustion engine, cryogenics and electricity generation. The development of thermodynamics both drove and was driven by atomic theory. It also, albeit in a subtle manner, motivated new directions in probability and statistics; see, for example, the timeline of thermodynamics.
Antiquity
The ancients viewed heat as that related to fire. In 3000 BC, the ancient Egyptians viewed heat as related to origin mythologies. The ancient Indian philosophy including Vedic philosophy believed that five classical elements (or pancha mahā bhūta) are the basis of all cosmic creations. In the Western philosophical tradition, after much debate about the primal element among earlier pre-Socratic philosophers, Empedocles proposed a four-element theory, in which all substances derive from earth, water, air, and fire. The Empedoclean element of fire is perhaps the principal ancestor of later concepts such as phlogiston and caloric. Around 500 BC, the Greek philosopher Heraclitus became famous as the "flux and fire" philosopher for his proverbial utterance: "All things are flowing." Heraclitus argued that the three principal elements in nature were fire, earth, and water.
Vacuum-abhorrence
The 5th century BC Greek philosopher Parmenides, in his only known work, a poem conventionally titled On Nature, uses verbal reasoning to postulate that a void, essentially what is now known as a vacuum, in nature could not occur. This view was supported by the arguments of Aristotle, but was |
https://en.wikipedia.org/wiki/Colony%20picker | A colony picker is an instrument used to automatically identify microbial colonies growing on a solid medium, pick them and duplicate them either onto solid or liquid media. It is used in research laboratories as well as in industrial environments such as food testing and in microbiological cultures.
Uses
In food safety and in clinical diagnosis colony picking is used to isolate individual colonies for identification. Colony pickers automate this procedure, saving costs and personnel and reducing human error. In the drug discovery process they are used for screening purposes by picking thousands of microbial colonies and transferring them for further testing.
Other uses include cloning procedures and DNA sequencing.
as add-on
Colony pickers are sold either as stand-alone instruments or as add-ons to liquid handling robots, using the robot as the actuator and adding a camera and image analysis capabilities. This strategy lowers the price of the system considerably and adds reusability as the robot can still be used for other purposes. |
https://en.wikipedia.org/wiki/Plexus | In neuroanatomy, a plexus (from the Latin term for "braid") is a branching network of vessels or nerves. The vessels may be blood vessels (veins, capillaries) or lymphatic vessels. The nerves are typically axons outside the central nervous system.
The standard plural form in English is plexuses. Alternatively, the Latin plural plexūs may be used.
Types
Nerve plexuses
The four primary nerve plexuses are the cervical plexus, brachial plexus, lumbar plexus, and the sacral plexus.
Cardiac plexus
Celiac plexus
Renal plexus
Venous plexus
Choroid plexus
The choroid plexus is a part of the central nervous system in the brain and consists of capillaries, brain ventricles, and ependymal cells.
Invertebrates
The plexus is the characteristic form of nervous system in the coelenterates and persists with modifications in the flatworms. The nerves of the radially symmetric echinoderms also take this form, where a plexus underlies the ectoderm of these animals and deeper in the body other nerve cells form plexuses of limited extent.
See also
Cranial nerve
Spinal nerve
Nerve plexus
Brachial nerve
List of anatomy mnemonics |
https://en.wikipedia.org/wiki/Gravitational%20lensing%20formalism | In general relativity, a point mass deflects a light ray with impact parameter by an angle approximately equal to
where G is the gravitational constant, M the mass of the deflecting object and c the speed of light. A naive application of Newtonian gravity can yield exactly half this value, where the light ray is assumed as a massed particle and scattered by the gravitational potential well. This approximation is good when is small.
In situations where general relativity can be approximated by linearized gravity, the deflection due to a spatially extended mass can be written simply as a vector sum over point masses. In the continuum limit, this becomes an integral over the density , and if the deflection is small we can approximate the gravitational potential along the deflected trajectory by the potential along the undeflected trajectory, as in the Born approximation in quantum mechanics. The deflection is then
where is the line-of-sight coordinate, and is the vector impact parameter of the actual ray path from the infinitesimal mass located at the coordinates .
Thin lens approximation
In the limit of a "thin lens", where the distances between the source, lens, and observer are much larger than the size of the lens (this is almost always true for astronomical objects), we can define the projected mass density
where is a vector in the plane of the sky. The deflection angle is then
As shown in the diagram on the right, the difference between the unlensed angular position and the observed position is this deflection angle, reduced by a ratio of distances, described as the lens equation
where is the distance from the lens to the source, is the distance from the observer to the source, and is the distance from the observer to the lens. For extragalactic lenses, these must be angular diameter distances.
In strong gravitational lensing, this equation can have multiple solutions, because a single source at can be lensed into multiple images.
Conv |
https://en.wikipedia.org/wiki/Minkowski%20distance | The Minkowski distance or Minkowski metric is a metric in a normed vector space which can be considered as a generalization of both the Euclidean distance and the Manhattan distance. It is named after the German mathematician Hermann Minkowski.
Definition
The Minkowski distance of order (where is an integer) between two points
is defined as:
For the Minkowski distance is a metric as a result of the Minkowski inequality. When the distance between and is but the point is at a distance from both of these points. Since this violates the triangle inequality, for it is not a metric. However, a metric can be obtained for these values by simply removing the exponent of The resulting metric is also an F-norm.
Minkowski distance is typically used with being 1 or 2, which correspond to the Manhattan distance and the Euclidean distance, respectively. In the limiting case of reaching infinity, we obtain the Chebyshev distance:
Similarly, for reaching negative infinity, we have:
The Minkowski distance can also be viewed as a multiple of the power mean of the component-wise differences between and
The following figure shows unit circles (the level set of the distance function where all points are at the unit distance from the center) with various values of :
See also
External links
Unit Balls for Different p-Norms in 2D and 3D at wolfram.com
Unit-Norm Vectors under Different p-Norms at wolfram.com
Simple IEEE 754 implementation in C++
NPM JavaScript Package/Module
Distance
Hermann Minkowski
Metric geometry
Normed spaces |
https://en.wikipedia.org/wiki/Decidual%20cells | Before the fertilized ovum reaches the uterus, the mucous membrane of the body of the uterus undergoes important changes and is then known as the decidua. The thickness and vascularity of the mucous membrane are greatly increased; its glands are elongated and open on its free surface by funnel-shaped orifices, while their deeper portions are tortuous and dilated into irregular spaces. The interglandular tissue is also increased in quantity, and is crowded with large round, oval, or polygonal cells, termed decidual cells. Their enlargement is due to glycogen and lipid accumulation in the cytoplasm allowing these cells to provide a rich source of nutrition for the developing embryo. Decidual cells are also thought to control the invasion of the endometrium by trophoblast cells.
Experimentally, human endometrial stromal cells can be decidualized in culture by using analogs of cAMP and progesterone. The cells will exhibit a decidualized phenotype and display upregulation of common decidualization markers such as prolactin and IGFBP1. |
https://en.wikipedia.org/wiki/David%20Eisenbud | David Eisenbud (born 8 April 1947 in New York City) is an American mathematician. He is a professor of mathematics at the University of California, Berkeley and former director of the then Mathematical Sciences Research Institute (MSRI), now known as Simons Laufer Mathematical Sciences Institute (SLMath). He served as Director of MSRI from 1997 to 2007, and then again from 2013 to 2022.
Biography
Eisenbud is the son of mathematical physicist Leonard Eisenbud, who was a student and collaborator of the renowned physicist Eugene Wigner. Eisenbud received his Ph.D. in 1970 from the University of Chicago, where he was a student of Saunders Mac Lane and, unofficially, James Christopher Robson. He then taught at Brandeis University from 1970 to 1997, during which time he had visiting positions at Harvard University, Institut des Hautes Études Scientifiques (IHÉS), University of Bonn, and Centre national de la recherche scientifique (CNRS). He joined the staff at MSRI in 1997, and took a position at Berkeley at the same time.
From 2003 to 2005 Eisenbud was President of the American Mathematical Society.
Eisenbud's mathematical interests include commutative and non-commutative algebra, algebraic geometry, topology, and computational methods in these fields. He has written over 150 papers and books with over 60 co-authors. Notable contributions include the theory of matrix factorizations for maximal Cohen–Macaulay modules over hypersurface rings, the Eisenbud–Goto conjecture on degrees of generators of syzygy modules, and the Buchsbaum–Eisenbud criterion for exactness of a complex. He also proposed the Eisenbud–Evans conjecture, which was later settled by the Indian mathematician Neithalath Mohan Kumar.
He has had 31 doctoral students, including Craig Huneke, Mircea Mustaţă, Irena Peeva, and Gregory G. Smith (winner of the Aisenstadt Prize in 2007).
Eisenbud's hobbies are juggling (he has written two papers on the mathematics of juggling) and music. He has appeared |
https://en.wikipedia.org/wiki/List%20of%20Indian%20state%20flags | As of 2023, there are no officially recognised flags for the individual states or union territories of India. No legal prohibitions to prevent states adopting distinctive flags exist in either the Emblems and Names (Prevention of Improper Use) Act, 1950, or the Prevention of Insults to National Honour Act, 1971. In a 1994 case before the Supreme Court of India, S. R. Bommai v. Union of India, the Supreme Court declared that there is no prohibition in the Constitution of India for a state to have its own flag. However, a state flag should not dishonour the national flag. The Flag code of India also permits other flags to be flown with the Flag of India, but not on the same flag pole or in a superior position to the national flag.
Former official state flags
The state of Jammu and Kashmir had an officially recognised state flag between 1952 and 2019 under the special status granted to the state by Article 370 of the Constitution of India.
Proposed state flags
The Government of Tamil Nadu proposed a design for the Flag of Tamil Nadu in 1970.
The Government of Karnataka proposed a design for the Flag of Karnataka in 2018 based on the traditional yellow-red Kannada bicolour. The new tricolour flag with the central white band and emblem, was designed to distance itself from regional political parties and emulate the structure of the Indian Tricolour. In August 2019, the Government of Karnataka announced it was no longer officially pursuing the proposal for an official state flag.
Banners of the states and union territories
When a distinctive banner is required to represent a state or union territory, the emblem of the state or union territory is usually displayed on a white field.
States
Union territories
See also
National Flag of India
Flag code of India
List of Indian flags
List of Indian state symbols
List of Indian state emblems
List of Indian state mottos
List of Indian state songs
List of Indian state foundation days
List of Indian state animals
List of Ind |
https://en.wikipedia.org/wiki/Oja%27s%20rule | Oja's learning rule, or simply Oja's rule, named after Finnish computer scientist Erkki Oja, is a model of how neurons in the brain or in artificial neural networks change connection strength, or learn, over time. It is a modification of the standard Hebb's Rule (see Hebbian learning) that, through multiplicative normalization, solves all stability problems and generates an algorithm for principal components analysis. This is a computational form of an effect which is believed to happen in biological neurons.
Theory
Oja's rule requires a number of simplifications to derive, but in its final form it is demonstrably stable, unlike Hebb's rule. It is a single-neuron special case of the Generalized Hebbian Algorithm. However, Oja's rule can also be generalized in other ways to varying degrees of stability and success.
Formula
Consider a simplified model of a neuron that returns a linear combination of its inputs using presynaptic weights :
Oja's rule defines the change in presynaptic weights given the output response of a neuron to its inputs to be
where is the learning rate which can also change with time. Note that the bold symbols are vectors and defines a discrete time iteration. The rule can also be made for continuous iterations as
Derivation
The simplest learning rule known is Hebb's rule, which states in conceptual terms that neurons that fire together, wire together. In component form as a difference equation, it is written
,
or in scalar form with implicit -dependence,
,
where is again the output, this time explicitly dependent on its input vector .
Hebb's rule has synaptic weights approaching infinity with a positive learning rate. We can stop this by normalizing the weights so that each weight's magnitude is restricted between 0, corresponding to no weight, and 1, corresponding to being the only input neuron with any weight. We do this by normalizing the weight vector to be of length one:
.
Note that in Oja's original paper, , correspondi |
https://en.wikipedia.org/wiki/Aspect%20ratio%20%28aeronautics%29 | In aeronautics, the aspect ratio of a wing is the ratio of its span to its mean chord. It is equal to the square of the wingspan divided by the wing area. Thus, a long, narrow wing has a high aspect ratio, whereas a short, wide wing has a low aspect ratio.
Aspect ratio and other features of the planform are often used to predict the aerodynamic efficiency of a wing because the lift-to-drag ratio increases with aspect ratio, improving the fuel economy in powered airplanes and the gliding angle of sailplanes.
Definition
The aspect ratio is the ratio of the square of the wingspan to the projected wing area , which is equal to the ratio of the wingspan to the standard mean chord :
Mechanism
As a useful simplification, an airplane in flight can be imagined to affect a circular cylinder of air with a diameter equal to the wingspan. A large wingspan affects a large cylinder of air, and a small wingspan affects a small cylinder of air. A small air cylinder must be pushed down with a greater power (energy change per unit time) than a large cylinder in order to produce an equal upward force (momentum change per unit time). This is because giving the same momentum change to a smaller mass of air requires giving it a greater velocity change, and a much greater energy change because energy is proportional to the square of the velocity while momentum is only linearly proportional to the velocity. The aft-leaning component of this change in velocity is proportional to the induced drag, which is the force needed to take up that power at that airspeed.
The interaction between undisturbed air outside the cylinder of air, and the downward-moving cylinder of air occurs at the wingtips and can be seen as wingtip vortices.
It is important to keep in mind that this is a drastic oversimplification, and an airplane wing affects a very large area around itself.
In aircraft
Although a long, narrow wing with a high aspect ratio has aerodynamic advantages like better lift-to-drag-ra |
https://en.wikipedia.org/wiki/Merge-insertion%20sort | In computer science, merge-insertion sort or the Ford–Johnson algorithm is a comparison sorting algorithm published in 1959 by L. R. Ford Jr. and Selmer M. Johnson. It uses fewer comparisons in the worst case than the best previously known algorithms, binary insertion sort and merge sort, and for 20 years it was the sorting algorithm with the fewest known comparisons. Although not of practical significance, it remains of theoretical interest in connection with the problem of sorting with a minimum number of comparisons. The same algorithm may have also been independently discovered by Stanisław Trybuła and Czen Ping.
Algorithm
Merge-insertion sort performs the following steps, on an input of elements:
Group the elements of into pairs of elements, arbitrarily, leaving one element unpaired if there is an odd number of elements.
Perform comparisons, one per pair, to determine the larger of the two elements in each pair.
Recursively sort the larger elements from each pair, creating a sorted sequence of of the input elements, in ascending order.
Insert at the start of the element that was paired with the first and smallest element of .
Insert the remaining elements of into , one at a time, with a specially chosen insertion ordering described below. Use binary search in subsequences of (as described below) to determine the position at which each element should be inserted.
The algorithm is designed to take advantage of the fact that the binary searches used to insert elements into are most efficient (from the point of view of worst case analysis) when the length of the subsequence that is searched is one less than a power of two. This is because, for those lengths, all outcomes of the search use the same number of comparisons as each other. To choose an insertion ordering that produces these lengths, consider the sorted sequence after step 4 of the outline above (before inserting the remaining elements),
and let denote the th element of this sorted seque |
https://en.wikipedia.org/wiki/Toltrazuril | Toltrazuril is a coccidiostat. |
https://en.wikipedia.org/wiki/Silane | Silane (Silicane) is an inorganic compound with chemical formula . It is a colourless, pyrophoric, toxic gas with a sharp, repulsive, pungent smell, somewhat similar to that of acetic acid. Silane is of practical interest as a precursor to elemental silicon. Silane with alkyl groups are effective water repellents for mineral surfaces such as concrete and masonry. Silanes with both organic and inorganic attachments are used as coupling agents. Silanes are commonly used to apply coatings to surfaces or as an adhesion promoter.
Production
Commercial-scale routes
Silane can be produced by several routes. Typically, it arises from the reaction of hydrogen chloride with magnesium silicide:
Mg2Si + 4 HCl -> 2 MgCl2 + SiH4
It is also prepared from metallurgical-grade silicon in a two-step process. First, silicon is treated with hydrogen chloride at about 300 °C to produce trichlorosilane, HSiCl3, along with hydrogen gas, according to the chemical equation
Si + 3 HCl -> HSiCl3 + H2
The trichlorosilane is then converted to a mixture of silane and silicon tetrachloride:
4 HSiCl3 -> SiH4 + 3 SiCl4
This redistribution reaction requires a catalyst.
The most commonly used catalysts for this process are metal halides, particularly aluminium chloride. This is referred to as a redistribution reaction, which is a double displacement involving the same central element. It may also be thought of as a disproportionation reaction, even though there is no change in the oxidation number for silicon (Si has a nominal oxidation number IV in all three species). However, the utility of the oxidation number concept for a covalent molecule, even a polar covalent molecule, is ambiguous. The silicon atom could be rationalized as having the highest formal oxidation state and partial positive charge in and the lowest formal oxidation state in , since Cl is far more electronegative than is H.
An alternative industrial process for the preparation of very high-purity silane, suitable for use |
https://en.wikipedia.org/wiki/Physicalization | Physicalization of computer hardware (the opposite of virtualization), is a way to place multiple physical machines in a rack unit.
It can be a way to reduce hardware costs, since in some cases, server processors cost more per core than energy efficient laptop processors, which may make up for added cost of board level integration. While Moore's law makes increasing integration less expensive, some jobs require much I/O bandwidth, which may be less expensive to provide using many less-integrated processors.
Applications and services that are I/O bound are likely to benefit from such physicalized environments. This ensures that each operating system instance is running on a processor that has its own network interface card, host bus and I/O sub-system unlike in the case of a multi-core servers where a single I/O sub-system is shared between all the cores / VMs.
A data physicalization (or simply physicalization) is a physical artefact whose geometry or material properties encode data. It has the main goals to engage people and to communicate data using computer-supported physical data representations.
History
Before the invention of computers and digital devices, the application of data physicalization already existed in ancient artifacts as a medium to represent abstract information. One example is Blombo ocher plaque which is estimated to be 70000 – 80000 years old. The geometric and iconographic shapes engraved at the surface of the artifact demonstrated the cognitive complexity of ancient humans. Moreover, since such representations were deliberately made and crafted, the evidences suggest that the geometric presentation of information is a popular methodology in the context of society. Although researchers still cannot decipher the specific type of information encoded in the artifact, there are several proposed interpretations. For example, the potential functions of the artifact are divided into four categories, categorized as "numerical", "functional", " |
https://en.wikipedia.org/wiki/Deinococcus%20apachensis | Deinococcus apachensis is a species of bacteria in the phylum Deinococcota. Strains of this species were isolated from soil samples from Arizona after exposure to more than 15 kGy of radiation. |
https://en.wikipedia.org/wiki/Alice%20Burks | Alice Burks (née Rowe, August 20, 1920 – November 21, 2017) was an American author of children's books and books about the history of electronic computers.
Early life and education
Burks was born Alice Rowe in East Cleveland, Ohio, in 1920. She began her undergraduate degree at Oberlin College on a competitive mathematics scholarship and transferred to the University of Pennsylvania in Philadelphia where she completed her B.A. in mathematics in 1944. During this period, she was employed as a human computer at the University of Pennsylvania's Moore School of Electrical Engineering.
Career
Burks retired from full-time employment after marrying Moore School lecturer Dr. Arthur Burks, a mathematician who served as one of the principal engineers in the construction of the ENIAC, the world's first general-purpose electronic digital computer, built at the Moore School between 1943 and 1946. Unlike some of the Moore School women computers, she never worked directly with the ENIAC.
At the conclusion of Arthur's work with the Moore School and at the Institute for Advanced Study in 1946, Burks moved with her husband to Ann Arbor, Michigan, where he joined the faculty of the University of Michigan and helped to found the computer science department. She returned to school, earning an M.S. in educational psychology in 1957 from Michigan.
Starting in the 1970s following the decision of Honeywell v. Sperry Rand, the federal court case that invalidated the ENIAC patent, she and husband Arthur championed the work of John Vincent Atanasoff, the Iowa State College physics professor whom the court had ruled invented the first electronic digital computer (a machine that came to be called the Atanasoff–Berry Computer) and from whom the subject matter of the ENIAC was ruled to be derived. In articles and two books, the first co-authored with Arthur, Mrs. Burks sought to bolster the judge's decision and highlight testimony and evidence from the case. This pitted the Burkses in a d |
https://en.wikipedia.org/wiki/Crepant%20resolution | In algebraic geometry, a crepant resolution of a singularity is a resolution that does not affect the canonical class of the manifold. The term "crepant" was coined by by removing the prefix "dis" from the word "discrepant", to indicate that the resolutions have no discrepancy in the canonical class.
The crepant resolution conjecture of states that the orbifold cohomology of a Gorenstein orbifold is isomorphic to a semiclassical limit of the quantum cohomology of a crepant resolution.
In 2 dimensions, crepant resolutions of complex Gorenstein quotient singularities (du Val singularities) always exist and are unique, in 3 dimensions they exist but need not be unique as they can be related by flops, and in dimensions greater than 3 they need not exist.
A substitute for crepant resolutions which always exists is a terminal model. Namely, for every variety X over a field of characteristic zero such that X has canonical singularities (for example, rational Gorenstein singularities), there is a variety Y with Q-factorial terminal singularities and a birational projective morphism f: Y → X which is crepant in the sense that KY = f*KX.
Notes |
https://en.wikipedia.org/wiki/Nitrogenase | Nitrogenases are enzymes () that are produced by certain bacteria, such as cyanobacteria (blue-green bacteria) and rhizobacteria. These enzymes are responsible for the reduction of nitrogen (N2) to ammonia (NH3). Nitrogenases are the only family of enzymes known to catalyze this reaction, which is a step in the process of nitrogen fixation. Nitrogen fixation is required for all forms of life, with nitrogen being essential for the biosynthesis of molecules (nucleotides, amino acids) that create plants, animals and other organisms. They are encoded by the Nif genes or homologs. They are related to protochlorophyllide reductase.
Classification and structure
Although the equilibrium formation of ammonia from molecular hydrogen and nitrogen has an overall negative enthalpy of reaction (), the activation energy is very high (). Nitrogenase acts as a catalyst, reducing this energy barrier such that the reaction can take place at ambient temperatures.
A usual assembly consists of two components:
The heterotetrameric MoFe protein, a nitrogenase which uses the electrons provided to reduce N2 to NH3. In some assemblies it is replaced by a homologous alternative.
The homodimeric Fe-only protein, the reductase which has a high reducing power and is responsible for the supply of electrons.
Reductase
The Fe protein, the dinitrogenase reductase or NifH, is a dimer of identical subunits which contains one [Fe4S4] cluster and has a mass of approximately 60-64kDa. The function of the Fe protein is to transfer electrons from a reducing agent, such as ferredoxin or flavodoxin to the nitrogenase protein. The transfer of electrons requires an input of chemical energy which comes from the binding and hydrolysis of ATP. The hydrolysis of ATP also causes a conformational change within the nitrogenase complex, bringing the Fe protein and MoFe protein closer together for easier electron transfer.
Nitrogenase
The MoFe protein is a heterotetramer consisting of two α subunits and two β su |
https://en.wikipedia.org/wiki/Recovery%20effect | The recovery effect is a phenomenon observed in battery usage where the available energy is less than the difference between energy charged and energy consumed. Intuitively, this is because the energy has been consumed from the edge of the battery and the charge has not yet diffused evenly around the battery.
When power is extracted continuously voltage decreases in a smooth curve, but the recovery effect can result in the voltage partially increasing if the current is interrupted.
The KiBaM battery model describes the recovery effect for lead-acid batteries and is also a good approximation to the observed effects in Li-ion batteries. In some batteries the gains from the recovery life can extend battery life by up to 45% by alternating discharging and inactive periods rather than constantly discharging. The size of the recovery effect depends on the battery load, recovery time and depth of discharge.
Even though the recovery effect phenomenon is more prominent in the lead acid battery chemistry, its existence in alkaline, Ni-MH and Li-Ion batteries is still questionable. For instance, a systematic experimental case study shows that an intermittent discharge current in case of alkaline, Ni-MH and Li-Ion batteries results in a decreased usable energy output compared to a continuous discharge current of the same average value. This is primarily due to the increased overpotential experienced due to the high peak currents of the intermittent discharge over the continuous discharge current of same average value.
See also
Capacity fading |
https://en.wikipedia.org/wiki/Vaginal%20branches%20of%20uterine%20artery | The uterine artery supplies branches to the cervix uteri and others which descend on the vagina; the latter anastomose with branches of the vaginal arteries and form with them two median longitudinal vessels—the vaginal branches of uterine artery (or azygos arteries of the vagina)—one of which runs down in front of and the other behind the vagina. |
https://en.wikipedia.org/wiki/Hairless | Hairless, also known as H, is a well-characterized Drosophila gene. Since Hairless is a dominant loss of function mutation, many mutations to Hairless are embryonic lethal, but there are several viable hairless mutants. This specific Drosophila gene is involved in the Notch signaling pathway (NSP) by acting as a suppressor of the organism's Notch signaling. This interaction of the NSP can be seen in Figure 1.
Hairless (H) encodes a hydrophilic protein that is composed of 1076 amino acids, and has a molecular weight of 110 kDa. The H protein is reported in insects, and is found in nearly all of the family Drosophilidae. It is also found in the family Culicidae, as well as the orders Lepidoptera, Hymenoptera, and Coleoptera.
In Drosophila, NSPs allow for communication intercellularly during embryo development, and the physiological activities of adult organisms. Furthermore, these pathways turn on periodically during devolvement to help determine cell fate functions of the cell. Since Hairless (H) is an antagonist of the NSP, as well as a “key member of the Su[H] repressor complex", it plays a significant part of embryo development in insects because it helps influence cell fate decisions during this time.
It is crucial that proteins HP120 and HP150 are present in the organism as they control the normal activity of H. This activity regulation also controls NSP, which allows the fly embryos to develop correctly, and if there is any fluctuation in the pathway or the proteins, it can change the outcome of the offspring. The two most common changes in the Drosophila phenotype as a result of Hairless mutation are bristle loss and vein gaps. These changes can be seen in Figure 2. |
https://en.wikipedia.org/wiki/Bonner%20sphere | A Bonner sphere is a device used to determine the energy spectrum of a neutron beam. The method was first described in 1960 by Rice University's Bramblett, Ewing and Tom W. Bonner and employs thermal neutron detectors embedded in moderating spheres of different sizes. Comparison of the neutrons detected by each sphere allows accurate determination of the neutron energy. This detector system utilizes a few channel unfolding techniques to determine the coarse, few group neutron spectrum. The original detector system was capable of measuring neutrons between thermal energies up to ~20 MeV. These detectors have been modified to provide additional resolution above 20 MeV to energies up to 1 GeV.
Bonner sphere spectroscopy
Because of the complexity with which neutrons interact with the environment, precise determination of the neutron energy is quite difficult. Bonner sphere spectroscopy (BSS) is one of the few methods that provide an accurate measure of the neutron spectrum.
Remball
A single Bonner sphere of an appropriate size can be used for dosimetry, as the sensitivity of the detector will approximate the radiation weighting factor across a range of neutron energies. Such Bonner spheres are sometimes known as a remball.
See also
Neutron detection |
https://en.wikipedia.org/wiki/Looking%20Glass%20server | Looking Glass servers (LG servers) are servers on the Internet running one of a variety of publicly available Looking Glass software implementations. They are commonly deployed by autonomous systems (AS) to offer access to their routing infrastructure in order to facilitate debugging network issues. A Looking Glass server is accessed remotely for the purpose of viewing routing information. Essentially, the server acts as a limited, read-only portal to routers of whatever organization is running the LG server.
Typically, Looking Glass servers are run by autonomous systems like Internet service providers (ISPs), Network Service Providers (NSPs), and Internet exchange points (IXPs).
Implementation
Looking glasses are web scripts directly connected to routers' admin interfaces such as telnet and SSH. These scripts are designed to relay textual commands from the web to the router and print back the response. They are often implemented in Perl PHP, and Python, and are publicly available on GitHub.
Security concerns
A 2014 paper demonstrated the potential security concerns of Looking Glass servers, noting that even an "attacker with very limited resources can exploit such flaws in operators' networks and gain access to core Internet infrastructure", resulting in anything from traffic disruption to global Border Gateway Protocol (BGP) route injection. This is due in part because looking glass servers are "an often overlooked critical part of an operator infrastructure" because it sits at the intersection of the public internet and "restricted admin consoles". As of 2014, most Looking Glass software was small and old, having last been updated in the early 2000's.
See also
Autonomous system (Internet)
Internet backbone |
https://en.wikipedia.org/wiki/Kendall%20tau%20distance | The Kendall tau rank distance is a metric (distance function) that counts the number of pairwise disagreements between two ranking lists. The larger the distance, the more dissimilar the two lists are. Kendall tau distance is also called bubble-sort distance since it is equivalent to the number of swaps that the bubble sort algorithm would take to place one list in the same order as the other list. The Kendall tau distance was created by Maurice Kendall.
Definition
The Kendall tau ranking distance between two lists and is
where and are the rankings of the element in and respectively.
will be equal to 0 if the two lists are identical and (where is the list size) if one list is the reverse of the other.
Kendall tau distance may also be defined as
where
P is the set of unordered pairs of distinct elements in and
= 0 if i and j are in the same order in and
= 1 if i and j are in the opposite order in and
Kendall tau distance can also be defined as the total number of discordant pairs.
Kendall tau distance in Rankings: A permutation (or ranking) is an array of N integers where each of the integers between 0 and N-1 appears exactly once.
The Kendall tau distance between two rankings is the number of pairs that are in different order in the two rankings. For example, the Kendall tau distance between 0 3 1 6 2 5 4 and 1 0 3 6 4 2 5 is four because the pairs 0-1, 3-1, 2-4, 5-4 are in different order in the two rankings, but all other pairs are in the same order.
The normalized Kendall tau distance is and therefore lies in the interval [0,1].
If Kendall tau distance function is performed as instead of (where and are the rankings of and elements respectively), then triangular inequality is not guaranteed. The triangular inequality fails sometimes also in cases where there are repetitions in the lists. So then we are not dealing with a metric anymore.
Generalised versions of Kendall tau distance have been proposed to give weights to |
https://en.wikipedia.org/wiki/Allium%20scorodoprasum | The sand leek (Allium scorodoprasum), also known as rocambole and Korean pickled-peel garlic, is a Eurasian species of wild onion with a native range extending across much of Europe, Middle East, and Korea. The species should not be confused with rocambole garlic, which is A. sativum var. ophioscorodon.
Description
The sand leek is a perennial plant with an egg-shaped bulb. The plant produces two to five unstalked leaves, the bases of which are sheath-like. Each leaf blade is linear, 7–20 mm wide, flat with a slight keel, an entire margin and parallel veins. The edges of the leaf and the central vein are rough to the touch. The flowering stem is cylindrical, growing to a height of and the upper half is leafless. The whole plant has an onion-like aroma. The inflorescence is a globular cluster surrounded by membranous bracts in bud which wither when the flowers open. Each individual flower is stalked and has a purple perianth long. There are six tepals, six stamens and a pistil formed from three fused carpels. Mixed with the flowers are a number of purple bulbils. The fruit is a capsule, but the seeds seldom set, and propagation usually takes place when the bulbils are knocked off and grow into new plants.
Distribution and habitat
The natural habitat of A. scorodoprasum is damp broad-leaved woodland, forest margins, shores, hillside meadows and hedgerows. It was at one time used as a kitchen herb and can sometimes be found near old habitations.
Cultivation
A. scorodoprasum is edible but seldom cultivated, and has a shorter flower stalk and fewer and more inconsistently shaped cloves than Rocambole garlic. Sand leek also has a dark violet bulb wrapper.
Elephant garlic (properly A. ampeloprasum var. ampeloprasum) is also sometimes incorrectly sold as A. scorodoprasum. |
https://en.wikipedia.org/wiki/Aleksandar%20Totic | Aleksandar Totic is one of the original developers of the Mosaic browser. He cofounded and was a partner at Netscape Communications Corporation.
He was born in Belgrade, Serbia, on 23 September 1966. He moved to America after his degree from Kuwait was not recognized by Yugoslav government, and currently lives in Palo Alto, CA San Francisco, CA.
External links
Mosaic - The First Global Web Browser
Software engineers
Serbian computer scientists
Computer programmers
Living people
Year of birth missing (living people)
Place of birth missing (living people) |
https://en.wikipedia.org/wiki/CE-ATA | Consumer Electronics ATA (CE-ATA) is an interface standard for the connection of storage devices and hosts in consumer electronic device such as mobile and handheld devices. One of the primary goals is to standardize connections for small form factor hard disk drives such as 1-inch Microdrives.
The standard is maintained by CE-ATA Workgroup.
History
The CE-ATA Specification was developed in 2005.
Interface
MMC
CE-ATA is electrically and physically compatible with MMC specification. CE-ATA uses MMC connector on host devices and matching flex cable or circuit connection on CE-ATA hard disk drives.
Pin Assignment
See also
AT Attachment (ATA)
MMC
External links
CE-ATA Workgroup
CE-ATA Digital Protocol Specification Revision 1.1
MultiMediaCard Association
AT Attachment |
https://en.wikipedia.org/wiki/Kinin%E2%80%93kallikrein%20system | The kinin–kallikrein system or simply kinin system is a poorly understood hormonal system with limited available research. It consists of blood proteins that play a role in inflammation, blood pressure control, coagulation and pain. Its important mediators bradykinin and kallidin are vasodilators and act on many cell types. Clinical symptoms include marked weakness, tachycardia, fever, leukocytosis and acceleration of ESR.
History
The system was discovered in 1909 when researchers discovered that injection with urine (high in kinins) led to hypotension (low blood pressure). The researchers Emil Karl Frey, Heinrich Kraut and Eugen Werle discovered high-molecular weight kininogen in urine around 1930.
Etymology
kinin [Gk] kīn(eîn) to move, set in motion. kallikrein [Gk ] kalli~ sweet and krein = kreos, flesh, named for the pancreatic extracts where it was first discovered
Members
The system consists of a number of large proteins, some small polypeptides and a group of enzymes that activate and deactivate the compounds.
Proteins
High-molecular weight kininogen (HMWK) and low-molecular weight kininogen (LMWK) are precursors of the polypeptides. They have no activity of themselves.
HMWK is produced by the liver together with prekallikrein (see below). It acts mainly as a cofactor on coagulation and inflammation, and has no intrinsic catalytic activity.
LMWK is produced locally by numerous tissues, and secreted together with tissue kallikrein.
Polypeptides
Bradykinin (BK), which acts on the B2 receptor and slightly on B1, is produced when kallikrein releases it from HMWK. It is a nonapeptide (9 amino acids) with the amino acid sequence Arg–Pro–Pro–Gly–Phe–Ser–Pro–Phe–Arg.
Kallidin (KD) is released from LMWK by tissue kallikrein. It is a decapeptide. KD has the same amino acid sequence as Bradykinin with the addition of a Lysine at the N-terminus, thus is sometimes referred to as Lys-Bradykinin.
HMWK and LMWK are formed by alternative splicing of the same gene.
|
https://en.wikipedia.org/wiki/The%20Elephant%20in%20the%20Brain | The Elephant in the Brain: Hidden Motives in Everyday Life is a 2018 nonfiction book by Kevin Simler and Robin Hanson. Simler is a writer and software engineer, while Hanson is an associate professor of economics at George Mason University. The book explores self-deception and hidden motives in human behaviour. The publisher's website describes the aim of the book as 'to track down the darker, unexamined corners of our psyches and blast them with floodlights'.
Summary
The main thesis of the book is that we are very often not aware of our real reasons for most of our behaviors. Our behaviors are optimised for living in a social group and very often, from the point of view of natural selection, it is useful if we are not consciously aware of our real motivations.
The book is split into two sections. The first, entitled 'Why We Hide Our Motives' includes an introduction to the subjects of animal behaviour, signalling, social norms and self-deception. In the second section, title 'Hidden Motives in Everyday Life' each chapter covers an aspect of human behaviour and describes how it can be explained through the framework of signalling and self-deception outlined in the first section. The chapters in this section cover body language, laughter, conversation, consumption, art, charity, education, medicine, religion and politics.
Body language
In this chapter, the authors claim that body language is largely a way of communicating status. Most of us will not consciously describe our body language as doing this, nonetheless, various status-related body positions can be observed in humans, similar to other primates.
Laughter
The authors argue that laughter is not just a response to humour, but is better thought of as a 'play signal' - an indication that we do not have feelings of hostility towards the other party. A wide range of data is used to support this hypothesis including the fact that people laugh more in groups than on their own, the fact that babies laugh more wi |
https://en.wikipedia.org/wiki/Lagrangian%20Grassmannian | In mathematics, the Lagrangian Grassmannian is the smooth manifold of Lagrangian subspaces of a real symplectic vector space V. Its dimension is n(n + 1) (where the dimension of V is 2n). It may be identified with the homogeneous space
,
where is the unitary group and the orthogonal group. Following Vladimir Arnold it is denoted by Λ(n). The Lagrangian Grassmannian is a submanifold of the ordinary Grassmannian of V.
A complex Lagrangian Grassmannian is the complex homogeneous manifold of Lagrangian subspaces of a complex symplectic vector space V of dimension 2n. It may be identified with the homogeneous space of complex dimension n(n + 1)
,
where is the compact symplectic group.
As a homogeneous space
To see that the Lagrangian Grassmannian Λ(n) can be identified with , note that is a 2n-dimensional real vector space, with the imaginary part of its usual inner product making it into a symplectic vector space. The Lagrangian subspaces of are then the real subspaces of real dimension n on which the imaginary part of the inner product vanishes. An example is . The unitary group acts transitively on the set of these subspaces, and the stabilizer of is the orthogonal group . It follows from the theory of homogeneous spaces that Λ(n) is isomorphic to as a homogeneous space of .
Topology
The stable topology of the Lagrangian Grassmannian and complex Lagrangian Grassmannian is completely understood, as these spaces appear in the Bott periodicity theorem: , and – they are thus exactly the homotopy groups of the stable orthogonal group, up to a shift in indexing (dimension).
In particular, the fundamental group of is infinite cyclic. Its first homology group is therefore also infinite cyclic, as is its first cohomology group, with a distinguished generator given by the square of the determinant of a unitary matrix, as a mapping to the unit circle. Arnold showed that this leads to a description of the Maslov index, introduced by V. P. Maslov.
For a |
https://en.wikipedia.org/wiki/Gamma-Re%20Transition%20Model | Gamma-Re (γ-Re) transition model is a two equation model used in Computational Fluid Dynamics (CFD) to modify turbulent transport equations to simulate laminar, laminar-to-turbulent and turbulence states in a fluid flow. The Gamma-Re model does not intend to model the physics of the problem but attempts to fit a wide range of experiments and transition methods into its formulation. The transition model calculated an intermittency factor that creates (or extinguishes) turbulence by slowly introducing turbulent production at the laminar-to-turbulent transition location.
Principle
The goal of developing the gamma-Re () transition model was to develop a transition model based on local variables which could be easily implemented into modern CFD code with unstructured grids and massive parallel execution. The majority of earlier transition models such as the model needs to know the structure of the boundary layer and the integration along it; both concepts are hard to implement in three dimensions along many subdivisions of a grid. Another key insight to the formulation of this model is that the Reynolds vorticity number can be related to the Reynolds transition onset number so there is a local way to determine the transition location. The gamma-Re transition model has two equations and is based on the two-equation turbulence models in the context of turbulence modeling. This way both local and global trends can be modelled. The intermittency or gamma determines the percentage of time the flow is turbulent (0 = fully laminar, 1 = fully turbulent). The intermittency acts on the production term of the turbulent kinetic energy transport equation in the SST model to simulate laminar/turbulence flows.
Standard Gamma-Theta model
For intermittency
For Transition Momentum Thickness Reynolds Number
Modification to SST Turbulence Model
Applications
The present model was appropriate for the prediction of an expansion swirl flow.
Other models
Following are some more mod |
https://en.wikipedia.org/wiki/Alex%20Verrijn%20Stuart | Adolf Alexander (Alex or Xander) Verrijn Stuart (Rotterdam, 22 October 1923 – Haarlem, 29 October 2004) was a Dutch computer scientist, and the first Professor in computer science at the Leiden University from 1969 tot 1991.
Biography
Alex Verrijn Stuart was born in Rotterdam, where his father was professor in economics at the Erasmus University Rotterdam. His grandfather was the economist Coenraad Alexander Verrijn Stuart, who in 1899 was the first president of the Statistics Netherlands.
Verrijn Stuart received an MA in physics at the Delft University of Technology, and he received a Ph.D. from the University of Michigan. Afterwards he worked at the operations research department of Royal Dutch Shell for 18 years. In 1969 he was appointed the first Professor in computer science in the Netherlands, at the Leiden University. From 1973 to 1974 he had been a fellow at the Netherlands Institute for Advanced Study. In 1976 he was one of the founding members of the IFIP TC8 Technical Committee of Information Systems. Among his graduate student were Dewald Roode and Sjaak Brinkkemper.
Alex Verrijn Stuart was also an ice skater and mountaineer. In 1940 he skated his first Elfstedentocht of the five to come. In 1977 he was the leader of an expedition to climb the Annapurna I. In 1982 he was part of an expedition to the Mount Everest that failed.
Publications
Books, a selection:
1982. Information Systems Design Methodologies: A Comparative Review Proceedings of the IFIP WG 8.1 Working Conference on Cooperative Review of Information Systems Design Methodologies, Noordwijkerhout, The Netherlands, 10–14 May 1982. Ed. with T. William Olle and Henk G. Sol. North-Holland.
1986. Trends in Information Systems: An Anthology of Papers from Conferences of the IFIP Technical Committee 8 "Information Systems" to Commemorate their Tenth Anniversary. With Börje Langefors and Giampio Bracchi. International Federation for Information Processing. Technical Committee for Information |
https://en.wikipedia.org/wiki/Gail%20F.%20Burrill | Gail F. Burrill is a mathematics educator who was president of the National Council of Teachers of Mathematics (NCTM) from 1996 to 1998. She works as a specialist in the Program in Mathematics Education at Michigan State University.
Burrill worked for nearly 30 years as a high school mathematics teacher; she is also the author of multiple textbooks, and has pushed to include statistics in the curriculum.
She won the Presidential Award for Excellence in Teaching Mathematics in 1985, and the lifetime achievement award of the NCTM in 2012. She was also given an honorary doctorate by the Rose-Hulman Institute of Technology in 2000. In 1994, she was elected as a fellow of the American Statistical Association. |
https://en.wikipedia.org/wiki/TS/A | TS/A is a mouse mammary adenocarcinoma cell line established in the early 1980s from a spontaneous tumor of a retired breeder BALB/c female mouse.
In the 1990s it was widely used as a recipient cell line for the transduction of cytokine genes and other genes of immunological interest, to produce recombinant experimental vaccines that induced anti-tumor immunological responses. |
https://en.wikipedia.org/wiki/Chlorpropham | Chlorpropham or CIPC is a plant growth regulator and herbicide used as a sprout suppressant for grass weeds, alfalfa, lima and snap beans, blueberries, cane fruit, carrots, cranberries, ladino clover, garlic, seed grass, onions, spinach, sugar beets, tomatoes, safflower, soybeans, gladioli and woody nursery stock. It is also used to inhibit potato sprouting and for sucker control in tobacco. Chlorpropham is available in emulsifiable concentrate and liquid formulations.
Chlorpropham is approved for use as a plant regulator and herbicide only on potatoes in the United States. The use of CIPC was banned in the EU and UK in 2019 after it was not reauthorised for use due to toxicity concerns, with sales prohibited from January 2020.
Uses
When it is used as an anti-sprouting agent for potatoes, the formulation is based on HN formulation, Hot Fogging.
Commercial names include Bud Nip, Taterpex, Preventol, Elbanil, Metoxon, Nexoval, Stickman Pistols, Preweed, Furloe, Stopgerme-S, Sprout Nip, Mirvale, Bygran, ChlorIPC, Spud-Nic, Spud-Nie, Chloro-IFK, Chloro-IPC, Keim-stop, Triherbicide CIPC, OORJA.
For herbicide, an EC formulation is used so that it is dissolvable in water for spray in the field.
Toxicity
Chlorpropham displays a low level toxicity profile, with no signs of acute toxicity after exposure of less than 1000 mg/kg/day. Long term exposure at high doses (≥ 1000 mg/kg/day) could cause reduction of body weight gain, decrease in hematocrit and hemoglobin, and increase in blood reticulocytes.
Regarding the carcinogenic risk, chlorpropham is classified by the EPA as group E (non-carcinogenic). One of its metabolites is 3-chloroaniline.
The acceptable daily intake ranges from 0.03 mg/kg (FAO 2001) to 0.05 mg/Kg (EPA 1996 and EC 2003).
Stability
Chlorpropham is partially degraded in the environment under aerobic conditions (15% to 30% after 100 days) and partially hydrolysed in water solution (90% after 59 to 130 days).
A study of the stability of chlorpropham in |
https://en.wikipedia.org/wiki/Schwarzian%20derivative | In mathematics, the Schwarzian derivative is an operator similar to the derivative which is invariant under Möbius transformations. Thus, it occurs in the theory of the complex projective line, and in particular, in the theory of modular forms and hypergeometric functions. It plays an important role in the theory of univalent functions, conformal mapping and Teichmüller spaces. It is named after the German mathematician Hermann Schwarz.
Definition
The Schwarzian derivative of a holomorphic function of one complex variable is defined by
The same formula also defines the Schwarzian derivative of a function of one real variable.
The alternative notation
is frequently used.
Properties
The Schwarzian derivative of any Möbius transformation
is zero. Conversely, the Möbius transformations are the only functions with this property. Thus, the Schwarzian derivative precisely measures the degree to which a function fails to be a Möbius transformation.
If is a Möbius transformation, then the composition has the same Schwarzian derivative as ; and on the other hand, the Schwarzian derivative of is given by the chain rule
More generally, for any sufficiently differentiable functions and
When and are smooth real-valued functions, this implies that all iterations of a function with negative (or positive) Schwarzian will remain negative (resp. positive), a fact of use in the study of one-dimensional dynamics.
Introducing the function of two complex variables
its second mixed partial derivative is given by
and the Schwarzian derivative is given by the formula:
The Schwarzian derivative has a simple inversion formula, exchanging the dependent and the independent variables. One has
or more explicitly, . This follows from the chain rule above.
Geometric interpretation
William Thurston interprets the Schwarzian derivative as a measure of how much a conformal map deviates from a Möbius transformation. Let be a conformal mapping in a neighborhood of Then |
https://en.wikipedia.org/wiki/Meredith%20L.%20Patterson | Meredith L. Patterson (born April 30, 1977) is an American technologist, science fiction writer, and journalist. She has spoken at numerous industry conferences on a wide range of topics. She is also a blogger and software developer, and a leading figure in the biopunk movement.
Raised in Houston, she pursued an education in linguistics and computer science, earning a Master's degree in linguistics and a PhD in computer science from the University of Iowa. Patterson's early career involved diverse roles like website design, technical writing, teaching math, and restaurant criticism for the Houston Press.
Her contributions in computer science include breakthroughs in computational linguistics applied to computer security. She introduced innovative techniques to counter SQL injection attacks and integrated data mining libraries into PostgreSQL databases, giving rise to her startup, Osogato. She also actively contributed to open-source projects and co-presented significant research on internet security vulnerabilities with her husband Len Sassaman. As a writer, Patterson has written science fiction stories and poetry, often inspired by her scientific interests. She has also engaged in blogging, addressing topics like copyright reform, biohacking, civil rights, and programming languages.
Early life
Patterson lived in and around Houston for 24 years before moving to Iowa City, Iowa, to pursue her Master's degree in linguistics and PhD in computer science. Patterson attended Kingwood High School from 1990 to 1994. She supported herself working as a website designer, technical writer, math teacher, and restaurant critic for the Houston Press. She served as the treasurer of the Mars Society Houston branch in 1999. That same year, at age 22, she traveled above the Arctic Circle as a NASA correspondent for a Mars simulation mission.
Computer science and academic career
Patterson is known for her work in computational linguistics and its applications to computer security |
https://en.wikipedia.org/wiki/Alexandrov%20topology | In topology, an Alexandrov topology is a topology in which the intersection of every family of open sets is open. It is an axiom of topology that the intersection of every finite family of open sets is open; in Alexandrov topologies the finite restriction is dropped.
A set together with an Alexandrov topology is known as an Alexandrov-discrete space or finitely generated space.
Alexandrov topologies are uniquely determined by their specialization preorders. Indeed, given any preorder ≤ on a set X, there is a unique Alexandrov topology on X for which the specialization preorder is ≤. The open sets are just the upper sets with respect to ≤. Thus, Alexandrov topologies on X are in one-to-one correspondence with preorders on X.
Alexandrov-discrete spaces are also called finitely generated spaces since their topology is uniquely determined by the family of all finite subspaces. Alexandrov-discrete spaces can thus be viewed as a generalization of finite topological spaces.
Due to the fact that inverse images commute with arbitrary unions and intersections, the property of being an Alexandrov-discrete space is preserved under quotients.
Alexandrov-discrete spaces are named after the Russian topologist Pavel Alexandrov. They should not be confused with the more geometrical Alexandrov spaces introduced by the Russian mathematician Aleksandr Danilovich Aleksandrov.
Characterizations of Alexandrov topologies
Alexandrov topologies have numerous characterizations. Let X = <X, T> be a topological space. Then the following are equivalent:
Open and closed set characterizations:
Open set. An arbitrary intersection of open sets in X is open.
Closed set. An arbitrary union of closed sets in X is closed.
Neighbourhood characterizations:
Smallest neighbourhood. Every point of X has a smallest neighbourhood.
Neighbourhood filter. The neighbourhood filter of every point in X is closed under arbitrary intersections.
Interior and closure algebraic characterizations:
Interior |
https://en.wikipedia.org/wiki/John%20Marioni | John Marioni is the Head of Research at the European Bioinformatics Institute and holds an appointment at the Wellcome Sanger Institute and the Cancer Research UK Cambridge Institute in Cambridge University. He is a computational biologist known for his research on statistical and computational methods for the analysis of genomics data, in particular single-cell biology and evolutionary genomics. He co-chairs the Human Cell Atlas Analysis Working Group.
Research and career
Marioni has conducted influential studies in whole-tissue and single-cell transcriptomics. He received his PhD from the University of Cambridge in Applied Mathematics in 2008. He conducted postdoctoral studies in the University of Chicago under Matthew Stephens.
Marioni is known for "pioneering the statistical analysis of gene expression patterns in individual cells, which has led to a radical paradigm shift in the field of transcriptomics."
Awards and honours
Blavatnik Awards Young Scientists 2021 United Kingdom Award Finalist
Elected Fellow of St Edmund's College, Cambridge (2016)
John Wishart Prize for Statistics, awarded by Statistical Laboratory of University of Cambridge (2004)
Select publications
R Argelaguet, SJ Clark, H Mohammed, et al., JC Marioni, W Reik. Multi-omics Profiling of Mouse Gastrulation at Single-cell Resolution. Nature. doi:10.1038/s41586-019-1825-8
L Haghverdi, ATL Lun, MD Morgan, JC Marioni. Batch Effects in Single-cell RNA-sequencing Data are Corrected by Matching Mutual Nearest Neighbors. Nature Biotechnology. doi:10.1038/nbt.4091
B Pijuan-Sala, JA Griffiths, C Guibentif, et al., JC Marioni, B Göttgens. A Single-cell Molecular Map of Mouse Gastrulation and Early Organogenesis. Nature. doi:10.1038/s41586-019-0933-9
T Lohoff, S Ghazanfar, A Missarova, N Koulena, N Pierson, JA Griffiths, ES Bardot, CL Eng, RCV Tyser, R Argelaguet, C Guibentif, S Srinivas, J Briscoe, BD Simons, AK Hadjantonakis, B Göttgens, W Reik, J Nichols, L Cai, JC Marioni. Integration |
https://en.wikipedia.org/wiki/Diophantine%20approximation | In number theory, the study of Diophantine approximation deals with the approximation of real numbers by rational numbers. It is named after Diophantus of Alexandria.
The first problem was to know how well a real number can be approximated by rational numbers. For this problem, a rational number a/b is a "good" approximation of a real number α if the absolute value of the difference between a/b and α may not decrease if a/b is replaced by another rational number with a smaller denominator. This problem was solved during the 18th century by means of continued fractions.
Knowing the "best" approximations of a given number, the main problem of the field is to find sharp upper and lower bounds of the above difference, expressed as a function of the denominator. It appears that these bounds depend on the nature of the real numbers to be approximated: the lower bound for the approximation of a rational number by another rational number is larger than the lower bound for algebraic numbers, which is itself larger than the lower bound for all real numbers. Thus a real number that may be better approximated than the bound for algebraic numbers is certainly a transcendental number.
This knowledge enabled Liouville, in 1844, to produce the first explicit transcendental number. Later, the proofs that and e are transcendental were obtained by a similar method.
Diophantine approximations and transcendental number theory are very close areas that share many theorems and methods. Diophantine approximations also have important applications in the study of Diophantine equations.
The 2022 Fields Medal was awarded to James Maynard for his work on Diophantine approximation.
Best Diophantine approximations of a real number
Given a real number , there are two ways to define a best Diophantine approximation of . For the first definition, the rational number is a best Diophantine approximation of if
for every rational number {{math|p/q' }} different from such that .
For the |
https://en.wikipedia.org/wiki/Chrystal%27s%20equation | In mathematics, Chrystal's equation is a first order nonlinear ordinary differential equation, named after the mathematician George Chrystal, who discussed the singular solution of this equation in 1896. The equation reads as
where are constants, which upon solving for , gives
This equation is a generalization of Clairaut's equation since it reduces to Clairaut's equation under certain condition as given below.
Solution
Introducing the transformation gives
Now, the equation is separable, thus
The denominator on the left hand side can be factorized if we solve the roots of the equation and the roots are , therefore
If , the solution is
where is an arbitrary constant. If , () then the solution is
When one of the roots is zero, the equation reduces to Clairaut's equation and a parabolic solution is obtained in this case, and the solution is
The above family of parabolas are enveloped by the parabola , therefore this enveloping parabola is a singular solution. |
https://en.wikipedia.org/wiki/Encyclopedia%20of%20Genetics | The Encyclopedia of Genetics () is a print encyclopedia of genetics edited by Sydney Brenner and Jeffrey H. Miller. It has four volumes and 1,700 entries. It is available online at http://www.sciencedirect.com/science/referenceworks/9780122270802.
Genetics
Genetics literature |
https://en.wikipedia.org/wiki/Phantasy%20Star%20Online%202 | is a free-to-play online action role-playing game in the Phantasy Star series, developed and published by Sega. It was created as a successor to Phantasy Star Online and Phantasy Star Universe, Phantasy Star Online 2 features gameplay elements and aesthetics reminiscent of previous Phantasy Star games while incorporating a few unique twists on the formula. The first version was released for Windows in Japan in July 2012.
A PlayStation Vita version was released in February 2013, but shut down in September 2020. A PlayStation 4 version was released in April 2016. A spin-off/companion game, was released for Android and iOS in 2014. A cloud version for Nintendo Switch, entitled was released in April 2018 and then on PC, released in December 2018. At E3 2019, it was announced that the game would be releasing outside of Asia for the first time on the Xbox One and Windows in early 2020.
A massive updated and separate "shared universe" game, ( in the cloud version), was released on June 9, 2021. It was released for Windows, Xbox One and Xbox Series X/S worldwide, with the PlayStation 4 version releasing on August 31, 2022, and for Windows (download and cloud), Nintendo Switch (cloud) and PlayStation 4 in Japan. Although it was released nearly a decade after Phantasy Star Online 2 launch and is substantially different in gameplay, graphics, and content, the creators decided to not name it "Phantasy Star Online 3" or focus on creating a game with no backwards compatibility with PSO2 content in order to continue support for current players and to not split up the player-base between two different simultaneously available games.
The game is a commercial success, making over 900 million dollars since its release in 2012.
Gameplay
Characters
Upon starting the game, the player can select one of several servers, known as "Ships", to play. Players create and customize their characters to be used in the game. The name, gender, race, character class, body and facial features |
https://en.wikipedia.org/wiki/Metastasis | Metastasis is a pathogenic agent's spread from an initial or primary site to a different or secondary site within the host's body; the term is typically used when referring to metastasis by a cancerous tumor. The newly pathological sites, then, are metastases (mets). It is generally distinguished from cancer invasion, which is the direct extension and penetration by cancer cells into neighboring tissues.
Cancer occurs after cells are genetically altered to proliferate rapidly and indefinitely. This uncontrolled proliferation by mitosis produces a primary heterogeneic tumour. The cells which constitute the tumor eventually undergo metaplasia, followed by dysplasia then anaplasia, resulting in a malignant phenotype. This malignancy allows for invasion into the circulation, followed by invasion to a second site for tumorigenesis.
Some cancer cells known as circulating tumor cells acquire the ability to penetrate the walls of lymphatic or blood vessels, after which they are able to circulate through the bloodstream to other sites and tissues in the body. This process is known (respectively) as lymphatic or hematogenous spread. After the tumor cells come to rest at another site, they re-penetrate the vessel or walls and continue to multiply, eventually forming another clinically detectable tumor. This new tumor is known as a metastatic (or secondary) tumor. Metastasis is one of the hallmarks of cancer, distinguishing it from benign tumors. Most cancers can metastasize, although in varying degrees. Basal cell carcinoma for example rarely metastasizes.
When tumor cells metastasize, the new tumor is called a secondary or metastatic tumor, and its cells are similar to those in the original or primary tumor. This means that if breast cancer metastasizes to the lungs, the secondary tumor is made up of abnormal breast cells, not of abnormal lung cells. The tumor in the lung is then called metastatic breast cancer, not lung cancer. Metastasis is a key element in cancer sta |
https://en.wikipedia.org/wiki/Self-perpetuation | Self-perpetuation, the capability of something to cause itself to continue to exist, is one of the main characteristics of life. Organisms' capability of reproduction leads to self-perpetuation of the species, if not to the individual. Populations self-perpetuate and grow. Entire ecosystems show homeostasis, and thus perpetuate themselves. The slow modifying effect of succession and similar shifts in the composition of the system can, however, not be neglected in the long run. Overall, life's object's capabilities of self-perpetuation are always accompanied by evolution, a perfect steady state of the biological system is never reached. Sexual reproduction is also a form of imperfect self-replication and thus imperfect self-perpetuation because of recombination and mutation. Organisms are not like self-replicating machine but amass random modifications from generation to generation. The property of self-perpetuation in the strict sense thus only applies to life itself.
In a social context, self-perpetuation is tied to reflexivity and (usually) positive feedback loops:
Depending on the time scope or the context, self-perpetuation either depends on self-sustainability, or is equivalent to it. While we may talk about the self-sustainability of an ecosystem, this depends amongst other factor on the self-perpetuation of its constituting species.
In computer science, self-reproducing programs constitute an incomplete metaphor for self-perpetuation. A better analogue can be seen in computer viruses which are actually able to self-reproduce - given a suitable computing environment. |
https://en.wikipedia.org/wiki/Medard%20W.%20Welch%20Award | The Medard W. Welch Award is given to scientists who demonstrated outstanding research in the fields pertinent to American Vacuum Society. It was established in 1969 in memory of Medard W. Welch, a founder of the American Vacuum Society.
List of recipients
See also
List of physics awards |
https://en.wikipedia.org/wiki/In-cell%20charge%20control | In-Cell Charge Control or I-C3 is a method for very rapid charging of a Nickel-metal hydride battery, patented by Rayovac. Batteries using this technology are commonly sold as "15-minute rechargeables".
The charge control consists of a pressure switch built into the cell, which disconnects the charging current when the internal cell pressure rises above a certain limit; usually to . This prevents overcharging and damage to the cell.
Sources
Battery charging |
https://en.wikipedia.org/wiki/Desktop%20virtualization | Desktop virtualization is a software technology that separates the desktop environment and associated application software from the physical client device that is used to access it.
Desktop virtualization can be used in conjunction with application virtualization and user profile management systems, now termed user virtualization, to provide a comprehensive desktop environment management system. In this mode, all the components of the desktop are virtualized, which allows for a highly flexible and much more secure desktop delivery model. In addition, this approach supports a more complete desktop disaster recovery strategy as all components are essentially saved in the data center and backed up through traditional redundant maintenance systems. If a user's device or hardware is lost, the restore is straightforward and simple, because the components will be present at login from another device. In addition, because no data are saved to the user's device, if that device is lost, there is much less chance that any critical data can be retrieved and compromised.
System architectures
Desktop virtualization implementations are classified based on whether the virtual desktop runs remotely or locally, on whether the access is required to be constant or is designed to be intermittent, and on whether or not the virtual desktop persists between sessions. Typically, software products that deliver desktop virtualization solutions can combine local and remote implementations into a single product to provide the most appropriate support specific to requirements. The degrees of independent functionality of the client device is necessarily interdependent with the server location and access strategy. And virtualization is not strictly required for remote control to exist. Virtualization is employed to present independent instances to multiple users and requires a strategic segmentation of the host server and presentation at some layer of the host's architecture. The enabling |
https://en.wikipedia.org/wiki/Erwin%20Kreyszig | Erwin Otto Kreyszig (January 6, 1922 in Pirna, Germany – December 12, 2008) was a German Canadian applied mathematician and the Professor of Mathematics at Carleton University in Ottawa, Ontario, Canada. He was a pioneer in the field of applied mathematics: non-wave replicating linear systems. He was also a distinguished author, having written the textbook Advanced Engineering Mathematics, the leading textbook for civil, mechanical, electrical, and chemical engineering undergraduate engineering mathematics.
Kreyszig received his PhD degree in 1949 at the University of Darmstadt under the supervision of Alwin Walther. He then continued his research activities at the universities of Tübingen and Münster. Prior to joining Carleton University in 1984, he held positions at Stanford University (1954/55), the University of Ottawa (1955/56), Ohio State University (1956–60, professor 1957) and he completed his habilitation at the University of Mainz. In 1960 he became professor at the Technical University of Graz and organized the Graz 1964 Mathematical Congress. He worked at the University of Düsseldorf (1967–71) and at the University of Karlsruhe (1971–73). From 1973 through 1984 he worked at the University of Windsor and since 1984 he had been at Carleton University. He was awarded the title of Distinguished Research Professor in 1991 in recognition of a research career during which he published 176 papers in refereed journals, and 37 in refereed conference proceedings.
Kreyszig was also an administrator, developing a Computer Centre at the University of Graz, and at the Mathematics Institute at the University of Düsseldorf. In 1964, he took a leave of absence from Graz to initiate a doctoral program in mathematics at Texas A&M University.
Kreyszig authored 14 books, including Advanced Engineering Mathematics, which was published in its 10th edition in 2011. He supervised 104 master's and 22 doctoral students as well as 12 postdoctoral researchers. Together with his so |
https://en.wikipedia.org/wiki/Machine%20learning%20in%20earth%20sciences | Applications of machine learning in earth sciences include geological mapping, gas leakage detection and geological features identification. Machine learning (ML) is a type of artificial intelligence (AI) that enables computer systems to classify, cluster, identify and analyze vast and complex sets of data while eliminating the need for explicit instructions and programming. Earth science is the study of the origin, evolution, and future of the planet Earth. The Earth system can be subdivided into four major components including the solid earth, atmosphere, hydrosphere and biosphere.
A variety of algorithms may be applied depending on the nature of the earth science exploration. Some algorithms may perform significantly better than others for particular objectives. For example, convolutional neural networks (CNN) are good at interpreting images, artificial neural networks (ANN) perform well in soil classification but more computationally expensive to train than support-vector machine (SVM) learning. The application of machine learning has been popular in recent decades, as the development of other technologies such as unmanned aerial vehicles (UAVs), ultra-high resolution remote sensing technology and high-performance computing units lead to the availability of large high-quality datasets and more advanced algorithms.
Significance
Complexity of earth science
Problems in earth science are often complex. It is difficult to apply well-known and described mathematical models to the natural environment, therefore machine learning is commonly a better alternative for such non-linear problems. Ecological data are commonly non-linear and consist of higher-order interactions, and together with missing data, traditional statistics may underperform as unrealistic assumptions such as linearity are applied to the model. A number of researchers found that machine learning outperforms traditional statistical models in earth science, such as in characterizing forest canopy stru |
https://en.wikipedia.org/wiki/Virosome | A virosome is a drug or vaccine delivery mechanism consisting of unilamellar phospholipid membrane (either a mono- or bi-layer) vesicle incorporating virus derived proteins to allow the virosomes to fuse with target cells. Viruses are infectious agents that can replicate in their host organism, however virosomes do not replicate. The properties that virosomes share with viruses are based on their structure; virosomes are essentially safely modified viral envelopes that contain the phospholipid membrane and surface glycoproteins. As a drug or vaccine delivery mechanism they are biologically compatible with many host organisms and are also biodegradable. The use of reconstituted virally derived proteins in the formation of the virosome allows for the utilization of what would otherwise be the immunogenic properties of a live-attenuated virus, but is instead a safely killed virus. A safely killed virus can serve as a promising vector because it won't cause infection and the viral structure allows the virosome to recognize specific components of its target cells.
Virosomes structure
Virosomes are vehicles that have a spherical shape with a phospholipid mono/bilayer membrane. Inside of the virosome, there is a central cavity that holds the therapeutic molecules such as nucleic acids, proteins, and drugs. On the surface of the virosome, there can be different types of glycoproteins. Glycoproteins are a type of protein that have an oligosaccharide chain bonded to amino acid chains. The different types of glycoproteins on the surface of the virosome increases the specificity of the target cells because the surface glycoproteins help with recognition as well as the attachments of the virosomes to their target cells. In the case of the influenza virosome, the glycoproteins are antigen, haemagglutinin, and neuraminidase. Antigens are molecules that triggers an immune response when targeted by a specific antibody that corresponds to the shape of the antigen. Haemagglutinin is |
https://en.wikipedia.org/wiki/Hybrid%20plasmid | Hybrid plasmid is a plasmid that contains an inserted piece of foreign DNA. |
https://en.wikipedia.org/wiki/Epizootiology | Epizootiology, epizoology, or veterinary epidemiology is the study of disease patterns within animal populations.
See also
Epizootic
Epidemiology |
https://en.wikipedia.org/wiki/Gluten%20challenge%20test | The gluten challenge test is a medical test in which gluten-containing foods are consumed and (re-)occurrence of symptoms is observed afterwards to determine whether and how much a person reacts to these foods. The test may be performed in people with suspected gluten-related disorders in very specific occasions and under medical supervision, for example in people who had started a gluten-free diet without performing duodenal biopsy.
Gluten challenge is discouraged before the age of 5 years and during pubertal growth.
Gluten challenge protocols have significant limitations because a symptomatic relapse generally precedes the onset of a serological and histological relapse, and therefore becomes unacceptable for most patients.
History
Before serological and biopsy-based diagnosis of coeliac disease was available, a gluten challenge test was a prerequisite for diagnosis of coeliac disease.
Today, with serological testing (determination of coeliac disease-specific antibodies in the blood) and duodenal biopsy with histological testing being available for diagnosing coeliac disease, patients with suspected coeliac disease are strongly advised to undergo both serological and biopsy testing before undertaking a gluten-free diet. People who present minor damage of the small intestine often have negative blood antibodies titers and many patients with coeliac disease are missed when a duodenal biopsy is not performed. Serologic tests have a high capacity to detect coeliac disease only in patients with total villous atrophy and have very low capacity to detect cases with partial villous atrophy or minor intestinal lesions with normal villi. Currently, gluten challenge is no longer required to confirm the diagnosis in patients with intestinal lesions compatible with coeliac disease and a positive response to a gluten-free diet. Nevertheless, in some cases, a gluten challenge with a subsequent biopsy may be useful to support the diagnosis, for example in people with a high s |
https://en.wikipedia.org/wiki/Ralstonia%20virus%20RSA1 | Ralstonia virus RSA1 is a virus of the family Myoviridae, genus Aresaunavirus.
As a member of the group I of the Baltimore classification, Ralstonia virus RSA1 is a dsDNA viruses. All the family Myoviridae members share a nonenveloped morphology consisting of a head and a tail separated by a neck. Its genome is linear. The propagation of the virions includes the attaching to a host cell (a bacterium, as Ralstonia virus RSA1 is a bacteriophage) and the injection of the double stranded DNA; the host transcribes and translates it to manufacture new particles. To replicate its genetic content requires host cell DNA polymerases and, hence, the process is highly dependent on the cell cycle.
Its genome is 38,760 base pairs long with 65.3% of GC content and 5′-extruding cohesive ends; contains 51 open reading frames. |
https://en.wikipedia.org/wiki/Count%20On | Count On is a major mathematics education project in the United Kingdom which was announced by education secretary David Blunkett at the end of 2000. It was the follow-on to Maths Year 2000 which was the UK's contribution to UNICEF's World Mathematical Year.
Count On had two main strands:
The website www.counton.org which won the 2002 BETT prize for best free online learning resource.
"MathFests", which were maths funfairs held around the country, aimed particularly at those who would not normally come into contact with mathematical ideas.
The MathFests were run largely by MatheMagic and the University of York.
The project has now been handed over to the NCETM.
Popularisation of Mathematics
Count On and Maths Year 2000 were some of the first big Popularisation of Mathematics projects. Others are listed below.
International
World Mathematical Year 2000
Statistics 2013
World Maths Day (orig. Australian) - next one is 6 March 2013
Australia
World Maths Day
India
National Mathematics Year
Ireland
Maths Week Ireland
Nigeria
National Mathematics Year
Spain
Matematica Vital
Paul Boron
United Kingdom
Maths Year 2000 Scotland
Maths Cymru (Wales)
United States
Steven Strogatz's blog |
https://en.wikipedia.org/wiki/Weightless%20%28wireless%20communications%29 | Weightless was a set of low-power wide-area network (LPWAN) wireless technology specifications for exchanging data between a base station and many of machines around it.
History
Cambridge Wireless held an event at the Moller Centre in Cambridge, United Kingdom on September 30, 2011. Neul, Landis+Gyr, Cable & Wireless, and ARM Holdings provided presentations. The Weightless Special Interest Group promoted the technology (SIG), announced on December 7, 2012. The group was led by William Webb, a professor at Cambridge and a founder of the company Neul. A subsequent event was held in September 2013, when version 1.0 was published.
The name Weightless was chosen to reflect the intention at a low overhead per transmission for devices that need to communicate just a low amounts of data. The Weightless logo appears in uppercase letters with the 'W' appearing in the top-right corner of a light blue box with a solid blue line above it.
In September 2014, Neul was acquired by Huawei for an estimated $25 million. By 2015, the company Nwave Technologies announced deployments in Copenhagen, Denmark, and Esbjerg, Denmark. However, observers noted no products on the market. A company called Ubiik, based in Taiwan, announced pre-orders in 2017.
Implementation
Weightless-N uses a differential binary phase shift keying (DBPSK) digital modulation scheme to transmit within narrow frequency bands using a frequency hopping algorithm for interference mitigation and enhanced security. It provides encryption and implicit authentication using a shared secret key regime to encode transmitted information via a 128-bit AES algorithm. The technology supports mobility with the network automatically routing terminal messages to the correct destination. Multiple networks, typically operated by different companies, are enabled and can be co-located. Each base station queries a central database to determine which network the terminal is registered to decode and route data accordingly.
Weightle |
https://en.wikipedia.org/wiki/Curie%E2%80%93Weiss%20law | In magnetism, the Curie–Weiss law describes the magnetic susceptibility of a ferromagnet in the paramagnetic region above the Curie temperature:
where is a material-specific Curie constant, is the absolute temperature, and is the Curie temperature, both measured in kelvin. The law predicts a singularity in the susceptibility at . Below this temperature, the ferromagnet has a spontaneous magnetization. The name is given after Pierre Curie and Pierre Weiss.
Background
A magnetic moment which is present even in the absence of the external magnetic field is called spontaneous magnetization. Materials with this property are known as ferromagnets, such as iron, nickel, and magnetite. However, when these materials are heated up, at a certain temperature they lose their spontaneous magnetization, and become paramagnetic. This threshold temperature below which a material is ferromagnetic is called the Curie temperature and is different for each material.
The Curie–Weiss law describes the changes in a material's magnetic susceptibility, , near its Curie temperature. The magnetic susceptibility is the ratio between the material's magnetization and the applied magnetic field.
Limitations
In many materials, the Curie–Weiss law fails to describe the susceptibility in the immediate vicinity of the Curie point, since it is based on a mean-field approximation. Instead, there is a critical behavior of the form
with the critical exponent . However, at temperatures the expression of the Curie–Weiss law still holds true, but with replaced by a temperature that is somewhat higher than the actual Curie temperature. Some authors call the Weiss constant to distinguish it from the temperature of the actual Curie point.
Classical approaches to magnetic susceptibility and Bohr–van Leeuwen theorem
According to the Bohr–van Leeuwen theorem, when statistical mechanics and classical mechanics are applied consistently, the thermal average of the magnetization is always zero. Magn |
https://en.wikipedia.org/wiki/Ramaria%20botrytis | Ramaria botrytis, commonly known as the clustered coral, the pink-tipped coral mushroom, or the cauliflower coral, is an edible species of coral fungus in the family Gomphaceae. Its robust fruit body can grow up to in diameter and tall, and resembles some marine coral. Its dense branches, which originate from a stout, massive base, are swollen at the tips and divided into several small branchlets. The branches are initially whitish but age to buff or tan, with tips that are pink to reddish. The flesh is thick and white. The spores, yellowish in deposit, are ellipsoid, feature longitudinal striations, and measure about 13.8 by 4.7 micrometers.
The type species of the genus Ramaria, R. botrytis was first described scientifically in 1797 by mycologist Christiaan Hendrik Persoon. A widely distributed species, it is found in North America, North Africa, central and eastern Europe, Australia, and Asia. The fungus is mycorrhizal with broadleaf trees, and fruits on the ground in wooded areas. There are several species of coral fungi that are superficially similar in appearance to R. botrytis, and although comparison of habitat or characteristics like color or branching morphology is often sufficient for identification, sometimes microscopy is required to definitively distinguish between them. Fruit bodies of Ramaria botrytis are edible, and young specimens have a mild, fruity taste. Some authors warn of laxative effects in susceptible individuals. The fungus contains several chemical compounds with in vitro biological activity, and fruit bodies have antimicrobial activity against several species and strains of drug-resistant bacteria that cause disease in humans.
Taxonomy and classification
The species was first named as Clavaria botrytis in 1797 by Christian Hendrik Persoon. In 1821, Elias Magnus Fries sanctioned the genus name Clavaria, and treated Ramaria as a section of Clavaria. It was given its current name in 1918 by Adalbert Ricken. Obsolete historical synonyms |
https://en.wikipedia.org/wiki/Therminol | Therminol is a synthetic heat transfer fluid produced by Eastman Chemical Company.
Therminol fluids are used in a variety of applications, including:
Hydrocarbon processing (oil and gas, refining, asphalt, gas-to-liquid, etc.)
Alternative energy and technologies (concentrated solar power, biofuel, organic Rankine cycle, desalination, etc.)
Plastics processing
Chemical processing (pharmaceutical, environmental test chambers, etc.)
Food and beverage processing
Heat transfer system maintenance
Prior to 1997, Therminol fluids were sold in Europe under the trade names SantoTherm and GiloTherm. Since 1997, all forms of Therminol fluid have been sold with the Therminol name and extension to define its uses.
Therminol Products From Eastman Chemical Company
Therminol 55 Heat Transfer Fluid
Therminol 59 Heat Transfer Fluid
Therminol FF (Flush Fluid)
History
Therminol heat transfer fluids were developed in 1963 by Monsanto. In 1997, the chemical businesses of Monsanto were spun off to form a new company called Solutia Inc. In 2012, Solutia was acquired by Eastman Chemical Company.
Polychlorinated biphenyl in Therminol
Prior to 1971, Monsanto marketed a series of polychlorinated biphenyl-(PCB)-containing heat transfer fluids designated as Therminol FR series in the United States and Santotherm FR series in Europe. FR series Therminol heat transfer fluids contained PCBs, which imparted fire resistance. Monsanto voluntarily ceased sales of these fluids in 1971. No form of Therminol heat transfer fluids have contained PCBs since that time. Polychlorinated biphenyl was banned by the United States Congress in 1979 and the Stockholm Convention on Persistent Organic Pollutants in 2001. |
https://en.wikipedia.org/wiki/Diacylglycerol%20oil | Diacylglycerol oil (DAG oil) is a cooking oil in which the ratio of triglycerides, also known as Triacylglycerols (TAGs), to diacylglycerols (DAGs) is shifted to contain mostly DAG, unlike conventional cooking oils, which are rich in TAGs. Vegetable DAG oil, for example, contains 80% DAG and is used as a 1:1 replacement for liquid vegetable oils in all applications.
How it works
DAGs and TAGs are natural components in all vegetable oils. Through an enzymatic process, the DAG content of a combination of soy and canola oils is significantly increased. Unlike TAG, which is stored as body fat, DAG is immediately burned as energy. With DAG-rich oil containing more than 80% DAG, less of the oil is stored as body fat than with traditional oils, which are rich in TAG. Excess calories consumed by the body are converted into fat and stored, regardless if it is consumed as DAG or TAG.
Study
According to a 2007 study,
Diacylglycerol (DAG) oil is present with vegetable oil. A study in 2004 indicated that DAG oil is effective for both fasting and postprandial hyperlipidemia; according to the same study, it helped prevent excess adiposity.
FDA designation
DAG oil was designated as generally recognized as safe (GRAS) by an outside panel of scientific experts, and their conclusion has been reviewed and accepted by the US Food and Drug Administration (FDA). This GRAS determination is for use in vegetable oil spreads and home cooking oil. In Japan, the Ministry of Health, Labor and Welfare has approved DAG oil to manage serum triglycerides after a meal, which leads to less build-up of body fat.
Side effects
Because DAG oil is digested the same way as conventional vegetable oils, the potential side effects are no different than those of conventional oil. In addition, studies with animals and human subjects have shown no adverse effects from single-dose or long-term consumption of DAG-rich oil. It has also been found that fat-soluble vitamins' status is not affected by the |
https://en.wikipedia.org/wiki/Protein%20splicing | Protein splicing is an intramolecular reaction of a particular protein in which an internal protein segment (called an intein) is removed from a precursor protein with a ligation of C-terminal and N-terminal external proteins (called exteins) on both sides. The splicing junction of the precursor protein is mainly a cysteine or a serine, which are amino acids containing a nucleophilic side chain. The protein splicing reactions which are known now do not require exogenous cofactors or energy sources such as adenosine triphosphate (ATP) or guanosine triphosphate (GTP). Normally, splicing is associated only with pre-mRNA splicing. This precursor protein contains three segments—an N-extein followed by the intein followed by a C-extein. After splicing has taken place, the resulting protein contains the N-extein linked to the C-extein; this splicing product is also termed an extein.
History
The first intein was discovered in 1988 through sequence comparison between the Neurospora crassa and carrot vacuolar ATPase (without intein) and the homologous gene in yeast (with intein) that was first described as a putative calcium ion transporter. In 1990 Hirata et al. demonstrated that the extra sequence in the yeast gene was transcribed into mRNA and removed itself from the host protein only after translation. Since then, inteins have been found in all three domains of life (eukaryotes, bacteria, and archaea) and in viruses.
Protein splicing was unanticipated and its mechanisms were discovered by two groups (Anraku and Stevens) in 1990. They both discovered a Saccharomyces cerevisiae VMA1 in a precursor of a vacuolar H+-ATPase enzyme. The amino acid sequence of the N- and C-termini corresponded to 70% DNA sequence of that of a vacuolar H+-ATPase from other organisms, while the amino acid sequence of the central position corresponded to 30% of the total DNA sequence of the yeast HO nuclease.
Many genes have unrelated intein-coding segments inserted at different positions. For |
https://en.wikipedia.org/wiki/Geometrized%20unit%20system | A geometrized unit system, geometric unit system or geometrodynamic unit system is a system of natural units in which the base physical units are chosen so that the speed of light in vacuum, c, and the gravitational constant, G, are set equal to unity.
The geometrized unit system is not a completely defined system. Some systems are geometrized unit systems in the sense that they set these, in addition to other constants, to unity, for example Stoney units and Planck units.
This system is useful in physics, especially in the special and general theories of relativity. All physical quantities are identified with geometric quantities such as areas, lengths, dimensionless numbers, path curvatures, or sectional curvatures.
Many equations in relativistic physics appear simpler when expressed in geometric units, because all occurrences of G and of c drop out. For example, the Schwarzschild radius of a nonrotating uncharged black hole with mass m becomes . For this reason, many books and papers on relativistic physics use geometric units. An alternative system of geometrized units is often used in particle physics and cosmology, in which instead. This introduces an additional factor of 8π into Newton's law of universal gravitation but simplifies the Einstein field equations, the Einstein–Hilbert action, the Friedmann equations and the Newtonian Poisson equation by removing the corresponding factor.
Practical measurements and computations are usually done in SI units, but conversions are generally quite straightforward.
Definition
In geometric units, every time interval is interpreted as the distance travelled by light during that given time interval. That is, one second is interpreted as one light-second, so time has the geometric units of length. This is dimensionally consistent with the notion that, according to the kinematical laws of special relativity, time and distance are on an equal footing.
Energy and momentum are interpreted as components of the four-mome |
https://en.wikipedia.org/wiki/Lola%20J.%20May | Lola J. May (October 29, 1923 – March 13, 2007) was a mathematics educator, consultant, author, producer of audio-visual materials, an early proponent of the new math educational process, and a household name among mathematics.
Life
Her father was a salesman and her mother was a homemaker. Her father taught her mathematics every night using a movable blackboard and a collection of coins. She found her early schooling boring and too strict, and she did not initially consider becoming a teacher.
A native of Kenosha, Wisconsin and a summa cum laude graduate of the University of Wisconsin–Madison in 1945, where she received her B.S. in mathematics and science. After teaching high school for three years, she studied and achieved her master's degree in mathematics at Northwestern University in 1950 and her doctorate in mathematics education from there in 1964. She taught mathematics at New Trier Township High School in the Chicago area until 1960, and was a mathematics consultant at the Winnetka, Illinois public schools until 1998. Her summers were often spent teaching at the university level, but she taught mathematics to all grades over the course of her career.
She promised herself to make her students laugh and ask questions. She did not want her students to be bored by or scared of mathematics. She succeeded; her students cheered when they figured out the answers to math problems and lamented when class time with Dr. May was over.
May explained, "The big thing I have going for me is my enthusiasm. There are people who are brighter than I am. There are people who may be better teachers-although I'm pretty good at teaching-and there are certainly people who are better writers. But I have enthusiasm." This enthusiasm was not unnoticed; teachers in the same hallway as her described how loud she was.
May died on March 13, 2007, in Evanston, Illinois, at the age of 83.
Contributions
Her authored works include her autobiography "Lola May Who?", the book "Teaching Math |
https://en.wikipedia.org/wiki/PlusCal | PlusCal (formerly called +CAL) is a formal specification language created by Leslie Lamport, which transpiles to TLA+. In contrast to TLA+'s action-oriented focus on distributed systems, PlusCal most resembles an imperative programming language and is better-suited when specifying sequential algorithms. PlusCal was designed to replace pseudocode, retaining its simplicity while providing a formally-defined and verifiable language. A one-bit clock is written in PlusCal as follows:
-- fair algorithm OneBitClock {
variable clock \in {0, 1};
{
while (TRUE) {
if (clock = 0)
clock := 1
else
clock := 0
}
}
}
See also
TLA+
Pseudocode |
https://en.wikipedia.org/wiki/Wiener%27s%20attack | The Wiener's attack, named after cryptologist Michael J. Wiener, is a type of cryptographic attack against RSA. The attack uses the continued fraction method to expose the private key d when d is small.
Background on RSA
Fictional characters Alice and Bob are people who want to communicate securely. More specifically, Alice wants to send a message to Bob which only Bob can read. First Bob chooses two primes p and q. Then he calculates the RSA modulus N = pq. This RSA modulus is made public together with the encryption exponent e. N and e form the public key pair (e, N). By making this information public, anyone can encrypt messages to Bob. The decryption exponent d satisfies , where denotes the Carmichael function, though sometimes , the Euler’s phi function, is used (note: this is the order of the multiplicative group , which is not necessarily a cyclic group). The encryption exponent e and also must be relatively prime so that there is a modular inverse. The factorization of N and the private key d are kept secret, so that only Bob can decrypt the message. We denote the private key pair as (d, N). The encryption of the message M is given by and the decryption of cipher text is given by (using Fermat's little theorem).
Using the Euclidean algorithm, one can efficiently recover the secret key d if one knows the factorization of N. By having the secret key d, one can efficiently factor the modulus of N.
Small private key
In the RSA cryptosystem, Bob might tend to use a small value of d, rather than a large random number to improve the RSA decryption performance. However, Wiener’s attack shows that choosing a small value for d will result in an insecure system in which an attacker can recover all secret information, i.e., break the RSA system. This break is based on Wiener’s Theorem, which holds for small values of d. Wiener has proved that the attacker may efficiently find d when .
Wiener's paper also presented some countermeasures against his attack th |
https://en.wikipedia.org/wiki/Cdc25 | Cdc25 is a dual-specificity phosphatase first isolated from the yeast Schizosaccharomyces pombe as a cell cycle defective mutant. As with other cell cycle proteins or genes such as Cdc2 and Cdc4, the "cdc" in its name refers to "cell division control".
Dual-specificity phosphatases are considered a sub-class of protein tyrosine phosphatases. By removing inhibitory phosphate residues from target cyclin-dependent kinases (Cdks), Cdc25 proteins control entry into and progression through various phases of the cell cycle, including mitosis and S ("Synthesis") phase.
Function in activating Cdk1
Cdc25 activates cyclin dependent kinases by removing phosphate from residues in the Cdk active site. In turn, the phosphorylation by M-Cdk (a complex of Cdk1 and cyclin B) activates Cdc25. Together with Wee1, M-Cdk activation is switch-like. The switch-like behavior forces entry into mitosis to be quick and irreversible. Cdk activity can be reactivated after dephosphorylation by Cdc25. The Cdc25 enzymes Cdc25A-C are known to control the transitions from G1 to S phase and G2 to M phase.
Structure
The structure of Cdc25 proteins can be divided into two main regions: the N-terminal region, which is highly divergent and contains sites for its phosphorylation and ubiquitination, which regulate the phosphatase activity; and the C-terminal region, which is highly homologous and contains the catalytic site.
Evolution and species distribution
Cdc25 enzymes are well conserved through evolution, and have been isolated from fungi such as yeasts as well as all metazoans examined to date, including humans. The exception among eukaryotes may be plants, as the purported plant Cdc25s have characteristics, (such as the use of cations for catalysis), that are more akin to serine/threonine phosphatases than dual-specificity phosphatases, raising doubts as to their authenticity as Cdc25 phosphatases. The Cdc25 family appears to have expanded in relation to the complexity of the cell-cycle and li |
https://en.wikipedia.org/wiki/IOS%2012 | iOS 12 is the twelfth major release of the iOS mobile operating system developed by Apple Inc. Aesthetically similar to its predecessor, iOS 11, it focuses more on performance than on new features, quality improvements and security updates. Announced at the company's Worldwide Developers Conference on June 4, 2018, iOS 12 was released to the public on September 17, 2018. It was succeeded for the iPhone and iPod Touch by iOS 13 on September 19, 2019 and for the iPad by iPadOS 13 on September 24, 2019. Security updates for iOS 12 continued for five years following the release of iOS 13 and iPadOS 13 for devices unable to run the newer versions. The last update, 12.5.7, was released on January 23, 2023.
Overview
iOS 12 was introduced by Craig Federighi at the Apple Worldwide Developers Conference keynote address on June 4, 2018. The first developer beta version was released after the keynote presentation, with the first public beta released on June 25, 2018. The initial release of version 12.0 was on September 17, 2018.
System features
Performance
Performance optimizations were made in order to speed up common tasks across all supported iOS devices. Tests done by Apple on an iPhone 6 Plus showed apps launching 40% faster, the system keyboard activating 50% faster, and the camera opening 70% faster compared to iOS 11.
Screen Time
Screen Time is a new feature in iOS 12 that records the amount of time a user spent on the device. The feature also displays the amount of time the user used particular apps, the amount of time the user used particular categories of apps (such as games), and the number of notifications the user received.
Screen Time also provides blocking features to limit usage of apps (with time limits) or set other restrictions such as on purchases or explicit content. It replaces Restrictions in the iOS Settings app, but can also be used by adults to limit their own usage. These features can be used with or without a passcode. Without setting a passc |
https://en.wikipedia.org/wiki/Cosmosoma%20fenestrata | Cosmosoma fenestrata is a moth of the family Erebidae. It was described by Dru Drury in 1773. It is found on Jamaica and Cuba.
Description
Upperside: Antennae nearly black and pectinated (comb like). Head and eyes black. Palpi small and long, and of a fine scarlet. Tongue spiral. Thorax blueish green, almost black. Abdomen dark brown. Anterior wings the same, the middle part being perfectly transparent like glass, wherein is a dark cloud which is joined to the anterior margin. Posterior wings small, transparent, with a dark brown narrow border running round their edges, which at the upper corners is broad where it becomes cloud like.
Underside: Breast dark brown. Legs and thighs scarlet, which colour extends along the middle of the abdomen, almost to the tail, where it becomes a little fainter, being crossed by the rings of the abdomen, which are black and very narrow. All the wings of the same colour as on the upperside. |
https://en.wikipedia.org/wiki/Y%20Chromosome%20Consortium | The Y Chromosome Consortium (YCC) was a collection of scientists who worked toward the understanding of human Y chromosomal phylogenetics and evolution. The consortium had the following objectives: web resources that communicate information relating to the non-recombinant region of the Y-chromosome including new variants and changes in the nomenclature. The consortium sponsored literature regarding updates in the phylogenetics and nomenclature.
See also
International Society of Genetic Genealogy (ISOGG)
Human Y-chromosome DNA haplogroup |
https://en.wikipedia.org/wiki/Malicious%20Software%20Removal%20Tool | Microsoft Windows Malicious Software Removal Tool (MSRT) is a freeware second-opinion malware scanner that Microsoft's Windows Update downloads and runs on Windows computers each month, independent of the install antivirus software. First released on January 13, 2005, MSRT does not offer real-time protection. It scans its host computer for specific, widespread malware, and tries to eliminate the infection. Outside its monthly deployment schedule, it can be separately downloaded from Microsoft.
Availability
Since its January 13, 2005, Microsoft releases the updated tool every second Tuesday of every month (commonly called "Patch Tuesday") through Windows Update, at which point it runs once automatically in the background and reports if malicious software is found. The tool is also available as a standalone download.
Since support for Windows 2000 ended on July 13, 2010, Microsoft stopped distributing the tool to Windows 2000 users via Windows Update. The last version of the tool that could run on Windows 2000 was 4.20, released on May 14, 2013. Starting with version 5.1, released on June 11, 2013, support for Windows 2000 was dropped altogether. Although Windows XP support ended on April 8, 2014, updates for the Windows XP version of the Malicious Software Removal Tool would be provided until August, 2016; version 5.39. The latest version of MSRT for Windows Vista is 5.47, released on 11 April 2017.
Despite Microsoft ending general support for the Windows 7 operating system in 2020, updates are still provided to Windows 7 users via the standard Windows Update delivery mechanism.
Operation
MSRT does not install a shortcut in the Start menu. Hence, users must manually execute %windir%\system32\mrt.exe. The tool records its results in a log file located at %windir%\debug\mrt.log.
The tool reports anonymized data about any detected infections to Microsoft. MSRT's EULA discloses this reporting behavior and explains how to disable it.
Impact
In a June 2006 Micro |
https://en.wikipedia.org/wiki/Impact%20driver | An impact driver is a tool that delivers a strong, sudden rotational force and forward thrust. The force can be delivered either by striking with a hammer in the case of manual impact drivers, or mechanically in the case of powered impact drivers.
It is often used by mechanics to loosen larger screws, bolts and nuts that are corrosively "frozen" or over-torqued. The direction can also be reversed for situations where screws have to be tightened with torque greater than a screwdriver can reasonably provide.
Manual impact drivers
Manual impact drivers consist of a heavy outer sleeve that surrounds an inner core that is splined to it. The spline is curved so that when the user strikes the outer sleeve with a hammer, its downward force works on the spline to produce turning force on the core and any socket or work bit attached to it. The tool translates the heavy rotational inertia of the sleeve to the lighter core to generate large amounts of torque. At the same time, the striking blow from the hammer forces the impact driver forward into the screw reducing or eliminating cam out. This attribute is beneficial for Phillips screws which are prone to cam out. It is also excellent for use with the Robertson square socket head screws that are in common use in Canada.
Powered impact drivers
Typical battery-powered impact drivers are similar to electric drills when used to drive screws or bolts, but additionally have a spring-driven mechanism that applies rotational striking blows once the torque required becomes too great for the motor alone. This shouldn't be confused with the hammer mechanism found on hammer drills, which is a longitudinal blow. Most impact drivers have a handle to make it easier to hold onto. The impact drivers can be used on various types of nuts and bolts, and surfaces. Various materials may be used with it, including steel, iron, wood, plastic, and more. An impact driver is more appropriate than a drill for tightening bolts.
Compared to an impa |
https://en.wikipedia.org/wiki/Frank%E2%80%93Tamm%20formula | The Frank–Tamm formula yields the amount of Cherenkov radiation emitted on a given frequency as a charged particle moves through a medium at superluminal velocity. It is named for Russian physicists Ilya Frank and Igor Tamm who developed the theory of the Cherenkov effect in 1937, for which they were awarded a Nobel Prize in Physics in 1958.
When a charged particle moves faster than the phase speed of light in a medium, electrons interacting with the particle can emit coherent photons while conserving energy and momentum. This process can be viewed as a decay. See Cherenkov radiation and nonradiation condition for an explanation of this effect.
Equation
The energy emitted per unit length travelled by the particle per unit of frequency is:
provided that . Here and are the frequency-dependent permeability and index of refraction of the medium respectively, is the electric charge of the particle, is the speed of the particle, and is the speed of light in vacuum.
Cherenkov radiation does not have characteristic spectral peaks, as typical for fluorescence or emission spectra. The relative intensity of one frequency is approximately proportional to the frequency. That is, higher frequencies (shorter wavelengths) are more intense in Cherenkov radiation. This is why visible Cherenkov radiation is observed to be brilliant blue. In fact, most Cherenkov radiation is in the ultraviolet spectrum; the sensitivity of the human eye peaks at green, and is very low in the violet portion of the spectrum.
The total amount of energy radiated per unit length is:
This integral is done over the frequencies for which the particle's speed is greater than speed of light of the media . The integral is convergent (finite) because at high frequencies the refractive index becomes less than unity and for extremely high frequencies it becomes unity.
Derivation of Frank–Tamm formula
Consider a charged particle moving relativistically along -axis in a medium with refraction index |
https://en.wikipedia.org/wiki/Adamkiewicz%20reaction | The Adamkiewicz reaction is part of a biochemical test used to detect the presence of the amino acid tryptophan in proteins. When concentrated sulfuric acid is combined with a solution of protein and glyoxylic acid, a red/purple colour is produced. It was named after its discoverer, Albert Wojciech Adamkiewicz. Pure sulphuric acid and a minimal amount of pure formaldehyde, along with an oxidizing agent introduced into the sulphuric acid, allow the reaction to proceed. |
https://en.wikipedia.org/wiki/18%20%28number%29 | 18 (eighteen) is the natural number following 17 and preceding 19.
In mathematics
Eighteen is the tenth composite number, its divisors being 1, 2, 3, 6 and 9. Three of these divisors (3, 6 and 9) add up to 18, hence 18 is a semiperfect number. Eighteen is the first inverted square-prime of the form p·q2.
In base ten, it is a Harshad number.
It is an abundant number, as the sum of its proper divisors is greater than itself (1+2+3+6+9 = 21). It is known to be a solitary number, despite not being coprime to this sum.
It is the number of one-sided pentominoes.
It is the only number where the sum of its written digits in base 10 (1+8 = 9) is equal to half of itself (18/2 = 9).
It is a Fine number.
In science
Chemistry
Eighteen is the atomic number of argon.
Group 18 of the periodic table is called the noble gases.
The 18-electron rule is a rule of thumb in transition metal chemistry for characterising and predicting the stability of metal complexes.
In religion and literature
The Hebrew word for "life" is (chai), which has a numerical value of 18. Consequently, the custom has arisen in Jewish circles to give donations and monetary gifts in multiples of 18 as an expression of blessing for long life.
In Judaism, in the Talmud; Pirkei Avot (5:25), Rabbi Yehudah ben Teime gives the age of 18 as the appropriate age to get married ("Ben shmonah esra lechupah", at eighteen years old to the Chupah (marriage canopy)). (See Coming of age, Age of majority).
Shemoneh Esrei (sh'MOH-nuh ES-ray) is a prayer that is the center of any Jewish religious service. Its name means "eighteen". The prayer is also known as the Amidah.
In Ancient Roman custom the number 18 can symbolise a blood relative.
Joseph Heller's novel Catch-22 was originally named Catch-18 because of the Hebrew meaning of the number, but was amended to the published title to avoid confusion with another war novel, Mila 18.
There are 18 chapters in the Bhagavad Gita, which is contained in the Mahabharata |
https://en.wikipedia.org/wiki/Trans%2Ccis-2%2C6-Nonadienal | trans,cis-2,6-Nonadienal is an organic compound that is classified as a doubly unsaturated derivative of nonanal. The molecule consists of a α,β-unsaturated aldehyde with an isolated alkene group. The compound has attracted attention as the essence of cucumbers, but it is also found in bread crust and freshly cut watermelon.
Biosynthesis
Isotopic labeling has indicated that nonadienal is formed from α-linolenic acid. Such reactions are typically catalyzed by hydroperoxide lyases.
See also
2-Nonenal - another potent odorant in cucumber |
https://en.wikipedia.org/wiki/MSI%20Barcode | MSI (also known as Modified Plessey) is a barcode symbology developed by the MSI Data Corporation, based on the original Plessey Code symbology. It is a continuous symbology that is not self-checking. MSI is used primarily for inventory control, marking storage containers and shelves in warehouse environments.
Character set and binary lookup
The MSI bar code represents only digits 0–9; it does not support letters or symbols.
Each digit is converted to 4 binary-coded decimal bits. Then a 1 bit is prepended and two 0 bits are appended.
Finally, each bit is printed as a bar/space pair totalling three modules wide. A 0 bit is represented as 1/3 bar followed by 2/3 space, while a 1 bit is represented as 2/3 bar followed by 1/3 space.
Binary mapping
Each digit and guard character is represented by a binary number, as shown in the table below.
To produce a barcode image from this map, one simply must consider the digit 1 to be a black bar and the digit 0 to be a white bar and produce an image accordingly.
Check digit calculation
The MSI barcode uses one of five possible schemes for calculating a check digit:
No check digit (least common)
Mod 10 (most common)
Mod 11
Mod 1010
Mod 1110
Mod 10 Check Digit
When using the Mod 10 check digit algorithm, a string to be encoded 1234567 will be printed with a check digit of 4:
12345674
The Mod 10 check digit algorithm
uses the Luhn algorithm.
Mod 11 Check Digit
1. Reverse the string to be encoded (in this case 1234567).
Let S be the reverse of the string to be encoded
S = 7654321
2. The string is then "weighted" using a repeating weighting factor pattern. There are two modulo 11 algorithms which use different repeated weighting factor patterns: the IBM algorithm which uses (2,3,4,5,6,7), and the NCR algorithm which uses (2,3,4,5,6,7,8,9). Get the sum of the string by looping through each character and multiply it by a weight from 2 to 7 (IBM) or 2 to 9 (NCR) depending on its position. If the weight's value |
https://en.wikipedia.org/wiki/Abstract%20semantic%20graph | In computer science, an abstract semantic graph (ASG) or term graph is a form of abstract syntax in which an expression of a formal or programming language is represented by a graph whose vertices are the expression's subterms. An ASG is at a higher level of abstraction than an abstract syntax tree (or AST), which is used to express the syntactic structure of an expression or program.
ASGs are more complex and concise than ASTs because they may contain shared subterms (also known as "common subexpressions"). Abstract semantic graphs are often used as an intermediate representation by compilers to store the results of performing common subexpression elimination upon abstract syntax trees. ASTs are trees and are thus incapable of representing shared terms. ASGs are usually directed acyclic graphs (DAG), although in some applications graphs containing cycles may be permitted. For example, a graph containing a cycle might be used to represent the recursive expressions that are commonly used in functional programming languages as non-looping iteration constructs. The mutability of these types of graphs, is studied in the field of graph rewriting.
The nomenclature term graph is associated with the field of term graph rewriting, which involves the transformation and processing of expressions by the specification of rewriting rules, whereas abstract semantic graph is used when discussing linguistics, programming languages, type systems and compilation.
Abstract syntax trees are not capable of sharing subexpression nodes because it is not possible for a node in a proper tree to have more than one parent. Although this conceptual simplicity is appealing, it may come at the cost of redundant representation and, in turn, possibly inefficiently duplicating the computation of identical terms. For this reason ASGs are often used as an intermediate language at a subsequent compilation stage to abstract syntax tree construction via parsing.
An abstract semantic graph is typicall |
https://en.wikipedia.org/wiki/Asset%20%28computer%20security%29 | In information security, computer security and network security, an asset is any data, device, or other component of the environment that supports information-related activities. Assets generally include hardware (e.g. servers and switches), software (e.g. mission critical applications and support systems) and confidential information. Assets should be protected from illicit access, use, disclosure, alteration, destruction, and/or theft, resulting in loss to the organization.
The CIA triad
The goal of information security is to ensure the confidentiality, integrity and availability (CIA) of assets from various threats. For example, a hacker might attack a system in order to steal credit card numbers by exploiting a vulnerability. Information Security experts must assess the likely impact of an attack and employ appropriate countermeasures. In this case they might put up a firewall and encrypt their credit card numbers.
Risk analysis
When performing risk assessment, it is important to weigh how much to spend protecting each asset against the cost of losing the asset. It is also important to take into account the chance of each loss occurring. Intangible costs must also be factored in. If a hacker makes a copy of all a company's credit card numbers it does not cost them anything directly but the loss in fines and reputation can be enormous.
See also
Countermeasure (computer)
Factor analysis of information risk
Information security management
IT risk
Risk factor
Risk management |
https://en.wikipedia.org/wiki/Ff%20phages | Ff phages (for F specific filamentous phages) is a group of almost identical filamentous phage (genus Inovirus) including phages f1, fd, M13 and ZJ/2, which infect bacteria bearing the F fertility factor. The virion (virus particle) is a flexible filament measuring about 6 by 900 nm, comprising a cylindrical protein tube protecting a single-stranded circular DNA molecule at its core. The phage codes for only 11 gene products, and is one of the simplest viruses known. It has been widely used to study fundamental aspects of molecular biology. George Smith and Greg Winter used f1 and fd for their work on phage display for which they were awarded a share of the 2018 Nobel Prize in Chemistry. Early experiments on Ff phages used M13 to identify gene functions, and M13 was also developed as a cloning vehicle, so the name M13 is sometimes used as an informal synonym for the whole group of Ff phages.
Structure
The virion is a flexible filament (worm-like chain) about 6 nm in diameter and 900 nm long. Several thousand copies of a small (50 amino-acid residues) elongated alpha-helical major coat protein subunit (the product of gene 8, or p8) in an overlapping shingle-like array form a hollow cylinder enclosing the circular single-stranded DNA genome. Each p8 subunit has a collection of basic residues near the C-terminus of the elongated protein and acidic residues near the N-terminus; these two regions are separated by about 20 hydrophobic (non-polar) residues. The shingle-like arrangement places the acidic residues of p8 near the outside surface of the cylinder, where they cause the virus particle to be negatively-charged; non-polar regions near non-polar regions of neighbouring p8 subunits, where non-polar interactions contribute to a notable physical stability of the virus particle; and basic residues near the centre of the cylinder, where they interact with the negatively-charged DNA phosphates at the core of the virion. Longer (or shorter) DNA molecules can be package |
https://en.wikipedia.org/wiki/Lead%20star | A lead star is a low-metallicity star with an overabundance of lead and bismuth as compared to other products of the S-process.
See also
Barium star |
https://en.wikipedia.org/wiki/Concentration%20ratio | In economics, concentration ratios are used to quantify market concentration and are based on companies' market shares in a given industry.
A concentration ratio (CR) is the sum of the percentage market shares of (a pre-specified number of) the largest firms in an industry. An n-firm concentration ratio is a common measure of market structure and shows the combined market share of the n largest firms in the market. For example, if n = 5, CR5 defines the combined market share of the five largest firms in an industry.
Calculation
The concentration ratio is calculated as follows:
where defines the market share of the th largest firm in an industry as a percentage of total industry market share, and defines the number of firms included in the concentration ratio calculation.
The and concentration ratios are commonly used. Concentration ratios show the extent of largest firms' market shares in a given industry. Specifically, a concentration ratio close to 0% denotes a low concentration industry, and a concentration ratio near 100% shows that an industry has high concentration.
Concentration levels
Concentration ratios range from 0%–100%. Concentration levels are explained as follows:
Benefits and shortfalls
Concentration ratios can readily be calculated from industry data, but they are a simplistic, single parameter statistic. They can be used to quantify market concentration in a given industry in a relevant and succinct manner, but do not capture all available information about the distribution of market shares. In particular, the definition of the concentration ratio does not use the market shares of all the firms in the industry and does not account for the distribution of firm size. Also, it does not provide much detail about competitiveness of an industry.
The following example exposes the aforementioned shortfalls of the concentration ratio.
Example
The table below shows the market shares of the largest firms in two different industries (Industry |
https://en.wikipedia.org/wiki/Relocatable%20user%20backup | Relocatable user backup (RUB) is the ability to restore or relocate backups to another system or site. It is one of the primary ways to mitigate the threat of service-provider lock-in in a software as a service model. It also provides a backup in the event that the service-provider suffers a catastrophic failure.
The concept requires 3 fundamental components :
The user must be able to initiate and receive backups from the service provider and store them locally
The user must be able to restore the backup to an environment other than that provided by the service provider, preferably locally
The backup must encapsulate all content, metadata, system and application state which allow the easy restoration of the backup into a standard framework either provided and maintained by the service provider or the wider software community.
The third component also contains the implication that either the application itself is provided with the backup, or there is a consistent and ongoing application upgrade path to prevent the situation where the restore of a RUB is prevented due to an obsolete version of application software. |
https://en.wikipedia.org/wiki/Crc%20%28protein%29 | The Catabolite repression control (Crc) protein participates in suppressing expression of several genes involved in utilization of carbon sources in Pseudomonas bacteria. Presence of organic acids triggers activation of Crc and in conjunction with the Hfq protein genes that metabolize a given carbon source are downregulated until another more favorable carbon source is depleted. Crc-mediated regulation impact processes such as biofilm formation, virulence and antibiotic susceptibility.
Interactions
Hfq and Crc bind to A-rich sequences in the ribosome binding sites of genes that code for carbon utilization enzymes and consequently suppress their translation. |
https://en.wikipedia.org/wiki/Vitruvian%20Man | The Vitruvian Man (; ) is a drawing by the Italian Renaissance artist and scientist Leonardo da Vinci, dated to . Inspired by the writings of the ancient Roman architect Vitruvius, the drawing depicts a nude man in two superimposed positions with his arms and legs apart and inscribed in both a circle and square. It was described by the art historian Carmen C. Bambach as "justly ranked among the all-time iconic images of Western civilization". Although not the only known drawing of a man inspired by the writings of Vitruvius, the work is a unique synthesis of artistic and scientific ideals and often considered an archetypal representation of the High Renaissance.
The drawing represents Leonardo's conception of ideal body proportions, originally derived from Vitruvius but influenced by his own measurements, the drawings of his contemporaries, and the De pictura treatise by Leon Battista Alberti. Leonardo produced the Vitruvian Man in Milan and the work was probably passed to his student Francesco Melzi. It later came into the possession of Venanzio de Pagave, who convinced the engraver Carlo Giuseppe Gerli to include it in a book of Leonardo's drawings, which widely disseminated the previously little-known image. It was later owned by Giuseppe Bossi, who wrote early scholarship on it, and eventually sold to the Gallerie dell'Accademia of Venice in 1822, where it has remained since. Due to its sensitivity to light, the drawing rarely goes on public display, but it was borrowed by the Louvre in 2019 for their exhibition marking the 500th anniversary of Leonardo's death.
Name
The drawing is described by Leonardo's notes as , variously translated as The Proportions of the Human Figure after Vitruvius, or Proportional Study of a Man in the Manner of Vitruvius. It is much better known as the Vitruvian Man. The art historian Carlo Pedretti lists it as Homo Vitruvius, study of proportions with the human figure inscribed in a circle and a square, and later as simply Homo Vit |
https://en.wikipedia.org/wiki/Triangular%20cupola | In geometry, the triangular cupola is one of the Johnson solids (). It can be seen as half a cuboctahedron.
Formulae
The following formulae for the volume (), the surface area () and the height () can be used if all faces are regular, with edge length a:
Dual polyhedron
The dual of the triangular cupola has 6 triangular and 3 kite faces:
Related polyhedra and honeycombs
The triangular cupola can be augmented by 3 square pyramids, leaving adjacent coplanar faces. This isn't a Johnson solid because of its coplanar faces. Merging those coplanar triangles into larger ones, topologically this is another triangular cupola with isosceles trapezoidal side faces. If all the triangles are retained and the base hexagon is replaced by 6 triangles, it generates a coplanar deltahedron with 22 faces.
The triangular cupola can form a tessellation of space with square pyramids and/or octahedra, the same way octahedra and cuboctahedra can fill space.
The family of cupolae with regular polygons exists up to n=5 (pentagons), and higher if isosceles triangles are used in the cupolae. |
https://en.wikipedia.org/wiki/Index%20to%20Marine%20%26%20Lacustrine%20Geological%20Samples | The Index to Marine & Lacustrine Geological Samples is a collaboration between multiple institutions and agencies that operate geological sample repositories. The purpose of the database is to help researchers locate sea floor and lakebed cores, grabs, dredges, and drill samples in their collections.
Sample material is available from participating institutions unless noted as unavailable.
Data include basic collection and storage information. Lithology, texture, age, principal investigator, province, weathering/metamorphism, glass remarks, and descriptive comments are included for some samples. Links are provided to related data and information at the institutions and at NCEI.
Data are coded by individual institutions, several of which receive funding from the US National Science Foundation. For more information see the NSF Division of Ocean Sciences Data and Sample Policy.
The Index is endorsed by the Intergovernmental Oceanographic Commission, Committee on International Oceanographic Data and Information Exchange (IODE-XIV.2).
The index is maintained by the National Centers for Environmental Information (NCEI), formerly the National Geophysical Data Center (NGDC), and collocated World Data Center for Geophysics, Boulder, Colorado. NCEI is part of the National Environmental Satellite, Data and Information Service of the National Oceanic & Atmospheric Administration, U. S. Department of Commerce.
Searches and data downloads are available via a JSP and an ArcIMS interface. Data selections can be downloaded in tab-delimited or shapefile form, depending on the interface used. Both WMS and WFS interfaces are also available.
The Index was created in 1977 in response to a meeting of Curators of Marine Geological Samples, sponsored by the U.S. National Science Foundation. The Curators' group continues to meet every 2–3 years.
Dataset Digital Object Identifier
DOI:10.7289/V5H41PB8
Web site
The Index to Marine and Lacustrine Geological Samples
Participating Ins |
https://en.wikipedia.org/wiki/Streit%27s | Aron Streit, Inc. (sold under the name Streit's) is a kosher food company founded in Manhattan, New York City, best known for its matzo. It is the only family-owned and operated matzo company in the United States, and distributes matzo in select international markets. Streit's and its major competitor, New Jersey based Manischewitz, together hold about 40 percent of the US matzo market.
The factory follows strict kosher laws. Only shomer Shabbat (Sabbath-observing) Jews are allowed to touch the dough. However, once the dough is baked, people of any religion and non-shomer Shabbat Jews are allowed to touch the matzo. The entire process of making the matzo is under Rabbinic supervision. During Passover, Jews are not allowed to eat leavened bread, so the dough must be baked within 18 minutes before it has had time to rise. If the dough sits for longer than this, it is considered chametz – no longer kosher for Passover – and must be discarded.
History
The company was founded in 1916 by Aron Streit, a Jewish immigrant from Austria who had made matzo there in the late 1800s. Its first factory was on Pitt Street in the Lower East Side of Manhattan. There, Streit and his business associate Rabbi Weinberger made each piece of matzo by hand. In 1925, with the growing number of Jewish immigrants congregating in the Lower East Side, Streit, along with his two sons, Jack and Irving, moved his business to nearby 150 Rivington Street. Soon thereafter they bought the adjacent buildings, where the company operated for 90 years, before moving in 2015.
Lower East Side factory
Streit's matzo factory, along with Katz's Delicatessen and Yonah Schimmel's Knish Bakery, was a surviving piece of the Lower East Side's Jewish heritage. At the turn of the 20th century, Jews, along with other European immigrants, were crammed into the many unsanitary tenements of the Lower East Side. In 1915 they made up 60 percent of the Lower East Side population. Because of the large Jewish presence, Je |
https://en.wikipedia.org/wiki/Unibus | The Unibus was the earliest of several computer bus and backplane designs used with PDP-11 and early VAX systems manufactured by the Digital Equipment Corporation (DEC) of Maynard, Massachusetts. The Unibus was developed around 1969 by Gordon Bell and student Harold McFarland while at Carnegie Mellon University.
The name refers to the unified nature of the bus; Unibus was used both as a system bus allowing the central processing unit to communicate with main memory, as well as a peripheral bus, allowing peripherals to send and receive data. Unifying these formerly separate busses allowed external devices to easily perform direct memory access (DMA) and made the construction of device drivers easier as control and data exchange was all handled through memory-mapped I/O.
Unibus was physically large, which led to the introduction of Q-bus, which multiplexed some signals to reduce pin count. Higher performance PDP systems used Fastbus, essentially two Unibusses in one. The system was later supplanted by Massbus, a dedicated I/O bus introduced on the VAX and late-model PDP-11s.
Technical specifications
The Unibus consists of 72 signals, usually connected via two 36-way edge connectors on each printed circuit board. When not counting the power and ground lines, it is usually referred to as a 56-line bus. It can exist within a backplane or on a cable. Up to 20 nodes (devices) can be connected to a single Unibus segment; additional segments can be connected via a bus repeater.
The bus is completely asynchronous, allowing a mixture of fast and slow devices. It allows the overlapping of arbitration (selection of the next bus master) while the current bus master is still performing data transfers. The 18 address lines allow the addressing of a maximum of . Typically, the top is reserved for the registers of the memory-mapped I/O devices used in the PDP-11 architecture.
The design deliberately minimizes the amount of redundant logic required in the system. For example, a |
https://en.wikipedia.org/wiki/TMEM125 | Transmembrane protein 125 is a protein that, in humans, is encoded by the TMEM125 gene. It has 4 transmembrane domains and is expressed in the lungs, thyroid, pancreas, intestines, spinal cord, and brain. Though its function is currently poorly understood by the scientific community, research indicates it may be involved in colorectal and lung cancer networks. Additionally, it was identified as a cell adhesion molecule in oligodendrocytes, suggesting it may play a role in neuron myelination.
Gene
The TMEM125 gene has no aliases, except for its encoded protein’s name. Its cytogenic location is at 1p34.2 on the plus strand and it spans from bases 43,272,723 to 43,273,379. TMEM125 comprises four exons.
Gene-level regulation
Five TMEM125 promoters were identified by Genomatix Gene2Promoter. The primary promoter (NM_001320244) is 1881 bp in length. It consists of binding sites for fork head domain factors and zinc finger transcription factors.
Transcript
TMEM125 has two variant transcripts that differ only in the 5' untranslated region (UTR), but both encode the same protein. mRNA variant 1 represents the longer of the two variants and is 1898 base pairs (bp) in length; variant 2 is 1797 bp long.
Transcript-level regulation
TMEM125 microarray-assessed expression patterns in normal human tissue demonstrate the primary tissues of expression are the pancreas, lungs, salivary glands, trachea, brain, prostate, spinal cord, and thyroid. Additionally, RNA-seq data illustrates transcript expression in the following additional tissue: colon, small intestines, prostate, and stomach.
Protein
Predicted structure
TMEM125 comprises 219 amino acids with four transmembrane domains. Its predicted isoelectric point is 8.32 and predicted molecular weight is 22.1 kDa. It is primarily leucine-rich, and secondarily alanine- and glycine-rich; TMEM125 is also arginine- and lysine-deficient. It has two core repeat blocks, VALL and TTSS, which both appear twice within the protein.
Th |
https://en.wikipedia.org/wiki/Actin-binding%20protein | Actin-binding proteins (also known as ABPs) are proteins that bind to actin. This may mean ability to bind actin monomers, or polymers, or both.
Many actin-binding proteins, including α-actinin, β-spectrin, dystrophin, utrophin and fimbrin, do this through the actin-binding calponin homology domain.
This is a list of actin-binding proteins in alphabetical order.
0–9
25kDa
25kDa ABP from aorta p185neu
30akDA 110 kD dimer ABP
30bkDa 110 kD (Drebrin)
34kDA
45kDa
p53
p58gag
p116rip
A
a-actinin
Abl
ABLIM Actin-Interacting MAPKKK
ABP120
ABP140
Abp1p
ABP280 (Filamin)
ABP50 (EF-1a)
Acan 125 (Carmil)
ActA
Actibind
Actin
Actinfilin
Actinogelin
Actin-regulating kinases
Actin-Related Proteins
Actobindin
Actolinkin
Actopaxin
Actophorin
Acumentin (= L-plastin)
Adducin
ADF/Cofilin
Adseverin (scinderin)
Afadin
AFAP-110
Affixin
Aginactin
AIP1
Aldolase
Angiogenin
Anillin
Annexins
Aplyronine
Archvillin
Arginine kinase
Arp2/3 complex
B
Band 4.1
Band 4.9(Dematin)
b-actinin
b-Cap73
Bifocal
Bistramide A
BPAG1
Brevin (Gelsolin)
C
c-Abl
Calpactin (Annexin)
CHO1
Cortactin
CamKinase II
Calponin
Chondramide
Cortexillin
CAP
Caltropin
CH-ILKBP
CPb3
Cap100
Calvasculin
Ciboulot
Coactosin
CAP23
CARMIL
Acan125
Cingulin
Cytovillin (Ezrin)
CapZ/Capping Protein
a-Catenin
Cofilin
CR16
Caldesmon
CCT
Comitin
Calicin
Centuarin
Coronin
D
DBP40
Drebrin
Dematin (Band 4.9)
Dynacortin
Destrin (ADF/cofilin)
Dystonins
Diaphanous
Dystroglycan
DNase I
Dystrophin
Doliculide
Dolastatins
E
EAST
Endossin
EF-1a (ABP50)
Eps15
EF-1b
EPLIN
EF-2
Epsin
EGF receptor
ERK
ENC-1
ERM proteins (ezrin, radixin, moesin, plus merlin)
END3p
Ezrin (the E of ERM protein family)
F
F17R
Fodrin (spectrin)
Fascin
Formins
Fessilin
Frabin
FHL3
Fragmin
Fhos
FLNA (filamin A)
Fimbrin (plastin)
G
GAP43
Glycogenins
Gas2
G-proteins
Gastrin-Binding Protein
Gelactins I-IV
Gelsolins
Girdin
Glucokinase
H
Harmonin b
Hrp36
Hexokinase
Hrp65-2
Hectochlorin
HS1 (actin binding protein)
Helicase II
Hsp27
HIP1 (Huntingtin Interacting protein 1)
H |
https://en.wikipedia.org/wiki/Credit%20Reporting%20Privacy%20Code%20%28New%20Zealand%29 | The Credit Reporting Privacy Code (CRPC) was issued by the Privacy Commissioner Marie Shroff on 6 December 2004. It is one of several Codes of Practice issued by the Privacy Commissioner under section 46 of the Privacy Act.
The Code has been amended 6 times, with a 7th amendment pending, with the amendments as follows:
Amendment No 1 – 1 April 2006 (now expired)
Amendment No 2 – 1 April 2006
Amendment No 3 – 22 February 2010;
Amendment No 4 – 1 October 2011 and 1 April 2012
Amendment No 5 – 1 December 2011 and 1 April 2012
Amendment No 6 – 1 April 2012
The Code replaces the Privacy Act's 12 Privacy Principles, with 12 Privacy Rules specifically customised for credit reporting matters.
Summary of the Code
The Code regulates into law matters related to credit reporting. The Code is however only limited to credit reporters that actually sell credit information, so at current the Code only applies to 3 credit reporting firms, Veda Limited, Dun & Bradstreet Limited, and recent newcomer Centrix Group Limited.
One of the most important aspects of the Code is that individuals now have the right to a free copy of their credit record. Not only that, individuals also have a right to a copy of all their Credit Information, which includes not only includes the normal credit record, but also other things such as a copy of your credit score, access log, and even the credit reporters internal file notes.
Other things the Code covers is procedures to handle complaints, maintaining an access log to your credit record, maximum time frames to report bad debts, credit inquiries, etc., having adequate subscriber agreements, allowing certain inquiries to your credit record without a consent being required, prohibiting debt collection agencies from bad debt listing debts under their account, requiring credit suppression where a person is a victim of identity fraud, and most recently, allowing positive reporting.
The 12 Privacy Rules of the Code
Rule 1: Purpose of Collection |
https://en.wikipedia.org/wiki/Erd%C5%91s%20conjecture%20on%20arithmetic%20progressions | Erdős' conjecture on arithmetic progressions, often referred to as the Erdős–Turán conjecture, is a conjecture in arithmetic combinatorics (not to be confused with the Erdős–Turán conjecture on additive bases). It states that if the sum of the reciprocals of the members of a set A of positive integers diverges, then A contains arbitrarily long arithmetic progressions.
Formally, the conjecture states that if A is a large set in the sense that
then A contains arithmetic progressions of any given length, meaning that for every positive integer k there are an integer a and a non-zero integer c such that .
History
In 1936, Erdős and Turán made the weaker conjecture that any set of integers with positive natural density contains infinitely many 3 term arithmetic progressions. This was proven by Klaus Roth in 1952, and generalized to arbitrarily long arithmetic progressions by Szemerédi in 1975 in what is now known as Szemerédi's theorem.
In a 1976 talk titled "To the memory of my lifelong friend and collaborator Paul Turán," Paul Erdős offered a prize of US$3000 for a proof of this conjecture. As of 2008 the problem is worth US$5000.
Progress and related results
Erdős' conjecture on arithmetic progressions can be viewed as a stronger version of Szemerédi's theorem. Because the sum of the reciprocals of the primes diverges, the Green–Tao theorem on arithmetic progressions is a special case of the conjecture.
The weaker claim that A must contain infinitely many arithmetic progressions of length 3 is a consequence of an improved bound in Roth's theorem. A 2016 paper by Bloom proved that if contains no non-trivial three-term arithmetic progressions then .
In 2020 a preprint by Bloom and Sisask improved the bound to for some absolute constant .
In 2023 a preprint by Kelley and Meka gave a new bound of and four days later Bloom and Sisask simplified the result and with a little improvement to .
See also
Problems involving arithmetic progressions
List of sums |
https://en.wikipedia.org/wiki/Frequency%20allocation | Frequency allocation (or spectrum allocation or spectrum management) is the allocation and regulation of the electromagnetic spectrum into radio frequency bands, normally done by governments in most countries. Because radio propagation does not stop at national boundaries, governments have sought to harmonise the allocation of RF bands and their standardization.
ITU definition
The International Telecommunication Union defines frequency allocation as being of "a given frequency band for the purpose of its use by one or more terrestrial or space radiocommunication services or the radio astronomy service under specified conditions".
Frequency allocation is also a special term, used in national frequency administration. Other terms are:
Bodies
Several bodies set standards for frequency allocation, including:
International Telecommunication Union (ITU)
European Conference of Postal and Telecommunications Administrations (CEPT)
Inter-American Telecommunication Commission (CITEL)
To improve harmonisation in spectrum utilisation, most service allocations are incorporated in national Tables of Frequency Allocations and Utilisations within the responsibility of the appropriate national administration. Allocations are:
primary
secondary
exclusive or shared utilization, within the responsibility of national administrations.
Allocations of military usage will be in accordance with the ITU Radio Regulations. In NATO countries, military mobile utilizations are made in accordance with the NATO Joint Civil/Military Frequency Agreement (NJFA).
Examples of frequency allocations
Some of the bands listed (e.g., amateur 1.8–29.7 MHz) have gaps / are not continuous allocations.
BCB is an abbreviation for broadcast band, for commercial radio news and music broadcasts.
See also
Spectrum management
Amateur radio frequency allocations |
https://en.wikipedia.org/wiki/Japan%20Collection%20of%20Microorganisms | Japan Collection of Microorganisms (JCM) is a culture collection of microorganisms in Japan. It is a semi-governmental collection maintained by RIKEN BioResource Center and it is located in Tsukuba, Ibaraki Prefecture. JCM was initiated in 1980 with the purpose of serving as a repository of microorganisms. As of April 2012 JCM has about 20,700 microbial strains, including approximately 8,300 strains of bacteria (including actinomycetes), 370 strains of archaea and 4,800 strains of fungi including yeasts (limited to organisms classified in Risk Group 1 or 2). JCM accepts deposition of strains based on their appearance in scientific journals or accepted manuscripts.
One can search the JCM database online for microorganisms using various key words such as scientific name, JCM accession number, IAM or other culture collection accession number, culture media, strain data etc. JCM publishes its own printed catalogue once every 3 year. The first catalogue was published in 1983 and the current edition is the 10th edition published in 2007.
JCM cultures are available for a fee. The culture fee, in case of an ampoule, for commercial organizations is 10800 Japanese yen and the fee for non-profit organizations is 5400 Japanese yen.
External links
Japan Collection of Microorganisms Website (English)
Microbiology organizations
Culture collections |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.