source stringlengths 32 209 | text stringlengths 18 1.5k |
|---|---|
https://en.wikipedia.org/wiki/Selenide | A selenide is a chemical compound containing a selenium with oxidation number of −2. Similar to sulfide, selenides occur both as inorganic compounds and as organic derivatives, which are called organoselenium compound.
Inorganic selenides
The parent inorganic selenide is hydrogen selenide (H2Se). It is a colorless, malodorous, toxic gas. It dissolves in aqueous solution, to give the hydrogenselenide ion HSe−. At higher pH, selenide forms. Solutions of hydrogen selenide and selenide are oxidized by air to give elemental selenium:
Most elements form selenides. They sometimes have salt-like properties, e.g. sodium selenide, but most exhibit covalent bonding, e.g. molybdenum diselenide. Their properties are diverse, mirroring the diverse properties of the corresponding sulfides.
As indicated by the fact that only a few thousand tons of selenium are produced annually, the subset of selenium compounds called selenides find few applications. Commercially significant is zinc selenide, which is used in some infrared optics. Cadmium selenide is a pigment but its use has been declining because of environmental considerations. Copper indium selenide () has attractive potential for photovoltaic devices, but these applications have not been implemented widely. Similarly, quantum dots based on metal selenides have been extensively investigated for their distinctive spectral properties. Core-shell alloys of cadmium sulfide and selenide are of interest in imaging and phototherapy. |
https://en.wikipedia.org/wiki/Royal%20Commission%20for%20the%20Exhibition%20of%201851 | The Royal Commission for the Exhibition of 1851 is an institution founded in 1850 to administer the Great Exhibition of the Works of Industry of all Nations, which was held in The Crystal Palace, London.
The founding President of the Commission was Prince Albert of Saxe-Coburg and Gotha and its chief administrator was Henry Cole. The current President is Anne, Princess Royal.
The exhibition was a popular and financial success, with a surplus of . An unusual decision was made to maintain the Royal Commission as a permanent administrative body and to use the profits for charitable purposes. Its revised charter charged the commission with "increasing the means of industrial education, and extending the influence of science and art upon productive industry".
South Kensington
The commission invested the profits from the 1851 Exhibition in the purchase of of land in South Kensington. The area was then developed as a centre for educational and cultural institutions, often known as "Albertopolis". These include:
Imperial College
the Natural History Museum
the Royal Albert Hall
the Royal College of Art
the Royal College of Music
the Science Museum
the Victoria and Albert Museum
The commission's headquarters are in Imperial College.
Since 1891 the role of the commission has been to provide postgraduate scholarships for students to study in Britain and abroad. Former recipients of scholarships include 13 Nobel Prize laureates.
The commission currently has capital assets o |
https://en.wikipedia.org/wiki/Splash | Splash or Splash! or The Splash may refer to:
Common meanings
Splash (fluid mechanics), sudden disturbances on the surface of water
Entertainment
Splash (film), a 1984 fantasy film starring Tom Hanks and Daryl Hannah
Splash, Too, the 1988 sequel
Reality television series based on the Celebrity Splash! franchise
Splash (American TV series), an American reality series
Splash! (Chinese TV series), the official English title of a Chinese reality series
Splash (South Korean TV series), a short-lived South Korean reality series
Splash! (UK TV series), a British reality TV series
Splash, the main character in the PBS Kids show Splash and Bubbles
Music
Artists
Splash (German band)
Splash (Hungarian band)
Splash (Japanese band)
Splash (South African band)
Jack Splash, American record producer
Albums
Splash (Flow album) (2003)
Splash (Freddie Hubbard album) (1981)
Splash (Jeremy Jay album) (2009)
Splash (Satomi Fukunaga album) (1986)
Splash (Sonia & Disappear Fear album) (2009)
Splashes (album), by Archie Shepp's Quartet (1987)
Songs
"Splash!" (B'z song), 2006
"Splash" (Sub Focus song), 2010
"Splash" (Colapesce & Dimartino song), 2023
"Splash", by Can, 1974
"Splash", by Chon from Grow, 2015
"Splash", by Doug Wimbish from the album Trippy Notes for Bass, 1999
"Splash", by Gwen Stefani from the album This Is What the Truth Feels Like, 2016
"Splash", by Ho99o9 from the album United States of Horror, 2017
"Splash", by John Legend, Jhené Aiko and Ty Dolla |
https://en.wikipedia.org/wiki/TBG | TBG may refer to:
Thyroxine-binding globulin, a transport protein
TBG AG, the investment arm of the Thyssen family
Traditional Britain Group, a far-right British pressure group
Tønsberg, a city in Norway |
https://en.wikipedia.org/wiki/Linear%20discriminant%20analysis | Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events. The resulting combination may be used as a linear classifier, or, more commonly, for dimensionality reduction before later classification.
LDA is closely related to analysis of variance (ANOVA) and regression analysis, which also attempt to express one dependent variable as a linear combination of other features or measurements. However, ANOVA uses categorical independent variables and a continuous dependent variable, whereas discriminant analysis has continuous independent variables and a categorical dependent variable (i.e. the class label). Logistic regression and probit regression are more similar to LDA than ANOVA is, as they also explain a categorical variable by the values of continuous independent variables. These other methods are preferable in applications where it is not reasonable to assume that the independent variables are normally distributed, which is a fundamental assumption of the LDA method.
LDA is also closely related to principal component analysis (PCA) and factor analysis in that they both look for linear combinations of variables which best explain the data. LDA explicitly attempts to model the difference between the classes of data. PC |
https://en.wikipedia.org/wiki/Linear%20complementarity%20problem | In mathematical optimization theory, the linear complementarity problem (LCP) arises frequently in computational mechanics and encompasses the well-known quadratic programming as a special case. It was proposed by Cottle and Dantzig in 1968.
Formulation
Given a real matrix M and vector q, the linear complementarity problem LCP(q, M) seeks vectors z and w which satisfy the following constraints:
(that is, each component of these two vectors is non-negative)
or equivalently This is the complementarity condition, since it implies that, for all , at most one of and can be positive.
A sufficient condition for existence and uniqueness of a solution to this problem is that M be symmetric positive-definite. If M is such that has a solution for every q, then M is a Q-matrix. If M is such that have a unique solution for every q, then M is a P-matrix. Both of these characterizations are sufficient and necessary.
The vector w is a slack variable, and so is generally discarded after z is found. As such, the problem can also be formulated as:
(the complementarity condition)
Convex quadratic-minimization: Minimum conditions
Finding a solution to the linear complementarity problem is associated with minimizing the quadratic function
subject to the constraints
These constraints ensure that f is always non-negative. The minimum of f is 0 at z if and only if z solves the linear complementarity problem.
If M is positive definite, any algorithm for solving (stri |
https://en.wikipedia.org/wiki/2-4-2 | Under the Whyte notation for the classification of steam locomotives, represents the wheel arrangement of two leading wheels on one axle, four powered and coupled driving wheels on two axles and two trailing wheels on one axle. The type is sometimes named Columbia after a Baldwin locomotive was showcased at the 1893 World's Columbian Exposition held at Chicago, Illinois.
Overview
The wheel arrangement was widely used on passenger tank locomotives during the last three decades of the nineteenth and the first decade of the twentieth centuries. The vast majority of 2-4-2 locomotives were tank engines, designated 2-4-2T. The symmetrical wheel arrangement was well suited for a tank locomotive that is used to work in either direction.
When the leading and trailing wheels are in swivelling trucks, the equivalent UIC classification is 1'B1'.
While a number of 2-4-2 tender locomotives were built, larger tender locomotive types soon became dominant.
Usage
Cape of Good Hope
In 1899, the Walvis Bay Railway in the British territory of Walvis Bay, a Cape of Good Hope exclave in Deutsch-Südwest-Afrika (German South West Africa), placed a single tank locomotive in service. The engine, named Hope and built by Kerr, Stuart and Company, remained in service until 1904 when operations on the railway were suspended. The line was abandoned in 1905, partly as a result of being buried by a sandstorm.
Finland
A 2-4-2 tank locomotive, built by Baldwin Locomotive Works in 1899 and used on the p |
https://en.wikipedia.org/wiki/Bismuthinite | Bismuthinite is a mineral consisting of bismuth sulfide (Bi2S3). It is an important ore for bismuth. The crystals are steel-grey to off-white with a metallic luster. It is soft enough to be scratched with a fingernail and rather dense.
Bismuthinite forms a series with the lead, copper, bismuth mineral aikinite (PbCuBiS3).
It occurs in hydrothermal veins with tourmaline-bearing copper veins associated with granite, in some high temperature gold veins, and in recent volcanic exhalation deposits. Associated minerals include native bismuth, aikinite, arsenopyrite, stannite, galena, pyrite, chalcopyrite, tourmaline, wolframite, cassiterite and quartz.
It was first reported in 1832 from the mines of Potosí, Bolivia.
References
Further reading
Bismuth minerals
Sulfide minerals
Orthorhombic minerals
Minerals in space group 62 |
https://en.wikipedia.org/wiki/Dawsonite | Dawsonite is a mineral composed of sodium aluminium carbonate hydroxide, chemical formula NaAlCO3(OH)2. It crystallizes in the orthorhombic crystal system. It is not mined for ore. It was discovered in 1874 during the construction of the Redpath Museum in a feldspathic dike on the campus of McGill University on the Island of Montreal, Canada. It is named after geologist Sir John William Dawson (1820–1899).
The type material is preserved in the collection of the Redpath Museum.
See also
List of minerals
List of minerals named after people
Dihydroxialumini sodium carbonate, the commercial (artificial) form, used as an antacid
References
Sodium minerals
Aluminium minerals
Carbonate minerals
Orthorhombic minerals
Minerals in space group 74
Luminescent minerals |
https://en.wikipedia.org/wiki/Rad%C3%B3%27s%20theorem%20%28harmonic%20functions%29 | See also Rado's theorem (Ramsey theory)
In mathematics, Radó's theorem is a result about harmonic functions, named after Tibor Radó. Informally, it says that any "nice looking" shape without holes can be smoothly deformed into a disk.
Suppose Ω is an open, connected and convex subset of the Euclidean space R2 with smooth boundary ∂Ω and suppose that D is the unit disk. Then, given any homeomorphism
μ : ∂D → ∂Ω, there exists a unique harmonic function u : D → Ω such that u = μ on ∂D and u is a diffeomorphism.
References
R. Schoen, S. T. Yau. (1997) Lectures on Harmonic Maps. International Press, Inc., Boston, Massachusetts. , page 4.
Theorems in harmonic analysis |
https://en.wikipedia.org/wiki/Metabolic%20ecology | Metabolic ecology is a field of ecology aiming to understand constraints on metabolic organization as important for understanding almost all life processes. Main focus is on the metabolism of individuals, emerging intra- and inter-specific patterns, and the evolutionary perspective.
Two main metabolic theories that have been applied in ecology are Kooijman's Dynamic energy budget (DEB) theory and the West, Brown, and Enquist (WBE) theory of ecology. Both theories have an individual-based metabolic underpinning, but have fundamentally different assumptions.
Models of individual's metabolism follow the energy uptake and allocation, and can focus on mechanisms and constraints of energy transport (transport models), or on dynamic use of stored metabolites (energy budget models).
References
Ecology
Metabolism |
https://en.wikipedia.org/wiki/Ovalbumin | Ovalbumin (abbreviated OVA) is the main protein found in egg white, making up approximately 55% of the total protein. Ovalbumin displays sequence and three-dimensional homology to the serpin superfamily, but unlike most serpins it is not a serine protease inhibitor. The function of ovalbumin is unknown, although it is presumed to be a storage protein.
Research
Ovalbumin is an important protein in several different areas of research, including:
general studies of protein structure and properties (because it is available in large quantities).
studies of serpin structure and function (the fact that ovalbumin does not inhibit proteases means that by comparing its structure with that of inhibitory serpins, the structural characteristics required for inhibition can be determined).
proteomics (chicken egg ovalbumin is commonly used as a molecular weight marker for calibrating electrophoresis gels).
immunology (commonly used to stimulate an allergic reaction in test subjects; e.g., established model allergen for airway hyper-responsiveness, AHR).
(For in vivo and in vitro studies based on ovalbumin it is important that the endotoxin content is less than 1 EU/mg.)
Structure
The ovalbumin protein of chickens consists of 385 amino acids, its relative molecular mass is 42.7kDa, and it adopts a serpin-like structure. Ovalbumin also has several modifications, including N-terminal acetylation (G1), phosphorylation (S68, S344), and glycosylation (N292). It has three isoforms, A1, A2 |
https://en.wikipedia.org/wiki/J%20Strother%20Moore | J Strother Moore (his first name is the alphabetic character "J" – not an abbreviated "J.") is a computer scientist. He is a co-developer of the Boyer–Moore string-search algorithm, Boyer–Moore majority vote algorithm, and the Boyer–Moore automated theorem prover, Nqthm. He made pioneering contributions to structure sharing including the piece table data structure and early logic programming. An example of the workings of the Boyer–Moore string search algorithm is given in Moore's website. Moore received his Bachelor of Science (BS) in mathematics at Massachusetts Institute of Technology in 1970 and his Doctor of Philosophy (Ph.D.) in computational logic at the University of Edinburgh in Scotland in 1973.
In addition, Moore is a co-author of the ACL2 automated theorem prover and its predecessors including Nqthm, for which he received, with Robert S. Boyer and Matt Kaufmann, the 2005 ACM Software System Award. He and others used ACL2 to prove the correctness of the floating point division operations of the AMD K5 microprocessor in the wake of the Pentium FDIV bug.
For his contributions to automated deduction, Moore received the 1999 Herbrand Award with Robert S. Boyer, and in 2006 he was inducted as a Fellow in the Association for Computing Machinery. Moore was elected a member of the National Academy of Engineering in 2007 for contributions to automated reasoning about computing systems. He is also a Fellow of the AAAI. He was elected a Corresponding Fellow of the Royal Soc |
https://en.wikipedia.org/wiki/Reference%20electrode | A reference electrode is an electrode that has a stable and well-known electrode potential. The overall chemical reaction taking place in a cell is made up of two independent half-reactions, which describe chemical changes at the two electrodes. To focus on the reaction at the working electrode, the reference electrode is standardized with constant (buffered or saturated) concentrations of each participant of the redox reaction.
There are many ways reference electrodes are used. The simplest is when the reference electrode is used as a half-cell to build an electrochemical cell. This allows the potential of the other half cell to be determined. An accurate and practical method to measure an electrode's potential in isolation (absolute electrode potential) has yet to be developed.
Aqueous reference electrodes
Common reference electrodes and potential with respect to the standard hydrogen electrode (SHE):
Standard hydrogen electrode (SHE) (E = 0.000 V) activity of H+ = 1 Molar
Normal hydrogen electrode (NHE) (E ≈ 0.000 V) concentration H+ = 1 Molar
Reversible hydrogen electrode (RHE) (E = 0.000 V - 0.0591 × pH) at 25°C
Saturated calomel electrode (SCE) (E = +0.241 V saturated)
Copper-copper(II) sulfate electrode (CSE) (E = +0.314 V)
Silver chloride electrode (E = +0.197 V in saturated KCl)
Silver chloride electrode (E = +0.210 V in 3.0 mol KCl/kg)
Silver chloride electrode (E = +0.22249 V in 3.0 mol KCl/L)
pH-electrode (in case of pH buffered solutions, see buffer |
https://en.wikipedia.org/wiki/Afrikan%20t%C3%A4hti | Afrikan tähti (; Finnish for 'Star of Africa'), known in Swedish as Den försvunna diamanten ('The Missing Diamond') or Afrikas stjärna ('The Star of Africa'), is a Finnish board game designed by Kari Mannerla originally in 1951. It has been one of the most popular board games in the Nordic countries for decades.
History
The board game was first published year 1951 in Helsinki, Finland. Stories of the world's largest diamond – the Star of Africa – had inspired the imagination of young Kari Mannerla. He managed to get a hold of a map of Africa in the English language and picked exotic sounding places. He then drafted land, sea and air routes arbitrarily across and around the continent. During his design process, an important innovation was that players could pick their route of choice, instead of following a pre-set course. A further revelation was the tokens that are shuffled before each game and placed randomly throughout the board. None of the players know which surprise is hidden under each token. The robbers and horseshoes were considered to add to the excitement of the game.
Afrikan tähti was the last of Kari Mannerla's games. He was then already an ‘experienced’ 19-year-old game designer and he intuitively sensed that this one was unique. He offered it to the large book publishing company Tilgmann. Price negotiations took a few years and finally he accepted a modest compensation for 10,000 printed games. His condition to the publisher was that any further prints would |
https://en.wikipedia.org/wiki/Synapsin | The synapsins are a family of proteins that have long been implicated in the regulation of neurotransmitter release at synapses. Specifically, they are thought to be involved in regulating the number of synaptic vesicles available for release via exocytosis at any one time. Synapsins are present in invertebrates and vertebrates and are strongly conserved across all species. They are expressed in highest concentration in the nervous system, although they also express in other body systems such as the reproductive organs, including both eggs and spermatozoa. Synapsin function also increases as the organism matures, reaching its peak at sexual maturity.
Current studies suggest the following hypothesis for the role of synapsin: synapsins bind synaptic vesicles to components of the cytoskeleton which prevents them from migrating to the presynaptic membrane and releasing neurotransmitter. During an action potential, synapsins are phosphorylated by PKA (cAMP dependent protein kinase), releasing the synaptic vesicles and allowing them to move to the membrane and release their neurotransmitter.
Gene knockout studies in mice (where the mouse is unable to produce synapsin) have had some surprising results. Consistently, knockout studies have shown that mice lacking one or more synapsins have defects in synaptic transmission induced by high‐frequency stimulation, suggesting that the synapsins may be one of the factors boosting release probability in synapses at high firing rates, such |
https://en.wikipedia.org/wiki/Context%20mixing | Context mixing is a type of data compression algorithm in which the next-symbol predictions of two or more statistical models are combined to yield a prediction that is often more accurate than any of the individual predictions. For example, one simple method (not necessarily the best) is to average the probabilities assigned by each model. The random forest is another method: it outputs the prediction that is the mode of the predictions output by individual models. Combining models is an active area of research in machine learning.
The PAQ series of data compression programs use context mixing to assign probabilities to individual bits of the input.
Application to Data Compression
Suppose that we are given two conditional probabilities, and , and we wish to estimate , the probability of event X given both conditions and . There is insufficient information for probability theory to give a result. In fact, it is possible to construct scenarios in which the result could be anything at all. But intuitively, we would expect the result to be some kind of average of the two.
The problem is important for data compression. In this application, and are contexts, is the event that the next bit or symbol of the data to be compressed has a particular value, and and are the probability estimates by two independent models. The compression ratio depends on how closely the estimated probability approaches the true but unknown probability of event . It is often the case that con |
https://en.wikipedia.org/wiki/Sex-determining%20region%20Y%20protein | Sex-determining region Y protein (SRY), or testis-determining factor (TDF), is a DNA-binding protein (also known as gene-regulatory protein/transcription factor) encoded by the SRY gene that is responsible for the initiation of male sex determination in therian mammals (placental mammals and marsupials). SRY is an intronless sex-determining gene on the Y chromosome. Mutations in this gene lead to a range of disorders of sex development with varying effects on an individual's phenotype and genotype.
SRY is a member of the SOX (SRY-like box) gene family of DNA-binding proteins. When complexed with the (SF-1) protein, SRY acts as a transcription factor that causes upregulation of other transcription factors, most importantly SOX9. Its expression causes the development of primary sex cords, which later develop into seminiferous tubules. These cords form in the central part of the yet-undifferentiated gonad, turning it into a testis. The now-induced Leydig cells of the testis then start secreting testosterone, while the Sertoli cells produce anti-Müllerian hormone. SRY gene effects normally take place 6–8 weeks after fetus formation which inhibits the female anatomical structural growth in males. It also works towards developing the secondary sexual characteristics of males.
Gene evolution and regulation
Evolution
SRY may have arisen from a gene duplication of the X chromosome bound gene SOX3, a member of the SOX family. This duplication occurred after the split between monotr |
https://en.wikipedia.org/wiki/Cerna%2C%20Croatia | Cerna () is a village and a municipality in eastern Croatia.
Geography
It is located half-way between the cities of Vinkovci and Županja. It is located on four rivers, Biđ, Bosut, Berava, Bitulja, Krajc channel and Kaluđer channel.
Population
According to 2011 Croatian census it has 4,616 inhabitants in 2 settlements:
Cerna - 3,791
Šiškovci - 804
98.96% of the citizens are Croats.
History
One Scordisci archaeological site in Cerna dating back to late La Tène culture was excavated in the 1970s and 1980s as a part of rescue excavations in eastern Croatia. Archaeological site was a part of the settlement network of Scordisci in the area of Vinkovci.
See also
Spačva basin
References
Municipalities of Croatia
Populated places in Syrmia
Populated places in Vukovar-Syrmia County
Archaeological sites in Croatia
La Tène culture |
https://en.wikipedia.org/wiki/Snowball%20sampling | In sociology and statistics research, snowball sampling (or chain sampling, chain-referral sampling, referral sampling) is a nonprobability sampling technique where existing study subjects recruit future subjects from among their acquaintances. Thus the sample group is said to grow like a rolling snowball. As the sample builds up, enough data are gathered to be useful for research. This sampling technique is often used in hidden populations, such as drug users or sex workers, which are difficult for researchers to access.
As sample members are not selected from a sampling frame, snowball samples are subject to numerous biases. For example, people who have many friends are more likely to be recruited into the sample. When virtual social networks are used, then this technique is called virtual snowball sampling.
It was widely believed that it was impossible to make unbiased estimates from snowball samples, but a variation of snowball sampling called respondent-driven sampling
has been shown to allow researchers to make asymptotically unbiased estimates from snowball samples under certain conditions. Snowball sampling and respondent-driven sampling also allows researchers to make estimates about the social network connecting the hidden population.
Description
Snowball sampling uses a small pool of initial informants to nominate, through their social networks, other participants who meet the eligibility criteria and could potentially contribute to a specific study. The term |
https://en.wikipedia.org/wiki/UK%20Singles%20Chart%20records%20and%20statistics | The UK Singles Chart was first compiled in 1969. However the records and statistics listed here date back to 1952 because the Official Charts Company counts a selected period of the New Musical Express chart (only from 1952 to 1960) and the Record Retailer chart from 1960 to 1969 as predecessors for the period prior to 11 February 1969, where multiples of competing charts coexisted side by side. For example, the BBC compiled its own chart based on an average of the music papers of the time; many songs announced as having reached number one on BBC Radio and Top of the Pops prior to 1969 may not be listed here as chart-toppers since they do not meet the legacy criteria of the Charts Company.
Number one hits
Most number ones
The following is a list of all the acts who are on eight or more UK number one songs with an individual credit (meaning, the main artist or named separately as a featured artist – being part of a group does not count towards an individual's total).
Simply playing or singing on a single without credit will not count, or the top positions would almost certainly belong to session musicians such as Clem Cattini who is reported to have played drums on over 40 number ones.
Most weeks at number one by artist
Most weeks at number one by single
The record for most non-consecutive weeks at number one is 18 by Frankie Laine's "I Believe" in 1953. It spent nine weeks at number one, dropped down for a week, returned to number one for six weeks, dropped down for a f |
https://en.wikipedia.org/wiki/Hunter%20Rouse | Hunter Rouse (March 29, 1906 – October 16, 1996) was a hydraulician known for his research on the mechanics of fluid turbulence.
Rouse was a faculty member at the Massachusetts Institute of Technology, Cambridge, from 1929 until 1933, when he moved to Columbia University. He was at the California Institute of Technology, Pasadena (1936–1939), and in 1939 he joined the staff of the University of Iowa, Iowa City, where he was dean of the college of engineering from 1966 to 1972. His work includes hydraulic studies of similitude, efflux and overflow, jet diffusion, boundary roughness, and sediment suspension.
Among his written works are Fluid Mechanics for Hydraulic Engineers (1938), Elementary Mechanics of Fluids (1946), Basic Mechanics of Fluids (1953), and History of Hydraulics (1957).
Hunter's eldest son Richard Rouse is one of the world's foremost specialists in Western European History in the Middle Ages.
References
External links
1906 births
1996 deaths
20th-century American physicists
Rouse, Hunter
Massachusetts Institute of Technology faculty
Columbia University faculty
California Institute of Technology faculty
University of Iowa faculty |
https://en.wikipedia.org/wiki/Flagship%20species | In conservation biology, a flagship species is a species chosen to raise support for biodiversity conservation in a given place or social context. Definitions have varied, but they have tended to focus on the strategic goals and the socio-economic nature of the concept, to support the marketing of a conservation effort. The species need to be popular, to work as symbols or icons, and to stimulate people to provide money or support.
Species selected since the idea was developed in 1980s include widely recognised and charismatic species like the black rhinoceros, the Bengal tiger, and the Asian elephant. Some species such as the Chesapeake blue crab and the Pemba flying fox, the former of which is locally significant to Northern America, have suited a cultural and social context.
Utilizing a flagship species has limitations. It can skew management and conservation priorities, which may conflict. Stakeholders may be negatively affected if the flagship species is lost. The use of a flagship may have limited effect, and the approach may not protect the species from extinction: all of the top ten charismatic groups of animal, including tigers, lions, elephants and giraffes, are endangered.
Definitions
The term flagship is linked to the metaphor of representation. In its popular usage, flagships are viewed as ambassadors or icons for a conservation project or movement. The geographer Maan Barua noted that metaphors influence what people understand and how they act; that mammal |
https://en.wikipedia.org/wiki/External%20sorting | External sorting is a class of sorting algorithms that can handle massive amounts of data. External sorting is required when the data being sorted do not fit into the main memory of a computing device (usually RAM) and instead they must reside in the slower external memory, usually a disk drive. Thus, external sorting algorithms are external memory algorithms and thus applicable in the external memory model of computation.
External sorting algorithms generally fall into two types, distribution sorting, which resembles quicksort, and external merge sort, which resembles merge sort. External merge sort typically uses a hybrid sort-merge strategy. In the sorting phase, chunks of data small enough to fit in main memory are read, sorted, and written out to a temporary file. In the merge phase, the sorted subfiles are combined into a single larger file.
Model
External sorting algorithms can be analyzed in the external memory model. In this model, a cache or internal memory of size and an unbounded external memory are divided into blocks of size , and the running time of an algorithm is determined by the number of memory transfers between internal and external memory. Like their cache-oblivious counterparts, asymptotically optimal external sorting algorithms achieve a running time (in Big O notation) of .
External merge sort
One example of external sorting is the external merge sort algorithm, which is a K-way merge algorithm. It sorts chunks that each fit in RAM, then merg |
https://en.wikipedia.org/wiki/XHFW-FM | XHFW-FM/XEFW-AM is a radio station in Tampico, operating on 88.5 FM. It's owned and operated by Multimedios Radio. 810 AM is a United States clear-channel frequency; KGO and WGY share Class A status on this frequency.
History
XEFW-AM 1240 received its concession in 1932. The 70-watt station owned by José Expedito Martínez until 1937, when the Flores family entered broadcasting by way of a joint partnership with Martínez. XEFW later moved to 810 kHz and ramped up its power to 50,000 watts day and night, earning it the distinction of being a border blaster in the eyes of some.
In the 1980s, XEFW cut its power back to 10,000 watts day and a mere 250 watts at night, later boosted back to 50,000/1,000. The station was now exclusively owned by the Flores family, which had built a local cluster featuring XEFW, AM sister XETW-AM 860, XHON-FM 96.1, and XHFW-TV channel 9.
In 2011, the Flores family took advantage of the government's program to migrate AM stations to FM and converted XEFW into XHFW-FM broadcasting on 106.3 MHz. XHFW was known as La Estrella.
In 2016, the Flores family ceded operation of its radio stations to Multimedios Radio, which converted them to their own formats. XHFW was given the Hits FM pop format. On April 26, 2018, XHFW-FM moved to 88.5 MHz in order to clear 106-108 MHz for community and indigenous stations, as a condition of the renewal of its concession in 2017.
On April 25, 2018, the Federal Telecommunications Institute approved the transfer of XHFW-F |
https://en.wikipedia.org/wiki/2-4-4T | Under the Whyte notation for the classification of steam locomotives by wheel arrangement, 2-4-4 is a steam locomotive with two unpowered leading wheels followed by four powered driving wheels and four unpowered trailing wheels. The wheel arrangement was only used for tank locomotives.
Equivalent classifications
Other equivalent classifications are:
UIC classification: 1B2 (also known as German classification and Italian classification)
French classification: 122
Turkish classification: 25
Swiss classification: 2/5
The equivalent UIC classification is 1′B2′ t (or (1′B)2′ t for a Mason Bogie).
Examples
This unusual wheel arrangement does not appear to have been used on the mainline railways in the UK. It was however one of the configurations used on the Mason Bogie articulated locomotives, in the USA during the 1870s and 1880s. Five examples were constructed at the Mason Machine Works for the narrow gauge Boston, Revere Beach and Lynn Railroad 1883–1887. The railway subsequently received twenty-one further examples between 1900 and 1914, constructed by the Taunton Locomotive Manufacturing Company, Manchester Locomotive Works, and ALCO. Developmentally, there are two logical ways of reaching this wheel formula: to add a forward axle to a Forney locomotive to improve its ability to negotiate curves, or to add a second trailing axle to a Columbia design, notably in a 2-4-4(T) configuration to expand its coal capacity.
Four 2-4-4T passenger locomotives were built by the Czec |
https://en.wikipedia.org/wiki/Hector%20Ruiz | Hector de Jesus Ruiz Cardenas (born December 25, 1945) is the chairman and CEO of Advanced Nanotechnology Solutions, Inc. and former CEO & executive chairman of semiconductor company Advanced Micro Devices, Inc. (AMD).
Ruiz is the author of Slingshot: AMD’s Fight To Free An Industry From The Ruthless Grip Of Intel, "a book that memorializes his bet-the-company decision in 2005 to file an antitrust case against its much larger rival." The book also "elaborates on his humble upbringing as well as advice and lessons learned from relatives and teachers."
Education
Ruiz was born in the border town of Piedras Negras, Coahuila, Mexico. As a teenager, he walked across the Mexico – United States border every day to attend a high school in nearby Eagle Pass, Texas, from which he graduated as valedictorian just three years after beginning to learn English. Ruiz earned a BS and MS in electrical engineering from the University of Texas at Austin in 1968 and 1970 respectively and a PhD from Rice University in 1973.
In 2012, Ruiz—along with former First Lady Laura Bush, Charles Matthews, Melinda Perrin, Julius Glickman and Admiral William H. McRaven—was named a Distinguished Alumnus of the University of Texas.
Career
Ruiz worked at Texas Instruments for six years and Motorola for 22 years, rising to become president of Motorola's Semiconductor Products Sector before being recruited in 2000 by AMD founder Jerry Sanders to serve as AMD's president and chief operating officer, and to beco |
https://en.wikipedia.org/wiki/Erna%20Hennicot-Schoepges | Erna Hennicot-Schoepges (born 24 July 1941 in Dudelange) is a Luxembourgish politician for the Christian Social People's Party. She was until 2009 a Member of the European Parliament, sitting as a CSV member of the European People's Party.
Erna Hennicot-Schoepges has an extensive history in cultural affairs. She was educated in Musical studies, for which she was awarded the gold medal as a pianist, as well as philosophy and literature at the Royal Conservatoire of Brussels in Brussels, Ecole Normale in Paris, the Mozarteum in Salzburg, and the Centre Universitaire in Luxembourg.
Her political career reflects this deep interest in culture as well as her commitment to civil affairs. The Hennicot-Schoepges was elected Mayor of Walferdange and then moved on to become the first woman President of Luxembourg’s Parliament followed by her appointment as Luxembourg’s Minister of Culture, Higher Education and Research and Public Works. Throughout this time, she was also highly engaged as a member of the Parliamentary Assembly of the Council of Europe, and of the European Parliament.
Hennicot-Schoepges has made many political contributions to the cultural field. On the national level, she is responsible for the creation of the University of Luxembourg in 2003 and the Luxembourg Philharmonic Hall Joséphine Charlotte in 2005, among other projects. On a European level, she was the rapporteur to the EP of the “European Year of Intercultural Dialogue, 2008.” She is also a member of the |
https://en.wikipedia.org/wiki/Menelaus%27s%20theorem | In Euclidean geometry, Menelaus's theorem, named for Menelaus of Alexandria, is a proposition about triangles in plane geometry. Suppose we have a triangle , and a transversal line that crosses at points respectively, with distinct from . A weak version of the theorem states that
where "| |" denotes absolute value (i.e., all segment lengths are positive).
The theorem can be strengthened to a statement about signed lengths of segments, which provides some additional information about the relative order of collinear points. Here, the length is taken to be positive or negative according to whether is to the left or right of in some fixed orientation of the line; for example, is defined as having positive value when is between and and negative otherwise. The signed version of Menelaus's theorem states
Equivalently,
Some authors organize the factors differently and obtain the seemingly different relation
but as each of these factors is the negative of the corresponding factor above, the relation is seen to be the same.
The converse is also true: If points are chosen on respectively so that
then are collinear. The converse is often included as part of the theorem. (Note that the converse of the weaker, unsigned statement is not necessarily true.)
The theorem is very similar to Ceva's theorem in that their equations differ only in sign. By re-writing each in terms of cross-ratios, the two theorems may be seen as projective duals.
Proofs
A standard proof
Fir |
https://en.wikipedia.org/wiki/Anisole | Anisole, or methoxybenzene, is an organic compound with the formula . It is a colorless liquid with a smell reminiscent of anise seed, and in fact many of its derivatives are found in natural and artificial fragrances. The compound is mainly made synthetically and is a precursor to other synthetic compounds. Structurally, it is an ether () with a methyl () and phenyl () group attached. Anisole is a standard reagent of both practical and pedagogical value.
It can be prepared by the Williamson ether synthesis; sodium phenoxide is reacted with a methyl halide to yield anisole.
Reactivity
Anisole undergoes electrophilic aromatic substitution reaction at a faster speed than benzene, which in turn reacts more quickly than nitrobenzene. The methoxy group is an ortho/para directing group, which means that electrophilic substitution preferentially occurs at these three sites. The enhanced nucleophilicity of anisole vs. benzene reflects the influence of the methoxy group, which renders the ring more electron-rich. The methoxy group strongly affects the pi cloud of the ring as a mesomeric electron donor, more so than as an inductive electron withdrawing group despite the electronegativity of the oxygen. Stated more quantitatively, the Hammett constant for para-substitution of anisole is –0.27.
Illustrative of its nucleophilicity, anisole reacts with acetic anhydride to give
Unlike most acetophenones, but reflecting the influence of the methoxy group, methoxyacetophenone underg |
https://en.wikipedia.org/wiki/Neck%20dissection | The neck dissection is a surgical procedure for control of neck lymph node metastasis from squamous cell carcinoma (SCC) of the head and neck. The aim of the procedure is to remove lymph nodes from one side of the neck into which cancer cells may have migrated. Metastasis of squamous cell carcinoma into the lymph nodes of the neck reduce survival and is the most important factor in the spread of the disease. The metastases may originate from SCC of the upper aerodigestive tract, including the oral cavity, tongue, nasopharynx, oropharynx, hypopharynx, and larynx, as well as the thyroid, parotid and posterior scalp.
History of neck dissections
1906 – George W. Crile of the Cleveland Clinic describes the radical neck dissection. The operation encompasses removal of all the lymph nodes on one side of the neck, and includes removal of the spinal accessory nerve (SAN), internal jugular vein (IJV) and sternocleidomastoid muscle (SCM).
1957 – Hayes Martin describes routine use of the radical neck dissection for control of neck metastases.
1967 – Oscar Suarez and E. Bocca describe a more conservative operation that preserves SAN, IJV and SCM.
Last 3 decades – Further operations have been described to selectively remove the involved regional lymph groups.
Division of the neck into levels and sublevels
To describe the lymph nodes of the neck for neck dissection, the neck is divided into 6 areas called Levels. The levels are identified by Roman numeral, increasing towards the che |
https://en.wikipedia.org/wiki/Marginal%20value%20theorem | The marginal value theorem (MVT) is an optimality model that usually describes the behavior of an optimally foraging individual in a system where resources (often food) are located in discrete patches separated by areas with no resources. Due to the resource-free space, animals must spend time traveling between patches. The MVT can also be applied to other situations in which organisms face diminishing returns.
The MVT was first proposed by Eric Charnov in 1976. In his original formulation: "The predator should leave the patch it is presently in when the marginal capture rate in the patch drops to the average capture rate for the habitat."
Definition
All animals must forage for food in order to meet their energetic needs, but doing so is energetically costly. It is assumed that evolution by natural selection results in animals utilizing the most economic and efficient strategy to balance energy gain and consumption. The Marginal Value Theorem is an optimality model that describes the strategy that maximizes gain per unit time in systems where resources, and thus rate of returns, decrease with time. The model weighs benefits and costs and is used to predict giving up time and giving up density. Giving up time (GUT) is the interval of time between when the animal last feeds and when it leaves the patch. Giving up density (GUD) is the food density within a patch when the animal will choose to move on to other food patches.
When an animal is foraging in a system where food so |
https://en.wikipedia.org/wiki/PFD |
Science, technology, and medicine
Personal flotation device
Pelvic floor dysfunction
Phase frequency detector in electronics
Primary flight display, in an aircraft
Probability of Failure on Demand, see Safety integrity level#Certification
Process flow diagram, in process engineering
Prepared for dyeing
Professional Disc, recordable optical disc format
PFD allowance in work systems
Partial fraction decomposition
Perfluorodecalin, a molecule capable of dissolving large amounts of gas
Pediatric Feeding Disorder, A unifying diagnostic term, “pediatric feeding disorder” encompassing medical, nutrition, feeding skill, and psychosocial domains
Organizations
Philadelphia Fire Department
Pigespejdernes Fællesråd Danmark, Guiding federation of Denmark
Peters, Fraser & Dunlop, an English literary and talent agency
Other uses
Permanent Fund Dividend of Alaska Permanent Fund
See also
PDF (disambiguation) |
https://en.wikipedia.org/wiki/14-3-3%20protein | 14-3-3 proteins are a family of conserved regulatory molecules that are expressed in all eukaryotic cells. 14-3-3 proteins have the ability to bind a multitude of functionally diverse signaling proteins, including kinases, phosphatases, and transmembrane receptors. More than 200 signaling proteins have been reported as 14-3-3 ligands.
Elevated amounts of 14-3-3 proteins in cerebrospinal fluid may be a sign of Creutzfeldt–Jakob disease.
Properties
Seven genes encode seven distinct 14-3-3 proteins in most mammals (See Human genes below) and 13-15 genes in many higher plants, though typically in fungi they are present only in pairs. Protists have at least one. Eukaryotes can tolerate the loss of a single 14-3-3 gene if multiple genes are expressed, but deletion of all 14-3-3s (as experimentally determined in yeast) results in death.
14-3-3 proteins are structurally similar to the Tetratrico Peptide Repeat (TPR) superfamily, which generally have 9 or 10 alpha helices, and usually form homo- and/or hetero-dimer interactions along their amino-termini helices. These proteins contain a number of known common modification domains, including regions for divalent cation interaction, phosphorylation & acetylation, and proteolytic cleavage, among others established and predicted.
14-3-3 binds to peptides. There are common recognition motifs for 14-3-3 proteins that contain a phosphorylated serine or threonine residue, although binding to non-phosphorylated ligands has also been repor |
https://en.wikipedia.org/wiki/Bekenstein%20bound | In physics, the Bekenstein bound (named after Jacob Bekenstein) is an upper limit on the thermodynamic entropy S, or Shannon entropy H, that can be contained within a given finite region of space which has a finite amount of energy—or conversely, the maximal amount of information required to perfectly describe a given physical system down to the quantum level. It implies that the information of a physical system, or the information necessary to perfectly describe that system, must be finite if the region of space and the energy are finite. In computer science this implies that non-finite models such as Turing machines are not realizable as finite devices.
Equations
The universal form of the bound was originally found by Jacob Bekenstein in 1981 as the inequality
where S is the entropy, k is the Boltzmann constant, R is the radius of a sphere that can enclose the given system, E is the total mass–energy including any rest masses, ħ is the reduced Planck constant, and c is the speed of light. Note that while gravity plays a significant role in its enforcement, the expression for the bound does not contain the gravitational constant G, and so, it ought to apply to quantum field theory in curved spacetime.
The Bekenstein–Hawking boundary entropy of three-dimensional black holes exactly saturates the bound. The Schwarzschild radius is given by
and so the two-dimensional area of the black hole's event horizon is
and using the Planck length
the Bekenstein–Hawking entropy is
O |
https://en.wikipedia.org/wiki/Structured%20program%20theorem | The structured program theorem, also called the Böhm–Jacopini theorem, is a result in programming language theory. It states that a class of control-flow graphs (historically called flowcharts in this context) can compute any computable function if it combines subprograms in only three specific ways (control structures). These are
Executing one subprogram, and then another subprogram (sequence)
Executing one of two subprograms according to the value of a boolean expression (selection)
Repeatedly executing a subprogram as long as a boolean expression is true (iteration)
The structured chart subject to these constraints, particularly the loop constraint implying a single exit (as described later in this article), may however use additional variables in the form of bits (stored in an extra integer variable in the original proof) in order to keep track of information that the original program represents by the program location. The construction was based on Böhm's programming language P′′.
The theorem forms the basis of structured programming, a programming paradigm which eschews goto commands and exclusively uses subroutines, sequences, selection and iteration.
Origin and variants
The theorem is typically credited to a 1966 paper by Corrado Böhm and Giuseppe Jacopini. David Harel wrote in 1980 that the Böhm–Jacopini paper enjoyed "universal popularity", particularly with proponents of structured programming. Harel also noted that "due to its rather technical style [the 1966 |
https://en.wikipedia.org/wiki/Tellurite%20%28mineral%29 | Tellurite is a rare oxide mineral composed of tellurium dioxide (TeO2).
It occurs as prismatic to acicular transparent yellow to white orthorhombic crystals. It occurs in the oxidation zone of mineral deposits in association with native tellurium, emmonsite and other tellurium minerals. Its name comes from Tellus, which is the Latin name for the planet Earth.
It was first described in 1842 because of an occurrence in Faţa Băii, Zlatna, Alba County, Romania.
References
Oxide minerals
Tellurite and selenite minerals
Orthorhombic minerals
Minerals in space group 61 |
https://en.wikipedia.org/wiki/SSD%20%28disambiguation%29 | A solid-state drive is a type of data storage device which uses semiconductor memory rather than magnetic media.
SSD may also refer to:
Science and technology
Saturated-surface-dry, aggregate or porous solid condition
Biology and medicine
Schizophrenia spectrum disorders
Signal-sensing domain, in molecular biology
Sterol-sensing domain, a protein domain
Speech sound disorder
Sexual size dimorphism
Single-sided deafness
Somatic symptom disorder
A brand name for Silver sulfadiazine antibacterial
Computing
Server-side decoration of windows, an alternative to client-side decoration
Single-shot multibox detection, computer vision object detection
System sequence diagram in software engineering
Mathematics
Schwartz sequential dropping, an electoral system
Other uses
Sardar Sarovar Dam, Gujarat, India
Scalextric Sport Digital, toy cars
Singapore School for the Deaf
South Sudan (ISO 3166-1 alpha-3 code: SSD)
SSD (band), Boston, US, 1981–1985
Siroi language (ISO code: ssd), of Papua New Guinea
Special School District of St. Louis County
Stansted Airport railway station (National Rail code: SSD)
Social Security Disability Insurance (SSD or SSDI), US
State Security Department, North Korean secret police
United States–Russia Strategic Stability Dialogue, meetings to reduce the risk of US–Russia nuclear war |
https://en.wikipedia.org/wiki/Atomic%20lattice | Atomic lattice may refer to:
In mineralogy, atomic lattice refers to the arrangement of atoms into a crystal structure.
In order theory, a lattice is called an atomic lattice if the underlying partial order is atomic.
In chemistry, atomic lattice refers to the arrangement of atoms in an atomic crystalline solid. |
https://en.wikipedia.org/wiki/Geometrical%20frustration | In condensed matter physics, the term geometrical frustration (or in short: frustration) refers to a phenomenon where atoms tend to stick to non-trivial positions or where, on a regular crystal lattice, conflicting inter-atomic forces (each one favoring rather simple, but different structures) lead to quite complex structures. As a consequence of the frustration in the geometry or in the forces, a plenitude of distinct ground states may result at zero temperature, and usual thermal ordering may be suppressed at higher temperatures. Much studied examples are amorphous materials, glasses, or dilute magnets.
The term frustration, in the context of magnetic systems, has been introduced by Gerard Toulouse in 1977. Frustrated magnetic systems had been studied even before. Early work includes a study of the Ising model on a triangular lattice with nearest-neighbor spins coupled antiferromagnetically, by G. H. Wannier, published in 1950. Related features occur in magnets with competing interactions, where both ferromagnetic as well as antiferromagnetic couplings between pairs of spins or magnetic moments are present, with the type of interaction depending on the separation distance of the spins. In that case commensurability, such as helical spin arrangements may result, as had been discussed originally, especially, by A. Yoshimori, T. A. Kaplan, R. J. Elliott, and others, starting in 1959, to describe experimental findings on rare-earth metals. A renewed interest in such spin syste |
https://en.wikipedia.org/wiki/Hementin | Hementin is an anticoagulant protease (fibrinogen lytic enzyme) from the salivary glands of the giant Amazon leech (Haementeria ghilianii). Hementin is a calcium-dependent protease with a molecular weight of 80-120 kDa, and it contains 39 amino acid sequences. Hementin is present in both the anterior and posterior salivary glands, however it is mostly produced from certain cells in the anterior glands. The secretion of hementin is limited to the lumen of the proboscis, which the Amazon leech inserts into the host to suck blood. Hementin dissolves platelet-rich blood clots and lets the blood flow through the proboscis. Hementin is able to dissolve a type of blood clots that cannot be dissolved by other compounds, such as streptokinase and urokinase.
The processes of blood anticoagulation by hementin includes the degradation of fibrinogen. It is capable of disrupting the function of fibrinogen, a glycoprotein responsible for blood clotting, by cleaving three peptide bonds in its structure. Hementin is also capable of segregating platelets by cleaving the fibrinogen cross-link amongst the platelets. Fibrinogen acts as a substrate for thrombin, which converts this protein into its functional form, fibrin. Cleavage of fibrinogen in its native conformation at AαAsn102-Asn103, BβLys130-Gln131, and ϓPro76-Asn77 yield three sets of products. Hementin may also cause platelet deaggregation, although not under all circumstances. Because of its anticoagulant effects, Hementin can be desc |
https://en.wikipedia.org/wiki/Montel%27s%20theorem | In complex analysis, an area of mathematics, Montel's theorem refers to one of two theorems about families of holomorphic functions. These are named after French mathematician Paul Montel, and give conditions under which a family of holomorphic functions is normal.
Locally uniformly bounded families are normal
The first, and simpler, version of the theorem states that a family of holomorphic functions defined on an open subset of the complex numbers is normal if and only if it is locally uniformly bounded.
This theorem has the following formally stronger corollary. Suppose that
is a family of
meromorphic functions on an open set . If is such that
is not normal at , and is a neighborhood of , then is dense
in the complex plane.
Functions omitting two values
The stronger version of Montel's Theorem (occasionally referred to as the Fundamental Normality Test) states that a family of holomorphic functions, all of which omit the same two values is normal.
Necessity
The conditions in the above theorems are sufficient, but not necessary for normality. Indeed,
the family is normal, but does not omit any complex value.
Proofs
The first version of Montel's theorem is a direct consequence of Marty's Theorem (which
states that a family is normal if and only if the spherical derivatives are locally bounded)
and Cauchy's integral formula.
This theorem has also been called the Stieltjes–Osgood theorem, after Thomas Joannes Stieltjes and William Fogg Osgood.
The Corollary st |
https://en.wikipedia.org/wiki/Robbie%20McEwen | Robbie McEwen (born 24 June 1972) is an Australian former professional road cyclist. McEwen is a three-time winner of the Tour de France points classification and, at the peak of his career, was considered the world's fastest sprinter.
He last rode for on the UCI World Tour.
A former Australian BMX champion, McEwen switched to road cycling in 1990 at 18 years of age. He raced as a professional from 1996 until 2012.
McEwen retired from the World Tour after riding the 2012 Tour of California and is now a cycling broadcast commentator on the Tour Down Under and the Tour de France.
Career
McEwen was born in Brisbane. After four years of moving through the regional, state and national levels of cycling, he started at the Australian Institute of Sport in Canberra under road cycling coach Heiko Salzwedel. The first signs of his sprinting prowess on the international stage were at the Peace Race, winning three stages for the Australian national team.
McEwen competed in the road race at the 1996 Atlanta Olympic Games (23rd) and the 2000 Sydney Olympic Games (19th). He was also included on the Australian team for the 1994 UCI Road Cycling World Championship in Italy, and the 2002 UCI Road Cycling World Championship in Belgium, where he won a silver medal. McEwen was again selected for Australia at the 2004 Athens Olympic Games (11th) as part of the road race team.
McEwen was named 2002 Australian Cyclist of the Year, 2002 Male Road Cyclist of the Year and 1999 Male Road Cyclis |
https://en.wikipedia.org/wiki/Cysteine%20dioxygenase | Cysteine dioxygenase (CDO) is a non-heme iron enzyme that catalyzes the conversion of L-cysteine to cysteine sulfinic acid (cysteine sulfinate). CDO plays an important role in cysteine catabolism, regulating intracellular levels of cysteine and responding changes in cysteine availability. As such, CDO is highly regulated and undergoes large changes in concentration and efficiency. It oxidizes cysteine to the corresponding sulfinic acid by activation of dioxygen, although the exact mechanism of the reaction is still unclear. In addition to being found in mammals, CDO also exists in some yeast and bacteria, although the exact function is still unknown. CDO has been implicated in various neurodegenerative diseases and cancers, which is likely related to cysteine toxicity.
Function
CDO is responsible for the first major step in metabolism of cysteine. CDO oxidizes to cysteine sulfinic acid (which exists predominantly in the anionic sulfinate form in vivo). Overall, CDO catalyzes the addition of dioxygen (O2) to a thiol, producing a sulfinic acid. More specifically, CDO is part of the group of non-heme iron oxygenases that employ oxygen as an electron acceptor. Cysteine sulfinic acid is then metabolized further via two divergent pathways: decarboxylated to hypotaurine by sulfinoalanine decarboxylase and oxidized to taurine by hypotaurine dehydrogenase; or transaminated to a putative 3-sulfinylpyruvate intermediate, which decomposes spontaneously into pyruvate and sulfite. Sulfit |
https://en.wikipedia.org/wiki/Arnaud%20Denjoy | Arnaud Denjoy (; 5 January 1884 – 21 January 1974) was a French mathematician.
Biography
Denjoy was born in Auch, Gers. His contributions include work in harmonic analysis and differential equations. His integral was the first to be able to integrate all derivatives. Among his students is Gustave Choquet. He is also known for the more general broad Denjoy integral, or Khinchin integral.
Denjoy was an Invited Speaker of the ICM with talk Sur une classe d'ensembles parfaits en relation avec les fonctions admettant une dérivée seconde généralisée in 1920 at Strasbourg and with talk Les equations differentielles periodiques in 1950 at Cambridge, Massachusetts. In 1931 he was the president of the Société Mathématique de France. In 1942 he was elected a member of the Académie des sciences and was its president in 1962.
Denjoy married in 1923 and was the father of three sons. He died in Paris in 1974. He was an atheist with a strong interest in philosophy, psychology, and social issues.
The asteroid (19349) Denjoy is named in his honor.
Selected publications
Une extension de l'intégrale de Lebesgue, Académie des Sciences, pp. 859–862 (1912)
Les continus cycliques et la représentation conforme, Bulletin de la Société Mathématique de France, pp. 97-124 (1942)
Sur les fonctions dérivées sommables., Bulletin de la Société Mathématique de France, pp. 161-248 (1915)
Introduction à la théorie de fonctions de variables réelles, vol. 1, Hermann 1937
Aspects actuels de la pensée mathémat |
https://en.wikipedia.org/wiki/Jaroslav%20Kurzweil | Jaroslav Kurzweil (, 7 May 1926, Prague – 17 March 2022) was a Czech mathematician.
Biography
Born in Prague, Czechoslovakia, he was a specialist in ordinary differential equations and defined the Henstock–Kurzweil integral in terms of Riemann sums, first published in 1957 in the Czechoslovak Mathematical Journal. Kurzweil has been awarded the highest possible scientific prize of Czechia, the "Czech Brain" of the year 2006, as an acknowledgement of his life achievements.
With limited opportunities of contact between mathematicians within the Iron Curtain and those from the West, Kurzweil and Ivo Babuška founded a series of international scientific conferences named EQUADIFF, being held every four years since 1962 alternately in Prague, Bratislava, and Brno. He was chief editor of Mathematica Bohemica (then called Časopis pro pěstování matematiky) from 1956 to 1970 and was in its editorial board until 2007. In 2007, Kurzweil delivered a New Year's toast on Czech Television.
See also
Henstock–Kurzweil integral
References
Kurzweil, Jaroslav (2012). Generalized Ordinary Differential Equations: Not Absolutely Continuous Solutions. Series in Real Analysis. 11. World Scientific Publishing Company. .
Jiří Jarník; Štefan Schwabik; Milan Tvrdý; Ivo Vrkoč, Sixty years of Jaroslav Kurzweil. Czechoslovak Mathematical Journal, Vol. 36 (1986), No. 1, 147–166
Kurzweil J.. Generalized ordinary differential equations and continuous dependence on a parameter. Czechoslovak Mathe. Journ |
https://en.wikipedia.org/wiki/Balchik | Balchik ( ; ) is a Black Sea coastal town and seaside resort in the Southern Dobruja area of northeastern Bulgaria. It is in Dobrich Province, 35 km southeast of Dobrich and 42 km northeast of Varna. It sprawls scenically along hilly terraces descending from the Dobruja plateau to the sea, and is often called "The White City" because of its white hills.
Etymology
Balchik is named after the medieval ruler Balik, brother of Dobrotitsa, after whom the city of Dobrich is named.
History
Founded as a Thracian settlement, it was later colonised by the Ionian ancient Greeks with the name Krounoi () (renamed as Dionysopolis (), after the discovery of a statue of Dionysus in the sea). Later became a Greek-Byzantine and Bulgarian fortress. Under the Ottoman Empire, the town came to be known with its present name, which perhaps derived from a Gagauz word meaning "small town". Another opinion is that its actual name derived from that of a local noble - Balik.
Karvuna is the old Bulgarian name of the ancient Dionysopol (now Balchik). The external resemblance to the name of the modern town of Kavarna is an occasion for some local historians to identify them, but the archeological and historical data are not in favor of this statement. Karvuna was the capital of the Karvuna region - so called Dobrogea in the Middle Ages until the arrival of the Turks. The remains of the castle of the boyars Balik and Dobrotitsa were found above the city hospital of Balchik in the "Horizon" district (Gem |
https://en.wikipedia.org/wiki/Mathematical%20Reviews | Mathematical Reviews is a journal published by the American Mathematical Society (AMS) that contains brief synopses, and in some cases evaluations, of many articles in mathematics, statistics, and theoretical computer science. The AMS also publishes an associated online bibliographic database called MathSciNet which contains an electronic version of Mathematical Reviews and additionally contains citation information for over 3.5 million items
Reviews
Mathematical Reviews was founded by Otto E. Neugebauer in 1940 as an alternative to the German journal Zentralblatt für Mathematik, which Neugebauer had also founded a decade earlier, but which under the Nazis had begun censoring reviews by and of Jewish mathematicians. The goal of the new journal was to give reviews of every mathematical research publication. As of November 2007, the Mathematical Reviews database contained information on over 2.2 million articles. The authors of reviews are volunteers, usually chosen by the editors because of some expertise in the area of the article. It and Zentralblatt für Mathematik are the only comprehensive resources of this type. (The Mathematics section of Referativny Zhurnal is available only in Russian and is smaller in scale and difficult to access.) Often reviews give detailed summaries of the contents of the paper, sometimes with critical comments by the reviewer and references to related work. However, reviewers are not encouraged to criticize the paper, because the author does no |
https://en.wikipedia.org/wiki/Beloslav | Beloslav ( ) is a small industrial town in Varna Province, Northeastern Bulgaria, located 19 km away to the west from Varna downtown and Bulgarian Black Sea coast. It is the administrative centre of the homonymous Beloslav Municipality. As of December 2009, the town has a population of 7,937 inhabitants. Beloslav is located along the eastern edge of Lake Beloslav and on both sides of the canal connecting it with Lake Varna. The two parts of town are connected by a vehicle and passenger ferry across the canal. A road bridge is to be built in the near future over the canal east of town carrying a bypass of the E87 international highway now passing through downtown Varna via the Asparuhov most bridge.
Beloslav is an ancient settlement populated probably about 12,000 years ago. The first settlers inhabited the cave near today's town, where 12,000-year-old items were found. Around the 5th century BC the Thracians settled in the region. The town was also known as Krushevo, Gebedzhe (), Belovo, and Belevo. The Ottoman name is still used today in certain situations.
Beloslav is a transportation and industrial hub of national significance, part of the Varna-Devnya Industrial Complex. Transportation facilities include the railway and ro-ro ferry terminal with lines to Ukraine, Russia and Georgia served by the Sindel-razpredelitelna marshalling yard, the seaports of Varna Power Plant and LesPort of the Port of Varna complex, and a couple of railway stations. The largest (1320 M |
https://en.wikipedia.org/wiki/Citizen%20Information%20Project | In the United Kingdom, the Citizen Information Project (CIP) was a plan by the Office for National Statistics to build a national population register.
On 18 April 2006 it was announced that instead of continuing as a separate project, it would be integrated into the National Identity Register, the database behind the proposed national identity cards. It has been estimated that this might add £200 million to the cost of the identity cards. The National Identity Register was destroyed as the Identity Cards Act 2006 was repealed in 2011.
Scope and purpose
The register was to have been used as a single reference point for government contact, for the exchange of personal contact data, and for the collection of statistics, so reducing duplication in government departments and agencies. Government databases would have been linked together using National Insurance or other personal numbers.
In late 2003 the project moved into a definition phase. It was hoped that the CIP would be able to use data from the proposed National Identity Register.
A report on preliminary testing was due in April 2005, and it had been expected that it would have been implemented before the end of 2007 if approval had been given by Government. Initial estimates in 2004 suggested that the costs might have been £1.2 - £2.4 billion (240 million annually for a period of 5 to 10 years).
References
External links
Citizen Information Project
Testing of technology involved in a UK population register is soon |
https://en.wikipedia.org/wiki/Battle%20of%20Churubusco | The Battle of Churubusco took place on August 20, 1847, while Santa Anna's army was in retreat from the Battle of Contreras or Battle of Padierna during the Mexican–American War. It was the battle where the San Patricio Battalion, made up largely of US deserters, made their last stand against U.S. forces. The U.S. Army was victorious, outnumbering more than six-to-one the defending Mexican troops. After the battle, the U.S. Army was only 5 miles (8 km) away from Mexico City. 50 Saint Patrick's Battalion members were officially executed by the U.S. Army, all but two by hanging. Collectively, this was the largest mass execution in United States history.
Background
Following their defeats at Contreras, Antonio López de Santa Anna ordered Major General Nicolás Bravo Rueda with the Army of the Center, to retreat from San Antonio to Churubusco. Santa Anna also ordered Major General Manuel Rincón to hold the Franciscan Convent of San Mateo in Churubusco, with earthworks and seven guns, and placed General Francisco Pérez at the tête de pont on the south bank of the river. Two regiments were placed along the river while the convent included the Bravo Battalions of the Mexico City National Guard and the San Patricio Battalion, plus Santa Anna formed a reserve along the highway to the north.
Battle
Scott sent David Twiggs and Gideon Johnson Pillow's Divisions from San Angel to Coyoacán, while he ordered William Jenkins Worth to turn the San Antonio position. Worth sent Colonel Newman |
https://en.wikipedia.org/wiki/Multicollinearity | In statistics, multicollinearity (also collinearity) is a phenomenon in which one predictor variable in a multiple regression model can be perfectly predicted from the others. In this situation, the coefficient estimates of the multiple regression may change erratically in response to small changes in the data or the procedure used to fit the model.
Contrary to popular belief, including collinear variables does not reduce the predictive power or reliability of the model as a whole, nor does it reduce how accurately coefficients are estimated. In fact, high collinearity indicates that it is exceptionally important to include all variables, as excluding any variable will cause strong confounding.
Note that in statements of the assumptions underlying regression analyses such as ordinary least squares, the phrase "no multicollinearity" usually refers to the absence of multicollinearity, which is an exact (non-stochastic) linear relation among the predictors. In such a case, the design matrix has less than full rank, and therefore the moment matrix cannot be inverted. Under these circumstances, for a general linear model , the ordinary least squares estimator does not exist.
Definition
Multicollinearity refers to a situation in which explanatory variables in a multiple regression model are highly linearly related. Mathematically, a set of variables is perfectly multicollinear if there exist one or more exact linear relationships among some of the variables. That is, for al |
https://en.wikipedia.org/wiki/TI%20InterActive%21 | TI InterActive! is a Texas Instruments computer program which combines the functionality of all of the TI graphing calculators with extra features into a text editor which allows you to save equations, graphs, tables, spreadsheets, and text onto a document. TI InterActive! also includes a web browser, but it is just an embedded version of Internet Explorer. It also works with TI Connect to share data with the TI Graphing Calculators.
References
Computer algebra systems |
https://en.wikipedia.org/wiki/List%20of%20neighborhoods%20in%20San%20Francisco | San Francisco, in the US state of California, has both major, well-known neighborhoods and districts as well as smaller, specific subsections and developments. While there is considerable fluidity among the sources, one guidebook identifies five major districts, corresponding to the four quadrants plus a south central district. These five broad districts, counterclockwise are: Central/downtown, Richmond, Sunset, Upper Market and beyond (south central) and Bernal Heights/Bayview and beyond (southeast). Within each of these five districts are located major neighborhoods, and again there is considerable fluidity seen in the sources. The San Francisco Planning Department officially identifies 36 neighborhoods. Within these 36 official neighborhoods are a large number of minor districts, some of which are historical, and some of which are overlapping.
Some of San Francisco's neighborhoods are also officially designated as "cultural districts".
Alamo Square
Alamo Square is a subset of the Western Addition neighborhood. Its boundaries are not well-defined, but are generally considered to be Webster Street on the east, Golden Gate Avenue on the north, Divisadero Street on the west, and Oak Street on the south. It is characterized by Victorian architecture that was left largely untouched by the urban renewal projects in other parts of the Western Addition. On a clear day, the Transamerica Pyramid building and the tops of the Golden Gate Bridge and Bay Bridge can be seen from the pa |
https://en.wikipedia.org/wiki/Poincar%C3%A9%20metric | In mathematics, the Poincaré metric, named after Henri Poincaré, is the metric tensor describing a two-dimensional surface of constant negative curvature. It is the natural metric commonly used in a variety of calculations in hyperbolic geometry or Riemann surfaces.
There are three equivalent representations commonly used in two-dimensional hyperbolic geometry. One is the Poincaré half-plane model, defining a model of hyperbolic space on the upper half-plane. The Poincaré disk model defines a model for hyperbolic space on the unit disk. The disk and the upper half plane are related by a conformal map, and isometries are given by Möbius transformations. A third representation is on the punctured disk, where relations for q-analogues are sometimes expressed. These various forms are reviewed below.
Overview of metrics on Riemann surfaces
A metric on the complex plane may be generally expressed in the form
where λ is a real, positive function of and . The length of a curve γ in the complex plane is thus given by
The area of a subset of the complex plane is given by
where is the exterior product used to construct the volume form. The determinant of the metric is equal to , so the square root of the determinant is . The Euclidean volume form on the plane is and so one has
A function is said to be the potential of the metric if
The Laplace–Beltrami operator is given by
The Gaussian curvature of the metric is given by
This curvature is one-half of the Ricci scalar c |
https://en.wikipedia.org/wiki/Schwarz%E2%80%93Ahlfors%E2%80%93Pick%20theorem | In mathematics, the Schwarz–Ahlfors–Pick theorem is an extension of the Schwarz lemma for hyperbolic geometry, such as the Poincaré half-plane model.
The Schwarz–Pick lemma states that every holomorphic function from the unit disk U to itself, or from the upper half-plane H to itself, will not increase the Poincaré distance between points. The unit disk U with the Poincaré metric has negative Gaussian curvature −1. In 1938, Lars Ahlfors generalised the lemma to maps from the unit disk to other negatively curved surfaces:
Theorem (Schwarz–Ahlfors–Pick). Let U be the unit disk with Poincaré metric ; let S be a Riemann surface endowed with a Hermitian metric whose Gaussian curvature is ≤ −1; let be a holomorphic function. Then
for all
A generalization of this theorem was proved by Shing-Tung Yau in 1973.
References
Hyperbolic geometry
Riemann surfaces
Theorems in complex analysis
Theorems in differential geometry |
https://en.wikipedia.org/wiki/Theory%20of%20the%20second%20best | In welfare economics, the theory of the second best (also known as the general theory of second best or the second best theorem) concerns the situation when one or more optimality conditions cannot be satisfied. The economists Richard Lipsey and Kelvin Lancaster showed in 1956, that if one optimality condition in an economic model cannot be satisfied, it is possible that the next-best solution involves changing other variables away from the values that would otherwise be optimal. Politically, the theory implies that if it is infeasible to remove a particular market distortion, introducing one or more additional market distortions in an interdependent market may partially counteract the first, and lead to a more efficient outcome.
Implications
In an economy with some uncorrectable market failure in one sector, actions to correct market failures in another related sector with the intent of increasing economic efficiency may actually decrease overall economic efficiency. In theory, at least, it may be better to let two market imperfections cancel each other out rather than making an effort to fix either one. Thus, it may be optimal for the government to intervene in a way that is contrary to usual policy. This suggests that economists need to study the details of the situation before jumping to the theory-based conclusion that an improvement in market perfection in one area implies a global improvement in efficiency.
Application
Even though the theory of the second best was d |
https://en.wikipedia.org/wiki/Fundamental%20theorems%20of%20welfare%20economics | There are two fundamental theorems of welfare economics. The first states that in economic equilibrium, a set of complete markets, with complete information, and in perfect competition, will be Pareto optimal (in the sense that no further exchange would make one person better off without making another worse off). The requirements for perfect competition are these:
There are no externalities and each actor has perfect information.
Firms and consumers take prices as given (no economic actor or group of actors has market power).
The theorem is sometimes seen as an analytical confirmation of Adam Smith's "invisible hand" principle, namely that competitive markets ensure an efficient allocation of resources. However, there is no guarantee that the Pareto optimal market outcome is socially desirable, as there are many possible Pareto efficient allocations of resources differing in their desirability (e.g. one person may own everything and everyone else nothing).
The second theorem states that any Pareto optimum can be supported as a competitive equilibrium for some initial set of endowments. The implication is that any desired Pareto optimal outcome can be supported; Pareto efficiency can be achieved with any redistribution of initial wealth. However, attempts to correct the distribution may introduce distortions, and so full optimality may not be attainable with redistribution.
The theorems can be visualized graphically for a simple pure exchange economy by means of the Edg |
https://en.wikipedia.org/wiki/Lateral%20ventricles | The lateral ventricles are the two largest ventricles of the brain and contain cerebrospinal fluid (CSF). Each cerebral hemisphere contains a lateral ventricle, known as the left or right lateral ventricle, respectively.
Each lateral ventricle resembles a C-shaped cavity that begins at an inferior horn in the temporal lobe, travels through a body in the parietal lobe and frontal lobe, and ultimately terminates at the interventricular foramina where each lateral ventricle connects to the single, central third ventricle. Along the path, a posterior horn extends backward into the occipital lobe, and an anterior horn extends farther into the frontal lobe.
Structure
Each lateral ventricle takes the form of an elongated curve, with an additional anterior-facing continuation emerging inferiorly from a point near the posterior end of the curve; the junction is known as the trigone of the lateral ventricle. The centre of the superior curve is referred to as the body, while the three remaining portions are known as horns (cornua in Latin); they are usually referred to by their position relative to the body (anterior, posterior, or inferior), or sometimes by the lobe of the cerebral cortex into which they extend. Though somewhat flat, the lateral ventricles have a vaguely triangular cross-section. Ependyma, which are neuroepithelial cells, line the ventricular system including the lateral ventricles.
Between the inferior horn and the main body of the ventricle is the putamen, which |
https://en.wikipedia.org/wiki/No-communication%20theorem | In physics, the no-communication theorem or no-signaling principle is a no-go theorem from quantum information theory which states that, during measurement of an entangled quantum state, it is not possible for one observer, by making a measurement of a subsystem of the total state, to communicate information to another observer. The theorem is important because, in quantum mechanics, quantum entanglement is an effect by which certain widely separated events can be correlated in ways that, at first glance, suggest the possibility of communication faster-than-light. The no-communication theorem gives conditions under which such transfer of information between two observers is impossible. These results can be applied to understand the so-called paradoxes in quantum mechanics, such as the EPR paradox, or violations of local realism obtained in tests of Bell's theorem. In these experiments, the no-communication theorem shows that failure of local realism does not lead to what could be referred to as "spooky communication at a distance" (in analogy with Einstein's labeling of quantum entanglement as requiring "spooky action at a distance" on the assumption of QM's completeness).
Informal overview
The no-communication theorem states that, within the context of quantum mechanics, it is not possible to transmit classical bits of information by means of carefully prepared mixed or pure states, whether entangled or not. The theorem is only a sufficient condition that states that if th |
https://en.wikipedia.org/wiki/Proprietary%20trading | Proprietary trading (also known as prop trading) occurs when a trader trades stocks, bonds, currencies, commodities, their derivatives, or other financial instruments with the firm's own money (instead of using depositors' money) in order to make a profit for itself. Proprietary trading can create potential conflicts of interest such as insider trading and front running.
Proprietary traders may use a variety of strategies such as index arbitrage, statistical arbitrage, merger arbitrage, fundamental analysis, volatility arbitrage, or global macro trading, much like a hedge fund. Many reporters and analysts believe that large banks purposely leave ambiguous the proportion of proprietary versus non-proprietary trading, because it is felt that proprietary trading is riskier and results in more volatile profits.
Arbitrage
One of the main strategies of trading, traditionally associated with banks, is arbitrage. In the most basic sense, arbitrage is defined as taking advantage of a price discrepancy through the purchase or sale of certain combinations of securities to lock in a market-neutral profit. The trade will remain subject to various non-market risks, such as settlement risk and other operational risks. Investment banks, which are often active in many markets around the world, constantly watch for arbitrage opportunities.
One of the more-notable areas of arbitrage, called risk arbitrage or merger arbitrage, evolved in the 1980s. When a company plans to buy another company, |
https://en.wikipedia.org/wiki/Homodyne%20detection | In electrical engineering, homodyne detection is a method of extracting information encoded as modulation of the phase and/or frequency of an oscillating signal, by comparing that signal with a standard oscillation that would be identical to the signal if it carried null information. "Homodyne" signifies a single frequency, in contrast to the dual frequencies employed in heterodyne detection.
When applied to processing of the reflected signal in remote sensing for topography, homodyne detection lacks the ability of heterodyne detection to determine the size of a static discontinuity in elevation between two locations. (If there is a path between the two locations with smoothly changing elevation, then homodyne detection may in principle be able to track the signal phase along the path if sampling is dense enough). Homodyne detection is more readily applicable to velocity sensing.
In optics
In optical interferometry, homodyne signifies that the reference radiation (i.e. the local oscillator) is derived from the same source as the signal before the modulating process. For example, in a laser scattering measurement, the laser beam is split into two parts. One is the local oscillator and the other is sent to the system to be probed. The scattered light is then mixed with the local oscillator on the detector. This arrangement has the advantage of being insensitive to fluctuations in the frequency of the laser. Usually the scattered beam will be weak, in which case the (nearly) |
https://en.wikipedia.org/wiki/Flavin%20adenine%20dinucleotide | In biochemistry, flavin adenine dinucleotide (FAD) is a redox-active coenzyme associated with various proteins, which is involved with several enzymatic reactions in metabolism. A flavoprotein is a protein that contains a flavin group, which may be in the form of FAD or flavin mononucleotide (FMN). Many flavoproteins are known: components of the succinate dehydrogenase complex, α-ketoglutarate dehydrogenase, and a component of the pyruvate dehydrogenase complex.
FAD can exist in four redox states, which are the flavin-N(5)-oxide, quinone, semiquinone, and hydroquinone. FAD is converted between these states by accepting or donating electrons. FAD, in its fully oxidized form, or quinone form, accepts two electrons and two protons to become FADH2 (hydroquinone form). The semiquinone (FADH·) can be formed by either reduction of FAD or oxidation of FADH2 by accepting or donating one electron and one proton, respectively. Some proteins, however, generate and maintain a superoxidized form of the flavin cofactor, the flavin-N(5)-oxide.
History
Flavoproteins were first discovered in 1879 by separating components of cow's milk. They were initially called lactochrome due to their milky origin and yellow pigment. It took 50 years for the scientific community to make any substantial progress in identifying the molecules responsible for the yellow pigment. The 1930s launched the field of coenzyme research with the publication of many flavin and nicotinamide derivative structures and the |
https://en.wikipedia.org/wiki/TNR | TNR may refer to:
TNR, the gene that encodes the protein Tenascin-R
Ivato International Airport, Antananarivo, Madagascar, IATA code
The New Republic, an American magazine of politics and the arts
Times New Roman, a popular typeface installed on most desktop computers
Train Navette Rapide, an express train service in Morocco
Trap-neuter-return, a method for managing feral cats
T.N.Rajarathnam Pillai (1898-1956), Indian musician |
https://en.wikipedia.org/wiki/Houston%20Cellular | Houston Cellular was a Houston-based cell phone company which provided AMPS and D-AMPS (TDMA) service in the Greater Houston area. It was formed in 1983 and was operated as a partnership between LIN Broadcasting Corp., Mobile Communication Corp. of America and BellSouth Co. Its headquarters were located in Houston, Texas.
Through a series of acquisitions and mergers, within 10 years, the company consisted of a two-way partnership between BellSouth and AT&T Wireless. This partnership was also known as BellSouth Mobility, LLC. Houston Cellular began providing service in May 1986.
History
Houston Cellular was formed as a result of the October 1983 Federal Communications Commission's ruling that set aside sufficient frequencies for the operation of two cellular systems in each metropolitan area. In the early 1990s, this ruling was expanded with the advent of the 800 and 1900 MHz protocols, which are also known as PCS). The ruling reserved one system for application by local phone companies. In Houston, the competing non-local company service was applied by GTE Mobilnet.
In June 2000, the above-mentioned FCC ruling was repealed and GTE merged with Bell Atlantic to form Verizon Wireless. At that time, AT&T Wireless sold its 55% stake in Houston Cellular to BellSouth. AT&T Wireless then bought 20 MHz of the CDMA spectrum owned by the outgoing PrimeCo (which was swallowed in the formation of Verizon Wireless). This purchase of frequencies was used to directly compete in the Houst |
https://en.wikipedia.org/wiki/Byala | Byala may refer to:
Byala, Varna Province, a town in Varna Province in northeastern Bulgaria
Byala, Ruse Province, a town in Ruse Province in northern Bulgaria
, a village in Sliven Municipality in southeastern Bulgaria
Nawarupa - a mythical creature in Burmese and Arakanese mythology
See also
Biala (disambiguation) |
https://en.wikipedia.org/wiki/Byala%2C%20Varna%20Province | Byala ( , , Aspros) is a small town and seaside resort in Eastern Bulgaria, located on the Bulgarian Black Sea Coast in Varna Province. It is the administrative centre of the homonymous Byala Municipality and lies in a semi-mountainous region in the easternmost branches of Stara Planina about 50 km south of the city of Varna and 70 km north of Burgas. As of June 2022, the town has a population of 2,205 inhabitants.
History
The first known settlement here, near Cape St. Athanasius (Sveti Atanas) just south of town, is dated back to the 6th century BC. Its Greek name was perhaps Larissa, later, Aspros (White). Later the Roman road service station Templum Iovis (Temple of Jupiter, today's Obzor) was erected with a fortress nearby. The region became part of the First Bulgarian Empire in 681 AD; the Bulgarian fortress of Vicha stood later perhaps on cape Beli Nos (Cape White) north of town.
In the Ottoman period (from the late 14th century through 1878), a village known as Akdere (White creek) existed at the town's present location; it was later renamed Byala. A customs office between the Principality of Bulgaria and Eastern Rumelia was temporarily established in 1878. Until the early 1900s, the majority of the population was ethnic Greek; with the population transfers between Bulgaria and Greece following World War I, about 75% of them are now ethnic Bulgarians tracing their ancestry to the village of Athira (Bulgarian: Bozets) in Greek Macedonia. Byala was declared town |
https://en.wikipedia.org/wiki/Primosome | In molecular biology, a primosome is a protein complex responsible for creating RNA primers on single stranded DNA during DNA replication.
The primosome consists of seven proteins: DnaG primase, DnaB helicase, DnaC helicase assistant, DnaT, PriA, Pri B, and PriC. At each replication fork, the primosome is utilized once on the leading strand of DNA and repeatedly, initiating each Okazaki fragment, on the lagging DNA strand. Initially the complex formed by PriA, PriB, and PriC binds to DNA. Then the DnaB-DnaC helicase complex attaches along with DnaT. This structure is referred to as the pre-primosome. Finally, DnaG will bind to the pre-primosome forming a complete primosome. The primosome attaches 1-10 RNA nucleotides to the single stranded DNA creating a DNA-RNA hybrid. This sequence of RNA is used as a primer to initiate DNA polymerase III. The RNA bases are ultimately replaced with DNA bases by RNase H nuclease (eukaryotes) or DNA polymerase I nuclease (prokaryotes). DNA Ligase then acts to join the two ends together.
Assembly of the Escherichia coli primosome requires six proteins, PriA, PriB, PriC, DnaB, DnaC, and DnaT, acting at a primosome assembly site (pas) on an SSBcoated single-stranded (8s) DNA. Assembly is initiated by interactions of PriA and PriB with ssDNA and the pas. PriC, DnaB, DnaC, and DnaT then act on the PriAPriB- DNA complex to yield the primosome.
Primosomes are nucleoproteins assemblies that activate DNA replication forks. Their primary role i |
https://en.wikipedia.org/wiki/HSLS | HSLS refer to:
Hazy Sighted Link State Routing Protocol, a wireless network routing algorithm
Slovak People's Party (Slovak: )
Croatian Social Liberal Party (Croatian: )
See also
HSL (disambiguation) |
https://en.wikipedia.org/wiki/Virtual%20colonoscopy | Virtual colonoscopy (VC, also called CT colonography or CT pneumocolon) is the use of CT scanning or magnetic resonance imaging (MRI) to produce two- and three-dimensional images of the colon (large intestine), from the lowest part, the rectum, to the lower end of the small intestine, and to display the images on an electronic display device. The procedure is used to screen for colon cancer and polyps, and may detect diverticulosis. A virtual colonoscopy can provide 3D reconstructed endoluminal views of the bowel. VC provides a secondary benefit of revealing diseases or abnormalities outside the colon.
Procedure
While preparations for VC vary, the patient will usually be asked to take laxatives or other oral agents at home the day before the procedure to clear stool from the colon. A suppository is also used to cleanse the rectum of any remaining fecal matter. The patient may also be given a solution designed to coat any residual faeces which may not have been cleared by the laxative, called 'faecal tagging'. This allows the physician (usually a consultant radiologist), viewing the 3D images to effectively subtract the left-over faeces, which may otherwise give false positive results.
VC takes place in the radiology department of a hospital or medical center. The examination takes about 10 minutes and does not require sedatives.
During the procedure:
The patient is placed in a supine position on the examination table
The patient may be given a dosage of Butylscopolam |
https://en.wikipedia.org/wiki/Methuselah%20%28cellular%20automaton%29 | In cellular automata, a methuselah is a small "seed" pattern of initial live cells that take a large number of generations in order to stabilize. More specifically, Martin Gardner defines them as patterns of fewer than ten live cells which take longer than 50 generations to stabilize, although some patterns that are larger than ten cells have also been called methuselahs. Patterns must eventually stabilize to be considered methuselahs. The term comes from the Biblical Methuselah, who lived for 969 years.
In Conway's Game of Life
In Conway's Game of Life, one of the smallest methuselahs is the R-pentomino, a pattern of five cells first considered by Conway himself, that takes 1103 generations before stabilizing with 116 cells. The acorn, a pattern of only seven live cells developed by Charles Corderman, takes 5206 generations to stabilize and produce a 633-cell pattern referred to as its "oak". Some other examples of methuselahs are called bunnies and rabbits.
References
External links
List of longest-lived methuselahs at LifeWiki
Cellular automaton patterns |
https://en.wikipedia.org/wiki/DCLS | DCLS may refer to:
Doctor Clinical Laboratory Science
Direct Current Level Shift, a modulation type in the IRIG timecode system
Dauphin County Library System, in Dauphin County, Pennsylvania
Doctorate of Clinical Laboratory Science. The DCLS is an advanced professional doctorate designed for practicing CLSs who wish to further their level of clinical expertise and to develop leadership and management skills |
https://en.wikipedia.org/wiki/Krabbe%20disease | Krabbe disease (KD) (also known as globoid cell leukodystrophy or galactosylceramide lipidosis) is a rare and often fatal lysosomal storage disease that results in progressive damage to the nervous system. KD involves dysfunctional metabolism of sphingolipids and is inherited in an autosomal recessive pattern. The disease is named after the Danish neurologist Knud Krabbe (1885–1961).
Signs and symptoms
Symptoms in asymptomatic infantile-onset (<12 months after birth) and later-onset Krabbe disease present themselves differently. Of individuals with infantile-onset Krabbe disease, 85–90% display progressive neurologic deterioration in infancy and death before the age of two. Symptoms include irritability, fevers, limb stiffness, seizures, feeding difficulties (like GERD), vomiting, staring episodes, and slowing of mental and motor development. In the first stages of the disease, doctors often mistake the symptoms for those of cerebral palsy. Other symptoms include muscle weakness, spasticity, deafness, optic atrophy, optic nerve enlargement, blindness, paralysis, and difficulty when swallowing. Prolonged weight loss may also occur.
10–15% of individuals with later-onset Krabbe disease have a much slower disease progression. These individuals may also display symptoms such as esotropia, slurred speech, and slow development or loss of motor milestones.
Causes
Krabbe disease is caused by mutations in the GALC gene located on chromosome 14 (14q31), which is inherited in an aut |
https://en.wikipedia.org/wiki/Pappus%27s%20hexagon%20theorem | In mathematics, Pappus's hexagon theorem (attributed to Pappus of Alexandria) states that
given one set of collinear points and another set of collinear points then the intersection points of line pairs and and and are collinear, lying on the Pappus line. These three points are the points of intersection of the "opposite" sides of the hexagon .
It holds in a projective plane over any field, but fails for projective planes over any noncommutative division ring. Projective planes in which the "theorem" is valid are called pappian planes.
If one restricts the projective plane such that the Pappus line is the line at infinity, one gets the affine version of Pappus's theorem shown in the second diagram.
If the Pappus line and the lines have a point in common, one gets the so-called little version of Pappus's theorem.
The dual of this incidence theorem states that given one set of concurrent lines , and another set of concurrent lines , then the lines defined by pairs of points resulting from pairs of intersections and and and are concurrent. (Concurrent means that the lines pass through one point.)
Pappus's theorem is a special case of Pascal's theorem for a conic—the limiting case when the conic degenerates into 2 straight lines. Pascal's theorem is in turn a special case of the Cayley–Bacharach theorem.
The Pappus configuration is the configuration of 9 lines and 9 points that occurs in Pappus's theorem, with each line meeting 3 of the points and each poi |
https://en.wikipedia.org/wiki/International%20Nucleotide%20Sequence%20Database%20Collaboration | The International Nucleotide Sequence Database Collaboration (INSDC) consists of a joint effort to collect and disseminate databases containing DNA and RNA sequences. It involves the following computerized databases: NIG's DNA Data Bank of Japan (Japan), NCBI's GenBank (USA) and the EMBL-EBI's European Nucleotide Archive (UK). New and updated data on nucleotide sequences contributed by research teams to each of the three databases are synchronized on a daily basis through continuous interaction between the staff at each the collaborating organizations.
All of the data in INSDC is available for free and unrestricted access, for any purpose, with no restrictions on analysis, redistribution, or re-publication of the data. This policy has been a foundational principle of the INSDC since its inception. The official policy statement can be found at http://www.insdc.org/. Since the 1990s, most of the world's major scientific journals have required that sequence data be deposited in an INSDC database as a pre-condition for publication.
The DDBJ/EMBL/GenBank synchronization is maintained according to a number of guidelines which are produced and published by an International Advisory Board. The guidelines consist of a common definition of the feature tables for the databases, which regulate the content and syntax of the database entries, in the form of a common DTD (Document Type Definition).
The syntax is called INSDSeq and its core consists of the letter sequence of the gene |
https://en.wikipedia.org/wiki/Intracellular%20pH | Intracellular pH (pHi) is the measure of the acidity or basicity (i.e., pH) of intracellular fluid. The pHi plays a critical role in membrane transport and other intracellular processes. In an environment with the improper pHi, biological cells may have compromised function. Therefore, pHi is closely regulated in order to ensure proper cellular function, controlled cell growth, and normal cellular processes. The mechanisms that regulate pHi are usually considered to be plasma membrane transporters of which two main types exist — those that are dependent and those that are independent of the concentration of bicarbonate (). Physiologically normal intracellular pH is most commonly between 7.0 and 7.4, though there is variability between tissues (e.g., mammalian skeletal muscle tends to have a pHi of 6.8–7.1). There is also pH variation across different organelles, which can span from around 4.5 to 8.0. pHi can be measured in a number of different ways.
Homeostasis
Intracellular pH is typically lower than extracellular pH due to lower concentrations of HCO3−. A rise of extracellular (e.g., serum) partial pressure of carbon dioxide (pCO2) above 45 mmHg leads to formation of carbonic acid, which causes a decrease of pHi as it dissociates:
H2O + CO2 H2CO3 H+ + HCO3–
Since biological cells contain fluid that can act as a buffer, pHi can be maintained fairly well within a certain range. Cells adjust their pHi accordingly upon an increase in acidity or basicity, usually with th |
https://en.wikipedia.org/wiki/Generalized%20extreme%20value%20distribution | In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel, Fréchet and Weibull families also known as type I, II and III extreme value distributions. By the extreme value theorem the GEV distribution is the only possible limit distribution of properly normalized maxima of a sequence of independent and identically distributed random variables. Note that a limit distribution needs to exist, which requires regularity conditions on the tail of the distribution. Despite this, the GEV distribution is often used as an approximation to model the maxima of long (finite) sequences of random variables.
In some fields of application the generalized extreme value distribution is known as the Fisher–Tippett distribution, named after Ronald Fisher and L. H. C. Tippett who recognised three different forms outlined below. However usage of this name is sometimes restricted to mean the special case of the Gumbel distribution. The origin of the common functional form for all 3 distributions dates back to at least Jenkinson, A. F. (1955), though allegedly it could also have been given by von Mises, R. (1936).
Specification
Using the standardized variable where the location parameter, can be any real number, and is the scale parameter; the cumulative distribution function of the GEV distribution is then
where the shape parameter, can be any real numb |
https://en.wikipedia.org/wiki/Anti-M%C3%BCllerian%20hormone | Anti-Müllerian hormone (AMH), also known as Müllerian-inhibiting hormone (MIH), is a glycoprotein hormone structurally related to inhibin and activin from the transforming growth factor beta superfamily, whose key roles are in growth differentiation and folliculogenesis. In humans, it is encoded by the gene, on chromosome 19p13.3, while its receptor is encoded by the gene on chromosome 12.
AMH is activated by SOX9 in the Sertoli cells of the male fetus. Its expression inhibits the development of the female reproductive tract, or Müllerian ducts (paramesonephric ducts), in the male embryo, thereby arresting the development of fallopian tubes, uterus, and upper vagina. AMH expression is critical to sex differentiation at a specific time during fetal development, and appears to be tightly regulated by nuclear receptor SF-1, transcription GATA factors, sex-reversal gene DAX1, and follicle-stimulating hormone (FSH). Mutations in both the AMH gene and the type II AMH receptor have been shown to cause the persistence of Müllerian derivatives in males that are otherwise normally masculinized.
AMH is also a product of granulosa cells of the preantral and small antral follicles in women. As such, AMH is only present in the ovary until menopause. Production of AMH regulates folliculogenesis by inhibiting recruitment of follicles from the resting pool in order to select for the dominant follicle, after which the production of AMH diminishes. As a product of the granulosa cells, which |
https://en.wikipedia.org/wiki/3Dc | 3Dc (FourCC : ATI2), also known as DXN, BC5, or Block Compression 5 is a lossy data compression algorithm for normal maps invented and first implemented by ATI. It builds upon the earlier DXT5 algorithm and is an open standard. 3Dc is now implemented by both ATI and Nvidia.
Target application
The target application, normal mapping, is an extension of bump mapping that simulates lighting on geometric surfaces by reading surface normals from a rectilinear grid analogous to a texture map - giving simple models the impression of increased complexity. This additional channel however increases the load on the graphics system's memory bandwidth. Pre-existing lossy compression algorithms implemented on consumer 3D hardware lacked the precision necessary for reproducing normal maps without excessive visible artefacts, justifying the development of 3Dc.
Algorithm
Surface normals are three-dimensional vectors of unit length. Because of the length constraint only two elements of any normal need to be stored. The input is therefore an array of two-dimensional values.
Compression is performed in 4×4 blocks. In each block the two components of each value are compressed separately. For each block, each of the two components have a palette of 8 values to choose from. The palettes are generated from two values representing the start and end of a line and the other six values being generated as linear combinations of the start and end values.
Compression is natively achieved by finding |
https://en.wikipedia.org/wiki/Intrinsic%20equation | In geometry, an intrinsic equation of a curve is an equation that defines the curve using a relation between the curve's intrinsic properties, that is, properties that do not depend on the location and possibly the orientation of the curve. Therefore an intrinsic equation defines the shape of the curve without specifying its position relative to an arbitrarily defined coordinate system.
The intrinsic quantities used most often are arc length , tangential angle , curvature or radius of curvature, and, for 3-dimensional curves, torsion . Specifically:
The natural equation is the curve given by its curvature and torsion.
The Whewell equation is obtained as a relation between arc length and tangential angle.
The Cesàro equation is obtained as a relation between arc length and curvature.
The equation of a circle (including a line) for example is given by the equation where is the arc length, the curvature and the radius of the circle.
These coordinates greatly simplify some physical problem. For elastic rods for example, the potential energy is given by
where is the bending modulus . Moreover, as , elasticity of rods can be given a simple variational form.
References
External links
Curves
Equations |
https://en.wikipedia.org/wiki/S%C3%A9tna | Sétna or Sedna was the name of two legendary High Kings of Ireland:
Sétna Airt, son of Artrí, son of Éber, son of Ír, son of Míl Espáine
Sétna Innarraid, son of Bres Rí
See also
Renu Setna, British actor
Sethna
Sedna (disambiguation) |
https://en.wikipedia.org/wiki/S%C3%ADrna%20S%C3%A1eglach | Sírna Sáeglach ("the long-lived"), son of Dian mac Demal, son of Demal mac Rothechtaid, son of Rothechtaid mac Main, was, according to medieval Irish legend and historical tradition, a High King of Ireland. He separated the province of Ulster from the authority of the High King, and is said to have made war against the Ulaid, who had killed his great grandfather, for a hundred years according to the Lebor Gabála Érenn, 150 years according to the Annals of the Four Masters, but Geoffrey Keating, citing an ancient poem, gives him only twenty-one years. According to one version of the Lebor Gabála, the Ulaid united with the Fomorians and gave him battle at Móin Trógaide in County Meath, but a plague fell on them and the leaders of both sides died. According to another version, agreed by Keating and the Four Masters, Sírna was killed by Rothechtaid Rotha at Alind. The Lebor Gabála synchronises the start of his reign with the reign of Deioces of the Medes (694–665 BC), and his death with his successor Phraortes (665–633 BC). The chronology of Keating's Foras Feasa ar Éirinn dates his reign to 814–794 BC, that of the Annals of the Four Masters to 1181–1031 BC.
References
Legendary High Kings of Ireland |
https://en.wikipedia.org/wiki/G%C3%ADallchad | Gíallchad, the son of Ailill Olcháin, son of Sírna Sáeglach, was, according to medieval Irish legend and historical tradition, the 37th High King of Ireland. Gíallchad took power after killing his predecessor, and the son of his grandfather's killer, Elim Olfínechta, in the battle of Comair Trí nUisce. He is said to have taken one out of every five men of Munster hostage. Gíallchad reigned for nine years, before being killed by Elim's son Art Imlech at Mag Muaide. The Lebor Gabála Érenn synchronises his reign with that of Phraortes of the Medes (665–633 BCE). The chronology of Geoffrey Keating's Foras Feasa ar Éirinn dates his reign to 786–777 BCE, that of the Annals of the Four Masters to 1023–1014 BCE.
Gíallchad's son was Nuadu Finn Fáil.
References
Legendary High Kings of Ireland |
https://en.wikipedia.org/wiki/List%20of%20census%20divisions%20of%20Saskatchewan | The province of Saskatchewan, Canada is divided into 18 census divisions according to Statistics Canada. Unlike in some other provinces, census divisions do not reflect the organization of local government in Saskatchewan. These areas exist solely for the purposes of statistical analysis and presentation; they have no government of their own.
Saskatchewan's census divisions consist of numerous census subdivisions which include subdivisions such as:
Urban municipalities (cities, towns, villages, and resort villages);
Rural municipalities;
Northern municipalities (northern towns, northern villages, and northern hamlets); and
Indian reserves
List of census divisions
See also
Administrative divisions of Canada
List of communities in Saskatchewan
List of cities in Saskatchewan
List of Indian reserves in Saskatchewan
List of resort villages in Saskatchewan
List of rural municipalities in Saskatchewan
List of towns in Saskatchewan
List of villages in Saskatchewan
Notes
References
Census divisions |
https://en.wikipedia.org/wiki/Solar%20vehicle | A solar vehicle or solar electric vehicle is an electric vehicle powered completely or significantly by direct solar energy. Usually, photovoltaic (PV) cells contained in solar panels convert the sun's energy directly into electric energy.
The term "solar vehicle" usually implies that solar energy is used to power all or part of a vehicle's propulsion. Solar power may also be used to provide power for communications or controls or other auxiliary functions.
Solar vehicles are not sold as practical day-to-day transportation devices at present, but are primarily demonstration vehicles and engineering exercises, often sponsored by government agencies. However, indirectly solar-charged vehicles are widespread and solar boats are available commercially.
Land
Solar cars
Solar cars are electric cars that use photovoltaic (PV) cells to convert sunlight into electrical power to charge the car's battery and to power the car's electric motors.
Solar cars have been designed for solar car races and for public use. Solar vehicles must be light and efficient to get the best range from their limited captured power. pound or even vehicles would be less practical because the limited solar power would not take them as far. Most student built solar cars lack the safety and convenience features of conventional vehicles and are thus not street legal.
The first solar family car, Stella, was built in 2013 by students in the Netherlands. This vehicle is capable of on one charge during sunl |
https://en.wikipedia.org/wiki/Kaprekar%27s%20routine | In number theory, Kaprekar's routine is an iterative algorithm named after its inventor, Indian mathematician D. R. Kaprekar. Each iteration starts with a number, sorts the digits into descending and ascending order, and calculates the difference between the two new numbers.
As an example, starting with the number 8991 in base 10:
6174, known as Kaprekar's constant, is a fixed point of this algorithm. Any four-digit number (in base 10) with at least two distinct digits will reach 6174 within seven iterations. The algorithm runs on any natural number in any given number base.
Definition and properties
The algorithm is as follows:
Choose any natural number in a given number base . This is the first number of the sequence.
Create a new number by sorting the digits of in descending order, and another number by sorting the digits of in ascending order. These numbers may have leading zeros, which can be ignored. Subtract to produce the next number of the sequence.
Repeat step 2.
The sequence is called a Kaprekar sequence and the function is the Kaprekar mapping. Some numbers map to themselves; these are the fixed points of the Kaprekar mapping, and are called Kaprekar's constants. Zero is a Kaprekar's constant for all bases , and so is called a trivial Kaprekar's constant. All other Kaprekar's constant are nontrivial Kaprekar's constants.
For example, in base 10, starting with 3524,
with 6174 as a Kaprekar's constant.
All Kaprekar sequences will eith |
https://en.wikipedia.org/wiki/Consistent%20estimator | In statistics, a consistent estimator or asymptotically consistent estimator is an estimator—a rule for computing estimates of a parameter θ0—having the property that as the number of data points used increases indefinitely, the resulting sequence of estimates converges in probability to θ0. This means that the distributions of the estimates become more and more concentrated near the true value of the parameter being estimated, so that the probability of the estimator being arbitrarily close to θ0 converges to one.
In practice one constructs an estimator as a function of an available sample of size n, and then imagines being able to keep collecting data and expanding the sample ad infinitum. In this way one would obtain a sequence of estimates indexed by n, and consistency is a property of what occurs as the sample size “grows to infinity”. If the sequence of estimates can be mathematically shown to converge in probability to the true value θ0, it is called a consistent estimator; otherwise the estimator is said to be inconsistent.
Consistency as defined here is sometimes referred to as weak consistency. When we replace convergence in probability with almost sure convergence, then the estimator is said to be strongly consistent. Consistency is related to bias; see bias versus consistency.
Definition
Formally speaking, an estimator Tn of parameter θ is said to be weakly consistent, if it converges in probability to the true value of the parameter:
i.e. if, for all ε > 0 |
https://en.wikipedia.org/wiki/Sylvania%20Electric%20Products | Sylvania Electric Products Inc. was an American manufacturer of diverse electrical equipment, including at various times radio transceivers, vacuum tubes, semiconductors, and mainframe computers such as MOBIDIC. They were one of the companies involved in the development of the COBOL programming language.
History
The Hygrade Sylvania Corporation was formed when NILCO, Sylvania and Hygrade Lamp Company merged into one company in 1931.
In 1939, Hygrade Sylvania started preliminary research on fluorescent technology, and later that year, demonstrated the first linear, or tubular, fluorescent lamp. It was featured at the 1939 New York World's Fair. Sylvania was also a manufacturer of both vacuum tubes and transistors. In 1942, the company changed its name to Sylvania Electric Products Inc.
During World War II, Sylvania was chosen from among several competing companies to manufacture the miniature vacuum tubes used in proximity fuze shells due to its quality standards and mass production capabilities.
In 1959, Sylvania Electronics merged with General Telephone to form General Telephone and Electronics (GTE).
Sylvania developed the earliest flash cubes for still cameras, later selling the technology to Eastman Kodak Company, and later a 10-flash unit called FlipFlash, as well as a line of household electric light bulbs, which continued during GTE's ownership, later sold off to the German manufacturer Osram, and is today marketed as Osram Sylvania.
In June 1964, Sylvania unve |
https://en.wikipedia.org/wiki/Consortium%20for%20the%20Barcode%20of%20Life | The Consortium for the Barcode of Life (CBOL) was an international initiative dedicated to supporting the development of DNA barcoding as a global standard for species identification. CBOL's Secretariat Office is hosted by the National Museum of Natural History, Smithsonian Institution, in Washington, DC. Barcoding was proposed in 2003 by Prof. Paul Hebert of the University of Guelph in Ontario as a way of distinguishing and identifying species with a short standardized gene sequence. Hebert proposed the 658 bases of the Folmer region of the mitochondrial gene cytochrome-C oxidase-1 as the standard barcode region. Hebert is the Director of the Biodiversity Institute of Ontario, the Canadian Centre for DNA Barcoding, and the International Barcode of Life Project (iBOL), all headquartered at the University of Guelph. The Barcode of Life Data Systems (BOLD) is also located at the University of Guelph.
CBOL was created in May 2004 with support of the Alfred P. Sloan Foundation, following two meetings in 2003, also funded by the Sloan Foundation, at the Banbury Center, Cold Spring Harbor Laboratory. Since then, more than 200 organizations from more than 50 countries have joined CBOL and agreed to put their barcode data in a public database. CBOL promotes DNA barcoding through workshops, working groups, international conferences, outreach meetings to developing countries, planning meetings for barcoding projects, and production of outreach material to raise awareness of bar |
https://en.wikipedia.org/wiki/Mellin%20inversion%20theorem | In mathematics, the Mellin inversion formula (named after Hjalmar Mellin) tells us conditions under
which the inverse Mellin transform, or equivalently the inverse two-sided Laplace transform, are defined and recover the transformed function.
Method
If is analytic in the strip ,
and if it tends to zero uniformly as for any real value c between a and b, with its integral along such a line converging absolutely, then if
we have that
Conversely, suppose is piecewise continuous on the positive real numbers, taking a value halfway between the limit values at any jump discontinuities, and suppose the integral
is absolutely convergent when . Then is recoverable via the inverse Mellin transform from its Mellin transform . These results can be obtained by relating the Mellin transform to the Fourier transform by a change of variables and then applying an appropriate version of the Fourier inversion theorem.
Boundedness condition
The boundedness condition on can be strengthened if
is continuous. If is analytic in the strip , and if , where K is a positive constant, then as defined by the inversion integral exists and is continuous; moreover the Mellin transform of is for at least .
On the other hand, if we are willing to accept an original which is a
generalized function, we may relax the boundedness condition on
to
simply make it of polynomial growth in any closed strip contained in the open strip .
We may also define a Banach space version of this theorem. |
https://en.wikipedia.org/wiki/John%20C.%20Hull%20%28economist%29 | John C. Hull is a Professor of Derivatives and Risk Management at the Rotman School of Management at the University of Toronto.
He is a respected researcher in the academic field of quantitative finance (see for example the Hull-White model) and is the author of two books on financial derivatives that are widely used texts for market practitioners: "Options, Futures, and Other Derivatives" and "Fundamentals of Futures and Options Markets". He has also written "Risk Management and Financial Institutions" and "Machine Learning in Business: An Introduction to the World of Data Science"
He studied Mathematics at Cambridge University (B.A. & M.A.), and holds an M.A. in Operational Research from Lancaster University and a Ph.D. in Finance from Cranfield University. In 1999, he was awarded the Financial Engineer of the Year Award, by the International Association of Financial Engineers. He has also won many teaching awards, such as the University of Toronto's prestigious Northrop Frye award.
He has twin sons named Peter and David, and a wife named Michelle.
Selected publications
A Neural Network Approach to Understanding Implied Volatility Movements" Quantitative Finance, 2020, forthcoming (with Jay Cao and Jacky Chen)
Funding Long Shots" Journal of Investment Management, 17, 4, 2019 : 1-33 (with Andrew Lo and Roger Stein)
Interest Rate Trees: Extensions and Applications, Quantitative Finance, 18, 7 (2018): 1199-1209 (with Alan White)
Optimal Delta Hedging for Options, Journal |
https://en.wikipedia.org/wiki/Amobarbital | Amobarbital (formerly known as amylobarbitone or sodium amytal as the soluble sodium salt) is a drug that is a barbiturate derivative. It has sedative-hypnotic properties. It is a white crystalline powder with no odor and a slightly bitter taste. It was first synthesized in Germany in 1923. It is considered a short to intermediate acting barbiturate. If amobarbital is taken for extended periods of time, physiological and psychological dependence can develop. Amobarbital withdrawal mimics delirium tremens and may be life-threatening. Amobarbital was manufactured by Eli Lilly and Company in the US under the brand name Amytal in bright blue bullet shaped capsules (known as Pulvules) or pink tablets (known as Diskets) containing 50, 100, or 200 milligrams of the drug. The drug was also manufactured generically. Amobarbital was widely misused, known as "Blue Heavens" on the street. Amytal, as well as Tuinal, a combination drug containing equal quantities of secobarbital and amobarbital, were both manufactured by Eli Lilly until the late-1990s. However, as the popularity of benzodiazepines increased, prescriptions for these medications became increasingly rare beginning in the mid to late-1980s.
Pharmacology
In an in vitro study in rat thalamic slices, amobarbital worked by activating GABAA receptors, which decreased input resistance, depressed burst and tonic firing, especially in ventrobasal and intralaminar neurons, while at the same time increasing burst duration and mean c |
https://en.wikipedia.org/wiki/Hardy%27s%20theorem | In mathematics, Hardy's theorem is a result in complex analysis describing the behavior of holomorphic functions.
Let be a holomorphic function on the open ball centered at zero and radius in the complex plane, and assume that is not a constant function. If one defines
for then this function is strictly increasing and is a convex function of .
See also
Maximum principle
Hadamard three-circle theorem
References
John B. Conway. (1978) Functions of One Complex Variable I. Springer-Verlag, New York, New York.
Theorems in complex analysis |
https://en.wikipedia.org/wiki/List%20of%20enzymes | Enzymes are listed here by their classification in the International Union of Biochemistry and Molecular Biology's Enzyme Commission (EC) numbering system:
:Category:Oxidoreductases (EC 1) (Oxidoreductase)
Dehydrogenase
Luciferase
DMSO reductase
:Category:EC 1.1 (act on the CH-OH group of donors)
:Category:EC 1.1.1 (with NAD+ or NADP+ as acceptor)
Alcohol dehydrogenase (NAD)
Alcohol dehydrogenase (NADP)
Homoserine dehydrogenase
Aminopropanol oxidoreductase
Diacetyl reductase
Glycerol dehydrogenase
Propanediol-phosphate dehydrogenase
glycerol-3-phoshitiendopene dehydrogenase (NAD+)
D-xylulose reductase
L-xylulose reductase
Lactate dehydrogenase
Malate dehydrogenase
Isocitrate dehydrogenase
HMG-CoA reductase
:Category:EC 1.1.2 (with a cytochrome as acceptor)
:Category:EC 1.1.3 (with oxygen as acceptor)
Glucose oxidase
L-gulonolactone oxidase
Thiamine oxidase
Xanthine oxidase
Category:EC 1.1.4 (with a disulfide as acceptor)
:Category:EC 1.1.5 (with a quinone or similar compound as acceptor)
:Category:EC 1.1.99 (with other acceptors)
:Category:EC 1.2 (act on the aldehyde or oxo group of donors)
:Category:EC 1.2.1 (with NAD+ or NADP+ as acceptor)
Acetaldehyde dehydrogenase
Glyceraldehyde 3-phosphate dehydrogenase
Pyruvate dehydrogenase
:Category:EC 1.2.4
Oxoglutarate dehydrogenase
:Category:EC 1.3 (act on the CH-CH group of donors)
:Category:EC 1.3.1 (with NAD+ or NADP+ as acceptor)
Biliverdin reductase
:Category:EC 1.3.2 (with a cytochrome as acceptor)
: |
https://en.wikipedia.org/wiki/Hadamard%20three-circle%20theorem | In complex analysis, a branch of mathematics, the
Hadamard three-circle theorem is a result about the behavior of holomorphic functions.
Let be a holomorphic function on the annulus
Let be the maximum of on the circle Then, is a convex function of the logarithm Moreover, if is not of the form for some constants and , then is strictly convex as a function of
The conclusion of the theorem can be restated as
for any three concentric circles of radii
History
A statement and proof for the theorem was given by J.E. Littlewood in 1912, but he attributes it to no one in particular, stating it as a known theorem. Harald Bohr and Edmund Landau attribute the theorem to Jacques Hadamard, writing in 1896; Hadamard published no proof.
Proof
The three circles theorem follows from the fact that for any real a, the function Re log(zaf(z)) is harmonic between two circles, and therefore takes its maximum value on one of the circles. The theorem follows by choosing the constant a so that this harmonic function has the same maximum value on both circles.
The theorem can also be deduced directly from Hadamard's three-lines theorem.
See also
Maximum principle
Logarithmically convex function
Hardy's theorem
Hadamard three-lines theorem
Borel–Carathéodory theorem
Phragmén–Lindelöf principle
Notes
References
E. C. Titchmarsh, The theory of the Riemann Zeta-Function, (1951) Oxford at the Clarendon Press, Oxford. (See chapter 14)
External links
"proof of Hadamard three- |
https://en.wikipedia.org/wiki/Projection-slice%20theorem | In mathematics, the projection-slice theorem, central slice theorem or Fourier slice theorem in two dimensions states that the results of the following two calculations are equal:
Take a two-dimensional function f(r), project (e.g. using the Radon transform) it onto a (one-dimensional) line, and do a Fourier transform of that projection.
Take that same function, but do a two-dimensional Fourier transform first, and then slice it through its origin, which is parallel to the projection line.
In operator terms, if
F1 and F2 are the 1- and 2-dimensional Fourier transform operators mentioned above,
P1 is the projection operator (which projects a 2-D function onto a 1-D line),
S1 is a slice operator (which extracts a 1-D central slice from a function),
then
This idea can be extended to higher dimensions.
This theorem is used, for example, in the analysis of medical
CT scans where a "projection" is an x-ray
image of an internal organ. The Fourier transforms of these images are
seen to be slices through the Fourier transform of the 3-dimensional
density of the internal organ, and these slices can be interpolated to build
up a complete Fourier transform of that density. The inverse Fourier transform
is then used to recover the 3-dimensional density of the object. This technique was first derived by Ronald N. Bracewell in 1956 for a radio-astronomy problem.
The projection-slice theorem in N dimensions
In N dimensions, the projection-slice theorem states that the
Fourier tr |
https://en.wikipedia.org/wiki/Hyperbolic%20motion%20%28relativity%29 | Hyperbolic motion is the motion of an object with constant proper acceleration in special relativity. It is called hyperbolic motion because the equation describing the path of the object through spacetime is a hyperbola, as can be seen when graphed on a Minkowski diagram whose coordinates represent a suitable inertial (non-accelerated) frame. This motion has several interesting features, among them that it is possible to outrun a photon if given a sufficient head start, as may be concluded from the diagram.
History
Hermann Minkowski (1908) showed the relation between a point on a worldline and the magnitude of four-acceleration and a "curvature hyperbola" (). In the context of Born rigidity, Max Born (1909) subsequently coined the term "hyperbolic motion" () for the case of constant magnitude of four-acceleration, then provided a detailed description for charged particles in hyperbolic motion, and introduced the corresponding "hyperbolically accelerated reference system" (). Born's formulas were simplified and extended by Arnold Sommerfeld (1910). For early reviews see the textbooks by Max von Laue (1911, 1921) or Wolfgang Pauli (1921). See also Galeriu (2015) or Gourgoulhon (2013), and Acceleration (special relativity)#History.
Worldline
The proper acceleration of a particle is defined as the acceleration that a particle "feels" as it accelerates from one inertial reference frame to another. If the proper acceleration is directed parallel to the line of motion, it is re |
https://en.wikipedia.org/wiki/Glass%20electrode | A glass electrode is a type of ion-selective electrode made of a doped glass membrane that is sensitive to a specific ion. The most common application of ion-selective glass electrodes is for the measurement of pH. The pH electrode is an example of a glass electrode that is sensitive to hydrogen ions. Glass electrodes play an important part in the instrumentation for chemical analysis and physicochemical studies. The voltage of the glass electrode, relative to some reference value, is sensitive to changes in the activity of a certain type of ions.
History
The first studies of glass electrodes (GE) found different sensitivities of different glasses to change the medium's acidity (pH), due to the effects of the alkali metal ions.
In 1906, M. Cremer, the father of Erika Cremer, determined that the electric potential that arises between parts of the fluid, located on opposite sides of the glass membrane is proportional to the concentration of acid (hydrogen ion concentration).
In 1909, S. P. L. Sørensen introduced the concept of pH, and in the same year F. Haber and Z. Klemensiewicz reported results of their research on the glass electrode in The Society of Chemistry in Karlsruhe.
In 1922, W. S. Hughes showed that the alkali-silicate GE are similar to hydrogen electrodes, reversible concerning H+.
In 1925, P.M. Tookey Kerridge developed the first glass electrode for analysis of blood samples and highlighted some of the practical problems with the equipment such as t |
https://en.wikipedia.org/wiki/Bass%20Islands%20%28French%20Polynesia%29 | The Bass Islands ( or ) consist primarily of Rapa Iti () and Marotiri (). They are usually considered to be the southernmost of the Austral Islands, although this classification is more one of geographic and political expediency than because of similarities between them and the rest of the Austral Islands. The Bass Islands, lying several degrees outside the tropics, are the southernmost islands in French Polynesia.
Culturally, the Bass Islands appear to have been colonized about the same time as Tahiti and the Marquesas, and the culture and language (Rapan) appear to have diverged about the same time as well, indicating that they developed in relative isolation almost from the time of first settlement.
Geography
It has an area of approximately 40.6 square kilometers. Geologically, the Bass Islands are distinguished from the Austral Islands in that their vulcanism appears to be much more recent.
Rapa, sometimes called Rapa Iti (Little Rapa, to distinguish it from "Rapa Nui" (Big Rapa), a name for Easter Island), is the largest and only inhabited island of the Bass Islands. An older name for the island is Oparo. Its area is 38.5 km2 with a population of 530 and a max elevation of 650 m. Its main town is Ahuréi.
Marotiri, with an area of 0,043 km2, is a group of four uninhabited volcanic rocks protruding from the sea (and several submerged rocks), forming the southeastern end of the Austral Islands of French Polynesia. Marotiri is also known as Bass Rocks (Îlots de Bass i |
https://en.wikipedia.org/wiki/Retail%20Price%20Index | In the United Kingdom, the Retail Prices Index or Retail Price Index (RPI) is a measure of inflation published monthly by the Office for National Statistics. It measures the change in the cost of a representative sample of retail goods and services.
As the RPI was held not to meet international statistical standards, since 2013, the Office for National Statistics no longer classifies it as a "national statistic", emphasising the Consumer Price Index instead. However, as of 2018, the UK Treasury still uses the RPI measure of inflation for various index-linked tax rises.
History
RPI was first introduced in 1956, replacing the previous Interim Index of Retail Prices that had been in use since June 1947. It was once the principal official measure of inflation. It has been superseded in that regard by the Consumer Price Index (CPI).
The RPI is still used by the government as a base for various purposes, such as the amounts payable on index-linked securities, including index-linked gilts, and social housing rent increases. Many employers also use it as a starting point in wage negotiation. Since 2003, it is no longer used by the government for the inflation target for the Bank of England's Monetary Policy Committee nor, from April 2011, as the basis for the indexation of pensions of former public sector employees. , the UK state pension is indexed by the highest of the increase in average earnings, CPI or 2.5% ("the triple lock").
The highest annual inflation since the introdu |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.