source stringlengths 32 209 | text stringlengths 18 1.5k |
|---|---|
https://en.wikipedia.org/wiki/List%20of%20European%20countries%20by%20membership%20in%20international%20organisations | This list depicts the membership of European countries in selected international organisations and treaties.
Table
Classification of countries according to the United Nations:
Classification of countries according to the World Bank:
Classification of countries according to the IMF:
1 These countries are currently not participating in the EU's single market (EEA), but the EU has common external Customs Union agreements with Turkey (EU-Turkey Customs Union in force since 1995), Andorra (since 1991) and San Marino (since 2002). Monaco participates in the EU customs union through its relationship with France; its ports are administered by the French. Vatican City has a customs union in effect with Italy.
2 Monaco, San Marino and Vatican City are not members of Schengen, but act as such via their open borders with France and Italy, respectively.
3 Switzerland is not an official member of EEA but has bilateral agreements largely with the same content, making it virtually a member.
4 Andorra, Monaco, San Marino and Vatican City are using the euro as their currency through a monetary agreement with the EU.
5 Montenegro and Kosovo unilaterally adopted the euro as their currency and, therefore, have no issuing rights.
6 Currency included in the ERM II mechanism.
Euler diagram
Maps
Notes
See also
Co-ordinated organisations
Euronest Parliamentary Assembly
European integration
International organization
List of countries in Europe
List of European countries by life expectancy
Lis |
https://en.wikipedia.org/wiki/Coefficient%20of%20determination | In statistics, the coefficient of determination, denoted R2 or r2 and pronounced "R squared", is the proportion of the variation in the dependent variable that is predictable from the independent variable(s).
It is a statistic used in the context of statistical models whose main purpose is either the prediction of future outcomes or the testing of hypotheses, on the basis of other related information. It provides a measure of how well observed outcomes are replicated by the model, based on the proportion of total variation of outcomes explained by the model.
There are several definitions of R2 that are only sometimes equivalent. One class of such cases includes that of simple linear regression where r2 is used instead of R2. When only an intercept is included, then r2 is simply the square of the sample correlation coefficient (i.e., r) between the observed outcomes and the observed predictor values. If additional regressors are included, R2 is the square of the coefficient of multiple correlation. In both such cases, the coefficient of determination normally ranges from 0 to 1.
There are cases where R2 can yield negative values. This can arise when the predictions that are being compared to the corresponding outcomes have not been derived from a model-fitting procedure using those data. Even if a model-fitting procedure has been used, R2 may still be negative, for example when linear regression is conducted without including an intercept, or when a non-linear function is u |
https://en.wikipedia.org/wiki/Elastic%20fiber | Elastic fibers (or yellow fibers) are an essential component of the extracellular matrix composed of bundles of proteins (elastin) which are produced by a number of different cell types including fibroblasts, endothelial, smooth muscle, and airway epithelial cells. These fibers are able to stretch many times their length, and snap back to their original length when relaxed without loss of energy. Elastic fibers include elastin, elaunin and oxytalan.
Elastic fibers are formed via elastogenesis, a highly complex process involving several key proteins including fibulin-4, fibulin-5, latent transforming growth factor β binding protein 4, and microfibril associated protein 4. In this process tropoelastin, the soluble monomeric precursor to elastic fibers is produced by elastogenic cells and chaperoned to the cell surface. Following excretion from the cell, tropoelastin self associates into ~200 nm particles by coacervation, an entropically driven process involving interactions between tropoelastin's hydrophobic domains, which is mediated by glycosaminoglycans, heparan, and other molecules. These particles then fuse to give rise to 1-2 micron spherules which continue to grow as they move down from the cells surface before being deposited onto fibrillin microfibrillar scaffolds.
Following deposition onto microfibrils tropoelastin is insolubilized via extensive crosslinking by members of the lysyl oxidase and lysyl oxidase like family of copper-dependent amine oxidases into amorpho |
https://en.wikipedia.org/wiki/K-space | K-space or k-space can refer to:
Another name for the spatial frequency domain of a spatial Fourier transform
Reciprocal space, containing the reciprocal lattice of a spatial lattice
Momentum space, or wavevector space, the vector space of possible values of momentum for a particle
k-space (magnetic resonance imaging)
Another name for a compactly generated space in topology
K-space (functional analysis) is an F-space such that every twisted sum by the real line splits
K-Space (band), a British-Siberian music ensemble |
https://en.wikipedia.org/wiki/Montmorillonite | Montmorillonite is a very soft phyllosilicate group of minerals that form when they precipitate from water solution as microscopic crystals, known as clay. It is named after Montmorillon in France. Montmorillonite, a member of the smectite group, is a 2:1 clay, meaning that it has two tetrahedral sheets of silica sandwiching a central octahedral sheet of alumina. The particles are plate-shaped with an average diameter around 1 μm and a thickness of 0.96 nm; magnification of about 25,000 times, using an electron microscope, is required to resolve individual clay particles. Members of this group include saponite, nontronite, beidellite, and hectorite.
Montmorillonite is a subclass of smectite, a 2:1 phyllosilicate mineral characterized as having greater than 50% octahedral charge; its cation exchange capacity is due to isomorphous substitution of Mg for Al in the central alumina plane. The substitution of lower valence cations in such instances leaves the nearby oxygen atoms with a net negative charge that can attract cations. In contrast, beidellite is smectite with greater than 50% tetrahedral charge originating from isomorphous substitution of Al for Si in the silica sheet.
The individual crystals of montmorillonite clay are not tightly bound hence water can intervene, causing the clay to swell, hence montmorillonite is a characteristic component of swelling soil. The water content of montmorillonite is variable and it increases greatly in volume when it absorbs water. Che |
https://en.wikipedia.org/wiki/Crystal%20momentum | In solid-state physics crystal momentum or quasimomentum is a momentum-like vector associated with electrons in a crystal lattice. It is defined by the associated wave vectors of this lattice, according to
(where is the reduced Planck's constant).
Frequently, crystal momentum is conserved like mechanical momentum, making it useful to physicists and materials scientists as an analytical tool.
Lattice symmetry origins
A common method of modeling crystal structure and behavior is to view electrons as quantum mechanical particles traveling through a fixed infinite periodic potential such that
where is an arbitrary lattice vector. Such a model is sensible because crystal ions that form the lattice structure are typically on the order of tens of thousands of times more massive than electrons,
making it safe to replace them with a fixed potential structure, and the macroscopic dimensions of a crystal are typically far greater than a single lattice spacing, making edge effects negligible. A consequence of this potential energy function is that it is possible to shift the initial position of an electron by any lattice vector without changing any aspect of the problem, thereby defining a discrete symmetry. Technically, an infinite periodic potential implies that the lattice translation operator commutes with the Hamiltonian, assuming a simple kinetic-plus-potential form.
These conditions imply Bloch's theorem, which states
,
or that an electron in a lattice, which can be mod |
https://en.wikipedia.org/wiki/Receptor-mediated%20endocytosis | Receptor-mediated endocytosis (RME), also called clathrin-mediated endocytosis, is a process by which cells absorb metabolites, hormones, proteins – and in some cases viruses – by the inward budding of the plasma membrane (invagination). This process forms vesicles containing the absorbed substances and is strictly mediated by receptors on the surface of the cell. Only the receptor-specific substances can enter the cell through this process.
Process
Although receptors and their ligands can be brought into the cell through a few mechanisms (e.g. caveolin and lipid raft), clathrin-mediated endocytosis remains the best studied. Clathrin-mediated endocytosis of many receptor types begins with the ligands binding to receptors on the cell plasma membrane. The ligand and receptor will then recruit adaptor proteins and clathrin triskelions to the plasma membrane around where invagination will take place. Invagination of the plasma membrane then occurs, forming a clathrin-coated pit. Other receptors can nucleate a clathrin-coated pit allowing formation around the receptor. A mature pit will be cleaved from the plasma membrane through the use of membrane-binding and fission proteins such as dynamin (as well as other BAR domain proteins), forming a clathrin-coated vesicle that then uncoats of clathrin and typically fuses to a sorting endosome. Once fused, the endocytosed cargo (receptor and/or ligand) can then be sorted to lysosomal, recycling, or other trafficking pathways.
Function |
https://en.wikipedia.org/wiki/Rendering%20equation | In computer graphics, the rendering equation is an integral equation in which the equilibrium radiance leaving a point is given as the sum of emitted plus reflected radiance under a geometric optics approximation. It was simultaneously introduced into computer graphics by David Immel et al. and James Kajiya in 1986. The various realistic rendering techniques in computer graphics attempt to solve this equation.
The physical basis for the rendering equation is the law of conservation of energy. Assuming that L denotes radiance, we have that at each particular position and direction, the outgoing light (Lo) is the sum of the emitted light (Le) and the reflected light (Lr). The reflected light itself is the sum from all directions of the incoming light (Li) multiplied by the surface reflection and cosine of the incident angle.
Equation form
The rendering equation may be written in the form
where
is the total spectral radiance of wavelength directed outward along direction at time , from a particular position
is the location in space
is the direction of the outgoing light
is a particular wavelength of light
is time
is emitted spectral radiance
is reflected spectral radiance
is an integral over
is the unit hemisphere centered around containing all possible values for where
is the bidirectional reflectance distribution function, the proportion of light reflected from to at position , time , and at wavelength
is the negative direction of the incoming light
|
https://en.wikipedia.org/wiki/Two-sided%20Laplace%20transform | In mathematics, the two-sided Laplace transform or bilateral Laplace transform is an integral transform equivalent to probability's moment generating function. Two-sided Laplace transforms are closely related to the Fourier transform, the Mellin transform, the Z-transform and the ordinary or one-sided Laplace transform. If f(t) is a real- or complex-valued function of the real variable t defined for all real numbers, then the two-sided Laplace transform is defined by the integral
The integral is most commonly understood as an improper integral, which converges if and only if both integrals
exist. There seems to be no generally accepted notation for the two-sided transform; the
used here recalls "bilateral". The two-sided transform
used by some authors is
In pure mathematics the argument t can be any variable, and Laplace transforms are used to study how differential operators transform the function.
In science and engineering applications, the argument t often represents time (in seconds), and the function f(t) often represents a signal or waveform that varies with time. In these cases, the signals are transformed by filters, that work like a mathematical operator, but with a restriction. They have to be causal, which means that the output in a given time t cannot depend on an output which is a higher value of t.
In population ecology, the argument t often represents spatial displacement in a dispersal kernel.
When working with functions of time, f(t) is called the |
https://en.wikipedia.org/wiki/DAPI | DAPI (pronounced 'DAPPY', /ˈdæpiː/), or 4′,6-diamidino-2-phenylindole, is a fluorescent stain that binds strongly to adenine–thymine-rich regions in DNA. It is used extensively in fluorescence microscopy. As DAPI can pass through an intact cell membrane, it can be used to stain both live and fixed cells, though it passes through the membrane less efficiently in live cells and therefore provides a marker for membrane viability.
History
DAPI was first synthesised in 1971 in the laboratory of Otto Dann as part of a search for drugs to treat trypanosomiasis. Although it was unsuccessful as a drug, further investigation indicated it bound strongly to DNA and became more fluorescent when bound. This led to its use in identifying mitochondrial DNA in ultracentrifugation in 1975, the first recorded use of DAPI as a fluorescent DNA stain.
Strong fluorescence when bound to DNA led to the rapid adoption of DAPI for fluorescent staining of DNA for fluorescence microscopy. Its use for detecting DNA in plant, metazoa and bacteria cells and virus particles was demonstrated in the late 1970s, and quantitative staining of DNA inside cells was demonstrated in 1977. Use of DAPI as a DNA stain for flow cytometry was also demonstrated around this time.
Fluorescence properties
When bound to double-stranded DNA, DAPI has an absorption maximum at a wavelength of 358 nm (ultraviolet) and its emission maximum is at 461 nm (blue). Therefore, for fluorescence microscopy, DAPI is excited with ultravi |
https://en.wikipedia.org/wiki/Vishtaspa | Vishtaspa ( ; ; ; ) is the Avestan-language name of a figure of Zoroastrian scripture and tradition, portrayed as an early follower of Zoroaster, and his patron, and instrumental in the diffusion of the prophet's message. Although Vishtaspa is not epigraphically attested, he is – like Zoroaster – traditionally assumed to have been a historical figure, although obscured by accretions from legend and myth.
In Zoroastrian tradition, which builds on allusions found in the Avesta, Vishtaspa is a righteous king who helped propagate and defend the faith. In the non-Zoroastrian Sistan cycle texts, Vishtaspa is a loathsome ruler of the Kayanian dynasty who intentionally sends his eldest son to a certain death. In Greco-Roman literature, Zoroaster's patron was the pseudo-anonymous author of a set of prophecies written under his name.
In scripture
Vishtaspa is referred to in the Gathas, the oldest texts of Zoroastrianism which were considered to have been composed by Zoroaster himself. In these hymns, the poet speaks of Vishtaspa as his ally (Yasna 46. 14), follower of the path of Vohu Manah (Y. 51. 16), and committed to spreading the prophet's message (Y. 51. 16, 46. 15, 53. 2). In Yasna 28. 1–28. 7, Zoroaster appeals to Mazda for several boons, including the power to vanquish their foes for Vishtaspa and himself. Considered collectively, the Gathas celebrate Vishtaspa as the "patron of Zoroaster and the establisher of the first Zoroastrian community. "
The Gathic allusions rec |
https://en.wikipedia.org/wiki/Possibility%20theory | Possibility theory is a mathematical theory for dealing with certain types of uncertainty and is an alternative to probability theory. It uses measures of possibility and necessity between 0 and 1, ranging from impossible to possible and unnecessary to necessary, respectively. Professor Lotfi Zadeh first introduced possibility theory in 1978 as an extension of his theory of fuzzy sets and fuzzy logic. Didier Dubois and Henri Prade further contributed to its development. Earlier, in the 1950s, economist G. L. S. Shackle proposed the min/max algebra to describe degrees of potential surprise.
Formalization of possibility
For simplicity, assume that the universe of discourse Ω is a finite set. A possibility measure is a function from to [0, 1] such that:
Axiom 1:
Axiom 2:
Axiom 3: for any disjoint subsets and .
It follows that, like probability on finite probability spaces, the possibility measure is determined by its behavior on singletons:
Axiom 1 can be interpreted as the assumption that Ω is an exhaustive description of future states of the world, because it means that no belief weight is given to elements outside Ω.
Axiom 2 could be interpreted as the assumption that the evidence from which was constructed is free of any contradiction. Technically, it implies that there is at least one element in Ω with possibility 1.
Axiom 3 corresponds to the additivity axiom in probabilities. However there is an important practical difference. Possibility theory is computat |
https://en.wikipedia.org/wiki/Cyberpunk%20derivatives | Since the advent of the cyberpunk genre, a number of cyberpunk derivatives have become recognized in their own right as distinct subgenres in speculative fiction, especially in science fiction. Rather than necessarily sharing the digitally and mechanically focused setting of cyberpunk, these derivatives can display other futuristic, or even retrofuturistic, qualities that are drawn from or analogous to cyberpunk: a world built on one particular technology that is extrapolated to a highly sophisticated level (this may even be a fantastical or anachronistic technology, akin to retrofuturism), a gritty transreal urban style, or a particular approach to social themes.
Steampunk, one of the most well-known of these subgenres, has been defined as a "kind of technological fantasy;" others in this category sometimes also incorporate aspects of science fantasy and historical fantasy. Scholars have written of the stylistic place of these subgenres in postmodern literature, as well as their ambiguous interaction with the historical perspective of postcolonialism.
Background
American author Bruce Bethke coined the term cyberpunk in his 1983 short story of that name, using it as a label for a generation of "punk" teenagers inspired by the perceptions inherent to the Information Age. The term was quickly appropriated as a label applied to the works of William Gibson, Bruce Sterling, John Shirley, Rudy Rucker, Michael Swanwick, Pat Cadigan, Lewis Shiner, Richard Kadrey, and others. Scien |
https://en.wikipedia.org/wiki/Dale%20Connelly | Dale Connelly (born 4 October 1955) was co-host (with Tom Keith, aka Jim Ed Poole) of The Morning Show on Minnesota Public Radio. The program was first carried on KSJN at frequency 91.1 (now KNOW-FM, later on KSJN at 99.5 and finally on KCMP 89.3 "The Current"; all in the Twin Cities and live on MPR's classical music network in outstate Minnesota. Connelly took over for Garrison Keillor, who was the original host along with Tom Keith. The show aired for more than 30 years. On October 15, 2008, Keith announced his intention to retire on December 11. The Morning Show was discontinued after a final live performance at the Fitzgerald Theater in St. Paul that morning. After the demise of the Morning Show, Connelly began hosting, directing, and producing a show in a similar genre on an Internet broadcast and HD Radio called "Radio Heartland." He also hosted Saturday evening broadcasts of recorded performances on the Minnesota News service of Minnesota Public Radio. Those programs were canceled by MPR on June 4, 2010, and Connelly was laid off as of June 30, 2010.
He authored many scripts for "advertisements", skits, and soap operas performed on The Morning Show. Among his notable creations were "The Sherpa from Intimida", "A Mighty Big Car", and Genway, "the supermarket for genetically engineered foods." Genway produced things like the Bullwinkle Salmon ("part salmon, part moose") and Chameleon Pears.
Other Connelly creations included the "Bowserbed" vibrating sleeping mattress f |
https://en.wikipedia.org/wiki/Neuroglycopenia | Neuroglycopenia is a shortage of glucose (glycopenia) in the brain, usually due to hypoglycemia. Glycopenia affects the function of neurons, and alters brain function and behavior. Prolonged or recurrent neuroglycopenia can result in loss of consciousness, damage to the brain, and eventual death.
Signs and symptoms
Abnormal mentation, impaired judgment
Nonspecific dysphoria, anxiety, moodiness, depression, crying, fear of dying, suicidal thoughts
Negativism, irritability, belligerence, combativeness, rage
Personality change, emotional lability
Fatigue, weakness, apathy, lethargy, daydreaming
Confusion, amnesia, dizziness, delirium
Staring, "glassy" look, blurred vision, double vision
Automatic behavior
Difficulty speaking, slurred speech
Ataxia, incoordination, sometimes mistaken for "drunkenness"
Focal or general motor deficit, paralysis, hemiparesis
Paresthesia, headache
Stupor, coma, abnormal breathing
Generalized or focal seizures
Plasma glucose 20 mg/dL (1.1 mmol/L) or lower
Not all of the above manifestations occur in every case of hypoglycemia. There is no consistent order to the appearance of the symptoms. Specific manifestations vary by age and by the severity of the hypoglycemia. In older children and adults, moderately severe hypoglycemia can resemble mania, mental illness, drug intoxication, or drunkenness. In the elderly, hypoglycemia can produce focal stroke-like effects or a hard-to-define malaise. The symptoms of a single person do tend to be similar from ep |
https://en.wikipedia.org/wiki/Robert%20Ammann | Robert Ammann (October 1, 1946 – May, 1994) was an amateur mathematician who made several significant and groundbreaking contributions to the theory of quasicrystals and aperiodic tilings.
Ammann attended Brandeis University, but generally did not go to classes, and left after three years. He worked as a programmer for Honeywell. After twelve years, his position was eliminated as part of a routine cutback, and Ammann ended up working as a mail sorter for a post office.
In 1975, Ammann read an announcement by Martin Gardner of new work by Roger Penrose. Penrose had discovered two simple sets of aperiodic tiles, each consisting of just two quadrilaterals. Since Penrose was taking out a patent, he wasn't ready to publish them, and Gardner's description was rather vague. Ammann wrote a letter to Gardner, describing his own work, which duplicated one of Penrose's sets, plus a foursome of "golden rhombohedra" that formed aperiodic tilings in space.
More letters followed, and Ammann became a correspondent with many of the professional researchers. He discovered several new aperiodic tilings, each among the simplest known examples of aperiodic sets of tiles. He also showed how to generate tilings using lines in the plane as guides for lines marked on the tiles, now called "Ammann bars".
The discovery of quasicrystals in 1982 changed the status of aperiodic tilings and Ammann's work from mere recreational mathematics to respectable academic research.
After more than ten years of |
https://en.wikipedia.org/wiki/Albert%20Claude | Albert Claude (; 24 August 1899 – 22 May 1983) was a Belgian-American cell biologist and medical doctor who shared the Nobel Prize in Physiology or Medicine in 1974 with Christian de Duve and George Emil Palade. His elementary education started in a comprehensive primary school at Longlier, his birthplace. He served in the British Intelligence Service during the First World War, and got imprisoned in concentration camps twice. In recognition of his service, he was granted enrolment at the University of Liège in Belgium to study medicine without any formal education required for the course. He earned his Doctor of Medicine degree in 1928. Devoted to medical research, he initially joined German institutes in Berlin. In 1929 he found an opportunity to join the Rockefeller Institute in New York. At Rockefeller University he made his most groundbreaking achievements in cell biology. In 1930 he developed the technique of cell fractionation, by which he discovered the agent of the Rous sarcoma, components of cell organelles such as mitochondrion, chloroplast, endoplasmic reticulum, Golgi apparatus, ribosome and lysosome. He was the first to employ the electron microscope in the field of biology. In 1945 he published the first detailed structure of cell. His collective works established the complex functional and structural properties of cells.
Claude served as director at Jules Bordet Institute for Cancer Research and Treatment and Laboratoire de Biologie Cellulaire et Cancérologie |
https://en.wikipedia.org/wiki/Bell%20407 | The Bell 407 is a four-blade, single-engine, civil utility helicopter. A derivative of the Bell 206L-4 LongRanger, the 407 uses the four-blade, soft-in-plane design rotor with composite hub developed for the United States Army's OH-58D Kiowa Warrior instead of the two-blade, semi-rigid, teetering rotor of the 206L-4.
Design and development
In 1993, Bell began the development of the New Light Aircraft as a replacement for its Model 206 series. The program resulted in the 407, a development of Bell's LongRanger. A 206L-3 LongRanger was modified to serve as the 407 demonstrator. The demonstrator used hardware for the 407 and added molded fairings to represent the 407's wider fuselage then under development.
The demonstrator was first flown in 1994, and the 407 program was publicly announced at the Heli-Expo in Las Vegas, Nevada, in January 1995. The first 407 prototype (C-GFOS) accomplished its maiden flight on June 29, 1995, and the second prototype (C-FORS) followed on July 13, 1995. After a short development program, the first production 407 (C-FWQY/N407BT) flew on November 10, 1995.
The Bell 407 features the four-blade main rotor developed for the OH-58D (Model 406). The blades and hub use composite construction without life limits, and provide better performance and a more comfortable ride. The 407's fuselage is wider, increasing internal cabin space, and includes 35% larger main cabin windows. The more powerful Rolls-Royce/Allison 250-C47 turboshaft allows an increa |
https://en.wikipedia.org/wiki/Phosphosilicate%20glass | Phosphosilicate glass, commonly referred to by the acronym PSG, is a silicate glass commonly used in semiconductor device fabrication for intermetal layers, i.e., insulating layers deposited between succeedingly higher metal or conducting layers, due to its effect in gettering alkali ions. Another common type of phosphosilicate glass is borophosphosilicate glass (BPSG).
Soda-lime phosphosilicate glasses also form the basis for bioactive glasses (e.g. Bioglass), a family of materials which chemically convert to mineralised bone (hydroxy-carbonate-apatite) in physiological fluid.
Bismuth doped phosphosilicate glasses are being explored for use as the active gain medium in fiber lasers for fiber-optic communication.
See also
Wafer (electronics)
References
Glass compositions
Semiconductor device fabrication |
https://en.wikipedia.org/wiki/Crystal%20violet | Crystal violet or gentian violet, also known as methyl violet 10B or hexamethyl pararosaniline chloride, is a triarylmethane dye used as a histological stain and in Gram's method of classifying bacteria. Crystal violet has antibacterial, antifungal, and anthelmintic (vermicide) properties and was formerly important as a topical antiseptic. The medical use of the dye has been largely superseded by more modern drugs, although it is still listed by the World Health Organization.
The name gentian violet was originally used for a mixture of methyl pararosaniline dyes (methyl violet), but is now often considered a synonym for crystal violet. The name refers to its colour, being like that of the petals of certain gentian flowers; it is not made from gentians or violets.
Production
A number of possible routes can be used to prepare crystal violet. The original procedure developed by the German chemists Kern and Caro involved the reaction of dimethylaniline with phosgene to give 4,4′-bis(dimethylamino)benzophenone (Michler's ketone) as an intermediate. This was then reacted with additional dimethylaniline in the presence of phosphorus oxychloride and hydrochloric acid.
The dye can also be prepared by the condensation of formaldehyde and dimethylaniline to give a leuco dye:
CH2O + 3 C6H5N(CH3)2 → CH(C6H4N(CH3)2)3 + H2O
Second, this colourless compound is oxidized to the coloured cationic form (hereafter with oxygen, but a typical oxidizing agent is manganese dioxide, MnO2):
CH(C6H4N |
https://en.wikipedia.org/wiki/Epigenome | An epigenome consists of a record of the chemical changes to the DNA and histone proteins of an organism; these changes can be passed down to an organism's offspring via transgenerational stranded epigenetic inheritance. Changes to the epigenome can result in changes to the structure of chromatin and changes to the function of the genome.
The epigenome is involved in regulating gene expression, development, tissue differentiation, and suppression of transposable elements. Unlike the underlying genome, which remains largely static within an individual, the epigenome can be dynamically altered by environmental conditions.
Cancer
Epigenetics is a currently active topic in cancer research. Human tumors undergo a major disruption of DNA methylation and histone modification patterns. The aberrant epigenetic landscape of the cancer cell is characterized by a global genomic hypomethylation, CpG island promoter hypermethylation of tumor suppressor genes, an altered histone code for critical genes and a global loss of monoacetylated and trimethylated histone H4.
Epigenome research projects
As a prelude to a potential Human Epigenome Project, the Human Epigenome Pilot Project aims to identify and catalogue Methylation Variable Positions (MVPs) in the human genome. Advances in sequencing technology now allow for assaying genome-wide epigenomic states by multiple molecular methodologies. Micro- and nanoscale devices have been constructed or proposed to investigate the epigenome.
An |
https://en.wikipedia.org/wiki/Hoechst%20stain | Hoechst stains are part of a family of blue fluorescent dyes used to stain DNA. These bis-benzimides were originally developed by Hoechst AG, which numbered all their compounds so that the dye Hoechst 33342 is the 33,342nd compound made by the company. There are three related Hoechst stains: Hoechst 33258, Hoechst 33342, and Hoechst 34580. The dyes Hoechst 33258 and Hoechst 33342 are the ones most commonly used and they have similar excitation–emission spectra.
Molecular characteristics
Both dyes are excited by ultraviolet light at around 350 nm, and both emit blue-cyan fluorescent light around an emission spectrum maximum at 461 nm. Unbound dye has its maximum fluorescence emission in the 510–540 nm range. Hoechst stains can be excited with a xenon- or mercury-arc lamp or with an ultraviolet laser. There is a considerable Stokes shift between the excitation and emission spectra that makes Hoechst dyes useful in experiments in which multiple fluorophores are used. The fluorescence intensity of Hoechst dyes also increases with the pH of the solvent.
Hoechst dyes are soluble in water and in organic solvents such as dimethyl formamide or dimethyl sulfoxide. Concentrations can be achieved of up to 10 mg/mL. Aqueous solutions are stable at 2–6 °C for at least six months when protected from light. For longterm storage the solutions are instead frozen at −20 °C or below.
The dyes bind to the minor groove of double-stranded DNA with a preference for sequences rich in adenine and |
https://en.wikipedia.org/wiki/Reticular%20formation | The reticular formation is a set of interconnected nuclei that are located throughout the brainstem. It is not anatomically well defined, because it includes neurons located in different parts of the brain. The neurons of the reticular formation make up a complex set of networks in the core of the brainstem that extend from the upper part of the midbrain to the lower part of the medulla oblongata. The reticular formation includes ascending pathways to the cortex in the ascending reticular activating system (ARAS) and descending pathways to the spinal cord via the reticulospinal tracts.
Neurons of the reticular formation, particularly those of the ascending reticular activating system, play a crucial role in maintaining behavioral arousal and consciousness. The overall functions of the reticular formation are modulatory and premotor,
involving somatic motor control, cardiovascular control, pain modulation, sleep and consciousness, and habituation. The modulatory functions are primarily found in the rostral sector of the reticular formation and the premotor functions are localized in the neurons in more caudal regions.
The reticular formation is divided into three columns: raphe nuclei (median), gigantocellular reticular nuclei (medial zone), and parvocellular reticular nuclei (lateral zone). The raphe nuclei are the place of synthesis of the neurotransmitter serotonin, which plays an important role in mood regulation. The gigantocellular nuclei are involved in motor coordina |
https://en.wikipedia.org/wiki/WaveBird%20Wireless%20Controller | The WaveBird Wireless Controller (stylized as WAVEBIRD, commonly abbreviated as WaveBird or WaveBird controller) is a radio frequency-based wireless controller manufactured by Nintendo for use with the GameCube home video game console. Its name is a reference to Dolphin, the GameCube's codename during development. The WaveBird was available for purchase separately as well as in bundles with either Metroid Prime or Mario Party 4, which were exclusive to Kmart in the US.
Development
Nintendo had attempted to create a reliable wireless controller since the development of the Famicom. Its first attempt was for the Advanced Video System (AVS), the precursor to the Nintendo Entertainment System (NES), which included two wireless controllers but was never released.
Nintendo later developed an infrared (IR) adapter called the NES Satellite for the NES. Released in 1989, it used infrared to extend the length of up to four wired controllers, which would plug into the base of the unit rather than the console. The base could then be positioned anywhere within a certain range of the NES without the need for a cable. However, the extension base still needed a direct line of sight with the NES console; line of sight is a significant limitation of IR technology, requiring a clear space between an IR port and controller.
Radio Frequency controllers were not possible in the late 1980s as the early digital RF links were bulky and used too much power to be useful in battery-powered devices. H |
https://en.wikipedia.org/wiki/MULTI-S01 | In cryptography, MULTI-S01 (pronounced multi-ess-zero-one), is an encryption algorithm based on a pseudorandom number generator (PRNG). MULTI-S01 is an encryption scheme preserving both confidentiality and data integrity. The scheme defines a pair of algorithms; the encryption, the corresponding decryption with verification. Coupling with an efficient keystream generator, such as Panama, MUGI, and RC4, the algorithm efficiently encrypts a message in the manner of a single path process, i.e. online algorithm. The decryption function cannot be used in such manner for keeping whole resultant data until successful verification.
The keysize of MULTI-S01 is determined by which keystream generator to use. MULTI-S01 takes a security parameter which determines the upperbound probability of successful forgery.
Since the calculation consists of addition and multiplication over the finite field, the algorithm is more suited to hardware implementation, although software implementation is still feasible.
MULTI-S01 with the PRNG Panama was among the cryptographic techniques recommended for Japanese government use by CRYPTREC in 2003, however, has been dropped to "candidate" by CRYPTREC revision in 2013. It has also been submitted to ISO/IEC 18033 Part 4 which defines stream-cipher standards.
The security of MULTI-S01 is based on that of underlying PRNG. If a secure PRNG is used, then the security of MULTI-S01 with respect to confidentiality and data integrity has been proven. As for the |
https://en.wikipedia.org/wiki/Potez%20630 | The Potez 630 and its derivatives were a family of twin-engined, multirole aircraft developed for the French Air Force in the late 1930s. The design was a contemporary of the British Bristol Blenheim (which was larger and designed purely as a bomber) and the German Messerschmitt Bf 110 (which was designed purely as a fighter).
The Potez 630 was in use by several operators during the Second World War. Following the Battle of France, both the Vichy French Air Force and Free French Air Forces used the type; a number of captured aircraft were operated by several air wings of the Axis powers. After the end of the conflict in 1945, a handful of aircraft were used for training purposes for some time.
Development
Origins
On 31 October 1934, the French Ministry of Air issued a specification for a heavy fighter. The specification demanded the aircraft be capable of performing three principal roles: fighter direction, in which it was required to lead formations of single-engine fighters with sufficient maneuverability; day attack, in which the type was also to escort friendly close air support and bomber aircraft; and nightfighter operations. Specified performance details included a maximum speed of 450 km/h at 4,000 meters, a 300 km/h cruising speed, and an endurance of at least four hours. Armament requirements included two fixed forward-firing 20 mm cannons and a single machine gun to the rear for self-defence. The sought aircraft was also required to accommodate two/three seats a |
https://en.wikipedia.org/wiki/Classifier | Classifier may refer to:
Classifier (linguistics), or measure word, especially in East Asian languages
Classifier handshape, in sign languages
Classifier (UML), in software engineering
Classification rule, in statistical classification, e.g.:
Hierarchical classifier
Linear classifier
Deductive classifier
Subobject classifier, in category theory
An air classifier or similar machine for sorting materials
Classifier (machine learning)
See also
Finite-state machine#Classifiers
Classification (disambiguation)
Classified (disambiguation) |
https://en.wikipedia.org/wiki/Mixed-data%20sampling | Econometric models involving data sampled at different frequencies are of general interest. Mixed-data sampling (MIDAS) is an econometric regression developed by Eric Ghysels with several co-authors. There is now a substantial literature on MIDAS regressions and their applications, including Ghysels, Santa-Clara and Valkanov (2006), Ghysels, Sinko and Valkanov, Andreou, Ghysels and Kourtellos (2010) and Andreou, Ghysels and Kourtellos (2013).
MIDAS Regressions
A MIDAS regression is a direct forecasting tool which can relate future low-frequency data with current and lagged high-frequency indicators, and yield different forecasting models for each forecast horizon. It can flexibly deal with data sampled at different frequencies and provide a direct forecast of the low-frequency variable. It incorporates each individual high-frequency data in the regression, which solves the problems of losing potentially useful information and including mis-specification.
A simple regression example has the independent variable appearing at a higher frequency than the dependent variable:
where y is the dependent variable, x is the regressor, m denotes the frequency – for instance if y is yearly is quarterly – is the disturbance and is a lag distribution, for instance the Beta function or the Almon Lag. For example .
The regression models can be viewed in some cases as substitutes for the Kalman filter when applied in the context of mixed frequency data. Bai, Ghysels and Wright (2013) |
https://en.wikipedia.org/wiki/Demographics%20of%20Serbia | Demographic features of the population of Serbia include vital statistics, ethnicity, religious affiliations, education level, health of the populace, and other aspects of the population.
History
Censuses in Serbia ordinarily take place every 10 years, organized by the Statistical Office of the Republic of Serbia. The Principality of Serbia had conducted the first population census in 1834; the subsequent censuses were conducted in 1841, 1843, 1846, 1850, 1854, 1859, 1863 and 1866 and 1874. During the era Kingdom of Serbia, six censuses were conducted in 1884, 1890, 1895, 1900, 1905 and the last one being in 1910. During the Kingdom of Yugoslavia, censuses were conducted in 1931 and 1921; the census in 1941 was never conducted due to the outbreak of World War II. Socialist Yugoslavia conducted censuses in 1948, 1953, 1961, 1971, 1981, and 1991. The two most recent censuses were held in 2011 and 2022.
The years since the first 1834 Census saw frequent border changes of Serbia, first amidst the disintegration of the Ottoman Empire and Austria-Hungary, then subsequent formation and later disintegration of Yugoslavia and, finally, 2008 partially recognized independence of Kosovo which affected territorial scope in which all these censuses have been conducted.
Total fertility rate 1860–1949
The total fertility rate is the number of children born per woman. It is based on fairly good data for the entire period. Sources: Our World In Data and Gapminder Foundation.
Vital statist |
https://en.wikipedia.org/wiki/Cohort%20%28statistics%29 | In statistics, epidemiology, marketing and demography, a cohort is a group of subjects who share a defining characteristic (typically subjects who experienced a common event in a selected time period, such as birth or graduation).
Cohort data can oftentimes be more advantageous to demographers than period data. Because cohort data is honed to a specific time period, it is usually more accurate. It is more accurate because it can be tuned to retrieve custom data for a specific study.
In addition, cohort data is not affected by tempo effects, unlike period data. However, cohort data can be disadvantageous in the sense that it can take a long amount of time to collect the data necessary for the cohort study. Another disadvantage of cohort studies is that it can be extremely costly to carry out, since the study will go on for a long period of time, demographers often require sufficient funds to fuel the study.
Demography often contrasts cohort perspectives and period perspectives. For instance, the total cohort fertility rate is an index of the average completed family size for cohorts of women, but since it can only be known for women who have finished child-bearing, it cannot be measured for currently fertile women. It can be calculated as the sum of the cohort's age-specific fertility rates that obtain as it ages through time. In contrast, the total period fertility rate uses current age-specific fertility rates to calculate the completed family size for a notional woman, |
https://en.wikipedia.org/wiki/Irnerius | Irnerius (– after 1125), sometimes referred to as lucerna juris ("lantern of the law"), was an Italian jurist, and founder of the School of Glossators and thus of the tradition of Medieval Roman Law.
He taught the newly recovered Roman lawcode of Justinian I, the Corpus Juris Civilis, among the liberal arts at the University of Bologna, his native city. The recovery and revival of Roman law, taught first at Bologna in the 1070s, was a momentous event in European cultural history. Irnerius' interlinear glosses on the Corpus Juris Civilis stand at the beginnings of a European law that was written, systematic, comprehensive and rational, and based on Roman law.
Life
He was born in Bologna about 1050.
At the urging of Countess Matilda of Tuscany he began to devote himself to the study of jurisprudence, taking the Justinian code as a guide. After teaching jurisprudence for a short while in Rome he returned to Bologna, where he founded a new school of jurisprudence in 1084 or 1088, which would rival the law school of Ravenna.
Some jurisprudence had been taught at Bologna, before Irnerius founded his school, by Pepo and a few others, and a tradition of jurisprudence had developed at Pavia since the mid-ninth century. He introduced the custom of explaining the Roman law by means of glosses, which originally were meagre interlinear elucidations of the text. But since the glosses were often too extensive to be inserted between the lines of the text, he began to write them on the m |
https://en.wikipedia.org/wiki/Anthropoid | Anthropoid means 'ape/human feature' and may refer to:
Simian, monkeys and apes (anthropoids, or suborder Anthropoidea, in earlier classifications)
Anthropoid apes, apes that are closely related to humans (e.g., former family Pongidae and sometimes also Hylobatidae and their extinct relatives)
Anthropoides, a genus of cranes
Operation Anthropoid, the codename for the assassination of Reinhard Heydrich, SS-Obergruppenführer and Reichsprotektor of Bohemia and Moravia
Operation Anthropoid Memorial, Libeň, Prague, Czech Republic
Anthropoid (film), a 2016 film based on Operation Anthropoid
In pelvimetry, one of four types of human female pelvis
Anthropoid animals, fictional creatures in the Japanese visual novel Wanko to Kurasō
Anthropoid robots, mostly referred to as androids meaning human-like robots
See also
Anthropology (disambiguation)
Anthropod (disambiguation)
Humanoid (disambiguation) |
https://en.wikipedia.org/wiki/Conexant | Conexant Systems, Inc. was an American-based software developer and fabless semiconductor company that developed technology for voice and audio processing, imaging and modems. The company began as a division of Rockwell International, before being spun off as a public company. Conexant itself then spun off several business units, creating independent public companies which included Skyworks Solutions and Mindspeed Technologies.
The company was acquired by computing interface technology company Synaptics, Inc. in July 2017.
History
In 1996, Rockwell International Corporation incorporated its semiconductor division as Rockwell Semiconductor Systems, Inc. On January 4, 1999, Rockwell spun off Conexant Systems, Inc. as a public company. It was listed on the NASDAQ under symbol CNXT on January 4, 1999. At that time, Conexant became the world's largest, standalone communications-IC company. Dwight W. Decker was its first chief executive officer and chairman of its board of directors. The company was based in Newport Beach, California.
In the early 2000s, Conexant spun off several standalone technology businesses to create public companies.
In March 2002, Conexant entered into a joint venture agreement with The Carlyle Group to share ownership of its wafer fabrication plant, called Jazz Semiconductor.
In June 2002, Conexant spun off its wireless communications division, which merged immediately following the spinoff with Massachusetts-based chip manufacturer Alpha Indus |
https://en.wikipedia.org/wiki/Order%20of%20a%20kernel | In statistics, the order of a kernel is the degree of the first non-zero moment of a kernel.
Definitions
The literature knows two major definitions of the order of a kernel:
Definition 1
Let be an integer. Then, is a kernel of order if the functions are integrable and satisfy
Definition 2
References
Nonparametric statistics |
https://en.wikipedia.org/wiki/Reduced%20form | In statistics, and particularly in econometrics, the reduced form of a system of equations is the result of solving the system for the endogenous variables. This gives the latter as functions of the exogenous variables, if any. In econometrics, the equations of a structural form model are estimated in their theoretically given form, while an alternative approach to estimation is to first solve the theoretical equations for the endogenous variables to obtain reduced form equations, and then to estimate the reduced form equations.
Let Y be the vector of the variables to be explained (endogeneous variables) by a statistical model and X be the vector of explanatory (exogeneous) variables. In addition let be a vector of error terms. Then the general expression of a structural form is , where f is a function, possibly from vectors to vectors in the case of a multiple-equation model. The reduced form of this model is given by , with g a function.
Structural and reduced forms
Exogenous variables are variables which are not determined by the system. If we assume that demand is influenced not only by price, but also by an exogenous variable, Z, we can consider the structural supply and demand model
supply:
demand:
where the terms are random errors (deviations of the quantities supplied and demanded from those implied by the rest of each equation). By solving for the unknowns (endogenous variables) P and Q, this structural model can be rewritten in the reduced form:
|
https://en.wikipedia.org/wiki/Tipp-Ex | Tipp-Ex is a brand of correction fluid and other related products that is popular throughout Europe. It was also the name of the German company (Tipp-Ex GmbH & Co. KG) that produced the products in the Tipp-Ex line. While Tipp-Ex is a trademark name for correction products, in some countries it has become a genericised trademark: to tippex or to tippex out means to erase, either generally or with correction fluid.
History
Tipp-Ex correction paper was invented by Wolfgang Dabisch from Eltville, West Germany, who filed a patent in 1958 on Colored film for the correction of typing errors (German: Tippfehler). He subsequently founded a company of the same name. Shortly after that a Tipp-Ex Sales & Distribution company (Tipp-Ex Vertrieb GmbH & Co. KG) was founded in Frankfurt by Otto Carls. This company still exists under the name of Tipp-Ex GmbH & Co. KG close to Frankfurt.
Tipp-Ex became a registered trademark with the German patent office in 1987.
Earlier, in 1951, Bette Nesmith Graham invented the first correction fluid in her kitchen and began marketing the product in 1956 as Mistake Out. Tipp-Ex GmbH only started to produce white correction fluid in 1965 under the brand Tipp-Ex, but also as C-fluid.
As a result of the invention of Tipp-Ex, it became possible to erase a typographical error made on a typewriter. The typewriter would be backspaced to the letter that was to be changed, the correction paper would be placed behind the ribbon, and the mistyped letter would be r |
https://en.wikipedia.org/wiki/Autoregressive%20integrated%20moving%20average | In statistics and econometrics, and in particular in time series analysis, an autoregressive integrated moving average (ARIMA) model is a generalization of an autoregressive moving average (ARMA) model. To better comprehend the data or to forecast upcoming series points, both of these models are fitted to time series data. ARIMA models are applied in some cases where data show evidence of non-stationarity in the sense of mean (but not variance/autocovariance), where an initial differencing step (corresponding to the "integrated" part of the model) can be applied one or more times to eliminate the non-stationarity of the mean function (i.e., the trend). When the seasonality shows in a time series, the seasonal-differencing could be applied to eliminate the seasonal component. Since the ARMA model, according to the Wold's decomposition theorem, is theoretically sufficient to describe a regular (a.k.a. purely nondeterministic) wide-sense stationary time series, we are motivated to make stationary a non-stationary time series, e.g., by using differencing, before we can use the ARMA model. Note that if the time series contains a predictable sub-process (a.k.a. pure sine or complex-valued exponential process), the predictable component is treated as a non-zero-mean but periodic (i.e., seasonal) component in the ARIMA framework so that it is eliminated by the seasonal differencing.
The autoregressive () part of ARIMA indicates that the evolving variable of interest is regressed on |
https://en.wikipedia.org/wiki/Bayesian%20search%20theory | Bayesian search theory is the application of Bayesian statistics to the search for lost objects. It has been used several times to find lost sea vessels, for example USS Scorpion, and has played a key role in the recovery of the flight recorders in the Air France Flight 447 disaster of 2009. It has also been used in the attempts to locate the remains of Malaysia Airlines Flight 370.
Procedure
The usual procedure is as follows:
Formulate as many reasonable hypotheses as possible about what may have happened to the object.
For each hypothesis, construct a probability density function for the location of the object.
Construct a function giving the probability of actually finding an object in location X when searching there if it really is in location X. In an ocean search, this is usually a function of water depth — in shallow water chances of finding an object are good if the search is in the right place. In deep water chances are reduced.
Combine the above information coherently to produce an overall probability density map. (Usually this simply means multiplying the two functions together.) This gives the probability of finding the object by looking in location X, for all possible locations X. (This can be visualized as a contour map of probability.)
Construct a search path which starts at the point of highest probability and 'scans' over high probability areas, then intermediate probabilities, and finally low probability areas.
Revise all the probabilities continuousl |
https://en.wikipedia.org/wiki/Desorption | Desorption is the physical process where adsorbed atoms or molecules are released from a surface into the surrounding vacuum or fluid. This occurs when a molecule gains enough energy to overcome the activation barrier and the binding energy that keep it attached to the surface.
Desorption is the reverse of the process of adsorption, which differs from absorption in that adsorption it refers to substances bound to the surface, rather than being absorbed into the bulk.
Desorption can occur from any of several processes, or a combination of them: it may result from heat (thermal energy); incident light such as infrared, visible, or ultraviolet photons; or a incident beam of energetic particles such as electrons. It may also occur following chemical reactions such as oxidation or reduction in an electrochemical cell or after a chemical reaction of a adsorbed compounds in which the surface may act as a catalyst.
Desorption mechanisms
Depending on the nature of the adsorbent-to-surface bond, there are a multitude of mechanisms for desorption. The surface bond of a sorbant can be cleaved thermally, through chemical reactions or by radiation, all which may result in desorption of the species.
Thermal desorption
Thermal desorption is the process by which an adsorbate is heated and this induces desorption of atoms or molecules from the surface. The first use of thermal desorption was by LeRoy Apker in 1948. It is one of the most frequently used modes of desorption, and can be |
https://en.wikipedia.org/wiki/Hankinson | Hankinson may refer to:
Hankinson (surname)
Hankinson, North Dakota, a city in Richland County, North Dakota, United States
Hankinson's equation, an equation for predicting the strength of wood
Hankinson-Moreau-Covenhoven House, a house located in Freehold, New Jersey, United States
Lake Hankinson, a lake within the catchment of the Waiau River, New Zealand |
https://en.wikipedia.org/wiki/Wald%27s%20equation | In probability theory, Wald's equation, Wald's identity or Wald's lemma is an important identity that simplifies the calculation of the expected value of the sum of a random number of random quantities. In its simplest form, it relates the expectation of a sum of randomly many finite-mean, independent and identically distributed random variables to the expected number of terms in the sum and the random variables' common expectation under the condition that the number of terms in the sum is independent of the summands.
The equation is named after the mathematician Abraham Wald. An identity for the second moment is given by the Blackwell–Girshick equation.
Basic version
Let be a sequence of real-valued, independent and identically distributed random variables and let be an integer-valued random variable that is independent of the sequence . Suppose that and the have finite expectations. Then
Example
Roll a six-sided dice. Take the number on the die (call it ) and roll that number of six-sided dice to get the numbers , and add up their values. By Wald's equation, the resulting value on average is
General version
Let be an infinite sequence of real-valued random variables and let be a nonnegative integer-valued random variable.
Assume that:
. are all integrable (finite-mean) random variables,
. for every natural number , and
. the infinite series satisfies
Then the random sums
are integrable and
If, in addition,
. all have the same expectation, and
. has |
https://en.wikipedia.org/wiki/Germplasm | Germplasm are genetic resources such as seeds, tissues, and DNA sequences that are maintained for the purpose of animal and plant breeding, conservation efforts, agriculture, and other research uses. These resources may take the form of seed collections stored in seed banks, trees growing in nurseries, animal breeding lines maintained in animal breeding programs or gene banks. Germplasm collections can range from collections of wild species to elite, domesticated breeding lines that have undergone extensive human selection. Germplasm collection is important for the maintenance of biological diversity, food security, and conservation efforts.
In the United States, germplasm resources are regulated by the National Genetic Resources Program (NGRP), created by the U.S. congress in 1990. In addition the web server The Germplasm Resources Information Network (GRIN) provides information about germplasms as they pertain to agriculture production.
Germplasm Regulation
In the United States, germplasm resources are regulated by the National Genetic Resources Program (NGRP), created by the U.S. congress in 1990. In addition the web server The Germplasm Resources Information Network (GRIN) provides information about germplasms as they pertain to agriculture production.
Specifically for plants, there is the U.S. National Plant Germplasm System (NPGS) which holds > 450,000 accessions with 10,000 species of the 85 most commonly grown crops. Many accessions held are international species, |
https://en.wikipedia.org/wiki/Fin%20field-effect%20transistor | A fin field-effect transistor (FinFET) is a multigate device, a MOSFET (metal–oxide–semiconductor field-effect transistor) built on a substrate where the gate is placed on two, three, or four sides of the channel or wrapped around the channel, forming a double or even multi gate structure. These devices have been given the generic name "FinFETs" because the source/drain region forms fins on the silicon surface. The FinFET devices have significantly faster switching times and higher current density than planar CMOS (complementary metal–oxide–semiconductor) technology.
FinFET is a type of non-planar transistor, or "3D" transistor. It is the basis for modern nanoelectronic semiconductor device fabrication. Microchips utilizing FinFET gates first became commercialized in the first half of the 2010s, and became the dominant gate design at 14 nm, 10 nm and 7 nm process nodes.
It is common for a single FinFET transistor to contain several fins, arranged side by side and all covered by the same gate, that act electrically as one, to increase drive strength and performance.
History
After the MOSFET was first demonstrated by Mohamed Atalla and Dawon Kahng of Bell Labs in 1960, the concept of a double-gate thin-film transistor (TFT) was proposed by H. R. Farrah (Bendix Corporation) and R. F. Steinberg in 1967. A double-gate MOSFET was later proposed by Toshihiro Sekigawa of the Electrotechnical Laboratory (ETL) in a 1980 patent describing the planar XMOS transistor. Sekigawa fabricat |
https://en.wikipedia.org/wiki/The%20First%20Generation%20of%20Postwar%20Writers | The First Generation of Postwar Writers''' is a classification in Modern Japanese literature used to group writers who appeared on the postwar literary scene between 1946 and 1947.
List of First Generation writers
Haniya Yutaka (埴谷雄高)
Nakamura Shin'ichirō (中村真一郎)
Noma Hiroshi (野間宏)
Shiina Rinzō (椎名麟三)
Takeda Taijun (武田泰淳)
Umezaki Haruo (梅崎春生)
Background of the Post-War Literature in Japan
During the beginning of the post-war period in Japan, the revolution of post-war literature in Japan became modern democratic as "Democracy", "Freedom", "class", and "individual". However, the influence of the emperor system made the revolution of post-war literature of Japan become contra-democratic. Therefore, the post-war literature in Japan had transferred to the management under the imperial institution of Japan.
Characteristics and Significance of the Post-War Literature
During the period post-war in Japan, trama was one of the representations as characteristics in literature and movie. During the literature in post-war Japan, the narration usually would be considered as the view of the "victim" in the war between Japan and other countries. Meanwhile, the reason for creating the characteristics of trauma and victim was the expectation to separate the past and present in Japanese military history.
Besides the characteristics of trauma and victim. The influence of the post-war literature in Japan can be demonstrated as "Body", "Individual", and "National Identity".
The body |
https://en.wikipedia.org/wiki/Second%20Generation%20of%20Postwar%20Writers | The is a classification in modern Japanese literature used for writers who appeared on the postwar literary scene between 1948 and 1949.
Exceptional in this generation of postwar writers are Mishima Yukio and Abe Kōbō, both of whom have received acclaim in Japan and abroad. At times, their reputation abroad has surpassed that of their reputation in Japan.
List of Second Generation writers
Mishima Yukio (三島由紀夫)
Abe Kōbō (安部公房)
Ōoka Shōhei (大岡昇平)
Shimao Toshio (島尾敏雄)
Hotta Yoshie (堀田善衛)
Inoue Mitsuharu (井上光晴)
See also
Japanese literature
The First Generation of Postwar Writers
The Third Generation of Postwar Writers
Second Generation of Postwar Writers, The
Japanese literary movements
Postwar Japan
20th-century Japanese literature |
https://en.wikipedia.org/wiki/Third%20Generation%20of%20Postwar%20Writers | The Third Generation of Postwar Writers (第三の新人, daisan no shinjin) is a classification in Modern Japanese literature used to group writers who appeared on the postwar literary scene between 1953 and 1955.
Shūsaku Endō, a member of the Third Generation once said, "In those days, although we had received the Akutagawa Prize one after another, hardly did anyone expect that we would become great writers. We were regarded as if we would soon be forgotten by the literary world. Precisely, almost all people did not start to know Akutagawa Prize until Ishihara Shintaro had won the prize and surfed away mass media and provoked public opinion into asunder, as the first manifesto from one of the Postwar Generation."
However, despite this, this generation has made a major mark on Japanese literature. The works of Endō in particularly have been translated into many languages and are widely read in the United States, France, and Germany.
At that same time, women writers such as Aya Kōda (幸田文), Minako Oba, and Sawako Ariyoshi also made their debuts.
After this generation, predominant and various writers like Shintaro Ishihara, Morio Kita, and Kenzaburō Ōe appeared.
List
Shūsaku Endō (遠藤周作)
Shōtarō Yasuoka (安岡章太郎)
Junnosuke Yoshiyuki (吉行淳之介)
Junzo Shono (庄野潤三)
Shumon Miura (三浦朱門)
Ayako Sono (曽野綾子)
Hiroyuki Agawa (阿川弘之)
Kojima Nobuo (小島信夫)
See also
Japanese literature
The First Generation of Postwar Writers
The Second Generation of Postwar Writers
Japanese literature
Litera |
https://en.wikipedia.org/wiki/Volvo%20B18%20engine | The B18 is a 1.8 L inline four cylinder automobile engine produced by Volvo from 1961 through 1968. A larger 2.0 L derivative called the B20 debuted in 1969.
Despite being a pushrod design, the engines can rev to 6,500 rpm. They are also reputed to be very durable. The world's highest mileage car, a 1966 Volvo P1800S, traveled more than on its original B18 engine.
B18
The B18 has a single cam-in-block, operating two overhead valves (OHV) per cylinder by pushrods and rocker arms. The crankshaft rides in five main bearings, making the B18 quite different in design from its predecessor, the three-bearing B16.
With a bore of and stroke of , the B18 displaces . The engine was used in Volvo's PV544, P210 Duett, 120 (Amazon), P1800 and 140 series. It could also be found in the L3314 and the Bandvagn 202 military vehicles. The B18 was fitted to many Volvo Penta sterndrive marine propulsion systems. It was also used in the Facel Vega Facel III and the Marcos 1800 GT.
There are four variations of this engine:
B18A: Single carburettor version.
B18B: Dual carburettor version with a higher compression ratio, fitted variously with dual sidedraft SU or Zenith/Stromberg carburettors.
B18C: Single carburettor version with a lower compression ratio and mechanical RPM regulator, fitted in the gasoline powered versions of the Volvo BM 320 tractor. This version was also used for the elevator in the PS-15 radar system.
B18D Dual carburettor version with a lower compression ratio.
DOH |
https://en.wikipedia.org/wiki/Helmert%E2%80%93Wolf%20blocking | The Helmert–Wolf blocking (HWB) is a least squares solution method for the solution of a sparse block system of linear equations. It was first reported by F. R. Helmert for use in geodesy problems in 1880; (1910–1994) published his direct semianalytic solution in 1978.
It is based on ordinary Gaussian elimination in matrix form or partial minimization form.
Description
Limitations
The HWB solution is very fast to compute but it is optimal only if observational errors do not correlate between the data blocks. The generalized canonical correlation analysis (gCCA) is the statistical method of choice for making those harmful cross-covariances vanish. This may, however, become quite tedious depending on the nature of the problem.
Applications
The HWB method is critical to satellite geodesy and similar large problems. The HWB method can be extended to fast Kalman filtering (FKF) by augmenting its linear regression equation system to take into account information from numerical forecasts, physical constraints and other ancillary data sources that are available in realtime. Operational accuracies can then be computed reliably from the theory of minimum-norm quadratic unbiased estimation (Minque) of C. R. Rao.
See also
Block matrix
Notes
Statistical algorithms
Least squares
Geodesy |
https://en.wikipedia.org/wiki/Michael%20I.%20Jordan | Michael Irwin Jordan (born February 25, 1956) is an American scientist, professor at the University of California, Berkeley and researcher in machine learning, statistics, and artificial intelligence.
Jordan was elected a member of the National Academy of Engineering in 2010 for contributions to the foundations and applications of machine learning.
He is one of the leading figures in machine learning, and in 2016 Science reported him as the world's most influential computer scientist.
In 2022, Jordan won the inaugural World Laureates Association Prize in Computer Science or Mathematics, "for fundamental contributions to the foundations of machine learning and its application."
Education
Jordan received his BS magna cum laude in Psychology in 1978 from the Louisiana State University, his MS in Mathematics in 1980 from Arizona State University and his PhD in Cognitive Science in 1985 from the University of California, San Diego. At the University of California, San Diego, Jordan was a student of David Rumelhart and a member of the Parallel Distributed Processing (PDP) Group in the 1980s.
Career and research
Jordan is the Pehong Chen Distinguished Professor at the University of California, Berkeley, where his appointment is split across EECS and Statistics. He was a professor at the Department of Brain and Cognitive Sciences at MIT from 1988 to 1998.
In the 1980s Jordan started developing recurrent neural networks as a cognitive model. In recent years, his work is less |
https://en.wikipedia.org/wiki/Aldolase%20A | Aldolase A (ALDOA, or ALDA), also known as fructose-bisphosphate aldolase, is an enzyme that in humans is encoded by the ALDOA gene on chromosome 16.
The protein encoded by this gene is a glycolytic enzyme that catalyzes the reversible conversion of fructose-1,6-bisphosphate to glyceraldehyde 3-phosphate (G3P) and dihydroxyacetone phosphate (DHAP). Three aldolase isozymes (A, B, and C), encoded by three different genes, are differentially expressed during development. Aldolase A is found in the developing embryo and is produced in even greater amounts in adult muscle. Aldolase A expression is repressed in adult liver, kidney and intestine and similar to aldolase C levels in brain and other nervous tissue. Aldolase A deficiency has been associated with myopathy and hemolytic anemia. Alternative splicing and alternative promoter usage results in multiple transcript variants. Related pseudogenes have been identified on chromosomes 3 and 10. [provided by RefSeq, Aug 2011]
Structure
ALDOA is a homotetramer and one of the three aldolase isozymes (A, B, and C), encoded by three different genes. The ALDOA gene contains 8 exons and the 5' UTR IB. Key amino acids responsible for its catalytic function have been identified. The residue Tyr363 functions as the acid–base catalyst for protonating C3 of the substrate, while Lys146 is proposed to stabilize the negative charge of the resulting conjugate base of Tyr363 and the strained configuration of the C-terminal. Residue Glu187 particip |
https://en.wikipedia.org/wiki/Defining%20length | In genetic algorithms and genetic programming defining length L(H) is the maximum distance between two defining symbols (that is symbols that have a fixed value as opposed to symbols that can take any value, commonly denoted as # or *) in schema H. In tree GP schemata, L(H) is the number of links in the minimum tree fragment including all the non-= symbols within a schema H.
Example
Schemata "00##0", "1###1", "01###", and "##0##" have defining lengths of 4, 4, 1, and 0, respectively. Lengths are computed by determining the last fixed position and subtracting from it the first fixed position.
In genetic algorithms as the defining length of a solution increases so does the susceptibility of the solution to disruption due to mutation or cross-over.
References
Genetic algorithms |
https://en.wikipedia.org/wiki/Stress-induced%20leakage%20current | Stress-induced leakage current (SILC) is an increase in the gate leakage current of a MOSFET, used in semiconductor physics. It occurs due to defects created in the gate oxide during electrical stressing. SILC is perhaps the largest factor inhibiting device miniaturization. Increased leakage is a common failure mode of electronic devices.
Oxide defects
The most well-studied defects assisting in the leakage current are those produced by charge trapping in the oxide. This model provides a point of attack and has stimulated researchers to develop methods to decrease the rate of charge trapping by mechanisms such as nitrous oxide (N2O) nitridation of the oxide.
Semiconductor device defects |
https://en.wikipedia.org/wiki/Training%2C%20validation%2C%20and%20test%20data%20sets | In machine learning, a common task is the study and construction of algorithms that can learn from and make predictions on data. Such algorithms function by making data-driven predictions or decisions, through building a mathematical model from input data. These input data used to build the model are usually divided into multiple data sets. In particular, three data sets are commonly used in different stages of the creation of the model: training, validation, and test sets.
The model is initially fit on a training data set, which is a set of examples used to fit the parameters (e.g. weights of connections between neurons in artificial neural networks) of the model. The model (e.g. a naive Bayes classifier) is trained on the training data set using a supervised learning method, for example using optimization methods such as gradient descent or stochastic gradient descent. In practice, the training data set often consists of pairs of an input vector (or scalar) and the corresponding output vector (or scalar), where the answer key is commonly denoted as the target (or label). The current model is run with the training data set and produces a result, which is then compared with the target, for each input vector in the training data set. Based on the result of the comparison and the specific learning algorithm being used, the parameters of the model are adjusted. The model fitting can include both variable selection and parameter estimation.
Successively, the fitted model is use |
https://en.wikipedia.org/wiki/Instrumental%20variables%20estimation | In statistics, econometrics, epidemiology and related disciplines, the method of instrumental variables (IV) is used to estimate causal relationships when controlled experiments are not feasible or when a treatment is not successfully delivered to every unit in a randomized experiment. Intuitively, IVs are used when an explanatory variable of interest is correlated with the error term, in which case ordinary least squares and ANOVA give biased results. A valid instrument induces changes in the explanatory variable but has no independent effect on the dependent variable, allowing a researcher to uncover the causal effect of the explanatory variable on the dependent variable.
Instrumental variable methods allow for consistent estimation when the explanatory variables (covariates) are correlated with the error terms in a regression model. Such correlation may occur when:
changes in the dependent variable change the value of at least one of the covariates ("reverse" causation),
there are omitted variables that affect both the dependent and explanatory variables, or
the covariates are subject to non-random measurement error.
Explanatory variables that suffer from one or more of these issues in the context of a regression are sometimes referred to as endogenous. In this situation, ordinary least squares produces biased and inconsistent estimates. However, if an instrument is available, consistent estimates may still be obtained. An instrument is a variable that does not itself |
https://en.wikipedia.org/wiki/Mating%20pool | A mating pool is a concept used in evolutionary computation, which refers to a family of algorithms used to solve optimization and search problems.
The mating pool is formed by candidate solutions that the selection operators deem to have the highest fitness in the current population. Solutions that are included in the mating pool are referred to as parents. Individual solutions can be repeatedly included in the mating pool, with individuals of higher fitness values having a higher chance of being included multiple times. Crossover operators are then applied to the parents, resulting in recombination of genes recognized as superior. Lastly, random changes in the genes are introduced through mutation operators, increasing the genetic variation in the gene pool. Those two operators improve the chance of creating new, superior solutions. A new generation of solutions is thereby created, the children, who will constitute the next population. Depending on the selection method, the total number of parents in the mating pool can be different to the size of the initial population, resulting in a new population that’s smaller. To continue the algorithm with an equally sized population, random individuals from the old populations can be chosen and added to the new population.
At this point, the fitness value of the new solutions is evaluated. If the termination conditions are fulfilled, processes come to an end. Otherwise, they are repeated.
The repetition of the steps result in can |
https://en.wikipedia.org/wiki/Premature%20convergence | In evolutionary algorithms (EA), the term of premature convergence means that a population for an optimization problem converged too early, resulting in being suboptimal. In this context, the parental solutions, through the aid of genetic operators, are not able to generate offspring that are superior to, or outperform, their parents. Premature convergence is a common problem found in evolutionary algorithms in general and genetic algorithms in particular, as it leads to a loss, or convergence of, a large number of alleles, subsequently making it very difficult to search for a specific gene in which the alleles were present. An allele is considered lost if, in a population, a gene is present, where all individuals are sharing the same value for that particular gene. An allele is, as defined by De Jong, considered to be a converged allele, when 95% of a population share the same value for a certain gene (see also convergence).
Strategies for preventing premature convergence
Strategies to regain genetic variation can be:
a mating strategy called incest prevention,
uniform crossover,
favored replacement of similar individuals (preselection or crowding),
segmentation of individuals of similar fitness (fitness sharing),
increasing population size.
The genetic variation can also be regained by mutation though this process is highly random.
One way to reduce the risk of premature convergence is to use structured populations instead of the commonly used panmictic ones, see b |
https://en.wikipedia.org/wiki/Brown%20note | The brown note, also sometimes called the brown frequency or brown noise, is a hypothetical infrasonic frequency capable of causing fecal incontinence by creating acoustic resonance in the human bowel. Considered an urban myth, the name is a metonym for the common color of human faeces. Attempts to demonstrate the existence of a "brown note" using sound waves transmitted through the air have failed.
Frequencies supposedly involved are between 5 and 9 Hz, which are below the lower frequency limit of human hearing. High-power sound waves below 20 Hz are felt in the body, not heard by the ear as sound.
Physiological effects of low frequency vibration
Air is a very inefficient medium for transferring low frequency vibration from a transducer to the human body. Mechanical connection of the vibration source to the human body, however, provides a potentially dangerous combination. The U.S. space program, worried about the harmful effects of rocket flight on astronauts, ordered vibration tests that used cockpit seats mounted on vibration tables to transfer "brown note" and other frequencies directly to the human subjects. Very high power levels of 160 dB were achieved at frequencies of 2–3 Hz. Test frequencies ranged from 0.5 Hz to 40 Hz. Test subjects suffered motor ataxia, nausea, visual disturbance, degraded task performance and difficulties in communication. These tests are assumed by researchers to be the nucleus of the current urban myth.
MythBusters testing
In February 20 |
https://en.wikipedia.org/wiki/Manaaki%20Whenua%20%E2%80%93%20Landcare%20Research | Manaaki Whenua – Landcare Research is a New Zealand Crown Research Institute whose focus of research is the environment, biodiversity, and sustainability.
History
Manaaki Whenua was originally part of the Department of Scientific and Industrial Research (DSIR), but was established as an independent organisation when the Crown Research Institutes were created in 1992. As part of that process, it was semi-commercialised, and now operates as a government-owned company rather than as a government department. The commercialisation has led to greater emphasis on financial viability, and Manaaki Whenua is employed by various private groups to provide advice and information. It is currently chaired by Acting Chair Dr Paul Reynolds QSO.
Locations
The main site is in Lincoln, near Christchurch. There are also other sites at Auckland on the Tamaki campus of Auckland University, Hamilton, Palmerston North, Wellington, Alexandra, and Dunedin.
Collections
Manaaki Whenua – Landcare Research holds several collections of organisms that are of significant national importance to New Zealand. Detailed information on all the specimens can be found though the Systematics Collections Data (SCD) website.
International collection of microorganisms from plants
The International Collection of Microorganisms from Plants in Auckland holds live bacterial and fungal specimens that are preserved under liquid nitrogen or in freeze dried ampoules. Currently there are over 20,000 specimens in the collecti |
https://en.wikipedia.org/wiki/Thermodynamic%20equations | Thermodynamics is expressed by a mathematical framework of thermodynamic equations which relate various thermodynamic quantities and physical properties measured in a laboratory or production process. Thermodynamics is based on a fundamental set of postulates, that became the laws of thermodynamics.
Introduction
One of the fundamental thermodynamic equations is the description of thermodynamic work in analogy to mechanical work, or weight lifted through an elevation against gravity, as defined in 1824 by French physicist Sadi Carnot. Carnot used the phrase motive power for work. In the footnotes to his famous On the Motive Power of Fire, he states: “We use here the expression motive power to express the useful effect that a motor is capable of producing. This effect can always be likened to the elevation of a weight to a certain height. It has, as we know, as a measure, the product of the weight multiplied by the height to which it is raised.” With the inclusion of a unit of time in Carnot's definition, one arrives at the modern definition for power:
During the latter half of the 19th century, physicists such as Rudolf Clausius, Peter Guthrie Tait, and Willard Gibbs worked to develop the concept of a thermodynamic system and the correlative energetic laws which govern its associated processes. The equilibrium state of a thermodynamic system is described by specifying its "state". The state of a thermodynamic system is specified by a number of extensive quantities, the |
https://en.wikipedia.org/wiki/Confluent%20hypergeometric%20function | In mathematics, a confluent hypergeometric function is a solution of a confluent hypergeometric equation, which is a degenerate form of a hypergeometric differential equation where two of the three regular singularities merge into an irregular singularity. The term confluent refers to the merging of singular points of families of differential equations; confluere is Latin for "to flow together". There are several common standard forms of confluent hypergeometric functions:
Kummer's (confluent hypergeometric) function , introduced by , is a solution to Kummer's differential equation. This is also known as the confluent hypergeometric function of the first kind. There is a different and unrelated Kummer's function bearing the same name.
Tricomi's (confluent hypergeometric) function introduced by , sometimes denoted by , is another solution to Kummer's equation. This is also known as the confluent hypergeometric function of the second kind.
Whittaker functions (for Edmund Taylor Whittaker) are solutions to Whittaker's equation.
Coulomb wave functions are solutions to the Coulomb wave equation.
The Kummer functions, Whittaker functions, and Coulomb wave functions are essentially the same, and differ from each other only by elementary functions and change of variables.
Kummer's equation
Kummer's equation may be written as:
with a regular singular point at and an irregular singular point at . It has two (usually) linearly independent solutions and .
Kummer's function |
https://en.wikipedia.org/wiki/Texture%20%28chemistry%29 | In physical chemistry and materials science, texture is the distribution of crystallographic orientations of a polycrystalline sample (it is also part of the geological fabric). A sample in which these orientations are fully random is said to have no distinct texture. If the crystallographic orientations are not random, but have some preferred orientation, then the sample has a weak, moderate or strong texture. The degree is dependent on the percentage of crystals having the preferred orientation.
Texture is seen in almost all engineered materials, and can have a great influence on materials properties. The texture forms in materials during thermo-mechanical processes, for example during production processes e.g. rolling. Consequently, the rolling process is often followed by a heat treatment to reduce the amount of unwanted texture. Controlling the production process in combination with the characterization of texture and the material's microstructure help to determine the materials properties, i.e. the processing-microstructure-texture-property relationship. Also, geologic rocks show texture due to their thermo-mechanic history of formation processes.
One extreme case is a complete lack of texture: a solid with perfectly random crystallite orientation will have isotropic properties at length scales sufficiently larger than the size of the crystallites. The opposite extreme is a perfect single crystal, which likely has anisotropic properties by geometric necessity.
Chara |
https://en.wikipedia.org/wiki/V%C3%ADctor%20A.%20Carre%C3%B1o | Víctor A. Carreño (born 1956) is a NASA aerospace engineer and aerospace technologist. He holds the patent for the Single Frequency Multitransmitter Telemetry System.
Early years
Carreño was born in Santo Domingo, Dominican Republic. His family moved to Puerto Rico when he was only a child and he was raised in the City of Guaynabo. Carreño became interested in electronics and the solution of mathematical problems as a child. After finishing his primary and secondary education, he attended the Margarita Janer Palacios High School and here he was a top mathematics and science student and graduated with honors in 1974.
In 1974, Carreño enrolled at the University of Puerto Rico and earned his Bachelor of Science degree in electrical engineering in 1979. Upon graduation, he applied to and was hired by the NASA Langley Research Center and assigned to the Aircraft Electronics System Branch, Flight Electronics Division. In 1983, he was reassigned to the digital system upset assessment team, in the Fault Tolerant Systems Branch. His work involved the development of techniques for analytically assessing digital system upset due to lightning-induced transients.
Career in NASA
Carreño is credited with inventing and developing the Single Frequency Multitransmitter Telemetry System in 1983. He also designed and conducted experiments for the real-time evaluation of the Viper single board computer while working in the instrumentation of the F-106 lightning research aircraft. Carreño co |
https://en.wikipedia.org/wiki/Cutaneous%20T-cell%20lymphoma | Cutaneous T-cell lymphoma (CTCL) is a class of non-Hodgkin lymphoma, which is a type of cancer of the immune system. Unlike most non-Hodgkin lymphomas (which are generally B-cell-related), CTCL is caused by a mutation of T cells. The cancerous T cells in the body initially migrate to the skin, causing various lesions to appear. These lesions change shape as the disease progresses, typically beginning as what appears to be a rash which can be very itchy and eventually forming plaques and tumors before spreading to other parts of the body.
Signs and symptoms
The presentation depends if it is mycosis fungoides or Sézary syndrome, the most common, though not the only types.
Among the symptoms for the aforementioned types are: enlarged lymph nodes, an enlarged liver and spleen, and non-specific dermatitis.
Cause
The cause of CTCL is unknown.
Diagnosis
A point-based algorithm for the diagnosis for early forms of cutaneous T-cell lymphoma was proposed by the International Society for Cutaneous Lymphomas in 2005.
Classification
Cutaneous T-cell lymphoma may be divided into the several subtypes. Mycosis fungoides is the most common form of CTCL and is responsible for half of all cases. A WHO-EORTC classification has been developed.
Treatment
There is no cure for CTCL, but there are a variety of treatment options available and some CTCL patients are able to live normal lives with this cancer, although symptoms can be debilitating and painful, even in earlier stages. FDA appr |
https://en.wikipedia.org/wiki/Jotun%20Hein | Jotun John Piet Hein (born 19 July 1956) is Professor of Bioinformatics at the Department of Statistics of the University of Oxford and a professorial fellow of University College, Oxford. Hein was previously Director of the Bioinformatics Research Centre at Aarhus University, Denmark.
Hein is the fourth son of Piet Hein, the Danish scientist, mathematician, inventor, designer, author, and poet who wrote the famed Grooks poetry collections and invented the Superegg and the Soma cube. When he was 12 years old, Jotun proved the Soma cube's "Basalt Rock" construction impossible, which was published in the puzzle's instruction manual as "Jotun's Proof."
Hein's research interests are in molecular evolution, molecular population genetics and bioinformatics.
Selected books
Hein, J; Schierup, M. H., and Wiuf, C. Gene Genealogies, Variation and Evolution – A Primer in Coalescent Theory. Oxford University Press, 2005. .
References
External links
Personal home page
1956 births
Living people
British statisticians
Fellows of University College, Oxford
20th-century British mathematicians
21st-century British mathematicians |
https://en.wikipedia.org/wiki/Br%C3%A9guet%20693 | The Bréguet 690 and its derivatives were a series of light twin-engine ground-attack aircraft that were used by the French Air Force in World War II. The aircraft was intended to be easy to maintain, forgiving to fly, and capable of at . The type's sturdy construction was frequently demonstrated and the armament was effective. French rearmament began two years later than that in Britain and none of these aircraft were available in sufficient numbers to make a difference in 1940.
Design and development
Bréguet 690
The Bréguet 690 had begun life in 1934 as the Bre 630, the Bréguet Aviation entry for the (STAé, Aeronautical Technical Service) specification of October 1934 along with the Hanriot H 220, Loire-Nieuport LN-20, Romano R.110 and the Potez 630. The Bréguet 630 was a twin-engined monoplane with twin tailplanes and Hispano-Suiza 14AB 02/03 (port and starboard) 14-cylinder air-cooled radial engines, both rotating inwards to limit torque problems if one engine failed. The aircraft was armed with two forward-firing Hispano-Suiza HS.404 cannon and a 7.5 mm MAC 1934 machine-gun firing rearwards for aft defence. The Potez 630 won the C3 competition but Bréguet began construction of the prototype Bréguet 690 in 1935, without an order from the , which was not placed until 26 March 1937. Completion of the Bréguet 690-01 was slowed by a ten-month delay in the delivery of its engines from Hispano-Suiza. The Bre 690-01 was finished in early 1938 and flew on 23 March, revealing |
https://en.wikipedia.org/wiki/%C3%93%20Flaithbheartaigh | O'Flaherty ( , ; ; ) is an Irish Gaelic clan based most prominently in what is today County Galway. The clan name originated in the 10th century as a derivative of its founder Flaithbheartach mac Eimhin. They descend in the paternal line from the Connachta's Uí Briúin Seóla. They were originally kings of Maigh Seóla and Muintir Murchada and as members of the Uí Briúin were kinsmen of the Ó Conchubhair and Mac Diarmada amongst others. After their king Cathal mac Tigernán lost out to Áed in Gai Bernaig in the 11th century, the family were pushed further west to Iar Connacht, a territory associated with Connemara today. They continued to rule this land until the 16th century. The name has been alternatively rendered into English in various forms, such as Flaherty, Fluharty, Faherty, Laverty, Flaverty, Lahiff, and Flahive.
Naming conventions
Overview
This Gaelic-Irish surname is written as "Ua Flaithbertach" (nominative) or "Ua Flaithbertaig" (genitive) in Old Irish and Middle Irish texts. In Modern Irish the surname is now generally spelt as Ó Flatharta.
The surname is commonly translated as "bright ruler" or more correctly "bright prince", flaith originally meaning prince in Irish. "O" or Ó comes from Ua, designating "grandson" or "descendant" of a (major) clan member. The prefix is often anglicised to O', using an apostrophe instead of the Irish síneadh fada: "'".
Maigh Seóla was the earliest O'Flaherty domain, to the east of Lough Corrib in the kingdom of Connacht, th |
https://en.wikipedia.org/wiki/Military%20brat%20%28U.S.%20subculture%29 | In the United States, a military brat (also known by various "brat" derivatives) is the child of a parent(s), adopted parent(s) or legal guardian(s) serving full-time in the United States Armed Forces, whether current or former. The term military brat can also refer to the subculture and lifestyle of such families.
The military brat lifestyle typically involves moving to new states or countries many times while growing up, as the child's military family is customarily transferred to new non-combat assignments; consequently, many military brats never have a home town. War-related family stresses are also a commonly occurring part of military brat life. There are also other aspects of military brat life that are significantly different in comparison to the civilian American population, often including living in foreign countries and or diverse regions within the U.S., exposure to foreign languages and cultures, and immersion in military culture.
The military brats subculture has emerged over the last 200 years. The age of the phenomenon has meant military brats have also been described by a number of researchers as one of America's oldest and yet least well-known and largely invisible subcultures. They have also been described as a "modern nomadic subculture".
Military brat is known in U.S. military culture as a term of endearment and respect. The term may also connote a military brat's experience of mobile upbringing, and may refer to a sense of worldliness. Research has sh |
https://en.wikipedia.org/wiki/Monocrystalline%20whisker | A monocrystalline whisker is a filament of material that is structured as a single, defect-free crystal. Some typical whisker materials are graphite, alumina, iron, silicon carbide and silicon. Single-crystal whiskers of these (and some other) materials are known for having very high tensile strength (on the order of 10–20 GPa). Whiskers are used in some composites, but large-scale fabrication of defect-free whiskers is very difficult.
Prior to the discovery of carbon nanotubes, single-crystal whiskers had the highest tensile strength of any materials known, and were featured regularly in science fiction as materials for fabrication of space elevators, arcologies, and other large structures. Despite showing great promise for a range of applications, their usage has been hindered by concerns over their effects on health when inhaled.
See also
Whisker (metallurgy) – Self-organizing metallic whisker-shaped structures that cause problems with electronics.
Laser-heated pedestal growth
References
"Mechanical and Physical Properties of Whiskers", CRC Handbook of Chemistry and Physics, 55th edition.
Materials |
https://en.wikipedia.org/wiki/Visvedevas | The visvedevas () refers to the designation used to address the entirety of the various deities featured in the Vedas. It also refers to a specific classification of deities in the Puranas. The visvedevas are sometimes regarded as the most comprehensive gathering of the gods, a classification in which no deity is stated to be omitted.
Literature
Rigveda
In the Rigveda a number of hymns are addressed to these deities, including (according to Griffith): 1.3,1.89,3.54-56, 4.55, 5.41-51, 6.49-52, 7.34-37, 39, 40, 42, 43, 8.27-30, 58, 83 10.31, 35, 36, 56, 57, 61-66, 92, 93, 100, 101, 109, 114, 126, 128, 137, 141, 157, 165, 181.
RV 3.54.17 addresses them as headed by Indra:
This is, ye Wise, your great and glorious title, that all ye Deities abide in Indra. (trans. Griffith)
The dichotomy between devas is not evident in these hymns, and the devas are invoked together such as Mitra and Varuna. Though many devas are named in the Rigveda, only 33 devas are counted, eleven of them present each in earth, space, and heaven.
Manusmriti
According to Manu (iii, 90, 121), offerings should be made daily to the visvedevas. These privileges were bestowed on them by Brahma and the Pitri as a reward for severe austerities they had performed on the Himalayas.
Puranas
In later Hinduism, the visvedevas form one of the nine ganadevatas (along with the adityas, vasus, tushitas, abhasvaras, anilas, maharajikas, sadhyas, and rudras). According to the Vishnu Purana and Padma Purana, they were |
https://en.wikipedia.org/wiki/Circular%20distribution | In probability and statistics, a circular distribution or polar distribution is a probability distribution of a random variable whose values are angles, usually taken to be in the range A circular distribution is often a continuous probability distribution, and hence has a probability density, but such distributions can also be discrete, in which case they are called circular lattice distributions. Circular distributions can be used even when the variables concerned are not explicitly angles: the main consideration is that there is not usually any real distinction between events occurring at the lower or upper end of the range, and the division of the range could notionally be made at any point.
Graphical representation
If a circular distribution has a density
it can be graphically represented as a closed curve
where the radius is set equal to
and where a and b are chosen on the basis of appearance.
Examples
By computing the probability distribution of angles along a handwritten ink trace,
a lobe-shaped polar distribution emerges. The main direction of the lobe in the
first quadrant corresponds to the slant of handwriting (see: graphonomics).
An example of a circular lattice distribution would be the probability of being born in a given month of the year, with each calendar month being thought of as arranged round a circle, so that "January" is next to "December".
See also
Circular mean
Circular uniform distribution
von Mises distribution
References
External link |
https://en.wikipedia.org/wiki/Hariri | Hariri (in Arabic حريري) is a surname and derivative of harir (in Arabic حرير meaning silk) which indicates a mercantile background at one point in that field.
People
Historic
Ali Hariri (1009-1079), Kurdish poet
Al-Hariri of Basra (1054–1122), Arab poet, scholar of the Arabic language and a high government official of the Seljuk Empire
Surname
Family of Rafic Hariri
Ayman Hariri (born 1978), Lebanese businessman, son of Rafic Hariri
Bahia Hariri (born 1952), Lebanese politician, sister of Rafic Hariri
Bahaa Hariri (born 1966), Lebanese business tycoon, son of Rafic Hariri
Fahd Hariri (born 1980/1981), Lebanese businessman and property developer, the son of Rafic Hariri
Hind Hariri (born 1984), daughter and youngest child of Rafic Hariri
Nazik Hariri, widow of Rafic Hariri
Rafic Hariri (1944–2005), business tycoon and Lebanese Prime Minister; assassinated
Saad Hariri (born 1970), politician, business tycoon, Lebanese Prime Minister, and son of Rafic Hariri
Other people
Abbas Hariri, Iranian wrestler
Abu Al-Izz Al-Hariri (1946–2014), Egyptian politician and member of parliament
Abdulhadi Al Hariri (born 1992), Syrian footballer
Fawzi Hariri (born 1958), Iraqi Minister of Industry and Minerals (since 2006)
Franso Hariri (1937–2001), Kurdish Iraqi politician
Lamia Al Hariri, Syrian diplomat
May Hariri (born 1972), Lebanese pop artist and actress
Naser al-Hariri, Syrian politician
Omar El-Hariri (c. 1944–2015), Libyan politician, minister, leading figur |
https://en.wikipedia.org/wiki/Suhas%20Patankar | Suhas V. Patankar (born 22 February 1941) is an Indian mechanical engineer. He is a pioneer in the field of computational fluid dynamics (CFD) and Finite volume method. He is currently a Professor Emeritus at the University of Minnesota. He is also president of Innovative Research, Inc. Patankar was born in Pune, Maharashtra, India.
Early life and education
Patankar received his bachelor's degree in mechanical engineering in 1962 from the College of Engineering, Pune, which is affiliated to the University of Pune and his Master of Technology degree in mechanical engineering from the Indian Institute of Technology Bombay in 1964. In 1967 he received his Ph.D. in mechanical engineering from the Imperial College, University of London.
Career
Patankar's most important contribution to the field of CFD is the SIMPLE algorithm that he developed along with his colleagues at Imperial College. Patankar is the author of a book in computational fluid dynamics titled Numerical Heat Transfer and Fluid Flow which was first published in 1980. This book has since been considered one of the groundbreaking contributions to computational fluid dynamics due to its emphasis on physical understanding and insight into the fluid flow and heat transfer phenomena. He is also one of the most cited authors in science and engineering.
References
1941 births
Living people
American mechanical engineers
Indian mechanical engineers
Computational fluid dynamicists
Savitribai Phule Pune University alumn |
https://en.wikipedia.org/wiki/Helly%E2%80%93Bray%20theorem | In probability theory, the Helly–Bray theorem relates the weak convergence of cumulative distribution functions to the convergence of expectations of certain measurable functions. It is named after Eduard Helly and Hubert Evelyn Bray.
Let F and F1, F2, ... be cumulative distribution functions on the real line. The Helly–Bray theorem states that if Fn converges weakly to F, then
for each bounded, continuous function g: R → R, where the integrals involved are Riemann–Stieltjes integrals.
Note that if X and X1, X2, ... are random variables corresponding to these distribution functions, then the Helly–Bray theorem does not imply that E(Xn) → E(X), since g(x) = x is not a bounded function.
In fact, a stronger and more general theorem holds. Let P and P1, P2, ... be probability measures on some set S. Then Pn converges weakly to P if and only if
for all bounded, continuous and real-valued functions on S. (The integrals in this version of the theorem are Lebesgue–Stieltjes integrals.)
The more general theorem above is sometimes taken as defining weak convergence of measures (see Billingsley, 1999, p. 3).
References
Probability theorems |
https://en.wikipedia.org/wiki/Nonmetricity%20tensor | In mathematics, the nonmetricity tensor in differential geometry is the covariant derivative of the metric tensor. It is therefore a tensor field of order three. It vanishes for the case of Riemannian geometry and can be
used to study non-Riemannian spacetimes.
Definition
By components, it is defined as follows.
It measures the rate of change of the components of the metric tensor along the flow of a given vector field, since
where is the coordinate basis of vector fields of the tangent bundle, in the case of having a 4-dimensional manifold.
Relation to connection
We say that a connection is compatible with the metric when its associated covariant derivative of the metric tensor (call it , for example) is zero, i.e.
If the connection is also torsion-free (i.e. totally symmetric) then it is known as the Levi-Civita connection, which is the only one without torsion and compatible with the metric tensor. If we see it from a geometrical point of view, a non-vanishing nonmetricity tensor for a metric tensor implies that the modulus of a vector defined on the tangent bundle to a certain point of the manifold, changes when it is evaluated along the direction (flow) of another arbitrary vector.
References
External links
Differential geometry |
https://en.wikipedia.org/wiki/Zeroth%20law | Zeroth law may refer to:
Zeroth law of black hole thermodynamics, about event horizons of black holes
Zeroth law of robotics, an addition to Isaac Asimov's Three Laws of Robotics
Zeroth law of thermodynamics, in relation to thermal equilibriums
See also
Zeroth (disambiguation) |
https://en.wikipedia.org/wiki/Collectively%20exhaustive%20events | In probability theory and logic, a set of events is jointly or collectively exhaustive if at least one of the events must occur. For example, when rolling a six-sided die, the events 1, 2, 3, 4, 5, and 6 balls of a single outcome are collectively exhaustive, because they encompass the entire range of possible outcomes.
Another way to describe collectively exhaustive events is that their union must cover all the events within the entire sample space. For example, events A and B are said to be collectively exhaustive if
where S is the sample space.
Compare this to the concept of a set of mutually exclusive events. In such a set no more than one event can occur at a given time. (In some forms of mutual exclusion only one event can ever occur.) The set of all possible die rolls is both mutually exclusive and collectively exhaustive (i.e., "MECE"). The events 1 and 6 are mutually exclusive but not collectively exhaustive. The events "even" (2,4 or 6) and "not-6" (1,2,3,4, or 5) are also collectively exhaustive but not mutually exclusive. In some forms of mutual exclusion only one event can ever occur, whether collectively exhaustive or not. For example, tossing a particular biscuit for a group of several dogs cannot be repeated, no matter which dog snaps it up.
One example of an event that is both collectively exhaustive and mutually exclusive is tossing a coin. The outcome must be either heads or tails, or p (heads or tails) = 1, so the outcomes are collectively exhaustiv |
https://en.wikipedia.org/wiki/Vertisol | A vertisol is a Soil Order in the USDA soil taxonomy and a Reference Soil Group in the World Reference Base for Soil Resources (WRB). It is also defined in many other soil classification systems. In the Australian Soil Classification it is called vertosol. Vertisols have a high content of expansive clay minerals, many of them belonging to the montmorillonites that form deep cracks in drier seasons or years. In a phenomenon known as argillipedoturbation, alternate shrinking and swelling causes self-ploughing, where the soil material consistently mixes itself, causing some vertisols to have an extremely deep A horizon and no B horizon. (A soil with no B horizon is called an A/C soil). This heaving of the underlying material to the surface often creates a microrelief known as gilgai.
Vertisols typically form from highly basic rocks, such as basalt, in climates that are seasonally humid or subject to erratic droughts and floods, or that impeded drainage. Depending on the parent material and the climate, they can range from grey or red to the more familiar deep black (known as "black earths" in Australia, "black gumbo" in East Texas, "black cotton" soils in East Africa, and "vlei soils" in South Africa).
Vertisols are found between 50°N and 45°S of the equator. Major areas where vertisols are dominant are eastern Australia (especially inland Queensland and New South Wales), the Deccan Plateau of India, and parts of southern Sudan, Ethiopia, Kenya, Chad (the Gezira), South Afric |
https://en.wikipedia.org/wiki/Nadvirna | Nadvírna, also referred to as Nadwirna or Nadvorna (, , , Nadvorna) is a city located in Ivano-Frankivsk Oblast in western Ukraine. It is the administrative centre of Nadvirna Raion. Nadvirna hosts the administration of Nadvirna urban hromada, one of the hromadas of Ukraine. Population:
From the mid-14th century until 1772 (see Partitions of Poland) Nadvirna, known in Polish as Nadwórna, was part of the Kingdom of Poland. In 1772, it was annexed by the Habsburg Empire, and remained in the province of Galicia until late 1918. In the inter-war years, the borders changed and the town became part of the Second Polish Republic. Following the 1939 Invasion of Poland, it was annexed into the Ukrainian SSR (see also Molotov–Ribbentrop pact). Nadvirna was occupied by the Germans in 1941 during World War II. After the war it was once again absorbed into the Ukrainian SSR. Since its independence in 1991, the city has been part of Ukraine.
The town is located in a slightly hilly, verdant area twenty miles () northeast of the Carpathian mountains. Major exports and raw materials from the town include salt, oil and petroleum products, and timber. The town was popular at the start of the 20th century as a summertime resort, with restaurants and hotels.
History
Evidence of the early settlement in the region around Nadvirna dates back to 2000 BC. Numerous finds of Bronze Age artifacts attest to a vibrant culture. The town was built around the Pniv castle. The Pniv (Polish: Pniów) Cas |
https://en.wikipedia.org/wiki/Nonelementary%20integral | In mathematics, a nonelementary antiderivative of a given elementary function is an antiderivative (or indefinite integral) that is, itself, not an elementary function (i.e. a function constructed from a finite number of quotients of constant, algebraic, exponential, trigonometric, and logarithmic functions using field operations). A theorem by Liouville in 1835 provided the first proof that nonelementary antiderivatives exist. This theorem also provides a basis for the Risch algorithm for determining (with difficulty) which elementary functions have elementary antiderivatives.
Examples
Examples of functions with nonelementary antiderivatives include:
(elliptic integral)
(logarithmic integral)
(error function, Gaussian integral)
and (Fresnel integral)
(sine integral, Dirichlet integral)
(exponential integral)
(in terms of the exponential integral)
(in terms of the logarithmic integral)
(incomplete gamma function); for the antiderivative can be written in terms of the exponential integral; for in terms of the error function; for any positive integer, the antiderivative elementary.
Some common non-elementary antiderivative functions are given names, defining so-called special functions, and formulas involving these new functions can express a larger class of non-elementary antiderivatives. The examples above name the corresponding special functions in parentheses.
Properties
Nonelementary antiderivatives can often be evaluated using Taylor series. Even if a |
https://en.wikipedia.org/wiki/Proper%20orthogonal%20decomposition | The proper orthogonal decomposition is a numerical method that enables a reduction in the complexity of computer intensive simulations such as computational fluid dynamics and structural analysis (like crash simulations). Typically in fluid dynamics and turbulences analysis, it is used to replace the Navier–Stokes equations by simpler models to solve.
It belongs to a class of algorithms called model order reduction (or in short model reduction). What it essentially does is to train a model based on simulation data. To this extent, it can be associated with the field of machine learning.
POD and PCA
The main use of POD is to decompose a physical field (like pressure, temperature in fluid dynamics or stress and deformation in structural analysis), depending on the different variables that influence its physical behaviors. As its name hints, it's operating an Orthogonal Decomposition along with the Principal Components of the field. As such it is assimilated with the principal component analysis from Pearson in the field of statistics, or the singular value decomposition in linear algebra because it refers to eigenvalues and eigenvectors of a physical field. In those domains, it is associated with the research of Karhunen and Loève, and their Karhunen–Loève theorem.
Mathematical expression
The first idea behind the Proper Orthogonal Decomposition (POD), as it was originally formulated in the domain of fluid dynamics to analyze turbulences, is to decompose a random vector f |
https://en.wikipedia.org/wiki/HomeRF | HomeRF was a wireless networking specification for home devices. It was developed in 1998 by the Home Radio Frequency Working Group, a consortium of mobile wireless companies that included Proxim Wireless, Intel, Siemens AG, Motorola, Philips and more than 100 other companies.
The group was disbanded in January 2003, after other wireless networks became accessible to home users and Microsoft began including support for them in its Windows operating systems. As a result, HomeRF fell into obsolescence.
Description
Initially called Shared Wireless Access Protocol (SWAP) and later just HomeRF, this open specification allowed PCs, peripherals, cordless phones and other consumer devices to share and communicate voice and data in and around the home without the complication and expense of running new wires. HomeRF combined several wireless technologies in the 2.4 GHz ISM band, including IEEE 802.11 FH (the frequency-hopping version of wireless data networking) and DECT (the most prevalent digital cordless telephony standard in the world) to meet the unique home networking requirements for security, quality of service (QoS) and interference immunity—issues that still plagued Wi-Fi (802.11b and g).
HomeRF used frequency hopping spread spectrum (FHSS) in the 2.4 GHz frequency band and in theory could achieve a maximum of 10 Mbit/s throughput; its nodes could travel within a 50-meter range of a wireless access point while remaining connected to the personal area network (PAN). Seve |
https://en.wikipedia.org/wiki/Ptolemy%27s%20theorem | In Euclidean geometry, Ptolemy's theorem is a relation between the four sides and two diagonals of a cyclic quadrilateral (a quadrilateral whose vertices lie on a common circle). The theorem is named after the Greek astronomer and mathematician Ptolemy (Claudius Ptolemaeus). Ptolemy used the theorem as an aid to creating his table of chords, a trigonometric table that he applied to astronomy.
If the vertices of the cyclic quadrilateral are A, B, C, and D in order, then the theorem states that:
This relation may be verbally expressed as follows:
If a quadrilateral is cyclic then the product of the lengths of its diagonals is equal to the sum of the products of the lengths of the pairs of opposite sides.
Moreover, the converse of Ptolemy's theorem is also true:
In a quadrilateral, if the sum of the products of the lengths of its two pairs of opposite sides is equal to the product of the lengths of its diagonals, then the quadrilateral can be inscribed in a circle i.e. it is a cyclic quadrilateral.
Corollaries on Inscribed Polygons
Equilateral triangle
Ptolemy's Theorem yields as a corollary a pretty theorem regarding an equilateral triangle inscribed in a circle.
Given An equilateral triangle inscribed on a circle and a point on the circle.
The distance from the point to the most distant vertex of the triangle is the sum of the distances from the point to the two nearer vertices.
Proof: Follows immediately from Ptolemy's theorem:
Square
Any square can be inscrib |
https://en.wikipedia.org/wiki/Formate | Formate (IUPAC name: methanoate) is the conjugate base of formic acid. Formate is an anion () or its derivatives such as ester of formic acid. The salts and esters are generally colorless.
Fundamentals
When dissolved in water, formic acid converts to formate:
Formate is a planar anion. The two oxygen atoms are equivalent and bear a partial negative charge. The remaining C-H bond is not acidic.
Biochemistry
Formate is a common C-1 source in living systems. It is formed from many precursors including choline, serine, and sarcosine. It provides a C-1 source in the biosynthesis of some nucleic acids. Formate (or formic acid) is invoked as a leaving group in the demethylation of some sterols.
These conversions are catalyzed by aromatase enzymes using O2 as the oxidant. Specific conversions include testosterone to estradiol and androstenedione to estrone.
Formate is reversibly oxidized by the enzyme formate dehydrogenase from Desulfovibrio gigas:
Formate esters
Formate esters have the formula HCOOR (alternative way of writing formula ROC(O)H or RO2CH). Many form spontaneously when alcohols dissolve in formic acid.
The most important formate ester is methyl formate, which is produced as an intermediate en route to formic acid. Methanol and carbon monoxide react in the presence of a strong base, such as sodium methoxide:
Hydrolysis of methyl formate gives formic acid and regenerates methanol:
Formic acid is used for many applications in industry.
Formate ester |
https://en.wikipedia.org/wiki/Schur%27s%20theorem | In discrete mathematics, Schur's theorem is any of several theorems of the mathematician Issai Schur. In differential geometry, Schur's theorem is a theorem of Axel Schur. In functional analysis, Schur's theorem is often called Schur's property, also due to Issai Schur.
Ramsey theory
In Ramsey theory, Schur's theorem states that for any partition of the positive integers into a finite number of parts, one of the parts contains three integers x, y, z with
For every positive integer c, S(c) denotes the smallest number S such that for every partition of the integers into c parts, one of the parts contains integers x, y, and z with . Schur's theorem ensures that S(c) is well-defined for every positive integer c. The numbers of the form S(c) are called Schur's number.
Folkman's theorem generalizes Schur's theorem by stating that there exist arbitrarily large sets of integers, all of whose nonempty sums belong to the same part.
Using this definition, the only known Schur numbers are S(n) 2, 5, 14, 45, and 161 () The proof that was announced in 2017 and took up 2 petabytes of space.
Combinatorics
In combinatorics, Schur's theorem tells the number of ways for expressing a given number as a (non-negative, integer) linear combination of a fixed set of relatively prime numbers. In particular, if is a set of integers such that , the number of different tuples of non-negative integer numbers such that when goes to infinity is:
As a result, for every set of relatively prim |
https://en.wikipedia.org/wiki/Education%20in%20North%20Korea | Education in North Korea is universal and state-funded schooling by the government. As of 2021, UNESCO Institute for Statistics does not report any data for North Korea's literacy rates. Some children go through one year of kindergarten, four years of primary education, six years of secondary education, and then on to university. The North Korean state claims its national literacy rate for citizens aged 15 and older is 100 percent.
In 1988, the United Nations Educational, Scientific, and Cultural Organization (UNESCO) reported that North Korea had 35,000 preprimary, 60,000 primary, 111,000 secondary, 23,000 college and university, and 4,000 other postsecondary teachers.
History
Formal education has played a central role in the social and cultural development of both traditional Korea and contemporary North Korea. During the Joseon Dynasty, the royal court established a system of schools that taught Confucian subjects in the provinces as well as in four central secondary schools in the capital. There was no state-supported system of primary education.
During the 15th century, state-supported schools declined in quality and were supplanted in importance by private academies, the seowon, centers of a Neo-Confucian revival in the 16th century. Higher education was provided by the Seonggyungwan, the Confucian national university, in Seoul. Its enrollment was limited to 200 students who had passed the lower civil-service examinations and were preparing for the highest examinat |
https://en.wikipedia.org/wiki/Wigner%E2%80%93Seitz%20cell | The Wigner–Seitz cell, named after Eugene Wigner and Frederick Seitz, is a primitive cell which has been constructed by applying Voronoi decomposition to a crystal lattice. It is used in the study of crystalline materials in crystallography.
The unique property of a crystal is that its atoms are arranged in a regular three-dimensional array called a lattice. All the properties attributed to crystalline materials stem from this highly ordered structure. Such a structure exhibits discrete translational symmetry. In order to model and study such a periodic system, one needs a mathematical "handle" to describe the symmetry and hence draw conclusions about the material properties consequent to this symmetry. The Wigner–Seitz cell is a means to achieve this.
A Wigner–Seitz cell is an example of a primitive cell, which is a unit cell containing exactly one lattice point. For any given lattice, there are an infinite number of possible primitive cells. However there is only one Wigner–Seitz cell for any given lattice. It is the locus of points in space that are closer to that lattice point than to any of the other lattice points.
A Wigner–Seitz cell, like any primitive cell, is a fundamental domain for the discrete translation symmetry of the lattice. The primitive cell of the reciprocal lattice in momentum space is called the Brillouin zone.
Overview
Background
The concept of Voronoi decomposition was investigated by Peter Gustav Lejeune Dirichlet, leading to the name Dirichlet |
https://en.wikipedia.org/wiki/Breakpoint%20%28disambiguation%29 | A breakpoint is an execution stop point in the code of a computer program.
Breakpoint or break point may also refer to:
BCR (gene), the gene that encodes the breakpoint cluster region protein
Break point, in tennis
Break Point, a 2002 novel by Rosie Rushton
Break Point (film), a 2015 U.S. comedy film
Breakpoint (demoparty), a German demoscene party
Breakpoint (meteorology), a location referred to by meteorologists when issuing watches, warnings, or advisories for specific areas
Breakpoint (novel) a 2007 novel by Richard A. Clarke
Breakpoint ("The Shield"), a 2003 episode of the television show The Shield
Breakpoint, an indicator of a microbial organism's susceptibility or resistance to a particular antimicrobial; see Minimum inhibitory concentration
"Breakpoint", a song by Megadeth on the 1995 album Hidden Treasures
See also
Break Point (disambiguation)
Breaking point (disambiguation)
Point Break, a 1991 action film
Tom Clancy's Ghost Recon Breakpoint, a 2019 online video game by Ubisoft |
https://en.wikipedia.org/wiki/Duophonic%20Records | Duophonic Ultra High Frequency Disks Limited (also known as Duophonic Records or Duophonic Super 45s) is a British independent record label formed by English-French rock band Stereolab in 1991. The label has two imprints: Duophonic Ultra High Frequency Disks for UK Stereolab releases licensed to various labels worldwide, and Duophonic Super 45s for releases of other artists and certain Stereolab UK-only releases. Duophonic's first release was Stereolab's debut EP Super 45 (1991), limited to 880 copies; of these, forty copies had handmade covers that were produced by Martin Pike in his father's garage.
Bands that have released records on Duophonic include Broadcast, the High Llamas, Labradford, Tortoise, Pram, Yo La Tengo, the Notwist, and Apparat Organ Quartet. Daft Punk, one of the most successful electronic bands of the 1990s, released their first songs under the name Darlin' on the 1993 Duophonic compilation Shimmies in Super 8. Duophonic's most successful release is Stereolab's Emperor Tomato Ketchup (1996), which was licensed to Elektra Records outside the UK and has sold over 60,000 copies worldwide.
Duophonic is managed by Martin Pike, and is owned by Tim Gane (34%), Laetitia Sadier (34%), and Pike (32%). Although founded in 1991, the label did not become a limited company until 25 August 1993, when Pike relocated from Horsham, West Sussex, to East Dulwich in the London Borough of Southwark. From there, Pike also runs Associated London Management [2008] Ltd, a compan |
https://en.wikipedia.org/wiki/Crystal%20%28song%29 | "Crystal" is a song by English rock band New Order. The song was released on 11 July 2001 as the first single from their seventh studio album, Get Ready (2001). "Crystal" entered the UK Singles Chart at number eight, attracting considerable attention and critical praise as the band's comeback single, their first original since 1993. The song also found success internationally, peaking at number three in Canada, number seven in Finland, and reaching the top 50 in Germany, Ireland, Italy, and Sweden. "Crystal" appears as the first track on the album in a different version than the single release, with an extended intro and coda.
Release
Singer-guitarist Bernard Sumner originally gave the song to German record label Mastermind for Success, and it was recorded by label artist Corvin Dalek. However, DJ Pete Tong heard the song and declared it to be the best New Order single since "Blue Monday", leading Sumner to reconsider the gift and have New Order record and release it.
A version of the single was also released in Japan to promote the release of the New Order DVD 316, and has a different cover that resembles the 316 cover. B-sides for the single were 4 live audio tracks taken from the DVD. The single was B-sided by a variety of remixes, and an original song titled "Behind Closed Doors". All versions feature extensive backing vocals from Dawn Zee, mostly wordless. Zee has continued to perform with New Order on all their successive studio albums.
After the song was released, a |
https://en.wikipedia.org/wiki/Generalized%20canonical%20correlation | In statistics, the generalized canonical correlation analysis (gCCA), is a way of making sense of cross-correlation matrices between the sets of random variables when there are more than two sets. While a conventional CCA generalizes principal component analysis (PCA) to two sets of random variables, a gCCA generalizes PCA to more than two sets of random variables. The canonical variables represent those common factors that can be found by a large PCA of all of the transformed random variables after each set underwent its own PCA.
Applications
The Helmert-Wolf blocking (HWB) method of estimating linear regression parameters can find an optimal solution only if all cross-correlations between the data blocks are zero. They can always be made to vanish by introducing a new regression parameter for each common factor. The gCCA method can be used for finding those harmful common factors that create cross-correlation between the blocks. However, no optimal HWB solution exists if the random variables do not contain enough information on all of the new regression parameters.
References
Afshin-Pour, B.; Hossein-Zadeh, G.A. Strother, S.C.; Soltanian-Zadeh, H. (2012), "Enhancing reproducibility of fMRI statistical maps using generalized canonical correlation analysis in NPAIRS framework", NeuroImage 60(4): 1970–1981.
Sun, Q.S., Liu, Z.D., Heng, P.A., Xia, D.S. (2005) "A Theorem on the Generalized Canonical Projective Vectors". Pattern Recognition 38 (3) 449
Kettenring, J. R. (1971 |
https://en.wikipedia.org/wiki/Red%20Hat%20Enterprise%20Linux%20derivatives | Red Hat Enterprise Linux derivatives are Linux distributions that are based on the source code of Red Hat Enterprise Linux (RHEL).
History
Red Hat Linux was one of the first and most popular Linux distributions. This was largely because, while a paid-for supported version was available, a freely downloadable version was also available. Since the only difference between the paid-for option and the free option was support, a great number of people chose to use the free version.
In 2003, Red Hat made the decision to split its Red Hat Linux product into two: Red Hat Enterprise Linux for customers who were willing to pay for it, and Fedora that was made available free of charge but gets updates for every release for approximately 13 months.
Fedora has its own beta cycle and has some issues fixed by contributors who include Red Hat staff. However, its quick and nonconservative release cycle means it might not be suitable for some users. Fedora is somewhat a test-bed for Red Hat, allowing them to beta test their new features before they get included in Red Hat Enterprise Linux. Since the release of Fedora, Red Hat has no longer made binary versions of its commercial product available free-of-charge.
Motivations
Red Hat does not make a compiled version of its Enterprise Linux product available for free download. However, as the license terms on which it is mostly based explicitly stipulate, Red Hat has made the entire source code available in RPM format via their network of ser |
https://en.wikipedia.org/wiki/Low-Frequency%20Array%20%28LOFAR%29 | The Low-Frequency Array, or LOFAR, is a large radio telescope, with an antenna network located mainly in the Netherlands, and spreading across 7 other European countries as of 2019. Originally designed and built by ASTRON, the Netherlands Institute for Radio Astronomy, it was first opened by Queen Beatrix of The Netherlands in 2010, and has since been operated on behalf of the International LOFAR Telescope (ILT) partnership by ASTRON.
LOFAR consists of a vast array of omnidirectional radio antennas using a modern concept, in which the signals from the separate antennas are not connected directly electrically to act as a single large antenna, as they are in most array antennas. Instead, the LOFAR dipole antennas (of two types) are distributed in stations, within which the antenna signals can be partly combined in analogue electronics, then digitised, then combined again across the full station. This step-wise approach provides great flexibility in setting and rapidly changing the directional sensitivity on the sky of an antenna station. The data from all stations are then transported over fiber to a central digital processor, and combined in software to emulate a conventional radio telescope dish with a resolving power corresponding to the greatest distance between the antenna stations across Europe. LOFAR is thus an interferometric array, using about 20,000 small antennas concentrated in 52 stations since 2019. 38 of these stations are distributed across the Netherlands, bui |
https://en.wikipedia.org/wiki/Battle%20of%20Dynekilen | The naval Battle of Dynekilen () took place on 8 July 1716 during the Great Northern War between a Dano-Norwegian fleet under Peter Tordenskjold and a Swedish fleet under Olof Strömstierna. The battle resulted in a Dano-Norwegian victory.
Background
On 28 October 1709 Frederik IV of Denmark, the Danish-Norwegian king declared war against Sweden. The war declaration came after the Swedish defeat at the Battle of Poltava, which resulted in a decisive victory for Peter I of Russia over Charles XII of Sweden.
In the naval enactment, a light Danish-Norwegian force of 7 ships under Peter Tordenskjold trapped and defeated a Swedish transport fleet of 44 ships in Dynekilen fjord, just north of Strömstad, on the west coast of Sweden. The Swedish transport fleet was transporting troops, ammunition and supplies from Göteborg, destined for the land forces under the command of Charles XII invading Norway.
Battle
The Danish-Norwegian flotilla ambushed the Swedish fleet while it was positioned in the harbour of Dynekilen. In the process, it overcame and destroyed a small island fort equipped with six 12-pounder guns positioned in the harbour entrance. The largest Swedish ship, Stenbock, a former ship of the line converted into a cannon barge, surrendered, after which the lighter vessels were run aground, and an attempt made to destroy most of them. The Dano-Norwegian forces worked to put out fires and salvage as many of the ships as possible. They managed to save and capture 30 ships, wh |
https://en.wikipedia.org/wiki/Heinrich%20Gustav%20Hotho | Heinrich Gustav Hotho (Berlin, May 22, 1802 – Berlin, December 25, 1873) was a German historian of art and Right Hegelian. He is famous for being the compiler and editor of Hegel's posthumous work Vorlesungen über die Ästhetik ("Lectures on Aesthetics").
Biography
During boyhood he was affected for two years with blindness consequent on an attack of measles. But recovering his sight he studied so hard as to take his degree at Berlin in 1826. A year of travel spent in visiting Paris, London and the Low Countries determined his vocation.
He came home delighted with the treasures which he had seen, worked laboriously for a higher examination and passed as "docent" in aesthetics and art history. In 1829 he was made professor at the university of Berlin. In 1833 GF Waagen accepted him as assistant in the museum of the Prussian capital; and in 1858 he was promoted to the directorship of the Berlin print-room.
During a long and busy life, in which his time was divided between literature and official duties, Hotho's ambition had always been to master the history of the schools of Germany and the Netherlands. Accordingly what he published was generally confined to those countries. In 1842-1843 he gave to the world his account of German and Flemish painting. From 1853 to 1858 he revised and published anew a part of this work, which he called "The school of Hubert van Eyck, with his German precursors and contemporaries."
His attempt later on to write a history of Christian painting |
https://en.wikipedia.org/wiki/Whey%20protein | Whey protein is a mixture of proteins isolated from whey, the liquid material created as a by-product of cheese production. The proteins consist of α-lactalbumin, β-lactoglobulin, serum albumin and immunoglobulins. Glycomacropeptide also makes up the third largest component but is not a protein. Whey protein is commonly marketed as a protein supplement, and various health claims have been attributed to it. A review published in 2010 in the European Food Safety Authority Journal concluded that the provided literature did not adequately support the proposed claims. For muscle growth, whey protein has been shown to be slightly better compared to other types of protein, such as casein or soy.
Production of whey
Whey is left over when milk is coagulated during the process of cheese production, and contains everything that is soluble from milk after the pH is dropped to 4.6 during the coagulation process. It is a 5% solution of lactose in water and contains the water-soluble proteins of milk as well as some lipid content. Processing can be done by simple drying, or the relative protein content can be increased by removing the lactose, lipids and other non-protein materials. For example, membrane filtration separates the proteins from lipids, lactose and minerals in whey, which is followed by spray drying.
Whey can be denatured by heat. High heat (such as the sustained high temperatures above 72 °C associated with the pasteurization process) denatures whey proteins. While native |
https://en.wikipedia.org/wiki/Clapp%20oscillator | The Clapp oscillator or Gouriet oscillator is an LC electronic oscillator that uses a particular combination of an inductor and three capacitors to set the oscillator's frequency. LC oscillators use a transistor (or vacuum tube or other gain element) and a positive feedback network. The oscillator has good frequency stability.
History
The Clapp oscillator design was published by James Kilton Clapp in 1948 while he worked at General Radio. According to Czech engineer Jiří Vackář, oscillators of this kind were independently developed by several inventors, and one developed by Gouriet had been in operation at the BBC since 1938.
Circuit
The Clapp oscillator uses a single inductor and three capacitors to set its frequency. The Clapp oscillator is often drawn as a Colpitts oscillator that has an additional capacitor () placed in series with the inductor.
The oscillation frequency in Hertz (cycles per second) for the circuit in the figure, which uses a field-effect transistor (FET), is
The capacitors and are usually much larger than , so the term dominates the other capacitances, and the frequency is near the series resonance of and . Clapp's paper gives an example where and are 40 times larger than ; the change makes the Clapp circuit about 400 times more stable than the Colpitts oscillator for capacitance changes of .
Capacitors , and form a voltage divider that determines the amount of feedback voltage applied to the transistor input.
Although, the Clapp circuit |
https://en.wikipedia.org/wiki/Assimilation%20%28biology%29 | ' is the process of absorption of vitamins, minerals, and other chemicals from food as part of the nutrition of an organism. In humans, this is always done with a chemical breakdown (enzymes and acids) and physical breakdown (oral mastication and stomach churning).chemical alteration of substances in the bloodstream by the liver or cellular secretions. Although a few similar compounds can be absorbed in digestion bio assimilation, the bioavailability of many compounds is dictated by this second process since both the liver and cellular secretions can be very specific in their metabolic action (see chirality). This second process is where the absorbed food reaches the cells via the liver.
Most foods are composed of largely indigestible components depending on the enzymes and effectiveness of an animal's digestive tract. The most well-known of these indigestible compounds is cellulose; the basic chemical polymer in the makeup of plant cell walls. Most animals, however, do not produce cellulase; the enzyme needed to digest cellulose. However some animal and species have developed symbiotic relationships with cellulase-producing bacteria (see termites and metamonads.) This allows termites to use the energy-dense cellulose carbohydrate. Other such enzymes are known to significantly improve bio-assimilation of nutrients. Because of the use of bacterial derivatives, enzymatic dietary supplements now contain such enzymes as amylase, glucoamylase, protease, invertase, peptidase |
https://en.wikipedia.org/wiki/Rodrigues%27%20rotation%20formula | In the theory of three-dimensional rotation, Rodrigues' rotation formula, named after Olinde Rodrigues, is an efficient algorithm for rotating a vector in space, given an axis and angle of rotation. By extension, this can be used to transform all three basis vectors to compute a rotation matrix in , the group of all rotation matrices, from an axis–angle representation. In terms of Lie theory, the Rodrigues' formula provides an algorithm to compute the exponential map from the Lie algebra to its Lie group .
This formula is variously credited to Leonhard Euler, Olinde Rodrigues, or a combination of the two. A detailed historical analysis in 1989 concluded that the formula should be attributed to Euler, and recommended calling it "Euler's finite rotation formula." This proposal has received notable support, but some others have viewed the formula as just one of many variations of the Euler–Rodrigues formula, thereby crediting both.
Statement
If is a vector in and is a unit vector describing an axis of rotation about which rotates by an angle according to the right hand rule, the Rodrigues formula for the rotated vector is
The intuition of the above formula is that the first term scales the vector down, while the second skews it (via vector addition) toward the new rotational position. The third term re-adds the height (relative to ) that was lost by the first term.
An alternative statement is to write the axis vector as a cross product of any two nonzero vectors and |
https://en.wikipedia.org/wiki/Ariel%20%28detergent%29 | Ariel is a brand of laundry detergent developed by P&G European Technology Centre in Belgium. The enzymes for the detergent are provided by Novozymes.
History
It was launched in multiple markets between 1967 and 1969. The brand is owned by US multinational Procter & Gamble and is popular in Mexico and India.
References
External links
Official UK Website
Official German Website
Products introduced in 1967
Cleaning product brands
Laundry detergents
Procter & Gamble brands
Consumer goods
British brands |
https://en.wikipedia.org/wiki/Heterokaryon | A heterokaryon is a multinucleate cell that contains genetically different nuclei. Heterokaryotic and heterokaryosis are derived terms. This is a special type of syncytium. This can occur naturally, such as in the mycelium of fungi during sexual reproduction, or artificially as formed by the experimental fusion of two genetically different cells, as e.g., in hybridoma technology.
Etymology
Heterokaryon is from neo-classic Greek hetero, meaning different, and karyon, meaning kernel or in this case nucleus.
The term was coined in 1965, independently by B. Ephrussi and M. Weiss, by H. Harris and J. F. Watkins, and by Y. Okada and F. Murayama.
Occurrence
Heterokaryons are found in the life cycle of yeasts, for example Saccharomyces cerevisiae, a genetic model organism. The heterokaryon stage is produced from the fusion of two haploid cells. This transient heterokaryon can produce further haploid buds, or cell nuclei can fuse and produce a diploid cell, which can then undergo mitosis.
Ciliate protozoans
The term was first used for ciliate protozoans such as Tetrahymena. This has two types of cell nuclei, a large, somatic macronucleus and a small, germline micronucleus. Both exist in a single cell at the same time and carry out different functions with distinct cytological and biochemical properties.
True fungi
Many fungi (notably the arbuscular mycorrhizal fungi) exhibit heterokaryosis. The haploid nuclei within a mycelium may differ from one another not merely by accumulating |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.