source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Victor%20Poor | Victor "Vic" Poor (July 12, 1933 – August 17, 2012) was an American engineer and computer pioneer. At Computer Terminal Corporation (later renamed Datapoint Corporation), he co-created the architecture that was ultimately implemented in the first successful computer microprocessor, the Intel 8008. Subsequently, Computer Terminal Corporation created the first personal computer, the Datapoint 2200 programmable terminal.
Early life
Victor Dale Poor was born in Los Angeles, California, the son of Pinckney Peyton Poor and Leona Lucille Poor (née Mallory). With a passion for radio, he built his own transceiver from collected discarded pieces, and qualified on amateur radio in 1951 (callsign W6JSO).
After high school, Poor joined the United States Navy. While attending electronics training at the Treasure Island Naval Base in 1952, he met his wife, Florence, in a church in San Francisco. On completing his training the couple married in November 1952, and he was then assigned to Ford Island naval base, Pearl Harbor, Hawaii.
Career
On leaving the Navy in 1955, Poor joined the telecommunications technology team at Stromberg-Carlson in San Diego, California. Trained in computer programming, he wrote his first programme in 1956 for the UNIVAC 1103. Recruited by Raytheon, the couple then moved to Massachusetts. Although Poor did not attend college, he took electronics training classes both in the Navy and at Raytheon. A quick learner, he soon knew more than his instructors and began teaching classes himself.
Poor then moved to Maryland to help form radio and telegraph equipment manufacturer Frederick Electronics. Quickly made an executive, he developed the idea of adapting radioteletype (RTTY) machines to send data wirelessly. These were then sold to both the United States Army and later commercial media customers, such as the Associated Press, to send affiliated news reports as data around the world.
Computer engineering
Poor continued his research and development, trying |
https://en.wikipedia.org/wiki/Refrigerator%20truck | A refrigerator truck or chiller lorry (also called a reefer), is a van or truck designed to carry perishable freight at low temperatures. Most long-distance refrigerated transport by truck is done in articulated trucks pulling refrigerated hardside (box) semi-trailers, although insulated curtainsiders are common in some countries. Occasionally, refrigerated trailers have been used as temporary morgues, and second-hand refrigerated trailers are frequently sold for use in tiny home conversions due to their insulation and existing status as a vehicle.
History
The first successful mechanically refrigerated trucks were made for the ice cream industry in 1925. American inventor Frederick McKinley Jones is known to be the first person to invent a refrigerated truck. There were around 4 million refrigerated road vehicles in use in 2010 worldwide.
Features
Like refrigerator cars, refrigerated trucks differ from simple insulated and ventilated vans (commonly used for transporting fruit), neither of which are fitted with cooling apparatus.
Refrigerator trucks can be ice-cooled, equipped with any one of a variety of mechanical refrigeration systems powered by small displacement diesel engines, or utilize carbon dioxide (either as dry ice or in liquid form) as a cooling agent. They are often equipped with small "vent doors" at the rear and front of the trailer. The purpose of these doors is to be kept open while hauling non-refrigerated cargo (often "backhaul"), so as to air out the trailer.
See also
Cold chain
Fuel cell auxiliary power unit
Reefer (container)
Reefer ship
Refrigerated van
Refrigerator car
Temperature data logger |
https://en.wikipedia.org/wiki/Object%20Linking%20and%20Embedding | Object Linking & Embedding (OLE) is a proprietary technology developed by Microsoft that allows embedding and linking to documents and other objects. For developers, it brought OLE Control Extension (OCX), a way to develop and use custom user interface elements. On a technical level, an OLE object is any object that implements the IOleObject interface, possibly along with a wide range of other interfaces, depending on the object's needs.
Overview
OLE allows an editing application to export part of a document to another editing application and then import it with additional content. For example, a desktop publishing system might send some text to a word processor or a picture to a bitmap editor using OLE. The main benefit of OLE is to add different kinds of data to a document from different applications, like a text editor and an image editor. This creates a Compound File Binary Format document and a master file to which the document makes reference. Changes to data in the master file immediately affect the document that references it. This is called "linking" (instead of "embedding").
OLE is also used for transferring data between different applications using drag and drop and clipboard operations.
History
OLE 1.0
OLE 1.0, released in 1990, was an evolution of the original Dynamic Data Exchange (DDE) concept that Microsoft developed for earlier versions of Windows. While DDE was limited to transferring limited amounts of data between two running applications, OLE was capable of maintaining active links between two documents or even embedding one type of document within another.
OLE servers and clients communicate with system libraries using virtual function tables, or VTBLs. The VTBL consists of a structure of function pointers that the system library can use to communicate with the server or client. The server and client libraries, and , were originally designed to communicate between themselves using the message.
OLE 1.0 later evolved to become an ar |
https://en.wikipedia.org/wiki/Canadian%20dollar | The Canadian dollar (symbol: $; code: CAD; ) is the currency of Canada. It is abbreviated with the dollar sign $. There is no standard disambiguating form, but the abbreviations Can$, CA$ and C$ are frequently used for distinction from other dollar-denominated currencies (though C$ remains ambiguous with the Nicaraguan córdoba). It is divided into 100 cents (¢).
Owing to the image of a common loon on its reverse, the dollar coin, and sometimes the unit of currency itself, may be referred to as the loonie by English-speaking Canadians and foreign exchange traders and analysts.
Accounting for approximately 2% of all global reserves, the Canadian dollar is the sixth-most held reserve currency in the world, behind the U.S. dollar, euro, yen, sterling, and renminbi. The Canadian dollar is popular with central banks because of Canada's relative economic soundness, the Canadian government's strong sovereign position, and the stability of the country's legal and political systems.
History
Colonial currencies
The 1850s in Canada were a decade of debate over whether to adopt a £sd-based monetary system or a decimal monetary system based on the US dollar. The British North American provinces, for reasons of practicality in relation to the increasing trade with the neighbouring United States, had a desire to assimilate their currencies with the American unit, but the imperial authorities in London still preferred sterling as the sole currency throughout the British Empire. The British North American provinces nonetheless gradually adopted currencies tied to the American dollar.
Province of Canada
In 1841, the Province of Canada adopted a new system based on the Halifax rating. The new Canadian pound was equal to four US dollars (92.88 grains gold), making £1 sterling equal to £1.4s.4d. Canadian. Thus, the new Canadian pound was worth 16 shillings and 5.3 pence sterling.
In 1851, the Parliament of the Province of Canada passed an act for the purposes of introducing a sterl |
https://en.wikipedia.org/wiki/Interleaving%20distance | In topological data analysis, the interleaving distance is a measure of similarity between persistence modules, a common object of study in topological data analysis and persistent homology. The interleaving distance was first introduced by Frédéric Chazal et al. in 2009. since then, it and its generalizations have been a central consideration in the study of applied algebraic topology and topological data analysis.
Definition
A persistence module is a collection of vector spaces indexed over the real line, along with a collection of linear maps such that is always an isomorphism, and the relation is satisfied for every . The case of indexing is presented here for simplicity, though the interleaving distance can be readily adapted to more general settings, including multi-dimensional persistence modules.
Let and be persistence modules. Then for any , a -shift is a collection of linear maps between the persistence modules that commute with the internal maps of and .
The persistence modules and are said to be -interleaved if there are -shifts and such that the following diagrams commute for all .
It follows from the definition that if and are -interleaved for some , then they are also -interleaved for any positive . Therefore, in order to find the closest interleaving between the two modules, we must take the infimum across all possible interleavings.
The interleaving distance between two persistence modules and is defined as .
Properties
Metric properties
It can be shown that the interleaving distance satisfies the triangle inequality. Namely, given three persistence modules , , and , the inequality is satisfied.
On the other hand, there are examples of persistence modules that are not isomorphic but that have interleaving distance zero. Furthermore, if no suitable exists then two persistence modules are said to have infinite interleaving distance. These two properties make the interleaving distance an extended pseudometric, which means |
https://en.wikipedia.org/wiki/Squeezed%20states%20of%20light | In quantum physics, light is in a squeezed state if its electric field strength Ԑ for some phases has a quantum uncertainty smaller than that of a coherent state. The term squeezing thus refers to a reduced quantum uncertainty. To obey Heisenberg's uncertainty relation, a squeezed state must also have phases at which the electric field uncertainty is anti-squeezed, i.e. larger than that of a coherent state. Since 2019, the gravitational-wave observatories LIGO and Virgo employ squeezed laser light, which has significantly increased the rate of observed gravitational-wave events.
Quantum physical background
An oscillating physical quantity cannot have precisely defined values at all phases of the oscillation. This is true for the electric and magnetic fields of an electromagnetic wave, as well as for any other wave or oscillation (see figure right). This fact can be observed in experiments and is described by quantum theory. For electromagnetic waves usually just the electric field is considered, because it is the one that mainly interacts with matter.
Fig. 1. shows five different quantum states that a monochromatic wave could be in. The difference of the five quantum states is given by different electric field excitations and by different distributions of the quantum uncertainty along the phase . For a displaced coherent state, the expectation (mean) value of the electric field shows an oscillation, with an uncertainty independent of the phase (a). Also the phase- (b) and amplitude-squeezed states (c) show an oscillation of the mean electric field, but here the uncertainty depends on phase and is squeezed for some phases. The vacuum state (d) is a special coherent state and is not squeezed. It has zero mean electric field for all phases and a phase-independent uncertainty. It has zero energy on average, i.e. zero photons, and is the ground state of the monochromatic wave we consider. Finally, a squeezed vacuum state has also a zero mean electric field but a pha |
https://en.wikipedia.org/wiki/Dictionary%20of%20Contemporary%20Finnish | Dictionary of Contemporary Finnish (, previously known as the New Dictionary of Modern Finnish) is the most recent dictionary of the modern Finnish language. It is edited by the Institute for the Languages of Finland. The current printed edition was first published in 2006 and is based on the 2004 digital version of the same name. The printed version lacks some features of the electronic version, such as the latter's index of Finnish place names (asuinpaikkahakemisto).
The dictionary is monolingual and attempts to "accurately reflect standard Finnish as it is used today". It explains the meanings of most headwords and provides information about their spelling, inflected forms, stylistic register, usage, and context.
Scope
The dictionary consists of almost 100,000 entries. It is an expanded and modernized version of the Suomen kielen perussanakirja (Basic Dictionary of the Finnish Language), printed between 1990 and 1994 (and digitalized as the CD-Perussanakirja in 1997). Compared to the Suomen kielen perussanakirja, the 2006 edition has about 6,500 new entries and about 20,000 modified entries. Together with the Nykysuomen sanakirja, the new dictionaries form a very comprehensive collection of information on the modern Finnish language.
See also
Nykysuomen sanakirja |
https://en.wikipedia.org/wiki/Proton%20coupled%20amino%20acid%20transporter | Proton-coupled amino acid transporters belong to the SLC26A5 family; they are protein receptors whose main function is the transmembrane movement of amino acids and their derivatives. This family of receptors is most commonly found within the luminal surface of the small intestine as well as in some lysosomes. The solute carrier family (SLC) of genes includes roughly 400 membrane proteins that are characterized by 66 families in total. The SLC36 family of genes maps to chromosome 11. The diversity of these receptors is vast, with the ability to transport both charged and uncharged amino acids along with their derivatives. In research and practice, SLC36A1/2 are both targets for drug-based delivery systems for a wide range of disorders.
Structure
The human protein acid transporter (hPAT1) is 5585 base pairs long and codes for a protein 476 amino acids long. The transporter has nine transmembrane regions where the amino terminus faces the cytoplasm. The rat protein acid transporter (rPAT1) has been widely studied and an 85% amino acid sequence match was found between hPAT1 and rPAT1. The hPAT1 gene is located on chromosome 5q31-33 and has 11 exons that are coding regions. Its translation site begins in exon 2 and exon 11 contains the termination site.
Proton-coupled amino acid transporters 1 and 2
The molecular weight of Proton-coupled amino acid transporter 1 is 53.28 kDA; the molecular weight of Proton-coupled amino acid transporter 2 is 53.22 kDA. PAT1 has been found in lysosomes in brain neurons but also in the apical membrane of intestinal epithelial cells where it is associated with the brush border. Proton-coupled amino acid transporter 1 has a higher affinity for proline than it does for glycine and alanine Proton-coupled amino acid transporter 2 is found subcellularly in the kidneys, lungs, spinal cord, and brain and likely has a role in myelinating neurons. It has an overall higher affinity for glycine, alanine, and proline than PAT1 but is more specific |
https://en.wikipedia.org/wiki/Lime%20sulfur | In horticulture, lime sulphur (American spelling lime sulfur) is mainly a mixture of calcium polysulfides and thiosulfate (plus other reaction by-products as sulfite and sulfate) formed by reacting calcium hydroxide with elemental sulfur, used in pest control. It can be prepared by boiling in water a suspension of poorly soluble calcium hydroxide (lime) and solid sulfur together with a small amount of surfactant to facilitate the dispersion of these solids in water. After elimination of any residual solids (flocculation, decantation and filtration), it is normally used as an aqueous solution, which is reddish-yellow in colour and has a distinctive offensive odour of hydrogen sulfide (H2S, rotten eggs).
Synthesis reaction
The exact chemical reaction leading to the synthesis of lime sulfur is poorly known and is generally written as:
as reported in a document of the US Department of Agriculture (USDA).
This vague reaction is puzzling because it involves the reduction of elemental sulfur and no reductant appears in the above mentioned equation while sulfur oxidation products are also mentioned. The initial pH of the solution imposed by poorly soluble hydrated lime is alkaline (pH = 12.5) while the final pH is in range 11–12, typical for sulfides which are also strong bases.
When the hydrolysis of calcium sulfide is accounted for, the individual reactions for each of the by-products are:
However, elemental sulfur can undergo a disproportionation reaction, also called dismutation. The first reaction resembles a disproportionation reaction. The inverse comproportionation reaction is the reaction occurring in the Claus process used for desulfurisation of oil and gas crude products in the refining industry:
By rewriting the last reaction in the inverse direction one obtains a reaction consistent with what is observed in the lime sulfur global reaction:
In alkaline conditions, it gives:
and after simplification, or more exactly recycling, of water molecules in the a |
https://en.wikipedia.org/wiki/Synaptonemal%20complex | The synaptonemal complex (SC) is a protein structure that forms between homologous chromosomes (two pairs of sister chromatids) during meiosis and is thought to mediate synapsis and recombination during prophase I during meiosis in eukaryotes. It is currently thought that the SC functions primarily as a scaffold to allow interacting chromatids to complete their crossover activities.
Composition
The synaptonemal complex is a tripartite structure consisting of two parallel lateral regions and a central element. This "tripartite structure" is seen during the pachytene stage of the first meiotic prophase, both in males and in females during gametogenesis. Previous to the pachytene stage, during leptonema, the lateral elements begin to form and they initiate and complete their pairing during the zygotene stage. After pachynema ends, the SC usually becomes disassembled and can no longer be identified.
In humans, three specific components of the synaptonemal complex have been characterized: SC protein-1 (SYCP1), SC protein-2 (SYCP2), and SC protein-3 (SYCP3). The SYCP1 gene is on chromosome 1p13; the SYCP2 gene is on chromosome 20q13.33; and the gene for SYCP3 is on chromosome 12q.
The synaptonemal complex was described by Montrose J. Moses in 1956 in primary spermatocytes of crayfish and by D. Fawcett in spermatocytes of pigeon, cat and man. As seen with the electron microscope, the synaptonemal complex is formed by two "lateral elements", mainly formed by SYCP3 and secondarily by SYCP2, a "central element" that contains at least two additional proteins and the amino terminal region of SYCP1, and a "central region" spanned between the two lateral elements, that contains the "transverse filaments" composed mainly by the protein SYCP1.
The SCs can be seen with the light microscope using silver staining or with immunofluorescence techniques that label the proteins SYCP3 or SYCP2.
Assembly and disassembly
Formation of the SC usually reflects the pairing or "synapsis" o |
https://en.wikipedia.org/wiki/ORFeome | In, molecular genetics, an ORFeome refers to the complete set of open reading frames (ORFs) in a genome. The term may also be used to describe a set of cloned ORFs. ORFs correspond to the protein coding sequences (CDS) of genes. ORFs can be found in genome sequences by computer programs such as GENSCAN and then amplified by PCR. While this is relatively trivial in bacteria the problem is non-trivial in eukaryotic genomes because of the presence of introns and exons as well as splice variants.
Use in research
The usage of complete ORFeomes reflects a new trend in biology that can be succinctly summarized as omics. ORFeomes are used for the study of protein-protein interactions, protein microarrays, the study of antigens, and other fields of study.
Cloned ORFeomes
Complete ORF sets have been cloned for a number of organisms including Brucella melitensis,
Chlamydia pneumoniae,
Escherichia coli,
Neisseria gonorrhoeae,
Pseudomonas aeruginosa, Schizosaccharomyces pombe, Staphylococcus aureus
and human herpesviruses
A partial human ORFeome has also been produced. |
https://en.wikipedia.org/wiki/Universal%20approximation%20theorem | In the mathematical theory of artificial neural networks, universal approximation theorems are results that put limits on what neural networks can theoretically learn, i.e. that establish the density of an algorithmically generated class of functions within a given function space of interest. Typically, these results concern the approximation capabilities of the feedforward architecture on the space of continuous functions between two Euclidean spaces, and the approximation is with respect to the compact convergence topology. What must be stressed, is that while some functions can be arbitrarily well approximated in a region, the proofs do not apply outside of the region, i.e. the approximated functions do not extrapolate outside of the region. That applies for all non-periodic activation functions, i.e. what's in practice used and most proofs assume.
However, there are also a variety of results between non-Euclidean spaces and other commonly used architectures and, more generally, algorithmically generated sets of functions, such as the convolutional neural network (CNN) architecture, radial basis functions, or neural networks with specific properties. Most universal approximation theorems can be parsed into two classes. The first quantifies the approximation capabilities of neural networks with an arbitrary number of artificial neurons ("arbitrary width" case) and the second focuses on the case with an arbitrary number of hidden layers, each containing a limited number of artificial neurons ("arbitrary depth" case). In addition to these two classes, there are also universal approximation theorems for neural networks with bounded number of hidden layers and a limited number of neurons in each layer ("bounded depth and bounded width" case).
Universal approximation theorems imply that neural networks can represent a wide variety of interesting functions with appropriate weights. On the other hand, they typically do not provide a construction for the weights, but m |
https://en.wikipedia.org/wiki/Advanced%20Digital%20Broadcast | Advanced Digital Broadcast (ADB) is a company which provides software, system and services to pay-TV and telecommunication operators, content distributors and property owners around the world. The company specializes also in the development of digital connectivity devices such as set-top boxes and residential gateways.
ADB's global headquarters are located in Bellevue, Switzerland. The company has research and development facilities in Poland and Italy and an Operations division in Taipei, Taiwan. ADB has local offices in several countries in Europe and the United States.
Founded in 1995, ADB initially focused on developing and marketing software for digital TV processors and expanded its business to the design and manufacture of digital TV equipment in 1997. The company sold its first set-top box in 1997 and since then has been delivering a number of set-top boxes, and residential gateways, together with advanced software platforms. ADB has sold over 64 million devices worldwide to cable, satellite, IPTV and broadband operators. ADB employs over 350 people, of which 70% are in engineering functions.
History
In 1995, ADB was founded by Andrew Rybicki, Janusz C. Szajna, Krzysztof Kolbuszewski, Mariusz Walkowiak, and Stanley Hemphill with an initial focus on developing and marketing software for advanced digital TV processors.
In 1997, ADB started designing and manufacturing of digital TV equipment. ADB designed its first commercial set-to box, and established its dedicated R&D facility in Poland and corporate headquarters in Taiwan.
Between 1998 and 2000 offices were opened in Australia, and Spain, and ADB sold its one millionth set-top box.
In 2001, ADB announced the development of an open standard set-top box middleware based on the Multimedia Home Platform (MHP) specification and established its worldwide headquarters in Geneva, Switzerland.
In 2002, the company opened its Americas headquarters in Chicago (which has since moved to Denver). and became the w |
https://en.wikipedia.org/wiki/Transducer | A transducer is a device that converts energy from one form to another. Usually a transducer converts a signal in one form of energy to a signal in another.
Transducers are often employed at the boundaries of automation, measurement, and control systems, where electrical signals are converted to and from other physical quantities (energy, force, torque, light, motion, position, etc.). The process of converting one form of energy to another is known as transduction.
Types
Mechanical transducers, so-called as they convert physical quantities into mechanical outputs or vice versa;
Electrical transducers however convert physical quantities into electrical outputs or signals. Examples of these are:
a thermocouple that changes temperature differences into a small voltage;
a linear variable differential transformer (LVDT), used to measure displacement (position) changes by means of electrical signals.
Sensors, actuators and transceivers
Transducers can be categorized by which direction information passes through them:
A sensor is a transducer that receives and responds to a signal or stimulus from a physical system. It produces a signal, which represents information about the system, which is used by some type of telemetry, information or control system.
An actuator is a device that is responsible for moving or controlling a mechanism or system. It is controlled by a signal from a control system or manual control. It is operated by a source of energy, which can be mechanical force, electrical current, hydraulic fluid pressure, or pneumatic pressure, and converts that energy into motion. An actuator is the mechanism by which a control system acts upon an environment. The control system can be simple (a fixed mechanical or electrical system), software-based (e.g. a printer driver, robot control system), a human, or any other input.
Bidirectional transducers can convert physical phenomena to electrical signals and electrical signals into physical phenomena. An exa |
https://en.wikipedia.org/wiki/ABTS | In biochemistry, ABTS (2,2'-azino-bis(3-ethylbenzothiazoline-6-sulfonic acid)) is a chemical compound used to observe the reaction kinetics of specific enzymes. A common use for it is in the enzyme-linked immunosorbent assay (ELISA) to detect the binding of molecules to each other.
It is commonly used as a substrate with hydrogen peroxide for a peroxidase enzyme (such as horseradish peroxidase) or alone with blue multicopper oxidase enzymes (such as laccase or bilirubin oxidase). Its use allows the reaction kinetics of peroxidases themselves to be followed. In this way it also can be used to indirectly follow the reaction kinetics of any hydrogen peroxide-producing enzyme, or to simply quantify the amount of hydrogen peroxide in a sample.
The formal reduction potentials for ABTS are high enough for it to act as an electron donor for the reduction of oxo species such as molecular oxygen and hydrogen peroxide, particularly at the less-extreme pH values encountered in biological catalysis. Under these conditions, the sulfonate groups are fully deprotonated and the mediator exists as a dianion.
ABTS–· + e– → ABTS2– E° = 0.67 V vs SHE
ABTS + e– → ABTS–· E° = 1.08 V vs SHE
This compound is chosen because the enzyme facilitates the reaction with hydrogen peroxide, turning it into a green and soluble end-product. Its new absorbance maximum of 420 nm light (ε = 3.6 × 104 M–1 cm–1) can easily be followed with a spectrophotometer, a common laboratory instrument.
It is sometimes used as part of a glucose estimating reagent when finding glucose concentrations of solutions such as blood serum.
ABTS is also frequently used by the food industry and agricultural researchers to measure the antioxidant capacities of foods. In this assay, ABTS is converted to its radical cation by addition of sodium persulfate. This radical cation is blue in color and absorbs light at 415, 645, 734 and 815 nm. The ABTS radical cation is reactive towards most antioxidants including phenolics, th |
https://en.wikipedia.org/wiki/Dereverberation | Dereverberation is the process by which the effects of reverberation are removed from sound, after such reverberant sound has been picked up by microphones. Dereverberation is a subtopic of acoustic digital signal processing and is most commonly applied to speech but also has relevance in some aspects of music processing. Dereverberation of audio (speech or music) is a corresponding function to blind deconvolution of images, although the techniques used are usually very different. Reverberation itself is caused by sound reflections in a room (or other enclosed space) and is quantified by the room reverberation time and the direct-to-reverberant ratio. The effect of dereverberation is to increase the direct-to-reverberant ratio so that the sound is perceived as closer and clearer.
A main application of dereverberation is in hands-free phones and desktop conferencing terminals because, in these cases, the microphones are not close to the source of sound – the talker’s mouth – but at arm’s length or further distance. As well as telecommunications, dereverberation is importantly applied in automatic speech recognition because speech recognizers are usually error-prone in reverberant scenarios.
Dereverberation became established as a topic of scientific research in the years 2000 to 2005., although a few notable early articles exist. The first scientific text book on the topic was published in 2010. A global scientific study sponsored by the IEEE Technical Committee for Audio and Acoustic Signal Processing took place in 2014.
Three different approaches can be followed to perform dereverberation. In the first approach, reverberation is cancelled by exploiting a mathematical model of the acoustic system (or room) and, after estimation of the room acoustic model parameters, forming an estimate for the original signal. In the second approach, reverberation is suppressed by treating it as a type of (convolutional) noise and performing a de-noising process specifically ad |
https://en.wikipedia.org/wiki/Fernandez%20reaction | The Fernandez reaction is a reaction that occurs to signal a positive result in the lepromin skin test for leprosy. The reaction occurs in the skin at the site of injection if the body possesses antibodies to the Dharmendra antigen, one of the antigens found in Mycobacterium leprae, the bacteria that causes leprosy. The reaction occurs via a delayed-type hypersensitivity mechanism. This reaction occurs within 48 hours of injection of lepromin and is seen in only tuberculoid forms of leprosy. This represents a delayed-type hypersensitivity reaction. In contrast, the Mitsuda reaction (delayed granulomatous lesion) occurs 3–4 weeks after injection of lepromin and is only seen in patients with the tuberculoid form of leprosy (not the lepromatous form, in which the body does not mount a strong response against the bacterium). In terms of mechanism of action and appearance, the reaction is similar to the tuberculin reaction of a positive Mantoux test for tuberculosis. |
https://en.wikipedia.org/wiki/Isotone | Two nuclides are isotones if they have the same neutron number N, but different proton number Z. For example, boron-12 and carbon-13 nuclei both contain 7 neutrons, and so are isotones. Similarly, 36S, 37Cl, 38Ar, 39K, and 40Ca nuclei are all isotones of 20 because they all contain 20 neutrons. Despite its similarity to the Greek for "same stretching", the term was formed by the German physicist K. Guggenheimer by changing the "p" in "isotope" from "p" for "proton" to "n" for "neutron".
The largest numbers of observationally stable nuclides exist for isotones 50 (five: 86Kr, 88Sr, 89Y, 90Zr, 92Mo) and 82 (six: 138Ba, 139La, 140Ce, 141Pr, 142Nd, 144Sm). Neutron numbers for which there are no stable isotones are 19, 21, 35, 39, 45, 61, 89, 115, 123, and 127 or more. In contrast, the proton numbers for which there are no stable isotopes are 43, 61, and 83 or more. This is related to nuclear magic numbers, the number of nucleons forming complete shells within the nucleus, e.g. 2, 8, 20, 28, 50, 82, and 126. No more than one stable nuclide has the same odd neutron number, except for 1 (2H and 3He), 5 (9Be and 10B), 7 (13C and 14N), 55 (97Mo and 99Ru), and 107 (179Hf and 180mTa). Odd neutron numbers for which there is a stable nuclide and a primordial radionuclide are 27 (50V), 65 (113Cd), 81 (138La), 85 (147Sm), and 105 (176Lu). Neutron numbers for which there are two primordial radionuclides are 88 (151Eu and 152Gd) and 112 (187Re and 190Pt).
See also
Isotopes are nuclides having the same number of protons: e.g. carbon-12 and carbon-13.
Isobars are nuclides having the same mass number (i.e. sum of protons plus neutrons): e.g. carbon-12 and boron-12.
Nuclear isomers are different excited states of the same type of nucleus. A transition from one isomer to another is accompanied by emission or absorption of a gamma ray, or the process of internal conversion. (Not to be confused with chemical isomers.)
Notes
Nuclear physics |
https://en.wikipedia.org/wiki/Polypide | The polypide in bryozoans encompasses most of the organs and tissues of each individual zooid. This includes the tentacles, tentacle sheath, U-shaped digestive tract, musculature and nerve cells. It is housed in the zooidal exoskeleton, which in cyclostomes is tubular and in cheilostomes is box-shaped.
See also
Bryozoan Anatomy |
https://en.wikipedia.org/wiki/Diplo%C3%AB | Diploë ( or ) is the spongy cancellous bone separating the inner and outer layers of the cortical bone of the skull. It is a subclass of trabecular bone.
In the cranial bones, the layers of compact cortical tissue are familiarly known as the tables of the skull; the outer one is thick and tough; the inner is thin, dense, and brittle, and hence is termed the vitreous table. The intervening cancellous tissue is called the diploë. In certain regions of the skull this becomes absorbed so as to leave spaces filled with liquid between the two tables.
Etymology
From Ancient Greek διπλόη (diplóē, “literally, a fold”), noun use of feminine of διπλόος (diplóos, “double”) |
https://en.wikipedia.org/wiki/Order%20%28biology%29 | Order () is one of the eight major hierarchical taxonomic ranks in Linnaean taxonomy. It is classified between family and class. In biological classification, the order is a taxonomic rank used in the classification of organisms and recognized by the nomenclature codes. An immediately higher rank, superorder, is sometimes added directly above order, with suborder directly beneath order. An order can also be defined as a group of related families.
What does and does not belong to each order is determined by a taxonomist, as is whether a particular order should be recognized at all. Often there is no exact agreement, with different taxonomists each taking a different position. There are no hard rules that a taxonomist needs to follow in describing or recognizing an order. Some taxa are accepted almost universally, while others are recognized only rarely.
The name of an order is usually written with a capital letter. For some groups of organisms, their orders may follow consistent naming schemes. Orders of plants, fungi, and algae use the suffix (e.g. Dictyotales). Orders of birds and fishes use the Latin suffix meaning 'having the form of' (e.g. Passeriformes), but orders of mammals and invertebrates are not so consistent (e.g. Artiodactyla, Actiniaria, Primates).
Hierarchy of ranks
Zoology
For some clades covered by the International Code of Zoological Nomenclature, several additional classifications are sometimes used, although not all of these are officially recognized.
In their 1997 classification of mammals, McKenna and Bell used two extra levels between superorder and order: grandorder and mirorder. Michael Novacek (1986) inserted them at the same position. Michael Benton (2005) inserted them between superorder and magnorder instead. This position was adopted by Systema Naturae 2000 and others.
Botany
In botany, the ranks of subclass and suborder are secondary ranks pre-defined as respectively above and below the rank of order. Any number of further ran |
https://en.wikipedia.org/wiki/White%20hole | In general relativity, a white hole is a hypothetical region of spacetime and singularity that cannot be entered from the outside, although energy-matter, light and information can escape from it. In this sense, it is the reverse of a black hole, from which energy-matter, light and information cannot escape. White holes appear in the theory of eternal black holes. In addition to a black hole region in the future, such a solution of the Einstein field equations has a white hole region in its past. This region does not exist for black holes that have formed through gravitational collapse, however, nor are there any observed physical processes through which a white hole could be formed.
Supermassive black holes (SMBHs) are theoretically predicted to be at the center of every galaxy and that possibly, a galaxy cannot form without one. Stephen Hawking and others have proposed that these supermassive black holes spawn a supermassive white hole.
Overview
Like black holes, white holes have properties like mass, charge, and angular momentum. They attract matter like any other mass, but objects falling towards a white hole would never actually reach the white hole's event horizon (though in the case of the maximally extended Schwarzschild solution, discussed below, the white hole event horizon in the past becomes a black hole event horizon in the future, so any object falling towards it will eventually reach the black hole horizon).
Imagine a gravitational field, without a surface. Acceleration due to gravity is the greatest on the surface of any body. But since black holes lack a surface, acceleration due to gravity increases exponentially, but never reaches a final value as there is no considered surface in a singularity.
In quantum mechanics, the black hole emits Hawking radiation and so it can come to thermal equilibrium with a gas of radiation (not compulsory). Because a thermal-equilibrium state is time-reversal-invariant, Stephen Hawking argued that the time reversa |
https://en.wikipedia.org/wiki/Platelet-derived%20growth%20factor | Platelet-derived growth factor (PDGF) is one among numerous growth factors that regulate cell growth and division. In particular, PDGF plays a significant role in blood vessel formation, the growth of blood vessels from already-existing blood vessel tissue, mitogenesis, i.e. proliferation, of mesenchymal cells such as fibroblasts, osteoblasts, tenocytes, vascular smooth muscle cells and mesenchymal stem cells as well as chemotaxis, the directed migration, of mesenchymal cells. Platelet-derived growth factor is a dimeric glycoprotein that can be composed of two A subunits (PDGF-AA), two B subunits (PDGF-BB), or one of each (PDGF-AB).
PDGF is a potent mitogen for cells of mesenchymal origin, including fibroblasts, smooth muscle cells and glial cells. In both mouse and human, the PDGF signalling network consists of five ligands, PDGF-AA through -DD (including -AB), and two receptors, PDGFRalpha and PDGFRbeta. All PDGFs function as secreted, disulphide-linked homodimers, but only PDGFA and B can form functional heterodimers.
Though PDGF is synthesized, stored (in the alpha granules of platelets), and released by platelets upon activation, it is also produced by other cells including smooth muscle cells, activated macrophages, and endothelial cells
Recombinant PDGF is used in medicine to help heal chronic ulcers and in orthopedic surgery and periodontics as an alternative to bone autograft to stimulate bone regeneration and repair.
Types and classification
There are five different isoforms of PDGF that activate cellular response through two different receptors. Known ligands include: PDGF-AA (PDGFA), -BB (PDGFB), -CC (PDGFC), and -DD (PDGFD), and -AB (a PDGFA and PDGFB heterodimer). The ligands interact with the two tyrosine kinase receptor monomers, PDGFRα (PDGFRA) and -Rβ (PDGFRB). The PDGF family also includes a few other members of the family, including the VEGF sub-family.
Mechanisms
The receptor for PDGF, PDGFR is classified as a receptor tyrosine kinase |
https://en.wikipedia.org/wiki/Temporal%20logic%20of%20actions | Temporal logic of actions (TLA) is a logic developed by Leslie Lamport, which combines temporal logic with a logic of actions.
It is used to describe behaviours of concurrent and distributed systems. It is the logic underlying the specification language TLA+.
Details
Statements in the temporal logic of actions are of the form , where A is an action and t contains a subset of the variables appearing in A. An action is an expression containing primed and non-primed variables, such as . The meaning of the non-primed variables is the variable's value in this state. The meaning of primed variables is the variable's value in the next state.
The above expression means the value of x today, plus the value of x tomorrow times the value of y today, equals the value of y tomorrow.
The meaning of is that either A is valid now, or the variables appearing in t do not change. This allows for stuttering steps, in which none of the program variables change their values.
See also
Temporal logic
PlusCal
TLA+ |
https://en.wikipedia.org/wiki/Edward%20B.%20Saff | Edward Barry Saff (born 2 January 1944 in New York City) is an American mathematician, specializing in complex analysis, approximation theory, numerical analysis, and potential theory.
Education and career
Saff received in 1964 his bachelor's degree from the Georgia Institute of Technology and in 1968 his PhD from the University of Maryland, College Park under Joseph L. Walsh with thesis Interpolation and Functions of Class H (k, a, 2). As a postdoc he was a Fulbright Fellow at Imperial College London from 1968 to 1969. At the University of South Florida he was from 1969 to 1971 an assistant professor, from 1971 to 1976 an associate professor, from 1976 to 1986 a full professor, and from 1986 to 2001 a distinguished research professor. At Vanderbilt University he is, since 2001, a professor and director of the Center for Constructive Approximation and was from 2004 to 2007 the Executive Dean of the College of Arts and Sciences.
His research deals with approximation of complex functions by polynomials and rational functions, approximate solutions of differential equations, Padé approximants, geometry of polynomials, special functions, Hardy spaces, conformal mappings (including numerical analysis). and potential theory (minima of energy under boundary value constraints or external fields). He is the author or coauthor of over 240 research articles, the coauthor of 9 books, and the coeditor of 11 volumes. He was the coeditor, with Theodore J. Rivlin, of a volume of Joseph L. Walsh's Selected Works published in 2000. Since 2007 he is an ISI Highly Cited Researcher.
For the academic year 1978–1979 Saff was a Guggenheim Fellow at the University of Oxford. In 2012 he was elected a Fellow of the American Mathematical Society. He was elected in 2013 a Foreign Member of the Bulgarian Academy of Sciences and made in 1987 an Honorary Professor of the Zhejiang Normal University in China.
Selected publications
Books as author
with A. D. Snider: Fundamentals of Complex Anal |
https://en.wikipedia.org/wiki/Pachicom | is a parlor video game released for multiple platforms that is themed around pachinko. It was released in 1985 exclusively to the Japanese market.
The game of pachinko is hugely popular in Japan, since it involves playing pinball with an element of slot machines mixed in along with gambling. There are four different gameplay modes to choose from (two of them assign a player with a pachinko machine while the other two allow players to choose their own machine).
Gameplay
The game is based on standard pachinko rules, replicating the experience of playing pachinko in a parlor. The game board has a layout with nails sticking out. Metal balls would either of or fall in between the holes created by the pegs in order to traverse the board. Certain compartments are added to the board. Should a metal ball find its way to one of these compartments, the player will receive points that are similar to credits on a slot machine. Players can expect to lose that ball (and a point) if it falls below the bottom of the board (similar to falling down a bottomless pit in an action game). Initially, the player starts off with one hundred balls (points).
In Mode A, the clock goes forward from 000000 seconds until it reaches 999999 seconds (the equivalent of either 16666.65 non-real time minutes, 2777.775 non-real time hours, or 115.740625 non-real time days). Once that occurs, the game is over and any remaining points are stored in the top three score list. Mode B is the same except that the clock goes backwards from 999999 seconds all the way back to 000000 seconds. It is very easy to mash the "A" button repeatedly around with the direction pad to the right in order to accumulate mass winnings from the pachinko games and "cheat" the system.
Hidden messages
There is a hidden message in the Famicom version that can be found in any hex editor. Almost five percent of the entire ROM (2.05 kilobytes out of the 41 kilobyte ROM image) is taken up by this otherwise inaccessible message. The m |
https://en.wikipedia.org/wiki/Application-layer%20framing | Application-layer framing or application-level framing (ALF) is a method of allowing an application to use its semantics for the design of its network protocols.
This procedure was first proposed by D. D. Clark and David L. Tennenhouse. It works as follows:
The application splits the data into useful segments.
These segments are called ADUs (application data units).
The ADUs can be processed in any order.
The lower layers keep the ADU borders.
This procedure simplifies the quality of service negotiation and provides a simpler method of error checking.
The Real-time Transport Protocol (RTP) is an example of where the semantics of the real-time application are used to segment the data. |
https://en.wikipedia.org/wiki/David%20E.%20Shaw | David Elliot Shaw (born March 29, 1951) is an American billionaire scientist and former hedge fund manager. He founded D. E. Shaw & Co., a hedge fund company which was once described by Fortune magazine as "the most intriguing and mysterious force on Wall Street". A former assistant professor in the computer science department at Columbia University, Shaw made his fortune exploiting inefficiencies in financial markets with the help of state-of-the-art high speed computer networks. In 1996, Fortune magazine referred to him as "King Quant" because of his firm's pioneering role in high-speed quantitative trading. In 2001, Shaw turned to full-time scientific research in computational biochemistry, more specifically molecular dynamics simulations of proteins.
Early life and education
Shaw was raised in Los Angeles, California. His father was a theoretical physicist who specialised in plasma and fluid flows, and his mother is an artist and educator. They divorced when he was 12. His stepfather, Irving Pfeffer, was professor of finance at University of California, Los Angeles, and the author of papers supporting the efficient market hypothesis.
Shaw earned a bachelor's degree summa cum laude from the University of California, San Diego, a PhD from Stanford University in 1980, and then became an assistant professor of the department of computer science at Columbia University. While at Columbia, Shaw conducted research in massively parallel computing with the NON-VON supercomputer. This supercomputer was composed of processing elements in a tree structure meant to be used for fast relational database searches. Earlier in his career, he founded Stanford Systems Corporation.
Investment career
In 1986, he joined Morgan Stanley, as Vice President for Technology in Nunzio Tartaglia's automated proprietary trading group. In 1994, Shaw was appointed by President Clinton to the President's Council of Advisors on Science and Technology, where he was chairman of the Panel on Educat |
https://en.wikipedia.org/wiki/Ettercap%20%28software%29 | Ettercap is a free and open source network security tool for man-in-the-middle attacks on a LAN. It can be used for computer network protocol analysis and security auditing. It runs on various Unix-like operating systems including Linux, Mac OS X, BSD and Solaris, and on Microsoft Windows. It is capable of intercepting traffic on a network segment, capturing passwords, and conducting active eavesdropping against a number of common protocols. Its original developers later founded Hacking Team.
Functionality
Ettercap works by putting the network interface into promiscuous mode and by ARP poisoning the target machines. Thereby it can act as a 'man in the middle' and unleash various attacks on the victims. Ettercap has plugin support so that the features can be extended by adding new plugins.
Features
Ettercap supports active and passive dissection of many protocols (including ciphered ones) and provides many features for network and host analysis. Ettercap offers four modes of operation:
IP-based: packets are filtered based on IP source and destination.
MAC-based: packets are filtered based on MAC address, useful for sniffing connections through a gateway.
ARP-based: uses ARP poisoning to sniff on a switched LAN between two hosts (full-duplex).
PublicARP-based: uses ARP poisoning to sniff on a switched LAN from a victim host to all other hosts (half-duplex).
In addition, the software also offers the following features:
Character injection into an established connection: characters can be injected into a server (emulating commands) or to a client (emulating replies) while maintaining a live connection.
SSH1 support: the sniffing of a username and password, and even the data of an SSH1 connection. Ettercap is the first software capable of sniffing an SSH connection in full duplex.
HTTPS support: the sniffing of HTTP SSL secured data—even when the connection is made through a proxy.
Remote traffic through a GRE tunnel: the sniffing of remote traffic through a |
https://en.wikipedia.org/wiki/Spice%20%28oceanography%29 | In oceanography, the term spice refers to spatial variations in the temperature and salinity of seawater whose effects on density cancel each other. Such density compensated thermohaline variability is ubiquitous in the upper ocean. Warmer, saltier water is more spicy while cooler, less salty water is more minty. For a density ratio of 1, all the thermohaline variability is spice, and there are no density fluctuations. |
https://en.wikipedia.org/wiki/Dorsal%20artery%20of%20the%20penis | The dorsal artery of the penis is an artery on the top surface of the penis. It is a branch of the internal pudendal artery. It runs forward on the dorsum of the penis to the glans, where it divides into two branches to the glans penis and the foreskin (prepuce).
The dorsal artery of the penis supplies the integument and fibrous sheath of the corpus cavernosum penis, the glans penis, the foreskin, and the skin of the distal shaft. It also branches with circumflex arteries that supply the corpus spongiosum. Its role in erectile function is unknown.
The dorsal artery of the penis may be damaged in traumatic amputation of the penis and repairing the dorsal artery surgically prevents skin loss, but it is not essential for sexual and urinary function. Its hemodynamics and blood pressure can be assessed to test for sexual impairment.
Structure
The dorsal artery of the penis is a branch of the internal pudendal artery. It ascends between the crus penis and the pubic symphysis of the pelvis. As it pierces the perineal membrane, it passes between the two layers of the suspensory ligament of the penis. It runs forward on the dorsum of the penis to the glans. At the glans, it divides into two branches to the glans penis and the foreskin (prepuce). On the penis, it lies between the dorsal nerve and deep dorsal vein.
The dorsal artery give perforators to the corpus cavernosum penis. It sends branches through the fibrous sheath of the corpus cavernosum penis to anastomose with the deep artery of the penis. It anastomoses with the artery of bulb of penis.
Function
The dorsal artery of the penis supplies the fibrous sheath of the corpus cavernosum penis. It gives branches to the glans penis, and the foreskin (prepuce). Through retrograde flow they help supply the skin of the distal shaft. It also gives branches to the circumflex arteries that supply the corpus spongiosum.
The dorsal artery of the penis gives perforators to the corpus cavernosum penis. Despite this, their |
https://en.wikipedia.org/wiki/Sensor-based%20sorting | Sensor-based sorting, is an umbrella term for all applications in which particles are detected using a sensor technique and rejected by an amplified mechanical, hydraulic or pneumatic process.
The technique is generally applied in mining, recycling and food processing and used in the particle size range between . Since sensor-based sorting is a single particle separation technology, the throughput is proportional to the average particle size and weight fed onto the machine.
Functional principle of sensor-based sorting
The main subprocesses of sensor-based sorting are material conditioning, material presentation, detection, data processing and separation.
Material conditioning includes all operations which prepare the particles for being detected by the sensor. All optical sensors need clean material to be able to detect optical characteristics. Conditioning includes screening and cleaning of the feed material.
The aim of the material presentation is the isolation of the particles by creating a single particle layer with the densest surface cover possible without particles touching each other and enough distance to each other allowing for a selective detection and rejection of each single particle.
There are two types of sensor-based sorters: the chute type and the belt type. For both types the first step in acceleration is spreading out the particles by a vibrating feeder followed by either a fast belt or a chute. On the belt type the sensor usually detects the particles horizontally while they pass it on the belt. For the chute type the material detection is usually done vertically while the material passes the sensor in a free fall. The data processing is done in real time by a computer. The computer transfers the result of the data processing to an ultra fast ejection unit which, depending on the sorting decision, ejects a particle or lets it pass.
Sensor-based ore sorting
Sensor-based ore sorting is the terminology used in the mining industry. It i |
https://en.wikipedia.org/wiki/Mark%20Kac | Mark Kac ( ; Polish: Marek Kac; August 3, 1914 – October 26, 1984) was a Polish American mathematician. His main interest was probability theory. His question, "Can one hear the shape of a drum?" set off research into spectral theory, the idea of understanding the extent to which the spectrum allows one to read back the geometry. (In the end, the answer was "no", in general.)
Biography
He was born to a Polish-Jewish family; their town, Kremenets (Polish: "Krzemieniec"), changed hands from the Russian Empire (by then Soviet Ukraine) to Poland after the Peace of Riga, when Kac was a child.
Kac completed his Ph.D. in mathematics at the Polish University of Lwów in 1937 under the direction of Hugo Steinhaus. While there, he was a member of the Lwów School of Mathematics. After receiving his degree, he began to look for a position abroad, and in 1938 was granted a scholarship from the Parnas Foundation, which enabled him to go work in the United States. He arrived in New York City in November 1938.
With the onset of World War II in Europe, Kac was able to stay in America, while his parents and brother, who had remained in Kremenets, were murdered by the Germans in mass executions in August 1942.
From 1939 to 1961, Kac taught at Cornell University, first as an instructor, then from 1943 as an assistant professor and from 1947 as a full professor. While there, he became a naturalized US citizen in 1943. From 1943 to 1945, he also worked in the MIT Radiation Laboratory, together with George Uhlenbeck. During the 1951–1952 academic year, Kac was on sabbatical at the Institute for Advanced Study. In 1952, Kac, with Theodore H. Berlin, introduced the spherical model of a ferromagnet (a variant of the Ising model) and, with J. C. Ward, found an exact solution of the Ising model using a combinatorial method. In 1961, Kac left Cornell and went to The Rockefeller University in New York City. In the early 1960s, he worked with George Uhlenbeck and P. C. Hemmer on the mathematic |
https://en.wikipedia.org/wiki/Genome%20project | Genome projects are scientific endeavours that ultimately aim to determine the complete genome sequence of an organism (be it an animal, a plant, a fungus, a bacterium, an archaean, a protist or a virus) and to annotate protein-coding genes and other important genome-encoded features. The genome sequence of an organism includes the collective DNA sequences of each chromosome in the organism. For a bacterium containing a single chromosome, a genome project will aim to map the sequence of that chromosome. For the human species, whose genome includes 22 pairs of autosomes and 2 sex chromosomes, a complete genome sequence will involve 46 separate chromosome sequences.
The Human Genome Project is a well known example of a genome project.
Genome assembly
Genome assembly refers to the process of taking a large number of short DNA sequences and reassembling them to create a representation of the original chromosomes from which the DNA originated. In a shotgun sequencing project, all the DNA from a source (usually a single organism, anything from a bacterium to a mammal) is first fractured into millions of small pieces. These pieces are then "read" by automated sequencing machines. A genome assembly algorithm works by taking all the pieces and aligning them to one another, and detecting all places where two of the short sequences, or reads, overlap. These overlapping reads can be merged, and the process continues.
Genome assembly is a very difficult computational problem, made more difficult because many genomes contain large numbers of identical sequences, known as repeats. These repeats can be thousands of nucleotides long, and occur different locations, especially in the large genomes of plants and animals.
The resulting (draft) genome sequence is produced by combining the information sequenced contigs and then employing linking information to create scaffolds. Scaffolds are positioned along the physical map of the chromosomes creating a "golden path".
Assembly soft |
https://en.wikipedia.org/wiki/Radial%20basis%20function%20network | In the field of mathematical modeling, a radial basis function network is an artificial neural network that uses radial basis functions as activation functions. The output of the network is a linear combination of radial basis functions of the inputs and neuron parameters. Radial basis function networks have many uses, including function approximation, time series prediction, classification, and system control. They were first formulated in a 1988 paper by Broomhead and Lowe, both researchers at the Royal Signals and Radar Establishment.
Network architecture
Radial basis function (RBF) networks typically have three layers: an input layer, a hidden layer with a non-linear RBF activation function and a linear output layer. The input can be modeled as a vector of real numbers . The output of the network is then a scalar function of the input vector, , and is given by
where is the number of neurons in the hidden layer, is the center vector for neuron , and is the weight of neuron in the linear output neuron. Functions that depend only on the distance from a center vector are radially symmetric about that vector, hence the name radial basis function. In the basic form, all inputs are connected to each hidden neuron. The norm is typically taken to be the Euclidean distance (although the Mahalanobis distance appears to perform better with pattern recognition) and the radial basis function is commonly taken to be Gaussian
.
The Gaussian basis functions are local to the center vector in the sense that
i.e. changing parameters of one neuron has only a small effect for input values that are far away from the center of that neuron.
Given certain mild conditions on the shape of the activation function, RBF networks are universal approximators on a compact subset of . This means that an RBF network with enough hidden neurons can approximate any continuous function on a closed, bounded set with arbitrary precision.
The parameters , , and are determined in a manner tha |
https://en.wikipedia.org/wiki/Virtual%20black%20hole | In quantum gravity, a virtual black hole is a hypothetical micro black hole that exists temporarily as a result of a quantum fluctuation of spacetime. It is an example of quantum foam and is the gravitational analog of the virtual electron–positron pairs found in quantum electrodynamics. Theoretical arguments suggest that virtual black holes should have mass on the order of the Planck mass, lifetime around the Planck time, and occur with a number density of approximately one per Planck volume.
The emergence of virtual black holes at the Planck scale is a consequence of the uncertainty relation
where is the radius of curvature of spacetime small domain, is the coordinate of the small domain, is the Planck length, is the reduced Planck constant, is the Newtonian constant of gravitation, and is the speed of light. These uncertainty relations are another form of Heisenberg's uncertainty principle at the Planck scale.
If virtual black holes exist, they provide a mechanism for proton decay. This is because when a black hole's mass increases via mass falling into the hole, and is theorized to decrease when Hawking radiation is emitted from the hole, the elementary particles emitted are, in general, not the same as those that fell in. Therefore, if two of a proton's constituent quarks fall into a virtual black hole, it is possible for an antiquark and a lepton to emerge, thus violating conservation of baryon number.
The existence of virtual black holes aggravates the black hole information loss paradox, as any physical process may potentially be disrupted by interaction with a virtual black hole.
See also
Quantum foam
Virtual particle
Quantum tunnelling |
https://en.wikipedia.org/wiki/Lambda-mu%20calculus | In mathematical logic and computer science, the lambda-mu calculus is an extension of the lambda calculus introduced by Michel Parigot. It introduces two new operators: the μ operator (which is completely different both from the μ operator found in computability theory and from the μ operator of modal μ-calculus) and the bracket operator. Proof-theoretically, it provides a well-behaved formulation of classical natural deduction.
One of the main goals of this extended calculus is to be able to describe expressions corresponding to theorems in classical logic. According to the Curry–Howard isomorphism, lambda calculus on its own can express theorems in intuitionistic logic only, and several classical logical theorems can't be written at all. However with these new operators one is able to write terms that have the type of, for example, Peirce's law.
Semantically these operators correspond to continuations, found in some functional programming languages.
Formal definition
We can augment the definition of a lambda expression to gain one in the context of lambda-mu calculus. The three main expressions found in lambda calculus are as follows:
, a variable, where V is any identifier.
, an abstraction, where V is any identifier and E is any lambda expression.
, an application, where E and E'''; are any lambda expressions.
For details, see the corresponding article.
In addition to the traditional λ-variables, the lambda-mu calculus includes a distinct set of μ-variables. These μ-variables can be used to name or freeze arbitrary subterms, allowing us to later abstract on those names. The set of terms contains unnamed (all traditional lambda expressions are of this kind) and named terms. The terms that are added by the lambda-mu calculus are of the form:
is a named term, where α is a μ-variable and t is an unnamed term.
is an unnamed term, where α is a μ-variable and E'' is a named term.
Reduction
The basic reduction rules used in the lambda-mu calculus are the follow |
https://en.wikipedia.org/wiki/Balding%E2%80%93Nichols%20model | In population genetics, the Balding–Nichols model is a statistical description of the allele frequencies in the components of a sub-divided population. With background allele frequency p the allele frequencies, in sub-populations separated by Wright's FST F, are distributed according to independent draws from
where B is the Beta distribution. This distribution has mean p and variance Fp(1 – p).
The model is due to David Balding and Richard Nichols and is widely used in the forensic analysis of DNA profiles and in population models for genetic epidemiology. |
https://en.wikipedia.org/wiki/Body%20load | Body load is the specific physical or tactile sensations brought on by psychoactive drugs, especially psychedelics. Generally, body load is an unpleasant physical sensation that is difficult to describe objectively either in terms of other sensations or in its specific location. However, it could be likened to an instinct of the body sensing it is about to be placed under exceptional stress, a state of pre-shock. Common symptoms include stomach ache, nausea, dizziness, feelings of being over-stimulated or "wired," shivering, feelings of excessive tension in the torso, or, in more severe cases, shortness of breath or a feeling of suffocation. Different drugs may cause different body load sensations which vary in intensity and duration.
In contrast, many drug users, and particularly users of cannabis, entactogens like MDMA or of certain synthetic phenethylamines (most notably the popular 2C-B) and tryptamines, also often report a "body high" or "body rush", which is similar to body load in many respects but is usually considered pleasant.
Causes
The causes of the experience of body load are unknown. However, one proposed mechanism is the stimulation of serotonergic 5-HT receptors, particularly those involved in tactile sensation and, equally importantly in many cases where nausea is experienced, those located along the lining of the digestive tract. Serotonin is heavily involved in appetite control, and over-stimulation of serotonergic receptors has been shown to cause nausea in overdoses of SSRIs or MDMA. Many psychedelics which can cause body load are partial serotonin agonists, which work by mimicking the structure of serotonin to varying degrees. |
https://en.wikipedia.org/wiki/Sounding%20board | A sounding board, also known as a tester and abat-voix is a structure placed above and sometimes also behind a pulpit or other speaking platform that helps to project the sound of the speaker. It is usually made of wood. The structure may be specially shaped to assist the projection, for example, being formed as a parabolic reflector. In the typical setting of a church building, the sounding board may be ornately carved or constructed. The term "abat-voix," from the French word for the same thing (abattre (“to beat down”) + voix (“voice”)), is also used in English.
The term “sounding board” is also used figuratively to describe a person who listens to a speech or proposal in order that the speaker may rehearse or explore the proposition more fully.
The term is also used inter-personally to describe one person listening to another, and especially to their ideas. When a person listens and responds with comments, they provide a perspective that otherwise would not be available through introspection or thought alone, and can potentially lead to Creative Hijack.
See also
Baldachin - canopy over altar or throne
Chhatri |
https://en.wikipedia.org/wiki/Sex-limited%20genes | Sex-limited genes are genes that are present in both sexes of sexually reproducing species but are expressed in only one sex and have no penetrance, or are simply 'turned off' in the other. In other words, sex-limited genes cause the two sexes to show different traits or phenotypes, despite having the same genotype. This term is restricted to autosomal traits, and should not be confused with sex-linked characteristics, which have to do with genetic differences on the sex chromosomes (see sex-determination system). Sex-limited genes are also distinguished from sex-influenced genes, where the same gene will show differential expression in each sex. Sex-influenced genes commonly show a dominant/recessive relationship, where the same gene will have a dominant effect in one sex and a recessive effect in the other (for example, male pattern baldness). However, the resulting phenotypes caused by sex-limited genes are present in only one sex and can be seen prominently in various species that typically show high sexual dimorphism.
Sex-limited genes are responsible for sexual dimorphism, which is a phenotypic (directly observable) difference between males and females of the same species regardless of genotype. These differences can be reflected in size, color, behavior (ex: levels of aggression), and morphology. An example of sex-limited genes are genes which control horn development in sheep: while both males and females possess the same genes controlling horn development, they are only expressed in males. Sex-limited genes are also responsible for some female beetles' inability to grow exaggerated mandibles, research that is discussed in detail later in this article.
Sex-limited genes were first hypothesized by Charles Darwin and though he was unsuccessful in distinguishing the previously mentioned sex-linked traits, his hypothesis was the starting point for future study of the subject. His studies on sex-limited traits have been further substantiated and supported over |
https://en.wikipedia.org/wiki/Random%20walk%20closeness%20centrality | Random walk closeness centrality is a measure of centrality in a network, which describes the average speed with which randomly walking processes reach a node from other nodes of the network. It is similar to the closeness centrality except that the farness is measured by the expected length of a random walk rather than by the shortest path.
The concept was first proposed by White and Smyth (2003) under the name Markov centrality.
Intuition
Consider a network with a finite number of nodes and a random walk process that starts in a certain node and proceeds from node to node along the edges. From each node, it chooses randomly the edge to be followed. In an unweighted network, the probability of choosing a certain edge is equal across all available edges, while in a weighted network it is proportional to the edge weights.
A node is considered to be close to other nodes, if the random walk process initiated from any node of the network arrives to this particular node in relatively few steps on average.
Definition
Consider a weighted network – either directed or undirected – with n nodes denoted by j=1, …, n; and a random walk process on this network with a transition matrix M. The element of M describes the probability of the random walker that has reached node i, proceeds directly to node j. These probabilities are defined in the following way.
where is the (i,j)th element of the weighting matrix A of the network. When there is no edge between two nodes, the corresponding element of the A matrix is zero.
The random walk closeness centrality of a node i is the inverse of the average mean first passage time to that node:
where is the mean first passage time from node j to node i.
Mean first passage time
The mean first passage time from node i to node j is the expected number of steps it takes for the process to reach node j from node i for the first time:
where P(i,j,r) denotes the probability that it takes exactly r steps to reach j from i for the fir |
https://en.wikipedia.org/wiki/Love%20lock | A love lock or love padlock is a padlock that couples lock to a bridge, fence, gate, monument, or similar public fixture to symbolize their love. Typically the sweethearts' names or initials, and perhaps the date, are inscribed on the padlock, and its key is thrown away (often into a nearby river) to symbolize unbreakable love.
Since the 2000s, love locks have proliferated at an increasing number of locations worldwide. They are treated by some municipal authorities as litter or vandalism, and there is some cost to their removal. However, there are other authorities who embrace them, and who use them as fundraising projects or tourist attractions.
History
In 2014, the New York Times reported that the history of love padlocks dates back at least 100 years to a melancholic Serbian tale of World War I, with an attribution for the bridge Most Ljubavi (lit. the Bridge of Love) in the spa town of Vrnjačka Banja. A local schoolmistress named Nada fell in love with a Serbian officer named Relja. After they committed to each other, Relja went to war in Greece, where he fell in love with a local woman from Corfu. As a consequence, Relja and Nada broke off their engagement. Nada never recovered from that devastating blow, and after some time she died due to heartbreak from her unfortunate love.
As young women from Vrnjačka Banja wanted to protect their own loves, they started writing down their names, with the names of their loved ones, on padlocks and affixing them to the railings of the bridge where Nada and Relja used to meet.
In the rest of Europe, love padlocks started appearing in the early 2000s as a ritual. The reasons love padlocks started to appear vary between locations and in many instances are unclear. However, in Rome, the ritual of affixing love padlocks to the bridge Ponte Milvio can be attributed to the 2006 book I Want You by Italian author Federico Moccia, who made a film adaptation in 2007.
Manufacturers
The German lock company ABUS manufacturers a |
https://en.wikipedia.org/wiki/Wild%20branch | In computer programming, a wild branch is a GOTO instruction where the target address is indeterminate, random or otherwise unintended. It is usually the result of a software bug causing the accidental corruption of a pointer or array index. It is "wild" in the sense that it cannot be predicted to behave consistently. In other words, a wild branch is a function pointer that is wild (dangling).
Detection of wild branches is frequently difficult; they are normally identified by erroneous results (where the unintended target address is nevertheless a valid instruction enabling the program to continue despite the error) or a hardware interrupt, which may change depending upon register contents. Debuggers and monitor programs such as Instruction set simulators can sometimes be used to determine the location of the original wild branch.
See also
Dangling pointer |
https://en.wikipedia.org/wiki/FLUIDEX | FLUIDEX is an online bibliographic database that covers fluids behavior, applications, and engineering in civil engineering and process engineering. It is published by Elsevier.
Indexing includes more than 900,000 past and current records pertaining to relevant trade magazines and scientific journals from 1966–present. Further topical coverage includes research and technology in fluid dynamics, separation processes, offshore platform issues, and new hydraulic and pneumatic equipment applications and operations. Abstracts summarize 98% of all documents.
Global literature coverage includes over 400 trade and peer-reviewed publications. Additionally, archives contain several hundred cataloged trade, peer-reviewed, and book titles. The database is annually updated with more than 40,000 abstracts and citations. OVID and Dialog are online access points. This database is also published in a printed paper format. Hence, the print counterparts are "Fluid Abstracts: Process Engineering" (), and "Fluid Abstracts: Civil engineering" (). |
https://en.wikipedia.org/wiki/International%20Conference%20on%20Digital%20Audio%20Effects | The annual International Conference on Digital Audio Effects or DAFx Conference is a meeting of enthusiasts working in research areas on audio signal processing, acoustics, and music related disciplines, who come together to present and discuss their findings. The conference evolved from an EU-COST-G6 project “Digital Audio Effects” in 1998.
The acronym DAFx stands for Digital Audio Effects and is also the name of a book
which was written by people in the community around the conference
A list of past and upcoming conferences together with an archive of all proceedings can be found at the website.
Scheduled and past conferences
DAFX, 2020 - Vienna, Austria
DAFX, 2019 - Birmingham, UK
DAFX, 2018 - Aveiro, Portugal
DAFX, 2017 - Edinburgh, UK
DAFX, 2016 - Brno, Czech Republic
DAFX, 2015 - Trondheim, Norway
DAFX, 2014 - Erlangen, Germany
DAFX, 2013 - Maynooth, Ireland
DAFX, 2012 - York, UK
DAFX, 2011 - Paris, France
DAFX, 2010 - Graz, Austria
DAFX, 2009 - Como, Italy
DAFX, 2008 - Espoo, Finland
DAFX, 2007 - Bordeaux, France
DAFX, 2006 - Montreal, Quebec, Canada
DAFX, 2005 - Madrid, Spain
DAFX, 2004 - Naples, Italy
DAFX, 2003 - London, UK
DAFX, 2002 - Hamburg, Germany
DAFX, 2001 - Limerick, Ireland
DAFX, 2000 - Verona, Italy
DAFX, 1999 - Trondheim, Norway
DAFX, 1998 - Barcelona, Spain
See also
International Society for Music Information Retrieval
Sound and Music Computing Conference |
https://en.wikipedia.org/wiki/Micropropagation | Micropropagation or tissue culture is the practice of rapidly multiplying plant stock material to produce many progeny plants, using modern plant tissue culture methods.
Micropropagation is used to multiply a wide variety of plants, such as those that have been genetically modified or bred through conventional plant breeding methods. It is also used to provide a sufficient number of plantlets for planting from seedless plants, plants that do not respond well to vegetative reproduction or where micropropagation is the cheaper means of propagating (e.g. Orchids). Cornell University botanist Frederick Campion Steward discovered and pioneered micropropagation and plant tissue culture in the late 1950s and early 1960s.
Steps
In short, steps of micropropagation can be divided into four stages:
Selection of mother plant
Multiplication
Rooting and acclimatizing
Transfer new plant to soil
Selection of mother plant
Micropropagation begins with the selection of plant material to be propagated. The plant tissues are removed from an intact plant in a sterile condition. Clean stock materials that are free of viruses and fungi are important in the production of the healthiest plants. Once the plant material is chosen for culture, the collection of explant(s) begins and is dependent on the type of tissue to be used; including stem tips, anthers, petals, pollen and other plant tissues. The explant material is then surface sterilized, usually in multiple courses of bleach and alcohol washes, and finally rinsed in sterilized water. This small portion of plant tissue, sometimes only a single cell, is placed on a growth medium, typically containing Macro and micro nutrients, water, sucrose as an energy source and one or more plant growth regulators (plant hormones). Usually the medium is thickened with a gelling agent, such as agar, to create a gel which supports the explant during growth. Some plants are easily grown on simple media, but others require more complicated media f |
https://en.wikipedia.org/wiki/Model%20robot | Model robots are model figures with origins in the Japanese anime genre of mecha. The majority of model robots are produced by Bandai and are based on the Gundam anime metaseries. This has given rise to the hobby's common name in Japan, Gunpla (or gan-pura, a Japanese portmanteau of "Gundam" and "plastic model"). Though there are exceptions, the model robot genre is dominated by anime tie-ins, with anime series and movies frequently serving as merchandising platform.
Construction
Gundam kits are the most common and popular variety of mecha models exemplifying the general characteristics of models in the genre. Gundam kits are typically oriented toward beginners, and most often feature simple assembly, simple designs, and rugged construction—less durable than a pre-assembled toy, but more durable than a true scale model. The result is that the majority of Gundam kits feature hands and other parts that favor poseability or easy assembly over accurate shape. They may also exhibit various draft-angle problems, and features like antennae that are oversized to prevent breakage. For the most part, other kit lines and other kit manufacturers in the genre follow suit, though there are exceptions.
Because the subjects of model robot kits are typically humanoid and/or possess limbs, joints are required in order to make the finished model poseable. For decades, poly-caps were and still are used for this purpose, although they tend to degrade over time and thus have been less frequently used since the 2010s. Hard plastic joints generally exhibit greater friction than polyvinyl joints, and are similarly more durable than polystyrene joints. ABS joints, however, require greater precision in tooling to ensure easy assembly, and in some cases, they require screws and a small gap between parts.
One distinctive feature of model robot kits since the 1990s, as opposed to most other plastic model kits, is that they are molded in color: each part generally is made of a colored plastic |
https://en.wikipedia.org/wiki/Comparison%20of%20chess%20video%20games | This is a comparison of chess video games.
General information
Gameplay
See also
List of chess software
Notes |
https://en.wikipedia.org/wiki/Service%20contour | In US broadcasting, service contour (or protected contour) refers to the area in which the Federal Communications Commission (FCC) predicts coverage.
The FCC calculates FM and TV contours based on effective radiated power (ERP) in a given direction, the radial height above average terrain (HAAT) in a given direction, the FCC's propagation curves, and the station's class. AM contours are based on the standard ground wave field strength pattern, the frequency, and the ground conductivity in the area. While the FCC makes FM and TV service contour data readily available, the AM, while unavailable as a separate data file, can be obtained through an AM Query in the resulting 'maps' section of each record (when using the 'detailed output' output option).
External links
FM and TV Service Contour Data (Official)
FM Database protected contour maps (Official)
AM Station Query (Official)
FM Station Query (Official)
TV Station Query (Official)
Plot predicted AM/FM coverage patterns (unofficial)
Broadcast engineering |
https://en.wikipedia.org/wiki/Orrery | An orrery is a mechanical model of the Solar System that illustrates or predicts the relative positions and motions of the planets and moons, usually according to the heliocentric model. It may also represent the relative sizes of these bodies; however, since accurate scaling is often not practical due to the actual large ratio differences, a subdued approximation may be used instead. Though the Greeks had working planetaria, the first orrery that was a planetarium of the modern era was produced in 1713, and one was presented to Charles Boyle, 4th Earl of Orrery – hence the name. They are typically driven by a clockwork mechanism with a globe representing the Sun at the centre, and with a planet at the end of each of the arms.
History
Ancient versions
The Antikythera mechanism, discovered in 1901 in a wreck off the Greek island of Antikythera in the Mediterranean Sea, exhibited the diurnal motions of the Sun, Moon, and the five planets known to the ancient Greeks. It has been dated between 205 to 87 BC. The mechanism is considered one of the first orreries. It was geocentric and used as a mechanical calculator to calculate astronomical positions.
Cicero, the Roman philosopher and politician writing in the first century BC, has references describing planetary mechanical models. According to him, the Greek polymaths Thales and Posidonius both constructed a device modeling celestial motion.
Early versions
In 1348, Giovanni Dondi built the first known clock driven mechanism of the system. It displays the ecliptic position of the Moon, Sun, Mercury, Venus, Mars, Jupiter and Saturn according to the complicated geocentric Ptolemaic planetary theories. The clock itself is lost, but Dondi left a complete description of its astronomic gear trains.
As late as 1650, P. Schirleus built a geocentric planetarium with the Sun as a planet, and with Mercury and Venus revolving around the Sun as its moons.
At the court of William IV, Landgrave of Hesse-Kassel two complicated a |
https://en.wikipedia.org/wiki/Normal%20convergence | In mathematics normal convergence is a type of convergence for series of functions. Like absolute-convergence, it has the useful property that it is preserved when the order of summation is changed.
History
The concept of normal convergence was first introduced by René Baire in 1908 in his book Leçons sur les théories générales de l'analyse.
Definition
Given a set S and functions (or to any normed vector space), the series
is called normally convergent if the series of uniform norms of the terms of the series converges, i.e.,
Distinctions
Normal convergence implies, but should not be confused with, uniform absolute convergence, i.e. uniform convergence of the series of nonnegative functions . To illustrate this, consider
Then the series is uniformly convergent (for any ε take n ≥ 1/ε), but the series of uniform norms is the harmonic series and thus diverges. An example using continuous functions can be made by replacing these functions with bump functions of height 1/n and width 1 centered at each natural number n.
As well, normal convergence of a series is different from norm-topology convergence, i.e. convergence of the partial sum sequence in the topology induced by the uniform norm. Normal convergence implies norm-topology convergence if and only if the space of functions under consideration is complete with respect to the uniform norm. (The converse does not hold even for complete function spaces: for example, consider the harmonic series as a sequence of constant functions).
Generalizations
Local normal convergence
A series can be called "locally normally convergent on X" if each point x in X has a neighborhood U such that the series of functions ƒn restricted to the domain U
is normally convergent, i.e. such that
where the norm is the supremum over the domain U.
Compact normal convergence
A series is said to be "normally convergent on compact subsets of X" or "compactly normally convergent on X" if for every compact subset K of X |
https://en.wikipedia.org/wiki/Alzheimer%27s%20Disease%20Neuroimaging%20Initiative | Alzheimer's Disease Neuroimaging Initiative (ADNI) is a multisite study that aims to improve clinical trials for the prevention and treatment of Alzheimer's disease (AD). This cooperative study combines expertise and funding from the private and public sector to study subjects with AD, as well as those who may develop AD and controls with no signs of cognitive impairment. Researchers at 63 sites in the US and Canada track the progression of AD in the human brain with neuroimaging, biochemical, and genetic biological markers. This knowledge helps to find better clinical trials for the prevention and treatment of AD. ADNI has made a global impact, firstly by developing a set of standardized protocols to allow the comparison of results from multiple centers, and secondly by its data-sharing policy which makes available all at the data without embargo to qualified researchers worldwide. To date, over 1000 scientific publications have used ADNI data. A number of other initiatives related to AD and other diseases have been designed and implemented using ADNI as a model. ADNI has been running since 2004 and is currently funded until 2021.
Primary goals
Detect the earliest signs of AD and to track the disease using biomarkers.
validate, standardize, and optimize biomarkers for clinical AD trials.
to make all data and samples available for sharing with clinical trial designers and scientists worldwide.
History and funding
The idea of a collaboration between public institutions and private pharmaceutical companies to fund a large biomarker project to study AD and to speed up progress toward effective treatments for the disease was conceived at the beginning of the millennium by Neil S. Buckholz at the National Institute on Aging (NIA) and Dr. William Potter, at Eli Lilly and Company. The Alzheimer's Disease Neuroimaging Initiative (ADNI) began in 2004 under the leadership of Dr. Michael W. Weiner, funded as a private – public partnership with $27 million contributed b |
https://en.wikipedia.org/wiki/Scripps%20Genomic%20Health%20Initiative | The Scripps Genomic Health Initiative (SGHI) is a large scale study aimed at understanding how personal genetic testing influences and improves health.
Led by Dr. Eric Topol, director of the San Diego-based Scripps Translational Science Institute, the 20-year initiative will determine whether patients make an effort to improve their lifestyle and get regular checkups after learning their genetic predisposition for many common diseases. Researchers will also assess the psychological impact of genomic testing, and whether those who do it are able to prevent or delay disease by taking action after getting their results.
The study was launched in October 2008 and will follow more than 10,000 adults. A consortium of health care, technology and research leaders have joined forces in the first-of-its-kind research study, including genetic test provider Navigenics Inc., Affymetrix and Microsoft Corp.
Study participants receive a scan of their genome and a detailed analysis of their genetic risk for more than 20 health conditions that may be changed by lifestyle, including type 2 diabetes, Alzheimer's disease, heart attack, obesity, and several types of cancer.
Said Peter Neupert, corporate VP for the Health Solutions Group at Microsoft:
"Personalized medicine stands to change the way people approach their health and wellness, as well as open up new genetic research opportunities."
See also
Full Genome Sequencing
Genetic testing
Personal genomics
Navigenics
Affymetrix
Microsoft
Eric Topol |
https://en.wikipedia.org/wiki/General%20Motors%20Local%20Area%20Network | General Motors Local Area Network (GMLAN) is an application- and transport-layer protocol using controller area network for lower layer services. It was standardized as SAE J2411 for use in OBD-II vehicle networks.
Transport-layer services
Transport-layer services include the transmission of multi-CAN-frame messages based on the ISO 15765-2 multi-frame messaging scheme. It was developed and is used primarily by General Motors for in-vehicle communication and diagnostics. GM's Tech2 uses the CANdi (Controller Area Network diagnostic interface) adapter to communicate over GMLAN.
Applications
Some software applications that allow interfacing to GMLAN are Intrepid Control Systems, Inc.'s Vehicle Spy 3; Vector's CANoe; Dearborn Group's Hercules, ETAS' ES-1222, ES590, ES715, and ES580; ScanTool.net's OBDLink MX; EControls by Enovation Controls' CANCapture; and GMLAN vehicle universal remote control GMRC for Android devices
Tesla uses J2411 (single-wire CAN over the Control Pilot) for their DC Supercharger (newer units are also capable of PLC over the control pilot) and AC Destination Charging. |
https://en.wikipedia.org/wiki/Antibody%20testing | Antibody testing may refer to:
Serological testing, tests that detect specific antibodies in the blood
Immunoassay, tests that use antibodies to detect substances
Antibody titer, tests that measure the amount of a specific antibody in a sample
Antibodies
Biological techniques and tools |
https://en.wikipedia.org/wiki/Moli%C3%A8re%20radius | The Molière radius is a characteristic constant of a material giving the scale of the transverse dimension of the fully contained electromagnetic showers initiated by an incident high energy electron or photon. By definition, it is the radius of a cylinder containing on average 90% of the shower's energy deposition. Two Molière radii contain 95% of the shower's energy deposition. It is related to the radiation length by the approximate relation , where is the atomic number. The Molière radius is useful in experimental particle physics in the design of calorimeters: a smaller Molière radius means better shower position resolution, and better shower separation due to a smaller degree of shower overlaps.
The Molière radius is named after German physicist Paul Friederich Gaspard Gert Molière (1909–64).
Molière radii for typical materials used in calorimetry
LYSO crystals: 2.07 cm
Lead tungstate crystals: 2.2 cm
Caesium iodide: 3.5 cm
Liquid krypton: 4.7 cm
Liquid argon: 9.04 cm
Earth's atmosphere at sea level: 79 m
Earth's atmosphere above ground: 91 m |
https://en.wikipedia.org/wiki/Diving%20reflex | The diving reflex, also known as the diving response and mammalian diving reflex, is a set of physiological responses to immersion that overrides the basic homeostatic reflexes, and is found in all air-breathing vertebrates studied to date. It optimizes respiration by preferentially distributing oxygen stores to the heart and brain, enabling submersion for an extended time.
The diving reflex is exhibited strongly in aquatic mammals, such as seals, otters, dolphins, and muskrats, and exists as a lesser response in other animals, including human babies up to 6 months old (see infant swimming), and diving birds, such as ducks and penguins. Adult humans generally exhibit a mild response, the dive-hunting Sama-Bajau people being a notable outlier.
The diving reflex is triggered specifically by chilling and wetting the nostrils and face while breath-holding, and is sustained via neural processing originating in the carotid chemoreceptors. The most noticeable effects are on the cardiovascular system, which displays peripheral vasoconstriction, slowed heart rate, redirection of blood to the vital organs to conserve oxygen, release of red blood cells stored in the spleen, and, in humans, heart rhythm irregularities. Although aquatic animals have evolved profound physiological adaptations to conserve oxygen during submersion, the apnea and its duration, bradycardia, vasoconstriction, and redistribution of cardiac output occur also in terrestrial animals as a neural response, but the effects are more profound in natural divers.
Physiological response
When the face is submerged and water fills the nostrils, sensory receptors sensitive to wetness within the nasal cavity and other areas of the face supplied by the fifth (V) cranial nerve (the trigeminal nerve) relay the information to the brain. The tenth (X) cranial nerve, (the vagus nerve) – part of the autonomic nervous system – then produces bradycardia and other neural pathways elicit peripheral vasoconstriction, restri |
https://en.wikipedia.org/wiki/K-Poincar%C3%A9%20group | In physics and mathematics, the κ-Poincaré group, named after Henri Poincaré, is a quantum group, obtained by deformation of the Poincaré group into a Hopf algebra.
It is generated by the elements and with the usual constraint:
where is the Minkowskian metric:
The commutation rules reads:
In the (1 + 1)-dimensional case the commutation rules between and are particularly simple. The Lorentz generator in this case is:
and the commutation rules reads:
The coproducts are classical, and encode the group composition law:
Also the antipodes and the counits are classical, and represent the group inversion law and the map to the identity:
The κ-Poincaré group is the dual Hopf algebra to the K-Poincaré algebra, and can be interpreted as its “finite” version. |
https://en.wikipedia.org/wiki/P-adic%20Hodge%20theory | In mathematics, p-adic Hodge theory is a theory that provides a way to classify and study p-adic Galois representations of characteristic 0 local fields with residual characteristic p (such as Qp). The theory has its beginnings in Jean-Pierre Serre and John Tate's study of Tate modules of abelian varieties and the notion of Hodge–Tate representation. Hodge–Tate representations are related to certain decompositions of p-adic cohomology theories analogous to the Hodge decomposition, hence the name p-adic Hodge theory. Further developments were inspired by properties of p-adic Galois representations arising from the étale cohomology of varieties. Jean-Marc Fontaine introduced many of the basic concepts of the field.
General classification of p-adic representations
Let K be a local field with residue field k of characteristic p. In this article, a p-adic representation of K (or of GK, the absolute Galois group of K) will be a continuous representation ρ : GK→ GL(V), where V is a finite-dimensional vector space over Qp. The collection of all p-adic representations of K form an abelian category denoted in this article. p-adic Hodge theory provides subcollections of p-adic representations based on how nice they are, and also provides faithful functors to categories of linear algebraic objects that are easier to study. The basic classification is as follows:
where each collection is a full subcategory properly contained in the next. In order, these are the categories of crystalline representations, semistable representations, de Rham representations, Hodge–Tate representations, and all p-adic representations. In addition, two other categories of representations can be introduced, the potentially crystalline representations Reppcris(K) and the potentially semistable representations Reppst(K). The latter strictly contains the former which in turn generally strictly contains Repcris(K); additionally, Reppst(K) generally strictly contains Repst(K), and is contained in RepdR( |
https://en.wikipedia.org/wiki/Linux%20on%20Apple%20devices | The Linux kernel can run on a variety of devices made by Apple, including devices where the unlocking of the bootloader is not possible with an official procedure, such as iPhones and iPads.
iPad
In June 2022, software developers Konrad Dybcio and Markuss Broks managed to run Linux kernel 5.18 on a iPad Air 2. The project made use of the Alpine Linux based Linux distribution called postmarketOS, which is primarily developed for Android devices. The developer suggested that they used the checkm8 exploit which was published back in 2019.
iPhone
In 2008, Linux kernel 2.6 was ported to the iPhone 3G, iPhone (1st generation), and iPod Touch (1st generation) using OpeniBoot.
Corellium's Project Sandcastle made it possible to run Android on an iPhone 7/7+ or an iPod Touch (7th generation) using the checkm8 exploit.
iPod
iPodLinux is a Linux distribution created specifically to run on Apple's iPod.
Mac
Motorola 68k Macs
Linux can be dual-booted on Macs that use Motorola 680x0 processors (only 68020 and higher, and only non-"EC" processor variants since an MMU is required). The Linux/mac68k community project provides resources to do so, and an m68k community port of the Debian Linux distribution is also available.
PowerPC Macs
PowerPC Macs can run Linux through both emulation and dual-booting ("bare metal"). The most popular PowerPC emulation tools for Mac OS/Mac OS X are Microsoft's Virtual PC, and the open-source QEMU.
Linux dual-booting is achieved by partitioning the boot drive, installing the Yaboot bootloader onto the Linux partition, and selecting that Linux partition as the Startup Disk. This results in users being prompted to select whether they want to boot into Mac OS or Linux when the machine starts.
By 2008, a number of major Linux distributions had official versions compatible with Mac PowerPC processors, including:
Gentoo
Debian (until Debian 8)
Ubuntu (until Ubuntu 16.10)
Fedora (until Fedora 17 for G3 and G4 processors, and Fedora 28 for |
https://en.wikipedia.org/wiki/Mind-wandering | Mind-wandering is loosely defined as thoughts that are not produced from the current task. Mind-wandering consists of thoughts that are task-unrelated and stimulus-independent. This can be in the form of three different subtypes: positive constructive daydreaming, guilty fear of failure, and poor attentional control.
In general, a folk explanation of mind-wandering could be described as the experience of thoughts not remaining on a single topic for a long period of time, particularly when people are engaged in an attention-demanding task.
One context in which mind-wandering often occurs is driving. This is because driving under optimal conditions becomes an almost automatic activity that can require minimal use of the task positive network, the brain network that is active when one is engaged in an attention-demanding activity. In situations where vigilance is low, people do not remember what happened in the surrounding environment because they are preoccupied with their thoughts. This is known as the decoupling hypothesis.
Studies using event-related potentials (ERPs) have quantified the extent that mind-wandering reduces the cortical processing of the external environment. When thoughts are unrelated to the task at hand, the brain processes both task-relevant and unrelated sensory information in a less detailed manner.
Mind-wandering appears to be a stable trait of people and a transient state. Studies have linked performance problems in the laboratory and in daily life. Mind-wandering has been associated with possible car accidents. Mind-wandering is also intimately linked to states of affect. Studies indicate that task-unrelated thoughts are common in people with low or depressed mood. Mind-wandering also occurs when a person is intoxicated via the consumption of alcohol.
Studies have demonstrated a prospective bias to spontaneous thought because individuals tend to engage in more future than past related thoughts during mind-wandering. The default mode |
https://en.wikipedia.org/wiki/Non-Hausdorff%20manifold | In geometry and topology, it is a usual axiom of a manifold to be a Hausdorff space. In general topology, this axiom is relaxed, and one studies non-Hausdorff manifolds: spaces locally homeomorphic to Euclidean space, but not necessarily Hausdorff.
Examples
Line with two origins
The most familiar non-Hausdorff manifold is the line with two origins, or bug-eyed line. This is the quotient space of two copies of the real line
and obtained by identifying points and whenever
An equivalent description of the space is to take the real line and replace the origin with two origins and The subspace retains its usual Euclidean topology. And a local base of open neighborhoods at each origin is formed by the sets with an open neighborhood of in
For each origin the subspace obtained from by replacing with is an open neighborhood of homeomorphic to Since every point has a neighborhood homeomorphic to the Euclidean line, the space is locally Euclidean. In particular, it is locally Hausdorff, in the sense that each point has a Hausdorff neighborhood. But the space is not Hausdorff, as every neighborhood of intersect every neighbourhood of It is however a T1 space.
The space is second countable.
The space exhibits several phenomena that don't happen in Hausdorff spaces:
The space is path connected but not arc connected. In particular, to get a path from one origin to the other one can first move left from to within the line through the first origin, and then move back to the right from to within the line through the second origin. But it is impossible to join the two origins with an arc, which is an injective path; intuitively, if one moves first to the left, one has to eventually backtrack and move back to the right.
The intersection of two compact sets need not be compact. For example, the sets and are compact, but their intersection is not.
The space is locally compact in the sense that every point has a local base of compact ne |
https://en.wikipedia.org/wiki/Geometric%20graph%20theory | Geometric graph theory in the broader sense is a large and amorphous subfield of graph theory, concerned with graphs defined by geometric means. In a stricter sense, geometric graph theory studies combinatorial and geometric properties of geometric graphs, meaning graphs drawn in the Euclidean plane with possibly intersecting straight-line edges, and topological graphs, where the edges are allowed to be arbitrary continuous curves connecting the vertices; thus, it can be described as "the theory of geometric and topological graphs" (Pach 2013). Geometric graphs are also known as spatial networks.
Different types of geometric graphs
A planar straight-line graph is a graph in which the vertices are embedded as points in the Euclidean plane, and the edges are embedded as non-crossing line segments. Fáry's theorem states that any planar graph may be represented as a planar straight line graph. A triangulation is a planar straight line graph to which no more edges may be added, so called because every face is necessarily a triangle; a special case of this is the Delaunay triangulation, a graph defined from a set of points in the plane by connecting two points with an edge whenever there exists a circle containing only those two points.
The 1-skeleton of a polyhedron or polytope is the set of vertices and edges of said polyhedron or polytope. The skeleton of any convex polyhedron is a planar graph, and the skeleton of any k-dimensional convex polytope is a k-connected graph. Conversely, Steinitz's theorem states that any 3-connected planar graph is the skeleton of a convex polyhedron; for this reason, this class of graphs is also known as the polyhedral graphs.
A Euclidean graph is a graph in which the vertices represent points in the plane, and each edge is assigned the length equal to the Euclidean distance between its endpoints. The Euclidean minimum spanning tree is the minimum spanning tree of a Euclidean complete graph. It is also possible to define graphs by c |
https://en.wikipedia.org/wiki/Blue%20Flag%20beach | The Blue Flag is a certification by the Foundation for Environmental Education (FEE) that a beach, marina, or sustainable boating tourism operator meets its standards.
The Blue Flag is a trademark owned by FEE, which is a not-for-profit non-governmental organisation consisting of 65 organisations in 77 member countries.
FEE's Blue Flag criteria include standards for quality, safety, environmental education and information, the provision of services and general environmental management criteria. The Blue Flag is sought for beaches, marinas, and sustainable boating tourism operators as an indication of their high environmental and quality standards.
Certificates, which FEE refers to as awards, are issued on an annual basis to beaches and marinas of FEE member countries. The awards are announced yearly on 5 June for Europe, Canada, Morocco, Tunisia, and other countries in a similar geographic location, and on 1 November for the Caribbean, New Zealand, South Africa, and other countries in the southern hemisphere.
In the European Union, the water quality standards are incorporated in the EC Water Framework Directive.
As of 2016 Spain has had more blue flag beaches than any other country every year since the awards began in 1987.
Blue Flags awarded
2015 Awards
As a result of the 2015 awards, a total of 4,154 Blue Flags are waving around the world.
Table of Blue Flags in force 2015
`The table below lists the Blue Flags (both for beaches and marinas) awarded and in force in 2015.
The table can be sorted to show the total number of Blue Flags per country and also the number of Blue Flags per population, per area or per the length of the coastline of each country.
History
The Blue Flag was created in France in 1985, as a pilot scheme from the Office of the Foundation for Environmental Education in Europe (Office français de la Fondation pour l'Education à l'Environnement en Europe) where French coastal municipalities were awarded the Blue Flag on the basis of criter |
https://en.wikipedia.org/wiki/Isostasy | Isostasy (Greek ísos 'equal', stásis 'standstill') or isostatic equilibrium is the state of gravitational equilibrium between Earth's crust (or lithosphere) and mantle such that the crust "floats" at an elevation that depends on its thickness and density. This concept is invoked to explain how different topographic heights can exist at Earth's surface. Although originally defined in terms of continental crust and mantle, it has subsequently been interpreted in terms of lithosphere and asthenosphere, particularly with respect to oceanic island volcanoes, such as the Hawaiian Islands.
Although Earth is a dynamic system that responds to loads in many different ways, isostasy describes the important limiting case in which crust and mantle are in static equilibrium. Certain areas (such as the Himalayas and other convergent margins) are not in isostatic equilibrium and are not well described by isostatic models.
The general term isostasy was coined in 1882 by the American geologist Clarence Dutton.
History of the concept
In the 17th and 18th centuries, French geodesists (for example, Jean Picard) attempted to determine the shape of the Earth (the geoid) by measuring the length of a degree of latitude at different latitudes (arc measurement). A party working in Ecuador was aware that its plumb lines, used to determine the vertical direction, would be deflected by the gravitational attraction of the nearby Andes Mountains. However, the deflection was less than expected, which was attributed to the mountains having low-density roots that compensated for the mass of the mountains. In other words, the low-density mountain roots provided the buoyancy to support the weight of the mountains above the surrounding terrain. Similar observations in the 19th century by British surveyors in India showed that this was a widespread phenomenon in mountainous areas. It was later found that the difference between the measured local gravitational field and what was expected for the altitu |
https://en.wikipedia.org/wiki/Graphitization | Graphitization is a process of transforming a carbonaceous material, such as coal, graphite or certain forms of iron alloys, into graphite.
Process
The graphitization process involves a restructuring of the molecular structure of the carbon material. In the initial state, these materials can have an amorphous structure or a crystalline structure different from graphite. Graphitization generally occurs at high temperatures (up to ), and can be accelerated by catalysts such as iron or nickel.
When carbonaceous material is exposed to high temperatures for an extended period of time, the carbon atoms begin to rearrange and form layered crystal planes. In the structure of graphite, carbon atoms are arranged in flat hexagonal sheets that are stacked on top of each other. These crystal planes give graphite its characteristic flake structure, giving it specific properties such as good electrical and thermal conductivity, low friction and excellent lubrication.
Interest
Graphitization can be observed in various contexts. For example, it occurs naturally during the formation of certain types of coal or graphite in the Earth's crust. It can also be artificially induced during the manufacture of specific carbon materials, such as graphite electrodes used in fuel cells, nuclear reactors or metallurgical applications.
Graphitization is of particular interest in the field of metallurgy. Some iron alloys, such as cast iron, can undergo graphitization heat treatment to improve their mechanical properties and machinability. During this process, the carbon dissolved in the iron alloy matrix separates and restructures as graphite, which gives the cast iron its specific characteristics, such as improved ductility and wear resistance.
Notes and references
Molecular physics
Metallurgy
Materials science
Materials |
https://en.wikipedia.org/wiki/Comparison%20of%20real-time%20operating%20systems | This is a list of real-time operating systems (RTOSs). This is an operating system in which the time taken to process an input stimulus is less than the time lapsed until the next input stimulus of the same type. |
https://en.wikipedia.org/wiki/Endianness | In computing, endianness is the order or sequence of bytes of a word of digital data in computer memory or data communication which is identified by describing the impact of the "first" bytes, meaning at the smallest address or sent first. Endianness is primarily expressed as big-endian (BE) or little-endian (LE). A big-endian system stores the most significant byte of a word at the smallest memory address and the least significant byte at the largest. A little-endian system, in contrast, stores the least-significant byte at the smallest address. Bi-endianness is a feature supported by numerous computer architectures that feature switchable endianness in data fetches and stores or for instruction fetches. Other orderings are generically called middle-endian or mixed-endian.
Endianness may also be used to describe the order in which the bits are transmitted over a communication channel, e.g., big-endian in a communications channel transmits the most significant bits first. Bit-endianness is seldom used in other contexts.
Etymology
Danny Cohen introduced the terms big-endian and little-endian into computer science for data ordering in an Internet Experiment Note published in 1980.
The adjective endian has its origin in the writings of 18th century Anglo-Irish writer Jonathan Swift. In the 1726 novel Gulliver's Travels, he portrays the conflict between sects of Lilliputians divided into those breaking the shell of a boiled egg from the big end or from the little end. As a boy, the grandfather of the emperor whom Gulliver met had cut his finger while opening an egg from the big end. The boy's father and emperor at the time published an imperial edict commanding all his subjects to break their eggs from the small end. The people resented the change, sparking six rebellions of "Big-Endians." Swift did not use the term Little-Endians in the work. Cohen makes the connection to Gulliver's Travels explicit in the appendix to his 1980 note.
The names and have someti |
https://en.wikipedia.org/wiki/Foreign%20body%20reaction | A foreign body reaction (FBR) is a typical tissue response to a foreign body within biological tissue. It usually includes the formation of a foreign body granuloma. Tissue-encapsulation of an implant is an example, as is inflammation around a splinter. Foreign body granuloma formation consists of protein adsorption, macrophages, multinucleated foreign body giant cells (macrophage fusion), fibroblasts, and angiogenesis. It has also been proposed that the mechanical property of the interface between an implant and its surrounding tissues is critical for the host response.
In the long term, the foreign body reaction results in encapsulation of the foreign body within a calcified shell. For example, a lithopedion is a rare phenomenon which occurs most commonly when a fetus dies during an abdominal pregnancy, is too large to be reabsorbed by the body, and calcifies.
Foreign body reaction to biomaterial implantation
Following biomaterial implantation, blood and body fluids contact the implant surface. Host blood proteins adsorb onto the implant surface and a fibrin matrix forms. Acute and chronic inflammation follow the initial blood protein deposition and matrix formation. Macrophages at the implant site fuse to form foreign body giant cells. Following the inflammatory response, granulation tissue form. The end stage of the foreign body reaction is the fibrous capsule formation around the implanted biomaterial. The biocompatibility of the device affects the severity of the foreign body reaction. The foreign body reaction can lead to device failure.
Protein adsorption
During blood-biomaterial interaction, blood proteins spontaneously adsorb to the biomaterial surface. The biomaterial surface properties affect the types, concentrations, and conformation of proteins that adsorb to the surface. The Vroman effect can describe the time-dependent behavior of this protein adsorption. Surface-adsorbed proteins regulate inflammatory cell interaction and adhesion. The deposit |
https://en.wikipedia.org/wiki/Stable%20roommates%20problem | In mathematics, economics and computer science, particularly in the fields of combinatorics, game theory and algorithms, the stable-roommate problem (SRP) is the problem of finding a stable matching for an even-sized set. A matching is a separation of the set into disjoint pairs ("roommates"). The matching is stable if there are no two elements which are not roommates and which both prefer each other to their roommate under the matching. This is distinct from the stable-marriage problem in that the stable-roommates problem allows matches between any two elements, not just between classes of "men" and "women".
It is commonly stated as:
In a given instance of the stable-roommates problem (SRP), each of 2n participants ranks the others in strict order of preference. A matching is a set of n disjoint pairs of participants. A matching M in an instance of SRP is stable if there are no two participants x and y, each of whom prefers the other to their partner in M. Such a pair is said to block M, or to be a blocking pair with respect to M.
Solution
Unlike the stable marriage problem, a stable matching may fail to exist for certain sets of participants and their preferences. For a minimal example of a stable pairing not existing, consider 4 people , , , and , whose rankings are:
A:(B,C,D), B:(C,A,D), C:(A,B,D), D:(A,B,C)
In this ranking, each of A, B, and C is the most preferable person for someone. In any solution, one of A, B, or C must be paired with D and the other two with each other (for example AD and BC), yet for anyone who is partnered with D, another member will have rated them highest, and D's partner will in turn prefer this other member over D. In this example, AC is a more favorable pairing than AD, but the necessary remaining pairing of BD then raises the same issue, illustrating the absence of a stable matching for these participants and their preferences.
Algorithm
An efficient algorithm was given in . The algorithm will determine, for any instance o |
https://en.wikipedia.org/wiki/Mutation%20Research%20%28journal%29 | Mutation Research is a peer-reviewed scientific journal that publishes research papers in the area of mutation research which focus on fundamental mechanisms underlying the phenotypic and genotypic expression of genetic damage. There are currently three sections:
Mutation Research/Fundamental and Molecular Mechanisms of Mutagenesis(,
Mutation Research/Genetic Toxicology and Environmental Mutagenesis(,
Mutation Research/Reviews in Mutation Research(,
Two previous sections
Mutation Research/DNA Repair Reports ()
Mutation Research/DNA Repair ()
are now continued as DNA Repair. |
https://en.wikipedia.org/wiki/Chinese%20Optical%20Society | Chinese Optical Society (; abbreviated COS) is a professional association of individuals with an interest in optics and photonics. It sponsored the Chinese Optics Letters, a monthly peer-reviewed scientific journal focusing on optics. As of 2019, the society has 21 specialized committees and 7 working committees with more than 15,000 individual members.
History
The Chinese Optical Society was established by Wang Daheng, Gong Zutong and Qian Linzhao on December 20, 1979. In 1987, it became a member of the International Commission for Optics (ICO).
Scientific publishing
Chinese Optics Letters
List of presidents |
https://en.wikipedia.org/wiki/OpenIO | OpenIO offered object storage for a wide range of high-performance applications. OpenIO was founded in 2015 by Laurent Denel (CEO), Jean-François Smigielski (CTO) and five other co-founders; it leveraged open source software, developed since 2006, based on a grid technology that enabled dynamic behaviour and supported heterogenous hardware. In October 2017 OpenIO completed a $5 million funding round. In July 2020 OpenIO had been acquired by OVH and withdrawn from the market to become the core technology of OVHcloud object storage offering.
Software
OpenIO is a software-defined object store that supports S3 and can be deployed on-premises, cloud-hosted or at the edge, on any hardware mix. It has been designed from the beginning for performance and cost-efficiency at any scale, and it has been optimized for Big Data, HPC and AI.
OpenIO stores objects within a flat structure within a massively distributed directory with indirections, which allows the data query path to be independent of the number of nodes and the performance not to be affected by the growth of capacity. Servers are organized as a grid of nodes massively distributed, where each node takes part in directory and storage services, which ensures that there is no single point of failure and that new nodes are automatically discovered and immediately available without the need to rebalance data.
The software is built on top of a technology that ensures optimal data placement based on real-time metrics and allows the addition or removal of storage devices with automatic performance and load impact optimization. For data protection OpenIO has synchronous and asynchronous replication with multiple copies, and an erasure coding implementation based on Reed-Solomon that can be deployed in one data center or geo-distributed or stretched clusters.
The software has a feature that catches all events that occur in the cluster and can pass them up in the stack or to applications running on OpenIO nodes. This enable |
https://en.wikipedia.org/wiki/Australasian%20Association%20of%20Clinical%20Geneticists | The Australasian Association of Clinical Geneticists (AACG) is a professional membership organization for medical specialists who are qualified to work in the field of clinical genetics. The Association was founded in 1995. As of 2021, the organization had approximately 180 members. The Association's members include fully qualified clinical geneticists from Australia and New Zealand as well as individuals training in the field from those jurisdictions. The Association is a Special Interest Group of the Human Genetics Society of Australasia (HGSA).
Mission
The AACG's mission is to:
serve as a special interest group of the HGSA.
provide representation for members of the HGSA who work in the field of clinical genetics in Australia and New Zealand to the HGSA and to other relevant bodies (e.g. the Royal Australasian College of Physicians (RACP)).
Purpose
The purpose of the AACG is to:
provide professional representation, on behalf of the HGSA, to external bodies on matters relating to clinical genetics.
provide clinical expertise and advice to government and policymakers on all matters relating to medical and clinical genetics.
develop and maintain professional standards in the field of clinical genetics.
promote and support teaching and education for individuals training and working in the field of clinical genetics in Australia and New Zealand.
AACG Biannual Meeting
The AACG's Biannual Meeting is the only National meeting of clinical geneticists in the region. The meeting brings specialists in the field together in the one place for education, clinical case discussions, and professional networking. The March meeting is held in an Australian capital city, on a rotational basis, and consists of an educational day for trainees plus a day for clinical case discussion for all members. A business meeting of the AACG is also held. The second meeting of the year is held concurrently with the Annual Scientific Meeting of the HGSA. The Association's annual general m |
https://en.wikipedia.org/wiki/Radial%20notch | The radial notch of the ulna (lesser sigmoid cavity) is a narrow, oblong, articular depression on the lateral side of the coronoid process; it receives the circumferential articular surface of the head of the radius.
It is concave from before backward, and its prominent extremities serve for the attachment of the annular ligament.
Additional images |
https://en.wikipedia.org/wiki/Aerospace%20materials | Aerospace materials are materials, frequently metal alloys, that have either been developed for, or have come to prominence through their use for aerospace purposes.
These uses often require exceptional performance, strength or heat resistance, even at the cost of considerable expense in their production or machining. Others are chosen for their long-term reliability in this safety-conscious field, particularly for their resistance to fatigue.
The field of materials engineering is an important one within aerospace engineering. Its practice is defined by the international standards bodies who maintain standards for the materials and processes involved. Engineers in this field may often have studied for degrees or post-graduate qualifications in it as a speciality.
History
Edwardian period
The first aerospace materials were those long-established and often naturally occurring materials used to construct the first aircraft. These included such mundane materials as timber for wing structures and fabric and dope to cover them. Their quality was of utmost importance and so the timber would be of carefully selected sitka spruce and the covering of irish linen. Standards were required for the selection, manufacture, and use of these materials. These standards were developed informally by manufacturers or government groups such as HM Balloon Factory, later to become RAE Farnborough, often with the assistance of university engineering departments.
The next stage in the development of aerospace materials was to adopt newly developed materials, such as Duralumin the first age hardening aluminum alloy. These offered attributes not previously available. Many of these new materials also required study to determine the extent of these new properties, their behavior and how to make the best use of them. This work was often carried out through the new government-funded national laboratories, such as the Reichsanstalt (German Imperial Institute) or the British National Physical |
https://en.wikipedia.org/wiki/Relative%20permeability | In multiphase flow in porous media, the relative permeability of a phase is a dimensionless measure of the effective permeability of that phase. It is the ratio of the effective permeability of that phase to the absolute permeability. It can be viewed as an adaptation of Darcy's law to multiphase flow.
For two-phase flow in porous media given steady-state conditions, we can write
where is the flux, is the pressure drop, is the viscosity. The subscript indicates that the parameters are for phase .
is here the phase permeability (i.e., the effective permeability of phase ), as observed through the equation above.
Relative permeability, , for phase is then defined from , as
where is the permeability of the porous medium in single-phase flow, i.e., the absolute permeability. Relative permeability must be between zero and one.
In applications, relative permeability is often represented as a function of water saturation; however, owing to capillary hysteresis one often resorts to a function or curve measured under drainage and another measured under imbibition.
Under this approach, the flow of each phase is inhibited by the presence of the other phases. Thus the sum of relative permeabilities over all phases is less than 1. However, apparent relative permeabilities larger than 1 have been obtained since the Darcean approach disregards the viscous coupling effects derived from momentum transfer between the phases (see assumptions below). This coupling could enhance the flow instead of inhibit it. This has been observed in heavy oil petroleum reservoirs when the gas phase flows as bubbles or patches (disconnected).
Modelling assumptions
The above form for Darcy's law is sometimes also called Darcy's extended law, formulated for horizontal, one-dimensional, immiscible multiphase flow in homogeneous and isotropic porous media. The interactions between the fluids are neglected, so this model assumes that the solid porous media and the other fluids form a new p |
https://en.wikipedia.org/wiki/Matroshka%20experiments | The Matroshka (or Matroschka or Phantom) experiments on the International Space Station (ISS) use a mannequin that has been used to study cosmic radiation dose types and rates that relate to the health of space travellers on long duration missions.
The two-part experiment is named for the Matryoshka dolls or Russian nested doll that has various layers of dolls, with the inner layers revealed when the outer layers are opened. In this experiment, the doll has measured the radiation doses of the separate components of the ionizing cosmic radiation at the skin surface and at different locations inside a realistic human torso, in order to establish the relation between skin doses and organ doses.
Details
Matroshka is a human torso (mannequin), a base structure and a container. The container is a carbon fibre structure and formed, with the base structure, a closed volume that contained a dry oxygen atmosphere and protected the torso against space vacuum, space debris, solar UV and material offgassing. It also acts more generally as a simulation of a space suit worn by astronauts during a space walk. Temperature, pressure and experiment data were collected during the mission and transferred to the onboard computer system of ISS, then to the Earth stations for transmission to the experimenters. The torso uses commercial parts common to the field of radiotherapy; various instrumented 'slices' were composed of natural bones embedded in plastics simulating tissue and lung.
The principal investigator is Guenter Reitz, Ph.D. of the German Aerospace Center (DLR), Cologne, Germany.
MTR-1
Matroshka-1 (MTR-1) was sent to the ISS aboard the Soyuz-U/Progress M1-11 supply vehicle launched from Baikonur Cosmodrome on 24 January 2004, and was placed on the outside of the Russian Zvezda module during a spacewalk during Expedition 8 by Alexander Kaleri and Michael Foale on 15 March 2004, and brought inside during Expedition 11 on 18 August 2005 then the experimental elements were r |
https://en.wikipedia.org/wiki/Ditkin%20set | In mathematics, a Ditkin set, introduced by , is a closed subset of the circle such that a function f vanishing on the set can be approximated by functions φnf with φ vanishing in a neighborhood of the set. |
https://en.wikipedia.org/wiki/Molecular%20genetics | Molecular genetics is a branch of biology that addresses how differences in the structures or expression of DNA molecules manifests as variation among organisms. Molecular genetics often applies an "investigative approach" to determine the structure and/or function of genes in an organism's genome using genetic screens.
The field of study is based on the merging of several sub-fields in biology: classical Mendelian inheritance, cellular biology, molecular biology, biochemistry, and biotechnology. It integrates these disciplines to explore things like genetic inheritance, gene regulation and expression, and the molcular mechanims behind various life processes.
A key goal of molecular genetics is to identfiy and study genetic mutations. Researchers search for mutations in a gene or induce mutations in a gene to link a gene sequence to a specific phenotype. Therefore molecular genetics is a powerful methodology for linking mutations to genetic conditions that may aid the search for treatments of various genetics diseases.
History
For molecular genetics to develop as a discipline, several scientific discoveries were necessary. The discovery of DNA as a means to transfer the genetic code of life from one cell to another and between generations was essential for identifying the molecule responsible for heredity. Molecular genetics arose initially from studies involving genetic transformation in bacteria. In 1944 Avery, McLeod and McCarthy isolated DNA from a virulent strain of S. pneumoniae, and using just this DNA were able to convert a harmless strain to virulence. They called the uptake, incorporation and expression of DNA by bacteria "transformation". This finding suggested that DNA is the genetic material of bacteria. Since its discovery in 1944 genetic transformation has been found to occur in numerous bacterial species including many species that are pathogenic to humans. Bacterial transformation is often induced by conditions of stress, and the func |
https://en.wikipedia.org/wiki/WebRTC | WebRTC (Web Real-Time Communication) is a free and open-source project providing web browsers and mobile applications with real-time communication (RTC) via application programming interfaces (APIs). It allows audio and video communication to work inside web pages by allowing direct peer-to-peer communication, eliminating the need to install plugins or download native apps.
Supported by Apple, Google, Microsoft, Mozilla, and Opera, WebRTC specifications have been published by the World Wide Web Consortium (W3C) and the Internet Engineering Task Force (IETF).
According to the webrtc.org website, the purpose of the project is to "enable rich, high-quality RTC applications to be developed for the browser, mobile platforms, and IoT devices, and allow them all to communicate via a common set of protocols".
History
In May 2010, Google bought Global IP Solutions or GIPS, a VoIP and videoconferencing software company that had developed many components required for RTC, such as codecs and echo cancellation techniques. Google open-sourced the GIPS technology and engaged with relevant standards bodies at the IETF and W3C to ensure industry consensus. In May 2011, Google released an open-source project for browser-based real-time communication known as WebRTC. This has been followed by ongoing work to standardize the relevant protocols in the IETF and browser APIs in the W3C.
In January 2011, Ericsson Labs built the first implementation of WebRTC using a modified WebKit library. In October 2011, the W3C published its first draft for the spec. WebRTC milestones include the first cross-browser video call (February 2013), first cross-browser data transfers (February 2014), and as of July 2014 Google Hangouts was "kind of" using WebRTC.
The W3C draft API was based on preliminary work done in the WHATWG. It was referred to as the ConnectionPeer API, and a pre-standards concept implementation was created at Ericsson Labs. The WebRTC Working Group expects this specification to |
https://en.wikipedia.org/wiki/Glycogen | Glycogen is a multibranched polysaccharide of glucose that serves as a form of energy storage in animals, fungi, and bacteria. It is the main storage form of glucose in the human body.
Glycogen functions as one of three regularly used forms of energy reserves, creatine phosphate being for very short-term, glycogen being for short-term and the triglyceride stores in adipose tissue (i.e., body fat) being for long-term storage. Protein, broken down into amino acids, is seldom used as a main energy source except during starvation and glycolytic crisis (see bioenergetic systems).
In humans, glycogen is made and stored primarily in the cells of the liver and skeletal muscle. In the liver, glycogen can make up 5–6% of the organ's fresh weight: the liver of an adult, weighing 1.5 kg, can store roughly 100–120 grams of glycogen. In skeletal muscle, glycogen is found in a low concentration (1–2% of the muscle mass): the skeletal muscle of an adult weighing 70 kg stores roughly 400 grams of glycogen. Small amounts of glycogen are also found in other tissues and cells, including the kidneys, red blood cells, white blood cells, and glial cells in the brain. The uterus also stores glycogen during pregnancy to nourish the embryo.
The amount of glycogen stored in the body mostly depends on oxidative type 1 fibres, physical training, basal metabolic rate, and eating habits. Different levels of resting muscle glycogen are reached by changing the number of glycogen particles, rather than increasing the size of existing particles though most glycogen particles at rest are smaller than their theoretical maximum.
Approximately 4 grams of glucose are present in the blood of humans at all times; in fasting individuals, blood glucose is maintained constant at this level at the expense of glycogen stores in the liver and skeletal muscle. Glycogen stores in skeletal muscle serve as a form of energy storage for the muscle itself; however, the breakdown of muscle glycogen impedes muscle |
https://en.wikipedia.org/wiki/DeepArt | DeepArt or DeepArt.io was a website that allowed users to create artistic images by using an algorithm to redraw one image using the stylistic elements of another image. with "A Neural Algorithm of Artistic Style" a Neural Style Transfer algorithm that was developed by several of its creators to separate style elements from a piece of art. The tool allows users to create imitation works of art using the style of various artists. The neural algorithm is used by the Deep Art website to create a representation of an image provided by the user by using the 'style' of another image provided by the user. A similar program, Prisma, is an iOS and Android app that was based on the open source programming that underlies DeepArt.
See also
Computational creativity |
https://en.wikipedia.org/wiki/Methaniazide/thioacetazone | Methaniazide/thioacetazone, sold under the brand name Neothetazone, is an antibiotic combination of methaniazide (neotizide) and thioacetazone that is or was very commonly used in the treatment of tuberculosis. It has been implicated as a cause of gigantomastia in a single 1970 case report, and, along with D-penicilliamine, bucillamine, ciclosporin, and indinavir, is one of the only drugs to have been associated with gigantomastia.
See also
Conteben
Isoniazid |
https://en.wikipedia.org/wiki/Phonetotopy | Phonetotopy is the concept that articulatory features as well as perceptual features of speech sounds are ordered in the brain in a similar way as tone (tonotopy), articulation and its somatosensory feedback (somatotopy), or visual location of an object (retinotopy). It is assumed that a phonetotopic ordering of speech sounds as well as of syllables can be found at a supramodal speech processing level (i.e. at a phonetic speech processing level) within the brain.
The concept of phonetotopy was introduced in Kröger et al. (2009) on the basis of modeling speech production, speech perception, as well as speech acquisition. Moreover, fMRI measurements on ordering of vowels with respect to phonetic features as well as EEG-array measurements on vowel and syllable articulation support this concept.
The concept of phonetotopy at least underpins the concept of distinctive features, which are phonetically based features of speech sounds (i.e. based in perceptual as well as in articulatory domain), but which as well are linguistically (or phonologically) relevant, and thus are realized in a language specific way in humans. |
https://en.wikipedia.org/wiki/Timeline%20of%20calculus%20and%20mathematical%20analysis | A timeline of calculus and mathematical analysis.
500BC to 1600
5th century BC - The Zeno's paradoxes,
5th century BC - Antiphon attempts to square the circle,
5th century BC - Democritus finds the volume of cone is 1/3 of volume of cylinder,
4th century BC - Eudoxus of Cnidus develops the method of exhaustion,
3rd century BC - Archimedes displays geometric series in The Quadrature of the Parabola. Archimedes also discovers a method which is similar to differential calculus.
3rd century BC - Archimedes develops a concept of the indivisibles—a precursor to infinitesimals—allowing him to solve several problems using methods now termed as integral calculus. Archimedes also derives several formulae for determining the area and volume of various solids including sphere, cone, paraboloid and hyperboloid.
Before 50 BC - Babylonian cuneiform tablets show use of the Trapezoid rule to calculate of the position of Jupiter.
3rd century - Liu Hui rediscovers the method of exhaustion in order to find the area of a circle.
4th century - The Pappus's centroid theorem,
5th century - Zu Chongzhi established a method that would later be called Cavalieri's principle to find the volume of a sphere.
600 - Liu Zhuo is the first person to use second-order interpolation for computing the positions of the sun and the moon.
665 - Brahmagupta discovers a second order Taylor interpolation for ,
862 - The Banu Musa brothers write the "Book on the Measurement of Plane and Spherical Figures",
9th century - Thābit ibn Qurra discusses the quadrature of the parabola and the volume of different types of conic sections.
12th century - Bhāskara II discovers a rule equivalent to Rolle's theorem for ,
14th century - Nicole Oresme proves of the divergence of the harmonic series,
14th century - Madhava discovers the power series expansion for , , and This theory is now well known in the Western world as the Taylor series or infinite series.
14th century - Parameshvara discovers a t |
https://en.wikipedia.org/wiki/Kedougou%20virus | Kedougou virus (KEDV) is an arbovirus of the Flaviviridae family, genus Flavivirus, belongs to the IV group of a ((+) ssRNA) viruses. The virus was first isolated from Aedes minutus mosquitoes in Senegal in 1971.
KEDV belongs to the Kedougou virus group. It is endemically present in the Kédougou Region, from which it takes its name, together with other arboviruses of the genus Flavivirus such as: Yellow fever virus (YFV), West Nile virus (WNV) and Zika virus (ZIKV). KEDV has also been isolated from Aedes circumluteolus mosquitoes collected in , KwaZulu-Natal, South Africa.
While there is serological evidence for previous infection in humans, and one isolate from humans identified in the Central African Republic, to date no disease has been reported. |
https://en.wikipedia.org/wiki/Graphotype%20%28machine%29 | Graphotype was a brand name used by the Addressograph-Multigraph Company for its range of metal plate embossing machines.
The machines were originally used to create address plates for the Addressograph system and mark military style identity tags and other industrial nameplates.
The machines came in a number of variants with sliding, hand wheel or keyboard selected letters. The keyboard models and some rotary select units were motorised to allow faster operation.
The machines used in making Addressograph plates would deboss (stamp into the plate) the letters into the plate resulting in a well defined printing surface resembling the typewriter fonts of the day on the reverse side that would be used to transfer the details (usually customer addresses) onto envelopes or form letters. The same style was used in the early 1940s to 1980s for the US military identification tags and the tag details could be transferred onto medical charts using a hand held imprinter in field hospital conditions.
These same machines also found use in marking other nameplates and rating plates in industry and for this an embossed (raised letters in the style found on contemporary credit cards) marking style was preferred for ease of reading and maintaining a flat surface on the back of the plate. Military tags moved over to this style when the imprinting use was deprecated in the late 1960s and new machines would only be supplied as embossing units as the address plate market had been taken over by the computer revolution. |
https://en.wikipedia.org/wiki/Wtfast | wtfast is a Canadian company that provides an optimized gaming network, also known as the Gamers Private Network (GPN), for MMO, FPS, and MOBA gamers. It is a platform that online gamers use to access gaming servers with an uninterrupted connection. The company is operated by AAA Internet Publishing, Inc. with headquarters in Kelowna, British Columbia.
History
The company was initially founded by Rob Bartlett in 1997, as AAA Internet Publishing Inc. to provide internet and technology services. However, the trademark name wtfast was registered in December 1, 2009, and the company has since operated under that name to provide lag reduction services for MMO games. In 2014, it was the top ten finalist of BCIC New Ventures Competition. In 2015, it was the top five national finalist for the TELUS and The Globe & Mail Small Business Competition. In 2016, wtfast was the finalist of Sir Richard Branson's Extreme Tech Challenge. Since 2017, some of the notable company partners include ASUS, MSI, Gigabyte, Linksys, HiNet, and ASI Networks. |
https://en.wikipedia.org/wiki/ZCCHC6 | Terminal uridylyltransferase 7 (TUT7), also known as "zinc finger, CCHC domain containing 6", is an enzyme that in humans is encoded by the ZCCHC6 gene located on chromosome 9. The ZCCHC6 protein mediates the terminal uridylation of RNA transcripts with short poly-A tails and is involved in mRNA and microRNA degradation
Structure
The ZCCHC6 gene contains 33 exons with at least six known isoforms due to alternative splicing. The ZCCHC6 gene encodes for a protein that is 171 kDa in molecular weight and is localized to the cytoplasm.
Function
It catalyzes the following reaction, requiring Mg2+ and Mn2+ as co-factors.
UTP + RNA(n) = diphosphate + RNA(n+1)
Uridylation catalyzed by ZCCHC6 takes place readily on deadenylated mRNAs inside the cells. Purified ZZHC6 selectively recognizes and uridylates RNA molecules possessing short poly(A) tails (less than 25 nucleotides) in vitro. In cells depleted of ZCCHC6, the majority of mRNAs lose the signature oligo(U) tails that are characteristic of ZCCHC6 reactivity, and the half-life of mRNA molecules are accordingly prolonged.
In addition to mRNA degradation, uridylation is also thought to function in pre-microRNA maturation, with some group II pre-microRNA requiring 3' mono-uridylation for Dicer processing. ZCCHC6 is thought to work in redundancy with ZCCHC11 to mediate the biogenesis of the let-7 microRNA through uridylation.
Genetic Inactivation of ZCCHC6 Suppresses Interleukin-6 Expression and Reduces the Severity of Experimental Osteoarthritis in Mice. |
https://en.wikipedia.org/wiki/Communication%20channel | A communication channel refers either to a physical transmission medium such as a wire, or to a logical connection over a multiplexed medium such as a radio channel in telecommunications and computer networking. A channel is used for information transfer of, for example, a digital bit stream, from one or several senders to one or several receivers. A channel has a certain capacity for transmitting information, often measured by its bandwidth in Hz or its data rate in bits per second.
Communicating an information signal across distance requires some form of pathway or medium. These pathways, called communication channels, use two types of media: Transmission line (e.g. twisted-pair, coaxial, and fiber-optic cable) and broadcast (e.g. microwave, satellite, radio, and infrared).
In information theory, a channel refers to a theoretical channel model with certain error characteristics. In this more general view, a storage device is also a communication channel, which can be sent to (written) and received from (reading) and allows communication of an information signal across time.
Examples
Examples of communications channels include:
A connection between initiating and terminating communication endpoints of a telecommunication circuit.
A single path provided by a transmission medium via either
physical separation, such as by multipair cable or
separation, such as by frequency-division or time-division multiplexing.
A path for conveying electrical or electromagnetic signals, usually distinguished from other parallel paths.
A data storage device which can communicate a message over time.
The portion of a storage medium, such as a track or band, that is accessible to a given reading or writing station or head.
A buffer from which messages can be put and got.
In a communications system, the physical or logical link that connects a data source to a data sink.
A specific radio frequency, pair or band of frequencies, usually named with a letter, number, or codewor |
https://en.wikipedia.org/wiki/Comparison%20of%20platform%20virtualization%20software | Platform virtualization software, specifically emulators and hypervisors, are software packages that emulate the whole physical computer machine, often providing multiple virtual machines on one physical platform. The table below compares basic information about platform virtualization hypervisors.
General
Features
Providing any virtual environment usually requires some overhead of some type or another. Native usually means that the virtualization technique does not do any CPU level virtualization (like Bochs), which executes code more slowly than when it is directly executed by a CPU. Some other products such as VMware and Virtual PC use similar approaches to Bochs and QEMU, however they use a number of advanced techniques to shortcut most of the calls directly to the CPU (similar to the process that JIT compiler uses) to bring the speed to near native in most cases. However, some products such as coLinux, Xen, z/VM (in real mode) do not suffer the cost of CPU-level slowdowns as the CPU-level instructions are not proxied or executing against an emulated architecture since the guest OS or hardware is providing the environment for the applications to run under. However access to many of the other resources on the system, such as devices and memory may be proxied or emulated in order to broker those shared services out to all the guests, which may cause some slow downs as compared to running outside of virtualization.
OS-level virtualization is described as "native" speed, however some groups have found overhead as high as 3% for some operations, but generally figures come under 1%, so long as secondary effects do not appear.
See for a paper comparing performance of paravirtualization approaches (e.g. Xen) with OS-level virtualization
Requires patches/recompiling.
Exceptional for lightweight, paravirtualized, single-user VM/CMS interactive shell: largest customers run several thousand users on even single prior models. For multiprogramming OSes like Linux |
https://en.wikipedia.org/wiki/Collider | A collider is a type of particle accelerator that brings two opposing particle beams together such that the particles collide. Colliders may either be ring accelerators or linear accelerators.
Colliders are used as a research tool in particle physics by accelerating particles to very high kinetic energy and letting them impact other particles. Analysis of the byproducts of these collisions gives scientists good evidence of the structure of the subatomic world and the laws of nature governing it. These may become apparent only at high energies and for tiny periods of time, and therefore may be hard or impossible to study in other ways.
Explanation
In particle physics one gains knowledge about elementary particles by accelerating particles to very high kinetic energy and letting them impact on other particles. For sufficiently high energy, a reaction occurs that transforms the particles into other particles. Detecting these products gives insight into the physics involved.
To do such experiments there are two possible setups:
Fixed target setup: A beam of particles (the projectiles) is accelerated with a particle accelerator, and as collision partner, one puts a stationary target into the path of the beam.
Collider: Two beams of particles are accelerated and the beams are directed against each other, so that the particles collide while flying in opposite directions. This process can be used to make strange and anti-matter.
The collider setup is harder to construct but has the great advantage that according to special relativity the energy of an inelastic collision between two particles approaching each other with a given velocity is not just 4 times as high as in the case of one particle resting (as it would be in non-relativistic physics); it can be orders of magnitude higher if the collision velocity is near the speed of light.
In the case of a collider where the collision point is at rest in the laboratory frame (i.e. ), the center of mass energy (the ene |
https://en.wikipedia.org/wiki/Shared%20secret | In cryptography, a shared secret is a piece of data, known only to the parties involved, in a secure communication. This usually refers to the key of a symmetric cryptosystem. The shared secret can be a password, a passphrase, a big number, or an array of randomly chosen bytes.
The shared secret is either shared beforehand between the communicating parties, in which case it can also be called a pre-shared key, or it is created at the start of the communication session by using a key-agreement protocol, for instance using public-key cryptography such as Diffie–Hellman or using symmetric-key cryptography such as Kerberos.
The shared secret can be used for authentication (for instance when logging into a remote system) using methods such as challenge–response or it can be fed to a key derivation function to produce one or more keys to use for encryption and/or MACing of messages.
To make unique session and message keys the shared secret is usually combined with an initialization vector (IV). An example of this is the derived unique key per transaction method.
It is also often used as an authentication measure in web APIs.
See also
Key stretching – a method to create a stronger key from a weak key or a weak shared secret
Security question – implementation method |
https://en.wikipedia.org/wiki/Abraham%20Wald | Abraham Wald (; , ; – ) was a Jewish Hungarian mathematician who contributed to decision theory, geometry and econometrics, and founded the field of sequential analysis. One of his well-known statistical works was written during World War II on how to minimize the damage to bomber aircraft and took into account the survivorship bias in his calculations. He spent his research career at Columbia University. He was the grandson of Rabbi Moshe Shmuel Glasner.
Life and career
Wald was born on 31 October 1902 in Kolozsvár, Transylvania, in the Kingdom of Hungary. A religious Jew, he did not attend school on Saturdays, as was then required by the Hungarian school system, and so he was homeschooled by his parents until college. His parents were quite knowledgeable and competent as teachers.
In 1928, he graduated in mathematics from the King Ferdinand I University. In 1927, he had entered graduate school at the University of Vienna, from which he graduated in 1931 with a Ph.D. in mathematics. His advisor there was Karl Menger.
Despite Wald's brilliance, he could not obtain a university position because of Austrian discrimination against Jews. However, Oskar Morgenstern created a position for Wald in economics. When Nazi Germany annexed Austria in 1938, the discrimination against Jews intensified. In particular, Wald and his family were persecuted as Jews. Wald immigrated to the United States at the invitation of the Cowles Commission for Research in Economics, to work on econometrics research.
During World War II, Wald was a member of the Statistical Research Group (SRG) at Columbia University, where he applied his statistical skills to various wartime problems. They included methods of sequential analysis and sampling inspection. One of the problems that the SRG worked on was to examine the distribution of damage to aircraft returning after flying missions to provide advice on how to minimize bomber losses to enemy fire. Wald derived a useful means of estimating the |
https://en.wikipedia.org/wiki/Neutron%20embrittlement | Neutron embrittlement, sometimes more broadly radiation embrittlement, is the embrittlement of various materials due to the action of neutrons. This is primarily seen in nuclear reactors, where the release of high-energy neutrons causes the long-term degradation of the reactor materials. The embrittlement is caused by the microscopic movement of atoms that are hit by the neutrons; this same action also gives rise to neutron-induced swelling causing materials to grow in size, and the Wigner effect causing energy buildup in certain materials that can lead to sudden releases of energy.
Neutron embrittlement mechanisms include:
Hardening and dislocation pinning due to nanometer features created by irradiation
Generation of lattice defects in collision cascades via the high-energy recoil atoms produced in the process of neutron scattering.
Diffusion of major defects, which leads to higher amounts of solute diffusion, as well as formation of nanoscale defect-solute cluster complexes, solute clusters, and distinct phases.
Embrittlement in Nuclear Reactors
Neutron irradiation embrittlement limits the service life of reactor-pressure vessels (RPV) in nuclear power plants due to the degradation of reactor materials. In order to perform at high efficiency and safely contain coolant water at temperatures around 290°C and pressures of ~7 MPa (for boiling water reactors) to 14 MPa (for pressurized water reactors), the RPV must be heavy-section steel. Due to regulations, RPV failure probabilities must be very low. To achieve sufficient safety, the design of the reactor assumes large cracks and extreme loading conditions. Under such conditions, a probable failure mode is rapid, catastrophic fracture if the vessel steel is brittle. Tough RPV base metals that are typically used are A302B, A533B plates, or A508 forgings; these are quenched and tempered, low-alloy steels with primarily tempered bainitic microstructures. Over the past few decades, RPV embrittlement has been addres |
https://en.wikipedia.org/wiki/Calculus%20on%20finite%20weighted%20graphs | In mathematics, calculus on finite weighted graphs is a discrete calculus for functions whose domain is the vertex set of a graph with a finite number of vertices and weights associated to the edges. This involves formulating discrete operators on graphs which are analogous to differential operators in calculus, such as graph Laplacians (or discrete Laplace operators) as discrete versions of the Laplacian, and using these operators to formulate differential equations, difference equations, or variational models on graphs which can be interpreted as discrete versions of partial differential equations or continuum variational models. Such equations and models are important tools to mathematically model, analyze, and process discrete information in many different research fields, e.g., image processing, machine learning, and network analysis.
In applications, finite weighted graphs represent a finite number of entities by the graph's vertices, any pairwise relationships between these entities by graph edges, and the significance of a relationship by an edge weight function. Differential equations or difference equations on such graphs can be employed to leverage the graph's structure for tasks such as image segmentation (where the vertices represent pixels and the weighted edges encode pixel similarity based on comparisons of Moore neighborhoods or larger windows), data clustering, data classification, or community detection in a social network (where the vertices represent users of the network, the edges represent links between users, and the weight function indicates the strength of interactions between users).
The main advantage of finite weighted graphs is that by not being restricted to highly regular structures such as discrete regular grids, lattice graphs, or meshes, they can be applied to represent abstract data with irregular interrelationships.
If a finite weighted graph is geometrically embedded in a Euclidean space, i.e., the graph vertices represent p |
https://en.wikipedia.org/wiki/Power%20Matters%20Alliance | Power Matters Alliance (PMA) was a global, not-for-profit, industry organization whose mission was to advance a suite of standards and protocols for wireless power transfer for mobile electronic devices (specifically a type of inductive charging that competes with the Qi standard). The organization was merged with Alliance for Wireless Power (A4WP) in 2015 to form AirFuel Alliance.
Founded by Procter & Gamble and Powermat Technologies in March 2012, PMA was networking technology companies in order to guarantee consumers interoperable devices which employed wireless power technology. Marked by the electron "P", PMA interface standard described analog power transfer (inductive and resonant), digital transceiver communication, cloud based power management, and environmental sustainability. The PMA board of directors included representatives from AT&T, Duracell, Starbucks, Powermat Technologies, Flextronics, Federal Communications Commission (FCC), and Energy Star. The membership of the PMA was made up of companies across the mobile device ecosystem, including handset providers, service providers, chipset suppliers, manufacturers, test labs and public establishments.
PMA standard and technology
The PMA's stated mission was to formulate and advance a suite of interface standards for smart and energy-efficient transfer of wireless power. The PMA was actively publishing a suite of standards based on inductive coupling technology to provide advanced inductive and resonant power. In addition, the PMA sought to add a digital layer providing policy-setting, monitoring, and extensible APIs. The PMA managed interoperability, certification and logo programs according to these specifications.
On February 11, 2014, the Alliance for Wireless Power (A4WP) and the PMA announced that they signed an agreement calling for the following immediate next steps:
PMA adopts the A4WP Rezence specification as the PMA magnetic resonance charging specification for both transmitters and recei |
https://en.wikipedia.org/wiki/Cram%C3%A9r%27s%20decomposition%20theorem | Cramér’s decomposition theorem for a normal distribution is a result of probability theory. It is well known that, given independent normally distributed random variables ξ1, ξ2, their sum is normally distributed as well. It turns out that the converse is also true. The latter result, initially announced by Paul Lévy, has been proved by Harald Cramér. This became a starting point for a new subfield in probability theory, decomposition theory for random variables as sums of independent variables (also known as arithmetic of probabilistic distributions).
The precise statement of the theorem
Let a random variable ξ be normally distributed and admit a decomposition as a sum ξ=ξ1+ξ2 of two independent random variables. Then the summands ξ1 and ξ2 are normally distributed as well.
A proof of Cramér's decomposition theorem uses the theory of entire functions.
See also
Raikov's theorem: Similar result for Poisson distribution. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.