source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Seaweed%20oil | Seaweed oil, also called algae oil or algal oil, is used for making food, with the purified product almost colorless and odorless. It is also under development as a possible alternative fuel and manufacturing agent.
Seaweed oil is also used as a source of fatty acid dietary supplement, as it contains mono- and polyunsaturated fats, in particular EPA and DHA, both of them omega-3 fatty acids. The supplement's DHA content is roughly equivalent to that of salmon-based fish oil supplement.
Seaweed oil is also used for biofuel, pharmaceutical manufacturing, massage oil, soaps, and lotions.
See also
List of omega-3 fatty acids
Edible seaweed |
https://en.wikipedia.org/wiki/K%20correction | K correction converts measurements of astronomical objects into their respective rest frames. The correction acts on that object's observed magnitude (or equivalently, its flux). Because astronomical observations often measure through a single filter or bandpass, observers only measure a fraction of the total spectrum, redshifted into the frame of the observer. For example, to compare measurements of stars at different redshifts viewed through a red filter, one must estimate K corrections to these measurements in order to make comparisons. If one could measure all wavelengths of light from an object (a bolometric flux), a K correction would not be required, nor would it be required if one could measure the light emitted in an emission line.
Carl Wilhelm Wirtz (1918), who referred to the correction as a Konstanten k (German for "constant") - correction dealing with the effects of redshift of in his work on Nebula. English-speaking claim for the origin of the term "K correction" is Edwin Hubble, who supposedly arbitrarily chose to represent the reduction factor in magnitude due to this same effect and who may not have been aware / given credit to the earlier work.
The K-correction can be defined as follows
I.E. the adjustment to the standard relationship between absolute and apparent magnitude required to correct for the redshift effect. Here, DL is the luminosity distance measured in parsecs.
The exact nature of the calculation that needs to be applied in order to perform a K correction depends upon the type of filter used to make the observation and the shape of the object's spectrum. If multi-color photometric measurements are available for a given object thus defining its spectral energy distribution (SED), K corrections then can be computed by fitting it against a theoretical or empirical SED template. It has been shown that K corrections in many frequently used broad-band filters for low-redshift galaxies can be precisely approximated using two-dimensio |
https://en.wikipedia.org/wiki/Gelsolin | Gelsolin is an actin-binding protein that is a key regulator of actin filament assembly and disassembly. Gelsolin is one of the most potent members of the actin-severing gelsolin/villin superfamily, as it severs with nearly 100% efficiency.
Cellular gelsolin, found within the cytosol and mitochondria, has a closely related secreted form, Plasma gelsolin, that contains an additional 24 AA N-terminal extension. Plasma gelsolin's ability to sever actin filaments helps the body recover from disease and injury that leaks cellular actin into the blood. Additionally it plays important roles in host innate immunity, activating macrophages and localizing of inflammation.
Structure
Gelsolin is an 82-kD protein with six homologous subdomains, referred to as S1-S6. Each subdomain is composed of a five-stranded β-sheet, flanked by two α-helices, one positioned perpendicular with respect to the strands and one positioned parallel. The β-sheets of the three N-terminal subdomains (S1-S3) join to form an extended β-sheet, as do the β-sheets of the C-terminal subdomains (S4-S6).
Regulation
Among the lipid-binding actin regulatory proteins, gelsolin (like cofilin) preferentially binds polyphosphoinositide (PPI). The binding sequences in gelsolin closely resemble the motifs in the other PPI-binding proteins.
Gelsolin's activity is stimulated by calcium ions (Ca2+). Although the protein retains its overall structural integrity in both activated and deactivated states, the S6 helical tail moves like a latch depending on the concentration of calcium ions. The C-terminal end detects the calcium concentration within the cell. When there is no Ca2+ present, the tail of S6 shields the actin-binding sites on one of S2's helices. When a calcium ion attaches to the S6 tail, however, it straightens, exposing the S2 actin-binding sites. The N-terminal is directly involved in the severing of actin. S2 and S3 bind to the actin before the binding of S1 severs actin-actin bonds and caps the |
https://en.wikipedia.org/wiki/Neuman%E2%80%93Stubblebine%20protocol | The Neuman–Stubblebine protocol is a computer network authentication protocol designed for use on insecure networks (e.g., the Internet). It allows individuals communicating over such a network to prove their identity to each other. This protocol utilizes time stamps, but does not depend on synchronized clocks.
The protocol
If Alice (A) initiates the communication to Bob (B) with S is a server trusted by both parties, the protocol can be specified as follows using security protocol notation:
A and B are identities of Alice and Bob respectively
M is a session identifier
is a symmetric key known only to A and S
is a symmetric key known only to B and S
and are nonces generated by A and B respectively
and are timestamps generated by A and B respectively
is a generated symmetric key, which will be the session key of the session between A and B
Alice notified Bob of intent to initiate secure communication.
Bob generates a times stamp and a nonce, and sends this to the trusted Server.
The trusted Server generates a session key and a message for Alice to forward to Bob.
Alice forwards the message and verifies is the same that she generated earlier. Bob will verify and have not changed when he receives the message.
Subsequent communications
An advantage provided by this protocol is that Alice can utilize the trusted Server's message to initiate authentication with Bob within some predetermined time limit without utilizing the trusted Server. The protocol proceeds as follows using the same definitions as above.
Alice sends the message the trusted Server sent her previously when communication with Bob.
Bob sends Alice a new nonce and her new nonce encrypted with session key that Alice resent from previous communication.
Alice returns Bob's nonce encrypted with the session key to allow him to verify the communication is successful.
Attack
Using the theorem prover SPASS it has been shown that this protocol can be attacked. This attack and two more f |
https://en.wikipedia.org/wiki/Celsius%20%28microarchitecture%29 | Celsius is the codename for a GPU microarchitecture developed by Nvidia, and released in 1999, as the successor to Fahrenheit (NV4, NV3...) microarchitecture. It was named with reference to Celsius and used with the GeForce 256 and GeForce 2 series.
Graphics features
DirectX 7.0
OpenGL 1.2 (1.5)
Max VRAM size bumped to 128MB
Chips
GeForce 256
NV10, 17 million transistor
GeForce 2 series
NV11, 20 million transistor
NV15, 25 million transistor
NV17, 29 million transistor
NV18, 29 million transistor
Crush11, 20 million transistor
Crush17, 29 million transistor
GPU list
GeForce 256
GeForce 2 series
See also
List of Nvidia graphics processing units
Scalable Link Interface (SLI)
Qualcomm Adreno |
https://en.wikipedia.org/wiki/Journal%20of%20the%20History%20of%20the%20Behavioral%20Sciences | The Journal of the History of the Behavioral Sciences is a quarterly peer-reviewed academic journal covering the history of social and behavioral sciences. It was established in 1965 and is published by John Wiley & Sons. The editor-in-chief is Alexandra Rutherford (York University). According to the Journal Citation Reports, the journal has a 2020 impact factor of 0.667, ranking it 22nd out of 34 journals in the category "History of Social Sciences". |
https://en.wikipedia.org/wiki/Pallister%E2%80%93Hall%20syndrome | Pallister–Hall syndrome (PHS) is a rare genetic disorder that affects various body systems. The main features are a non-cancerous mass on the hypothalamus (hypothalamic hamartoma) and extra digits (polydactylism). The prevalence of Pallister-Hall Syndrome is unknown; about 100 cases have been reported in publication.
History
The syndrome was originally described by American and Canadian geneticists Philip Pallister and Judith Hall in their research of newborn deaths due to pituitary failure. Subsequent discovery of living children and adults expanded the understanding of the syndrome and established the transmission pattern within families.
Presentation
The main characteristics of the syndrome are extra fingers and/or toes (polydactyly), with the skin between some fingers or toes potentially fused or "webbed" (cutaneous syndactyly), and a benign mass or lesion in the brain called a hypothalamic hamartoma. This benign tumor may not cause any medical problems; however, some hypothalamic hamartomas lead to seizures or hormone abnormalities. Other features of Pallister–Hall syndrome include a split opening of the airway called bifid epiglottis, laryngeal cleft, blockage of the anal opening (imperforate anus), and kidney abnormalities. Signs and symptoms of this disorder vary from mild to severe.
Seizures
The most common type of seizure from this disorder of that occur is known as gelastic epilepsy or "laughing" seizures. Seizures may begin at any age but usually before three or four years of age. The seizures usually start with laughter described as being "hollow" or "empty" and unpleasant. The laughter occurs suddenly for no obvious reason and is often out of place. Other traditional seizure types such as tonic-clonic and absence seizures may also develop due to temporal lobe epilepsy. People with Pallister-Hall may experience less severe seizures than people with a hypothalamic hamartoma only.
Transmission
Pallister-Hall Syndrome occurs due to a mutation in the G |
https://en.wikipedia.org/wiki/Origin%20of%20avian%20flight | Around 350 BCE, Aristotle and other philosophers of the time attempted to explain the aerodynamics of avian flight. Even after the discovery of the ancestral bird Archaeopteryx which lived over 150 million years ago, debates still persist regarding the evolution of flight. There are three leading hypotheses pertaining to avian flight: Pouncing Proavis model, Cursorial model, and Arboreal model.
In March 2018, scientists reported that Archaeopteryx was likely capable of flight, but in a manner substantially different from that of modern birds.
Flight characteristics
For flight to occur, four physical forces (thrust and drag, lift and weight) must be favorably combined. In order for birds to balance these forces, certain physical characteristics are required. Asymmetrical wing feathers, found on all flying birds with the exception of hummingbirds, help in the production of thrust and lift. Anything that moves through the air produces drag due to friction. The aerodynamic body of a bird can reduce drag, but when stopping or slowing down a bird will use its tail and feet to increase drag. Weight is the largest obstacle birds must overcome in order to fly. An animal can more easily attain flight by reducing its absolute weight. Birds evolved from other theropod dinosaurs that had already gone through a phase of size reduction during the Middle Jurassic, combined with rapid evolutionary changes. Flying birds during their evolution further reduced relative weight through several characteristics such as the loss of teeth, shrinkage of the gonads out of mating season, and fusion of bones. Teeth were replaced by a lightweight bill made of keratin, the food being processed by the bird's gizzard. Other advanced physical characteristics evolved for flight are a keel for the attachment of flight muscles and an enlarged cerebellum for fine motor coordination. These were gradual changes, though, and not strict conditions for flight: the first birds had teeth, at best a small keel |
https://en.wikipedia.org/wiki/Creep%20and%20shrinkage%20of%20concrete | Creep and shrinkage of concrete are two physical properties of concrete. The creep of concrete, which originates from the calcium silicate hydrates (C-S-H) in the hardened Portland cement paste (which is the binder of mineral aggregates), is fundamentally different from the creep of metals and polymers. Unlike the creep of metals, it occurs at all stress levels and, within the service stress range, is linearly dependent on the stress if the pore water content is constant. Unlike the creep of polymers and metals, it exhibits multi-months aging, caused by chemical hardening due to hydration which stiffens the microstructure, and multi-year aging, caused by long-term relaxation of self-equilibrated micro-stresses in the nano-porous microstructure of the C-S-H. If concrete is fully dried, it does not creep, but it is next to impossible to dry concrete fully without severe cracking.
Changes of pore water content due to drying or wetting processes cause significant volume changes of concrete in load-free specimens. They are called the shrinkage (typically causing strains between 0.0002 and 0.0005, and in low strength concretes even 0.0012) or swelling (< 0.00005 in normal concretes, < 0.00020 in high strength concretes). To separate shrinkage from creep, the compliance function , defined as the stress-produced strain (i.e., the total strain minus shrinkage) caused at time t by a unit sustained uniaxial stress applied at age , is measured as the strain difference between the loaded and load-free specimens.
The multi-year creep evolves logarithmically in time (with no final asymptotic value), and over the typical structural lifetimes it may attain values 3 to 6 times larger than the initial elastic strain. When a deformation is suddenly imposed and held constant, creep causes relaxation of critically produced elastic stress. After unloading, creep recovery takes place, but it is partial, because of aging.
In practice, creep during drying is inseparable from shrinkage. |
https://en.wikipedia.org/wiki/Chromosome%20jumping | Chromosome jumping is a tool of molecular biology that is used in the physical mapping of genomes. It is related to several other tools used for the same purpose, including chromosome walking.
Chromosome jumping is used to bypass regions difficult to clone, such as those containing repetitive DNA, that cannot be easily mapped by chromosome walking, and is useful in moving along a chromosome rapidly in search of a particular gene. Unlike chromosome walking, chromosome jumping is able to start on one point of the chromosome in order to traverse potential distant point of the same chromosome without cloning the intervening sequences. The ends of a large DNA fragment is the target cloning section of the chromosome jumping while the middle section gets removed by sequences of chemical manipulations prior to the cloning step.
Process
Chromosome jumping enables two ends of a DNA sequence to be cloned without the middle section. Genomic DNA may be partially digested using restriction endonuclease and with the aid of DNA ligase, the fragments are circularized at low concentration. From a known sequence, a primer is designed to sequence across the circularized junction. This primer is used to jump 100 kb-300 kb intervals: a sequence 100 kb away would have come near the known sequence on circularization, it permits jumping and sequencing in an alternative manner. Thus, sequences not reachable by chromosome walking can be sequenced. Chromosome walking can also be used from the new jump position (in either direction) to look for gene-like sequences, or additional jumps can be used to progress further along the chromosome. Combining chromosome jumping to chromosome walking through the chromosome allows bypassing repetitive DNA for the search of the target gene.
Library
Chromosome jumping library is different from chromosome walking due to the manipulations executed before the cloning step. In order to construct the library of chromosome jumping, individual clones originate |
https://en.wikipedia.org/wiki/Clinical%20ecology | Clinical ecology was the name given by proponents in the 1960s to a claim that exposure to low levels of certain chemical agents harm susceptible people, causing multiple chemical sensitivity and other disorders. Clinical ecologists are people that support and promote this offshoot of conventional medicine. They often have a background in the field of allergy or otorhinolaryngology, and the theoretical approach is derived in part from classic concepts of allergic responses, first articulated by Theron Randolph and developed by Richard Mackarness.
Clinical ecologists support a cause-and-effect relationship for non-specific symptoms reported by some people after low-dose exposure to chemical, biologic, or physical agents. This pattern of low-dose reaction is not generally accepted by toxicologists. Although some of the mainstream medical community continue to reject these claims, the concept is gaining some recognition under the modern and more clearly articulated classification of environmental medicine.
Training and qualifications
"Clinical Ecologist" is an environmental approach that is consistent with the practice of holistic medicine. Practitioners with this orientation do not use the term "Clinical Ecologist," although those opposed to this complementary medicine approach to illness often still do. Unlike terms such as physician or nurse, the term clinical ecologist is not legally regulated in any jurisdiction, which means that any person may legally claim to be a clinical ecologist. If wanted, they may obtain an extralegal certification or membership from the unregulated private organization American Academy of Environmental Medicine upon payment of a fee.
Many clinical ecologists are traditionally licensed healthcare professionals who hold advanced traditional medical certifications. Others may have a more alternative training.
History
Randolph published a number of books to promote clinical ecology and environmental medicine, including:
In 196 |
https://en.wikipedia.org/wiki/Directional%20symmetry%20%28time%20series%29 | In statistical analysis of time series and in signal processing, directional symmetry is a statistical measure of a model's performance in predicting the direction of change, positive or negative, of a time series from one time period to the next.
Definition
Given a time series with values at times and a model that makes predictions for those values , then the directional symmetry (DS) statistic is defined as
Interpretation
The DS statistic gives the percentage of occurrences in which the sign of the change in value from one time period to the next is the same for both the actual and predicted time series. The DS statistic is a measure of the performance of a model in predicting the direction of value changes. The case would indicate that a model perfectly predicts the direction of change of a time series from one time period to the next.
See also
Statistical finance
Notes and references
Drossu, Radu, and Zoran Obradovic. "INFFC data analysis: lower bounds and testbed design recommendations." Computational Intelligence for Financial Engineering (CIFEr), 1997., Proceedings of the IEEE/IAFE 1997. IEEE, 1997.
Lawrance, A. J., "Directionality and Reversibility in Time Series", International Statistical Review, 59 (1991), 67–79.
Tay, Francis EH, and Lijuan Cao. "Application of support vector machines in financial time series forecasting." Omega 29.4 (2001): 309–317.
Xiong, Tao, Yukun Bao, and Zhongyi Hu. "Beyond one-step-ahead forecasting: Evaluation of alternative multi-step-ahead forecasting models for crude oil prices." Energy Economics 40 (2013): 405–415.
Symmetry
Signal processing |
https://en.wikipedia.org/wiki/Reduce%20%28computer%20algebra%20system%29 | Reduce is a general-purpose computer algebra system geared towards applications in physics.
The development of the Reduce computer algebra system was started in the 1960s by Anthony C. Hearn. Since then, many scientists from all over the world have contributed to its development under his direction.
Reduce is written entirely in its own LISP dialect called Portable Standard Lisp, expressed in an ALGOL-like syntax called RLISP. The latter is used as a basis for Reduce's user-level language.
Implementations of Reduce are available on most variants of Unix, Linux, Microsoft Windows, or Apple Macintosh systems by using an underlying Portable Standard Lisp or Codemist Standard LISP implementation. The Julia package Reduce.jl uses Reduce as a backend and implements its semantics in Julia style.
Reduce was open sourced in December 2008 and is available for free under a modified BSD license on SourceForge. Previously it had cost $695.
See also
Comparison of computer algebra systems
ALTRAN
REDUCE Meets CAMAL - REDUCE Computer Algebra System - J. P. Fitch |
https://en.wikipedia.org/wiki/Renal%20artery | The renal arteries are paired arteries that supply the kidneys with blood. Each is directed across the crus of the diaphragm, so as to form nearly a right angle.
The renal arteries carry a large portion of total blood flow to the kidneys. Up to a third of total cardiac output can pass through the renal arteries to be filtered by the kidneys.
Structure
The renal arteries normally arise at a 90° angle off of the left interior side of the abdominal aorta, immediately below the superior mesenteric artery. They have a radius of approximately 0.25 cm, 0.26 cm at the root. The measured mean diameter can differ depending on the imaging method used. For example, the diameter was found to be 5.04 ± 0.74 mm using ultrasound but 5.68 ± 1.19 mm using angiography.
Due to the anatomical position of the aorta, the inferior vena cava, and the kidneys, the right renal artery is normally longer than the left renal artery.
The right passes behind the inferior vena cava, the right renal vein, the head of the pancreas, and the descending part of the duodenum. It’s somewhat lower than the left one.
Left artery lies behind the left renal vein, the body of the pancreas and the splenic vein, and is crossed by the inferior mesenteric vein.
Branches
Before reaching the hilus of the kidney, each artery divides into four or five branches. The anterior branches (the upper, middle, lower and apical segmental arteries) lie between the renal vein and ureter, the vein being in front, the ureter behind. The posterior branches, which are fewer in number and include the posterior segmental artery, are usually situated behind the ureter.
Each vessel gives off some small inferior suprarenal branches to the suprarenal gland, the ureter, and the surrounding cellular tissue and muscles.
One or two accessory renal arteries are frequently found, especially on the left side since they usually arise from the aorta, and may come off above (more common) or below the main artery. Instead of entering the ki |
https://en.wikipedia.org/wiki/%C3%98ystein%20Fischer | Øystein Fischer (born 9 March 1942 in Bergen, died 19 September 2013) was a Norwegian physicist and specialist in the field of superconductivity. He was a professor of the Faculty of Science of the University of Geneva. He was also the founder and director of the Swiss National Center of Competence in Research MaNEP (Materials with Novel Electronic Properties), dedicated to exploring materials of the future.
Career
After having worked as a technical research assistant for the laboratory Nera A/S in Bergen, Norway, Fischer studied physics at the Swiss Federal Institute of Technology in Zurich. He joined the University of Geneva in 1967 and obtained his PhD in 1971. He was appointed assistant professor at the University of Geneva in the same year. In 1977 he became a full professor.
Research
In 1975, he synthesized the first superconducting compounds containing a regular lattice of magnetic ions, a discovery opening up a decade of international research concerning the interaction between magnetism and superconductivity. This work was highlighted by his discovery in 1984 of superconductivity induced via magnetic field.
With his team, Fischer launched the first artificial superlattices of superconductor cuprates, the pioneering work of many developments in new areas of thin films and oxide interfaces.
From 1986, Fischer assigned a part of his team to work in scanning tunneling microscopy which allowed him to probe the fundamental properties of high temperature superconductors.
In 2001, he founded and became director of the NCCR (PRN) MaNEP dedicated to the study of materials with novel electronic properties.
Fischer initiated the Geneva Creativity Center whose purpose is to stimulate discussion between the academic and industrial sectors and to find innovative solutions for future technological challenges.
Fischer was also the head of the project "Centre for astronomical, physical and mathematical sciences of Geneva".
Over the past 20 years, Fischer has focused h |
https://en.wikipedia.org/wiki/Oyster%20wave%20energy%20converter | The Oyster is a hydro-electric wave energy device that uses the motion of ocean waves to generate electricity. It is made up of a Power Connector Frame (PCF), which is bolted to the seabed, and a Power Capture Unit (PCU). The PCU is a hinged buoyant flap that moves back and forth with movement of the waves. The movement of the flap drives two hydraulic pistons that feed high-pressured water to an onshore hydro-electric turbine, which drives a generator to make electricity. Oyster was stationed at the European Marine Energy Centre (EMEC) at its Billia Croo site in Orkney, Scotland until the company ceased trading in 2015.
Aquamarine Power installed Oyster at the EMEC in August 2009. On 20 November 2009, Oyster was officially launched and connected to the National Grid (UK) by the First Minister of Scotland, Alex Salmond.
Development work was started to build a more efficient and powerful second-generation device, Oyster 2.
History
Oyster was developed by Edinburgh-based Aquamarine Power, a company that focuses on wave energy. The concept originated from research at Queen's University, Belfast, led by professor Trevor Whittaker, Head of the Wave Power Research Centre at Queen's. Aquamarine Power also teamed up with Renewable Technology Ventures Ltd (STVL), a subsidiary of Scottish and Southern Energy (SEE), to fund the Oyster project. Aquamarine Power was able to secure a £6.3m investment from Scottish Enterprise. In addition, Scottish Enterprises awarded Aquamarine Power a £3.15 million grant from the Wave and Tidal Energy: Research, Development and Demonstration Support fund (WATERS). Aquamarine Power also received £1.5m from Sigma Capital Group plc. Altogether, Aquamarine Power was able to raise £11 million to stage this project.
In June 2009, Aquamarine Power signed a £2.5 million contract with Fugro Seacore to install the Oyster device at the European Marine Energy Centre test site at Billia Croo. Oyster was installed 400 metres offshore, west of the Orkney m |
https://en.wikipedia.org/wiki/Photoluminescence | Photoluminescence (abbreviated as PL) is light emission from any form of matter after the absorption of photons (electromagnetic radiation). It is one of many forms of luminescence (light emission) and is initiated by photoexcitation (i.e. photons that excite electrons to a higher energy level in an atom), hence the prefix photo-. Following excitation, various relaxation processes typically occur in which other photons are re-radiated. Time periods between absorption and emission may vary: ranging from short femtosecond-regime for emission involving free-carrier plasma in inorganic semiconductors up to milliseconds for phosphoresence processes in molecular systems; and under special circumstances delay of emission may even span to minutes or hours.
Observation of photoluminescence at a certain energy can be viewed as an indication that an electron populated an excited state associated with this transition energy.
While this is generally true in atoms and similar systems, correlations and other more complex phenomena also act as sources for photoluminescence in many-body systems such as semiconductors. A theoretical approach to handle this is given by the semiconductor luminescence equations.
Forms
Photoluminescence processes can be classified by various parameters such as the energy of the exciting photon with respect to the emission.
Resonant excitation describes a situation in which photons of a particular wavelength are absorbed and equivalent photons are very rapidly re-emitted. This is often referred to as resonance fluorescence. For materials in solution or in the gas phase, this process involves electrons but no significant internal energy transitions involving molecular features of the chemical substance between absorption and emission. In crystalline inorganic semiconductors where an electronic band structure is formed, secondary emission can be more complicated as events may contain both coherent contributions such as resonant Rayleigh scattering wher |
https://en.wikipedia.org/wiki/Force%20field%20%28technology%29 | In speculative fiction, a force field, sometimes known as an energy shield, force shield, energy bubble, or deflector shield, is a barrier produced by something like energy, negative energy, dark energy, electromagnetic fields, gravitational fields, electric fields, quantum fields, plasma, particles, radiation, solid light, or pure force. It protects a person, area, or object from attacks or intrusions, or even deflects energy attacks back at the attacker. This fictional technology is created as a field of energy without matter that acts as a wall, so that objects affected by the particular force relating to the field are unable to pass through the field and reach the other side, instead being deflected or destroyed. Actual research in the 21st century has looked into the potential to deflect radiation or cosmic rays, as well as more extensive shielding.
This concept has become a staple of many science-fiction works, so much so that authors frequently do not even bother to explain or justify them to their readers, treating them almost as established fact and attributing whatever capabilities the plot requires. The ability to create force fields has become a frequent superpower in superhero media.
History
The concept of a force field goes back at least as far as early 20th century. The Encyclopedia of Science Fiction suggests that the first use of the term in science fiction was in 1931, in Spacehounds of IPC by E.E. 'Doc' Smith.
An early precursor of what is now called "force field" may be found in William Hope Hodgson's The Night Land (1912), where the Last Redoubt, the fortress of the remnants of a far-future humanity, is kept safe by "The Air Clog" generated by the burning "Earth-Current".
In Isaac Asimov's Foundation universe, personal shields have been developed by scientists specializing in the miniaturization of planet-based shields. As they are primarily used by Foundation Traders, most other inhabitants of the Galactic Empire do not know about this tec |
https://en.wikipedia.org/wiki/Mayer%27s%20relation | In the 19th century, German chemist and physicist Julius von Mayer derived a relation between specific heat at constant pressure and the specific heat at constant volume for an ideal gas. Mayer's relation states that
where is the molar specific heat at constant pressure, is the molar specific heat at constant volume and is the gas constant.
For more general homogeneous substances, not just ideal gases, the difference takes the form,
(see relations between heat capacities), where is the molar volume, is the temperature, is the thermal expansion coefficient and is the isothermal compressibility.
From this latter relation, several inferences can be made:
Since the isothermal compressibility is positive for nearly all phases, and the square of thermal expansion coefficient is always either a positive quantity or zero, the specific heat at constant pressure is nearly always greater than or equal to specific heat at constant volume: There are no known exceptions to this principle for gases or liquids, but certain solids are known to exhibit negative compressibilities and presumably these would be (unusual) cases where .
For incompressible substances, and are identical. Also for substances that are nearly incompressible, such as solids and liquids, the difference between the two specific heats is negligible.
As the absolute temperature of the system approaches zero, since both heat capacities must generally approach zero in accordance with the Third Law of Thermodynamics, the difference between and also approaches zero. Exceptions to this rule might be found in systems exhibiting residual entropy due to disorder within the crystal. |
https://en.wikipedia.org/wiki/Untranslated%20region | In molecular genetics, an untranslated region (or UTR) refers to either of two sections, one on each side of a coding sequence on a strand of mRNA. If it is found on the 5' side, it is called the 5' UTR (or leader sequence), or if it is found on the 3' side, it is called the 3' UTR (or trailer sequence). mRNA is RNA that carries information from DNA to the ribosome, the site of protein synthesis (translation) within a cell. The mRNA is initially transcribed from the corresponding DNA sequence and then translated into protein. However, several regions of the mRNA are usually not translated into protein, including the 5' and 3' UTRs.
Although they are called untranslated regions, and do not form the protein-coding region of the gene, uORFs located within the 5' UTR can be translated into peptides.
The 5' UTR is upstream from the coding sequence. Within the 5' UTR is a sequence that is recognized by the ribosome which allows the ribosome to bind and initiate translation. The mechanism of translation initiation differs in prokaryotes and eukaryotes. The 3' UTR is found immediately following the translation stop codon. The 3' UTR plays a critical role in translation termination as well as post-transcriptional modification.
These often long sequences were once thought to be useless or junk mRNA that has simply accumulated over evolutionary time. However, it is now known that the untranslated region of mRNA is involved in many regulatory aspects of gene expression in eukaryotic organisms. The importance of these non-coding regions is supported by evolutionary reasoning, as natural selection would have otherwise eliminated this unusable RNA.
It is important to distinguish the 5' and 3' UTRs from other non-protein-coding RNA. Within the coding sequence of pre-mRNA, there can be found sections of RNA that will not be included in the protein product. These sections of RNA are called introns. The RNA that results from RNA splicing is a sequence of exons. The reason why intr |
https://en.wikipedia.org/wiki/Cemetech | Cemetech is a programming and hardware development group and developer community founded in 2000. Its primary software focus is calculator programming for TI and Casio graphing calculators, and its primary hardware focus is on mobile and wearable computing hardware. Among its most notable projects are the Doors CS shell for the TI-83+ series of graphing calculators, the Clove 2 dataglove, the Ultimate Calculator, and the CALCnet / globalCALCnet system for networking graphing calculators and connecting them to the Internet. The Cemetech website hosts tools for calculator programmers, including the SourceCoder TI-BASIC IDE and the jsTIfied TI-83+/84+ emulator. The founder of the site, Dr. Christopher Mitchell ("Kerm Martian"), began the site to showcase his personal projects, but since its early days, it has branched out to become one of the several major sites of the TI calculator hobbyist community and a source for hardware and programming development assistance. It has incubated many software and hardware projects beginning in the calculator community at its roots but including microprocessor development, general electrical engineering, desktop applications, and mobile/web applications.
History
Cemetech began as a personal website hosted on Homestead and later GeoCities, publishing personal software and hardware projects. In 2004 the site expanded on shared hosting with a PhpBB-based forum, and in March 2005 moved to Cemetech.net. The site spent the following three years consolidating its presence in the TI graphing calculator enthusiast community, attracting programmers who began publishing their own independent software projects on the site. Early projects were primarily calculator-related, later branching out into computer, web, and embedded programming. In mid-2006, Cemetech lost several hundred posts when hosting provider Jatol disappeared overnight, stranding hundreds of customers without websites or backups. From 2008, Cemetech expanded further i |
https://en.wikipedia.org/wiki/Windows%2098 | Windows 98 is a consumer-oriented operating system developed by Microsoft as part of its Windows 9x family of Microsoft Windows operating systems. The second operating system in the 9x line, it is the successor to Windows 95, and was released to manufacturing on May 15, 1998, and generally to retail on June 25, 1998. Like its predecessor, it is a hybrid 16-bit and 32-bit monolithic product with the boot stage based on MS-DOS.
Windows 98 is a web-integrated operating system that bears numerous similarities to its predecessor. Most of its improvements were cosmetic or designed to improve the user experience, but there were also a handful of features introduced to enhance system functionality and capabilities, including improved USB support and accessibility, as well as support for hardware advancements such as DVD players. Windows 98 was the first edition of Windows to adopt the Windows Driver Model, and introduced features that would become standard in future generations of Windows, such as Disk Cleanup, Windows Update, multi-monitor support, and Internet Connection Sharing.
Microsoft had marketed Windows 98 as a "tune-up" to Windows 95, rather than an entirely improved next generation of Windows. Upon release, it received a positive reception for its web-integrated interface and ease of use, as well as its addressing of issues present in Windows 95, although some pointed out that it was not significantly more stable than its predecessor. Windows 98 sold an estimated 58 million licenses and saw one major update, known as Windows 98 Second Edition (SE), released on May 5, 1999. After the release of its successor, Windows Me in 2000, mainstream support for Windows 98 and 98 SE ended on June 30, 2002, followed by extended support on July 11, 2006.
Development
Following the success of Windows 95, the development of Windows 98 began, initially under the development codename "Memphis." The first test version, Windows Memphis Developer Release, was released in January 19 |
https://en.wikipedia.org/wiki/Community-based%20clinical%20trial | Community-based clinical trials are clinical trials conducted directly through doctors and clinics rather than academic research facilities. They are designed to be administered through primary care physicians, community health centers and local outpatient facilities. In 1986, the Community Consortium held the first such trials in the United States to determine the efficiency of preventive treatments after the onset of Pneumocystis pneumonia. The trials give patients access to new medications and keep doctors involved with new developments in research. However, critics state that drug company payments to doctors for patients enrolled in such studies present a conflict of interest and potential for abuse. Community-based trials are becoming prevalent in human-testing stage pharmaceutical research.
See also
Clinical trial |
https://en.wikipedia.org/wiki/Windows%207 | Windows 7 is a major release of the Windows NT operating system developed by Microsoft. It was released to manufacturing on July 22, 2009, and became generally available on October 22, 2009. It is the successor to Windows Vista, released nearly three years earlier. Windows 7's server counterpart, Windows Server 2008 R2, was released at the same time. Windows 7 remained an operating system for use on personal computers, including home and business desktops, laptops, tablet PCs and media center PCs, and itself was replaced in November 2012 by Windows 8, the name spanning more than three years of the product.
Extended support ended on January 14, 2020, over ten years after the release of Windows 7, after which the operating system ceased receiving further updates. A paid support program was available for enterprises, providing security updates for Windows 7 for up to three years since the official end of life.
Windows 7 was intended to be an incremental upgrade to Microsoft Windows, addressing Windows Vista's poor critical reception while maintaining hardware and software compatibility. Windows 7 continued improvements on the Windows Aero user interface with the addition of a redesigned taskbar that allows pinned applications, and new window management features. Other new features were added to the operating system, including libraries, the new file-sharing system HomeGroup, and support for multitouch input. A new "Action Center" was also added to provide an overview of system security and maintenance information, and tweaks were made to the User Account Control system to make it less intrusive. Windows 7 also shipped with updated versions of several stock applications, including Internet Explorer 8, Windows Media Player, and Windows Media Center.
Unlike Windows Vista, Windows 7 received critical acclaim, with critics considering the operating system to be a major improvement over its predecessor because of its improved performance, its more intuitive interface, few |
https://en.wikipedia.org/wiki/IOIO | IOIO (pronounced yo-yo) is a series of open source PIC microcontroller-based boards that allow Android mobile applications to interact with external electronics. The device was invented by Ytai Ben-Tsvi in 2011, and was first manufactured by SparkFun Electronics. The name "IOIO" is inspired by the function of the device, which enables applications to receive external input ("I") and produce external output ("O").
Features
The IOIO board contains a single PIC MCU that acts as a USB host/USB slave and communicates with an Android app running on a connected Android device. The board provides connectivity via USB, USB-OTG or Bluetooth, and is controllable from within an Android application using the Java API.
In addition to basic digital input/output and analog input, the IOIO library also handles PWM, I2C, SPI, UART, Input capture, Capacitive sensing and advanced motor control. To connect to older Android devices that use USB 2.0 in slave mode, newer IOIO models use USB On-The-Go to act as a host for such devices. Some models also support the Google Open Accessory USB protocol.
The IOIO motor control API can drive up to 9 motors and any number of binary actuators in synchronization and cycle-accurate precision. Developers may send a sequence of high-level commands to the IOIO, which performs the low-level waveform generation on-chip. The IOIO firmware supports 3 different kinds of motors; stepper motors, DC motors and servo motors.
Device firmware may be updated on-site by the user. For first-generation devices updating is performed using an Android device and the IOIO Manager application available on Google Play. Second-generation IOIO-OTG devices must be updated using a desktop computer running the IOIODude application.
The IOIO supports both computers and Android devices as first-class hosts, and provides the exact API on both types of devices. First-generation devices can only communicate with PCs over Bluetooth, while IOIO-OTG devices can use either Bluetooth |
https://en.wikipedia.org/wiki/Apollo%20PRISM | PRISM (Parallel Reduced Instruction Set Multiprocessor) was Apollo Computer's high-performance CPU used in their DN10000 series workstations. It was for some time the fastest microprocessor available, a high fraction of a Cray-1 in a workstation. Hewlett-Packard purchased Apollo in 1989, ending development of PRISM, although some of PRISM's ideas were later used in HP's own HP-PA Reduced instruction set computer (RISC) and Itanium processors.
PRISM was based on what would be known today as a VLIW-design, while most efforts of the era, 1988, were based on a more "pure" RISC approach. In early RISC designs, the core processor was simplified as much as possible in order to allow more of the chip's real-estate to be used for registers and simplifying the addition of instruction pipelines for improved performance.
Compilers
The compilers used with the systems were expected to dedicate more time during compilation to making effective use of the registers and cleaning the instruction stream. By doing instruction scheduling in the compiler, this design avoided the problems and complexity of dynamic instruction scheduling (where instructions for multiple functional units must be selected carefully in order to avoid interdependencies between intermediate values) encountered in superscalar designs such as Digital Equipment Corporation's Alpha.
In some respects, the VLIW design can be thought of as "super-RISCy", as it offloads the instruction selection process to the compiler as well. In the VLIW design, the compiler examines the code and selects instructions that are known to be "safe", and then packages them into longer instruction words. For instance, for a CPU with two functional units, like the PRISM, the compiler would find pairs of safe instructions and stuff them into a single larger word. Inside the CPU, the instructions are simply split apart again, and fed into the selected units.
This design minimizes logical changes to the CPU as functional units are added, |
https://en.wikipedia.org/wiki/Polynomial%20mapping | In algebra, a polynomial map or polynomial mapping between vector spaces over an infinite field k is a polynomial in linear functionals with coefficients in k; i.e., it can be written as
where the are linear functionals and the are vectors in W. For example, if , then a polynomial mapping can be expressed as where the are (scalar-valued) polynomial functions on V. (The abstract definition has an advantage that the map is manifestly free of a choice of basis.)
When V, W are finite-dimensional vector spaces and are viewed as algebraic varieties, then a polynomial mapping is precisely a morphism of algebraic varieties.
One fundamental outstanding question regarding polynomial mappings is the Jacobian conjecture, which concerns the sufficiency of a polynomial mapping to be invertible.
See also
Polynomial functor |
https://en.wikipedia.org/wiki/Initial%20coin%20offering | An initial coin offering (ICO) or initial currency offering is a type of funding using cryptocurrencies. It is often a form of crowdfunding, although a private ICO which does not seek public investment is also possible. In an ICO, a quantity of cryptocurrency is sold in the form of "tokens" ("coins") to speculators or investors, in exchange for legal tender or other (generally established and more stable) cryptocurrencies such as Bitcoin or Ether. The tokens are promoted as future functional units of currency if or when the ICO's funding goal is met and the project successfully launches.
An ICO can be a source of capital for startup companies. ICOs can allow startups to avoid regulations that prevent them from seeking investment directly from the public, and intermediaries such as venture capitalists, banks, and stock exchanges, which may demand greater scrutiny and some percentage of future profits or joint ownership. ICOs may fall outside existing regulations, depending on the nature of the project, or be banned altogether in some jurisdictions, such as China and South Korea.
Due to the lack of regulation and enforcement of securities law, ICOs have been the vehicle for scams and fraud. Fewer than half of all ICOs survive four months after the offering, while almost half of ICOs sold in 2017 failed by February 2018. Despite their record of failure and the falling prices of cryptocurrencies, a record $7 billion was raised via ICO from January–June 2018.
History
The first token sale (also known as an ICO) was held by Mastercoin in July 2013. Ethereum raised money with a token sale in 2014, raising around 31,000 BTC in July, equal to approximately $18.3 million at the time.
ICOs and token sales became popular in 2017. There were at least 18 websites tracking ICOs before mid-year. In May, the ICO for a new web browser called Brave generated about $35 million in under 30 seconds. Messaging app developer Kik's September 2017 ICO raised nearly $100 million. At the s |
https://en.wikipedia.org/wiki/How%20to%20Create%20a%20Mind | How to Create a Mind: The Secret of Human Thought Revealed is a non-fiction book about brains, both human and artificial, by the inventor and futurist Ray Kurzweil. First published in hardcover on November 13, 2012 by Viking Press it became a New York Times Best Seller. It has received attention from The Washington Post, The New York Times and The New Yorker.
Kurzweil describes a series of thought experiments which suggest to him that the brain contains a hierarchy of pattern recognizers. Based on this he introduces his Pattern Recognition Theory of Mind (PRTM). He says the neocortex contains 300 million very general pattern recognition circuits and argues that they are responsible for most aspects of human thought. He also suggests that the brain is a "recursive probabilistic fractal" whose line of code is represented within the 30-100 million bytes of compressed code in the genome.
Kurzweil then explains that a computer version of this design could be used to create an artificial intelligence more capable than the human brain. It would employ techniques such as hidden Markov models and genetic algorithms, strategies Kurzweil used successfully in his years as a commercial developer of speech recognition software. Artificial brains will require massive computational power, so Kurzweil reviews his law of accelerating returns, which explains how the compounding effects of exponential growth will deliver the necessary hardware in only a few decades.
Critics felt the subtitle of the book, The Secret of Human Thought Revealed, overpromises. Some protested that pattern recognition does not explain the "depth and nuance" of mind including elements like emotion and imagination. Others felt Kurzweil's ideas might be right, but they are not original, pointing to existing work as far back as the 1980s. Yet critics admire Kurzweil's "impressive track record" and say that his writing is "refreshingly clear", containing "lucid discussions" of computing history.
Background
K |
https://en.wikipedia.org/wiki/Journal%20of%20Cellular%20Biochemistry | The Journal of Cellular Biochemistry publishes descriptions of original research in which complex cellular, pathogenic, clinical, or animal model systems are studied by biochemical, molecular, genetic, epigenetic, or quantitative ultrastructural approaches.
History
The journal was previously called the Journal of Supramolecular Structure (1972–1980) and the Journal of Supramolecular Structure and Cellular Biochemistry (1981).
Abstracting and indexing
The Journal of Cellular Biochemistry is indexed and/or abstracted in the following databases: BIOBASE, Biochemistry and Biophysics Citation Index, Biological Abstracts, BIOSIS Previews, CAB Abstracts, Cambridge Scientific Abstracts, Chemical Abstracts Service/SciFinder, CSA Biological Sciences Database, Current Awareness in Biological Sciences, Current Contents/Life Sciences, EMBASE, EORTC Database, Index Medicus/MEDLINE/PubMed, Reference Update, Science Citation Index, and Scopus. |
https://en.wikipedia.org/wiki/Feedback-controlled%20electromigration | Feedback-controlled electromigration (FCE) is an experimental technique to investigate the phenomenon known as electromigration. By controlling the voltage applied as the conductance varies it is possible to keep the voltage at a critical level for electromigration.
Theory
FCE has been shown to be reversible, demonstrating the fact that the electrons are moving rather than thermomigration or sublimation. The migration occurs due to the electronic wind force experienced by the metallic adatom. The electromigration occurs at a critical power dissipation in the neck of the bridge. This leads to Electromigrated Nanogaps.
Uses
FCE is often used in forming nanogaps in metallic bridges.
Problems
Thermal runaway can occur when the neck is narrower than about 20 nm. |
https://en.wikipedia.org/wiki/Michael%20Spivak | Michael David Spivak (May 25, 1940October 1, 2020) was an American mathematician specializing in differential geometry, an expositor of mathematics, and the founder of Publish-or-Perish Press. Spivak was the author of the five-volume A Comprehensive Introduction to Differential Geometry.
Biography
Spivak was born in Queens, New York. He received his Bachelor of Arts (A.B.) from Harvard University in 1960, and in 1964 he received his Ph.D. from Princeton University under the supervision of John Milnor, with thesis On Spaces Satisfying Poincaré Duality. In 1985, Spivak received the Leroy P. Steele Prize.
Spivak lectured on elementary physics. Spivak's book, Physics for Mathematicians: Mechanics I (published December 6, 2010), contains the material that these lectures stemmed from and more. Spivak was also the designer of the MathTime Professional 2 fonts (which are widely used in academic publishing) and the creator of Science International.
Writing
His five-volume A Comprehensive Introduction to Differential Geometry (Publish or Perish Inc., 1970; 2nd ed., 1979; 3rd ed., 1999, revised 2005) is among his most influential and celebrated works. The distinctive pedagogical aim of the work, as stated in its preface, was to elucidate for graduate students the often obscure relationship between classical differential geometry—geometrically intuitive but imprecise—and its modern counterpart, replete with precise but unintuitive algebraic definitions. On several occasions, most prominently in Volume 2, Spivak "translates" the classical language that Gauss or Riemann would be familiar with to the abstract language that a modern differential geometer might use. The Leroy P. Steele Prize was awarded to Spivak in 1985 for his authorship of the work.
Spivak also authored several well-known undergraduate textbooks. Among them, his textbook Calculus (W. A. Benjamin Inc., 1967; Publish or Perish, 4th ed., 2008) takes a rigorous and theoretical approach to introductory calculu |
https://en.wikipedia.org/wiki/Cucurbitacin | Cucurbitacins are a class of biochemical compounds that some plants – notably members of the pumpkin and gourd family, Cucurbitaceae – produce and which function as a defense against herbivores. Cucurbitacins and their derivatives have also been found in many other plant families (including Brassicaceae, Cucurbitaceae, Scrophulariaceae, Begoniaceae, Elaeocarpaceae, Datiscaceae, Desfontainiaceae, Polemoniaceae, Primulaceae, Rubiaceae, Sterculiaceae, Rosaceae, and Thymelaeaceae), in some mushrooms (including Russula and Hebeloma) and even in some marine mollusks.
Cucurbitacins may be a taste deterrent in plants foraged by some animals and in some edible plants preferred by humans, such as cucumbers and zucchinis. In laboratory research, cucurbitacins have cytotoxic properties and are under study for their potential biological activity.
Cucurbitacins are chemically classified as triterpenes, formally derived from cucurbitane, a triterpene hydrocarbon – specifically, from the unsaturated variant cucurbit-5-ene, or 19(10→9β)-abeo-10α-lanost-5-ene. They often occur as glycosides. Most cucurbitacins are tetracyclic except some have an extra ring due to formal cyclization between C16 and C24 as in cucurbitacin S and cucurbitacin T.
Biosynthesis
The biosynthesis of cucurbitacin C has been described. Zhang et al. (2014) identified nine cucumber genes in the pathway for biosynthesis of cucurbitacin C and elucidated four catalytic steps. These authors also discovered the transcription factors Bl (Bitter leaf) and Bt (Bitter fruit) that regulate this pathway in leaves and fruits, respectively. The Bi gene confers bitterness to the entire plant and is genetically associated with an operon-like gene cluster, similar to the gene cluster involved in thalianol biosynthesis in Arabidopsis. Fruit bitterness requires both Bi and the dominant Bt (Bitter fruit) gene. Nonbitterness of cultivated cucumber fruit is conferred by bt, an allele selected during domestication. Bi is a memb |
https://en.wikipedia.org/wiki/Paracrine%20regulator | A paracrine regulator is a molecule or hormone produced by a tissue to regulate activity in that same tissue. Paracrine regulators are distinct from endocrine regulators, which secrete substances directly into the blood stream, thus accessing other tissues as well. Some paracrine regulators can also be autocrine regulators, which are produced by cells to induce changes within themselves. |
https://en.wikipedia.org/wiki/Pinger | Pinger, Inc. is a US Telecom provider for free texts, pictures, calls, and voicemails. Pinger was founded in 2005 by former Palm, Inc. managers Greg Woock (CEO of Pinger, Inc) and Joe Sipher. The company was founded in 2006 and is headquartered in San Jose, California.
Products
The company has released several apps, mostly being available on iOS, Android, Microsoft Windows and Macintosh.
Textfree
Textfree is an app that allows user to call and text using a real phone number for free. The app has a free ad-supported version and a paid version. The app allows you to text anyone in the world.
Sideline
Sideline is an app that adds a second phone number to your phone using technology similar to Textfree. It's available for iOS, Android, Microsoft Windows and Macintosh. Their official website is sideline.com. |
https://en.wikipedia.org/wiki/Bachelor%20herd | A bachelor herd is a herd of (usually) juvenile male animals who are still sexually immature or 'harem'-forming animals who have been thrown out of their parent groups but not yet formed a new family group. It may also refer to a group of males who are not currently territorial or mating with females.
Examples include seals, dolphins, lions, and many herbivores such as deer, horses, and elephants. Bachelor herds are thought to provide useful protection for social animals against more established herd competition or aggressive, dominant males. Males in bachelor herds are sometimes closely related to each other. Some animals, for example New Zealand fur seals, live in a bachelor herd all year except for the mating season, when there is a substantial increase in aggression and competition.
In many species, males and females move in separate groups, often coming together at mating time, or to fight for territory or mating partners. In many species it is common for males to leave or be driven from the group as they mature, and they may wander as lone animals or form a bachelor group for the time being. This arrangement may be long term and stable, or short term until they find a new group to join.
Types
The social structure, aggression level, population size, and duration of presence of these herds across species varies greatly. Bachelor herds are most often found in mammals and are especially common in the grasslands.
Impala
Male impala form small bachelor herds during both the wet and dry seasons. These bachelor herds are generally smaller than herds of females, numbering around 4 members, compared to upwards of 10. Juvenile males begin to join bachelor herds at 8 months of age. In the Serengeti, immature or older males will usually form their own bachelor herds, while males of reproductive age are more often in mixed groups with females.
Being actively territorial in the Serengeti is physically demanding for male impala, so males occupy this role for about 3 mo |
https://en.wikipedia.org/wiki/Diagnosis%20%28artificial%20intelligence%29 | As a subfield in artificial intelligence, diagnosis is concerned with the development of algorithms and techniques that are able to determine whether the behaviour of a system is correct. If the system is not functioning correctly, the algorithm should be able to determine, as accurately as possible, which part of the system is failing, and which kind of fault it is facing. The computation is based on observations, which provide information on the current behaviour.
The expression diagnosis also refers to the answer of the question of whether the system is malfunctioning or not, and to the process of computing the answer. This word comes from the medical context where a diagnosis is the process of identifying a disease by its symptoms.
Example
An example of diagnosis is the process of a garage mechanic with an automobile. The mechanic will first try to detect any abnormal behavior based on the observations on the car and his knowledge of this type of vehicle. If he finds out that the behavior is abnormal, the mechanic will try to refine his diagnosis by using new observations and possibly testing the system, until he discovers the faulty component; the mechanic plays an important role in the vehicle diagnosis.
Expert diagnosis
The expert diagnosis (or diagnosis by expert system) is based on experience with the system. Using this experience, a mapping is built that efficiently associates the observations to the corresponding diagnoses.
The experience can be provided:
By a human operator. In this case, the human knowledge must be translated into a computer language.
By examples of the system behaviour. In this case, the examples must be classified as correct or faulty (and, in the latter case, by the type of fault). Machine learning methods are then used to generalize from the examples.
The main drawbacks of these methods are:
The difficulty acquiring the expertise. The expertise is typically only available after a long period of use of the system |
https://en.wikipedia.org/wiki/Lipoprotein-associated%20phospholipase%20A2 | Lipoprotein-associated phospholipase A2 (Lp-PLA2) also known as platelet-activating factor acetylhydrolase (PAF-AH) is a phospholipase A2 enzyme that in humans is encoded by the PLA2G7 gene. Lp-PLA2 is a 45-kDa protein of 441 amino acids. It is one of several PAF acetylhydrolases.
Function
In the blood Lp-PLA2 travels mainly with low-density lipoprotein (LDL). Less than 20% is associated with high-density lipoprotein HDL. Several lines of evidence suggest that HDL-associated Lp-PLA2 may substantially contribute to the HDL antiatherogenic activities. It is an enzyme produced by inflammatory cells and hydrolyzes oxidized phospholipids in LDL.
Lp-PLA2 is platelet-activating factor (PAF) acetylhydrolase (EC 3.1.1.47), a secreted enzyme that catalyzes the degradation of PAF to inactive products by hydrolysis of the acetyl group at the sn-2 position, producing the biologically inactive products LYSO-PAF and acetate.
Clinical significance
Lp-PLA2 is involved in the development of atherosclerosis, an observation that has prompted interest as a possible therapeutic target (see, e.g. the investigational drug Darapladib). In human atherosclerotic lesions, 2 main sources of Lp-PLA2 can be identified, including that which is brought into the intima bound to LDL (from the circulation), and that which is synthesized de novo by plaque inflammatory cells (macrophages, T cells, mast cells)."
It is used as a marker for cardiac disease.
A meta-analysis involving a total of 79,036 participants in 32 prospective studies found that Lp-PLA2 levels are positively correlated with increased risk of developing coronary heart disease and stroke.
See also
Darapladib |
https://en.wikipedia.org/wiki/Behavior%20tree | Behavior trees are a formal, graphical modelling language used primarily in systems and software engineering. Behavior trees employ a well-defined notation to unambiguously represent the hundreds or even thousands of natural language requirements that are typically used to express the stakeholder needs for a large-scale software-integrated system.
Overview
The amount of detail in the large number of natural language requirements for a large-scale system causes short-term memory overload and may create a barrier that prevents anyone from gaining a deep, accurate and holistic understanding of the system needs. Also, because of the use of natural language, there are likely to be many ambiguities, aliases, inconsistencies, redundancies and incompleteness problems associated with the requirements information. This adds further to the uncertainty and complexity. Generally, at best, a few people understand parts of the system or situation well, but no one has other than a superficial understanding of the whole – that is, the detailed integrated behavior of the system.
The behavior tree representation, (with the help of the composition tree representation that resolves alias and other vocabulary problems with large sets of requirements) allows people to avoid short-term memory overload and produce a deep, accurate, holistic representation of system needs that can be understood by all stakeholders because it strictly uses the vocabulary of the original requirements. Because the behavior tree notation uses a formal semantics, for any given example, it already is, or can be made executable.
Behavior tree forms
Single and composite or integrated behavior tree forms are both important in the application of behavior trees in systems and software engineering.
Requirement behavior trees: Initially, individual requirement behavior trees (RBTs) are used to capture all the fragments of behavior in each individual natural language requirement by a process of rigorous, intent-p |
https://en.wikipedia.org/wiki/AEGIS%20SecureConnect | AEGIS SecureConnect (or simply AEGIS) is the former name of a network authentication system used in IEEE 802.1X networks. It was developed by Meetinghouse Data Communications, Inc.; the system was renamed "Cisco Secure Services Client" when Meetinghouse was acquired by Cisco Systems. The AEGIS Protocol is an 802.1X supplicant (i.e. handles authentication for wired and wireless networks, such as those that use WPA-PSK, WPA-Radius, or Certificate-based authentication), and is commonly installed along with a Network Interface Card's (NIC) or VPN drivers. |
https://en.wikipedia.org/wiki/Smart%20Technologies | SMART Technologies (also known as "SMART") is a Canadian company headquartered in Calgary, Alberta, Canada and wholly owned by Foxconn. Founded in 1987, SMART is best known as the developer of interactive whiteboards branded as the "SMART Board" popularly used in education and business.
History
SMART was founded in 1987 by husband and wife David Martin and Nancy Knowlton.
In 1991, SMART introduced its interactive whiteboard, branded the "SMART Board".
In 2003, SMART developed and later patented DViT (Digital Vision Touch) technology which was an important feature of the SMART Board.
In 2009, SMART filed a patent infringement lawsuit against New Zealand-based NextWindow before acquiring the company outright in 2010.
On July 15, 2010 SMART Technologies placed an initial public offering on the NASDAQ stock exchange and Toronto Stock Exchange (TSX), raising upwards of $660 million. Later that year, several class action complaints were filed against SMART in the U.S. District Courts in New York and Illinois. Filed on behalf of the purchasers of the Class A Subordinate Voting Shares in Smart's IPO in July 2010, the complaints alleged that the company did not disclose important information prior to the IPO. A settlement was reached in March 2013.
As of the 2011 fiscal year end, SMART Board interactive whiteboards led the interactive whiteboard category with a 63% share in the United States, 44% share in EMEA (Europe, Middle East and Africa) and 47% share globally.
In 2011, SMART employed about 1,700 people.
In April 2012, amid a declining stock price and sales resulting from decreased school board spending, both Martin and Knowlton resigned from their executive roles at Smart before leaving the board in April 2014.
On December 11, 2012, SMART Technologies announced it would begin corporate restructuring. The company proceeded to lay off 25% of its employees. Then-CEO Neil Gaydon stated that it would save the company approximately $40 million, and cited competition |
https://en.wikipedia.org/wiki/Tryptophan-rich%20sensory%20protein | Tryptophan-rich sensory proteins (TspO) are a family of proteins that are involved in transmembrane signalling. In either prokaryotes or mitochondria they are localized to the outer membrane, and have been shown to bind and transport dicarboxylic tetrapyrrole intermediates of the haem biosynthetic pathway. They are associated with the major outer membrane porins (in prokaryotes) and with the voltage-dependent anion channel (in mitochondria).
TspO of Rhodobacter sphaeroides is involved in signal transduction, functioning as a negative regulator of the expression of some photosynthesis genes (PpsR/AppA repressor/antirepressor regulon). This down-regulation is believed to be in response to oxygen levels. TspO works through (or modulates) the PpsR/AppA system and acts upstream of the site of action of these regulatory proteins. It has been suggested that the TspO regulatory pathway works by regulating the efflux of certain tetrapyrrole intermediates of the haem/bacteriochlorophyll biosynthetic pathways in response to the availability of molecular oxygen, thereby causing the accumulation of a biosynthetic intermediate that serves as a corepressor for the regulated genes. A homologue of the TspO protein in Sinorhizobium meliloti is involved in regulating expression of the ndi locus in response to stress conditions.
In animals, the peripheral benzodiazepine receptor is a mitochondrial protein (located in the outer mitochondrial membrane) characterised by its ability to bind with nanomolar affinity to a variety of benzodiazepine-like drugs, as well as to dicarboxylic tetrapyrrole intermediates of the haem biosynthetic pathway. Depending upon the tissue, it was shown to be involved in steroidogenesis, haem biosynthesis, apoptosis, cell growth and differentiation, mitochondrial respiratory control, and immune and stress response, but the precise function of the PBR remains unclear. The role of PBR in the regulation of cholesterol transport from the outer to the inner mitoch |
https://en.wikipedia.org/wiki/Insertional%20mutagenesis | In molecular biology, insertional mutagenesis is the creation of mutations in DNA by the addition of one or more base pairs. Such insertional mutations can occur naturally, mediated by viruses or transposons, or can be artificially created for research purposes in the lab.
Signature tagged mutagenesis
This is a technique used to study the function of genes. A transposon such as the P element of Drosophila melanogaster is allowed to integrate at random locations in the genome of the organism being studied. Mutants generated by this method are then screened for any unusual phenotypes. If such a phenotype is found then it can be assumed that the insertion has caused the gene relating to the usual phenotype to be inactivated. Because the sequence of the transposon is known, the gene can be identified, either by sequencing the whole genome and searching for the sequence, or by using the polymerase chain reaction to amplify specifically that gene.
Virus insertional mutagenesis
Because many viruses integrate their own genomes into the genomes of their host cells in order to replicate, mutagenesis caused by viral infections is a fairly common occurrence. Not all integrating viruses cause insertional mutagenesis, however.
Some DNA insertions will lead to no noticeable mutation. Historically, lentiviral vectors included strong viral promoters which had a side effect of insertional mutagenesis, nuclear DNA mutations that effect the function of a gene. These strong viral promotors were shown to be the main cause of cancer formation. As a result, viral promotors have been replaced by cellular promotors and regulatory sequences. For those viruses such as gammaretroviruses that tend to integrate their DNA in genetically unfavorable locations, the severity of any ensuing mutation depends entirely on the location within the host's genome wherein the viral DNA is inserted. If the DNA is inserted into the middle of an essential gene, the effects on the cell will be drastic. Ad |
https://en.wikipedia.org/wiki/Desulfovibrio | Desulfovibrio is a genus of Gram-negative sulfate-reducing bacteria. Desulfovibrio species are commonly found in aquatic environments with high levels of organic material, as well as in water-logged soils, and form major community members of extreme oligotrophic habitats such as deep granitic fractured rock aquifers.
High amounts of Desulfovibrio bacteria have been associated with inflammatory bowel disease, bacteremia infections and Parkinson's disease.
Some Desulfovibrio species have in recent years been shown to have bioremediation potential for toxic radionuclides such as uranium by a reductive bioaccumulation process, such as converting highly water-soluble U(VI) to relatively insoluble U(IV) precipitate, thus removing the toxic uranium from contaminated water.
Phylogeny
The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI)
Unassigned species:
"D. caledoniensis" Tardy-Jacquenod et al. 1996
"D. cavernae" Sass & Cypionka 2004
"D. diazotrophica" Sayavedra et al. 2021
"D. halohydrocarbonoclasticus" Zobell 1947
"D. hontreensis" Tarasov et al. 2015
"D. lacusfryxellense" Sattley & Madigan 2010
"D. multispirans" Czechowski et al. 1984
"D. oliviopondense" Qatabi Sr. 2007
"D. oryzae" Ouattara et al. 2000b
"D. rubentschikii" (Baars 1930) ZoBell 1948
"D. singaporenus" Sheng et al. 2007
See also
List of bacterial orders
List of bacteria genera |
https://en.wikipedia.org/wiki/A.%20J.%20Ayer | Sir Alfred Jules "Freddie" Ayer ( ; 29 October 1910 – 27 June 1989), usually cited as A. J. Ayer, was an English philosopher known for his promotion of logical positivism, particularly in his books Language, Truth, and Logic (1936) and The Problem of Knowledge (1956).
Ayer was educated at Eton College and the University of Oxford, after which he studied the philosophy of logical positivism at the University of Vienna. From 1933 to 1940 he lectured on philosophy at Christ Church, Oxford.
During the Second World War Ayer was a Special Operations Executive and MI6 agent.
Ayer was Grote Professor of the Philosophy of Mind and Logic at University College London from 1946 until 1959, after which he returned to Oxford to become Wykeham Professor of Logic at New College. He was president of the Aristotelian Society from 1951 to 1952 and knighted in 1970. He was known for his advocacy of humanism, and was the second president of the British Humanist Association (now known as Humanists UK).
Ayer was president of the Homosexual Law Reform Society for a time; he remarked, "as a notorious heterosexual I could never be accused of feathering my own nest."
Life
Ayer was born in St John's Wood, in north west London, to Jules Louis Cyprien Ayer and Reine (née Citroen), wealthy parents from continental Europe. His mother was from the Dutch-Jewish family that founded the Citroën car company in France; his father was a Swiss Calvinist financier who worked for the Rothschild family, including for their bank and as secretary to Alfred Rothschild.
Ayer was educated at Ascham St Vincent's School, a former boarding preparatory school for boys in the seaside town of Eastbourne in Sussex, where he started boarding at the relatively early age of seven for reasons to do with the First World War, and at Eton College, where he was a King's Scholar. At Eton Ayer first became known for his characteristic bravado and precocity. Though primarily interested in his intellectual pursuits, he was |
https://en.wikipedia.org/wiki/Chi-square%20automatic%20interaction%20detection | Chi-square automatic interaction detection (CHAID) is a decision tree technique based on adjusted significance testing (Bonferroni correction, Holm-Bonferroni testing). The technique was developed in South Africa and was published in 1980 by Gordon V. Kass, who had completed a PhD thesis on this topic. CHAID can be used for prediction (in a similar fashion to regression analysis, this version of CHAID being originally known as XAID) as well as classification, and for detection of interaction between variables. CHAID is based on a formal extension of AID (Automatic Interaction Detection) and THAID (THeta Automatic Interaction Detection) procedures of the 1960s and 1970s, which in turn were extensions of earlier research, including that performed by Belson in the UK in the 1950s. A history of earlier supervised tree methods together with a detailed description of the original CHAID algorithm and the exhaustive CHAID extension by Biggs, De Ville, and Suen, can be found in Ritschard.
In practice, CHAID is often used in the context of direct marketing to select groups of consumers to predict how their responses to some variables affect other variables, although other early applications were in the fields of medical and psychiatric research.
Like other decision trees, CHAID's advantages are that its output is highly visual and easy to interpret. Because it uses multiway splits by default, it needs rather large sample sizes to work effectively, since with small sample sizes the respondent groups can quickly become too small for reliable analysis.
One important advantage of CHAID over alternatives such as multiple regression is that it is non-parametric.
See also
Chi-squared distribution
Bonferroni correction
Latent class model
Structural equation modeling
Market segment
Decision tree learning
Multiple comparisons |
https://en.wikipedia.org/wiki/Pressure%20flow%20hypothesis | The pressure flow hypothesis, also known as the mass flow hypothesis, is the best-supported theory to explain the movement of sap through the phloem. It was proposed by Ernst Münch, a German plant physiologist in 1930.
A high concentration of organic substances, particularly sugar, inside cells of the phloem at a source, such as a leaf, creates a diffusion gradient (osmotic gradient) that draws water into the cells from the adjacent xylem. This creates turgor pressure, also known as hydrostatic pressure, in the phloem. Movement of phloem sap occurs by bulk flow (mass flow) from sugar sources to sugar sinks. The movement in phloem is bidirectional, whereas, in xylem cells, it is unidirectional (upward). Because of this multi-directional flow, coupled with the fact that sap cannot move with ease between adjacent sieve-tubes, it is not unusual for sap in adjacent sieve-tubes to be flowing in opposite directions.
Sources and sinks
A sugar source is any part of the plant that is producing or releasing sugar.
During the plant's growth period, usually during the spring, storage organs such as the roots are sugar sources, and the plant's many growing areas are sugar sinks.
After the growth period, when the meristems are dormant, the leaves are sources, and storage organs are sinks. Developing seed-bearing organs (such as fruit) are always sinks.
Mechanisms
While movement of water and minerals through the xylem is driven by negative pressures (tension) most of the time, movement through the phloem is driven by positive hydrostatic pressure. This process is termed translocation, and is accomplished by a process called phloem loading and unloading. Cells in a sugar source "load" a sieve-tube element by actively transporting solute molecules into it. This causes water to move into the sieve-tube element by osmosis, creating pressure that pushes the sap down the tube. In sugar sinks, cells actively transport solutes out of the sieve-tube elements, producing the exactly oppos |
https://en.wikipedia.org/wiki/Exception%20handling | In computing and computer programming, exception handling is the process of responding to the occurrence of exceptions – anomalous or exceptional conditions requiring special processing – during the execution of a program. In general, an exception breaks the normal flow of execution and executes a pre-registered exception handler; the details of how this is done depend on whether it is a hardware or software exception and how the software exception is implemented.
Exception handling, if provided, is facilitated by specialized programming language constructs, hardware mechanisms like interrupts, or operating system (OS) inter-process communication (IPC) facilities like signals. Some exceptions, especially hardware ones, may be handled so gracefully that execution can resume where it was interrupted.
Definition
The definition of an exception is based on the observation that each procedure has a precondition, a set of circumstances for which it will terminate "normally". An exception handling mechanism allows the procedure to raise an exception if this precondition is violated, for example if the procedure has been called on an abnormal set of arguments.
The exception handling mechanism then handles the exception.
The precondition, and the definition of exception, is subjective. The set of "normal" circumstances is defined entirely by the programmer, e.g. the programmer may deem division by zero to be undefined, hence an exception, or devise some behavior such as returning zero or a special "ZERO DIVIDE" value (circumventing the need for exceptions). Common exceptions include an invalid argument (e.g. value is outside of the domain of a function), an unavailable resource (like a missing file, a hard disk error, or out-of-memory errors), or that the routine has detected a normal condition that requires special handling, e.g., attention, end of file.
Exception handling solves the semipredicate problem, in that the mechanism distinguishes normal return values from |
https://en.wikipedia.org/wiki/Waldspurger%27s%20theorem | In mathematics, Waldspurger's theorem, introduced by , is a result that identifies Fourier coefficients of modular forms of half-integral weight k+1/2 with the value of an L-series at s=k/2. |
https://en.wikipedia.org/wiki/Algorithmic%20management | Algorithmic management is a term used to describe certain labor management practices in the contemporary digital economy. In scholarly uses, the term was initially coined in 2015 by Min Kyung Lee, Daniel Kusbit, Evan Metsky, and Laura Dabbish to describe the managerial role played by algorithms on the Uber and Lyft platforms, but has since been taken up by other scholars to describe more generally the managerial and organisational characteristics of platform economies. However, digital direction of labor was present in manufacturing already since the 1970s and algorithmic management is becoming increasingly widespread across a wide range of industries.
The concept of algorithmic management can be broadly defined as the delegation of managerial functions to algorithmic and automated systems. Algorithmic management has been enabled by recent advances in digital technologies which allow for the real-time and "large-scale collection of data" which is then used to "improve learning algorithms that carry out learning and control functions traditionally performed by managers".
In the contemporary workplace, firms employ an ecology of accounting devices, such as “rankings, lists, classifications, stars and other symbols’ in order to effectively manage their operations and create value without the need for traditional forms of hierarchical control.” Many of these devices fall under the label of what is called algorithmic management, and were first developed by companies operating in the sharing economy or gig economy, functioning as effective labor and cost cutting measures. The Data&Society explainer of the term, for example, describes algorithmic management as ‘a diverse set of technological tools and techniques that structure the conditions of work and remotely manage workforces. Data&Society also provides a list of five typical features of algorithmic management:
Prolific data collection and surveillance of workers through technology;
Real-time responsiveness to |
https://en.wikipedia.org/wiki/Ted%20Livingston%20%28Kik%29 | Ted Livingston (born 1987) is founder of Kik Messenger that provides mobile messaging service.
Early life
Livingston was born in Toronto. As a child, Livingston had a passion for building Lego structures and automated machines in his basement. He continues following this interest on the school robotics team.
From 1997 to 2005, Livingston attended the Crescent School, in Toronto. He later stated that Crescent "provided an education which couldn't be more perfect for me as an individual. It allowed me to grow and develop my own philosophies and directions in life."
2005–Present: UWaterloo and Kik
Livingston took a mechatronics bachelor's degree at the University of Waterloo from 2005 to 2009 in order to "pursue his dream of building robots." He participated in the Waterloo co-op program there, with placements at Honda and the City of Toronto government. Livingston eventually landed a co-op position as a system engineering project coordinator at Research in Motion.
In December 2007, eight months into his co-op, Livingston was promoted by Research in Motion to "Technical Product Management Coordinator". In 2008, he turned down a full-time job to return to school at the advice of his manager.
Back at Waterloo, Livingston decided to take part in the Velocity program, a startup-focused community where he founded Kik Messenger instead of completing his degree.
Kik now competes with Facebook and WeChat. At its peak in 2015 it boasted over 275 million registered users, and it was used by approximately 40 percent of American teens. In 2015 Chinese Internet giant Tencent invested $50 million into Kik.
In 2014, Forbes placed him on their "Top 30 under 30" in their technology list.
In November 2015, Torontolife ranked Livingston at #20 on their "Toronto's 50 Most Influential,". |
https://en.wikipedia.org/wiki/Zeal%20%28web%29 | Zeal was a volunteer-built web directory launched by Brian Goler and Kevin Berk in 1999, and then acquired by LookSmart in October 2000 for $20 million. Zeal combined the work of Looksmart's paid editors with that of volunteers who profiled websites and placed them in a hierarchy of subcategories. The resulting categories and profiles were downloaded at intervals by LookSmart and its partners, other search companies such as MSN, Lycos, and Altavista, for use in their own systems with or without modification.
Paid editors attended to commercial sites and oversaw the voluntary work on non-commercial sites.
Volunteers worked under a defined set of Guidelines and were required to pass an introductory level test on those Guidelines before submitting site profiles or edits. As points and experience were acquired, volunteers could elect to take a further exam which allowed them to "adopt" and create topic categories of special interest. They could then move up the organizational structure from Community Member to Zealot to Expert Zealot, acquiring additional tools and oversight responsibility at each level. Expert Zealots, who could move or delete some whole categories, monitored the day-to-day operations of the non-commercial portion of the directory and acted as mentors to new members.
Active volunteers were found in many English-speaking countries (particularly North America, United Kingdom, India, Australia, and New Zealand) and some other countries such as Spain, Switzerland, and Japan.
By March 2003, Zeal had passed the 250,000 listings mark; eventually it passed the 400,000 mark due, in part, to the Zeal Charity Drive contest of October 2003, which saw over $25,000 distributed around prominent charities such as the WWF.
After Looksmart's acquisition of Zeal, its internet traffic as measured by Alexa fluctuated considerably; after MSN withdrew from the related partnership, Zeal traffic declined from "usually better than 2000th" (mid-2003) to "about 5000th" |
https://en.wikipedia.org/wiki/Degenerative%20disease | Degenerative disease is the result of a continuous process based on degenerative cell changes, affecting tissues or organs, which will increasingly deteriorate over time.
In neurodegenerative diseases, cells of the central nervous system stop working or die via neurodegeneration. An example of this is Alzheimer's disease. The other two common groups of degenerative diseases are those that affect circulatory system (e.g. coronary artery disease) and neoplastic diseases (e.g. cancers).
Many degenerative diseases exist and some are related to aging. Normal bodily wear or lifestyle choices (such as exercise or eating habits) may worsen degenerative diseases, but this depends on the disease. Sometimes the main or partial cause behind such diseases is genetic. Thus some are clearly hereditary like Huntington's disease. Sometimes the cause is viruses, poisons or other chemicals. The cause may also be unknown.
Some degenerative diseases can be cured. In those that can not, it may be possible to alleviate the symptoms.
Examples
Alzheimer's disease (AD)
Amyotrophic lateral sclerosis (ALS, Lou Gehrig's disease)
Cancers
Charcot–Marie–Tooth disease (CMT)
Chronic traumatic encephalopathy
Cystic fibrosis
Some cytochrome c oxidase deficiencies (often the cause of degenerative Leigh syndrome)
Ehlers–Danlos syndrome
Fibrodysplasia ossificans progressiva
Friedreich's ataxia
Frontotemporal dementia (FTD)
Some cardiovascular diseases (e.g. atherosclerotic ones like coronary artery disease, aortic stenosis, congenital defects etc.)
Huntington's disease
Infantile neuroaxonal dystrophy
Keratoconus (KC)
Keratoglobus
Leukodystrophies
Macular degeneration (AMD)
Marfan's syndrome (MFS)
Some mitochondrial myopathies
Mitochondrial DNA depletion syndrome
Mueller–Weiss syndrome
Multiple sclerosis (MS)
Multiple system atrophy
Muscular dystrophies (MD)
Neuronal ceroid lipofuscinosis
Niemann–Pick diseases
Osteoarthritis
Osteoporosis
Parkinson's disease
Pulmonary ar |
https://en.wikipedia.org/wiki/IBM%204764 | The IBM 4764 Cryptographic Coprocessor is a secure cryptoprocessor that performs cryptographic operations used by application programs and by communications such as SSL private key transactions associated with SSL digital certificates.
Details
Each coprocessor includes a tamper-responding hardware security module (HSM) that provides secure storage for storing master keys and other sensitive data. The HSM has been certified to meet FIPS 140-2 Level 4 security requirements. The coprocessor is supported on all IBM server platforms including System z (mainframe), System i, System p, and System x (Linux or Microsoft Windows). On System z, it is called the "Crypto Express2".
Applications may include financial PIN transactions, bank-to-clearing-house transactions, EMV transactions for integrated circuit (chip) based credit cards, and general-purpose cryptographic applications using symmetric key algorithms, hashing algorithms, and public key algorithms.
The operational keys (symmetric or RSA private) are generated in the coprocessor and are then saved either in a keystore file or in application memory, encrypted under the master key of that coprocessor. Any coprocessor with an identical master key can use those keys.
As of June 2005, the IBM 4764 superseded the IBM 4758 that was discontinued.
At the end of December 2011, the IBM 4764 was discontinued. The successor to the 4764 was introduced on each of the IBM server platforms between 2009 and 2011:
November 2009 for the System z (mainframe), where it is called the Crypto Express3
April 2010 for Power Systems, where it is available as feature codes 4807, 4808, and 4809
May 2011 for System x (PC) servers, where it is called the 4765 |
https://en.wikipedia.org/wiki/Reichert%27s%20membrane | Reichert's membrane is an extraembryonic membrane that forms during early mammalian embryonic development. It forms as a thickened basement membrane to cover the embryo immediately following implantation to give protection to the embryo from the uterine pressures exerted. Reichert's membrane is also important for the maternofetal exchange of nutrients. The membrane collapses once the placenta has fully developed.
Structure
Reichert's membrane is a multilayered, non-vascular, specialised thickened basement membrane that forms on the inner surface of the trophoblast around the time of implantation, and during the formation of the placenta.
It is composed of an extracellular matrix that includes laminin, type IV collagen, and nidogen, and is secreted by embryonic cells in the distal parietal endoderm.
The synthesis of laminin 111 in the embryo contributes to the formation of Reichert's membrane.
Function
Reichert's membrane functions as a buffer space between the embryo and the decidua. This space provides protection to the embryo from varying uterine pressures exerted by smooth muscle contractions of the myometrium.
During post gastrulation Reichert's membrane is necessary for the maternofetal exchange of nutrients. Reichert's membrane encloses the embryo until the amnion develops, and when the placenta is fully developed the membrane collapses.
A major difference in the early formation of the mouse embryo, and that of the human embryo is that in the mouse following implantation the epiblast takes on an egg or cylindrical shape; in the human the epiblast forms into a horizontal, disc-shape the bilaminar disc. A study that looked at this morphological difference between a human embryo initial development and a mouse embryo, concluded that it is likely that Reichert’s membrane is the key regulator of the epiblast’s horizontal growth. |
https://en.wikipedia.org/wiki/Thomas%20Greenhill%20%28surgeon%29 | Thomas Greenhill (1669?–1740) was a surgeon who worked in London and was also author of a book Νεκροκηδεία (Greek, literally Dead-funeral) or The Art of Embalming on embalming. He was also surgeon to Henry Howard, 7th Duke of Norfolk. He was born after the death of his father, William Greenhill, the last of 39 children born to Elizabeth Greenhill.
Life
According to the Oxford Dictionary of National Biography, "Most of the basic details of his life are a matter for conjecture. Essential parish papers have been mutilated ... Papers concerning his family shed no direct light on his education, practice, family, or death." This lack of records seems to be due to discontinuing the recording of baptisms and other records when Cromwell was in power during the Commonwealth following the English Civil War. The Dictionary of National Biography observes that "he was born ..after his father's death, probably at Abbot's Langley, Hertfordshire, as his father died there". He then lived in King Street Bloomsbury in London. His father was William Greenhill, Secretary to General Monck, and his mother was Elizabeth Greenhill (1615–1679) of Harrow in Middlesex and Abbots Langley.
Embalming
His book Νεκροκηδεία or The Art of Embalming wherein is shewn the right of burial, and funeral ceremonies, especially that of preserving bodies after the Egyptian method was published in 1705. Its main concern was to advocate the importance of embalming for the burial of the aristocracy, and make it a task limited to surgeons and not "undertakers and quacks". The argument is based upon appeals to Antiquary, Greek and Roman Classics and Scripture and takes the form of three letters: the first to Charles Bernard, Serjeant Surgeon to Queen Anne, the second, John Lawson, the former president of the Royal College of Physicians and the third to Hans Sloane, secretary to the Royal Society. Its dedication is to Thomas Herbert, 8th Earl of Pembroke, the same person who was the dedicatee of John Locke's An Es |
https://en.wikipedia.org/wiki/TRISPHAT | TRISPHAT (full name tris(tetrachlorocatecholato)phosphate(1−)) is an inorganic anion with the formula often prepared as the tributylammonium () or tetrabutylammonium ( salt. The anion features phosphorus(V) bonded to three tetrachlorocatecholate () ligands. This anion can be resolved into the axially chiral enantiomers, which are optically stable (the picture shows the Δ enantiomer).
The TRISPHAT anion has been used as a chiral shift reagent for cations. It improves the resolution of 1H NMR spectra by forming diastereomeric ion pairs.
Preparation
The anion is prepared by treatment of phosphorus pentachloride with tetrachlorocatechol followed by a tertiary amine gives the anion:
PCl5 + 3 C6Cl4(OH)2 → H[P(O2C6Cl4)3] + 5 HCl
H[P(O2C6Cl4)3] + Bu3N → Bu3NH+ [P(O2C6Cl4)3]−
Using a chiral amine, the anion can be readily resolved. |
https://en.wikipedia.org/wiki/List%20of%20plant%20genera%20named%20for%20people%20%28A%E2%80%93C%29 | Since the first printing of Carl Linnaeus's Species Plantarum in 1753, plants have been assigned one epithet or name for their species and one name for their genus, a grouping of related species. Thousands of plants have been named for people, including botanists and their colleagues, plant collectors, horticulturists, explorers, rulers, politicians, clerics, doctors, philosophers and scientists. Even before Linnaeus, botanists such as Joseph Pitton de Tournefort, Charles Plumier and Pier Antonio Micheli were naming plants for people, sometimes in gratitude for the financial support of their patrons.
Early works researching the naming of plant genera include an 1810 glossary by and an etymological dictionary in two editions (1853 and 1856) by Georg Christian Wittstein. Modern works include The Gardener's Botanical by Ross Bayton, Index of Eponymic Plant Names and Encyclopedia of Eponymic Plant Names by Lotte Burkhardt, Plants of the World by Maarten J. M. Christenhusz (lead author), Michael F. Fay and Mark W. Chase, The A to Z of Plant Names by Allan J. Coombes, the four-volume CRC World Dictionary of Plant Names by Umberto Quattrocchi, and Stearn's Dictionary of Plant Names for Gardeners by William T. Stearn; these supply the seed-bearing genera listed in the first column below. Excluded from this list are genus names not accepted (as of January 2021) at Plants of the World Online, which includes updates to Plants of the World (2017).
Key
Ba = listed in Bayton's The Gardener's Botanical
Bt = listed in Burkhardt's Encyclopedia of Eponymic Plant Names
Bu = listed in Burkhardt's Index of Eponymic Plant Names
Ch = listed in Christenhusz's Plants of the World
Co = listed in Coombes's The A to Z of Plant Names
Qu = listed in Quattrocchi's CRC World Dictionary of Plant Names
St = listed in Stearn's Dictionary of Plant Names for Gardeners
In addition, Burkhardt's Index is used as a reference for every row in the table not cited to Stearn.
Genera
See also
List o |
https://en.wikipedia.org/wiki/Heterodont | In anatomy, a heterodont (from Greek, meaning 'different teeth') is an animal which possesses more than a single tooth morphology.
In vertebrates, heterodont pertains to animals where teeth are differentiated into different forms. For example, members of the Synapsida generally possess incisors, canines ("dogteeth"), premolars, and molars. The presence of heterodont dentition is evidence of some degree of feeding and or hunting specialization in a species. In contrast, homodont or isodont dentition refers to a set of teeth that possess the same tooth morphology.
In invertebrates, the term heterodont refers to a condition where teeth of differing sizes occur in the hinge plate, a part of the Bivalvia. |
https://en.wikipedia.org/wiki/Costal%20cartilage | The costal cartilages are bars of hyaline cartilage that serve to prolong the ribs forward and contribute to the elasticity of the walls of the thorax. Costal cartilage is only found at the anterior ends of the ribs, providing medial extension.
Differences from Ribs 1-12
The first seven pairs are connected with the sternum; the next three are each articulated with the lower border of the cartilage of the preceding rib; the last two have pointed extremities, which end in the wall of the abdomen.
Like the ribs, the costal cartilages vary in their length, breadth, and direction. They increase in length from the first to the seventh, then gradually decrease to the twelfth.
Their breadth, as well as that of the intervals between them, diminishes from the first to the last. They are broad at their attachments to the ribs, and taper toward their sternal extremities, excepting the first two, which are of the same breadth throughout, and the sixth, seventh, and eighth, which are enlarged where their margins are in contact.
They also vary in direction: the first descends a little to the sternum, the second is horizontal, the third ascends slightly, while the others are angular, following the course of the ribs for a short distance, and then ascending to the sternum or preceding cartilage.
Structure
Each costal cartilage presents two surfaces, two borders, and two extremities.
Surfaces
The anterior surface is convex, and looks forward and upward: that of the first gives attachment to the costoclavicular ligament and the subclavius muscle; those of the first six or seven at their sternal ends, to the pectoralis major. The others are covered by, and give partial attachment to, some of the flat muscles of the abdomen.
The posterior surface is concave, and directed backward and downward; that of the first gives attachment to the sternothyroideus, those of the third to the sixth inclusive to the transversus thoracis muscle, and the six or seven inferior ones to the transvers |
https://en.wikipedia.org/wiki/Drift%20current | In condensed matter physics and electrochemistry, drift current is the electric current, or movement of charge carriers, which is due to the applied electric field, often stated as the electromotive force over a given distance. When an electric field is applied across a semiconductor material, a current is produced due to the flow of charge carriers.
The drift velocity is the average velocity of the charge carriers in the drift current. The drift velocity, and resulting current, is characterized by the mobility; for details, see electron mobility (for solids) or electrical mobility (for a more general discussion).
See drift–diffusion equation for the way that the drift current, diffusion current, and carrier generation and recombination are combined into a single equation.
Overview
Drift current is the electric current caused by particles getting pulled by an electric field. The term is most commonly used in the context of electrons and holes in semiconductors, although the same concept also applies to metals, electrolytes, and so on.
Drift current is caused by the electric force: Charged particles get pushed by an electric field. Electrons, being negatively charged, get pushed in the opposite direction to the electric field, while holes get pushed in the same direction as the electric field, but the resulting conventional current points in the same direction as the electric field in both cases.
If an electric field is applied to an electron in a vacuum, the electron will accelerate faster and faster, in approximately a straight line. A drift current looks very different than that up close. Typically, electrons are moving randomly in all directions (Brownian motion), frequently changing direction when they collide with grain boundaries or other disturbances. Between collisions, the electric field subtly accelerates them in one direction. So over time, they move at the drift velocity on average, but at any instant the electrons are moving at the (typically much |
https://en.wikipedia.org/wiki/CEBus | CEBus(r), short for Consumer Electronics Bus, also known as EIA-600, is a set of electrical standards and communication protocols for electronic devices to transmit commands and data. It is suitable for devices in households and offices to use, and might be useful for utility interface and light industrial applications.
History
In 1984, members of the Electronic Industries Alliance (EIA) identified a need for standards that included more capability than the de facto home automation standard X10. X10 provided blind transmission of the commands ON, OFF, DIM, BRIGHT, ALL LIGHTS ON, and ALL UNITS OFF over powerline carrier, and later infrared and short range radio mediums. Over a six-year period, engineers representing international companies met on a regular basis and developed a proposed standard. They called this standard CEBus (pronounced "see bus"). The CEBus standard was released in September 1992.
CEBus is an open architecture set of specification documents which define protocols for products to communicate through power line wire, low voltage twisted pair wire, coaxial cable, infrared, RF, and fiber optics.
The CEBus Standard was developed on the foundation of an IR (infrared) protocol developed by GE (General Electric). This work was transferred to the EIA at the beginning of the EIA's involvement, under the plan that it would be expanded then maintained by the EIA.
Technology
Powerline carrier
The CEBus standard includes such things as spread spectrum modulation on the power line. Spread spectrum involves starting a modulation at one frequency, and altering the frequency during its cycle. The CEBus power line standard begins each burst at 100 kHz, and increases linearly to 400 kHz during a 100 microsecond duration. Both the bursts (referred to as "superior" state) and the absence of burst (referred to as the "inferior" state) create similar digits, so a pause in between is not necessary.
A digit 1 is created by an inferior or superior state that l |
https://en.wikipedia.org/wiki/Monitoring%20Maintenance%20Lifecycle | The Monitoring Maintenance Lifecycle (MML) is a monitoring development process to lower maintenance expenses and improve the dependability of IT infrastructure with regard to service recovery-related issues. It is based on the classical Waterfall model.
Monitoring Maintenance Lifecycle are methods and standards for improving and mastering maintenance processes, supporting processes and management processes throughout the monitoring lifecycle.
The quest for the optimized mix of processes has resulted in different standards throughout the history. One of the latest which was published is the ISO/IEC 12207 standard. This standard was proposed in 1988 and published in August 1995. It was created to establish a common international framework to acquire, supply, develop, operate, and maintain.
ISO/IEC 12207 consists of three types of processes:
Primary lifecycle processes;
Supporting lifecycle processes;
Organizational lifecycle processes.
Supply
During the supply phase a project management plan is developed based on RUP. This plan contains information about the project such as different milestones that need to be reached. This project management plan is needed in the start up phase and throughout the monitoring and maintenance lifecycle.
Software development process |
https://en.wikipedia.org/wiki/Zond%206 | Zond 6 was a formal member of the Soviet Zond program, and an unpiloted version of the Soyuz 7K-L1 crewed Moon-flyby spacecraft. It was launched on a lunar flyby mission on November 10, 1968, from a parent satellite (68-101B) in Earth parking orbit. The spacecraft carried a biological payload of turtles, flies, and bacteria, scientific probes including cosmic ray and micrometeoroid detectors, and as photographic equipment.
The mission was a precursor to a crewed circumlunar flight which the Soviets hoped could occur in December 1968, thus beating the American Apollo 8. However, after rounding the Moon on November 14, Zond 6 crashed on its return to Earth, due to a parachute failure.
Mission
Zond 6 was the official designation for Soyuz 7K-L1 s/n 12. It was supposed to photograph the Moon in colour and in black and white, from 8,000 km and 2,600 km ranges, then return to Earth, landing at Tyuratam, only 16 km from the launch pad. It had been a long and difficult road to develop the L1 guidance system, but it worked perfectly that time.
Zond 6 flew around the Moon on 14 November 1968, at a minimum distance of 2,420 km. Photographs of the lunar near side and far side were obtained with panchromatic film. Each photo was . Some of the views allowed for stereo pictures. The photos were taken from distances of approximately 11,000 km and 3,300 km.
The flight was tracked by Jodrell Bank Observatory, which picked up telemetry data and voice transmissions. The telemetry data appear to have been simulated sensor readings; the voices were either from a tape recorder or being relayed through the probe. In 2018, a recording of the signals was found in Jodrell Bank's archive.
During the craft's return from the Moon, concerns arose about the falling temperatures of the hydrogen peroxide tanks. This issue was addressed by rotating Zond 6 so that the tanks were in direct sunlight. While this was successful in raising the temperatures of the tanks, it also damaged the seal around |
https://en.wikipedia.org/wiki/Differential-linear%20attack | Introduced by Martin Hellman and Susan K. Langford in 1994, the differential-linear attack is a mix of both linear cryptanalysis and differential cryptanalysis.
The attack utilises a differential characteristic over part of the cipher with a probability of 1 (for a few rounds—this probability would be much lower for the whole cipher). The rounds immediately following the differential characteristic have a linear approximation defined, and we expect that for each chosen plaintext pair, the probability of the linear approximation holding for one chosen plaintext but not the other will be lower for the correct key. Hellman and Langford have shown that this attack can recover 10 key bits of an 8-round DES with only 512 chosen plaintexts and an 80% chance of success.
The attack was generalised by Eli Biham et al. to use differential characteristics with probability less than 1. Besides DES, it has been applied to FEAL, IDEA, Serpent, Camellia, and even the stream cipher Phelix. |
https://en.wikipedia.org/wiki/Genetic%20fallacy | The genetic fallacy (also known as the fallacy of origins or fallacy of virtue) is a fallacy of irrelevance in which arguments or information are dismissed or validated based solely on their source of origin rather than their content. In other words, a claim is ignored or given credibility based on its source rather than the claim itself.
The fallacy therefore fails to assess the claim on its merit. The first criterion of a good argument is that the premises must have bearing on the truth or falsity of the claim in question. Genetic accounts of an issue may be true, and they may help illuminate the reasons why the issue has assumed its present form, but they are not conclusive in determining its merits.
In The Oxford Companion to Philosophy (1995) it is asserted that the term originated in Morris Raphael Cohen and Ernest Nagel's book Logic and Scientific Method (1934). However, in a book review published in The Nation in 1926, Mortimer J. Adler complained that The Story of Philosophy by Will Durant was guilty throughout of "the fallacy of genetic interpretation." Adler characterized the genetic fallacy generally as "the substitution of psychology for logic."
Examples
From Attacking Faulty Reasoning by T. Edward Damer, Third Edition 36:
There are numerous motives explaining why people choose to wear wedding rings, but it would be a fallacy to presume those who continue the tradition are promoting sexism.
Another example would be from How to Win Every Argument: The Use and Abuse of Logic (2006) by Madsen Pirie, p. 82:
As the author points out, private developers may well have legitimate and knowledgeable opinions on such a matter.
See also
Appeal to noveltyThe argument that a newer idea is superior
Chronological snobberyThe argument that an older idea is inferior
The argument that an older idea is superior
Etymological fallacyAn assertion that the historical meaning of the word is its only true meaning
"Not invented here"A dismissal of "f |
https://en.wikipedia.org/wiki/TAN-1057%20C | TAN-1057 C and TAN-1057 D are organic compounds found in the Flexibacter sp. PK-74 bacterium. TAN-1057 C and D are closely related structurally as diastereomers. Also related are TAN-1057 A and TAN-1057 B, isolated from the same bacteria. The four compounds have been shown to be an effective antibiotics against methicillin-resistant strains of Staphylococcus aureus which act through the inhibition of protein biosynthesis. |
https://en.wikipedia.org/wiki/Liminal%20BioSciences | Liminal BioSciences Inc., formerly known as Prometic Life Sciences Inc., is a Canadian biopharmaceutical company.
History
Liminal was founded in 1988, as a commercial spinoff of research at the University of Cambridge on affinity chromatography. Its founder was the current President and CEO, Pierre Laurin, whose research formed the basis for the company's products. Liminal was incorporated on October 14, 1994 under the Canada Business Corporations Act, or the CBCA, under the name Innovon Life Sciences Holdings Limited.
In August 2016, the company announced the acquisition of Telesta Therapeutics.
It changed its name to “Prometic Life Sciences Inc.” on May 19, 1998 and subsequently rebranded to “Liminal BioSciences Inc.” on October 3, 2019. On July 28, 1998, the Company's common shares began trading on the Toronto Stock Exchange, or the TSX, under the trading symbol “PLI” and, on October 7, 2019, began trading under the trading symbol “LMNL.”
On November 18, 2019, Liminal's common shares began trading on the Nasdaq Global Market, or the Nasdaq, under the trading symbol “LMNL.”
On August 5, 2020, it voluntarily delisted its common shares from the TSX. On May 19, 1998, the Corporation went public through a listing on the Toronto Stock Exchange.
In September 2023, the company was taken private by Thomvest Asset Management Ltd and delisted from Nasdaq.
Business
Liminal BioSciences development programs target seven transmembrane GPCRs (7TM GPCR’s), where the receptor protein passes through the cell membrane seven times. These receptors are easily accessible to hydrophilic drugs due to their presence on the cell surface, and their non-uniform expression enables selectivity in modulating physiological processes. Agonists and antagonists of 7TM GPCRs receptors are utilized for treating various diseases in all organ systems. An agonist is a drug that binds to a target and mimics the action of the natural ligand. An antagonist is a drug that binds to a receptor an |
https://en.wikipedia.org/wiki/Mathematical%20analysis | Analysis is the branch of mathematics dealing with continuous functions, limits, and related theories, such as differentiation, integration, measure, infinite sequences, series, and analytic functions.
These theories are usually studied in the context of real and complex numbers and functions. Analysis evolved from calculus, which involves the elementary concepts and techniques of analysis.
Analysis may be distinguished from geometry; however, it can be applied to any space of mathematical objects that has a definition of nearness (a topological space) or specific distances between objects (a metric space).
History
Ancient
Mathematical analysis formally developed in the 17th century during the Scientific Revolution, but many of its ideas can be traced back to earlier mathematicians. Early results in analysis were implicitly present in the early days of ancient Greek mathematics. For instance, an infinite geometric sum is implicit in Zeno's paradox of the dichotomy. (Strictly speaking, the point of the paradox is to deny that the infinite sum exists.) Later, Greek mathematicians such as Eudoxus and Archimedes made more explicit, but informal, use of the concepts of limits and convergence when they used the method of exhaustion to compute the area and volume of regions and solids. The explicit use of infinitesimals appears in Archimedes' The Method of Mechanical Theorems, a work rediscovered in the 20th century. In Asia, the Chinese mathematician Liu Hui used the method of exhaustion in the 3rd century CE to find the area of a circle. From Jain literature, it appears that Hindus were in possession of the formulae for the sum of the arithmetic and geometric series as early as the 4th century BCE.
Ācārya Bhadrabāhu uses the sum of a geometric series in his Kalpasūtra in . In Indian mathematics, particular instances of arithmetic series have been found to implicitly occur in Vedic Literature as early as .
Medieval
Zu Chongzhi established a method that would later be |
https://en.wikipedia.org/wiki/TRPN | TRPN is a member of the transient receptor potential channel family of ion channels, which is a diverse group of proteins thought to be involved in mechanoreception. The TRPN gene was given the name no mechanoreceptor potential C (nompC) when it was first discovered in fruit flies, hence the N in TRPN. Since its discovery in fruit flies, TRPN homologs have been discovered and characterized in worms, frogs, and zebrafish.
Structure
A structure of NOMPC was published in 2017, solved using electron cryo-microscopy. X-ray crystallography studies of channel segments cloned from fruit flies and zebrafish have led to the hypothesis that multiple ankyrin repeats at TRPN's N-terminus are involved in the gating of the channel pore. Crystallography studies of TRPY1, a yeast TRP homolog, have shown that aromatic residues conserved across TRP family members, including TRPN, in the sixth transmembrane domain are critical to the gating mechanism as well.
Function
As a mechanoreceptor, TRPN responds to impinging mechanical forces. Studies in TRPN deficient adult fruit flies and larvae have shown that these null mutants have severe difficulty moving, which suggests a role for TRPN in proprioception. This hypothesis is further strengthened by immunostaining studies in fruit flies that have shown TRPN localization in the cilia of campaniform sensilla and chordotonal organs in Johnston's organ. Further immunostaining studies in fruit flies have identified, with higher resolution techniques, that TRPN is localized at the distal end of motile mechanosensory cilia in Johnston's organ. However, TRPN is not required for transduction of mechanical stimuli in larvae or adult flies, suggesting that the TRPV channels nanchung and inactive may also serve a mechanosensory function.
Studies in worms have shown that TRPN mutants have locomotion defects, as well as a decreased basal slowing response, which is a reduction in rate of motion that is induced by contact with a food source. This result |
https://en.wikipedia.org/wiki/Sun%20Yat-sen | Sun Yat-sen (, , 12 November 1866 – 12 March 1925) was a Chinese revolutionary statesman, physician, and political philosopher who served as the first provisional president of the Republic of China and the first leader of the Kuomintang (Nationalist Party of China). He is called the "Father of the Nation" in the present-day Republic of China (Taiwan) and the "Forerunner of the Revolution" in the People's Republic of China for his instrumental role in the overthrow of the Qing dynasty during the 1911 Revolution. Sun is unique among 20th-century Chinese leaders for being widely revered by both the Communist Party in Mainland China and the Nationalist Party in Taiwan.
Educated overseas, Sun is considered to be one of the greatest and most important leaders of modern China, but his political life was one of constant struggle and frequent exile. After the success of the revolution in 1911, he quickly resigned as president of the newly founded Republic of China and relinquished the position to Yuan Shikai. He soon went to exile in Japan for safety but returned to found a revolutionary government in Southern China, as a challenge to the warlords who controlled much of the nation. In 1923, he invited representatives of the Communist International to Canton (Guangzhou) to reorganize his party and formed a brittle alliance with the Chinese Communist Party. He did not live to see his party unify the country under his successor, Chiang Kai-shek, in the Northern Expedition. He died in Peking (Beijing) of gallbladder cancer in 1925.
Sun's chief legacy is his political philosophy known as the Three Principles of the People: Mínzú () or nationalism (independence from foreign domination), Mínquán () or "rights of the people" (sometimes translated as "democracy"), and Mínshēng () or people's livelihood (sometimes translated as "communitarianism" or "welfarism").
Names
Sun's genealogical name was Sun Deming (Syūn Dāk-mìhng; ). As a child, his pet name was Tai Tseung (Dai-jeuhng; |
https://en.wikipedia.org/wiki/Paleomerus | Paleomerus is a genus of strabopid, a group of extinct arthropods. It has been found in deposits from the Cambrian period (Atdabanian epoch). It is classified in the family Strabopidae of the monotypic order Strabopida. It contains two species, P. hamiltoni from Sweden and P. makowskii from Poland. The generic name is composed by the Ancient Greek words παλαιός (palaiós), meaning "ancient", and μέρος (méros), meaning "part" (and therefore, "ancient part").
Paleomerus is one of the oldest arthropods, being sometimes interpreted as the model of the first arachnomorphs. It is part of the order Strabopida, a poorly known group closely related to the aglaspidids of uncertain affinities, often being ignored by researchers and authors due to the poor preservation and abundance of their fossils. It has been suggested that Paleomerus and the closely related Strabops could be synonymous with each other, since they differ only in the size of the telson (the posteriormost division of the body) and the position of the eyes. These two genera were originally deferred by a hypothetical twelfth segment in Paleomerus, but after the discovery and description of a fourth specimen of P. hamiltoni, it has been shown that this segment actually represents the tail of the animal.
Description
As the other strabopids, Paleomerus was a small-sized arthropod. The bigger species was P. hamiltoni at 9.3 centimetres (3.7 inches), while the smaller P. makowskii reached only 7.3 cm (2.9 in). Within Strabopida, the closely related Strabops thacheri exceeded the length of Paleomerus with 11 cm (4.3 in), while Parapaleomerus sinensis reached a total length of 9.2 cm (3.6 in).
Like some other arthropod clades, the strabopids possessed segmented bodies and jointed appendages (limbs) covered in a cuticle composed of proteins and chitin. The arthropod body is divided into two tagmata (sections); the frontal prosoma (head) and posterior opisthosoma (abdomen). The appendages were attached to the prosoma, |
https://en.wikipedia.org/wiki/Iain%20Rice | Iain Alexander Rice (11 October 1947 – 8 October 2022) was a British model railway enthusiast, writer and illustrator who authored more than 20 books and 400 articles on model railways over five decades. Railway Modeller described Rice as one of the hobby's "most influential and prominent enthusiasts of the last fifty years".
Biography
Iain Alexander Rice was born 11 October 1947 in Crouch End, North London, and grew up in Chingford in East London. He constructed his first model railway when he was eight. Rice was educated at Parmiter's School in Bethnal Green followed by Chingford County High School, then Reigate College of Art in Surrey. In 1971 Rice began working as a firefighter in the Northamptonshire Fire Brigade, before moving to Devon in 1977 and subsequently serving 26 years as a retained firefighter with the Devon Fire Brigade.
Rice began contributing to the model railway press in the 1970s, with his first books published in the 1980s for Wild Swan Publications. His writings were often accompanied by his own artwork. In the 1990s Rice helped launch and edit Modelling Railways Illustrated followed by the short-lived periodical Rail Model Digest. In the 2000s he also wrote a series of books on model railroads for the American publisher Kalmbach Books.
Books |
https://en.wikipedia.org/wiki/Ion%20gun | An Ion Gun typically refers to an instrument that generates a beam of heavy ions with a well defined energy distribution. The ion beam is produced from a plasma that has been confined within a volume. Ions of a particular energy are extracted, accelerated, collimated and/or focused. The ion gun is composed of an ion source, extraction grid structure and a collimation/lensing structure. The plasma can be made up of an inert or reactive gas (e.g. N+ and O+) or an easily condensable substance (e.g. C+ and B+). The plasma can be formed from molecules that contain the substance which will form the beam, in which case, these molecules must be fragmented then ionized (e.g. H and CH4 can together be fragmented and ionized to create a beam for depositing diamond-like carbon films).
The ion current density (or similarly the ion flux), the ion energy spread, and the resolution of the ion beam are key factors in ion gun design. The ion current density is controlled by the ion source, the energy spread is determined primarily by the extraction grid, and the resolution is determined primarily by the optical column.
The ion gun is an important component in surface science in that it provides the scientist with a means to sputter etch a surface and generate an elemental or chemical depth profile. Modern ion guns can produce beam energies from 10eV to more than 10keV.
Measurement and Detection
A Nanocoulombmeter in combination with a Faraday cup can be used
to detect and measure the beams emitted from ion guns.
See also
The term "ion gun" might also refer to an accelerator of any charged particle. See the following:
Coronal Ionizer gun
Soft X-ray Ionizer
Electron gun
Duoplasmatron
Capillaritron
Particle accelerator
Ion ray gun
Focused ion beam |
https://en.wikipedia.org/wiki/Jose%20Emilio%20Esteban | Jose Emilio Esteban is a Mexican-American food scientist and former veterinarian who is serving as under secretary of agriculture for food safety in the Biden administration since January 2023.
Education
Esteban earned a Doctor of Veterinary Medicine from the National Autonomous University of Mexico. He later earned a Master of Business Administration, Master of Science in preventive veterinary medicine, and PhD in epidemiology from the University of California, Davis.
Career
Esteban began his career as a veterinarian in Mexico. After graduating from University of California, Davis, he joined the Centers for Disease Control and Prevention, serving as a staff epidemiologist, officer with the Epidemic Intelligence Service, and assistant director of the Food Safety Office. He joined the Food Safety and Inspection Service in 2006 and has served as a science advisor, laboratory director, and chief scientist in the department. He is also chair of the Codex Alimentarius Commission Committee on Food Hygiene and vice president of the International Association for Food Protection. |
https://en.wikipedia.org/wiki/Volute | A volute is a spiral, scroll-like ornament that forms the basis of the Ionic order, found in the capital of the Ionic column. It was later incorporated into Corinthian order and Composite column capitals. Four are normally to be found on an Ionic capital, eight on Composite capitals and smaller versions (sometimes called helix) on the Corinthian capital.
The word derives from the Latin voluta ("scroll"). It has been suggested that the ornament was inspired by the curve of a ram's horns, or perhaps was derived from the natural spiral found in the ovule of a common species of clover native to Greece. Alternatively, it may simply be of geometrical origin.
The ornament can be seen in Renaissance and Baroque architecture and is a common decoration in furniture design, silverware and ceramics. A method of drawing the complex geometry was devised by the ancient Roman architect Vitruvius through the study of classical buildings and structures.
See also
Scrollwork
Ionic order
Spiral |
https://en.wikipedia.org/wiki/Beijing%20World%20Park | Beijing World Park () is a theme park that attempts to give visitors the chance to see the world without having to leave Beijing. The park covers 46.7 hectares and is located in the southwestern Fengtai District of Beijing. It is about 17 km from Tiananmen, the City center, and 40 km from the Capital International Airport. The park opened in 1993 and is estimated to receive 1.5 million visitors annually.
The park
The entrance to the park is made up of a Gothic castle, Roman corridor, and granite relief sculptures. Immediately inside the gate is an Italian-style terrace garden with grand staircases, fountains, and sculptures inspired by originals from the European Renaissance. More lawns and gardens are scattered throughout the park. On these lawns are miniature models of around 100 of the world's most famous statues including the American Statue of Liberty, Copenhagen's Little Mermaid, Michelangelo's David, and the Venus de Milo. Once inside the gates, Beijing World Park consists of two main parts: the scenic portion and the shopping, dining, and entertainment area.
Scenic area
The scenic area of the park models itself after the naturalistic layout of the globe, representing the four major oceans, and focusing on five continents: Asia, Africa, Europe, North America, and South America. The park contains about a hundred (109) scaled-down replicas of famous landmarks from nearly 40 countries and regions around the world, including the Tower Bridge in London, the Eiffel Tower in Paris, and the Great Pyramids in Egypt. There is even a miniature Manhattan, complete with twin towers of the World Trade Center. Each landmark represents its country or region of origin and is situated in the park according to its location on the map.
Close attention to detail was paid in modeling these landmarks after their originals. For example, detailed carvings and ornamentations are included. Even the materials used are modeled after their originals to create the most authentic look po |
https://en.wikipedia.org/wiki/TerraNet%20AB | Terranet AB is a company that develops technology for Advanced driver-assistance systems (ADAS) and autonomous vehicles. It is headquartered in Sweden and has an office in Stuttgart, Germany. Terranet is currently led by acting CEO Magnus Andersson.
Terranet previously focused on delivering mobile telephony and data services via a peer-to-peer mobile mesh network of handsets and light infrastructure. Since 2018, the company is focused on developing technology for advanced driver assistance and autonomous vehicles.
Terranet addresses the fast-growing global ADAS market, which is projected to reach USD 84 billion by 2025 - an increase of 150% from 2021.
The company's primary focus is to develop and commercialize their BlincVision product, a new type of anti-collision system for advanced driver assistance for motorized vehicles based on laser scanning, event cameras and three-dimensional image analysis. BlincVision is based on Voxelflow, a patented software for advanced three-dimensional image analysis of moving objects. In May 2022, Terranet shared that BlincVision is expected to be production ready in a couple of years.
Terranet Holding AB (publ) is listed on Nasdaq First North Premier Growth Market since 2017 (Nasdaq: TERRNT B).
History
Terranet AB was founded in 2004 by the inventor Anders Carlius, a serial entrepreneur from Lund, Skåne County. Carlius, who came from a background in chip manufacturing with Switchcore and also worked for web portal operator Spray Network, served as the first CEO until 2010. Since its inception Terranet has been headquartered at Lund's Ideon Science Park, which is best known as the home of Ericsson Radio Systems.
Carlius says he came up with the idea while travelling on safari in east Africa with his wife Emma. The first extensive pilot project was carried out in the autumn of 2005 on a farming co-operative in Botswana. Other trials included an agreement with Indian operator Bharat Sanchar Nigam to test 50 handsets in late 2 |
https://en.wikipedia.org/wiki/Self-consistency%20principle%20in%20high%20energy%20physics | The self-consistency principle was established by Rolf Hagedorn in 1965 to explain the thermodynamics of fireballs in high energy physics collisions. A thermodynamical approach to the high energy collisions first proposed by E. Fermi.
Partition function
The partition function of the fireballs can be written in two forms, one in terms of its density of states, , and the other in terms of its mass spectrum, .
The self-consistency principle says that both forms must be asymptotically equivalent for energies or masses sufficiently high (asymptotic limit). Also, the density of states and the mass spectrum must be asymptotically equivalent in the sense of the weak constraint proposed by Hagedorn as
.
These two conditions are known as the self-consistency principle or bootstrap-idea. After a long mathematical analysis Hagedorn was able to prove that there is in fact and satisfying the above conditions, resulting in
and
with and related by
.
Then the asymptotic partition function is given by
where a singularity is clearly observed for →. This singularity determines the limiting temperature in Hagedorn's theory, which is also known as Hagedorn temperature.
Hagedorn was able not only to give a simple explanation for the thermodynamical aspect of high energy particle production, but also worked out a formula for the hadronic mass spectrum and predicted the limiting temperature for hot hadronic systems.
After some time this limiting temperature was shown by N. Cabibbo and G. Parisi to be related to a phase transition, which characterizes by the deconfinement of quarks at high energies. The mass spectrum was further analyzed by Steven Frautschi.
Q-exponential function
The Hagedorn theory was able to describe correctly the experimental data from collision with center-of-mass energies up to approximately 10 GeV, but above this region it failed. In 2000 I. Bediaga, E. M. F. Curado and J. M. de Miranda proposed a phenomenological generalization of Hagedorn's theor |
https://en.wikipedia.org/wiki/Manycore%20processor | Manycore processors are special kinds of multi-core processors designed for a high degree of parallel processing, containing numerous simpler, independent processor cores (from a few tens of cores to thousands or more). Manycore processors are used extensively in embedded computers and high-performance computing.
Contrast with multicore architecture
Manycore processors are distinct from multi-core processors in being optimized from the outset for a higher degree of explicit parallelism, and for higher throughput (or lower power consumption) at the expense of latency and lower single-thread performance.
The broader category of multi-core processors, by contrast, are usually designed to efficiently run both parallel and serial code, and therefore place more emphasis on high single-thread performance (e.g. devoting more silicon to out of order execution, deeper pipelines, more superscalar execution units, and larger, more general caches), and shared memory. These techniques devote runtime resources toward figuring out implicit parallelism in a single thread. They are used in systems where they have evolved continuously (with backward compatibility) from single core processors. They usually have a 'few' cores (e.g. 2, 4, 8) and may be complemented by a manycore accelerator (such as a GPU) in a heterogeneous system.
Motivation
Cache coherency is an issue limiting the scaling of multicore processors. Manycore processors may bypass this with methods such as message passing, scratchpad memory, DMA, partitioned global address space, or read-only/non-coherent caches. A manycore processor using a network on a chip and local memories gives software the opportunity to explicitly optimise the spatial layout of tasks (e.g. as seen in tooling developed for TrueNorth).
Manycore processors may have more in common (conceptually) with technologies originating in high-performance computing such as clusters and vector processors.
GPUs may be considered a form of manycore process |
https://en.wikipedia.org/wiki/Perceptual-based%203D%20sound%20localization | Perceptual-based 3D sound localization is the application of knowledge of the human auditory system to develop 3D sound localization technology.
Motivation and Applications
Human listeners combine information from two ears to localize and separate sound sources originating in different locations in a process called binaural hearing. The powerful signal processing methods found in the neural systems and brains of humans and other animals are flexible, environmentally adaptable, and take place rapidly and seemingly without effort. Emulating the mechanisms of binaural hearing can improve recognition accuracy and signal separation in DSP algorithms, especially in noisy environments. Furthermore, by understanding and exploiting biological mechanisms of sound localization, virtual sound scenes may be rendered with more perceptually relevant methods, allowing listeners to accurately perceive the locations of auditory events. One way to obtain the perceptual-based sound localization is from the sparse approximations of the anthropometric features. Perceptual-based sound localization may be used to enhance and supplement robotic navigation and environment recognition capability. In addition, it is also used to create virtual auditory spaces which is widely implemented in hearing aids.
Problem Statement and Basic Concepts
While the relationship between human perception of sound and various attributes of the sound field is not yet well understood, DSP algorithms for sound localization are able to employ several mechanisms found in neural systems, including the interaural time difference (ITD, the difference in arrival time of a sound between two locations), the interaural intensity difference (IID, the difference in intensity of a sound between two locations), artificial pinnae, the precedence effect, and head-related transfer functions (HRTF). When localizing 3D sound in spatial domain, one could take into account that the incoming sound signal could be reflected, diffrac |
https://en.wikipedia.org/wiki/Urbach%E2%80%93Wiethe%20disease | Urbach–Wiethe disease is a very rare recessive genetic disorder, with approximately 400 reported cases since its discovery. It was first officially reported in 1929 by Erich Urbach and Camillo Wiethe, although cases may be recognized dating back as early as 1908.
The symptoms of the disease vary greatly from individual to individual. They may include a hoarse voice, lesions and scarring on the skin, easily damaged skin with poor wound healing, dry, wrinkly skin, and beading of the papules around the eyelids. All of these are results of a general thickening of the skin and mucous membranes. In some cases there is also a hardening of brain tissue in the medial temporal lobes, which can lead to epilepsy and neuropsychiatric abnormalities. The disease is typically not life-threatening and patients do not show a decreased life span.
Because Urbach–Wiethe disease is an autosomal recessive condition individuals can be carriers of the disease but show no symptoms. The disease is caused by loss-of-function mutations to chromosome 1 at 1q21, the extracellular matrix protein 1 (ECM1) gene. The dermatological symptoms are caused by a buildup of a hyaline material in the dermis and the thickening of the basement membranes in the skin. Urbach–Wiethe disease is typically diagnosed by its clinical dermatological manifestations, particularly the beaded papules on the eyelids. The discovery of the mutations within the ECM1 gene has allowed the use of genetic testing to confirm an initial clinical diagnosis. Periodic acid-Schiff (PAS) and immunohistochemical staining may also be used for diagnosis.
Currently, there is no cure for Urbach–Wiethe disease although there are ways to individually treat many of its symptoms. The discovery of the mutations of the ECM1 gene has opened the possibility of gene therapy or a recombinant ECM1 protein for Urbach–Wiethe disease treatment, but neither of these options are currently available. Some researchers are examining patients with Urbach–Wie |
https://en.wikipedia.org/wiki/Samia%20ricini | Samia ricini, the Eri silkmoth, is a species of insect, a member of the family Saturniidae which includes the giant silk moths. This moth is a domestic polyhybrid that has been bred for centuries due to the silk it makes. The name is based on the host plant used for feeding the caterpillars, castor, Ricinus communis. This moth is derived from several different species within the genus including Samia cynthia and Samia canningi.
Gallery |
https://en.wikipedia.org/wiki/List%20of%20personal%20standards%20of%20the%20kings%20of%20Portugal | This is a list of personal standards of the kings of Portugal.
See also
Flag of Portugal
Portuguese vexillology
Portugal
Personal standards
Portugal |
https://en.wikipedia.org/wiki/Convex%20embedding | In geometric graph theory, a convex embedding of a graph is an embedding of the graph into a Euclidean space, with its vertices represented as points and its edges as line segments, so that all of the vertices outside a specified subset belong to the convex hull of their neighbors. More precisely, if is a subset of the vertices of the graph, then a convex -embedding embeds the graph in such a way that every vertex either belongs to or is placed within the convex hull of its neighbors. A convex embedding into -dimensional Euclidean space is said to be in general position if every subset of its vertices spans a subspace of dimension .
Convex embeddings were introduced by W. T. Tutte in 1963. Tutte showed that if the outer face of a planar graph is fixed to the shape of a given convex polygon in the plane, and the remaining vertices are placed by solving a system of linear equations describing the behavior of ideal springs on the edges of the graph, then the result will be a convex -embedding. More strongly, every face of an embedding constructed in this way will be a convex polygon, resulting in a convex drawing of the graph.
Beyond planarity, convex embeddings gained interest from a 1988 result of Nati Linial, László Lovász, and Avi Wigderson that a graph is -vertex-connected if and only if it has a -dimensional convex -embedding in general position, for some of of its vertices, and that if it is -vertex-connected then such an embedding can be constructed in polynomial time by choosing to be any subset of vertices, and solving Tutte's system of linear equations.
One-dimensional convex embeddings (in general position), for a specified set of two vertices, are equivalent to bipolar orientations of the given graph. |
https://en.wikipedia.org/wiki/Kr00k | Kr00k (also written as KrØØk) is a security vulnerability that allows some WPA2 encrypted WiFi traffic to be decrypted. The vulnerability was originally discovered by security company ESET in 2019 and assigned on August 17, 2019. ESET estimates that this vulnerability affects over a billion devices.
Discovery
Kr00k was discovered by ESET Experimental Research and Detection Team, most prominently ESET security researcher Miloš Čermák.
It was named Kr00k by Robert Lipovsky and Stefan Svorencik. It was discovered when trying variations of the KRACK attack.
Initially found in chips made by Broadcom and Cypress, similar vulnerabilities have been found in other implementations, including those by Qualcomm and MediaTek.
Patches
The vulnerability is known to be patched in:
iOS 13.2 and iPadOS 13.2 - October 28, 2019
macOS Catalina 10.15.1, Security Update 2019–001, and Security Update 2019-006 - October 29, 2019
Vulnerable devices
During their research, ESET confirmed over a dozen popular devices were vulnerable.
Cisco has found several of their devices to be vulnerable and are working on patches. They are tracking the issue with advisory id cisco-sa-20200226-wi-fi-info-disclosure.
Known vulnerable devices include:
Amazon Echo 2nd gen
Amazon Kindle 8th gen
Apple iPad mini 2
Apple iPhone 6, 6S, 8, XR
Apple MacBook Air Retina 13-inch 2018
Asus wireless routers (RT-AC1200G+, RT-AC68U), but fixed in firmware Version 3.0.0.4.382.5161220 during March 2020
Google Nexus 5
Google Nexus 6
Google Nexus 6P
Raspberry Pi 3
Samsung Galaxy S4 (GT-I9505)
Samsung Galaxy S8
Xiaomi Redmi 3S |
https://en.wikipedia.org/wiki/Joan%20A.%20Steitz | Joan Elaine Argetsinger Steitz (born January 26, 1941) is Sterling Professor of Molecular Biophysics and Biochemistry at Yale University and Investigator at the Howard Hughes Medical Institute. She is known for her discoveries involving RNA, including ground-breaking insights into how ribosomes interact with messenger RNA by complementary base pairing and that introns are spliced by small nuclear ribonucleic proteins (snRNPs), which occur in eukaryotes. In September 2018, Steitz won the Lasker-Koshland Award for Special Achievement in Medical Science. The Lasker award is often referred to as the 'American Nobel' because 87 of the former recipients have gone on to win Nobel prizes.
Early life and education
Steitz was born in Minneapolis, Minnesota. She grew up in Minnesota in the 1950s and 60s and attended the then all-girls Northrop Collegiate School for high school.
In 1963, Steitz received her Bachelor of Science degree in chemistry from Antioch College, Ohio, where she first became interested in molecular biology at Alex Rich's Massachusetts Institute of Technology laboratory as an Antioch "coop" intern.
After completing her undergraduate degree, Steitz applied to medical school rather than graduate school since she knew of female medical doctors but not women scientists. She was accepted to Harvard Medical School, but having been excited by a summer working as a bench scientist in the laboratory of Joseph Gall at the University of Minnesota, she declined the invitation to Harvard Medical School and instead applied to Harvard's new program in biochemistry and molecular biology. There, she was the first female graduate student to join the laboratory of Nobel Laureate James Watson, with whom she first worked on bacteriophage RNA.
Career
Steitz completed postdoctoral research at the Medical Research Council (MRC) Laboratory of Molecular Biology (LMB) at the University of Cambridge (UK), where she collaborated with Francis Crick, Sydney Brenner, and Mark Bret |
https://en.wikipedia.org/wiki/Cloud%20storage | Cloud storage is a model of computer data storage in which the digital data is stored in logical pools, said to be on "the cloud". The physical storage spans multiple servers (sometimes in multiple locations), and the physical environment is typically owned and managed by a hosting company. These cloud storage providers are responsible for keeping the data available and accessible, and the physical environment secured, protected, and running. People and organizations buy or lease storage capacity from the providers to store user, organization, or application data.
Cloud storage services may be accessed through a colocated cloud computing service, a web service application programming interface (API) or by applications that use the API, such as cloud desktop storage, a cloud storage gateway or Web-based content management systems.
History
Cloud computing is believed to have been invented by J. C. R. Licklider in the 1960s with his work on ARPANET to connect people and data from anywhere at any time.
In 1983, CompuServe offered its consumer users a small amount of disk space that could be used to store any files they chose to upload.
In 1994, AT&T launched PersonaLink Services, an online platform for personal and business communication and entrepreneurship. The storage was one of the first to be all web-based, and referenced in their commercials as, "you can think of our electronic meeting place as the cloud." Amazon Web Services introduced their cloud storage service Amazon S3 in 2006, and has gained widespread recognition and adoption as the storage supplier to popular services such as SmugMug, Dropbox, and Pinterest. In 2005, Box announced an online file sharing and personal cloud content management service for businesses.
Architecture
Cloud storage is based on highly virtualized infrastructure and is like broader cloud computing in terms of interfaces, near-instant elasticity and scalability, multi-tenancy, and metered resources. Cloud storage services can b |
https://en.wikipedia.org/wiki/Amplituhedron | In mathematics and theoretical physics (especially twistor string theory), an amplituhedron is a geometric structure introduced in 2013 by Nima Arkani-Hamed and Jaroslav Trnka. It enables simplified calculation of particle interactions in some quantum field theories. In planar N = 4 supersymmetric Yang–Mills theory, also equivalent to the perturbative topological B model string theory in twistor space, an amplituhedron is defined as a mathematical space known as the positive Grassmannian.
Amplituhedron theory challenges the notion that spacetime locality and unitarity are necessary components of a model of particle interactions. Instead, they are treated as properties that emerge from an underlying phenomenon.
The connection between the amplituhedron and scattering amplitudes is a conjecture that has passed many non-trivial checks, including an understanding of how locality and unitarity arise as consequences of positivity. Research has been led by Nima Arkani-Hamed. Edward Witten described the work as "very unexpected" and said that "it is difficult to guess what will happen or what the lessons will turn out to be".
Description
When subatomic particles interact, different outcomes are possible. The evolution of the various possibilities is called a "tree" and the probability amplitude of a given outcome is called its scattering amplitude. According to the principle of unitarity, the sum of the probabilities (the squared moduli of the probability amplitudes) for every possible outcome is 1.
The on-shell scattering process "tree" may be described by a positive Grassmannian, a structure in algebraic geometry analogous to a convex polytope, that generalizes the idea of a simplex in projective space. A polytope is the n-dimensional analogue of a 3-dimensional polyhedron, the values being calculated in this case are scattering amplitudes, and so the object is called an amplituhedron.
Using twistor theory, Britto–Cachazo–Feng–Witten recursion (BCFW recursion) relati |
https://en.wikipedia.org/wiki/999%20phone%20charging%20myth | The 999 phone charging myth is an urban legend that claims that if a mobile phone has low battery, then dialling 999 (or any regional emergency telephone number) charges the phone so it has more power. This was confirmed as untrue by several British police forces who publicly cited the dangers of making such calls.
Basis
The basis for the belief was a feature of BlackBerry phones: if the battery level was too low, the phone automatically locked down phone features and shut down the phone radio for all calls except to emergency services. People discovered that if they dialled 999 then immediately hung up, it would override the shutdown for several minutes so that phone calls could be made. The belief seems to have originated in BlackBerry forums around 2012.
A related belief arose in 2015 that telling Siri on an iPhone to "Charge my phone to 100%" would cause the phone to call emergency services as a secret safety code. This was later traced to a bug in Apple programming that was fixed within a day. The myth continued to spread on social media as a prank.
Response
In 2013, Derbyshire Constabulary released a press release telling people not to believe the claim that calling 999 charges the battery. They cited that for every silent or aborted 999 call received, the operators have to call the person back to make sure there is no emergency. These silent calls waste police time that could potentially block responses to real emergencies. Bedfordshire Police also released information asking people not to call 999 except for an emergency as they stated that in the last six months of 2013 they had an increase in hoax 999 calls from people believing the urban legend. Other sources supplemented these press releases by stating that misusing the 999 number is illegal. They also stated that the police could cut off telephones being used to abuse the 999 service. |
https://en.wikipedia.org/wiki/The%20Inner%20World%20of%20Aphasia | The Inner World of Aphasia is a 1968 medical training film by co-directors Edward Feil and Naomi Feil of Edward Feil Productions for case w. It portrays patients with aphasia and follows their experiences through their recoveries. Notable for its innovative artistic direction and empathetic approach to patients with aphasia, the film was added to the National Film Registry in 2015.
Plot
The film follows Marge Nelson (portrayed by Naomi Feil), a nurse who treats patients with aphasia. Marge is overworked and unsympathetic towards her patients, experiencing frustration at their behavior and apparent inability to communicate.
Upon having an accident and sustaining brain damage, Marge herself experiences aphasia and undergoes treatment at the same hospital where she previously worked. While struggling to deal with the emotions around her trauma and new issues with processing information and communicating, Marge also discovers a startling lack of empathy from hospital staff and therapists treating her.
Marge finds a new friend in a second patient who has aphasia. The film shows us that, like Marge, this second patient has also experienced a lack of empathy from others, except in his case it is his family members who are unable to understand his illness.
Through this new friendship and her own determination, Marge changes the emotional direction of her life and takes the first steps towards her recovery.
Reception
The Inner World of Aphasia was praised by contemporary sources for its attention to patient experience, particularly for drawing attention to the ways in which patients treated for aphasia in the 1960s were dehumanized by hospital staff and family members.
Compared to other media portrayals of aphasia the time, The Inner World of Aphasia was unique for its focus on portraying "the complex subjective feel of what it is like to be an aphasic." Its use of flashback, optical distortions, and audio editing techniques were noted by medical professionals as acc |
https://en.wikipedia.org/wiki/Crossed%20module | In mathematics, and especially in homotopy theory, a crossed module consists of groups and , where acts on by automorphisms (which we will write on the left, , and a homomorphism of groups
that is equivariant with respect to the conjugation action of on itself:
and also satisfies the so-called Peiffer identity:
Origin
The first mention of the second identity for a crossed module seems to be in footnote 25 on p. 422 of J. H. C. Whitehead's 1941 paper cited below, while the term 'crossed module' is introduced in his 1946 paper cited below. These ideas were well worked up in his 1949 paper 'Combinatorial homotopy II', which also introduced the important idea of a free crossed module. Whitehead's ideas on crossed modules and their applications are developed and explained in the book by Brown, Higgins, Sivera listed below. Some generalisations of the idea of crossed module are explained in the paper of Janelidze.
Examples
Let be a normal subgroup of a group . Then, the inclusion
is a crossed module with the conjugation action of on .
For any group G, modules over the group ring are crossed G-modules with d = 0.
For any group H, the homomorphism from H to Aut(H) sending any element of H to the corresponding inner automorphism is a crossed module.
Given any central extension of groups
the surjective homomorphism
together with the action of on defines a crossed module. Thus, central extensions can be seen as special crossed modules. Conversely, a crossed module with surjective boundary defines a central extension.
If (X,A,x) is a pointed pair of topological spaces (i.e. is a subspace of , and is a point in ), then the homotopy boundary
from the second relative homotopy group to the fundamental group, may be given the structure of crossed module. The functor
satisfies a form of the van Kampen theorem, in that it preserves certain colimits.
The result on the crossed module of a pair can also be phrased as: if
is a pointed fibration |
https://en.wikipedia.org/wiki/Half%20sphere%20exposure | Half Sphere exposure (HSE) is a protein solvent exposure measure that was first introduced by .
Like all solvent exposure measures it measures how buried amino acid residues are in a protein. It is found by counting the number of amino acid neighbors within two half spheres of chosen radius around the amino acid. The calculation of HSE is found by dividing a contact number (CN) sphere in two halves by the plane perpendicular to the
Cβ-Cα vector. This simple division of the CN sphere results in
two strikingly different measures, HSE-up and HSE-down. HSE-up is defined as the number of Cα atoms in the
upper half (containing the pseudo-Cβ atom) and analogously HSE-down is defined as the number of Cα atoms
in the opposite sphere.
If only Cα atoms are available (as is the case for many simplified representations of protein structure), a related measure, called HSEα, can be used. HSEα uses a pseudo-Cβ instead of the real Cβ atom for its
calculation. The position of this pseudo-Cβ atom (pCβ) is derived from the positions of preceding
Cα−1 and the following Cα+1. The Cα-pCβ vector is calculated by adding the
Cα−1-Cα0 and Cα+1-Cα0 vectors.
HSE is used in predicting discontinuous B-cell epitopes. Song et al. have developed an online webserver termed HSEpred to predict half-sphere exposure from protein primary sequences. HSEpred server can achieve the correlation coefficients of 0.72 and 0.68 between the predicted and observed HSE-up and HSE-down measures, respectively, when evaluated on a well-prepared non-homologous protein structure dataset. Moreover, residue contact number (CN) can also be accurately predicted by HSEpred webserver using the summation of the predicted HSE-up and HSE-down values, which has further enlarged the application of this new solvent exposure measure.
Recently, Heffernan et al. has developed the most accurate predictor for both HSEα and HSEβ based on a big dataset by using multiple-step iterative deep neural-network learning. The predicted HSEa |
https://en.wikipedia.org/wiki/Kinetic-segregation%20model%20of%20T%20cell%20activation | Kinetic-segregation is a model proposed for the mechanism of T-cell receptor (TCR) triggering. It offers an explanation for how TCR binding to its ligand triggers T-cell activation, based on size-sensitivity for the molecules involved. Simon J. Davis and Anton van der Merwe, University of Oxford, proposed this model in 1996.
According to the model, TCR signalling is initiated by segregation of phosphatases with large extracellular domains from the TCR complex when binding to its ligand, allowing small kinases to phosphorylate intracellular domains of the TCR without inhibition. Its might also be applicable to other receptors of the Non-catalytic tyrosine-phosphorylated receptors family such as CD28.
Mechanism
On plasma membrane of a T cell there is the T-cell receptor (consists of α,β chains and multiple CD3 adaptor proteins), as well as molecules that induce signalling (tyrosine kinase Lck phosphorylates ITAMs in CD3 complex) and factors that inhibit signalling (tyrosine phosphatases CD45 and CD148). In the resting T-cell, all molecules are repeatedly colliding by means of diffusion.
The TCR/CD3 complex is constantly being phosphorylated by Lck. Because of an abundance of CD45 and CD148 in the cell membrane, phosphorylations are readily removed before they can recruit downstream signalling molecules. Overall phosphorylation of the TCR is low and tonic TCR signalling is avoided.
The TCR/peptide-MHC complex, formed when a T cell recognises its ligand on an antigen presenting cell (APC) and the T-cell-APC contact occurs, spans a short length. This results in the formation of close contact zones between the membranes of the T cell and antigen presenting cell (~15 nm apart) around the TCR/peptide-MHC complex. Phosphatases CD45 and CD148 with much larger ectodomains than TCR are sterically excluded from the close contact zones, while the region is still accessible for the small kinase Lck. This perturbs the balance of kinase activity to phosphatase activity and ITAM |
https://en.wikipedia.org/wiki/Eberlein%E2%80%93%C5%A0mulian%20theorem | In the mathematical field of functional analysis, the Eberlein–Šmulian theorem (named after William Frederick Eberlein and Witold Lwowitsch Schmulian) is a result that relates three different kinds of weak compactness in a Banach space.
Statement
Eberlein–Šmulian theorem: If X is a Banach space and A is a subset of X, then the following statements are equivalent:
each sequence of elements of A has a subsequence that is weakly convergent in X
each sequence of elements of A has a weak cluster point in X
the weak closure of A is weakly compact.
A set A can be weakly compact in three different ways:
Sequential compactness: Every sequence from A has a convergent subsequence whose limit is in A.
Limit point compactness: Every infinite subset of A has a limit point in A.
Compactness (or Heine-Borel compactness): Every open cover of A admits a finite subcover.
The Eberlein–Šmulian theorem states that the three are equivalent on a weak topology of a Banach space.
While this equivalence is true in general for a metric space, the weak topology is not metrizable in infinite dimensional vector spaces, and so the Eberlein–Šmulian theorem is needed.
Applications
The Eberlein–Šmulian theorem is important in the theory of PDEs, and particularly in Sobolev spaces. Many Sobolev spaces are reflexive Banach spaces and therefore bounded subsets are weakly precompact by Alaoglu's theorem. Thus the theorem implies that bounded subsets are weakly sequentially precompact, and therefore from every bounded sequence of elements of that space it is possible to extract a subsequence which is weakly converging in the space. Since many PDEs only have solutions in the weak sense, this theorem is an important step in deciding which spaces of weak solutions to use in solving a PDE.
See also
Banach–Alaoglu theorem
Bishop–Phelps theorem
Mazur's lemma
James' theorem
Goldstine theorem |
https://en.wikipedia.org/wiki/Operational%20taxonomic%20unit | An operational taxonomic unit (OTU) is an operational definition used to classify groups of closely related individuals. The term was originally introduced in 1963 by Robert R. Sokal and Peter H. A. Sneath in the context of numerical taxonomy, where an "operational taxonomic unit" is simply the group of organisms currently being studied. In this sense, an OTU is a pragmatic definition to group individuals by similarity, equivalent to but not necessarily in line with classical Linnaean taxonomy or modern evolutionary taxonomy.
Nowadays, however, the term "OTU" is commonly used in a different context and refers to clusters of (uncultivated or unknown) organisms, grouped by DNA sequence similarity of a specific taxonomic marker gene (originally coined as mOTU; molecular OTU). In other words, OTUs are pragmatic proxies for "species" (microbial or metazoan) at different taxonomic levels, in the absence of traditional systems of biological classification as are available for macroscopic organisms. For several years, OTUs have been the most commonly used units of diversity, especially when analysing small subunit 16S (for prokaryotes) or 18S rRNA (for eukaryotes) marker gene sequence datasets.
Sequences can be clustered according to their similarity to one another, and operational taxonomic units are defined based on the similarity threshold (usually 97% similarity; however also 100% similarity is common, also known as single variants) set by the researcher. It remains debatable how well this commonly-used method recapitulates true microbial species phylogeny or ecology. Although OTUs can be calculated differently when using different algorithms or thresholds, recent research by Schmidt et al. (2014) demonstrated that microbial OTUs were generally ecologically consistent across habitats and several OTU clustering approaches. The number of OTUs defined may be inflated due to errors in DNA sequencing.
OTU clustering approaches
There are three main approaches to clusterin |
https://en.wikipedia.org/wiki/Dodecahedral%20molecular%20geometry | In chemistry, the dodecahedral molecular geometry describes the shape of compounds where eight atoms or groups of atoms or ligands are arranged around a central atom defining the vertices of a snub disphenoid (also known as a trigonal dodecahedron). This shape has D2d symmetry and is one of the three common shapes for octacoordinate transition metal complexes, along with the square antiprism and the bicapped trigonal prism.
One example of the dodecahedral molecular geometry is the ion. |
https://en.wikipedia.org/wiki/Construction%20and%20Analysis%20of%20Distributed%20Processes | CADP (Construction and Analysis of Distributed Processes) is a toolbox for the design of communication protocols and distributed systems. CADP is developed by the CONVECS team (formerly by the VASY team) at INRIA Rhone-Alpes and connected to various complementary tools. CADP is maintained, regularly improved, and used in many industrial projects.
The purpose of the CADP toolkit is to facilitate the design of reliable systems by use of formal description techniques together with software tools for simulation, rapid application development, verification, and test generation.
CADP can be applied to any system that comprises asynchronous concurrency, i.e., any system whose behavior can be modeled as a set of parallel processes governed by interleaving semantics. Therefore, CADP can be used to design hardware architecture, distributed algorithms, telecommunications protocols, etc.
The enumerative verification (also known as explicit state verification) techniques implemented in CADP, though less general that theorem proving, enable an automatic, cost-efficient detection of design errors in complex systems.
CADP includes tools to support use of two approaches in formal methods, both of which are needed for reliable systems design:
Models provide mathematical representations for parallel programs and related verification problems. Examples of models are automata, networks of communicating automata, Petri nets, binary decision diagrams, boolean equation systems, etc. From a theoretical point of view, research on models seeks for general results, independent of any particular description language.
In practice, models are often too elementary to describe complex systems directly (this would be tedious and error-prone). A higher level formalism known as process algebra or process calculus is needed for this task, as well as compilers that translate high-level descriptions into models suitable for verification algorithms.
History
Work began on CADP in 1986, when the de |
https://en.wikipedia.org/wiki/Sony%20TV8-301 | The TV8-301 was a small black and white television made by Sony. It is notable for being the world's first non-projection type all-transistor television. It had an eight-inch screen. It was also portable, having a bay in the back for two 6 volt lead acid batteries. It was priced high as it was innovative in many ways, so, to the average consumer it was something of a luxury item and not a practical buy. Additionally, this television was rather prone to malfunction, which led to it being called Sony's “frail little baby". Released on the market in 1960, it was discontinued in 1962.
Technical data
Transistors: 23 (silicon and germanium)
Diodes: 15 + 2 high-voltage (17) |
https://en.wikipedia.org/wiki/Josiah%20Willard%20Gibbs%20Lectureship | The Josiah Willard Gibbs Lectureship (also called the Gibbs Lecture) of the American Mathematical Society is an annually awarded mathematical prize, named in honor of Josiah Willard Gibbs. The prize is intended not only for mathematicians, but also for physicists, chemists, biologists, physicians, and other scientists who have made important applications of mathematics. The purpose of the prize is to recognize outstanding achievement in applied mathematics and "to enable the public and the academic community to become aware of the contribution that mathematics is making to present-day thinking and to modern civilization."
The prize winner gives a lecture, which is subsequently published in the Bulletin of the American Mathematical Society.
Prize winners
See also
Colloquium Lectures (AMS)
List of mathematics awards |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.