source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Hotspot%20Ecosystem%20Research%20and%20Man%27s%20Impact%20On%20European%20Seas | Hotspot Ecosystem Research and Man's Impact On European Seas (HERMIONE) is an international multidisciplinary project, started in April 2009, that studies deep-sea ecosystems. HERMIONE scientists study the distribution of hotspot ecosystems, how they function and how they interconnect, partially in the context of how these ecosystems are being affected by climate change and impacted by humans through overfishing, resource extraction, seabed installations (oil platforms, etc.) and pollution. Major aims of the project are to understand how humans are affecting the deep-sea environment and to provide policy makers with accurate scientific information, enabling effective management strategies to protect deep sea ecosystems. The HERMIONE project is funded by the European Commission's Seventh Framework Programme, and is the successor to the HERMES project, which concluded in March 2009.
Introduction
Europe's deep-ocean margin, from the Arctic to the Iberian Margin, and across the Mediterranean to the Black Sea, spans a distance of over 15,000 km and hosts a number of diverse habitats and ecosystems. Deep water coral reefs, undersea mountains populated by a multitude of organisms, vast submarine canyon systems, and hydrothermal vents are some of the features contained therein. The traditional view of the deep-sea realm as a hostile and barren place was discredited long ago, and scientists now know that much of Europe's deep sea is rich and diverse.
However, the deep sea is increasingly threatened by humans: most of this deep-ocean frontier lies within Europe's Exclusive Economic Zone (EEZ) and has significant potential for the exploitation of biological, energy, and mineral resources. Research and exploration over the last two decades has shown clear signs of direct and indirect anthropogenic impacts in the deep sea, resulting from such activities as overfishing, littering and pollution. This raises concerns because deep-sea processes and ecosystems are not only impo |
https://en.wikipedia.org/wiki/Googol | A googol is the large number 10100. In decimal notation, it is written as the digit 1 followed by one hundred zeroes: 10,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000. Its systematic name is 10 duotrigintillion. (The short scale names are standard in the English-speaking world.) Its prime factorization is
Etymology
The term was coined in 1920 by 9-year-old Milton Sirotta (1911–1981), nephew of U.S. mathematician Edward Kasner. He may have been inspired by the contemporary comic strip character Barney Google. Kasner popularized the concept in his 1940 book Mathematics and the Imagination. Other names for this quantity include ten duotrigintillion on the short scale, ten thousand sexdecillion on the long scale, or ten sexdecilliard on the Peletier long scale.
Size
A googol has no special significance in mathematics. However, it is useful when comparing with other very large quantities such as the number of subatomic particles in the visible universe or the number of hypothetical possibilities in a chess game. Kasner used it to illustrate the difference between an unimaginably large number and infinity, and in this role it is sometimes used in teaching mathematics. To put in perspective the size of a googol, the mass of an electron, just under , can be compared to the mass of the visible universe, estimated at between and . It is a ratio in the order of about 1080 to 1090, or at most one ten-billionth of a googol (0.00000001% of a googol).
Another way of illustrating the immense size of a googol is to picture the Frontier supercomputer, which as of 2022 is the most powerful supercomputer in the world and measures 680 m2 (7,300 sq ft), almost exactly the same size of a basketball court with run-offs and sidelines. The Frontier is capable of making 1,102,000 TFLOPs (1.1 quintillion calculations per second). If the supercomputer was shrunk down to the size of an atom (for referenc |
https://en.wikipedia.org/wiki/Lobenzarit | Lobenzarit (INN) is a drug used in the treatment of arthritis. It is an immunomodulator. |
https://en.wikipedia.org/wiki/Faggot%20cell | Faggot cells are cells normally found in the hypergranular form of acute promyelocytic leukemia (FAB - M3). These promyelocytes (not blast cells) have numerous Auer rods in the cytoplasm which gives the appearance of a bundle of sticks, from which the cells are given their name.
See also
Buttock cell |
https://en.wikipedia.org/wiki/Holmgren%27s%20uniqueness%20theorem | In the theory of partial differential equations, Holmgren's uniqueness theorem, or simply Holmgren's theorem, named after the Swedish mathematician Erik Albert Holmgren (1873–1943), is a uniqueness result for linear partial differential equations with real analytic coefficients.
Simple form of Holmgren's theorem
We will use the multi-index notation:
Let ,
with standing for the nonnegative integers;
denote and
.
Holmgren's theorem in its simpler form could be stated as follows:
Assume that P = ∑|α| ≤m Aα(x)∂ is an elliptic partial differential operator with real-analytic coefficients. If Pu is real-analytic in a connected open neighborhood Ω ⊂ Rn, then u is also real-analytic.
This statement, with "analytic" replaced by "smooth", is Hermann Weyl's classical lemma on elliptic regularity:
If P is an elliptic differential operator and Pu is smooth in Ω, then u is also smooth in Ω.
This statement can be proved using Sobolev spaces.
Classical form
Let be a connected open neighborhood in , and let be an analytic hypersurface in , such that there are two open subsets and in , nonempty and connected, not intersecting nor each other, such that .
Let be a differential operator with real-analytic coefficients.
Assume that the hypersurface is noncharacteristic with respect to at every one of its points:
.
Above,
the principal symbol of .
is a conormal bundle to , defined as
.
The classical formulation of Holmgren's theorem is as follows:
Holmgren's theorem
Let be a distribution in such that in . If vanishes in , then it vanishes in an open neighborhood of .
Relation to the Cauchy–Kowalevski theorem
Consider the problem
with the Cauchy data
Assume that is real-analytic with respect to all its arguments in the neighborhood of
and that are real-analytic in the neighborhood of .
Theorem (Cauchy–Kowalevski)
There is a unique real-analytic solution in the neighborhood of .
Note that the Cauchy–Kowalevski theorem does not exclude the exist |
https://en.wikipedia.org/wiki/Premack%27s%20principle | The Premack principle, or the relativity theory of reinforcement, states that more probable behaviors will reinforce less probable behaviors.
Origin and description
The Premack principle was derived from a study of Cebus monkeys by David Premack. It was found that parameters can be understood in which the monkey operates. However, it has explanatory and predictive power when applied to humans, and it has been used by therapists practicing applied behavior analysis. The Premack principle suggests that if a person wants to perform a given activity, the person will perform a less desirable activity to get at the more desirable activity; that is, activities may themselves be reinforcers. An individual will be more motivated to perform a particular activity if they know that they will partake in a more desirable activity as a consequence. Stated objectively, if high-probability behaviors (more desirable behaviors) are made contingent upon lower-probability behaviors (less desirable behaviors), then the lower-probability behaviors are more likely to occur. More desirable behaviors are those that individuals spend more time doing if permitted; less desirable behaviors are those that individuals spend less time doing when free to act. Just as "reward" was commonly used to alter behavior long before "reinforcement" was studied experimentally, the Premack principle has long been informally understood and used in a wide variety of circumstances. An example is a mother who says, "You have to finish your vegetables (low frequency) before you can eat any ice cream (high frequency)."
Experimental evidence
David Premack and his colleagues, and others have conducted several experiments to test the effectiveness of the Premack principle in humans. One of the earliest studies was conducted with young children. Premack gave the children two response alternatives, eating candy or playing a pinball machine, and determined which of these behaviors was more probable for each child. So |
https://en.wikipedia.org/wiki/Radiogenomics | The term radiogenomics is used in two contexts: either to refer to the study of genetic variation associated with response to radiation (radiation genomics) or to refer to the correlation between cancer imaging features and gene expression (imaging genomics).
Radiation genomics
In radiation genomics, radiogenomics is used to refer to the study of genetic variation associated with response to radiation therapy. Genetic variation, such as single nucleotide polymorphisms, is studied in relation to a cancer patient's risk of developing toxicity following radiation therapy. It is also used in the context of studying the genomics of tumor response to radiation therapy.
The term radiogenomics was coined in 2002 by Andreassen et al. (2002) as an analogy to pharmacogenomics, which studies the genetic variation associated with drug responses. See also West et al. (2005) and Bentzen (2006).
The Radiogenomics Consortium
In 2009, a Radiogenomics Consortium (RGC) was established to facilitate and promote multi-centre collaboration of researchers linking genetic variants with response to radiation therapy. The Radiogenomics Consortium (http://epi.grants.cancer.gov/radiogenomics/) is a Cancer Epidemiology Consortium supported by the Epidemiology and Genetics Research Program of the National Cancer Institute of the National Institutes of Health. RGC researchers have completed numerous clinical studies that identified genetic variants associated with radiation toxicities in patients with prostate, breast, lung, head and neck, and other cancers.
Past meetings
2009 - Manchester, UK. Consortium proposed.
2010 - New York, USA.
2011 - London, UK.
2012 - Boston, USA.
2013 - Cambridge (also REQUITE launch), UK.
2014 - Heidelberg, Germany.
2015 - Montpellier, France.
2016 - Maastricht, Netherlands.
2017 - Barcelona, Spain.
2018 - Manchester, UK.
2019 - Rochester, USA.
2020 - Online.
2021 - Online.
2022 - Groningen, Netherlands.
2023 - Manchester, UK.
|
https://en.wikipedia.org/wiki/Chikungunya | Chikungunya is an infection caused by the Chikungunya virus (CHIKV). Symptoms include fever and joint pains. These typically occur two to twelve days after exposure. Other symptoms may include headache, muscle pain, joint swelling, and a rash. Symptoms usually improve within a week; however, occasionally the joint pain may last for months or years. The risk of death is around 1 in 1,000. The very young, old, and those with other health problems are at risk of more severe disease.
The virus is spread between people by two types of mosquitos: Aedes albopictus and Aedes aegypti. They mainly bite during the day. The virus may circulate within a number of animals including birds and rodents. Diagnosis is by either testing the blood for the virus's RNA or antibodies to the virus. The symptoms can be mistaken for those of dengue fever and Zika fever. It is believed most people become immune after a single infection.
The best means of prevention is overall mosquito control and the avoidance of bites in areas where the disease is common. This may be partly achieved by decreasing mosquito access to water and with the use of insect repellent and mosquito nets. There is no vaccine and no specific treatment as of 2016. Recommendations include rest, fluids, and medications to help with fever and joint pain.
While the disease typically occurs in Africa and Asia, outbreaks have been reported in Europe and the Americas since the 2000s. In 2014 more than a million suspected cases occurred. In 2014 it was occurring in Florida in the continental United States but as of 2016 there were no further locally acquired cases. The disease was first identified in 1952 in Tanzania. The term is from the Kimakonde language and means "to become contorted".
Signs and symptoms
Around 85% of people infected with Chikungunya virus experience symptoms, typically beginning with a sudden high fever above . The fever is soon followed by severe muscle and joint pain. Pain usually affects multiple joints |
https://en.wikipedia.org/wiki/Blob%20Tree | The Blob Tree is a visual tool designed to allow individuals to express themselves and their emotions in a non-verbal way through the use of human figures known as Blobs who are genderless, ageless, and culture-less.
The original Blob Tree was created in the early 1980s by Pip Wilson and Ian Long as a way of communicating with young people and adults who found reading difficult.
The Blob Tree collection consists of a set of illustrations of blob figures in various poses and expressions, each representing a different emotion or feeling. These illustrations are intended to be used as prompts for individuals to identify and express their own emotions, or as a way to start a conversation about emotions and feelings.
The Blob Tree has been adopted by professionals in a variety of fields, including counselling, therapy, education, and youth work, and in a variety of settings, including schools, hospitals, community centers, and prisons. The tool's success came from its simplicity and universality where the blob figures were easily recognizable and relatable, making it easy for individuals to connect with the illustrations and express their own emotions.
Inventors
The Blob Tree was created by Ian Long and Pip Wilson. Recognising the need for a non-verbal, universally accessible tool for emotional expression and communication, they developed the Blob Tree as a way to bridge language and cultural barriers and make emotional expression more accessible to people of different ages and backgrounds.
Pip Wilson
Pip Wilson, born in 1939, was an author and British psychologist with specialism in the field of practical training and the development of emotional intelligence. The intentions for the art according to Wilson was that they can "help facilitate and stimulate meaningful discussions about difficult issues or situations. Individuals or groups can start discussions by identifying themselves, or others, with an individual or group of blobs whose actions or feelings repres |
https://en.wikipedia.org/wiki/National%20Center%20for%20Integrative%20Biomedical%20Informatics | The National Center for Integrative Biomedical Informatics (NCIBI) is one of seven National Centers for Biomedical Computing funded by the National Institutes of Health's (NIH) Roadmap for Medical Research. The center is based at the University of Michigan and is part of the Center for Computational Medicine and Bioinformatics. NCIBI's mission is to create targeted knowledge environments for molecular biomedical research to help guide experiments and enable new insights from the analysis of complex diseases. It was established in October 2005.
The Center develops computational methods to effectively access and integrate biological data. Driving Biological Projects (DBPs) provide a starting point from which tool development is informed, launched, and tested. Current DBPs include gene fusion in cancers, major organ-specific complications of diabetes, nutrition and obesity, and co-morbid disease associations of bipolar disorder. In addition to testing tools for function, a separate team is dedicated to testing usability and user interaction.
Once tools are developed and validated, the Center disseminates data and software throughout the University and the broader biomedical research community. Various mechanisms such as training videos, tutorials, and demonstrations and presentations at prominent scientific conferences are used to share NCIBI data and software nationally and internationally.
Available tools
In addition to the tools listed below, the enriched data contained in many NCIBI databases is available from the Databases tab on the Try Our Tools page.
Exploratory analysis
Conceptual literature searching
See also
Bioinformatics
Biomedical Informatics
Notes |
https://en.wikipedia.org/wiki/Guifi.net | Guifi.net is a free, open and neutral, mostly wireless community network, with over 37,000 active nodes and about 71,000 km of wireless links (as of December 2021). The majority of these nodes are located in Catalonia and the Valencian Community, in Spain, but the network is growing in other parts of the world. The network is self-organized and operated by the users using unlicensed wireless links and open optical fiber links.
The nodes of the network are contributed by individuals, companies and administrations that freely connect to an open telecommunications network and extend the network wherever the infrastructure and content might not otherwise be accessible. Nodes join the network following the self-provision model since the whole structure is explicitly open to facilitate understanding how it is structured, so that everyone can create new sections as required. That results in a network infrastructure commons that provides abundant connectivity.
Guifi.net is supported by the Guifi.net Foundation, which has been registered as an operator with the Spanish Telecommunications Market Commission (CMT) since April 2009. In August 2009, the first deployment of optical fiber was started, known as the Fiber From The Farms (FFTF) Broadband Initiative, covering about 2 km and linking dozens of farms and farmhouses in the town of Gurb.
Since early 2011, guifi.net has been connected to the Catalonia Neutral Internet Exchange Point (CATNIX), which exchanges data with other international telecommunications operators such as Cogent Communications and Hurricane Electric. This Internet connection is used by several associations that offer their members high-speed Internet access at low cost, which other Internet service providers currently do not offer. The model is oriented to cost sharing, the compensation mechanism.
The basic principle of operation is based on the Wireless Commons License.
See also
Mesh networking
Freifunk
NYC Mesh
internet (lowercase "i"), a coll |
https://en.wikipedia.org/wiki/Mixed-species%20foraging%20flock | A mixed-species feeding flock, also termed a mixed-species foraging flock, mixed hunting party or informally bird wave, is a flock of usually insectivorous birds of different species that join each other and move together while foraging. These are different from feeding aggregations, which are congregations of several species of bird at areas of high food availability.
While it is currently unknown how mixed-species foraging flocks originate, researchers have proposed a few mechanisms for their initiation. Many believe that nuclear species play a vital role in mixed-species flock initiation. Additionally, the forest structure is hypothesized to play a vital role in these flocks' formation. In Sri Lanka, for example, vocal mimicry by the greater racket-tailed drongo might have a key role in the initiation of mixed-species foraging flocks, while in parts of the American tropics packs of foraging golden-crowned warblers might play the same role.
Composition
Mixed-species foraging flocks tend to form around a "nuclear" species. Researchers believe nuclear species both stimulate the formation of a mixed-species flock and maintain the cohesion between bird species. They tend to have a disproportionately large influence on the flock. Nuclear species have a few universal qualities. Typically, they are both generalists that employ a gleaning foraging strategy and intraspecifically social birds. "Associate" or "attendant" species are birds that trail the flock only after it has entered their territory. Researchers have shown that these species tend to have a higher fitness following mixed-species foraging flocks. The third class of birds found in mixed-species flocks have been termed "sentinel" species. Unlike nuclear species, sentinels are fly-catching birds that are rarely gregarious. Their role is to alert the other birds in the mixed-species flock to the arrival of potential predators.
Benefits
Ecologists generally assume that species in the same ecological niche comp |
https://en.wikipedia.org/wiki/Greenberger%E2%80%93Horne%E2%80%93Zeilinger%20state | In physics, in the area of quantum information theory, a Greenberger–Horne–Zeilinger state (GHZ state) is a certain type of entangled quantum state that involves at least three subsystems (particle states, qubits, or qudits). The four-particle version was first studied by Daniel Greenberger, Michael Horne and Anton Zeilinger in 1989, and the three-particle version was introduced by N. David Mermin in 1990. Extremely non-classical properties of the state have been observed. GHZ states for large numbers of qubits are theorized to give enhanced performance for metrology compared to other qubit superposition states.
Definition
The GHZ state is an entangled quantum state for 3 qubits and its state is
Generalization
The generalized GHZ state is an entangled quantum state of subsystems. If each system has dimension , i.e., the local Hilbert space is isomorphic to , then the total Hilbert space of an -partite system is . This GHZ state is also called an -partite qudit GHZ state.
Its formula as a tensor product is
.
In the case of each of the subsystems being two-dimensional, that is for a collection of M qubits, it reads
Properties
There is no standard measure of multi-partite entanglement because different, not mutually convertible, types of multi-partite entanglement exist. Nonetheless, many measures define the GHZ state to be maximally entangled state.
Another important property of the GHZ state is that taking the partial trace over one of the three systems yields
which is an unentangled mixed state. It has certain two-particle (qubit) correlations, but these are of a classical nature. On the other hand, if we were to measure one of the subsystems in such a way that the measurement distinguishes between the states 0 and 1, we will leave behind either or , which are unentangled pure states. This is unlike the W state, which leaves bipartite entanglements even when we measure one of its subsystems.
The GHZ state is non-biseparable and is the representative of o |
https://en.wikipedia.org/wiki/Pitthea%20famula | Pitthea famula is a species of moth in the family Geometridae. It was first described by Dru Drury in 1773 from Calabar, in what is now Nigeria. It is found in Angola, Benin, Cameroon, the Republic of the Congo, the Democratic Republic of the Congo, Equatorial Guinea (Bioko), Nigeria, Sierra Leone and Zambia.
Description
Upperside: antennae long and pectinated (comb like). Thorax spiral. Neck orange. Thorax and abdomen dusky grey. Anterior wings about halfway from the tips black, but at the base are of a pellucid (transparent) white; being surrounded along the anterior edge and part of the posterior with black; an oblong white spot is placed near the tips on the black part. Posterior wings black and white; the white entirely surrounded by the black, which on the anterior and abdominal edges is very narrow.
Underside: palpi orange, black at the tips. Neck, breast, and sides orange. Feet black. Thighs white. Abdomen white, annulated with dusky grey. Anterior wings as on the upperside, the black parts being of a russet hue. Posterior wings differ a little, the white part running down to the middle of the external edges, with a white spot at the upper corners. Margins of the wings entire.
Wingspan 2 inches (50 mm). |
https://en.wikipedia.org/wiki/Phenotype%20mixing | Phenotype mixing is a form of interaction between two viruses each of which holds its own unique genetic material. The two particles "share" coat proteins, therefore each has a similar assortment of identifying surface proteins, while having different genetic material.
In other words; non-genetic interaction in which virus particles released from a cell that is infected with two different viruses have components from both the infecting agents, but with a genome from one of them. |
https://en.wikipedia.org/wiki/European%20Union%20wine%20growing%20zones | European Union wine growing zones are used in the common European Union wine regulations to regulate certain aspects of winemaking. The zones differ in terms of climate and examples of what is regulated by wine growing zone are required grape maturity at harvest and allowed levels of chaptalisation. There are 21 wine producing countries in the European Union, with 14 of them having significant levels of production. In the 2004-2005 vintage, total European Union wine production was around 184 million hectoliters (4.8 billion gallons) which accounted for nearly 70% of total worldwide production. Of that total, nearly 55% was classified as table wine with 4% used in the production of grape-based distilled spirits such as Armagnac and Cognac.
The remaining 41% were produced as "quality wine"-wine that produced under one of the quality wine designation in a country's appellation systems such as Germany's QmP & QbA classifications, France's Appellation d'origine contrôlée (AOC), Spain's Denominación de Origen (DO) and Italy's Denominazione di origine controllata (DOC) systems. In 1997, the European Union had over 3.4 million hectare (8.4 million acres) planted under vine which accounted for nearly 44% of the world's wine, table and raisin grape production.
Wine zones
The wine growing zones and the wine regions that belong to them are as follows:
Zone A (the coldest), comprising Germany except Baden, Luxembourg, Belgium, The Netherlands, United Kingdom, the Čechy region of the Czech Republic and those countries in northern Europe where commercial winemaking is a very marginal business.
Zone B, comprising Baden in Germany, Austria, the French regions of Alsace, Champagne, Jura, Loire, Lorraine and Savoie and parts of South Moravian Region of the Czech Republic, Poland, Slovakia, Slovenia and Romania.
Zone C (the warmest), which is subdivided into:
Zone C I, comprising the French regions of Bordeaux, Burgundy, Provence, Rhône and Sud-Ouest, some areas in the far north |
https://en.wikipedia.org/wiki/Butyrivibrio | Butyrivibrio is a genus of bacteria in Class Clostridia. Bacteria of this genus are common in the gastrointestinal systems of many animals. Genus Butyrivibrio was first described by Bryant and Small (1956) as anaerobic, butyric acid-producing, curved rods (or vibroids). Butyrivibrio cells are small, typically 0.4 – 0.6 µm by 2 – 5 µm. They are motile, using a single polar or subpolar monotrichous flagellum. They are commonly found singly or in short chains but it is not unusual for them to form long chains. Despite historically being described as Gram-negative, their cell walls contain derivatives of teichoic acid, and electron microscopy indicates that bacteria of this genus have a Gram-positive cell wall type. It is thought that they appear Gram-negative when Gram stained because their cell walls thin to 12 to 18 nm as they reach stationary phase.
Butyrivibrio species are common in the rumens of ruminant animals such as cows, deer and sheep, where they are involved in a number of ruminal functions of agricultural importance in addition to butyrate production. These include fibre degradation, protein breakdown, biohydrogenation of lipids and the production of microbial inhibitors. Of particular importance to ruminant digestion, and therefore productivity, is their contribution to the degradation of plant structural carbohydrates, principally hemicellulose.
Butyrivibrio species are metabolically versatile and are able to ferment a wide range of sugars and cellodextrins. Some strains have been reported to break down cellulose, although their ability to sustain growth on cellulose appears to be lost during in vitro culturing. Most isolates are amylolytic and are able to degrade xylan by producing xylanolytic and esterase enzymes. The induction of xylanase enzymes varies between strains; in group D1 strains (49, H17c, 12) xylanase expression appears to be constitutively expressed, while groups B1 (113) and C (CF3) are induced only by growth on xylan, and those of gr |
https://en.wikipedia.org/wiki/Effects%20of%20high%20altitude%20on%20humans | The effects of high altitude on humans are mostly the consequences of reduced partial pressure of oxygen in the atmosphere. The medical problems that are direct consequence of high altitude are caused by the low inspired partial pressure of oxygen, which is caused by the reduced atmospheric pressure, and the constant gas fraction of oxygen in atmospheric air over the range in which humans can survive. The other major effect of altitude is due to lower ambient temperature.
The oxygen saturation of hemoglobin determines the content of oxygen in blood. After the human body reaches around above sea level, the saturation of oxyhemoglobin begins to decrease rapidly. However, the human body has both short-term and long-term adaptations to altitude that allow it to partially compensate for the lack of oxygen. There is a limit to the level of adaptation; mountaineers refer to the altitudes above as the death zone, where it is generally believed that no human body can acclimatize. At extreme altitudes, the ambient pressure can drop below the vapor pressure of water at body temperature, but at such altitudes even pure oxygen at ambient pressure cannot support human life, and a pressure suit is necessary. A rapid depressurisation to the low pressures of high altitudes can trigger altitude decompression sickness.
The physiological responses to high altitude include hyperventilation, polycythemia, increased capillary density in muscle and hypoxic pulmonary vasoconstriction–increased intracellular oxidative enzymes. There are a range of responses to hypoxia at the cellular level, shown by discovery of hypoxia-inducible factors (HIFs), which determine the general responses of the body to oxygen deprivation. Physiological functions at high altitude are not normal and evidence also shows impairment of neuropsychological function, which has been implicated in mountaineering and aviation accidents. Methods of mitigating the effects of the high altitude environment include oxygen en |
https://en.wikipedia.org/wiki/Korn%27s%20inequality | In mathematical analysis, Korn's inequality is an inequality concerning the gradient of a vector field that generalizes the following classical theorem: if the gradient of a vector field is skew-symmetric at every point, then the gradient must be equal to a constant skew-symmetric matrix. Korn's theorem is a quantitative version of this statement, which intuitively says that if the gradient of a vector field is on average not far from the space of skew-symmetric matrices, then the gradient must not be far from a particular skew-symmetric matrix. The statement that Korn's inequality generalizes thus arises as a special case of rigidity.
In (linear) elasticity theory, the symmetric part of the gradient is a measure of the strain that an elastic body experiences when it is deformed by a given vector-valued function. The inequality is therefore an important tool as an a priori estimate in linear elasticity theory.
Statement of the inequality
Let be an open, connected domain in -dimensional Euclidean space , . Let be the Sobolev space of all vector fields on that, along with their (first) weak derivatives, lie in the Lebesgue space . Denoting the partial derivative with respect to the ith component by , the norm in is given by
Then there is a constant , known as the Korn constant of , such that, for all ,
where denotes the symmetrized gradient given by
Inequality is known as Korn's inequality.
See also
Linear elasticity
Hardy inequality
Poincaré inequality |
https://en.wikipedia.org/wiki/Annals%20of%20Human%20Biology | Annals of Human Biology is a bimonthly academic journal that publishes review articles on human population biology, nature, development and causes of human variation. It is published by Taylor & Francis on behalf of the Society for the Study of Human Biology, of which it is the official journal.
Coverage Includes
Global health
Ageing
Epidemiology
Ecology
Environmental physiology
Human genetics
Auxology
Population biology
Society
Annals of Human Biology is the official journal of the Society for the Study of Human Biology
Editors-in-Chief
Noël Cameron, Olga Rickards, and Babette Zemel are the Editors-in-Chief of Annals of Human Biology.
Publication Format
Annals of Human Biology publishes six issues per year in simultaneous print and online editions. |
https://en.wikipedia.org/wiki/Attack%20patterns | In computer science, attack patterns are a group of rigorous methods for finding bugs or errors in code related to computer security.
Attack patterns are often used for testing purposes and are very important for ensuring that potential vulnerabilities are prevented. The attack patterns themselves can be used to highlight areas which need to be considered for security hardening in a software application. They also provide, either physically or in reference, the common solution pattern for preventing the attack. Such a practice can be termed defensive coding patterns.
Attack patterns define a series of repeatable steps that can be applied to simulate an attack against the security of a system.
Categories
There are several different ways to categorize attack patterns. One way is to group them into general categories, such as: Architectural, Physical, and External (see details below). Another way of categorizing attack patterns is to group them by a specific technology or type of technology (e.g. database attack patterns, web application attack patterns, network attack patterns, etc. or SQL Server attack patterns, Oracle Attack Patterns, .Net attack patterns, Java attack patterns, etc.)
Using general categories
Architectural attack patterns are used to attack flaws in the architectural design of the system. These are things like weaknesses in protocols, authentication strategies, and system modularization. These are more logic-based attacks than actual bit-manipulation attacks.
Physical attack patterns are targeted at the code itself. These are things such as SQL injection attacks, buffer overflows, race conditions, and some of the more common forms of attacks that have become popular in the news.
External attack patterns include attacks such as trojan horse attacks, viruses, and worms. These are not generally solvable by software-design approaches because they operate relatively independently from the attacked program. However, vulnerabilities in a pie |
https://en.wikipedia.org/wiki/Hemagglutinin | In molecular biology, hemagglutinins (or haemagglutinin in British English) (from the Greek , 'blood' + Latin , 'glue') are receptor-binding membrane fusion glycoproteins produced by viruses in the Paramyxoviridae family. Hemagglutinins are responsible for binding to receptors on red blood cells to initiate viral attachment and infection. The agglutination of red cells occurs when antibodies on one cell bind to those on others, causing amorphous aggregates of clumped cells.Hemagglutinins recognize cell-surface glycoconjugates containing sialic acid on the surface of host red blood cells with a low affinity, and use them to enter the endosome of host cells. In the endosome, hemagglutinins are activated at a pH of 5 - 6.5 to undergo conformational changes that enable viral attachment through a fusion peptide.
Agglutination and hemagglutinins were discovered by virologist George K. Hirst in 1941. Alfred Gottschalk proved in 1957 that hemagglutinins bind a virus to a host cell by attaching to sialic acids on carbohydrate side chains of cell-membrane glycoproteins and glycolipids.
Types
Influenza hemagglutinin or haemagglutinin: a homotrimeric glycoprotein that is found on the surface of influenza viruses which is responsible for their infectivity. Influenza strains are named for the specific hemagglutinin variant they produce, along with the specific variant of another surface protein, neuraminidase.
Measles hemagglutinin: a hemagglutinin produced by the measles virus which encodes six structural proteins, of which, hemagglutinin and fusion are surface glycoproteins involved in attachment and entry.
Parainfluenza hemagglutinin-neuraminidase: a type of hemagglutinin-neuraminidase produced by parainfluenza which is closely associated with both human and veterinary disease.
Mumps hemagglutinin-neuraminidase: a kind of hemagglutinin that the mumps virus (MuV) produces, which is the virus that causes mumps.
The PH-E form of phytohaemagglutinin.
Structure
Hemagglutini |
https://en.wikipedia.org/wiki/Five-year%20survival%20rate | The five-year survival rate is a type of survival rate for estimating the prognosis of a particular disease, normally calculated from the point of diagnosis. Lead time bias from earlier diagnosis can affect interpretation of the five-year survival rate.
There are absolute and relative survival rates, but the latter are more useful and commonly used.
Relative and absolute rates
Five-year relative survival rates are more commonly cited in cancer statistics. Five-year absolute survival rates may sometimes also be cited.
Five-year absolute survival rates describe the percentage of patients alive five years after the disease is diagnosed.
Five-year relative survival rates describe the percentage of patients with a disease alive five years after the disease is diagnosed, divided by the percentage of the general population of corresponding sex and age alive after five years. Typically, cancer five-year relative survival rates are well below 100%, reflecting excess mortality among cancer patients compared to the general population. In contrast to five-year absolute survival rates, five-year relative survival rates may also equal or even exceed 100% if cancer patients have the same or even higher survival rates than the general population. The pattern may occur if cancer patients can generally be cured, or patients diagnosed with cancer have greater socioeconomic wealth or access to medical care than the general population.
The fact that relative survival rates above 100% were estimated for some groups of patients appears counterintuitive on first view. It is unlikely that occurrence of prostate cancer would increase chances of survival, compared to the general population. A more plausible explanation is that the pattern reflects a selection effect of PSA screening, as screening tests tend to be used less often by socially disadvantaged population groups, who, in general, also have higher mortality.
Uses
Five-year survival rates can be used to compare the effectiven |
https://en.wikipedia.org/wiki/PSOS%20%28real-time%20operating%20system%29 | pSOS (Portable Software On Silicon) is a real-time operating system (RTOS), created in about 1982 by Alfred Chao, and developed and marketed for the first part of its life by his company Software Components Group (SCG). In the 1980s, pSOS rapidly became the RTOS of choice for all embedded systems based on the Motorola 68000 series family architecture, because it was written in 68000 assembly language and was highly optimised from the start. It was also modularised, with early support for OS-aware debugging, plug-in device drivers, Internet protocol suite (TCP/IP) stacks, language libraries, and disk subsystems. Later came source code level debugging, multiprocessing support, and further computer networking extensions.
In about 1991, Software Components Group was acquired by Integrated Systems Inc. (ISI) which further developed pSOS, then renamed as pSOS+, for other microprocessor families, by rewriting most of it in the programming language C. Attention was also paid to supporting successively more integrated development environments, culminating in pRISM+.
In July 1994, ISI acquired Digital Research's modular real-time multi-tasking operating system FlexOS from Novell.
In 1995, ISI offered a pSOSystem/NEST package for Novell Embedded Systems Technology (NEST).
In February 2000, ISI was acquired by Wind River Systems, the originators of the rival RTOS VxWorks. Despite initial reports that pSOS support would continue, development was halted. Wind River announced plans for a 'convergence' version of VxWorks which will support pSOS system calls, and that no further releases of pSOS will occur.
NXP Semiconductors acquired pSOS for TriMedia from Wind River and continued to support this OS for the TriMedia very long instruction word (VLIW) core.
Migration away from pSOS
In March 2000, rival company Express Logic released their Evaluation Kit for pSOS+ users, designed to provide a migration path to its ThreadX RTOS.
During August 2000, MapuSoft Technologies Inc. cam |
https://en.wikipedia.org/wiki/MLOps | MLOps or ML Ops is a paradigm that aims to deploy and maintain machine learning models in production reliably and efficiently. The word is a compound of "machine learning" and the continuous development practice of DevOps in the software field. Machine learning models are tested and developed in isolated experimental systems. When an algorithm is ready to be launched, MLOps is practiced between Data Scientists, DevOps, and Machine Learning engineers to transition the algorithm to production systems. Similar to DevOps or DataOps approaches, MLOps seeks to increase automation and improve the quality of production models, while also focusing on business and regulatory requirements. While MLOps started as a set of best practices, it is slowly evolving into an independent approach to ML lifecycle management. MLOps applies to the entire lifecycle - from integrating with model generation (software development lifecycle, continuous integration/continuous delivery), orchestration, and deployment, to health, diagnostics, governance, and business metrics. According to Gartner, MLOps is a subset of ModelOps. MLOps is focused on the operationalization of ML models, while ModelOps covers the operationalization of all types of AI models.
Definition
MLOps is a paradigm, including aspects like best practices, sets of concepts, as well as a development culture when it comes to the end-to-end conceptualization, implementation, monitoring, deployment, and scalability of machine learning products. Most of all, it is an engineering practice that leverages three contributing disciplines: machine learning, software engineering (especially DevOps), and data engineering. MLOps is aimed at productionizing machine learning systems by bridging the gap between development (Dev) and operations (Ops). Essentially, MLOps aims to facilitate the creation of machine learning products by leveraging these principles: CI/CD automation, workflow orchestration, reproducibility; versioning of data, model, |
https://en.wikipedia.org/wiki/Tucktonia | Tucktonia was a late 1970s theme park located on Stour Road, Christchurch, Dorset, England. It was officially opened on 23 May 1976 by Arthur Askey. It originally occupied of the Tuckton Park Leisure Complex. The park was closed down in 1986. The site has since been redeveloped for residential use.
General information
The park was best known for its large model village layout, which included a representation of London.
The narrow gauge ride-on steam train and some additional fixtures and fittings, were moved to the Moors Valley Railway in the Moors Valley Country Park near Ringwood, Hampshire.
It is rumoured that just prior to the closure of the park, the owners wanted to build a roller coaster at the rear of the site, but were refused permission by the local council. Following the refusal, the park closed shortly afterwards.
Bekonscot Model Village in Buckinghamshire provided much inspiration for the designers, one of whom had his office near Bekonscot and was a frequent visitor. The bulk of the models were built by KLF Ltd, who later went on to design similar models at Britannia Park in Derbyshire (known as "The American Adventure" theme park by the time of its demise in 2006). Tucktonia was the brainchild of former double British Formula 3 champion Harry Stiller who was still living in the area in 2006, before his death in 2018.
Some sources claim that the miniature London landmarks were destroyed during the making of the 1985 alien invasion movie Lifeforce, and that the entire model village was destroyed when the park closed. However, other sources claim that, while the model village was used as the miniature set for the filming of Lifeforce, the visual effects crew actually used their own destructible miniatures in the Tucktonia streets. The model of Buckingham Palace survives; this was acquired, restored and put on display at the Wimborne Model Town, Wimborne Minster in 2002, and moved to Merrivale Model Village, Great Yarmouth in 2006.
There are nu |
https://en.wikipedia.org/wiki/Deep%20palmar%20arch | The deep palmar arch (deep volar arch) is an arterial network found in the palm. It is usually primarily formed from the terminal part of the radial artery. The ulnar artery also contributes through an anastomosis. This is in contrast to the superficial palmar arch, which is formed predominantly by the ulnar artery.
Structure
The deep palmar arch is usually primarily formed from the radial artery. The ulnar artery also contributes through an anastomosis.
The deep palmar arch lies upon the bases of the metacarpal bones and on the interossei of the hand. It is deep to the oblique head of the adductor pollicis muscle, the flexor tendons of the fingers, and the lumbricals of the hand.
Alongside of it, but running in the opposite direction—toward the radial side of the hand—is the deep branch of the ulnar nerve.
The superficial palmar arch is more distally located than the deep palmar arch. If one were to fully extend the thumb and draw a line from the distal border of the thumb across the palm, this would be the level of the superficial palmar arch (Boeckel's line). The deep palmar arch is about a finger width proximal to this. The connection between the deep and superficial palmar arterial arches is an example of anastomosis. This anastomosis can be tested for using Allen's test.
The palmar metacarpal arteries arise from the deep palmar arch.
Function
The deep palmar arch supplies the thumb and the lateral side of the index finger.
See also
Superficial palmar arch
Palmar carpal arch
Dorsal carpal arch |
https://en.wikipedia.org/wiki/Coccinelle%20%28software%29 | Coccinelle (French for ladybug) is an open-source utility for matching and transforming the source code of programs written in the C programming language.
Utility
Coccinelle was initially used to aid the evolution of the Linux kernel, providing support for changes to library application programming interfaces (APIs) such as renaming a function, adding a function argument whose value is somehow context-dependent, and reorganizing a data structure.
It can also be used to find defective programming patterns in code (i.e., pieces of code that are erroneous with high probability such as possible NULL pointer dereference) without transforming them. Then coccinelles role is close to that of static analysis tools. Examples of such use are provided by the applications of the herodotos tool, which keeps track of warnings generated by coccinelle.
Support for Coccinelle is provided by IRILL. Funding for the development has been provided by the Agence Nationale de la Recherche (France), the Danish Research Council for Technology and Production Sciences, and INRIA.
The source code of Coccinelle is licensed under the terms of version 2 of the GNU General Public License (GPL).
Semantic Patch Language
The source code to be matched or replaced is specified using a "semantic patch" syntax based on the patch syntax. The Semantic Patch Language (SmPL') pattern resembles a unified diff with C-like declarations.
Example
@@
expression lock, flags;
expression urb;
@@
spin_lock_irqsave(lock, flags);
<...
- usb_submit_urb(urb)
+ usb_submit_urb(urb, GFP_ATOMIC)
...>
spin_unlock_irqrestore(lock, flags);
@@
expression urb;
@@
- usb_submit_urb(urb)
+ usb_submit_urb(urb, GFP_KERNEL) |
https://en.wikipedia.org/wiki/CLCN4 | H(+)/Cl(-) exchange transporter 4 is a protein that in humans is encoded by the CLCN4 gene.
Function
The CLCN family of voltage-dependent chloride channel genes comprises nine members (CLCN1-7, Ka and Kb) which demonstrate quite diverse functional characteristics while sharing significant sequence homology. Chloride channel 4 has an evolutionary conserved CpG island and is conserved in both mouse and hamster. This gene is mapped in close proximity to APXL (Apical protein Xenopus laevis-like) and OA1 (Ocular albinism type I), which are both located on the human X chromosome at band p22.3. The physiological role of chloride channel 4 remains unknown but may contribute to the pathogenesis of neuronal disorders.
Clinical significance
Mutations in this gene have been linked to cases of early onset epilepsy
See also
Chloride channel |
https://en.wikipedia.org/wiki/FAM237A | FAM237A is a protein coding gene which encodes a protein of the same name. Within Homo sapiens, FAM237A is believed to be primarily expressed within the brain, with moderate heart and lesser testes expression,. FAM237A is hypothesized to act as a specific activator of receptor GPR83.
Gene
FAM237A is alternatively known as HCG1657980 and LOC200726. Homo sapiens FAM237A's sequence resides on chromosome 2’s + strand, and extends from bases 207486904 to 207514174. Homo sapiens FAM237A sequence contains 13 exons unspliced.
Transcripts
Homo sapiens FAM237A is predicted to produce six unique transcripts, of which four are spliced.
Proteins
Homo sapiens FAM237A is associated with three unnamed protein isoforms. FAM237A's most-researched isoform is 181 amino acids long, and is predicted to contain a transmembrane domain. FAM237A's second protein isoform is predicted to be 417 amino acids long; it contains a transmembrane domain and an upstream open reading frame. The last protein isoform of FAM237A is made up of 158 amino acids and contains a transmembrane domain; this isoform is predicted to localize within the membrane. Several databases, including NCBI, only recognize FAM237A's 181 amino acid isoform. Given the relative abundance of literature surrounding it, the remainder of this page's findings only discuss FAM237A's 181 amino acid isoform.
The theoretical molecular weight of this isoform is 20.56 kDA. Its theoretical isoelectric point is 8.96. Homo sapiens FAM237A amino acid composition is predicted to be relatively standard. It notably contains a repeat LFWD motif at amino acids 90 and 97.
FAM237A's transmembrane domain is generally predicted to reside on amino acids 14-32 within the protein. However, structure prediction tool Phyre2 predicts that the protein's transmembrane domain resides on amino acids 91–106.
Regulation
Three promoters of Homo sapiens FAM237A are predicted: GXP_8991091, GXP_7539237, and GXP_8991092. Of these, GXP_8991091 has the greate |
https://en.wikipedia.org/wiki/Georg%20Kreisel | Georg Kreisel FRS (September 15, 1923 – March 1, 2015) was an Austrian-born mathematical logician who studied and worked in the United Kingdom and America.
Biography
Kreisel was born in Graz and came from a Jewish background; his family sent him to the United Kingdom before the Anschluss in 1938. He studied mathematics at Trinity College, Cambridge, and then, during World War II, worked on military subjects. Kreisel never took a Ph.D., though much later, in 1962, he was awarded the Cambridge degree of Sc.D., a `higher doctorate' given on the basis of published research.
He taught at the University of Reading from 1949 until 1954 and then worked at the Institute for Advanced Study from 1955 to 1957. He returned to Reading in 1957, but then taught at Stanford University from 1958-1959. Then back at Reading for the year 1959-1960, and then the University of Paris 1960-1962. Kreisel was appointed a professor at Stanford University in 1962 and remained on the faculty there until he retired in 1985.
Kreisel worked in various areas of logic, and especially in proof theory, where he is known for his so-called "unwinding" program, whose aim was to extract constructive content from superficially non-constructive proofs.
Kreisel was elected to the Royal Society in 1966; Kreisel remained a close friend of Francis Crick whom he had met in the Royal Navy during WWII.
While a student at Cambridge, Kreisel was the student most respected by Ludwig Wittgenstein. Ray Monk writes, "In 1944--when Kreisel was still only twenty-one--Wittgenstein shocked Rush Rhees by declaring Kreisel to be the most able philosopher he had ever met who was also a mathematician."
Kreisel was also a close friend of the Anglo-Irish philosopher and novelist Iris Murdoch. They met at Cambridge in 1947 during Murdoch's year of study there. Peter Conradi reports that Murdoch transcribed Kreisel's letters into her journals over the next fifty years. According to Conradi, "For half a century she none |
https://en.wikipedia.org/wiki/Hybrid%20growth%20disorders | Hybrid growth disorders refer to reduced growth or overgrowth in an organism that is a hybrid of two different species. In some sense, it is a type of hybrid dysgenesis when the growth disorder proves deleterious, making it the opposite of heterosis or hybrid vigour.
Hybrid growth disorders may be referred to as a growth dysplasia, especially when resulting in overgrowth, although this terminology may be confusing since the term dysplasia is commonly used to imply an impending cancer. However, a hybrid growth disorder is not caused by cancer.
Hybrid growth disorders are exhibited among a variety organisms, including ligers, tigons, hybrid mice, and hybrid dwarf hamsters.
A study on hybrid mice which investigated the possible causes for hybrid growth disorders reveals genomic imprinting to have a major effect. Paternal imprinting may increase growth to maximize maternal resources allocated to his progeny, while maternal imprinting may suppress growth in favor of ensuring her own survival and equal allocation of resources between offspring. This suggests that the extent of a disorder depends on the combination of parental species and their respective sexes, as demonstrated by the Vrana study. The study concludes that hybrid growth disorders most commonly affect the heterozygous sex, as expected by Haldane's rule. This would also explain why hybrid growth disorders often appear to affect one sex more than the other.
Similarly, a study of hybrids between dwarf hamster species Phodopus campbelli and Phodopus sungorus suggests that gene imprinting causes abnormal interactions between growth-promoting and growth-repressing genes which regulate placental and embryonic growth.
See also
F1 hybrid |
https://en.wikipedia.org/wiki/WordUp%20%28program%29 | WordUp is a word processor for the Atari ST platform released by Neocept in 1988. It was one of the first word processors on the platform to offer a true what you see is what you get (WYSIWYG) display, using GDOS to work with multiple fonts and embedded graphics. Most previous word processors on the platform were either entirely text-based, like Atari's own ST Writer, or did not use GDOS and did not support multiple fonts and effects on-screen.
Overall, the program was relatively simple, similar to MacWrite. It did offer some more powerful features like the ability to generate a glossary and good control over typography. Reviews were generally positive, especially over its ability to easily perform layout and editing in rich documents, but the lack of a spell checker and the very slow printing was a notable concern in most reviews.
Description
GDOS
The Graphics Environment Manager, or GEM, formed the basis of the Atari ST's graphical user interface (GUI). GEM included a system known as GDOS, short for Graphics Device Operating System, which was designed to virtualize graphics output in the same fashion that CP/Ms BDOS allowed different input/output devices to be virtualized. This meant that graphics generated for one device could be sent to any other GDOS device, for instance, from the screen to a printer. GDOS introduced significant overhead which noticeably effected the speed of all applications on the system, not just those using it.
For performance reasons, few word processors on the ST made use of GDOS and instead called the underlying graphics routines directly. This meant they lacked the features like multiple fonts and WYSIWYG layout that would be seen on the Macintosh even in simple programs like MacWrite. Their GUIs were limited to issuing commands through the menu system and interaction using dialog boxes. Among the most popular word processors on the ST was Atari's own ST Writer, which took this to its extreme and removed any GUI at all, opening in a |
https://en.wikipedia.org/wiki/Statistics%20Online%20Computational%20Resource | The Statistics Online Computational Resource (SOCR) is an online multi-institutional research and education organization. SOCR designs, validates and broadly shares a suite of online tools for statistical computing, and interactive materials for hands-on learning and teaching concepts in data science, statistical analysis and probability theory. The SOCR resources are platform agnostic based on HTML, XML and Java, and all materials, tools and services are freely available over the Internet.
The core SOCR components include interactive distribution calculators, statistical analysis modules, tools for data modeling, graphics visualization, instructional resources, learning activities and other resources.
All SOCR resources are licensed under either the Lesser GNU Public License or CC BY; peer-reviewed, integrated internally and interoperate with independent digital libraries developed by other professional societies and scientific organizations like NSDL, Open Educational Resources, Mathematical Association of America, California Digital Library, LONI Pipeline, etc.
See also
List of statistical packages
Comparison of statistical packages
External links
SOCR University of Michigan site
SOCR UCLA site |
https://en.wikipedia.org/wiki/Scientific%20Information%20Database | Scientific Information Database (SID) of Academic Center for Education, Culture and Research () is an Iranian free accessible website for indexing academic journals and access to full text or metadata of Academic publishing.
Features
Institute for Scientific Information indexed articles
See also
List of academic databases and search engines
Notes
Citation indices
Science and technology in Iran
Official website not in Wikidata
Scholarly search services
Online databases
Bibliographic databases and indexes
Iranian websites
Iranian online encyclopedias |
https://en.wikipedia.org/wiki/Acidulant | Acidulants are chemical compounds that give a tart, sour, or acidic flavor to foods or enhance the perceived sweetness of foods. Acidulants can also function as leavening agents and emulsifiers in some kinds of processed foods. Though acidulants can lower pH they do differ from acidity regulators, which are food additives specifically intended to modify the stability of food or enzymes within it. Typical acidulants are acetic acid (e.g. in pickles) and citric acid. Many beverages, such as colas, contain phosphoric acid. Sour candies often are formulated with malic acid. Other acidulants used in food production include: fumaric acid, tartaric acid, lactic acid and gluconic acid.
See also
Food additive
List of food additives
Sour sanding |
https://en.wikipedia.org/wiki/Pinch%20%28plasma%20physics%29 | A pinch (or: Bennett pinch (after Willard Harrison Bennett), electromagnetic pinch, magnetic pinch, pinch effect, or plasma pinch.) is the compression of an electrically conducting filament by magnetic forces, or a device that does such. The conductor is usually a plasma, but could also be a solid or liquid metal. Pinches were the first type of device used for experiments in controlled nuclear fusion power.
Pinches occur naturally in electrical discharges such as lightning bolts, planetary auroras, current sheets, and solar flares.
Basic mechanism
Types
Pinches exist in nature and in laboratories. Pinches differ in their geometry and operating forces. These include:
Uncontrolled – Any time an electric current moves in large amounts (e.g., lightning, arcs, sparks, discharges) a magnetic force can pull together plasma. This can be insufficient for fusion.
Sheet pinch – An astrophysical effect, this arises from vast sheets of charged particles.
Z-pinch – The current runs down the axis, or walls, of a cylinder while the magnetic field is azimuthal
Theta pinch – The magnetic field runs down the axis of a cylinder, while the electric field is in the azimuthal direction (also called a thetatron)
Screw pinch – A combination of a Z-pinch and theta pinch (also called a stabilized Z-pinch, or θ-Z pinch)
Reversed field pinch or toroidal pinch – This is a Z-pinch arranged in the shape of a torus. The plasma has an internal magnetic field. As distance increases from the center of this ring, the magnetic field reverses direction.
Inverse pinch – An early fusion concept, this device consisted of a rod surrounded by plasma. Current traveled through the plasma and returned along the center rod. This geometry was slightly different than a z-pinch in that the conductor was in the center, not the sides.
Cylindrical pinch
Orthogonal pinch effect
Ware pinch – A pinch that occurs inside a Tokamak plasma, when particles inside the banana orbit condense together.
Magnetized Li |
https://en.wikipedia.org/wiki/Broselow%20tape | The Broselow Tape, also called the Broselow pediatric emergency tape, is a color-coded length-based tape measure that is used throughout the world for pediatric emergencies. The Broselow Tape relates a child's height as measured by the tape to their weight to provide medical instructions including medication dosages, the size of the equipment that should be used, and the level of energy when using a defibrillator. Particular to children is the need to calculate all these therapies for each child individually. In an emergency, the time required to do this detracts from valuable time needed to evaluate, initiate, and monitor patient treatment. The Broselow Tape is designed for children up to approximately 12 years of age who have a maximum weight of roughly . The Broselow Tape is recognized in most medical textbooks and publications as a standard for the emergency treatment of children.
Usage
To use the Broselow Tape effectively, the child must be lying down. Use one hand to hold the red end of the tape, so it is even with the child's head. (Remember: "red to head"). While maintaining one hand on the red portion at the top of the child's head, use your free hand to run the tape down the length of the child's body until it is even with their heels (not toes). The tape that is level with the child's heels will provide their approximate weight in kilograms and their color zone.
As the tape is not completely accurate, care is required with its use.
Accuracy
The Broselow Tape is based on the relationship between weight and length ; each color zone estimates the 50th percentile weight for length, which for practical purposes estimates the ideal body weight (IBW) for emergency dosing. Because of the recent obesity epidemic, concerns have been raised as to the accuracy of the tape to determine acceptable weights and subsequently acceptable doses of emergency medications.
The most recent version of the Broselow Tape incorporates updated length/weight zones based on the mo |
https://en.wikipedia.org/wiki/Ehud%20Tenenbaum | Ehud "Udi" Tenenbaum (; born August 29, 1979), also known as The Analyzer, is an Israeli hacker.
Biography
Tenenbaum was born in Hod HaSharon in 1979. He became famous in 1998 when he was arrested for hacking computers belonging to NASA, The Pentagon, the U.S. Air Force, the U.S. Navy, the Knesset, MIT, among other high-profile organizations. He also hacked into the computers of Palestinian groups and claimed to have destroyed the website of Hamas. To do this, Tenenbaum installed packet analyzer and trojan horse software on some of the hacked servers.
The then-US Deputy Defense Secretary John Hamre stated that the attack was "the most organized and systematic attack to date" on US military systems. The military had thought that they were witnessing sophisticated Iraqi 'information warfare'. In an effort to stop the attack, the United States government assembled agents from the FBI, the Air Force Office of Special Investigations, NASA, the US Department of Justice, the Defense Information Systems Agency, the NSA, and the CIA. The government was so worried that the warning and briefings went all the way up to the President of the United States. The investigation, code-named "Solar Sunrise," eventually snared two California teenagers. After their arrest, a subsequent probe led US investigators to Tenenbaum, who was arrested after Israeli police were given evidence of Tenenbaum's activities. Later, the FBI sent agents to Israel to question Tenenbaum.
Before he was sentenced, Tenenbaum served briefly in the Israel Defense Forces, but was released soon thereafter after he was involved in a traffic collision.
In 2001, Tenenbaum pleaded guilty, while stating that he was not attempting to infiltrate the computer systems to get a hold of secrets but rather to prove that the systems were flawed. Tenenbaum was sentenced to a year and a half in prison, of which he served only 8 months following the "Deri Law". After the attack, the FBI made a short 18 minutes training vide |
https://en.wikipedia.org/wiki/Multi-fractional%20order%20estimator | The multi-fractional order estimator (MFOE) is a straightforward, practical, and flexible alternative to the Kalman filter (KF) for tracking targets. The MFOE is focused strictly on simple and pragmatic fundamentals along with the integrity of mathematical modeling. Like the KF, the MFOE is based on the least squares method (LSM) invented by Gauss and the orthogonality principle at the center of Kalman's derivation. Optimized, the MFOE yields better accuracy than the KF and subsequent algorithms such as the extended KF and the interacting multiple model (IMM).
The MFOE is an expanded form of the LSM, which effectively includes the KF and ordinary least squares (OLS) as subsets (special cases). OLS is revolutionized in for application in econometrics. The MFOE also intersects with signal processing, estimation theory, economics, finance, statistics, and the method of moments. The MFOE offers two major advances: (1) minimizing the mean squared error (MSE) with fractions of estimated coefficients (useful in target tracking) and (2) describing the effect of deterministic OLS processing of statistical inputs (of value in econometrics)
Description
Consider equally time spaced noisy measurement samples of a target trajectory described by
where n represents both the time samples and the index; the polynomial describing the trajectory is of degree J-1; and is zero mean, stationary, white noise (not necessarily Gaussian) with variance .
Estimating x(t) at time with the MFOE is described by
where the hat (^) denotes an estimate, N is the number of samples in the data window, is the time of the desired estimate, and the data weights are
The are orthogonal polynomial coefficient estimators. (a function detailed in) projects the estimate of the polynomial coefficient to the desired estimation time . The MFOE parameter 0≤≤1 can apply a fraction of the projected coefficient estimate.
The combined terms effectively constitute a novel set of expansion functions with coef |
https://en.wikipedia.org/wiki/Metal-induced%20gap%20states | In bulk semiconductor band structure calculations, it is assumed that the crystal lattice (which features a periodic potential due to the atomic structure) of the material is infinite. When the finite size of a crystal is taken into account, the wavefunctions of electrons are altered and states that are forbidden within the bulk semiconductor gap are allowed at the surface. Similarly, when a metal is deposited onto a semiconductor (by thermal evaporation, for example), the wavefunction of an electron in the semiconductor must match that of an electron in the metal at the interface. Since the Fermi levels of the two materials must match at the interface, there exists gap states that decay deeper into the semiconductor.
Band-bending at the metal-semiconductor interface
As mentioned above, when a metal is deposited onto a semiconductor, even when the metal film as small as a single atomic layer, the Fermi levels of the metal and semiconductor must match. This pins the Fermi level in the semiconductor to a position in the bulk gap. Shown to the right is a diagram of band-bending interfaces between two different metals (high and low work functions) and two different semiconductors (n-type and p-type).
Volker Heine was one of the first to estimate the length of the tail end of metal electron states extending into the semiconductor's energy gap. He calculated the variation in surface state energy by matching wavefunctions of a free-electron metal to gapped states in an undoped semiconductor, showing that in most cases the position of the surface state energy is quite stable regardless of the metal used.
Branching point
It is somewhat crude to suggest that the metal-induced gap states (MIGS) are tail ends of metal states that leak into the semiconductor. Since the mid-gap states do exist within some depth of the semiconductor, they must be a mixture (a Fourier series) of valence and conduction band states from the bulk. The resulting positions of these states, as |
https://en.wikipedia.org/wiki/Blanet | A blanet is a member of a hypothetical class of exoplanets that directly orbit black holes.
Blanets are fundamentally similar to planets; they have enough mass to be rounded by their own gravity, but are not massive enough to start thermonuclear fusion, just like planets that orbit stars. In 2019, a team of astronomers and exoplanetologists showed that there is a safe zone around a supermassive black hole that could harbor thousands of blanets in orbit around it.
Etymology
The team led by Keiichi Wada of Kagoshima University in Japan has given this name to black hole planets. The word is a portmanteau of black hole and planet.
Formation
Blanets are suspected to form in the accretion disk that orbits a sufficiently large black hole.
In fiction
In Interstellar (2014), two of the 3 terrestrial planets orbiting supermassive black hole Garguantua are proper blanets. The other one orbits a Main-sequence star named Pantagruel. |
https://en.wikipedia.org/wiki/Endomorphism%20ring | In mathematics, the endomorphisms of an abelian group X form a ring. This ring is called the endomorphism ring of X, denoted by End(X); the set of all homomorphisms of X into itself. Addition of endomorphisms arises naturally in a pointwise manner and multiplication via endomorphism composition. Using these operations, the set of endomorphisms of an abelian group forms a (unital) ring, with the zero map as additive identity and the identity map as multiplicative identity.
The functions involved are restricted to what is defined as a homomorphism in the context, which depends upon the category of the object under consideration. The endomorphism ring consequently encodes several internal properties of the object. As the resulting object is often an algebra over some ring R, this may also be called the endomorphism algebra.
An abelian group is the same thing as a module over the ring of integers, which is the initial object in the category of rings. In a similar fashion, if R is any commutative ring, the endomorphisms of an R-module form an algebra over R by the same axioms and derivation. In particular, if R is a field, its modules M are vector spaces and the endomorphism ring of each is an algebra over the field R.
Description
Let be an abelian group and we consider the group homomorphisms from A into A. Then addition of two such homomorphisms may be defined pointwise to produce another group homomorphism. Explicitly, given two such homomorphisms f and g, the sum of f and g is the homomorphism . Under this operation End(A) is an abelian group. With the additional operation of composition of homomorphisms, End(A) is a ring with multiplicative identity. This composition is explicitly . The multiplicative identity is the identity homomorphism on A.
If the set A does not form an abelian group, then the above construction is not necessarily additive, as then the sum of two homomorphisms need not be a homomorphism. This set of endomorphisms is a canonical examp |
https://en.wikipedia.org/wiki/Ethanol%20fermentation | Ethanol fermentation, also called alcoholic fermentation, is a biological process which converts sugars such as glucose, fructose, and sucrose into cellular energy, producing ethanol and carbon dioxide as by-products. Because yeasts perform this conversion in the absence of oxygen, alcoholic fermentation is considered an anaerobic process. It also takes place in some species of fish (including goldfish and carp) where (along with lactic acid fermentation) it provides energy when oxygen is scarce.
Ethanol fermentation is the basis for alcoholic beverages, ethanol fuel and bread dough rising.
Biochemical process of fermentation of sucrose
The chemical equations below summarize the fermentation of sucrose (C12H22O11) into ethanol (C2H5OH). Alcoholic fermentation converts one mole of glucose into two moles of ethanol and two moles of carbon dioxide, producing two moles of ATP in the process.
C6H12O6 → 2 C2H5OH + 2 CO2
Sucrose is a sugar composed of a glucose linked to a fructose. In the first step of alcoholic fermentation, the enzyme invertase cleaves the glycosidic linkage between the glucose and fructose molecules.
C12H22O11 + H2O + invertase → 2 C6H12O6
Next, each glucose molecule is broken down into two pyruvate molecules in a process known as glycolysis. Glycolysis is summarized by the equation:
C6H12O6 + 2 ADP + 2 Pi + 2 NAD+ → 2 CH3COCOO− + 2 ATP + 2 NADH + 2 H2O + 2 H+
CH3COCOO− is pyruvate, and Pi is inorganic phosphate. Finally, pyruvate is converted to ethanol and CO2 in two steps, regenerating oxidized NAD+ needed for glycolysis:
1. CH3COCOO− + H+ → CH3CHO + CO2
catalyzed by pyruvate decarboxylase
2. CH3CHO + NADH + H+ → C2H5OH + NAD+
This reaction is catalyzed by alcohol dehydrogenase (ADH1 in baker's yeast).
As shown by the reaction equation, glycolysis causes the reduction of two molecules of NAD+ to NADH. Two ADP molecules are also converted to two ATP and two water molecules via substrate-level phosphorylation.
Related processes
Ferment |
https://en.wikipedia.org/wiki/Protein%20subfamily | Protein subfamily is a level of protein classification, based on their close evolutionary relationship. It is below the larger levels of protein superfamily and protein family.
Proteins typically share greater sequence and function similarities with other subfamily members than they do with members of their wider family. For example, in the Structural Classification of Proteins database classification system, members of a subfamily share the same interaction interfaces and interaction partners. These are stricter criteria than for a family, where members have similar structures, but may be more distantly related and so have different interfaces. Subfamilies are assigned by a variety of methods, including sequence similarity, motifs linked to function, or phylogenetic clade. There is no exact and consistent distinction between a subfamily and a family. The same group of proteins may sometimes be described as a family or a subfamily, depending on the context. |
https://en.wikipedia.org/wiki/Symmetry%20breaking%20and%20cortical%20rotation | Symmetry breaking in biology is the process by which uniformity is broken, or the number of points to view invariance are reduced, to generate a more structured and improbable state. Symmetry breaking is the event where symmetry along a particular axis is lost to establish a polarity. Polarity is a measure for a biological system to distinguish poles along an axis. This measure is important because it is the first step to building complexity. For example, during organismal development, one of the first steps for the embryo is to distinguish its dorsal-ventral axis. The symmetry-breaking event that occurs here will determine which end of this axis will be the ventral side, and which end will be the dorsal side. Once this distinction is made, then all the structures that are located along this axis can develop at the proper location. As an example, during human development, the embryo needs to establish where is ‘back’ and where is ‘front’ before complex structures, such as the spine and lungs, can develop in the right location (where the lungs are placed ‘in front’ of the spine). This relationship between symmetry breaking and complexity was articulated by P.W. Anderson. He speculated that increasing levels of broken symmetry in many-body systems correlates with increasing complexity and functional specialization. In a biological perspective, the more complex an organism is, the higher number of symmetry-breaking events can be found.
The importance of symmetry breaking in biology is also reflected in the fact that it's found at all scales. Symmetry breaking can be found at the macromolecular level, at the subcellular level and even at the tissues and organ level. It's also interesting to note that most asymmetry on a higher scale is a reflection of symmetry breaking on a lower scale. Cells first need to establish a polarity through a symmetry-breaking event before tissues and organs themselves can be polar. For example, one model proposes that left-right bo |
https://en.wikipedia.org/wiki/Fascin | Fascin is an actin bundling protein.
Species and tissue distribution
It is a 54-58 kilodalton monomeric actin filament bundling protein originally isolated from sea urchin egg but also found in Drosophila and vertebrates, including humans. Fascin (from the Latin for bundle) is spaced at 11 nanometre intervals along the filament. The bundles in cross section are seen to be hexagonally packed, and the longitudinal spacing is compatible with a model where fascin cross-links at alternating 4 and 5 actins. It is calcium insensitive and monomeric. Three forms of fascin are found in vertebrates: Fascin1, widely found in the nervous system and elsewhere; fascin2 found in the retinal photoreceptor cells; fascin3, which is only found in the testes.
Function
Fascin binds beta-catenin, and colocalizes with it at the leading edges and borders of epithelial and endothelial cells. The role of Fascin in regulating cytoskeletal structures for the maintenance of cell adhesion, coordinating motility and invasion through interactions with signalling pathways is an active area of research especially from the cancer biology perspective. Fascin localizes to actin-rich protrusions at the cell surface called filopodia. Recent study shows that fascin also localizes to invadopodia, membrane protrusions formed at the adherent cell surface that facilitate extracellular matrix (ECM) invasion, this provide a potential molecular mechanism for how fascin increases the invasiveness of cancer cells since fascin expression is upregulated in a spectrum of cancers. Studies have also shown that Fascin plays a major role in immune suppression. T regulatory cell adhesion to antigen presenting dendritic cell causes sequestration of Fascin-1, an actin-bundling protein essential for immunological synapse formation, and skews Fascin-1–dependent actin polarization in antigen presenting dendritic cells toward the T reg cell adhesion zone. Although it is reversible upon T regulatory cell disengagement, t |
https://en.wikipedia.org/wiki/Mass%20flux | In physics and engineering, mass flux is the rate of mass flow. Its SI units are kg m−2 s−1. The common symbols are j, J, q, Q, φ, or Φ (Greek lower or capital Phi), sometimes with subscript m to indicate mass is the flowing quantity. Mass flux can also refer to an alternate form of flux in Fick's law that includes the molecular mass, or in Darcy's law that includes the mass density.
Sometimes the defining equation for mass flux in this article is used interchangeably with the defining equation in mass flow rate. For example, Fluid Mechanics, Schaum's et al uses the definition of mass flux as the equation in the mass flow rate article.
Definition
Mathematically, mass flux is defined as the limit
where
is the mass current (flow of mass per unit time ) and is the area through which the mass flows.
For mass flux as a vector , the surface integral of it over a surface S, followed by an integral over the time duration to , gives the total amount of mass flowing through the surface in that time ():
The area required to calculate the flux is real or imaginary, flat or curved, either as a cross-sectional area or a surface.
For example, for substances passing through a filter or a membrane, the real surface is the (generally curved) surface area of the filter, macroscopically - ignoring the area spanned by the holes in the filter/membrane. The spaces would be cross-sectional areas. For liquids passing through a pipe, the area is the cross-section of the pipe, at the section considered.
The vector area is a combination of the magnitude of the area through which the mass passes through, A, and a unit vector normal to the area, . The relation is .
If the mass flux passes through the area at an angle θ to the area normal , then
where is the dot product of the unit vectors. That is, the component of mass flux passing through the surface (i.e. normal to it) is . While the component of mass flux passing tangential to the area is given by , there is no mass flux act |
https://en.wikipedia.org/wiki/List%20of%20shapes%20with%20known%20packing%20constant | The packing constant of a geometric body is the largest average density achieved by packing arrangements of congruent copies of the body. For most bodies the value of the packing constant is unknown. The following is a list of bodies in Euclidean spaces whose packing constant is known. Fejes Tóth proved that in the plane, a point symmetric body has a packing constant that is equal to its translative packing constant and its lattice packing constant. Therefore, any such body for which the lattice packing constant was previously known, such as any ellipse, consequently has a known packing constant. In addition to these bodies, the packing constants of hyperspheres in 8 and 24 dimensions are almost exactly known. |
https://en.wikipedia.org/wiki/Old%20Baldy%20%28horse%29 | Old Baldy (ca. 1852 – December 16, 1882) was the horse ridden by Union Major General George G. Meade at the Battle of Gettysburg and in many other important battles of the American Civil War.
Early life and Civil War service
Baldy was born and raised on the western frontier and at the start of the Civil War was owned by Maj. Gen. David Hunter. His name during this period is unknown. It is said that he was wounded anywhere from five to 14 times during the war, starting at the First Battle of Bull Run, where he was struck in the nose by a piece of an artillery shell. Soon after, in September 1861, he was purchased from the government by Meade in Washington, D.C., for $150 and named Baldy because of his white face.
Despite Baldy's unusual, uncomfortable pace, Meade became quite devoted to him and rode him in all of his battles through 1862 and the spring of 1863. The horse was wounded in the right hind leg at the Second Battle of Bull Run, and at Antietam, he was wounded through the neck and left for dead on the field. He survived and was treated. At Gettysburg, on July 2, 1863, Baldy was hit by a bullet that entered his stomach after passing through Meade's right trouser leg. He staggered and refused to move forward, defying all of Meade's directions. Meade commented, "Baldy is done for this time. This is the first time he has refused to go forward under fire." Baldy was sent to the rear for recuperation. In 1864, having returned to duty for the Overland Campaign and the Siege of Petersburg, he was struck in the ribs by a shell at the Weldon Railroad, and Meade decided that Old Baldy should be retired.
Retirement and death
Baldy was sent north to Philadelphia and then to the farm of Meade's staff quartermaster, Captain Sam Ringwalt, in Downingtown, Pennsylvania. He was later relocated to the Meadow Bank Farm, owned by a friend of the Meade family, where he remained for several years. He was moderately active in retirement and Meade rode the horse in several me |
https://en.wikipedia.org/wiki/Tetanolysin | Tetanolysin is a toxin produced by Clostridium tetani bacteria. Its function is unknown, but it is believed to contribute to the pathogenesis of tetanus. The other C. tetani toxin, tetanospasmin, is more definitively linked to tetanus. It is sensitive to oxygen.
Tetanolysin belongs to a family of protein toxins known as thiol-activated cytolysins, which bind to cholesterol. It is related to streptolysin O and the θ-toxin of Clostridium perfringens. Cytolysins form pores in the cytoplasmic membrane that allows for the passage of ions and other molecules into the cell. The molecular weight of tetanolysin is around 55,000 daltons. |
https://en.wikipedia.org/wiki/Panniculus%20adiposus | The panniculus adiposus is the fatty layer of the subcutaneous tissues, superficial to a deeper vestigial layer of muscle, the panniculus carnosus.
It includes structures that are considered fascia by some sources but not by others. Some examples include the fascia of Camper and the superficial cervical fascia.
A group of disorders of inflammation of this layer is called panniculitis. |
https://en.wikipedia.org/wiki/Neural%20circuit%20reconstruction | Neural circuit reconstruction is the reconstruction of the detailed circuitry of the nervous system (or a portion of the nervous system) of an animal. It is sometimes called EM reconstruction since the main method used is the electron microscope (EM). This field is a close relative of reverse engineering of human-made devices, and is part of the field of connectomics, which in turn is a sub-field of neuroanatomy.
Model systems
Some of the model systems used for circuit reconstruction are the fruit fly, the mouse, and the nematode C. elegans.
Sample preparation
The sample must be fixed, stained, and embedded in plastic.
Imaging
The sample may be cut into thin slices with a microtome, then imaged using transmission electron microscopy. Alternatively, the sample may be imaged with a scanning electron microscope, then the surface abraded using a focused ion beam, or trimmed using an in-microscope microtome. Then the sample is re-imaged, and the process repeated until the desired volume is processed.
Image processing
The first step is to align the individual images into a coherent three dimensional volume.
The volume is then annotated using one of two main methods. The first manually identifies the skeletons of each neurite. The second techniques uses computer vision software to identify voxels belonging to the same neuron, which are then corrected in the process of proofreading.
Notable examples
The connectome of C. elegans was the seminal work in this field. This circuit was obtained with great effort using manually cut sections and purely manual annotation on photographic film. For many years this was the only circuit reconstruction available.
The central brain of the fruit fly Drosophila Melanogaster was released in 2020. This data release introduced the first on-line tools to query the connectome.
Querying the connectome
Connectomes of higher organism's brains requires considerable data. For the fruit fly, for example, roughly 10 terabytes of image |
https://en.wikipedia.org/wiki/Posterior%20cortex | Posterior cortex usually means the posterior (back) part of the complete cerebral cortex and includes the occipital, parietal, and temporal cortices. In other words, the posterior cortex includes all the cerebral cortex without the frontal cortex.
In combination with specific cortical areas, 'posterior cortex' usually refers to the posterior (back) part of that cortical area. For example: the posterior parietal cortex is the posterior part of the parietal cortex and the posterior cingulate cortex is the posterior part of the cingulate cortex.
Function of the posterior cortex
The posterior cortex is the “sensory” cortex, much as the frontal cortex is the “action” cortex. The posterior cortex is responsible for encoding the sensory content (visual, auditory, and tactile) of any experience (both real and imaginary experience). The posterior cortex with the exception of the primary sensory areas (Primary visual cortex (V1), primary auditory cortex, and somatosensory cortex) was called by Christof Koch and colleagues the posterior cortical hot zone for its close association with the minimal neural substrate essential for conscious perception. |
https://en.wikipedia.org/wiki/Western%20Latin%20character%20sets%20%28computing%29 | Several 8-bit character sets (encodings) were designed for binary representation of common Western European languages (Italian, Spanish, Portuguese, French, German, Dutch, English, Danish, Swedish, Norwegian, and Icelandic), which use the Latin alphabet, a few additional letters and ones with precomposed diacritics, some punctuation, and various symbols (including some Greek letters). These character sets also happen to support many other languages such as Malay, Swahili, and Classical Latin.
This material is technically obsolete, having been functionally replaced by Unicode. However it continues to have historical interest.
Summary
The ISO-8859 series of 8-bit character sets encodes all Latin character sets used in Europe, albeit that the same code points have multiple uses that caused some difficulty (including mojibake, or garbled characters, and communication issues). The arrival of Unicode, with a unique code point for every glyph, resolved these issues.
ISO/IEC 8859-1 or Latin-1 is the most used and also defines the first 256 codepoints in Unicode.
ISO/IEC 8859-15 modifies ISO-8859-1 to fully support Estonian, Finnish and French and add the euro sign.
Windows-1252 is a superset of ISO-8859-1 that includes the printable characters from ISO/IEC 8859-15 and popular punctuation such as curved quotation marks (also known as smart quotes, such as in Microsoft Word settings and similar programs). It is common that web page tools for Windows use Windows-1252 but label the web page as using ISO-8859-1, this has been addressed in HTML5, which mandates that pages labeled as ISO-8859-1 must be interpreted as Windows-1252.
IBM CP437, being intended for English only, has very little in the way of accented letters (particularly uppercase) but has far more graphics characters than the other IBM code pages listed here and also some mathematical and Greek characters that are useful as technical symbols.
IBM CP850 has all the printable characters that ISO-8859-1 has |
https://en.wikipedia.org/wiki/Pneumolysin | Pneumolysin is a virulence factor of the Gram-positive bacteria Streptococcus pneumoniae.
It is a pore-forming toxin of 53 kDa composed of 471 amino acids. It has a range of biological activity, including the ability to lyse and interfere with the function of cells and soluble molecules of the immune system.
Released pneumolysin will aid the bacteria during colonization by facilitating adherence to the host, during invasion by damaging host cells, and during infection by interfering with the host immune response.
The presence of pneumolysin in sputum, urine, CSF and blood can be indicative of an S. pneumoniae infection. |
https://en.wikipedia.org/wiki/Bapat%E2%80%93Beg%20theorem | In probability theory, the Bapat–Beg theorem gives the joint probability distribution of order statistics of independent but not necessarily identically distributed random variables in terms of the cumulative distribution functions of the random variables. Ravindra Bapat and Beg published the theorem in 1989, though they did not offer a proof. A simple proof was offered by Hande in 1994.
Often, all elements of the sample are obtained from the same population and thus have the same probability distribution. The Bapat–Beg theorem describes the order statistics when each element of the sample is obtained from a different statistical population and therefore has its own probability distribution.
Statement
Let be independent real valued random variables with cumulative distribution functions respectively . Write for the order statistics. Then the joint probability distribution of the order statistics (with and ) is
where
is the permanent of the given block matrix. (The figures under the braces show the number of columns.)
Independent identically distributed case
In the case when the variables are independent and identically distributed with cumulative probability distribution function for all i the theorem reduces to
Remarks
No assumption of continuity of the cumulative distribution functions is needed.
If the inequalities x1 < x2 < ... < xk are not imposed, some of the inequalities "may be redundant and the probability can be evaluated after making the necessary reduction."
Complexity
Glueck et al. note that the Bapat‒Beg formula is computationally intractable, because it involves an exponential number of permanents of the size of the number of random variables. However, when the random variables have only two possible distributions, the complexity can be reduced to . Thus, in the case of two populations, the complexity is polynomial in for any fixed number of statistics . |
https://en.wikipedia.org/wiki/Newtonian%20motivations%20for%20general%20relativity | Some of the basic concepts of general relativity can be outlined outside the relativistic domain. In particular, the idea that mass–energy generates curvature in space and that curvature affects the motion of masses can be illustrated in a Newtonian setting. We use circular orbits as our prototype. This has the advantage that we know the kinetics of circular orbits. This allows us to calculate curvature of orbits in space directly and compare the results with dynamical forces.
The equivalence of gravitational and inertial mass
A unique feature of the gravitational force is that all massive objects accelerate in the same manner in a gravitational field. This is often expressed as "The gravitational mass is equal to the inertial mass." This allows us to think of gravity as a curvature of spacetime.
Test for flatness in spacetime
If initially parallel paths of two particles on nearby geodesics remain parallel within some accuracy, then spacetime is flat to within that accuracy. [Ref. 2, p. 30]
Two nearby particles in a radial gravitational field
Newtonian mechanics for circular orbits
The geodesic and field equations for circular orbits
Consider the situation in which there are two particles in nearby circular polar orbits of the Earth at radius and speed . Since the orbits are circular, the gravitational force on the particles must equal the centripetal force,
where G is the gravitational constant and is the mass of the earth.
The particles execute simple harmonic motion about the earth and with respect to each other. They are at their maximum distance from each other as they cross the equator. Their trajectories intersect at the poles.
From Newton's Law of Gravitation the separation vector can be shown to be given by the "geodesic equation"
where is the curvature of the trajectory and is the speed of light c times the time.
The curvature of the trajectory is generated by the mass of the earth . This is represented by the "field equation"
In this |
https://en.wikipedia.org/wiki/Daniel%20McCartney | Daniel McCartney (September 10, 1817 – November 15, 1887) was an American who had (what is now known as) Highly Superior Autobiographical Memory (HSAM).
McCartney was born in Westmoreland County, Pennsylvania. He was legally blind and lived with relatives throughout his lifetime. For a large part of his life, he resided in Morrow County, Ohio, before his final days in Muscatine, Iowa. He never married.
McCartney was famous for his mental ability in two specific areas, memory and calculation.
He remembered every single day in his life from the age of nine until his death. Given any specific calendar date, McCartney could, in seconds, give the day of the week, describe the weather conditions, describe what he did during the day, describe what he ate during the day, and provide details of local, regional and national events on that day. In a similar case to McCartney's, noted neurobiologist James McGaugh of the University of California, Irvine, one of the world's leading experts on human memory, reported a woman, Jill Price, with the astonishing ability to clearly remember events that happened to her decades before. McGaugh labeled this one-of-a-kind ability as Hyperthymesia (National Public Radio, 2006). McCartney's mental aptitude appears to be nearly identical to this recent case reported in McGaugh's study. McCartney, however, had an additional mental ability: mathematical computation.
McCartney could mentally compute difficult mathematical computations in seconds, and extremely difficult ones in minutes. McCartney was tested several times by panels of university mathematicians in which he was given a battery of mathematical questions. On one such examination in July 1870 in Salem, Ohio, McCartney was asked to take 89 to the sixth power, which he mentally computed in ten minutes, giving the correct answer of 496,981,290,961. On another examination he was asked to provide the cube root of 4,741,632 for which he answered correctly in three minutes-(168); and 389, |
https://en.wikipedia.org/wiki/Valence%20and%20conduction%20bands | In solid-state physics, the valence band and conduction band are the bands closest to the Fermi level, and thus determine the electrical conductivity of the solid. In nonmetals, the valence band is the highest range of electron energies in which electrons are normally present at absolute zero temperature, while the conduction band is the lowest range of vacant electronic states. On a graph of the electronic band structure of a semiconducting material, the valence band is located below the Fermi level, while the conduction band is located above it.
The distinction between the valence and conduction bands is meaningless in metals, because conduction occurs in one or more partially filled bands that take on the properties of both the valence and conduction bands.
Band gap
In semiconductors and insulators the two bands are separated by a band gap, while in conductors the bands overlap. A band gap is an energy range in a solid where no electron states can exist due to the quantization of energy. Within the concept of bands, the energy gap between the valence band and the conduction band is the band gap. Electrical conductivity of non-metals is determined by the susceptibility of electrons to be excited from the valence band to the conduction band.
Electrical conductivity
Semiconductor band structureSee electrical conduction and semiconductor for a more detailed description of band structure.
In solids, the ability of electrons to act as charge carriers depends on the availability of vacant electronic states. This allows the electrons to increase their energy (i.e., accelerate) when an electric field is applied. Similarly, holes (empty states) in the almost filled valence band also allow for conductivity.
As such, the electrical conductivity of a solid depends on its capability to flow electrons from the valence to the conduction band. Hence, in the case of a semimetal with an overlap region, the electrical conductivity is high. If there is a small band gap (Eg), |
https://en.wikipedia.org/wiki/Scale%20co-occurrence%20matrix | Scale co-occurrence matrix (SCM) is a method for image feature extraction within scale space after wavelet transformation, proposed by Wu Jun and Zhao Zhongming (Institute of Remote Sensing Application, China). In practice, we first do discrete wavelet transformation for one gray image and get sub images with different scales. Then we construct a series of scale based concurrent matrices, every matrix describing the gray level variation between two adjacent scales. Last we use selected functions (such as Harris statistical approach) to calculate measurements with SCM and do feature extraction and classification.
One basis of the method is the fact: way texture information changes from one scale to another can represent that texture in some extent thus it can be used as a criterion for feature extraction. The matrix captures the relation of features between different scales rather than the features within a single scale space, which can represent the scale property of texture better. Also, there are several experiments showing that it can get more accurate results for texture classification than the traditional texture classification.
Background
Texture can be regarded as a similarity grouping in an image. Traditional texture analysis can be divided into four major issues: feature extraction, texture discrimination, texture classification and shape from texture(to reconstruct 3D surface geometry from texture information). For tradition feature extraction, approaches are usually categorized into structural, statistical, model based and transform.
Wavelet transformation is a popular method in numerical analysis and functional analysis, which captures both frequency and location information. Gray level co-occurrence matrix provides an important basis for SCM construction.
SCM based on discrete wavelet frame transformation make use of both correlations and feature information so that it combines structural and statistical benefits.
Discrete wavelet frame (DWF)
In |
https://en.wikipedia.org/wiki/Paden%E2%80%93Kahan%20subproblems | Paden–Kahan subproblems are a set of solved geometric problems which occur frequently in inverse kinematics of common robotic manipulators. Although the set of problems is not exhaustive, it may be used to simplify inverse kinematic analysis for many industrial robots. Beyond the three classical subproblems several others have been proposed.
Simplification strategies
For a structure equation defined by the product of exponentials method, Paden–Kahan subproblems may be used to simplify and solve the inverse kinematics problem. Notably, the matrix exponentials are non-commutative.
Generally, subproblems are applied to solve for particular points in the inverse kinematics problem (e.g., the intersection of joint axes) in order to solve for joint angles.
Eliminating revolute joints
Simplification is accomplished by the principle that a rotation has no effect on a point lying on its axis. For example, if the point is on the axis of a revolute twist , its position is unaffected by the actuation of the twist. To wit:
Thus, for a structure equationwhere , and are all zero-pitch twists, applying both sides of the equation to a point which is on the axis of (but not on the axes of or ) yieldsBy the cancellation of , this yieldswhich, if and intersect, may be solved by Subproblem 2.
Norm
In some cases, the problem may also be simplified by subtracting a point from both sides of the equation and taking the norm of the result.
For example, to solvefor , where and intersect at the point , both sides of the equation may be applied to a point that is not on the axis of . Subtracting and taking the norm of both sides yields
This may be solved using Subproblem 3.
List of subproblems
Each subproblem is presented as an algorithm based on a geometric proof. Code to solve a given subproblem, which should be written to account for cases with multiple solutions or no solution, may be integrated into inverse kinematics algorithms for a wide range of robots.
Subpr |
https://en.wikipedia.org/wiki/Ragulator-Rag%20complex | The Ragulator-Rag complex is a regulator of lysosomal signalling and trafficking in eukaryotic cells, which plays an important role in regulating cell metabolism and growth in response to nutrient availability in the cell. The Ragulator-Rag Complex is composed of five LAMTOR subunits, which work to regulate MAPK and mTOR complex 1. The LAMTOR subunits form a complex with Rag GTPase and v-ATPase, which sits on the cell’s lysosomes and detects the availability of amino acids. If the Ragulator complex receives signals for low amino acid count, it will start the process of catabolizing the cell. If there is an abundance of amino acids available to the cell, the Ragulator complex will signal that the cell can continue to grow.
Ragulator proteins come in two different forms: Rag A/Rag B and Rag C/Rag D. These interact to form heterodimers with one another.
History
mTORC1 is a complex within the lysosome membrane that initiates growth when promoted by a stimulus, such as growth factors. A GTPase is a key component in cell signaling, and there were, in 2010, four RAG complexes discovered within the lysosomes of cells. In 2008, it was thought that these RAG complexes would slow down autophagy and activate cell growth by interacting with mTORC1. However, in 2010, the Ragulator was discovered. Researchers determined that the function of this Ragulator was to interact with the RAG A, B, C, and D complexes to promote cell growth. This discovery also led to the first use of the term “Rag-Ragulator” complex, because of the interaction between these two.
The amino acid level, cell growth, and other important factors are influenced by the mTOR Complex 1 pathway. On the lysosomal surface, the amino acids signal the activation of the four Rag proteins (RagA, RagB, RagC, and RagD) to translocate mTORC1 to the site of activation.
A 2014 study noted that AMPK (AMP-activated protein kinase) and mTOR play important roles in managing different metabolic programs. It was also found that |
https://en.wikipedia.org/wiki/Structure%20validation | Macromolecular structure validation is the process of evaluating reliability for 3-dimensional atomic models of large biological molecules such as proteins and nucleic acids. These models, which provide 3D coordinates for each atom in the molecule (see example in the image), come from structural biology experiments such as x-ray crystallography or nuclear magnetic resonance (NMR). The validation has three aspects: 1) checking on the validity of the thousands to millions of measurements in the experiment; 2) checking how consistent the atomic model is with those experimental data; and 3) checking consistency of the model with known physical and chemical properties.
Proteins and nucleic acids are the workhorses of biology, providing the necessary chemical reactions, structural organization, growth, mobility, reproduction, and environmental sensitivity. Essential to their biological functions are the detailed 3D structures of the molecules and the changes in those structures. To understand and control those functions, we need accurate knowledge about the models that represent those structures, including their many strong points and their occasional weaknesses.
End-users of macromolecular models include clinicians, teachers and students, as well as the structural biologists themselves, journal editors and referees, experimentalists studying the macromolecules by other techniques, and theoreticians and bioinformaticians studying more general properties of biological molecules. Their interests and requirements vary, but all benefit greatly from a global and local understanding of the reliability of the models.
Historical summary
Macromolecular crystallography was preceded by the older field of small-molecule x-ray crystallography (for structures with less than a few hundred atoms). Small-molecule diffraction data extends to much higher resolution than feasible for macromolecules, and has a very clean mathematical relationship between the data and the atomic model. The |
https://en.wikipedia.org/wiki/Polygonal%20number | In mathematics, a polygonal number is a number represented as dots or pebbles arranged in the shape of a regular polygon. The dots are thought of as alphas (units). These are one type of 2-dimensional figurate numbers.
Definition and examples
The number 10 for example, can be arranged as a triangle (see triangular number):
{|
| align="center" style="line-height: 0;" |
|}
But 10 cannot be arranged as a square. The number 9, on the other hand, can be (see square number):
{|
| align="center" style="line-height: 0;" |
|}
Some numbers, like 36, can be arranged both as a square and as a triangle (see square triangular number):
{| cellpadding="5"
|- align="center" valign="bottom"
| style="line-height: 0; display: inline-block;"|
| style="line-height: 0; display: inline-block"|
|}
By convention, 1 is the first polygonal number for any number of sides. The rule for enlarging the polygon to the next size is to extend two adjacent arms by one point and to then add the required extra sides between those points. In the following diagrams, each extra layer is shown as in red.
Triangular numbers
Square numbers
Polygons with higher numbers of sides, such as pentagons and hexagons, can also be constructed according to this rule, although the dots will no longer form a perfectly regular lattice like above.
Pentagonal numbers
Hexagonal numbers
Formula
If is the number of sides in a polygon, the formula for the th -gonal number is
or
The th -gonal number is also related to the triangular numbers as follows:
Thus:
For a given -gonal number , one can find by
and one can find by
.
Every hexagonal number is also a triangular number
Applying the formula above:
to the case of 6 sides gives:
but since:
it follows that:
This shows that the th hexagonal number is also the th triangular number . We can find every hexagonal number by simply taking the odd-numbered triangular numbers:
1, 3, 6, 10, 15, 21, 28, 36, 45, 55, 66, ...
Table of values
The first 6 v |
https://en.wikipedia.org/wiki/Museum%20of%20the%20Gorge%2C%20Ironbridge | The Museum of the Gorge, originally the Severn Warehouse, is one of the ten museums of the Ironbridge Gorge Museum Trust. It portrays the history of the Ironbridge Gorge and the surrounding area of Coalbrookdale, Shropshire, England.
History
The River Severn was a major transport route, especially before the building of the railway. Severn trows were used to bring raw materials to the forges of the Gorge and to take the finished goods away. At this time, before the management of the river by weirs, water levels in the Severn were highly seasonal. During the summer the river was too low to be navigable and so finished goods were held in warehouses until there was once again enough water for passage.
The site is at the Wharfage, just west of the village of Ironbridge. This location is also the confluence of the main manufacturing area of Coalbrookdale, and its non-navigable river, with the valley of the Severn.
Around 1840 a warehouse was constructed here for the Coalbrookdale Company, to plans by the architect Samuel Cookson. Its architectural style is highly distinctive and most unusual for a warehouse. It follows the Gothic Revival architecture recently made fashionable by Pugin and already made use of locally for St Luke's Church, Ironbridge. St Luke's is in the simple Commissioners' Gothic style, by local architect Samuel Smith of Madeley. That style is close to the original medieval Gothic and follows Pugin's ideas. The style of the warehouse owes far more to Horace Walpole's Strawberry Hill than to Pugin. The parapet of the roof is crenellated at each end and decorated with pinnacles. The Eastward, riverward face is extended with a church-like apse, flanked by two narrow towers decorated with cross-shaped arrow loops, but actually hiding chimneys. Construction is of local red brick, with yellow brick dressing. The main roof is simple and warehouse-like, comprising four tiled bays with simple gable ends. Most of the walls are blind, with only high windows i |
https://en.wikipedia.org/wiki/Cantharellus%20spectaculus | Cantharellus spectaculus is a species of fungus in the genus Cantharellus. Found in North America, it was described as new to science in 2013. |
https://en.wikipedia.org/wiki/CGHS%20model | The Callan–Giddings–Harvey–Strominger model or CGHS model in short is a toy model of general relativity in 1 spatial and 1 time dimension.
Overview
General relativity is a highly nonlinear model, and as such, its 3+1D version is usually too complicated to analyze in detail. In 3+1D and higher, propagating gravitational waves exist, but not in 2+1D or 1+1D. In 2+1D, general relativity becomes a topological field theory with no local degrees of freedom, and all 1+1D models are locally flat. However, a slightly more complicated generalization of general relativity which includes dilatons will turn the 2+1D model into one admitting mixed propagating dilaton-gravity waves, as well as making the 1+1D model geometrically nontrivial locally. The 1+1D model still does not admit any propagating gravitational (or dilaton) degrees of freedom, but with the addition of matter fields, it becomes a simplified, but still nontrivial model. With other numbers of dimensions, a dilaton-gravity coupling can always be rescaled away by a conformal rescaling of the metric, converting the Jordan frame to the Einstein frame. But not in two dimensions, because the conformal weight of the dilaton is now 0. The metric in this case is more amenable to analytical solutions than the general 3+1D case. And of course, 0+1D models cannot capture any nontrivial aspect of relativity because there is no space at all.
This class of models retains just enough complexity to include among its solutions black holes, their formation, FRW cosmological models, gravitational singularities, etc. In the quantized version of such models with matter fields, Hawking radiation also shows up, just as in higher-dimensional models.
Action
A very specific choice of couplings and interactions leads to the CGHS model.
where g is the metric tensor, is the dilaton field, fi are the matter fields, and λ2 is the cosmological constant. In particular, the cosmological constant is nonzero, and the matter fields are massless |
https://en.wikipedia.org/wiki/Jos%C3%A9%20Luis%20Massera | José Luis Massera (Genoa, Italy, June 8, 1915 – Montevideo, September 9, 2002) was a Uruguayan dissident and mathematician who researched the stability of differential equations.
Massera's lemma is named after him. He published over 40 papers during 1940–1970. A militant Communist, he was a political prisoner during 1975–1984. In the 1930s, Julio Rey Pastor gave regular weekend lectures on topology in Montevideo to a group that included Massera. Stimulated by contact with Argentine mathematics, the 1950s saw Uruguay develop a fine school in mathematics, of which Massera was very much a part.
Massera developed new notions of stability, and published several foundational papers and an influential textbook. His results in on periodic differential equations have been heavily cited and are referred to as Massera's theorem. His work in and on the converse to Lyapunov's criterion is also influential, and contain the well known Massera's lemma. His textbook is also heavily cited.
After military intervention in Uruguay in 1973, Massera was arrested on October 22, 1975 in Montevideo and was held in solitary confinement for nearly a year. During this time he was subjected to repeated torture resulting in injuries including a fractured pelvis. In October 1976 he was taken from solitary confinement, tried and convicted for "subversive association", and given a 24-year prison sentence. On June 22, 1979, as a consequence of a proposal put forward by Gaetano Fichera and unanimously approved by the whole Mathematics Faculty Council of the Sapienza University of Rome, he was awarded the laurea honoris causa while still being under conviction. He was released in 1984.
Honors
The outer main-belt asteroid 10690 Massera, discovered by American astronomer Schelte Bus at the Australian Siding Spring Observatory in 1981, was named in his memory on 13 April 2017 ().
Selected works
.
.
. |
https://en.wikipedia.org/wiki/Post-traumatic%20embitterment%20disorder | Post-traumatic embitterment disorder (PTED) is defined as a pathological reaction to a negative life event, which those affected experienced as a grave insult, humiliation, betrayal, or injustice. Prevalent emotions of PTED are embitterment, anger, fury, and hatred, especially against the triggering stressor, often accompanied by fantasies of revenge. The disorder commences immediately and without time delay at the moment of the triggering event. If left untreated, the prognosis of PTED presents as rather unfavorable, since patients find themselves trapped in a vicious circle of strong negative emotions constantly intensifying one another and eventually leading into a self-destructive downward spiral. People affected by PTED are more likely to put fantasies of revenge into action, making them a serious threat to the stressor.
The concept of PTED as a distinct clinical disorder has been first described by the German psychiatrist and psychologist Michael Linden in 2003, who remains its most involved researcher. Even though it has been backed up by empirical research in the past years, it remains disputed as to whether embitterment should be included among psychological disorders. Therefore, PTED currently does not hold its own category in the ICD-10 but is categorized under F43.8 “Other reactions to severe stress”. It cannot be categorized as an adjustment disorder under F43.2, since “ordinary” adjustment disorders normally subside within six months, while PTED is much more likely to become chronic. A condition similar to PTED has already been described by Emil Kraepelin as early as 1915 by the name querulous paranoia as a form of traumatic neuroses, explicitly demarcating it from personality disorders.
Bitterness and embitterment
Bitterness (also called resentment) is defined as a basic human reaction in response to experiences of injustice, betrayal, or humiliation, consisting of emotions such as anger, wrath, hostility, disappointment, disgust, and shame. Howeve |
https://en.wikipedia.org/wiki/Connecting%20stalk | The connecting stalk, or body stalk is an embryonic structure that is formed by the third week of development and connects the embryo to its shell of trophoblasts. The connecting stalk is derived from the extraembryonic mesoderm. Initially it lies caudally to the trilaminar germ disc, but, with subsequent embryonic folding, the body stalk assume a more ventral position. Progressive expansion of the amnion from the umbilical ring (surrounding the roots of the vitelline duct and connecting stalk) creates a tube with a covering of amniotic membrane with allantois and umbilical vessels as its content and mesoderm of the connecting stalk as the ground substance. This extraembryonic mesodermal ground substance forms the future wharton's jelly. The amniotic membrane and its contents form the umbilical cord that connects the embryo and the placenta.
The root of the connecting stalk contains the allantois as a diverticulum of hindgut endoderm along with umbilical vessels.
Anomalies are usually referred to as body stalk anomalies and occur in approximately 1 in 15,000 births. They are due to defects in the formation of the cephalic, caudal, and lateral embryonic body folds, that result in a reduced or absent umbilical cord.
Additional images |
https://en.wikipedia.org/wiki/Paleofauna%20of%20the%20Messel%20Formation | This is an overview of the of the Eocene Messel Formation as explored by the Messel Pit excavations in Germany. A former quarry and now UNESCO World Heritage Site, the Messel Formation preserves what once were a series of anoxic lakes surrounded by a sub-tropical rainforest during the Middle Eocene, approximately 47 Ma.
Sponges
Molluscs
Crustaceans
Arachnids
Araneae
Opiliones
Insects
Coleoptera
Dictyopterans
Dipterans
Hemiptera
Hymenoptera
Lepidoptera
Neuroptera
Odonata
Phasmatodea
"Fish"
Amiiformes
Anguilliformes
Lepisosteiformes
Perciformes
Thaumaturidae
Amphibians
Caudata
Anura
Squamata
Testudinata
Crocodyliformes
Birds
Palaeognathae
Anseriformes
Galliformes
Mirandornithes
Cuculiformes
Strisores
Charadriiformes
Gruimorphae
Suliformes
Pelecaniformes
Strigiformes
Coraciimorphae
Cariamiformes (?)
Several groups of Messel birds share characteristics with the modern seriemas, which has led to them being placed within the clade Cariamae in the past. However, this placement typically occurred under the assumption that they are a group within gruiformes, which has been disputed by more recent analysis. Instead more recent publications consider Cariamae (or Cariamiformes) as basal members of Australaves.
Falconiformes
Psittacopasserae
Incertae sedis
Mammals
Apatotheria
Artiodactyla
Chiroptera
Cimolesta
Eulipotyphla
Leptictida
Metatheria
Pan-Carnivora
Pholidota
Perissodactyla
Primates
Rodentia |
https://en.wikipedia.org/wiki/Minimum%20polynomial%20extrapolation | In mathematics, minimum polynomial extrapolation is a sequence transformation used for convergence acceleration of vector sequences, due to Cabay and Jackson.
While Aitken's method is the most famous, it often fails for vector sequences. An effective method for vector sequences is the minimum polynomial extrapolation. It is usually phrased in terms of the fixed point iteration:
Given iterates in , one constructs the matrix whose columns are the differences. Then, one computes the vector where denotes the Moore–Penrose pseudoinverse of . The number 1 is then appended to the end of , and the extrapolated limit is
where is the matrix whose columns are the iterates starting at 2.
The following 4 line MATLAB code segment implements the MPE algorithm:
U = x(:, 2:end - 1) - x(:, 1:end - 2);
c = - pinv(U) * (x(:, end) - x(:, end - 1));
c(end + 1, 1) = 1;
s = (x(:, 2:end) * c) / sum(c); |
https://en.wikipedia.org/wiki/Semiochemical | A semiochemical, from the Greek σημεῖον (semeion), meaning "signal", is a chemical substance or mixture released by an organism that affects the behaviors of other individuals. Semiochemical communication can be divided into two broad classes: communication between individuals of the same species (intraspecific) or communication between different species (interspecific).
It is usually used in the field of chemical ecology to encompass pheromones, allomones, kairomones, attractants and repellents.
Many insects, including parasitic insects, use semiochemicals. Pheromones are intraspecific signals that aid in finding mates, food and habitat resources, warning of enemies, and avoiding competition. Interspecific signals known as allomones and kairomones have similar functions.
In nature
Pheromone
A pheromone (from Greek phero "to bear" + hormone from Greek – "impetus") is a secreted or excreted chemical factor that triggers a social response in members of the same species. Pheromones are chemicals capable of acting outside the body of the secreting individual to impact the behavior of the receiving individual. There are alarm pheromones, food trail pheromones, sex pheromones, and many others that affect behavior or physiology. Their use among insects has been particularly well documented. In addition, some vertebrates and plants communicate by using pheromones. A notable example of pheromone usage to indicate sexual receptivity in insects can be seen in the female Dawson's burrowing bee, which uses a particular mixture of cuticular hydrocarbons to signal sexual receptivity to mating, and then another mixture to indicate sexual disinterest. These hydrocarbons, in association with other chemical signals produced in the Dufour's gland, have been implicated in male repulsion signaling as well.
The term "pheromone" was introduced by Peter Karlson and Martin Lüscher in 1959, based on the Greek word pherein (to transport) and hormone (to stimulate). They are also sometim |
https://en.wikipedia.org/wiki/Mathematical%20Biosciences%20Institute | The Mathematical Biosciences Institute (MBI) was an institution of higher learning affiliated with the Ohio State University in Columbus, Ohio. MBI received major funding from the National Science Foundation.
History
Under the leadership of founding director Avner Friedman, MBI opened in September 2002, holding its first workshop, hosting its first visiting researchers, and starting its first cohort of postdocs in that month. MBI holds 10–12 scientific workshops each year, and hosts about 25 postdoctoral and visiting researchers in residence at any given time. Through its collective events and programs, MBI draws over 1000 visits by researchers in the broadly defined area of mathematical biology throughout the year. MBI's long term planning is overseen by its Directorate and its Board of Trustees, while its scientific activities are overseen by its Directorate and its Scientific Advisory Committee.
MBI programs
Workshops
MBI organizes Emphasis Semesters consisting of three or four week-long workshops. Emphasis Semesters are organized around selected themes which have included Mathematical Neuroscience, Cancer and Its Environment, and Analysis of Complex Data in Biological Systems. Outside of Emphasis Semesters, Current Topic Workshops focus on emerging topics in the mathematical biosciences. Most of the institute's programs are conducted on The Ohio State University campus, but MBI also sponsors conferences and workshops at its academic Institute Partners. MBI accepts proposals for future programs.
Postdoctoral fellows
MBI postdoctoral fellows engage in an integrated program of tutorials, working seminars, workshops, and interactions with their mathematical and bioscience mentors. These activities are geared toward providing the tools to pursue an independent research program with an emphasis on collaborative research in the mathematical biosciences.
Long-term visitors
MBI has a program of support for visitors to spend an extended period of time in residence. D |
https://en.wikipedia.org/wiki/Brachytherapy | Brachytherapy is a form of radiation therapy where a sealed radiation source is placed inside or next to the area requiring treatment. Brachy is Greek for short. Brachytherapy is commonly used as an effective treatment for cervical, prostate, breast, esophageal and skin cancer and can also be used to treat tumours in many other body sites. Treatment results have demonstrated that the cancer-cure rates of brachytherapy are either comparable to surgery and external beam radiotherapy (EBRT) or are improved when used in combination with these techniques. Brachytherapy can be used alone or in combination with other therapies such as surgery, EBRT and chemotherapy.
Brachytherapy contrasts with unsealed source radiotherapy, in which a therapeutic radionuclide (radioisotope) is injected into the body to chemically localize to the tissue requiring destruction. It also contrasts to External Beam Radiation Therapy (EBRT), in which high-energy x-rays (or occasionally gamma-rays from a radioisotope like cobalt-60) are directed at the tumour from outside the body. Brachytherapy instead involves the precise placement of short-range radiation-sources (radioisotopes, iodine-125 or caesium-131 for instance) directly at the site of the cancerous tumour. These are enclosed in a protective capsule or wire, which allows the ionizing radiation to escape to treat and kill surrounding tissue but prevents the charge of radioisotope from moving or dissolving in body fluids. The capsule may be removed later, or (with some radioisotopes) it may be allowed to remain in place.
A feature of brachytherapy is that the irradiation affects only a very localized area around the radiation sources. Exposure to radiation of healthy tissues farther away from the sources is therefore reduced. In addition, if the patient moves or if there is any movement of the tumour within the body during treatment, the radiation sources retain their correct position in relation to the tumour. These characteristics of b |
https://en.wikipedia.org/wiki/Khurpa | A khurpa (alternatively called a khurpi) is a short handled cutting tool similar to a Trowel with a flat blade used for digging soil and weeding in small gardens or vegetable farms. It is commonly use in small farms or in ridges or rows of vegetables to hoewing or earth up the weeds. It is traditionally used while in a squatting posture. The work khurpa is a word in the Punjabi language.
The khurpa is used in Punjab (as well as in other areas in India) for small-scale gardening processes such as bed preparation, digging, tilling, and weeding. |
https://en.wikipedia.org/wiki/Berry%20paradox | The Berry paradox is a self-referential paradox arising from an expression like "The smallest positive integer not definable in under sixty letters" (a phrase with fifty-seven letters).
Bertrand Russell, the first to discuss the paradox in print, attributed it to G. G. Berry (1867–1928), a junior librarian at Oxford's Bodleian Library. Russell called Berry "the only person in Oxford who understood mathematical logic". The paradox was called "Richard's paradox" by Jean-Yves Girard.
Overview
Consider the expression:
"The smallest positive integer not definable in under sixty letters."
Since there are only twenty-six letters in the English alphabet, there are finitely many phrases of under sixty letters, and hence finitely many positive integers that are defined by phrases of under sixty letters. Since there are infinitely many positive integers, this means that there are positive integers that cannot be defined by phrases of under sixty letters. If there are positive integers that satisfy a given property, then there is a smallest positive integer that satisfies that property; therefore, there is a smallest positive integer satisfying the property "not definable in under sixty letters". This is the integer to which the above expression refers. But the above expression is only fifty-seven letters long, therefore it is definable in under sixty letters, and is not the smallest positive integer not definable in under sixty letters, and is not defined by this expression. This is a paradox: there must be an integer defined by this expression, but since the expression is self-contradictory (any integer it defines is definable in under sixty letters), there cannot be any integer defined by it.
Perhaps another helpful analogy to Berry's Paradox would be the phrase "indescribable feeling". If the feeling is indeed indescribable, then no description of the feeling would be true. But if the word "indescribable" communicates something about the feeling, then it may be cons |
https://en.wikipedia.org/wiki/Dadda%20multiplier | The Dadda multiplier is a hardware binary multiplier design invented by computer scientist Luigi Dadda in 1965. It uses a selection of full and half adders to sum the partial products in stages (the Dadda tree or Dadda reduction) until two numbers are left. The design is similar to the Wallace multiplier, but the different reduction tree reduces the required number of gates (for all but the smallest operand sizes) and makes it slightly faster (for all operand sizes).
Dadda and Wallace multipliers have the same three steps for two bit strings and of lengths and respectively:
Multiply (logical AND) each bit of , by each bit of , yielding results, grouped by weight in columns
Reduce the number of partial products by stages of full and half adders until we are left with at most two bits of each weight.
Add the final result with a conventional adder.
As with the Wallace multiplier, the multiplication products of the first step carry different weights reflecting the magnitude of the original bit values in the multiplication. For example, the product of bits has weight .
Unlike Wallace multipliers that reduce as much as possible on each layer, Dadda multipliers attempt to minimize the number of gates used, as well as input/output delay. Because of this, Dadda multipliers have a less expensive reduction phase, but the final numbers may be a few bits longer, thus requiring slightly bigger adders.
Description
To achieve a more optimal final product, the structure of the reduction process is governed by slightly more complex rules than in Wallace multipliers.
The progression of the reduction is controlled by a maximum-height sequence , defined by:
, and
This yields a sequence like so:
The initial value of is chosen as the largest value such that , where and are the number of bits in the input multiplicand and multiplier. The lesser of the two bit lengths will be the maximum height of each column of weights after the first stage of multiplication. |
https://en.wikipedia.org/wiki/Smithsonian%20Tropical%20Research%20Institute | The Smithsonian Tropical Research Institute (STRI, ) is located in Panama and is the only bureau of the Smithsonian Institution based outside of the United States. It is dedicated to understanding the past, present, and future of tropical ecosystems and their relevance to human welfare. STRI grew out of a small field station established in 1923 on Barro Colorado Island in the Panama Canal Zone to become one of the world's leading tropical research organizations. STRI's facilities provide for long-term ecological studies in the tropics and are used by some 1,200 visiting scientists from academic and research institutions around the world every year.
History
Smithsonian scientists first came to Panama during the construction of the Panama Canal from 1904 to 1914.
The Secretary of the Smithsonian Institution, Charles Doolittle Walcott, reached an agreement with Federico Boyd to conduct a biological inventory of the new Canal Zone in 1910, and this survey was subsequently extended to include all of Panama. Thanks largely to their efforts, the governor of the Canal Zone declared Barro Colorado Island (BCI) a biological reserve in 1923, making it one of the earliest biological reserves in the Americas. During the 1920s and 1930s BCI, in Gatun Lake, became an outdoor laboratory for scientists from U.S. universities and the Smithsonian Institution. By 1940, when BCI was designated the Canal Zone Biological Area (CZBA), more than 300 scientific publications had described the biota of BCI. In the Government Reorganization Act of 1946, BCI became a bureau of the Smithsonian Institution. The Smithsonian Tropical Research Institute (STRI) was created in 1966.
With the establishment of STRI, permanent staff scientists were hired and fellowship programs were initiated to support aspiring tropical biologists. The first director after the name change was Martin Humphrey Moynihan. A strong relationship with the Republic of Panama was formalized in the Panama Canal Treaties of 197 |
https://en.wikipedia.org/wiki/Recognition-primed%20decision | Recognition-primed decision (RPD) is a model of how people make quick, effective decisions when faced with complex situations. In this model, the decision maker is assumed to generate a possible course of action, compare it to the constraints imposed by the situation, and select the first course of action that is not rejected. RPD has been described in diverse groups including trauma nurses, fireground commanders, chess players, and stock market traders. It functions well in conditions of time pressure, and in which information is partial and goals poorly defined. The limitations of RPD include the need for extensive experience among decision-makers (in order to correctly recognize the salient features of a problem and model solutions) and the problem of the failure of recognition and modeling in unusual or misidentified circumstances. It appears, as discussed by Gary A. Klein in Sources of Power, to be a valid model for how human decision-makers make decisions.
Overview
The RPD model identifies a reasonable reaction as the first one that is immediately considered. RPD combines two ways of developing a decision; the first is recognizing which course of action makes sense, and the second, evaluating the course of action through imagination to see if the actions resulting from that decision make sense. However, the difference of being experienced or inexperienced plays a major factor in the decision-making processes.
RPD reveals a critical difference between experts and novices when presented with recurring situations. Experienced people will generally be able to come up with a quicker decision because the situation may match a prototypical situation they have encountered before. Novices, lacking this experience, must cycle through different possibilities, and tend to use the first course of action that they believe will work. The inexperienced also have the tendencies of using trial and error through their imagination.
Variations
There are three variations in RPD |
https://en.wikipedia.org/wiki/Multiverse%20%28set%20theory%29 | In mathematical set theory, the multiverse view is that there are many models of set theory, but no "absolute", "canonical" or "true" model. The various models are all equally valid or true, though some may be more useful or attractive than others. The opposite view is the "universe" view of set theory in which all sets are contained in some single ultimate model.
The collection of countable transitive models of ZFC (in some universe) is called the hyperverse and is very similar to the "multiverse".
A typical difference between the universe and multiverse views is the attitude to the continuum hypothesis. In the universe view the continuum hypothesis is a meaningful question that is either true or false though we have not yet been able to decide which. In the multiverse view it is meaningless to ask whether the continuum hypothesis is true or false before selecting a model of set theory. Another difference is that the statement "For every transitive model of ZFC there is a larger model of ZFC in which it is countable" is true in some versions of the multiverse view of mathematics but is false in the universe view. |
https://en.wikipedia.org/wiki/Durability | Durability is the ability of a physical product to remain functional, without requiring excessive maintenance or repair, when faced with the challenges of normal operation over its design lifetime. There are several measures of durability in use, including years of life, hours of use, and number of operational cycles. In economics, goods with a long usable life are referred to as durable goods.
Requirements for product durability
Product durability is predicated by good repairability and regenerability in conjunction with maintenance. Every durable product must be capable of adapting to technical, technological and design developments. This must be accompanied by a willingness on the part of consumers to forgo having the "very latest" version of a product.
In the United Kingdom, durability as a characteristic relating to the quality of goods that can be demanded by consumers was not clearly established until an amendment of the Sale of Goods Act 1979 relating to the quality standards for supplied goods in 1994.
Product life spans and sustainable consumption
The lifespan of household goods is a significant factor in sustainable consumption. Longer product life spans can contribute to eco-efficiency and sufficiency, thus slowing consumption in order to progress towards a sustainable level of consumption. Cooper (2005) proposed a model to demonstrate the crucial role of product lifespans to sustainable production and consumption.
Types of durability
Durability can encompass several specific physical properties of designed products, including:
Ageing (of polymers)
Dust resistance
Resistance to fatigue
Fire resistance
Radiation hardening
Thermal resistance
Rot-proofing
Rustproofing
Toughness
Waterproofing
See also
Availability
Consumables
Disposable product
Durable good
Interchangeable parts
Maintainability
Product life
Product stewardship
Throwaway society
Waste minimization |
https://en.wikipedia.org/wiki/List%20of%20clubmosses%20and%20mosses%20of%20Montana | There are at least 23 species of clubmosses and 153 species of mosses found in the state of Montana in the United States. The Montana Natural Heritage Program has identified a number of clubmoss and moss species as species of concern.
Clubmosses
Clubmosses
Class: Lycopodiopsida
Order: Lycopodiales Family: Lycopodiaceae
Alpine clubmoss, Diphasiastrum alpinum
Chinese clubmoss, Huperzia chinensis
Common clubmoss, Lycopodium clavatum
Northern bog clubmoss, Lycopodiella inundata
One-cone clubmoss, Lycopodium lagopus
Pacific clubmoss, Huperzia haleakalae
Sitka clubmoss, Diphasiastrum sitchense
Stiff clubmoss, Spinulum annotinum
Trailing clubmoss, Diphasiastrum complanatum
Tree groundpine, Dendrolycopodium dendroideum
Western clubmoss, Huperzia occidentalis
Quillworts
Class: Isoetopsida
Order: Isoetales, Family: Isoetaceae
Bolander's quillwort, Isoetes bolanderi
Howell's quillwort, Isoetes howellii
Nuttall's quillwort, Isoetes nuttallii
Spiny-spored quillwort, Isoetes echinospora
Western quillwort, Isoetes occidentalis
Spike-mosses
Class: Isoetopsida
Order: Selaginellales, Family: Selaginellaceae
Lesser spikemoss, Selaginella densa
Low spikemoss, Selaginella selaginoides
Wallace's spikemoss, Selaginella wallacei
Watson's spikemoss, Selaginella watsonii
Mosses
Granite mosses
Class: Andreaeopsida Order: Andreaeales, Family: Andreaeaceae
Blytt's andreaea moss, Andreaea blyttii
Peat mosses
Class: Sphagnopsida
Order: Sphagnales, Family: Sphagnaceae
Contorted sphagnum, Sphagnum contortum
Flat-leaved bogmoss, Sphagnum platyphyllum
Flat-topped bogmoss, Sphagnum fallax
Fringed bogmoss, Sphagnum fimbriatum
Girgensohn's sphagnum, Sphagnum girgensohnii
Low sphagnum, Sphagnum compactum
Magellan's peat moss, Sphagnum magellanicum
Mendocino peat moss, Sphagnum mendocinum
Barrowleaf peat moss, Sphagnum angustifolium
Rusty peat moss, Sphagnum fuscum
Sphagnum moss, Sphagnum centrale
Streamside sphagnum moss, Sphagnum riparium
Wulf's peat moss, Spha |
https://en.wikipedia.org/wiki/Segregate%20%28taxonomy%29 | In taxonomy, a segregate, or a segregate taxon is created when a taxon is split off from another taxon. This other taxon will be better known, usually bigger, and will continue to exist, even after the segregate taxon has been split off. A segregate will be either new or ephemeral: there is a tendency for taxonomists to disagree on segregates, and later workers often reunite a segregate with the 'mother' taxon.
If a segregate is generally accepted as a 'good' taxon it ceases to be a segregate. Thus, this is a way of indicating change in the taxonomic status. It should not be confused with, for example, the subdivision of a genus into subgenera.
For example, the genus Alsobia is a segregate from the genus Episcia; The genera Filipendula and Aruncus are segregates from the genus Spiraea.
External links
A more detailed explanation, with multiple examples on mushrooms.
Botanical nomenclature
Plant taxonomy
Taxonomy (biology) |
https://en.wikipedia.org/wiki/Gyrokinetic%20ElectroMagnetic | Gyrokinetic ElectroMagnetic (GEM) is a gyrokinetic plasma turbulence simulation that uses the particle-in-cell method. It is used to study waves, instabilities and nonlinear behavior of tokamak fusion plasmas. Information about GEM can be found at the GEM web page.
There are two versions of GEM, one is a flux-tube version and the other one is a global general geometry
version. Both versions of GEM use a field-aligned coordinate system. Ions are treated kinetically, but averaged over their gyro-obits and electrons are treated as drift-kinetic.
The modeling of the tokamak plasmas
GEM solves the electromagnetic gyrokinetic equations which are the appropriate equations for well magnetized plasmas. The plasma is treated statistically as a kinetic distribution function. The distribution function depends on the three-dimensional position, the energy and magnetic moment. The time evolution of the distribution function is described by gyrokinetic theory which simply averages the Vlasov-Maxwell system of equations over the fast gyromotion associated with particles exhibiting cyclotron motion about the magnetic field lines. This eliminates fast time scales associated with the gyromotion and reduces the dimensionality of the problem from six down to five.
Algorithm to solve the equations
GEM uses the delta-f particle-in-cell (PIC) plasma simulation method. An expansion about an adiabatic response is made for electrons to overcome the limit of small time step, which is caused by the fast motion of electrons. GEM uses a novel electromagnetic algorithm allowing direct numerical simulation of the electromagnetic problem at high plasma pressures. GEM uses a two-dimensional domain decomposition (see domain decomposition method) of the grid and particles to obtain good performance on massively parallel computers. A Monte Carlo method is used to model small angle Coulomb collisions.
Applications
GEM is used to study nonlinear physics associated with tokamak plasma turbule |
https://en.wikipedia.org/wiki/Write%20protection | Write protection is any physical mechanism that prevents writing, modifying, or erasing data on a device. Most commercial software, audio and video on writeable media is write-protected when distributed.
Examples
IBM -inch magnetic tape reels, introduced in the 1950s, had a circular groove on one side of the reel, into which a soft plastic ring had to be placed in order to write on the tape. ("No ring, no write.")
Audio cassettes and VHS videocassettes have tabs on the top/rear edge that can be broken off (uncovered = protected).
8 and -inch floppies can have, respectively, write-protect and write-enable notches on the right side (8-inch punched = protected; -inch covered/notch not present = protected). A common practice with single-sided floppies was to punch a second notch on the opposite side of the disk to enable use of both sides of the media, creating a flippy disk, so called because one originally had to flip the disk over to use the other side.
-inch floppy disks have a sliding tab in a window on the right side (open = protected).
Iomega Zip disks were write-protected using the IomegaWare software.
Syquest EZ-drive (135 & 250 MB) disks were write-protected using a small metal switch on the rear of the disk at the bottom.
VHS-C, Video8, Hi8, and DV videocassettes have a sliding tab on the rear edge.
Iomega ditto tape cartridges had a small sliding tab on the top left hand corner on the front face of the cartridge.
USB flash drives sometimes have a small switch, though this has become uncommon. An example of a USB flash drive that supported write protection via a switch is the Transcend JetFlash series.
Secure Digital (SD) cards have a write-protect tab on the left side.
Extensively, media that, by means of design, can't operate outside from this mode: CD-R, DVD-R, Vinyl records, etc.
These mechanisms are intended to prevent only accidental data loss or attacks by computer viruses. A determined user can easily circumvent them either by covering a |
https://en.wikipedia.org/wiki/Monoamine%20nuclei | Monoamine nuclei are clusters of cells that primarily use monoamine neurotransmitters to communicate. The raphe nuclei, ventral tegmental area, and locus coeruleus have been included in texts about monoamine nuclei. These nuclei receive a variety of inputs including from other monoamines, as well as from glutaminergic, GABAergic, and substance p related pathways. The catacholaminergic pathways mainly project upwards into the cortical and limbic regions, power sparse descending axons have been observed in animals models. Both ascending and descending serotonergic pathways project from the raphe nuclei. Raphe nuclei in the obscurus, pallid us, and magnus descend into the brainstem and spinal cord, while the raphe ponds, raphe dorsals, and nucleus centralism superior projected up into the medial forebrain bundle before branching off. Monoamine nuclei have been studied in relation to major depressive disorder, with some abnormalities observed, however MAO-B levels appear to be normal during depression in these regions. |
https://en.wikipedia.org/wiki/Giuseppe%20Cocconi | Giuseppe Cocconi (1914–2008) was an Italian physicist who was director of the Proton Synchrotron at CERN in Geneva.
He is known for his work in particle physics and for his involvement with SETI where he wrote, "[t]he probability of success is difficult to estimate; but if we never search, the chance of success is zero."
Life
Cocconi was born in Como, Kingdom of Italy in 1914. He went to study physics at the University of Milan, and then in February 1938, went to the Sapienza University of Rome on the invitation of Edoardo Amaldi. There he met physicists Enrico Fermi, and Gilberto Bernardini. With Fermi, he built a Wilson chamber to study the disintegration of mesons. In August of that year, Cocconi laid the foundation of cosmic ray research in Milan. While at Milan, Cocconi supervised Vanna Tongiorgi, who picked cosmic rays as her thesis' subject, and later married her in 1945.
In 1942, Cocconi was nominated professor at University of Catania, but was engaged by the Italian army to research infrared phenomena for the Royal Italian Air Force until the end of World War II, in late 1944. He taught at Catania until 1947, when Hans Bethe made a request that he would join Cornell University. During his stay at Cornell, Cocconi and his wife performed many experiments there and in Echo Lake located in the Rocky Mountains, where they demonstrated the galactic and extragalactic origins of cosmic rays. In 1955, he was awarded a Guggenheim Fellowship. While at Cornell he also wrote, with Philip Morrison, his most famous paper "Searching for Interstellar Communications", on the 21 cm Hydrogen line, which turned out to be of vital importance in the SETI program.
During his sabbatical of 1959–1961, Cocconi helped kick-start the Proton Synchrotron research program at CERN, and conducted a series of experiment on proton-proton scattering, and on the cross section of protons and neutrons. He also continued this research at Brookhaven National Laboratory (BNL). In 1963 he returne |
https://en.wikipedia.org/wiki/Computer%20security%20incident%20management | In the fields of computer security and information technology, computer security incident management involves the monitoring and detection of security events on a computer or computer network, and the execution of proper responses to those events. Computer security incident management is a specialized form of incident management, the primary purpose of which is the development of a well understood and predictable response to damaging events and computer intrusions.
Incident management requires a process and a response team which follows this process. This definition of computer security incident management follows the standards and definitions described in the National Incident Management System (NIMS). The incident coordinator manages the response to an emergency security incident. In a Natural Disaster or other event requiring response from Emergency services, the incident coordinator would act as a liaison to the emergency services incident manager.
Overview
Computer security incident management is an administrative function of managing and protecting computer assets, networks and information systems. These systems continue to become more critical to the personal and economic welfare of our society. Organizations (public and private sector groups, associations and enterprises) must understand their responsibilities to the public good and to the welfare of their memberships and stakeholders. This responsibility extends to having a management program for “what to do, when things go wrong.” Incident management is a program which defines and implements a process that an organization may adopt to promote its own welfare and the security of the public.
Components of an incident
Events
An event is an observable change to the normal behavior of a system, environment, process, workflow or person (components). There are three basic types of events:
Normal—a normal event does not affect critical components or require change controls prior to the implementation |
https://en.wikipedia.org/wiki/Federa%C3%A7%C3%A3o%20das%20Sociedades%20de%20Biologia%20Experimental | The Federação das Sociedades de Biologia Experimental (Federation of Experimental Biology Societies, abbreviated FeSBE) is a Brazilian scientific association which runs a number of the mainstream specialized societies in experimental biology and medicine. It was founded in 1985 and currently has the following member societies:
Brazilian Society of Physiology (SBFis)
Brazilian Society of Biophysics (SBBf)
Brazilian Society of Biochemistry and Molecular Biology (SBBq)
Brazilian Society of Pharmacology and Experimental Therapeutics (SBFTE)
Brazilian Society of Immunology (SBI)
Brazilian Society of Neurosciences and Behavior (SBNeC)
Brazilian Society of Clinical Investigation (SBIC)
There are also 4 associate member societies:
Brazilian Society of Endocrinology and Metabology (SBEM)
Brazilian Society of Cell Biology (SBBC)
Brazilian Society of Nuclear Biosciences (SBBN)
Brazilian Research Association in Vision (BRAVO)
FeSBE holds an annual meeting with the societies of physiology, pharmacology, neurosciences, biophysics and clinical investigation every August, the three first societies responding for 83% of all attendees. The remaining societies biochemistry and molecular biology, immunology, endocrinology, cell biology and nuclear biosciences) have their own separate meetings.
The Annual FeSBE Meeting has the same importance of its American counterpart, the Federation of Experimental Biology Societies (FEBS), although smaller in size. In 2005 the Meeting has had 3,768 participants (74% of which were undergrad or graduate students) and 2,940 posters were presented. About 90% of the participants came from states from the South and Southeast regions.
FeSBE is affiliated with the Brazilian Society for Advancement of Science. Its current president is Dr. Gerhard Malnic. Former presidents were Eduardo Krieger, Sérgio Ferreira, Dora Fix Ventura and Antonio Carlos Campos de Carvalho.
See also
Science and technology in Brazil
External links
FESBE Home Page
Sc |
https://en.wikipedia.org/wiki/Jahn%E2%80%93Teller%20effect | The Jahn–Teller effect (JT effect or JTE) is an important mechanism of spontaneous symmetry breaking in molecular and solid-state systems which has far-reaching consequences in different fields, and is responsible for a variety of phenomena in spectroscopy, stereochemistry, crystal chemistry, molecular and solid-state physics, and materials science. The effect is named for Hermann Arthur Jahn and Edward Teller, who first reported studies about it in 1937.
Simplified overview
The Jahn–Teller effect, sometimes also referred to as Jahn–Teller distortion, describes the geometrical distortion of molecules and ions that results from certain electron configurations. The Jahn–Teller theorem essentially states that any non-linear molecule with a spatially degenerate electronic ground state will undergo a geometrical distortion that removes that degeneracy, because the distortion lowers the overall energy of the species. For a description of another type of geometrical distortion that occurs in crystals with substitutional impurities see article off-center ions.
Transition metal chemistry
The Jahn–Teller effect is most often encountered in octahedral complexes of the transition metals. The phenomenon is very common in six-coordinate copper(II) complexes. The d9 electronic configuration of this ion gives three electrons in the two degenerate eg orbitals, leading to a doubly degenerate electronic ground state. Such complexes distort along one of the molecular fourfold axes (always labelled the z axis), which has the effect of removing the orbital and electronic degeneracies and lowering the overall energy. The distortion normally takes the form of elongating the bonds to the ligands lying along the z axis, but occasionally occurs as a shortening of these bonds instead (the Jahn–Teller theorem does not predict the direction of the distortion, only the presence of an unstable geometry). When such an elongation occurs, the effect is to lower the electrostatic repulsion between |
https://en.wikipedia.org/wiki/Intracellular%20bacteria | Intracellular bacteria are bacteria that have the capability to enter and survive within the cells of the host organism. Many of them are capable of growth extracellularly, but some of them can grow and reproduce only intracellularly (obligate intracellular parasites). Besides bacteria, there are other kinds of intracellular microorganisms.
Examples of non-obligate intracellular bacteria include members of the genera Brucella, Legionella, Listeria, and Mycobacterium. Examples of obligate intracellular bacteria include members of the order Rickettsiales and members of the genus Mycoplasma.
See also
Endosymbiont |
https://en.wikipedia.org/wiki/Paul%20F.%20Forman%20Team%20Engineering%20Excellence%20Award | The Paul F. Forman Team Engineering Excellence Award was first introduced as the Engineering Excellence Award by the Optical Society in 1989 and was awarded individually, or shared among individuals. In 2007 it was named in honor of Paul F. Forman. This award recognizes technical achievements in optical engineering as well as contributions to society such as engineering education. It award is sponsored by Zygo Corporation, Canon Inc, Optical Solutions Group at Synopsys, Cambridge Research & Instrumentation, and several individual contributors.
List of winners
Source: The Optical Society
2020 2-Photon Optical Clock Collaboration
2019 Headwall Photonics, Inc., Special Projects Team
2018 Adaptive Optics Facility on the Very Large Telescope (VLT) at European Southern Observatory’s Paranal Observatory
2017 Guide Star Alliance
2016 Advanced LIGO Engineering Team
2015 Logic Analysis Tool Team (LAT Team)
2014 Intel® Silicon Photonics Solutions Group
2013 ZygoLOT Automotive Precision Optical Team
2012 Georgia Tech Research Institute (GTRI) Lidar Team
2011 Tomasz S. Tkaczyk
2010 Alan E. Willner
2009 Stephen A. Boppart
2008 Michael J. Bechtold
2007 Ming C. Wu
2006 Jean-Claude Diels
2005 René-Jean Essiambre
2005 Michael G. Littman
2004 Shun-Lien Chuang, S. Chandrasekhar
2003 L. Ramdas Ram-Mohan, Paul R. Dumas, Mark E. Lowry, Alan H. Gnauck
2002 Tim Day, Christopher R. Doerr, David W. Peckham
2001 Henry A. Blauvelt, Michael A. Klug, David G. Mehuys, Dale E. Morton
2000 J. W. Anello, Jr., A. Erstling, A. C. Tam, R. L. Hartman, L. J. P. Ketelsen, J. A. Grenko, W-T. Tzang
1999 Denis Barbier, Valentin Gapontsev, Igor Koltchanov, Olaf Lenzmann, Herman Reedy
1998 Lee R. Shiozawa, Kenneth L. Walker
1997 Donald M. Combs, John J. Mader, Jeffrey W. Roblee, Edward A. Yobaccio, Paul R. Yoder, Jr.
1996 Gary Blough, Teddi C. Laurin
1995 Francisco J. Duarte, John D. Gonglewski, Gary Guenther, Melvyn H. Kreitze |
https://en.wikipedia.org/wiki/Cisco%20Valet%20routers | On March 30, 2010 Cisco unveiled a new series of home networking products called "Valet". The focus of this series is to simplify the installation and configuration of a typical wireless home network by shipping routers partially pre-configured and bundling with them software that aids the user in setting up their network with a step-by-step wizard.
While the early Valet routers are based on variants of Cisco's Linksys routers, the Valet series is not tied to the Linksys brand and is marketed separately.
Specifications and versions
Valet M10
The Valet M10 along with the Valet Plus M20 were Cisco's first routers in the Valet series. The M10 is a 2.4 GHz single-band 802.11n wireless router featuring 10/100 LAN connectivity. The v1 of this model is equivalent to the Linksys E1000 v1 and WRT160N v3, sharing the same hardware and specifications. It is white in color and features a light blue trim.
Third Party Firmware: DD-WRT & Tomato Firmware releases have been known to work on the v1 unit.
Valet Plus M20
The M20 is a 2.4 GHz single-band 802.11n wireless router featuring gigabit LAN connectivity. This model is equivalent to the Linksys WRT310N v2, also sharing the same hardware and specifications. It is white in color and features a silver trim.
Default settings
IP address: 192.168.1.1
Web interface username: "admin"
Password: "admin" |
https://en.wikipedia.org/wiki/Cord%20factor | Cord factor, or trehalose dimycolate (TDM), is a glycolipid molecule found in the cell wall of Mycobacterium tuberculosis and similar species. It is the primary lipid found on the exterior of M. tuberculosis cells. Cord factor influences the arrangement of M. tuberculosis cells into long and slender formations, giving its name. Cord factor is virulent towards mammalian cells and critical for survival of M. tuberculosis in hosts, but not outside of hosts. Cord factor has been observed to influence immune responses, induce the formation of granulomas, and inhibit tumor growth. The antimycobacterial drug SQ109 is thought to inhibit TDM production levels and in this way disrupts its cell wall assembly.
Structure
A cord factor molecule is composed of a sugar molecule, trehalose (a disaccharide), composed of two glucose molecules linked together. Trehalose is esterified to two mycolic acid residues. One of the two mycolic acid residues is attached to the sixth carbon of one glucose, while the other mycolic acid residue is attached to the sixth carbon of the other glucose. Therefore, cord factor is also named trehalose-6,6'-dimycolate. The carbon chain of the mycolic acid residues vary in length depending on the species of bacteria it is found in, but the general range is 20 to 80 carbon atoms. Cord factor's amphiphilic nature leads to varying structures when many cord factor molecules are in close proximity. On a hydrophobic surface, they spontaneously form a crystalline monolayer. This crystalline monolayer is extremely durable and firm; it is stronger than any other amphiphile found in biology. This monolayer also forms in oil-water, plastic-water, and air-water surfaces. In an aqueous environment free of hydrophobic surfaces, cord factor forms a micelle. Furthermore, cord factor interlocks with lipoarabinomannan (LAM), which is found on the surface of M. tuberculosis cells as well, to form an asymmetrical bilayer. These properties cause bacteria that produce cord |
https://en.wikipedia.org/wiki/Net%20force | In mechanics, the net force is the sum of all the forces acting on an object. For example, if two forces are acting upon an object in opposite directions, and one force is greater than the other, the forces can be replaced with a single force that is the difference of the greater and smaller force. That force is the net force.
When forces act upon an object, they change its acceleration. The net force is the combined effect of all the forces on the object's acceleration, as described by Newton's second law of motion.
When the net force is applied at a specific point on an object, the associated torque can be calculated. The sum of the net force and torque is called the resultant force, which causes the object to rotate in the same way as all the forces acting upon it would if they were applied individually.
It is possible for all the forces acting upon an object to produce no torque at all. This happens when the net force is applied along the line of action.
In some texts, the terms resultant force and net force are used as if they mean the same thing. This is not always true, especially when in complex topics like the motion of spinning objects or situations where everything is perfectly balanced, known as static equilibrium. In these cases, it's important to understand that "net force" and "resultant force" can have distinct meanings.
The Concept of Total Force
In physics, a force is considered a vector quantity. This means that it not only has a size (or magnitude) but also a direction in which it acts. We typically represent force with the symbol F in boldface, or sometimes, we place an arrow over the symbol to indicate its vector nature, like this: .
When we need to visually represent a force, we draw a line segment. This segment starts at a point A, where the force is applied, and ends at another point B. This line not only gives us the direction of the force (from A to B) but also its magnitude: the longer the line, the stronger the force.
One of the |
https://en.wikipedia.org/wiki/Medical%20test | A medical test is a medical procedure performed to detect, diagnose, or monitor diseases, disease processes, susceptibility, or to determine a course of treatment. Medical tests such as, physical and visual exams, diagnostic imaging, genetic testing, chemical and cellular analysis, relating to clinical chemistry and molecular diagnostics, are typically performed in a medical setting.
Types of tests
By purpose
Medical tests can be classified by their purposes, the most common of which are diagnosis, screening and evaluation.
Diagnostic
A diagnostic test is a procedure performed to confirm or determine the presence of disease in an individual suspected of having a disease, usually following the report of symptoms, or based on other medical test results. This includes posthumous diagnosis. Examples of such tests are:
Using nuclear medicine to examine a patient suspected of having a lymphoma.
Measuring the blood sugar in a person suspected of having diabetes mellitus after periods of increased urination.
Taking a complete blood count of an individual experiencing a high fever to check for a bacterial infection.
Monitoring electrocardiogram readings on a patient with chest pain to diagnose or determine any heart irregularities.
Screening
Screening refers to a medical test or series of tests used to detect or predict the presence of disease in at-risk individuals within a defined group such as a population, family, or workforce. Screenings may be performed to monitor disease prevalence, manage epidemiology, aid in prevention, or strictly for statistical purposes.
Examples of screenings include measuring the level of TSH in the blood of a newborn infant as part of newborn screening for congenital hypothyroidism, checking for Lung cancer in non-smoking individuals who are exposed to second-hand smoke in an unregulated working environment, and Pap smear screening for prevention or early detection of cervical cancer.
Monitoring
Some medical tests are used to mon |
https://en.wikipedia.org/wiki/Artturi%20Ilmari%20Virtanen | Artturi Ilmari Virtanen (; 15 January 1895 – 11 November 1973) was a Finnish chemist and recipient of the 1945 Nobel Prize in Chemistry "for his research and inventions in agricultural and nutrition chemistry, especially for his fodder preservation method".
He invented AIV silage which improved milk production and a method of preserving butter, the AIV salt, which led to increased Finnish butter exports.
Personal life
Artturi Ilmari Virtanen was born on 15 January, 1895, in Helsinki, Finland . He is the son of Kaarlo Virtanen, a railway engine driver and Serafina Isotalo.
He completed his school education at the Classical Lyceum in Viipuri, Finland. He married the botanist Lilja Moisio (1894-1972) in 1920 and had two sons with her.
In 1933, he bought a farm near Helsinki where he tested some of his scientific results in practice. He saw in the overproduction of food only a temporary phenomenon. He lived a simple life, he never had a car of his own, never smoked and never consumed alcohol. He died of pneumonia in November 1973, following a broken femur from a fall few weeks prior. He was buried at the Hietaniemi Cemetery.
Academics
Virtanen began his studies at the University of Helsinki in chemistry 1913 earning his Master and in 1918 his PhD in organic chemistry. In 1919 he started to work in the laboratories of Valio, a large producer of dairy products and became director of the laboratory in 1920. Feeling not fully qualified and following his interest in botany and zoology led him to further scientific education and so he left Valio and studied at the ETH, the University of Münster and the University of Stockholm he studied physical chemistry, soil chemistry and microbiology. In 1923 in Sweden he worked with Hans von Euler-Chelpin, who was awarded with the Nobel Prize in Chemistry in 1929. Back in Finland he became lecturer at the University of Helsinki in 1924, known for his lectures on chemistry of life. He worked in the laboratory of the Butter Export As |
https://en.wikipedia.org/wiki/Death%20effector%20domain | The death-effector domain (DED) is a protein interaction domain found only in eukaryotes that regulates a variety of cellular signalling pathways. The DED domain is found in inactive procaspases (cysteine proteases) and proteins that regulate caspase activation in the apoptosis cascade such as FAS-associating death domain-containing protein (FADD). FADD recruits procaspase 8 and procaspase 10 into a death induced signaling complex (DISC). This recruitment is mediated by a homotypic interaction between the procaspase DED and a second DED that is death effector domain in an adaptor protein that is directly associated with activated TNF receptors. Complex formation allows proteolytic activation of procaspase into the active caspase form which results in the initiation of apoptosis (cell death). Structurally the DED domain are a subclass of protein motif known as the death fold and contains 6 alpha helices, that closely resemble the structure of the Death domain (DD).
Structure
DED is a subfamily of the DD superfamily (other recognizable domains in this superfamily are: caspase-recruitment domain (CARD), pyrin domain (PYD) and death domain (DD)). The subfamilies resemble structurally one another, all of them (and DED in particular) are composed of a bundle of 6 alpha-helices, but they diverge in the surface features.
The complete primary structure of this proteic domain has not been consensually defined. Some studies described residues 2-184, but C-terminus and N-terminus residues are not identified yet. The presence of amino acids that determine the solubility and aggregation to DED allowed to identify DED's in different proteins, such as caspase-8 and MC159. The secondary structure of the domain, as said, is built by 6 alpha-helices.
The tertiary structure of the domain has been described from the crystallization of caspase 8 in the human. The method used to describe the structure was X-RAY diffraction and the resolution obtained is 2.2 Å. DEDs in this protein |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.