source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Lymphocyte-variant%20hypereosinophilia | Lymphocyte-variant hypereosinophilia is a rare disorder in which eosinophilia or hypereosinophilia (i.e. a large or extremely large increase in the number of eosinophils in the blood circulation) is caused by an aberrant population of lymphocytes. These aberrant lymphocytes function abnormally by stimulating the proliferation and maturation of bone marrow eosinophil-precursor cells termed colony forming unit-Eosinophils or CFU-Eos.
The overly stimulated CFU-Eos cells mature to apparently normal appearing but possibly overactive eosinophils which enter the circulation and may accumulate in and damage various tissues. The disorder is usually indolent or slowly progressive but may proceed to a leukemic phase sometimes classified as acute eosinophilic leukemia. Lymphocyte-variant hypereosinophilia can therefore be regarded as a precancerous disorder.
The disorder merits therapeutic intervention to avoid or reduce eosinophil-induced tissue injury and treat its leukemic phase. The latter phase is aggressive and typically responds relatively poorly to anti-leukemia chemotherapeutic drug regimens.
Presentation
The typical patient with lymphocyte-variant hypereosinophilia presents with an extended history of hypereosinophilia and cutaneous allergy-like symptoms. Skin symptoms, which occur in >75% of patients, include erythroderma, pruritus, eczema, Poikiloderma, urticarial, and episodic angioedema. The symptom of episodic angioedema (i.e. soft tissue swelling of the face, tongue, larynx, abdomen, arms, or legs) in lymphocyte-variant hypereosinophilia resembles that occurring in Gleich's syndrome, a rare disease that is accompanied by secondary hypereosinophilia plus a sub-population of CD3(-), CD4(+) T cells; this involvement of the latter cell types supports the notion that Gleich's syndrome is a subtype of lymphocyte-variant hypereosiophilia. Biopsies of skin lesions commonly find prominent accumulations of eosinophils. Other presentations include:
a) lymphadenopath |
https://en.wikipedia.org/wiki/Yakov%20Perelman | Yakov Isidorovich Perelman (; – 16 March 1942) was a Russian Empire and Soviet science writer and author of many popular science books, including Physics Can Be Fun and Mathematics Can Be Fun (both translated from Russian into English).
Life and work
Perelman was born in 1882 in the town of Białystok, Russian Empire. He obtained the Diploma in Forestry from the Imperial Forestry Institute (Now Saint-Petersburg State Forestry University) in Saint Petersburg, in 1909. He was influenced by Ernst Mach and probably the Russian Machist Alexander Bogdanov in his pedagogical approach to popularising science. After the success of "Physics for Entertainment", Perelman set out to produce other books, in which he showed himself to be an imaginative populariser of science. Especially popular were "Arithmetic for entertainment", "Mechanics for entertainment", "Geometry for Entertainment", "Astronomy for entertainment", "Lively Mathematics", " Physics Everywhere", and "Tricks and Amusements".
His famous books on physics and astronomy were translated into various languages by the erstwhile Soviet Union.
The scientist Konstantin Tsiolkovsky thought highly of Perelman's talents and creative genius, writing of him in the preface of Interplanetary Journeys: "The author has long been known by his popular, witty and quite scientific works on physics, astronomy and mathematics, which are, moreover written in a marvelous language and are very readable."
Perelman has also authored a number of textbooks and articles in Soviet popular science magazines.
In addition to his educational and scientific writings, he also worked as an editor of science magazines, including Nature and People and In the Workshop of Nature.
Perelman died from starvation in 1942, during the German Siege of Leningrad. The siege started on 9 September 1941 and lasted 872 days, until
27 January 1944. The Siege of Leningrad was one of the longest, most destructive sieges of a major city in modern history and one of |
https://en.wikipedia.org/wiki/Brienomyrus | Brienomyrus is a genus of small elephantfish in the family Mormyridae from Africa. Usually available in the pet trade, these fish are commercially referred to as baby whales or baby whalefish.
Species
There are currently three recognized species in this genus:
Brienomyrus adustus (Fowler 1936)
Brienomyrus brachyistius (T. N. Gill 1862) (whale African black reg)
Brienomyrus longianalis (Boulenger 1901) (Niger delta elephantfish) |
https://en.wikipedia.org/wiki/Disulfide%20oxidoreductase%20D | The Disulfide bond oxidoreductase D (DsbD) family is a member of the Lysine Exporter (LysE) Superfamily. A representative list of proteins belonging to the DsbD family can be found in the Transporter Classification Base.
Homology
Homologues include:
(1) several thiol-disulfide exchange proteins (i.e., TC# 5.A.1.1.1)
(2) the cytochrome c-type biogenesis proteins, CcdA (TC# 5.A.1.2.1) of Paracoccus pantotrophus and Bacillus subtilis.
(3) the methylamine utilization proteins, MauF (TC# 5.A.1.3.1) of Paracoccus denitrificans and P. versutus.
(4) the mercury resistance proteins (TC# 5.A.1.4.1; possibly Hg2+ transporters) of Mycobacterium tuberculosis and Streptomyces lividans.
(5) suppressors of copper sensitivity (TC# 5.A.1.5.1; copper tolerance proteins) of Salmonella typhimurium and Vibrio cholerae.
(6) components of peroxide reduction pathways (TC# 5.A.1.5.2), and
(7) components of sulfenic acid reductases.
Disulfide bond oxidoreductase D (DsbD)
The best characterized member of the DsbD family is DsbD of E. coli (TC# 5.A.1.1.1). The DsbD protein is membrane-embedded with a putative N-terminal transmembrane segment (TMS) plus 8 additionalTMSs. The smallest homologues (190 aas with 6 putative TMSs) are found in archaea, while the largest are found in both Gram-negative bacteria (758 aas with 9 putative TMSs) and Gram-positive bacteria (695 aas with 6 putative TMSs).
The overall vectorial electron transfer reaction catalyzed by DsbD is:
2 e → 2 e
Structure
DsbB contains 4 essential cysteine residues, reversibly forming two disulfide bonds. Although DsbA displays no proofreading activity for repair of wrongly paired disulfides, DsbC, DsbE and DsbG have been found to demonstrate proofreading activity. Therefore, the two transmembrane pathways involving DsbD and DsbB together catalyze extracellular disulfide reduction (DsbD) and oxidation (DsbB) in a superficially reversible process that allows dithiol/disulfide exchange.
System reduction pathway
In the E. |
https://en.wikipedia.org/wiki/Intelligent%20Mail%20barcode | The Intelligent Mail Barcode (IMb) is a 65-bar barcode for use on mail in the United States. The term "Intelligent Mail" refers to services offered by the United States Postal Service for domestic mail delivery. The IM barcode is intended to provide greater information and functionality than its predecessors POSTNET and PLANET. An Intelligent Mail barcode has also been referred to as a One Code Solution and a 4-State Customer Barcode, abbreviated 4CB, 4-CB or USPS4CB. The complete specification can be found in USPS Document USPS-B-3200. It effectively incorporates the routing ZIP Code and tracking information included in previously used postal barcode standards.
The barcode is applied by the sender; the Postal Service required use of the Intelligent Mail barcode to qualify for automation prices beginning January 28, 2013. Use of the barcode provides increased overall efficiency, including improved deliverability, and new services.
Symbology
The Intelligent Mail barcode is a height-modulated barcode that encodes up to 31 decimal digits of mail-piece data into 65 vertical bars.
The code is made up of four distinct symbols, which is why it was once referred to as the 4-State Customer Barcode. Each bar contains the central "tracker" portion, and may contain an ascender, descender, neither, or both (a "full bar").
The 65 bars represent 130 bits (or 39.13 decimal digits), grouped as ten 13-bit characters. Each character has 2, 5, 8, or 11 of its 13 bits set to one. The Hamming distance between characters is at least 2. Consequently, single-bit errors in a character can be detected (toggling one bit results in an invalid character). The characters are interleaved throughout the symbol.
The number of characters can be calculated from the binomial coefficient.
The total number of characters is two times 1365, or 2730. Log2(2730) is 11.41469 bits per group. So the 65 bars (or 130 bits) encode a 114-bit message.
The encoding includes an eleven-bit cyclic redundancy c |
https://en.wikipedia.org/wiki/Faddeev%E2%80%93LeVerrier%20algorithm | In mathematics (linear algebra), the Faddeev–LeVerrier algorithm is a recursive method to calculate the coefficients of the characteristic polynomial of a square matrix, , named after Dmitry Konstantinovich Faddeev and Urbain Le Verrier. Calculation of this polynomial yields the eigenvalues of as its roots; as a matrix polynomial in the matrix itself, it vanishes by the Cayley–Hamilton theorem. Computing the characteristic polynomial directly from the definition of the determinant is computationally cumbersome insofar as it introduces a new symbolic quantity ; by contrast, the Faddeev-Le Verrier algorithm works directly with coefficients of matrix .
The algorithm has been independently rediscovered several times in different forms. It was first published in 1840 by Urbain Le Verrier, subsequently redeveloped by P. Horst, Jean-Marie Souriau, in its present form here by Faddeev and Sominsky, and further by J. S. Frame, and others. (For historical points, see Householder. An elegant shortcut to the proof, bypassing Newton polynomials, was introduced by Hou. The bulk of the presentation here follows Gantmacher, p. 88.)
The Algorithm
The objective is to calculate the coefficients of the characteristic polynomial of the matrix ,
where, evidently, = 1 and 0 = (−1)n det .
The coefficients are determined by induction on , using an auxiliary sequence of matrices
Thus,
etc.,
...;
Observe terminates the recursion at . This could be used to obtain the inverse or the determinant of .
Derivation
The proof relies on the modes of the adjugate matrix, , the auxiliary matrices encountered.
This matrix is defined by
and is thus proportional to the resolvent
It is evidently a matrix polynomial in of degree . Thus,
where one may define the harmless ≡0.
Inserting the explicit polynomial forms into the defining equation for the adjugate, above,
Now, at the highest order, the first term vanishes by =0; whereas at the bottom order (constant in , fro |
https://en.wikipedia.org/wiki/Haloplanus%20natans | Haloplanus natans is a halophilic Archaeon in the family of Halobacteriaceae and the type species of the genus Haloplanus. It was isolated from controlled mesocosms with a mixture of water from the Dead Sea and the Red Sea. |
https://en.wikipedia.org/wiki/Vertebra | Each vertebra (: vertebrae) is an irregular bone with a complex structure composed of bone and some hyaline cartilage, that make up the vertebral column or spine, of vertebrates. The proportions of the vertebrae differ according to their spinal segment and the particular species.
The basic configuration of a vertebra varies; the bone is the body, and the central part of the body
is the centrum. The upper and lower surfaces of the vertebra body give attachment to the intervertebral discs. The posterior part of a vertebra forms a vertebral arch, in eleven parts, consisting of two pedicles (pedicle of vertebral arch), two laminae, and seven processes. The laminae give attachment to the ligamenta flava (ligaments of the spine). There are vertebral notches formed from the shape of the pedicles, which form the intervertebral foramina when the vertebrae articulate. These foramina are the entry and exit conduits for the spinal nerves. The body of the vertebra and the vertebral arch form the vertebral foramen, the larger, central opening that accommodates the spinal canal, which encloses and protects the spinal cord.
Vertebrae articulate with each other to give strength and flexibility to the spinal column, and the shape at their back and front aspects determines the range of movement. Structurally, vertebrae are essentially alike across the vertebrate species, with the greatest difference seen between an aquatic animal and other vertebrate animals. As such, vertebrates take their name from the vertebrae that compose the vertebral column.
Structure
General structure
In the human vertebral column the size of the vertebrae varies according to placement in the vertebral column, spinal loading, posture and pathology. Along the length of the spine the vertebrae change to accommodate different needs related to stress and mobility. Each vertebra is an irregular bone.
Every vertebra has a body (vertebral body), which consists of a large anterior middle portion called the cen |
https://en.wikipedia.org/wiki/Dysphania%20botrys | Dysphania botrys (syn. Chenopodium botrys), the Jerusalem oak goosefoot, sticky goosefoot or feathered geranium, is a flowering plant in the genus Dysphania (the glandular goosefoots). It is native to the Mediterranean region.
Jerusalem oak goosefoot was formerly classed in the genus Ambrosia, with the binomial name Ambrosia mexicana. It is naturalised in the United States and Mexico, the old species synonym deriving from the latter.
Cultivation
The plant has a strong scent, reminiscent of stock cubes, and can be used as a flavouring in cooking. It is cultivated as a hardy annual by gardeners. |
https://en.wikipedia.org/wiki/National%20Society%20of%20Black%20Physicists | The National Society of Black Physicists (NSBP), established in the United States in 1977, is a non-profit professional organization with the goal to promote the professional well-being of African Diaspora physicists and physics students within the international scientific community and the world community at large.
History
A group of involved physicists met at Fisk University in 1972 to honor three well known African-American physicists: Dr. Donald Edwards, Dr. John McNeile Hunter, and Dr. Halson V. Eagleson. On April 28, 1977, the Society was established at Morgan State University, with its founding co-chairs being Walter E. Massey and James Davenport. As of 2023, the NSBP relocated its headquarters from Arlington, Virginia to the Optica headquarters in Washington, DC.
Activities
The organization holds its annual conference in February. More recently, it has jointly held these conferences with the National Society of Hispanic Physicists. Attendance at these conferences is upwards of 500 persons, drawing people from all over the world, from Kenya to California. This conference has a cutting-edge scientific program as well as a student professional development program that includes mentor-protégé match-making and a recruiting fair.
Throughout its history NSBP has had an active interest in physics related activities outside of the United States. Twenty years ago under the leadership of the late Nobel Laureate, Abdus Salam, and the late Charles Brown, several NSBP members organized the Edward Bouchet Abdus Salam Institute (EBASI). EBASI (1) provides mechanisms for synergistic scientific and technical collaborations between African and American physical scientists, engineers, and technologists, (2) enhances the impact of science and technology on the sustainable development of the countries on the African continent, and (3) increases the technical manpower pool working in Africa today by facilitating the training of Ph.D. students from African universities. M |
https://en.wikipedia.org/wiki/Nordic%20Data%20Grid%20Facility | The Nordic Data Grid Facility, or NDGF, is a common e-Science infrastructure provided by the Nordic countries (Denmark, Finland, Norway, Sweden and Iceland) for scientific computing and data storage. It is the first and so far only internationally distributed WLCG Tier1 center, providing computing and storage services to experiments at CERN.
History
Nordic Data Grid Facility traces its history back to end-2001, being intrinsically related to the NorduGrid project. Success of the latter indicated need for a larger pan-Nordic facility, with storage resources being of high priority. This need has been addressed by establishing a pilot NDGF infrastructure, which was operational in 2002-2005, and provided distributed storage in addition to the NorduGrid computing resources. During this phase, NDGF committed to provide a Nordic Tier1 (regional computing center) for the Worldwide LHC Computing Grid project at CERN. Specifics of this Tier1 are such that it has to be an internationally distributed Facility. The Nordic Data Grid Facility in its present function as a provider of the Nordic Grid Infrastructure was established in April 2006 by the Nordic Research Councils. It came into operation on June 1, 2006, and its initial priority is to live up to the original commitment of establishing the Nordic Tier1, with the traditional focus on storage facilities. NDGF team includes software experts who take part in various Grid middleware development.
In 2012 NDGF became a part of a wider initiative, the Nordic e-Infrastructure Collaboration.
Users and operations
NDGF Tier1 is a production Grid facility that leverages existing, national computational resources and Grid infrastructures.
To qualify for support research groups should form a Virtual Organization, a VO. The VO provides compute resources for sharing and NDGF Tier1 operates a Grid interface for the sharing of these resources.
Currently, most computational resources of NDGF Tier1 are accessible through ARC middlewa |
https://en.wikipedia.org/wiki/Antigen%20transfer%20in%20the%20thymus | Antigen transfer in the thymus is the transmission of self-antigens between thymic antigen-presenting cells which contributes to the establishment of T cell central tolerance.
Thymus represents an origin of T cell development and its responsibility is to select functional but also safe T cells which will not attack self tissues. Self-harmful T cells, further referred to as autoreactive T cells, originate in the thymus because of the stochastic process called V(D)J recombination which conducts the generation of T cell receptors (TCRs) and enables their limitless variability. Two processes of central tolerance take place in thymic medulla, namely clonal deletion (recessive tolerance) and T Regulatory cells selection (dominant tolerance) which force autoreactive T cells to apoptosis or skew them into suppressor T regulatory cells (TRegs), respectively, in order to protect body against manifestations of autoimmunity.
These processes are mediated especially by unique subset of stromal cells called Medullary thymic epithelial cells (mTECs) via presentation of Tissue restricted antigens (TRAs) that represent self tissues from almost all parts of the body.
mTECs
mTECs are not only capable to present TRAs as efficient APCs. They are also potent in production of these TRAs via unique process called promiscuous gene expression (PGE) and might serve as their reservoir.
Drawbacks of antigen presentation
mTECs as APCs reveal some drawbacks on population level. Their numbers in thymic medulla reach only 100,000 per 2-week-old thymus. Furthermore, average lifespan of mTECs does not exceed 2–3 days, probably due to only known PGE activator Autoimmune regulator (Aire), which requires for its proper function generation of DNA double strand breaks. And last but not least, each TRA is expressed only by 1-3% of mTEC population. These facts decrease the chance of efficient recessive or dominant tolerance.
Relevance of antigen transfer
Unidirectional spreading of mTEC-derived TRAs |
https://en.wikipedia.org/wiki/Vigilante%20%28video%20game%29 | is a 1988 beat 'em up arcade video game developed and published by Irem in Japan and Europe, and published in North America by Data East. It is considered as a spiritual sequel to Irem's earlier Kung-Fu Master (1984).
Plot
The game takes place in downtown New York City. The game's plot involves a lone, professional martial artist who became a vigilante to fight an evil gang called the Skinheads ruled by a man known as the Giant Devil, in order to protect his "turf" and save a female hostage named Madonna, who was kidnapped by them.
Gameplay
Players control the titular character using punches and kicks to defeat the Skinheads in a 2D platform manner, while sometimes picking up and using nunchaku against them. If players get hurt while holding nunchuku, they become unarmed. There are five stages in order of appearance: a street, a junkyard, the Brooklyn Bridge, a back street scene and on top of a building that is under construction. Skinheads with Mohawk or spiked hairdo attack the vigilante with knives, chains, motorbikes, guns and other kinds of weapons. They will also choke him if he lets them get too close.
Development
An arcade sequel to Kung-Fu Master called Beyond Kung-Fu: Return of the Master was developed by Irem and underwent location testing in 1987, but was shelved after it underperformed. The Kung-Fu sequel was then revamped into Vigilante, after Irem decided to give the game a more Americanized setting, which was released in 1988.
Ports
The arcade game was later ported to several different home computers and consoles. The Master System version was published exclusively in North America and Europe by Sega, and is one of several games in the console to include an FM sound switch for enhanced music quality. In the Master System version, Madonna was renamed "Maria" and the Skinheads were called the "Rogues".
The ones for Commodore 64, ZX Spectrum, Atari ST, Amiga and the Amstrad CPC were reprogrammed by Emerald Software and published by U.S. Gold mostl |
https://en.wikipedia.org/wiki/Bernoulli%20quadrisection%20problem | In triangle geometry, the Bernoulli quadrisection problem asks how to divide a given triangle into four equal-area pieces by two perpendicular lines. Its solution by Jacob Bernoulli was published in 1687. Leonhard Euler formulated a complete solution in 1779.
As Euler proved, in a scalene triangle, it is possible to find a subdivision of this form so that two of the four crossings of the lines and the triangle lie on the middle edge of the triangle, cutting off a triangular area from that edge and leaving the other three areas as quadrilaterals. It is also possible for some triangles to be subdivided differently, with two crossings on the shortest of the three edges; however, it is never possible for two crossings to lie on the longest edge. Among isosceles triangles, the one whose height at its apex is 8/9 of its base length is the only one with exactly two perpendicular quadrisections. One of the two uses the symmetry axis as one of the two perpendicular lines, while the other has two lines of slope , each crossing the base and one side.
This subdivision of a triangle is a special case of a theorem of Richard Courant and Herbert Robbins that any plane area can be subdivided into four equal parts by two perpendicular lines, a result that is related to the ham sandwich theorem. Although the triangle quadrisection has a solution involving the roots of low-degree polynomials, the more general quadrisection of Courant and Robbins can be significantly more difficult: for any computable number there exist convex shapes whose boundaries can be accurately approximated to within any desired error in polynomial time, with a unique perpendicular quadrisection whose construction computes .
In 2022, the first place in an Irish secondary school science competition, the Young Scientist and Technology Exhibition, went to a project by Aditya Joshi and Aditya Kumar using metaheuristic methods to find numerical solutions to the Bernoulli quadrisection problem.
Notes and referenc |
https://en.wikipedia.org/wiki/Essential%20Biodiversity%20Variables | Essential Biodiversity Variables (EBVs) is a putative set of parameters intended to be the minimum set of broadly agreed upon necessary and sufficient biodiversity variables for at least national to global monitoring, researching, and forecasting of biodiversity. They are being developed by an interdisciplinary group of governmental and academic research partners. The initiative aims for a harmonised global biodiversity monitoring system. EBVs would be used to inform biodiversity change indicators, such as the CBD Biodiversity Indicators for the Aichi Targets.
The concept is partly based on the earlier Essential Climate Variables. It can be generalised as the minimum set of variables for describing and predicting a system's state and dynamics. Areas with more developed EV lists include climate, ocean, and biodiversity.
EBV Classes / Categories
The current candidate EBVs occupy six classes of Essential Biodiversity Variable: genetic composition, species populations, species traits, community composition, ecosystem structure, and ecosystem function. Within each class are a few to several variables.
Associated projects and organisations
As of 2017, participants in the project consist of the GlobDiversity project (funded by the European Space Agency) under GEO BON (Group on Earth Observations Biodiversity Observation Network; a cooperative project of international universities), and the GLOBIS-B project (Global Infrastructures for Supporting Biodiversity Research; funded by the EU Horizon 2020 programme)
Development
The concept was first proposed in 2012 and developed in the following years.
The GLOBIS-B global cooperation project, aimed to advance the challenge of practical implementation of EBVs by supporting interoperability and cooperation activities among diverse biodiversity infrastructures, started in 2015. The GlobDiversity project of GEO BON, led by the University of Zurich, started in 2017, focusing on specification and engineering of three RS-enabled |
https://en.wikipedia.org/wiki/DC%20Universe%20Online | DC Universe Online (DCUO) is a free-to-play action combat massively multiplayer online game set in the fictional universe of DC Comics. Developed by Dimensional Ink Games and co-published by Daybreak Game Company and WB Games, the game was initially released in January 2011 for Windows and PlayStation 3. It would later release in November 2013 for PlayStation 4, April 2016 for Xbox One and August 2019 for Nintendo Switch.
DC Universe Online is Arleen Sorkin's final voice performance as Harley Quinn, before her death in August 24, 2023.
Gameplay
The player creates a new, original character that interacts with the iconic heroes and villains of DC Comics. Players choose their character's faction (Hero or Villain), gender (male or female), body type (short, medium, tall, spry, athletic, large), personality (comical, flirty, powerful, primal, serious), movement mode (flight, acrobatics, speed or skimming), weapon, and power (fire, ice, gadgets, mental, nature, sorcery, earth, light, electricity, rage, quantum, celestial, munitions, atomic, and water). Numerous hair, skin, and costume types are available, and up to 4 colors can be applied to the color scheme palette. Pre-built templates, inspired by some key DC characters, are available to expedite the character creation process.
When the player's character is named, they are thrown into the world of DC Universe Online with the first experience having to fight their way out of a Brainiac spaceship. The tutorial teaches basic movement and abilities, counter mechanics, and using power and skill points. Eventually the player's character is made an official member of the Justice League (heroes) or The Society (villains), where they continue on their missions to increase their level and skill in various content. Daybreak Game Company is working to make DC Universe Online more interactive than standard MMO games, while trying to keep their key elements which include a leveling system, raid instances, endgame progression and |
https://en.wikipedia.org/wiki/Glycine%20receptor | The glycine receptor (abbreviated as GlyR or GLR) is the receptor of the amino acid neurotransmitter glycine. GlyR is an ionotropic receptor that produces its effects through chloride current. It is one of the most widely distributed inhibitory receptors in the central nervous system and has important roles in a variety of physiological processes, especially in mediating inhibitory neurotransmission in the spinal cord and brainstem.
The receptor can be activated by a range of simple amino acids including glycine, β-alanine and taurine, and can be selectively blocked by the high-affinity competitive antagonist strychnine. Caffeine is a competitive antagonist of GlyR. Cannabinoids enhance the function.
The protein Gephyrin has been shown to be necessary for GlyR clustering at inhibitory synapses. GlyR is known to colocalize with the GABAA receptor on some hippocampal neurons. Nevertheless, some exceptions can occur in the central nervous system where the GlyR α1 subunit and gephyrin, its anchoring protein, are not found in dorsal root ganglion neurons despite the presence of GABAA receptors.
History
Glycine and its receptor were first suggested to play a role in inhibition of cells in 1965. Two years later, experiments showed that glycine had a hyperpolarizing effect on spinal motor neurons due to increased chloride conductance through the receptor. Then, in 1971, glycine was found to be localized in the spinal cord using autoradiography. All of these discoveries resulted in the conclusion that glycine is a primary inhibitory neurotransmitter of the spinal cord that works via its receptor.
Arrangement of subunits
Strychnine-sensitive GlyRs are members of a family of ligand-gated ion channels. Receptors of this family are arranged as five subunits surrounding a central pore, with each subunit composed of four α helical transmembrane segments. There are presently four known isoforms of the ligand-binding α-subunit (α1-4) of GlyR (GLRA1, GLRA2, GLRA3, GLRA4) and a |
https://en.wikipedia.org/wiki/Glossary%20of%20arithmetic%20and%20diophantine%20geometry | This is a glossary of arithmetic and diophantine geometry in mathematics, areas growing out of the traditional study of Diophantine equations to encompass large parts of number theory and algebraic geometry. Much of the theory is in the form of proposed conjectures, which can be related at various levels of generality.
Diophantine geometry in general is the study of algebraic varieties V over fields K that are finitely generated over their prime fields—including as of special interest number fields and finite fields—and over local fields. Of those, only the complex numbers are algebraically closed; over any other K the existence of points of V with coordinates in K is something to be proved and studied as an extra topic, even knowing the geometry of V.
Arithmetic geometry can be more generally defined as the study of schemes of finite type over the spectrum of the ring of integers. Arithmetic geometry has also been defined as the application of the techniques of algebraic geometry to problems in number theory.
A
B
C
D
E
F
G
H
I
K
L
M
N
O
Q
R
S
T
U
V
W
See also
Arithmetic topology
Arithmetic dynamics |
https://en.wikipedia.org/wiki/Domestication%20of%20the%20goat | Goat evolution is the process by which domestic goats came to exist through evolution by natural selection. Wild goats — medium-sized mammals which are found in noticeably harsh environments, particularly forests and mountains, in the Middle East and Central Asia — were one of the first species domesticated by modern humans, with the date of domestication generally considered to be 8,000 BCE. Goats are part of the family Bovidae, a broad and populous group which includes a variety of ruminants such as bison, cows and sheep. Bovids all share many traits, such as hooves and a herbivorous diet and all males, along with many females, have horns. Bovids began to diverge from deer and giraffids during the early Miocene epoch. The subfamily Caprinae, which includes goats, ibex and sheep, are considered to have diverged from the rest of Bovidae as early as the late Miocene, with the group reaching its greatest diversity in the ice ages.
The tribe Caprini would subsequently develop from Caprids who arrived in the mountainous areas of Eurasia and split into goats and sheep in response to a further geographic separation. The ancestors of sheep remained in the foothills and the ancestors of goats went to higher altitudes. This divergence resulted in the adaption of the ancestors of goats to a mountainous environment, producing many of the traits considered peculiar to the species. During the ice ages a genus called Capri evolved which would then diverge into the modern goat species, along with several species of ibex.
It is commonly held that the earliest domestication was of the bezoar ibex in the Zagros Mountains. These earliest domesticated goats were used to produce meat and milk for Neolithic farmers, along with providing many of the materials required to build residences and tools. Following the domestication of goats over 300 breeds have been established for a variety of purposes, including for the maximation of milk production and for meat. Domestication and the selec |
https://en.wikipedia.org/wiki/Fredholm%20determinant | In mathematics, the Fredholm determinant is a complex-valued function which generalizes the determinant of a finite dimensional linear operator. It is defined for bounded operators on a Hilbert space which differ from the identity operator by a trace-class operator. The function is named after the mathematician Erik Ivar Fredholm.
Fredholm determinants have had many applications in mathematical physics, the most celebrated example being Gábor Szegő's limit formula, proved in response to a question raised by Lars Onsager and C. N. Yang on the spontaneous magnetization of the Ising model.
Definition
Let be a Hilbert space and the set of bounded invertible operators on of the form
, where is a trace-class operator. is a group because
so is trace class if is. It has a natural metric given by
, where is the trace-class norm.
If is a Hilbert space with inner product , then so too is the th exterior power
with inner product
In particular
gives an orthonormal basis of if is an orthonormal basis of .
If is a bounded operator on , then functorially defines a bounded operator on by
If is trace-class, then is also trace-class with
This shows that the definition of the Fredholm determinant given by
makes sense.
Properties
If is a trace-class operator
defines an entire function such that
The function is continuous on trace-class operators, with
One can improve this inequality slightly to the following, as noted in Chapter 5 of Simon:
If and are trace-class then
The function defines a homomorphism of into the multiplicative group of nonzero complex numbers (since elements of are invertible).
If is in and is invertible,
If is trace-class, then
Fredholm determinants of commutators
A function from into is said to be differentiable if is differentiable as a map into the trace-class operators, i.e. if the limit
exists in trace-class norm.
If is a differentiable function with values in trace-class operators, then so |
https://en.wikipedia.org/wiki/Landau%27s%20function | In mathematics, Landau's function g(n), named after Edmund Landau, is defined for every natural number n to be the largest order of an element of the symmetric group Sn. Equivalently, g(n) is the largest least common multiple (lcm) of any partition of n, or the maximum number of times a permutation of n elements can be recursively applied to itself before it returns to its starting sequence.
For instance, 5 = 2 + 3 and lcm(2,3) = 6. No other partition of 5 yields a bigger lcm, so g(5) = 6. An element of order 6 in the group S5 can be written in cycle notation as (1 2) (3 4 5). Note that the same argument applies to the number 6, that is, g(6) = 6. There are arbitrarily long sequences of consecutive numbers n, n + 1, …, n + m on which the function g is constant.
The integer sequence g(0) = 1, g(1) = 1, g(2) = 2, g(3) = 3, g(4) = 4, g(5) = 6, g(6) = 6, g(7) = 12, g(8) = 15, ... is named after Edmund Landau, who proved in 1902 that
(where ln denotes the natural logarithm). Equivalently (using little-o notation), .
The statement that
for all sufficiently large n, where Li−1 denotes the inverse of the logarithmic integral function, is equivalent to the Riemann hypothesis.
It can be shown that
with the only equality between the functions at n = 0, and indeed
Notes |
https://en.wikipedia.org/wiki/Index%20of%20meteorology%20articles | This is a list of meteorology topics. The terms relate to meteorology, the interdisciplinary scientific study of the atmosphere that focuses on weather processes and forecasting. (see also: List of meteorological phenomena)
A
advection
aeroacoustics
aerobiology
aerography (meteorology)
aerology
air parcel (in meteorology)
air quality index (AQI)
airshed (in meteorology)
American Geophysical Union (AGU)
American Meteorological Society (AMS)
anabatic wind
anemometer
annular hurricane
anticyclone (in meteorology)
apparent wind
Atlantic Oceanographic and Meteorological Laboratory (AOML)
Atlantic hurricane season
atmometer
atmosphere
Atmospheric Model Intercomparison Project (AMIP)
Atmospheric Radiation Measurement (ARM)
(atmospheric boundary layer [ABL]) planetary boundary layer (PBL)
atmospheric chemistry
atmospheric circulation
atmospheric convection
atmospheric dispersion modeling
atmospheric electricity
atmospheric icing
atmospheric physics
atmospheric pressure
atmospheric sciences
atmospheric stratification
atmospheric thermodynamics
atmospheric window (see under Threats)
B
ball lightning
balloon (aircraft)
baroclinity
barotropity
barometer ("to measure atmospheric pressure")
berg wind
biometeorology
blizzard
bomb (meteorology)
buoyancy
Bureau of Meteorology (in Australia)
C
Canada Weather Extremes
Canadian Hurricane Centre (CHC)
Cape Verde-type hurricane
capping inversion (in meteorology) (see "severe thunderstorms" in paragraph 5)
carbon cycle
carbon fixation
carbon flux
carbon monoxide (see under Atmospheric presence)
ceiling balloon ("to determine the height of the base of clouds above ground level")
ceilometer ("to determine the height of a cloud base")
celestial coordinate system
celestial equator
celestial horizon (rational horizon)
celestial navigation (astronavigation)
celestial pole
Celsius
Center for Analysis and Prediction of Storms (CAPS) (in Oklahoma in the US)
Center for the Study o |
https://en.wikipedia.org/wiki/Lactarius%20salmoneus | Lactarius salmoneus is an edible species of fungus belonging to the genus Lactarius, and classified under the family Russulaceae. It is native to North America. L. salmoneus is not to be confused with Lactarius salmonicolor, a species native to Europe.
Description
The mushroom has a white cap, with vividly orange lamellas that can be decurrent or adnate. The cap is slightly depressed in the center, and becomes reddish when bruised. The stipe is white and velvety but has orange flesh inside. It grows usually in waterlogged soil. |
https://en.wikipedia.org/wiki/High-frequency%20Active%20Auroral%20Research%20Program | The High-frequency Active Auroral Research Program (HAARP) is a University of Alaska Fairbanks program which researches the ionosphere — the highest, ionized part of Earth's atmosphere.
The most prominent instrument at HAARP is the Ionospheric Research Instrument (IRI), a high-power radio frequency transmitter facility operating in the high frequency (HF) band. The IRI is used to temporarily excite a limited area of the ionosphere. Other instruments, such as a VHF and a UHF radar, a fluxgate magnetometer, a digisonde (an ionospheric sounding device), and an induction magnetometer, are used to study the physical processes that occur in the excited region.
Initially HAARP was jointly funded by the U.S. Air Force, the U.S. Navy, the University of Alaska Fairbanks, and the Defense Advanced Research Projects Agency (DARPA). It was designed and built by BAE Advanced Technologies. Its original purpose was to analyze the ionosphere and investigate the potential for developing ionospheric enhancement technology for radio communications and surveillance. Since 2015 it has been operated by the University of Alaska Fairbanks.
Work on the HAARP facility began in 1993. The current working IRI was completed in 2007; its prime contractor was BAE Systems Advanced Technologies. As of 2008, HAARP had incurred around $250 million in tax-funded construction and operating costs. In May 2014, it was announced that the HAARP program would be permanently shut down later in the year. After discussions between the parties, ownership of the facility was transferred to the University of Alaska Fairbanks in August 2015.
HAARP is a target of conspiracy theorists, who claim that it is capable of weather manipulation. Commentators and scientists say that advocates of this and other conspiracy theories are wrong, as claims made fall well outside the abilities of the facility, if not the scope of natural science.
History
The High-frequency Active Auroral Research Program began in 1990. Ted Stev |
https://en.wikipedia.org/wiki/Kite%20%28geometry%29 | In Euclidean geometry, a kite is a quadrilateral with reflection symmetry across a diagonal. Because of this symmetry, a kite has two equal angles and two pairs of adjacent equal-length sides. Kites are also known as deltoids, but the word deltoid may also refer to a deltoid curve, an unrelated geometric object sometimes studied in connection with quadrilaterals. A kite may also be called a dart, particularly if it is not convex.
Every kite is an orthodiagonal quadrilateral (its diagonals are at right angles) and, when convex, a tangential quadrilateral (its sides are tangent to an inscribed circle). The convex kites are exactly the quadrilaterals that are both orthodiagonal and tangential. They include as special cases the right kites, with two opposite right angles; the rhombi, with two diagonal axes of symmetry; and the squares, which are also special cases of both right kites and rhombi.
The quadrilateral with the greatest ratio of perimeter to diameter is a kite, with 60°, 75°, and 150° angles. Kites of two shapes (one convex and one non-convex) form the prototiles of one of the forms of the Penrose tiling. Kites also form the faces of several face-symmetric polyhedra and tessellations, and have been studied in connection with outer billiards, a problem in the advanced mathematics of dynamical systems.
Definition and classification
A kite is a quadrilateral with reflection symmetry across one of its diagonals. Equivalently, it is a quadrilateral whose four sides can be grouped into two pairs of adjacent equal-length sides. A kite can be constructed from the centers and crossing points of any two intersecting circles. Kites as described here may be either convex or concave, although some sources restrict kite to mean only convex kites. A quadrilateral is a kite if and only if any one of the following conditions is true:
The four sides can be split into two pairs of adjacent equal-length sides.
One diagonal crosses the midpoint of the other diagonal at a righ |
https://en.wikipedia.org/wiki/Ellman%27s%20reagent | Ellman's reagent (5,5′-dithiobis-(2-nitrobenzoic acid) or DTNB) is a colorogenic chemical used to quantify the number or concentration of thiol groups in a sample. It was developed by George L. Ellman.
Preparation
In Ellman's original paper, he prepared this reagent by oxidizing 2-nitro-5-chlorobenzaldehyde to the carboxylic acid, introducing the thiol via sodium sulfide, and coupling the monomer by oxidization with iodine. Today, this reagent is readily available commercially.
Ellman's test
Thiols react with this compound, cleaving the disulfide bond to give 2-nitro-5-thiobenzoate (TNB−), which ionizes to the TNB2− dianion in water at neutral and alkaline pH. This TNB2− ion has a yellow color.
This reaction is rapid and stoichiometric, with the addition of one mole of thiol releasing one mole of TNB. The TNB2− is quantified in a spectrophotometer by measuring the absorbance of visible light at 412 nm, using an extinction coefficient of 14,150 M−1 cm−1 for dilute buffer solutions, and a coefficient of 13,700 M−1 cm−1 for high salt concentrations, such as 6 M guanidinium hydrochloride or 8 M urea. Ellman's original 1959 publication estimated the molar extinction at 13,600 M−1 cm−1, and this value can be found in some modern applications of the method despite improved determinations. Commercial DTNB may not be completely pure, so may require recrystallization to obtain completely accurate and reproducible results.
Ellman's reagent can be used for measuring low-molecular mass thiols such as glutathione in both pure solutions and biological samples, such as blood. It can also measure the number of thiol groups on proteins. |
https://en.wikipedia.org/wiki/Sal%20Khan | Salman "Sal" Amin Khan (born October 11, 1976) is an American educator and the founder of Khan Academy, a free online non-profit educational platform with which he has produced over 6,500 video lessons teaching a wide spectrum of academic subjects, originally focusing on mathematics and science. He is also the founder of Khan Lab School, a private brick-and-mortar school in Mountain View, California.
, the Khan Academy channel on YouTube has 7.94 million subscribers, and its videos have been viewed more than two billion times. In 2012, Khan was named in the annual publication of Time 100. In the same year, he was featured on the cover of Forbes, with the tagline "The $1 Trillion Opportunity."
Early life and family
Khan was born in Metairie, Louisiana, on October 11, 1976, into a Bengali Muslim family. His father Fakhrul Amin Khan (d. 1990) was a physician, originally hailing from the village of Rahmatpur in Babuganj Upazila, Bangladesh. His mother Masuda Khan is from Murshidabad in West Bengal, India. He grew up poor, recalling that his mother made $16,000 in 1993—he knew this because he had to do her taxes to get financial aid. They are descendants of Rahmat Khan, a 16th-century Pashtun chieftain who was killed in battle with Kandarpanarayan Rai, the erstwhile Raja of Chandradwip. Khan's grandfather Abdul Wahab Khan was a prominent Pakistani politician who served as the country's second official Speaker of the National Assembly.
Education
He attended Grace King High School, where, as he recalls, "a few classmates were fresh out of jail and others were bound for top universities." He was a cartoonist for the high school's newspaper. Khan took upper-level mathematics courses at the University of New Orleans while he was in high school and graduated as valedictorian in 1994.
He attended the Massachusetts Institute of Technology (MIT), graduating in 1998 with Bachelor's and Master of Science degrees in electrical engineering and computer science, and another bach |
https://en.wikipedia.org/wiki/Gyrokinetics | Gyrokinetics is a theoretical framework to study plasma behavior on perpendicular spatial scales comparable to the gyroradius and frequencies much lower than the particle cyclotron frequencies.
These particular scales have been experimentally shown to be appropriate for modeling plasma turbulence. The trajectory of charged particles in a magnetic field is a helix that winds around the field line. This trajectory can be decomposed into a relatively slow motion of the guiding center along the field line and a fast circular motion, called gyromotion. For most plasma behavior, this gyromotion is irrelevant. Averaging over this gyromotion reduces the equations to six dimensions (3 spatial, 2 velocity, and time) rather than the seven (3 spatial, 3 velocity, and time). Because of this simplification, gyrokinetics governs the evolution of charged rings with a guiding center position, instead of gyrating charged particles.
Derivation of the gyrokinetic equation
Fundamentally, the gyrokinetic model assumes the plasma is strongly magnetized ( ), the perpendicular spatial scales are comparable to the gyroradius ( ), and the behavior of interest has low frequencies ( ). We must also expand the distribution function, , and assume the perturbation is small compared to the background (). The starting point is the Fokker–Planck equation and Maxwell's equations. The first step is to change spatial variables from the particle position to the guiding center position . Then, we change velocity coordinates from to the velocity parallel , the magnetic moment , and the gyrophase angle . Here parallel and perpendicular are relative to , the direction of the magnetic field, and is the mass of the particle. Now, we can average over the gyrophase angle at constant guiding center position, denoted by , yielding the gyrokinetic equation.
The electrostatic gyrokinetic equation, in the absence of large plasma flow, is given by
.
Here the first term represents the change in the perturb |
https://en.wikipedia.org/wiki/Decoupling%20%28cosmology%29 | In cosmology, decoupling is a period in the development of the universe when different types of particles fall out of thermal equilibrium with each other. This occurs as a result of the expansion of the universe, as their interaction rates decrease (and mean free paths increase) up to this critical point. The two verified instances of decoupling since the Big Bang which are most often discussed are photon decoupling and neutrino decoupling, as these led to the cosmic microwave background and cosmic neutrino background, respectively.
Photon decoupling is closely related to recombination, which occurred about 378,000 years after the Big Bang (at a redshift of z = ), when the universe was a hot opaque ("foggy") plasma. During recombination, free electrons became bound to protons (hydrogen nuclei) to form neutral hydrogen atoms. Because direct recombinations to the ground state (lowest energy) of hydrogen are very inefficient, these hydrogen atoms generally form with the electrons in a high energy state, and the electrons quickly transition to their low energy state by emitting photons. Because the neutral hydrogen that formed was transparent to light, those photons which were not captured by other hydrogen atoms were able, for the first time in the history of the universe, to travel long distances. They can still be detected today, although they now appear as radio waves, and form the cosmic microwave background ("CMB"). They reveal crucial clues about how the universe formed.
Photon decoupling
Photon decoupling occurred during the epoch known as the recombination. During this time, electrons combined with protons to form hydrogen atoms, resulting in a sudden drop in free electron density. Decoupling occurred abruptly when the rate of Compton scattering of photons was approximately equal to the rate of expansion of the universe , or alternatively when the mean free path of the photons was approximately equal to the horizon size of the universe . After this photon |
https://en.wikipedia.org/wiki/Modeshape | In applied mathematics, mode shapes are a manifestation of eigenvectors which describe the relative displacement of two or more elements in a mechanical system or wave front.
A mode shape is a deflection pattern related to a particular natural frequency and represents the relative displacement of all parts of a structure for that particular mode.
See also
Normal mode
Harmonic oscillator |
https://en.wikipedia.org/wiki/Materiomics | Materiomics is the holistic study of material systems. Materiomics examines links between physicochemical material properties and material characteristics and function. The focus of materiomics is system functionality and behavior, rather than a piecewise collection of properties, a paradigm similar to systems biology. While typically applied to complex biological systems and biomaterials, materiomics is equally applicable to non-biological systems. Materiomics investigates the material properties of natural and synthetic materials by examining fundamental links between processes, structures and properties at multiple scales, from nano to macro, by using systematic experimental, theoretical or computational methods.
The term has been independently proposed with slightly different definitions in 2004 by T. Akita et al. (AIST/Japan), in 2008 by Markus J. Buehler (MIT/USA), and Clemens van Blitterswijk, Jan de Boer and Hemant Unadkat (University of Twente/The Netherlands) in analogy to genomics, the study of an organism's entire genome. Similarly, materiomics refers to the study of the processes, structures and properties of materials from a fundamental, systematic perspective by incorporating all relevant scales, from nano to macro, in the synthesis and function of materials and structures. The integrated view of these interactions at all scales is referred to as a material's materiome.
New techniques for evaluating materials at the tissue level, such as reference point indentation (RPI) and raman spectroscopy are lending insight into the nature of these highly complex, functional relationships.
Materiomics is related to proteomics, where the difference is the focus on material properties, stability, failure and mechanistic insight into multi-scale phenomena.
See also
Genomics
Bionanotechnology
Universality-diversity paradigm
Notes
Other References
: Buehler, M.J., Materiomics: Materials Science of Biological Protein Materials, from Nano to Macro. The |
https://en.wikipedia.org/wiki/List%20of%20artificial%20objects%20leaving%20the%20Solar%20System | The artificial objects leaving the Solar System are all space probes and the upper stages of their launch vehicles, all launched by NASA. Three of the probes, Voyager 1, Voyager 2, and New Horizons are still functioning and are regularly contacted by radio communication, while Pioneer 10 and Pioneer 11 are now derelict. In addition to these spacecraft, some upper stages and de-spin weights are leaving the Solar System, assuming they continue on their trajectories.
These objects are leaving the Solar System because their velocity and direction are taking them away from the Sun, and at their distance from the Sun, its gravitational pull is not sufficient to pull these objects back or into orbit. They are not impervious to the gravitational pull of the Sun and are being slowed, but are still traveling in excess of escape velocity to leave the Solar System and coast into interstellar space.
Planetary exploration probes
Pioneer 10 – launched in 1972, flew past Jupiter in 1973 and is heading in the direction of Aldebaran (65 light years away) in the constellation of Taurus. Contact was lost in January 2003, and it is estimated to have passed 134 astronomical units (AU; one AU is roughly the average distance between Earth and the Sun: 150 million kilometers (93 million miles)).
Pioneer 11 – launched in 1973, flew past Jupiter in 1974 and Saturn in 1979. Contact was lost in November 1995, and it is estimated to be at 111 AU. The spacecraft is heading toward the constellation of Aquila, northwest of the constellation of Sagittarius. Barring an incident, Pioneer 11 will pass near one of the stars in the constellation in about 4 million years.
Voyager 2 – launched in August 1977, flew past Jupiter in 1979, Saturn in 1981, Uranus in 1986, and Neptune in 1989. The probe left the heliosphere for interstellar space at 119 AU on 5 November 2018. Voyager 2 is still active. It is not heading toward any particular star, although in roughly 40,000 years it should pass 1.7 light-yea |
https://en.wikipedia.org/wiki/Squirrel%20%28programming%20language%29 | Squirrel is a high level imperative, object-oriented programming language, designed to be a lightweight scripting language that fits in the size, memory bandwidth, and real-time requirements of applications like video games.
MirthKit, a simple toolkit for making and distributing open source, cross-platform 2D games, uses Squirrel for its platform. It is used extensively by Code::Blocks for scripting and was also used in Final Fantasy Crystal Chronicles: My Life as a King. It is also used in Left 4 Dead 2, Portal 2 and Thimbleweed Park for scripted events and in NewDark, an unofficial Thief 2: The Metal Age engine update, to facilitate additional, simplified means of scripting mission events, aside of the regular C scripting.
Language features
Dynamic typing
Delegation
Classes, inheritance
Higher order functions
Generators
Cooperative threads (coroutines)
Tail recursion
Exception handling
Automatic memory management (mainly reference counting with backup garbage collector)
Weak references
Both compiler and virtual machine fit together in about 7k lines of C++ code
Optional 16-bit character strings
Syntax
Squirrel uses a C-like syntax.
Factorial in Squirrel
function factorial(x)
{
if (x <= 1) {
return 1;
}
else {
return x * factorial(x-1);
}
}
Generators
function not_a_random_number_generator(max) {
local last = 42;
local IM = 139968;
local IA = 3877;
local IC = 29573;
for(;;) { // loops forever
yield (max * (last = (last * IA + IC) % IM) / IM);
}
}
local randtor = not_a_random_number_generator(100);
for(local i = 0; i < 10; i += 1)
print(">"+resume randtor+"\n");
Classes and inheritance
class BaseVector {
constructor(...)
{
if(vargv.len() >= 3) {
x = vargv[0];
y = vargv[1];
z = vargv[2];
}
}
x = 0;
y = 0;
z = 0;
}
class Vector3 extends BaseVector {
function _add(other)
{
if(other instanceof ::Vector3)
return ::Vector3(x+other.x,y+other.y,z+other.z);
else
throw "wro |
https://en.wikipedia.org/wiki/Fenchel%27s%20duality%20theorem | In mathematics, Fenchel's duality theorem is a result in the theory of convex functions named after Werner Fenchel.
Let ƒ be a proper convex function on Rn and let g be a proper concave function on Rn. Then, if regularity conditions are satisfied,
where ƒ * is the convex conjugate of ƒ (also referred to as the Fenchel–Legendre transform) and g * is the concave conjugate of g. That is,
Mathematical theorem
Let X and Y be Banach spaces, and be convex functions and be a bounded linear map. Then the Fenchel problems:
satisfy weak duality, i.e. . Note that are the convex conjugates of f,g respectively, and is the adjoint operator. The perturbation function for this dual problem is given by .
Suppose that f,g, and A satisfy either
f and g are lower semi-continuous and where is the algebraic interior and , where h is some function, is the set , or
where are the points where the function is continuous.
Then strong duality holds, i.e. . If then supremum is attained.
One-dimensional illustration
In the following figure, the minimization problem on the left side of the equation is illustrated. One seeks to vary x such that the vertical distance between the convex and concave curves at x is as small as possible. The position of the vertical line in the figure is the (approximate) optimum.
The next figure illustrates the maximization problem on the right hand side of the above equation. Tangents are drawn to each of the two curves such that both tangents have the same slope p. The problem is to adjust p in such a way that the two tangents are as far away from each other as possible (more precisely, such that the points where they intersect the y-axis are as far from each other as possible). Imagine the two tangents as metal bars with vertical springs between them that push them apart and against the two parabolas that are fixed in place.
Fenchel's theorem states that the two problems have the same solution. The points having the minimum vertical sepa |
https://en.wikipedia.org/wiki/Nocturnal%20enuresis | Nocturnal enuresis, also informally called bedwetting, is involuntary urination while asleep after the age at which bladder control usually begins. Bedwetting in children and adults can result in emotional stress. Complications can include urinary tract infections.
Most bedwetting is a developmental delay—not an emotional problem or physical illness. Only a small percentage (5 to 10%) of bedwetting cases have a specific medical cause. Bedwetting is commonly associated with a family history of the condition. Nocturnal enuresis is considered primary when a child has not yet had a prolonged period of being dry. Secondary nocturnal enuresis is when a child or adult begins wetting again after having stayed dry.
Treatments range from behavioral therapy, such as bedwetting alarms, to medication, such as hormone replacement, and even surgery such as urethral dilatation. Since most bedwetting is simply a developmental delay, most treatment plans aim to protect or improve self-esteem. Treatment guidelines recommend that the physician counsel the parents, warning about psychological consequences caused by pressure, shaming, or punishment for a condition children cannot control.
Bedwetting is the most common childhood complaint.
Impact
A review of medical literature shows doctors consistently stressing that a bedwetting child is not at fault for the situation. Many medical studies state that the psychological impacts of bedwetting are more important than the physical considerations. "It is often the child's and family members' reaction to bedwetting that determines whether it is a problem or not."
Self-esteem
Whether bedwetting causes low self-esteem remains a subject of debate, but several studies have found that self-esteem improved with management of the condition.
Children questioned in one study ranked bedwetting as the third most stressful life event, after "parental war of words", divorce and parental fighting. Adolescents in the same study ranked bedwetting as tie |
https://en.wikipedia.org/wiki/Environmental%20Change%20Network | The Environmental Change Network (ECN) was established in 1992 by the Natural Environment Research Council (NERC) to monitor long-term environmental change and its effects on ecosystems at a series of sites throughout Great Britain and Northern Ireland. Measurements made include a wide range of physical, chemical and biological variables.
See also
Climate change
External links
Environmental Change Network website
Environment of the United Kingdom
Natural Environment Research Council |
https://en.wikipedia.org/wiki/Elisabeth%20Larsson%20%28scientific%20computing%29 | Elisabeth Larsson (born December 30, 1971) is a Swedish applied mathematician and numerical analyst. She is a professor in the Department of Information Technology of Uppsala University, and the director of the Uppsala Multidisciplinary Centre for Advanced Computational Science.
Research
Larsson's research involves the applications of radial basis functions to scientific computing. It has included work on the propagation of sound waves through water, pricing of financial options, and simulation of the earth's climate.
Education and career
Larsson was born on December 30, 1971 in Ljusdal, and went to high school in Ljusdal. She earned a master's degree in engineering physics at Uppsala University in 1994, and completed a Ph.D. in numerical analysis at Uppsala University in 2000. Her dissertation was Domain Decomposition and Preconditioned Iterative Methods for the Helmholtz Equation. Her doctorate was supervised by Kurt Otto, with as outside examiner.
She became a junior researcher in the Department of Information Technology at Uppsala University in 2001, and an assistant professor in 2007. She was promoted to senior lecturer (associate professor) in 2011 and professor in 2020.
Recognition
In 2007, Larsson was one of two winners of the Göran Gustafsson Award for outstanding young Swedish scientists. |
https://en.wikipedia.org/wiki/General%20Agri-Food%20Federation | The General Agri-Food Federation (, FGA) is a trade union representing workers in the agricultural and food processing sectors in France.
The union was founded in 1980, when the General Federation of Agriculture merged with the General Federation of Food. Like its predecessors, the union affiliated to the French Democratic Confederation of Labour. By 1994, the union claimed 37,300 members, and by 2017, this was almost unchanged, at 37,428.
External links |
https://en.wikipedia.org/wiki/Philippine%20condiments | The generic term for condiments in the Filipino cuisine is sawsawan (Philippine Spanish: sarsa). Unlike sauces in other Southeast Asian regions, most sawsawan are not prepared beforehand, but are assembled on the table according to the preferences of the diner.
Description
In the Philippines, the common condiments aside from salt and pepper are vinegar, soy sauce, calamansi, and patis. The combination and different regional variations of these simple sauces make up the various common dipping sauces in the region.
The most common type of sawsawan is the toyomansi (or toyo't kalamansi), which is a mixture of soy sauce, calamansi, and native Siling labuyo. It can also be seasoned with vinegar and patis (fish sauce). This sauce is typically served with roasted meat dishes.
A similar dipping sauce used for grilled meats like inihaw is toyo, suka, at sili (literally "soy sauce, vinegar, and chili"). It is made of soy sauce, vinegar, and siling labuyo with some opting to add diced onions and/or garlic and a seasoning of sugar and/or black pepper. For serving with grilled fish, it is typically garnished with diced tomatoes, patis (fish sauce), or more rarely, bagoong (fermented shrimp or fish).
The simplest dipping sauce, for example, is vinegar mixed with another ingredient like siling labuyo (sukang may sili), garlic (suka't bawang), soy sauce (sukang may toyo), and so on. This can be elaborated further by adding a range of spices and even fruits, resulting in dipping sauces like sinamak (spiced vinegar). Suka Pinakurat is a popular brand of spiced vinegar in the Philippines.
All of these do not have set recipes, however, and can use ingredients and proportions interchangeably according to what is available and to the preference of the diner.
Other notable ingredients added to these kinds of sawsawan include shallots, whole black peppercorns, sugar, siling haba, wansoy (cilantro), ginger, and so on. Sawsawan are also unique in that they can function as marinades.
S |
https://en.wikipedia.org/wiki/Poynting%20effect | The Poynting effect may refer to two unrelated physical phenomena. Neither should be confused with the Poynting–Robertson effect. All of these effects are named after John Henry Poynting, an English physicist.
Solid mechanics
In solid mechanics, the Poynting effect is a Finite strain theory effect observed when an elastic cube is sheared between two plates and stress is developed in the direction normal to the sheared faces, or when a cylinder is subjected to torsion and the axial length changes. The Poynting phenomenon in torsion was noticed experimentally by J. H. Poynting.
Chemistry and thermodynamics
In thermodynamics, the Poynting effect generally refers to the change in the fugacity of a liquid when a non-condensable gas is mixed with the vapor at saturated conditions.
Equivalently in terms of vapor pressure, if one assumes that the vapor and the non-condensable gas behave as ideal gases and an ideal mixture, it can be shown that:
where
is the modified vapor pressure
is the unmodified vapor pressure
is the liquid molar volume
is the liquid/vapor's gas constant
is the temperature
is the total pressure (vapor pressure + non-condensable gas)
A common example is the production of the medicine Entonox, a high-pressure mixture of nitrous oxide and oxygen. The ability to combine and at high pressure while remaining in the gaseous form is due to the Poynting effect. |
https://en.wikipedia.org/wiki/Formylglycine-generating%20enzyme | Formylglycine-generating enzyme (FGE), located at 3p26.1 in humans, is the name for an enzyme present in the endoplasmic reticulum that catalyzes the conversion of cysteine to formylglycine (fGly). There are two main classes of FGE, aerobic and anaerobic. FGE activates sulfatases, which are essential for the degradation of sulfate esters. The catalytic activity of sulfatases is dependent upon a formylglycine (sometimes called oxoalanine) residue in the active site.
Aerobic
The aerobic enzyme has a structure homologous to the complex alpha/beta topology found in the gene product of human sulfatase-modifying factor 1 (SUMF1). Aerobic FGE converts a cysteine residue in the highly conserved consensus sequence CXPXR to fGly. To do so, FGE “activates” its target by utilizing mononuclear copper. The substrate first binds to copper, increasing reactivity of the substrate-copper complex with oxygen. Activation is then accomplished through oxidation of a cysteine residue in the substrate-copper complex. Due to the nature of this reaction, FGE is termed a “copper-dependent metalloenzyme.
Anaerobic
The most well-studied anaerobic FGE is the bacterial AtsB, an iron-sulfur cluster containing enzyme present in Klebsiella pneumoniae, that is able to convert either cysteine or serine to fGly with a distinctly different mechanism than the aerobic form. While AtsB can convert either, its activity increases four fold when in the presence of cysteine over serine. AtsB is 48% similar to an enzyme present in Clostridium perfringens. Both enzymes possess the Cx3Cx2C motif unique to the radical S-adenosyl methionine superfamily and are able to use a reduction reaction to cleave S-adenosyl methionine. These two enzymes fall into a larger group called the anaerobic Sulfatase Maturing Enzymes, which are able to convert cysteine into fGly without the use of oxygen.
Protein domain
In molecular biology, "formylglycine-generating enzyme" (sometimes annotated as formylglycine-generating sulfata |
https://en.wikipedia.org/wiki/Optical%20aberration | In optics, aberration is a property of optical systems, such as lenses, that causes light to be spread out over some region of space rather than focused to a point. Aberrations cause the image formed by a lens to be blurred or distorted, with the nature of the distortion depending on the type of aberration. Aberration can be defined as a departure of the performance of an optical system from the predictions of paraxial optics. In an imaging system, it occurs when light from one point of an object does not converge into (or does not diverge from) a single point after transmission through the system. Aberrations occur because the simple paraxial theory is not a completely accurate model of the effect of an optical system on light, rather than due to flaws in the optical elements.
An image-forming optical system with aberration will produce an image which is not sharp. Makers of optical instruments need to correct optical systems to compensate for aberration.
Aberration can be analyzed with the techniques of geometrical optics. The articles on reflection, refraction and caustics discuss the general features of reflected and refracted rays.
Overview
With an ideal lens, light from any given point on an object would pass through the lens and come together at a single point in the image plane (or, more generally, the image surface). Real lenses do not focus light exactly to a single point, however, even when they are perfectly made. These deviations from the idealized lens performance are called aberrations of the lens.
Aberrations fall into two classes: monochromatic and chromatic. Monochromatic aberrations are caused by the geometry of the lens or mirror and occur both when light is reflected and when it is refracted. They appear even when using monochromatic light, hence the name.
Chromatic aberrations are caused by dispersion, the variation of a lens's refractive index with wavelength. Because of dispersion, different wavelengths of light come to focus at differ |
https://en.wikipedia.org/wiki/Tuple%20space | A tuple space is an implementation of the associative memory paradigm for parallel/distributed computing. It provides a repository of tuples that can be accessed concurrently. As an illustrative example, consider that there are a group of processors that produce pieces of data and a group of processors that use the data. Producers post their data as tuples in the space, and the consumers then retrieve data from the space that match a certain pattern. This is also known as the blackboard metaphor. Tuple space may be thought as a form of distributed shared memory.
Tuple spaces were the theoretical underpinning of the Linda language developed by David Gelernter and Nicholas Carriero at Yale University in 1986.
Implementations of tuple spaces have also been developed for Java (JavaSpaces), Lisp, Lua, Prolog, Python, Ruby, Smalltalk, Tcl, and the .NET Framework.
Object Spaces
Object Spaces is a paradigm for development of distributed computing applications. It is characterized by the existence of logical entities, called Object Spaces. All the participants of the distributed application share an Object Space. A provider of a service encapsulates the service as an Object, and puts it in the Object Space. Clients of a service then access the Object Space, find out which object provides the needed service, and have the request serviced by the object.
Object Spaces, as a computing paradigm, was put forward in the 1980s by David Gelernter at Yale University. Gelernter developed a language called Linda to support the concept of global object coordination.
Object Space can be thought of as a virtual repository, shared amongst providers and accessors of network services, which are themselves abstracted as objects. Processes communicate among each other using these shared objects — by updating the state of the objects as and when needed.
An object, when deposited into a space, needs to be registered with an Object Directory in the Object Space. Any processes can then i |
https://en.wikipedia.org/wiki/Distributed%20shared%20memory | In computer science, distributed shared memory (DSM) is a form of memory architecture where physically separated memories can be addressed as a single shared address space. The term "shared" does not mean that there is a single centralized memory, but that the address space is shared—i.e., the same physical address on two processors refers to the same location in memory. Distributed global address space (DGAS), is a similar term for a wide class of software and hardware implementations, in which each node of a cluster has access to shared memory in addition to each node's private (i.e., not shared) memory.
Overview
A distributed-memory system, often called a multicomputer, consists of multiple independent processing nodes with local memory modules which is connected by a general interconnection network. Software DSM systems can be implemented in an operating system, or as a programming library and can be thought of as extensions of the underlying virtual memory architecture. When implemented in the operating system, such systems are transparent to the developer; which means that the underlying distributed memory is completely hidden from the users. In contrast, software DSM systems implemented at the library or language level are not transparent and developers usually have to program them differently. However, these systems offer a more portable approach to DSM system implementations. A DSM system implements the shared-memory model on a physically distributed memory system.
DSM can be achieved via software as well as hardware. Hardware examples include cache coherence circuits and network interface controllers. There are three ways of implementing DSM:
Page-based approach using virtual memory
Shared-variable approach using routines to access shared variables
Object-based approach, ideally accessing shared data through object-oriented discipline
Advantages
Scales well with a large number of nodes
Message passing is hidden
Can handle complex and large data |
https://en.wikipedia.org/wiki/Doublet%20state | In quantum mechanics, a doublet is a composite quantum state of a system with an effective spin of 1/2, such that there are two allowed values of the spin component, −1/2 and +1/2. Quantum systems with two possible states are sometimes called two-level systems. Essentially all occurrences of doublets in nature arise from rotational symmetry; spin 1/2 is associated with the fundamental representation of the Lie group SU(2).
History and applications
The term "doublet" dates back to the 19th century, when it was observed that certain spectral lines of an ionized, excited gas would split into two under the influence of a strong magnetic field, in an effect known as the anomalous Zeeman effect. Such spectral lines were observed not only in the laboratory, but also in astronomical spectroscopy observations, allowing astronomers to deduce the existence of, and measure the strength of magnetic fields around the Sun, stars and galaxies. Conversely, it was the observation of doublets in spectroscopy that allowed physicists to deduce that the electron had a spin, and that furthermore, the magnitude of the spin had to be 1/2. See the history section of the article on Spin (physics) for greater detail.
Doublets continue to play an important role in physics. For example, the healthcare technology of magnetic resonance imaging is based on nuclear magnetic resonance. In this technology, a spectroscopic doublet occurs in a spin-1/2 atomic nucleus, whose doublet splitting is in the radio-frequency range. By applying both a magnetic field and carefully tuning a radio-frequency transmitter, the nuclear spins will flip and re-emit radiation, in an effect known as the Rabi cycle. The strength and frequency of the emitted radio waves allows the concentration of such nuclei to be measured. Another potential application is the use of doublets as the emitting layer in light emitting diodes (LEDs). These materials have the advantage of having 100% theoretical quantum efficiency based on spi |
https://en.wikipedia.org/wiki/Windows%20Neptune |
Neptune was the codename for a version of Microsoft Windows under development in 1999. Based on Windows 2000, it was originally to replace the Windows 9x series and was scheduled to be the first home consumer-oriented version of Windows built on Windows NT code. Internally, the project's name was capitalized as NepTune.
History
Neptune largely resembled Windows 2000, but some new features were introduced. Neptune included a logon screen similar to that later used in Windows XP. A firewall new to Neptune was later integrated into Windows XP as the Windows Firewall. Neptune also experimented with a new HTML and Win32-based user interface originally intended for Windows Me, called Activity Centers, for task-centered operations.
Only one alpha build of Neptune, 5111, was released to testers under a non-disclosure agreement, and later made its way to various beta collectors' sites and virtual museums in 2000. Other builds of Neptune are known to exist due to information in beta builds of Windows Me and Windows XP. In November 2015, a build 5111.6 disk was shown in a Microsoft Channel 9 video; version 5111 was the last build of Neptune that was sent to external testers, with the .1 or .6 after the build number stands for variant, not for compile. It is the only build of Neptune that made its way to the public. Build 5111 included Activity Centers, which could be installed by copying ACCORE.DLL from the installation disk to the hard drive and then running regsvr32 on ACCORE.DLL. The centers contained traces of Windows Me, then code-named Millennium, but were broken due to JavaScript errors, missing links and executables to the Game, Photo, and Music Centers. In response, some Windows enthusiasts have spent years fixing Activity Centers in build 5111 close to what Microsoft intended.
In early 2000, Microsoft merged the team working on Neptune with that developing Odyssey, the successor to Windows 2000 for business customers. The combined team worked on a new project c |
https://en.wikipedia.org/wiki/Polar%20decomposition | In mathematics, the polar decomposition of a square real or complex matrix is a factorization of the form , where is a unitary matrix and is a positive semi-definite Hermitian matrix ( is an orthogonal matrix and is a positive semi-definite symmetric matrix in the real case), both square and of the same size.
Intuitively, if a real matrix is interpreted as a linear transformation of -dimensional space , the polar decomposition separates it into a rotation or reflection of , and a scaling of the space along a set of orthogonal axes.
The polar decomposition of a square matrix always exists. If is invertible, the decomposition is unique, and the factor will be positive-definite. In that case, can be written uniquely in the form , where is unitary and is the unique self-adjoint logarithm of the matrix . This decomposition is useful in computing the fundamental group of (matrix) Lie groups.
The polar decomposition can also be defined as where is a symmetric positive-definite matrix with the same eigenvalues as but different eigenvectors.
The polar decomposition of a matrix can be seen as the matrix analog of the polar form of a complex number as , where is its absolute value (a non-negative real number), and is a complex number with unit norm (an element of the circle group).
The definition may be extended to rectangular matrices by requiring to be a semi-unitary matrix and to be a positive-semidefinite Hermitian matrix. The decomposition always exists and is always unique. The matrix is unique if and only if has full rank.
Intuitive interpretation
A real square matrix can be interpreted as the linear transformation of that takes a column vector to . Then, in the polar decomposition , the factor is an real orthonormal matrix. The polar decomposition then can be seen as expressing the linear transformation defined by into a scaling of the space along each eigenvector of by a scale factor (the action of ), followed by a r |
https://en.wikipedia.org/wiki/Roy%20Kerr | Roy Patrick Kerr (; born 16 May 1934) is a New Zealand mathematician who discovered the Kerr geometry, an exact solution to the Einstein field equation of general relativity. His solution models the gravitational field outside an uncharged rotating massive object, including a rotating black hole. His solution to Einstein's equations predicted spinning black holes before they were discovered.
Early life and education
Kerr was born in 1934 in Kurow, New Zealand. He was born into a dysfunctional family, and his mother was forced to leave when he was three. When his father went to war, he was sent to a farm. After his father's return from war, they moved to Christchurch. He was accepted to St Andrew's College, a private school, as his father had served under a former headmaster. Kerr's mathematical talent was first recognised while he was still a student at St Andrew's College. Although there was no mathematics teacher there at the time, he was able in 1951 to go straight into the third year of mathematics at Canterbury University College, a constituent of the University of New Zealand and the precursor to the University of Canterbury. Their regulations did not permit him to graduate until 1954 and so it was not until September 1955 that he moved to the University of Cambridge, where he earned his PhD in 1959. His dissertation concerned the equations of motion in general relativity.
Career and research
After a postdoctoral fellowship at Syracuse University, where Einstein's collaborator Peter Bergmann was a professor, he spent some time working for the United States Air Force at Wright-Patterson Air Force Base. Kerr speculated that the "main reason why the US Air Force had created a General Relativity section was probably to show the U.S. Navy that they could also do pure research."
Work at Texas and Canterbury
In 1962, Kerr joined Alfred Schild and his Relativity Group at the University of Texas at Austin. As Kerr wrote in 2009:
By the summer of 1963, Maarten Schmi |
https://en.wikipedia.org/wiki/Wet%20meadow | A wet meadow is a type of wetland with soils that are saturated for part or all of the growing season which prevents the growth of trees and brush. Debate exists whether a wet meadow is a type of marsh or a completely separate type of wetland. Wet prairies and wet savannas are hydrologically similar.
Hydrology and ecology
Wet meadows may occur because of restricted drainage or the receipt of large amounts of water from rain or melted snow. They may also occur in riparian zones and around the shores of large lakes.
Unlike a marsh or swamp, a wet meadow does not have standing water present except for brief to moderate periods during the growing season. Instead, the ground in a wet meadow fluctuates between brief periods of inundation and longer periods of saturation. Wet meadows often have large numbers of wetland plant species, which frequently survive as buried seeds during dry periods, and then regenerate after flooding. Wet meadows therefore do not usually support aquatic life such as fish. They typically have a high diversity of plant species, and may attract large numbers of birds, small mammals and insects including butterflies.
Vegetation in a wet meadow usually includes a wide variety of herbaceous species including sedges, rushes, grasses and a wide diversity of other plant species. A few of many possible examples include species of Rhexia, Parnassia, Lobelia, many species of wild orchids (e.g. Calopogon and Spiranthes), and carnivorous plants such as Sarracenia and Drosera. Woody plants, if present, account for a minority of the total area cover. High water levels are one of the important factors that prevent invasion by woody plants; in other cases, fire is important. In areas with low frequencies of fire, or reduced water level fluctuations, or higher fertility, plant diversity will decline.
Conservation
Wet meadows were once common in wetland types around the world. They remain an important community type in wet savannas and flatwoods. The also su |
https://en.wikipedia.org/wiki/Primitive%20ring | In the branch of abstract algebra known as ring theory, a left primitive ring is a ring which has a faithful simple left module. Well known examples include endomorphism rings of vector spaces and Weyl algebras over fields of characteristic zero.
Definition
A ring R is said to be a left primitive ring if it has a faithful simple left R-module. A right primitive ring is defined similarly with right R-modules. There are rings which are primitive on one side but not on the other. The first example was constructed by George M. Bergman in . Another example found by Jategaonkar showing the distinction can be found in .
An internal characterization of left primitive rings is as follows: a ring is left primitive if and only if there is a maximal left ideal containing no nonzero two-sided ideals. The analogous definition for right primitive rings is also valid.
The structure of left primitive rings is completely determined by the Jacobson density theorem: A ring is left primitive if and only if it is isomorphic to a dense subring of the ring of endomorphisms of a left vector space over a division ring.
Another equivalent definition states that a ring is left primitive if and only if it is a prime ring with a faithful left module of finite length (, Ex. 11.19, p. 191).
Properties
One-sided primitive rings are both semiprimitive rings and prime rings. Since the product ring of two or more nonzero rings is not prime, it is clear that the product of primitive rings is never primitive.
For a left Artinian ring, it is known that the conditions "left primitive", "right primitive", "prime", and "simple" are all equivalent, and in this case it is a semisimple ring isomorphic to a square matrix ring over a division ring. More generally, in any ring with a minimal one sided ideal, "left primitive" = "right primitive" = "prime".
A commutative ring is left primitive if and only if it is a field.
Being left primitive is a Morita invariant property.
Examples
Every simple r |
https://en.wikipedia.org/wiki/Linksys | Linksys Holdings, Inc., is an American brand of data networking hardware products mainly sold to home users and small businesses. It was founded in 1988 by the couple Victor and Janie Tsao, both Taiwanese immigrants to the United States. Linksys products include Wi-Fi routers, mesh Wi-Fi systems, Wifi extenders, access points, network switches, and Wi-Fi networking. It is headquartered in Irvine, California.
Linksys products are sold direct-to-consumer from its website, through online retailers and marketplaces, as well as off-the-shelf in consumer electronics and big-box retail stores. As of 2020, Linksys products are sold in retail locations and value-added resellers in 64 countries and was the first router company to ship 100 million products.
History
In 1988, spouses Janie and Victor Tsao founded DEW International, later renamed Linksys, in the garage of their Irvine, California home. The Tsaos were immigrants from Taiwan who held second jobs as consultants specializing in pairing American technology vendors with manufacturers in Taiwan. The founders used Taiwanese manufacturing to achieve its early success. The company's first products were printer sharers that connected multiple PCs to printers. The company expanded into Ethernet hubs, network cards, and cords. In 1992, the Tsaos began running Linksys full time and moved the company and its growing staff to a formal office. By 1994, it had grown to 55 employees with annual revenues of $6.5 million.
Linksys received a major boost in 1995, when Microsoft released Windows 95 with built-in networking functions that expanded the market for its products. Linksys established its first U.S. retail channels with Fry's Electronics (1995) and Best Buy (1996). In the late 1990s, Linksys released the first affordable multiport router, popularizing Linksys as a home networking brand. By 2003, when the company was acquired by Cisco, it had 305 employees and revenues of more than $500 million.
Cisco expanded the company |
https://en.wikipedia.org/wiki/Jan%20Dietz | Jean Leonardus Gerardus (Jan) Dietz (born 20 June 1945) is a Dutch Information Systems researcher, Professor Emeritus of Information Systems Design at the Delft University of Technology, known for the development of the Design & Engineering Methodology for Organisations. and his work on Enterprise Engineering.
Biography
Born in Brunssum, Dietz studied at the Eindhoven University of Technology, where in 1970 he obtained his MSc in Electrical Engineering (control systems) and in 1987 his Doctoral Degree on the thesis titled "Modelleren en specificeren van informatiesystemen" (Modelling and Specifying Information Systems) under supervision of Theo Bemelmans and Kees van Hee.
Dietz has been practitioner in the field of automation and information systems from 1970 to 1980. He (co-)developed one of the first relational model based production control systems at Philips Factories, a state of the art computer accounting system at Eindhoven University of Technology, and a terminal-based, interactive theatre reservation system. In 1980 he returned to academia. In 1988 he was appointed Professor of Management Information Systems at the University of Maastricht (Faculty of Economics and Business Administration) where he started the development of the DEMO methodology. From September 1994 to Oct 2009 he was Professor of Information Systems Design at Delft University of Technology. Since 2010 he is visiting professor of Enterprise Engineering at the Instituto Superior Técnico (IST - Technical University of Lisbon), as well as visiting professor of Enterprise Engineering at Czech Technical University in Prague (Faculty of Informatics, Center for Conceptual Modelling and Implementations).
Dietz has been chairman of the Dutch professional association of informaticians (VRI) and board member of the Dutch association for IT architecture (NAF). He has also been editorial board member of several journals, and has been in the program committee of - and has chaired - numerous conf |
https://en.wikipedia.org/wiki/Natural%20prolongation%20principle | The natural prolongation principle or principle of natural prolongation is a legal concept introduced in maritime claims submitted to the United Nations.
The phrase denotes a concept of political geography and international law that a nation's maritime boundary should reflect the 'natural prolongation' of where its land territory reaches the coast.
Oceanographic descriptions of the land mass under coastal waters became conflated and confused with criteria that are deemed relevant in border delimitation. The concept was developed in the process of settling disputes if the borders of adjacent nations were located on a contiguous continental shelf.
An unresolved issue is whether a natural prolongation defined scientifically, without reference to equitable principles, is to be construed as a "natural prolongation" for the purpose of maritime border delimitation or maritime boundary disputes.
History
The phrase natural prolongation was established as a concept in the North Sea Continental Cases in 1969.
The relevance and importance of natural prolongation as a factor in delimitation disputes and agreements has declined during the period in which international acceptance of UNCLOS III has expanded.
The Malta/Libya Case in 1985 is marked as the eventual demise of the natural prolongation principle being used in delimiting between adjoining national maritime boundaries.
The Bay of Bengal cases in the early 2010s (Bangladesh v Myanmar) and (Bangladesh v India) likewise dealt a blow to natural prolongation as the guiding principle for delimitation of the continental shelf more than 200 nautical miles beyond baselines.
See also
Equidistance principle |
https://en.wikipedia.org/wiki/Spinous%20cell | Spinous cells, or prickle cells, are keratin producing epidermal cells owing their prickly appearance to their numerous intracellular connections. They make up the stratum spinosum (prickly layer) of the epidermis and provide a continuous net-like layer of protection for underlying tissue. They are susceptible to mutations caused by sunlight and can become malignant.
Location
Spinous cells are found in the superficial layers of the skin. They are found in the stratum spinosum (prickly layer, spinosum layer), which lies above the stratum basale (basal layer) and below the stratum granulosum (granular layer) of the epidermis. The spinous cells are arranged several layers thick to form a net-like covering.
Origin
Spinous cells originate through mitosis in the basal layer (also known as the germinative layer). They are pushed upward into the stratum spinosum by the continuous formation of new cells in the basal layer. They reach the outmost layer of the skin as flattened dead flaking skin cells we shed daily. The journey from origin to shed takes 25 to 45 days.
Function
Spinous cells serve “as a physical and biological barrier to the environment, preventing penetration by irritants and allergens and loss of water while maintaining internal homeostasis. They accomplish this in two ways. First, they are keratinocytes (keratin cells) whose primary function is to produce keratin, a strong structural protein. The keratin accumulates within each spinous cell as it moves upward through the epidermis layers, until the cell is almost completely filled with hardening keratin (keratinisation). Second, the cells are bound together across their cytoplasm by keratin filaments that form cell-to-cell connections (desmosomes).
Cancer
Squamous cell carcinoma (prickle cell carcinoma) is relatively common in people over age 60, with fair skin, and a history of longer term sun exposure. It is not as commonly known as other skin cancers because it is less likely to metastasize, |
https://en.wikipedia.org/wiki/Nominal%20level | Nominal level is the operating level at which an electronic signal processing device is designed to operate. The electronic circuits that make up such equipment are limited in the maximum signal they can handle and the low-level internally generated electronic noise they add to the signal. The difference between the internal noise and the maximum level is the device's dynamic range. The nominal level is the level that these devices were designed to operate at, for best dynamic range and adequate headroom. When a signal is chained with improper gain staging through many devices, clipping may occur or the system may operate with reduced dynamic range.
In audio, a related measurement, signal-to-noise ratio, is usually defined as the difference between the nominal level and the noise floor, leaving the headroom as the difference between nominal and maximum output. It is important to realize that the measured level is a time average, meaning that the peaks of audio signals regularly exceed the measured average level. The headroom measurement defines how far the peak levels can stray from the nominal measured level before clipping. The difference between the peaks and the average for a given signal is the crest factor.
Standards
VU meters are designed to represent the perceived loudness of a passage of music, or other audio content, measuring in volume units. Devices are designed so that the best signal quality is obtained when the meter rarely goes above nominal. The markings are often in dB instead of "VU", and the reference level should be defined in the device's manual. In most professional recording and sound reinforcement equipment, the nominal level is . In semi-professional and domestic equipment, the nominal level is usually −10 dBV. This difference is due to the cost required to create larger power supplies and output higher levels.
In broadcasting equipment, this is termed the Maximum Permitted Level, which is defined by European Broadcasting Union stand |
https://en.wikipedia.org/wiki/Yarkovsky%20effect | The Yarkovsky effect is a force acting on a rotating body in space caused by the anisotropic emission of thermal photons, which carry momentum. It is usually considered in relation to meteoroids or small asteroids (about 10 cm to 10 km in diameter), as its influence is most significant for these bodies.
History of discovery
The effect was discovered by the Polish-Russian civil engineer Ivan Osipovich Yarkovsky (1844–1902), who worked in Russia on scientific problems in his spare time. Writing in a pamphlet around the year 1900, Yarkovsky noted that the daily heating of a rotating object in space would cause it to experience a force that, while tiny, could lead to large long-term effects in the orbits of small bodies, especially meteoroids and small asteroids. Yarkovsky's insight would have been forgotten had it not been for the Estonian astronomer Ernst J. Öpik (1893–1985), who read Yarkovsky's pamphlet sometime around 1909. Decades later, Öpik, recalling the pamphlet from memory, discussed the possible importance of the Yarkovsky effect on movement of meteoroids about the Solar System.
Mechanism
The Yarkovsky effect is a consequence of the fact that change in the temperature of an object warmed by radiation (and therefore the intensity of thermal radiation from the object) lags behind changes in the incoming radiation. That is, the surface of the object takes time to become warm when first illuminated, and takes time to cool down when illumination stops. In general there are two components to the effect:
Diurnal effect: On a rotating body illuminated by the Sun (e.g. an asteroid or the Earth), the surface is warmed by solar radiation during the day, and cools at night. Due to the thermal properties of the surface, there is a lag between the absorption of radiation from the Sun, and the emission of radiation as heat, so the warmest point on a rotating body occurs around the "2 PM" site on the surface, or slightly after noon. This results in a difference between |
https://en.wikipedia.org/wiki/Michael%20McQuillan%20%28mathematician%29 | Michael Liam McQuillan is a Scottish mathematician studying algebraic geometry. As of 2019 he is Professor at the University of Rome Tor Vergata.
Career
Michael McQuillan received the doctorate in 1992 at Harvard University under Barry Mazur ("Division points on semi-Abelian varieties").
In 1995, McQuillan proved the Mordell–Lang conjecture. In 1996, MacQuillan gave a new proof of a conjecture of André Bloch (1926) about holomorphic curves in closed subvarieties of Abelian varieties, proved a conjecture of Shoshichi Kobayashi (about the Kobayashi-hyperbolicity of generic hypersurfaces of high degree in projective n-dimensional space) in the three-dimensional case and achieved partial results on a conjecture of Mark Green and Phillip Griffiths (which states that a holomorphic curve on an algebraic surface of general type with cannot be Zariski-dense).
From 1996 to 2001 he was a post-doctoral Research Fellow at All Souls College of the University of Oxford and in 2009 was Professor at the University of Glasgow as well as Advanced Research Fellow of the British Engineering and Physical Sciences Research Council. As of 2019 he is Professor at the University of Rome Tor Vergata and an editor of the European Journal of Mathematics.
Awards
In 2000 McQuillan received the EMS Prize, which was announced from the European Congress of Mathematics in July 2000, for his work:
In 2001 he was awarded the Whitehead Prize of the London Mathematical Society. In 2002 he was invited speaker at the International Congress of Mathematicians in Beijing (Integrating ). In 2001 he received the Whittaker Prize. |
https://en.wikipedia.org/wiki/Trace%20fossil%20classification | Trace fossils are classified in various ways for different purposes. Traces can be classified taxonomically (by morphology), ethologically (by behavior), and toponomically, that is, according to their relationship to the surrounding sedimentary layers. Except in the rare cases where the original maker of a trace fossil can be identified with confidence, phylogenetic classification of trace fossils is an unreasonable proposition.
Taxonomic classification
The taxonomic classification of trace fossils parallels the taxonomic classification of organisms under the International Code of Zoological Nomenclature. In trace fossil nomenclature a Latin binomial name is used, just as in animal and plant taxonomy, with a genus and specific epithet. However, the binomial names are not linked to an organism, but rather just a trace fossil. This is due to the rarity of association between a trace fossil and a specific organism or group of organisms. Trace fossils are therefore included in an ichnotaxon separate from Linnaean taxonomy. When referring to trace fossils, the terms ichnogenus and ichnospecies parallel genus and species respectively.
The most promising cases of phylogenetic classification are those in which similar trace fossils show details complex enough to deduce the makers, such as bryozoan borings, large trilobite trace fossils such as Cruziana, and vertebrate footprints. However, most trace fossils lack sufficiently complex details to allow such classification.
Ethologic classification
The Seilacherian System
Adolf Seilacher was the first to propose a broadly accepted ethological basis for trace fossil classification. He recognized that most trace fossils are created by animals in one of five main behavioural activities, and named them accordingly:
Cubichnia are the traces of organisms left on the surface of a soft sediment. This behaviour may simply be resting as in the case of a starfish, but might also evidence the hiding place of prey, or even the ambus |
https://en.wikipedia.org/wiki/Photobioreactor | A photobioreactor (PBR) refers to any cultivation system designed for growing photoautotrophic organisms using artificial light sources or solar light to facilitate photosynthesis. PBRs are typically used to cultivate microalgae, cyanobacteria, and some mosses. PBRs can be open systems, such as raceway ponds, which rely upon natural sources of light and carbon dioxide. Closed PBRs are flexible systems that can be controlled to the physiological requirements of the cultured organism, resulting in optimal growth rates and purity levels. PBRs are typically used for the cultivation of bioactive compounds for biofuels, pharmaceuticals, and other industrial uses.
Open systems
The first approach for the controlled production of phototrophic organisms was a natural open pond or artificial raceway pond. Therein, the culture suspension, which contains all necessary nutrients and carbon dioxide, is pumped around in a cycle, being directly illuminated from sunlight via the liquid's surface. Raceway ponds are still commonly used in industry due to their low operational cost in comparison to closed PBRs. However, they offer an insufficient control of reaction conditions due to their reliance on environmental light supply and carbon dioxide, as well as possible contamination from other microorganisms. Using open technologies also result in losses of water due to evaporation into the atmosphere.
Closed systems
The construction of closed PBRs avoids system-related water losses and minimises contamination. Though closed systems have better productivity compared to open systems due to the advantages mentioned, they still need to be improved to make them suitable for production of low price commodities as cell density remains low due to several limiting factors. All modern photobioreactors have tried to balance between a thin layer of culture suspension, optimized light application, low pumping energy consumption, capital expenditure and microbial purity. However, light attenuation a |
https://en.wikipedia.org/wiki/Posterior%20spinal%20veins | Posterior spinal veins are small veins which receive blood from the dorsal spinal cord. |
https://en.wikipedia.org/wiki/Whiteshift | Whiteshift: Populism, Immigration and the Future of White Majorities is a 2018 non-fiction book written by Eric Kaufmann, a professor at Birkbeck College, University of London. Described by The Economist as a "monumental study of ethno-demographic change", Whiteshift covers politics in both Europe and North America and looks into the political views of the populist right. Kaufmann argues that the rise of Donald Trump in America and the populist right in Europe is a reaction to sweeping demographic change rather than to "economic anxiety".
Reviews
On release, The Times made Whiteshift the 'Book of the Week' but with a sceptical review by David Aaronovitch, who called it "a big controversial book about a big controversial subject". Publishers Weekly said it was "likely to make a big splash", and The Financial Times listed it as one of the 'Best books of 2018' in the politics genre. The New Yorker wrote that Kaufmann and Whiteshift were defending white identity politics. Daniel Trilling, in the London Review of Books, was critical of the book, describing Kaufmann's frame of reference as "both too broad and too narrow".
Kenan Malik wrote that "Whiteshift is a hefty work crammed with data and graphs. The trouble with viewing the world primarily in demographic terms, though, is that, for all the facts and figures, it is easy to be blind to the social context."
In a review symposium about Whiteshift published in the journal Ethnicities, political scientist Rob Ford wrote that "There is much to admire here. Kaufmann is methodologically catholic and draws on a rich range of different resources to examine and interrogate evolving white identity politics." However, he also noted "Kaufmann’s rather Manichean account of white ethnic politics involves some curious omissions and misunderstandings" and that "lack of balance is a recurring feature of Kaufmann’s discussions about the competing claims of ethno-cultural whites, cosmopolitan whites and ethnic minorities." Sociologist |
https://en.wikipedia.org/wiki/Bette%20Korber | Bette Korber is an American computational biologist focusing on the molecular biology and population genetics of the HIV virus that causes infection and eventually AIDS. She has contributed heavily to efforts to obtain an effective HIV vaccine. She created a database at Los Alamos National Laboratory that has enabled her to design novel mosaic HIV vaccines, one of which is currently in human testing in Africa. The database contains thousands of HIV genome sequences and related data.
Korber is a scientist in theoretical biology and biophysics at Los Alamos National Laboratory. She has received the Ernest Orlando Lawrence Award, the Department of Energy's highest award for scientific achievement. She has also received several other awards including the Elizabeth Glaser Award for pediatric AIDS research and the Richard Feynman Award for Innovation.
Early life and education
Bette Korber grew up in Southern California. She earned her B.S. in chemistry in 1981 from California State University, Long Beach, where her father was a sociology professor, her mother graduated in nursing, and her sister graduated in journalism. From 1981 to 1988, she was in the graduate program at the California Institute of Technology (Caltech), where she worked with Iwona Stroynowski in Leroy Hood's laboratory, receiving her PhD in chemistry in 1988. Her work focused on regulation of the expression of major histocompatibility complex type 1 genes, producing cell surface proteins that participate in the rejection of tissue transplants, by interferon induced by viral infections.
She then became a postdoctoral fellow with Myron Essex, working on the molecular epidemiology of the AIDS/HIV virus and HTLV-1, the human leukemia virus, at the Harvard School of Public Health until 1990. There, Korber used polymerase chain reaction (PCR) to show both complete and deleted versions of viral genomes in leukemic cells. Her work on these viral partial and complete genomes was influential and widely cit |
https://en.wikipedia.org/wiki/Vertiginous%20question | Benj Hellie's vertiginous question asks why, of all the subjects of experience out there, this one—the one corresponding to the human being referred to as Benj Hellie—is the one whose experiences are live? (The reader is supposed to substitute their own case for Hellie's.)
A simple response is that this question reduces to "Why are Hellie's experiences live from Hellie's perspective," which is trivial to answer. However Hellie argues, through a parable, that this response leaves something out. His parable describes two situations, one reflecting a broad global constellation view of the world and everyone's phenomenal features, and one describing an embedded view from the perspective of a single subject. The former seems to align better with the simple response above, but the latter seems a better description of consciousness.
Hellie's argument is closely related to Caspar Hare's theories of egocentric presentism and perspectival realism, of which several other philosophers have written reviews.
Similar questions are also asked repeatedly by J.J. Valberg in justifying his horizonal view of the self.
David Chalmers wrote a response to Hellie, but it did not address the question itself.
See also
Binding problem
Centered world
Further facts
Indexicality
Metaphysical subjectivism
Personal horizon
Personal identity
Open individualism
Subjective idealism
Subjectivity and objectivity (philosophy)
Solipsism |
https://en.wikipedia.org/wiki/CASP | Critical Assessment of Structure Prediction (CASP), sometimes called Critical Assessment of Protein Structure Prediction, is a community-wide, worldwide experiment for protein structure prediction taking place every two years since 1994. CASP provides research groups with an opportunity to objectively test their structure prediction methods and delivers an independent assessment of the state of the art in protein structure modeling to the research community and software users. Even though the primary goal of CASP is to help advance the methods of identifying protein three-dimensional structure from its amino acid sequence many view the experiment more as a “world championship” in this field of science. More than 100 research groups from all over the world participate in CASP on a regular basis and it is not uncommon for entire groups to suspend their other research for months while they focus on getting their servers ready for the experiment and on performing the detailed predictions.
Selection of target proteins
In order to ensure that no predictor can have prior information about a protein's structure that would put them at an advantage, it is important that the experiment be conducted in a double-blind fashion: Neither predictors nor the organizers and assessors know the structures of the target proteins at the time when predictions are made. Targets for structure prediction are either structures soon-to-be solved by X-ray crystallography or NMR spectroscopy, or structures that have just been solved (mainly by one of the structural genomics centers) and are kept on hold by the Protein Data Bank. If the given sequence is found to be related by common descent to a protein sequence of known structure (called a template), comparative protein modeling may be used to predict the tertiary structure. Templates can be found using sequence alignment methods (e.g. BLAST or HHsearch) or protein threading methods, which are better in finding distantly related templates. Othe |
https://en.wikipedia.org/wiki/Marc%20Van%20Montagu | Marc, Baron Van Montagu (born 10 November 1933 in Ghent) is a Belgian molecular biologist. He was full professor and director of the Laboratory of Genetics at the faculty of Sciences at Ghent University (Belgium) and scientific director of the genetics department of the Flanders Interuniversity Institute for Biotechnology (VIB). Together with Jozef Schell he founded the biotech company Plant Genetic Systems Inc. (Belgium) in 1982, of which he was scientific director and member of the board of directors. Van Montagu was also involved in founding the biotech company CropDesign, of which he was a board member from 1998 to 2004. He is president of the Public Research and Regulation Initiative (PRRI).
Van Montagu and his colleagues were credited with the discovery of the Ti plasmid. They described the gene transfer mechanism between Agrobacterium and plants, which resulted in the development of methods to alter Agrobacterium into an efficient delivery system for gene engineering and to create transgenic plants. They developed plant molecular genetics, in particular molecular mechanisms for cell proliferation and differentiation and response to abiotic stresses (high light, ozone, cold, salt and drought) and constructed transgenic crops (tobacco, rape seed, corn) resistant to insect pest and tolerant to novel herbicides. His work with poplar trees resulted in engineering of trees with improved pulping qualities.
After his retirement as director of the Laboratory of Genetics at Ghent University, Marc Van Montagu created IPBO - International Plant Biotechnology Outreach, VIB-Ghent University, with the mission to foster biotechnological solutions to global agriculture. In 2015 IPBO launched the “Marc and Nora Van Montagu (MNVM) Fund” with focus on sustainable agriculture and agro-industry to the African continent.
Honors
Van Montagu has been a foreign associate of the United States National Academy of Sciences since 1986, the agricultural Academy of Russia and France, the |
https://en.wikipedia.org/wiki/JPEG%202000 | JPEG 2000 (JP2) is an image compression standard and coding system. It was developed from 1997 to 2000 by a Joint Photographic Experts Group committee chaired by Touradj Ebrahimi (later the JPEG president), with the intention of superseding their original JPEG standard (created in 1992), which is based on a discrete cosine transform (DCT), with a newly designed, wavelet-based method. The standardized filename extension is .jp2 for ISO/IEC 15444-1 conforming files and .jpx for the extended part-2 specifications, published as ISO/IEC 15444-2. The registered MIME types are defined in RFC 3745. For ISO/IEC 15444-1 it is image/jp2.
The JPEG 2000 project was motivated by Ricoh's submission in 1995 of the CREW (Compression with Reversible Embedded Wavelets) algorithm to the standardization effort of JPEG-LS. Ultimately the LOCO-I algorithm was selected as the basis for JPEG-LS, but many of the features of CREW ended up in the JPEG 2000 standard.
JPEG 2000 codestreams are regions of interest that offer several mechanisms to support spatial random access or region of interest access at varying degrees of granularity. It is possible to store different parts of the same picture using different quality.
JPEG 2000 is a compression standard based on a discrete wavelet transform (DWT). The standard could be adapted for motion imaging video compression with the Motion JPEG 2000 extension. JPEG 2000 technology was selected as the video coding standard for digital cinema in 2004. However, JPEG 2000 is still not widely supported in web browsers (other than Safari) and hence is not generally used on the World Wide Web.
Design goals
While there is a modest increase in compression performance of JPEG 2000 compared to JPEG, the main advantage offered by JPEG 2000 is the significant flexibility of the codestream. The codestream obtained after compression of an image with JPEG 2000 is scalable in nature, meaning that it can be decoded in a number of ways; for instance, by truncating th |
https://en.wikipedia.org/wiki/MOSH%20%28Nokia%29 | MOSH was a user defined distribution channel for mobile content initiated by Nokia. The name "MOSH" comes from "Mobilize and Share". The channel could have been used to both download and upload various content for mobile phones or other platforms. File types that were handled were: audio, images, applications, games, videos, documents.
A person could create an account on MOSH that comes with the following stats: creator (for content created by that person), collector (for the number of objects added to the collection in the account) and sharer (for uploading files that are not necessarily the work of that person). MOSH was known for having been filled with explicit and copyright-infringing content as it had no screening process.
MOSH was launched by Nokia in August 2007 and remained in Beta until its closure in 2009, by which time it had 137 million downloads. It was replaced by Ovi. |
https://en.wikipedia.org/wiki/Visual%20cryptography | Visual cryptography is a cryptographic technique which allows visual information (pictures, text, etc.) to be encrypted in such a way that the decrypted information appears as a visual image.
One of the best-known techniques has been credited to Moni Naor and Adi Shamir, who developed it in 1994. They demonstrated a visual secret sharing scheme, where an image was broken up into n shares so that only someone with all n shares could decrypt the image, while any shares revealed no information about the original image. Each share was printed on a separate transparency, and decryption was performed by overlaying the shares. When all n shares were overlaid, the original image would appear. There are several generalizations of the basic scheme including k-out-of-n visual cryptography, and using opaque sheets but illuminating them by multiple sets of identical illumination patterns under the recording of only one single-pixel detector.
Using a similar idea, transparencies can be used to implement a one-time pad encryption, where one transparency is a shared random pad, and another transparency acts as the ciphertext. Normally, there is an expansion of space requirement in visual cryptography. But if one of the two shares is structured recursively, the efficiency of visual cryptography can be increased to 100%.
Some antecedents of visual cryptography are in patents from the 1960s. Other antecedents are in the work on perception and secure communication.
Visual cryptography can be used to protect biometric templates in which decryption does not require any complex computations.
Example
In this example, the image has been split into two component images. Each component image has a pair of pixels for every pixel in the original image. These pixel pairs are shaded black or white according to the following rule: if the original image pixel was black, the pixel pairs in the component images must be complementary; randomly shade one ■□, and the other □■. When these comple |
https://en.wikipedia.org/wiki/Circumflex%20branch%20of%20left%20coronary%20artery | The circumflex branch of left coronary artery (also known as the left circumflex artery, or circumflex artery) is a branch of the left coronary artery. It winds around the left side of the heart along the atrioventricular groove (coronary sulcus). It supplies the posterolateral portion of the left ventricle.
In a minority of individuals, the left circumflex artery gives rise to the posterior interventricular artery, in which cases such a heart is deemed left dominant.
Anatomy
The left circumflex artery follows the left part of the coronary sulcus, running first to the left and then to the right, reaching nearly as far as the posterior longitudinal sulcus. There have been multiple anomalies described, for example the left circumflex having an aberrant course from the right coronary artery.
Branches
The circumflex artery curves to the left around the heart within the coronary sulcus, giving rise to one or more left marginal arteries (also called obtuse marginal branches) as it curves toward the posterior surface of the heart. It helps form the posterior left ventricular branch or posterolateral artery. The circumflex artery ends at the point where it joins to form to the posterior interventricular artery in 15% of all cases, which lies in the posterior interventricular sulcus. In the other 85% of all cases the posterior interventricular artery comes out of the right coronary artery. When the left circumflex supplies the posterior descending artery in those 15% of cases, it is known as a left dominant circulation.
Distribution
The circumflex artery supplies the posterolateral left ventricle and the anterolateral papillary muscle.
It also supplies the sinoatrial nodal artery in 38% of people.
It supplies 15–25% of the left ventricle in right-dominant systems. If the coronary anatomy is left-dominant, the circumflex artery supplies 40–50% of the left ventricle.
Additional images |
https://en.wikipedia.org/wiki/FAM122A%20%28gene%29 | Family with sequence similarity 122A is a protein that in humans is encoded by the FAM122A gene. |
https://en.wikipedia.org/wiki/Atomic%20and%20molecular%20astrophysics | Atomic astrophysics is concerned with performing atomic physics calculations that will be useful to astronomers and using atomic data to interpret astronomical observations. Atomic physics plays a key role in astrophysics as astronomers' only information about a particular object comes through the light that it emits, and this light arises through atomic transitions.
Molecular astrophysics, developed into a rigorous field of investigation by theoretical astrochemist Alexander Dalgarno beginning in 1967, concerns the study of emission from molecules in space. There are 110 currently known interstellar molecules. These molecules have large numbers of observable transitions. Lines may also be observed in absorption—for example the highly redshifted lines seen against the gravitationally lensed quasar PKS1830-211. High energy radiation, such as ultraviolet light, can break the molecular bonds which hold atoms in molecules. In general then, molecules are found in cool astrophysical environments. The most massive objects in our galaxy are giant clouds of molecules and dust known as giant molecular clouds. In these clouds, and smaller versions of them, stars and planets are formed. One of the primary fields of study of molecular astrophysics is star and planet formation. Molecules may be found in many environments, however, from stellar atmospheres to those of planetary satellites. Most of these locations are relatively cool, and molecular emission is most easily studied via photons emitted when the molecules make transitions between low rotational energy states. One molecule, composed of the abundant carbon and oxygen atoms, and very stable against dissociation into atoms, is carbon monoxide (CO). The wavelength of the photon emitted when the CO molecule falls from its lowest excited state to its zero energy, or ground, state is 2.6mm, or 115 gigahertz. This frequency is a thousand times higher than typical FM radio frequencies. At these high frequencies, molecules in th |
https://en.wikipedia.org/wiki/Electrohydrogenesis | Electrohydrogenesis or biocatalyzed electrolysis is the name given to a process for generating hydrogen gas from organic matter being decomposed by bacteria. This process uses a modified fuel cell to contain the organic matter and water. A small amount, 0.2–0.8 V of electricity is used, the original article reports an overall energy efficiency of 288% can be achieved (this is computed relative to the amount of electricity used, waste heat lowers the overall efficiency). This work was reported by Cheng and Logan.
See also
Biohydrogen
Electrochemical reduction of carbon dioxide
Electromethanogenesis
Fermentative hydrogen production
Microbial fuel cell |
https://en.wikipedia.org/wiki/Tapetum%20lucidum | The ; ; ) is a layer of tissue in the eye of many vertebrates and some other animals. Lying immediately behind the retina, it is a retroreflector. It reflects visible light back through the retina, increasing the light available to the photoreceptors (although slightly blurring the image). The tapetum lucidum contributes to the superior night vision of some animals. Many of these animals are nocturnal, especially carnivores, while others are deep sea animals.
Similar adaptations occur in some species of spiders. Haplorhine primates, including humans, are diurnal and lack a tapetum lucidum.
Function and mechanism
Presence of a tapetum lucidum enables animals to see in dimmer light than would otherwise be possible. The tapetum lucidum, which is iridescent, reflects light roughly on the interference principles of thin-film optics, as seen in other iridescent tissues. However, the tapetum lucidum cells are leucophores, not iridophores.
The tapetum functions as a retroreflector which reflects light directly back along the light path. This serves to match the original and reflected light, thus maintaining the sharpness and contrast of the image on the retina. The tapetum lucidum reflects with constructive interference, thus increasing the quantity of light passing through the retina. In the cat, the tapetum lucidum increases the sensitivity of vision by 44%, allowing the cat to see light that is imperceptible to human eyes.
It has been speculated that some flashlight fish may use eyeshine both to detect and to communicate with other flashlight fish. American scientist Nathan H. Lents has proposed that the tapetum lucidum evolved in vertebrates, but not in cephalopods, which have a very similar eye, because of the backwards-facing nature of vertebrate photoreceptors. The tapetum boosts photosensitivity under conditions of low illumination, thus compensating for the suboptimal design of the vertebrate retina.
Classification
A classification of anatomical variants of |
https://en.wikipedia.org/wiki/Neurohypophysial%20hormone | The neurohypophysial hormones form a family of structurally and functionally related peptide hormones. Their representatives in humans are oxytocin and vasopressin. They are named after the location of their release into the blood, the neurohypophysis (another name for the posterior pituitary).
Most of the circulating oxytocin and vasopressin hormones are synthesized in magnocellular neurosecretory cells of the supraoptic nucleus and paraventricular nucleus of the hypothalamus. They are then transported in neurosecretory granules along axons within the hypothalamo-neurohypophysial tract by axoplasmic flow to axon terminals forming the pars nervosa of the posterior pituitary. There, they are stored in Herring bodies and can be released into the circulation on the basis of hormonal and synaptic signals with assistance from pituicytes.
Oxytocin mediates contraction of the smooth muscle of the uterus and mammary gland, while vasopressin has antidiuretic action on the kidney, and mediates vasoconstriction of the peripheral vessels. Due to the similarity of the two hormones, there is cross-reaction: oxytocin has a slight antidiuretic function, and high levels of AVP can cause uterine contractions. In common with most active peptides, both hormones are synthesised as larger protein precursors that are enzymatically converted to their mature forms.
Members of this family are found in birds, fish, reptiles and amphibians (mesotocin, isotocin, valitocin, glumitocin, aspargtocin, vasotocin, seritocin, asvatocin, phasvatocin), in worms (annetocin, nematocin), octopuses (cephalotocin, octopressin), insects (locupressin, inotocin) and in molluscs (conopressins G and S). Animals that lack a hormone from this family include fruit flies, and at least some mosquitos, silkworms, and honeybees. |
https://en.wikipedia.org/wiki/Scandosorbus%20intermedia | Scandosorbus intermedia or, formerly, Sorbus intermedia, the Swedish whitebeam, is a species of whitebeam found in southern Sweden, with scattered occurrences in Estonia, Latvia, easternmost Denmark (Bornholm), the far southwest of Finland, and northern Poland.
Description
It is a medium-sized deciduous tree growing to tall with a stout trunk usually up to , but sometimes as much as in diameter, and grey bark; the crown is dome-shaped, with stout horizontal branches. The leaves are green above, and densely hairy with pale grey-white hairs beneath, long and broad, with four to seven oval lobes on each side of the leaf, broadest near the middle, rounded at the apex, and finely serrated margins. The autumn colour is dull yellowish to grey-brown. The flowers are diameter, with five white petals and 20 yellowish-white stamens; they are produced in corymbs diameter in late spring. The fruit is an oval pome long and in diameter, orange-red to red, maturing in mid autumn. The fruit is dryish, and eaten by thrushes and waxwings, which disperse the seeds.
Sorbus intermedia is a triple hybrid between S. aucuparia, S. torminalis, and either S. aria or one of its close relatives. It is closely related to Sorbus hybrida (Finnish whitebeam), another species of hybrid origin, which differs in having the leaves more deeply lobed, with the basal two pairs cut right to the midrib as separate leaflets. Both are tetraploid apomictic species which breed true without pollination.
Habitat, cultivation and uses
In the Nordic countries, the tree typically grows in forests, pastures or forest edges.
It is widely grown as an ornamental tree in northern Europe, valued for its tolerance of urban street conditions; it is very commonly used in avenues and urban parks. It is frequently naturalised in the British Isles. In recent years, much new planting of "Swedish whitebeam" has actually been of the related Sorbus mougeotii (Vosges whitebeam), another apomictic species from further |
https://en.wikipedia.org/wiki/Computer%20cartography | Computer cartography (also called digital cartography) is the art, science, and technology of making and using maps with a computer. This technology represents a paradigm shift in how maps are produced, but is still fundamentally a subset of traditional cartography. The primary function of this technology is to produce maps, including creation of accurate representations of a particular area such as, detailing major road arteries and other points of interest for navigation, and in the creation of thematic maps. Computer cartography is one of the main functions of geographic information systems (GIS), however, GIS is not necessary to facilitate computer cartography and has functions beyond just making maps. The first peer-reviewed publications on using computers to help in the cartographic process predate the introduction of full GIS by several years.
Computer cartography is employed to facilitate a variety of computer applications, often through integration with the Global Positioning System (GPS) satellite network. This can allow real-time automated map generation for tasks such as automotive navigation systems.
History
From paper to paperless
In the 1950, Waldo Tobler published a paper titled "Automation and Cartography" that established the first use case for computers as aids in cartography. In this paper, Tobler established what he referred to as a "map in–map out" (MIMO) system, which facilitated digitization of traditional maps, changing them, and reproducing them. The MIMO system, while simple, established the use of computers for map making in the literature and set the stage for more advanced Geographic information systems in later years by geographers such as Roger Tomlinson. The rapid acceleration that followed lead to a rapid paradigm shift in cartography, where traditional cartography was replaced by computer-aided cartography. This was predicted in 1985, when Mark Monmonier speculated in his book "Technological Transition in Cartography" that com |
https://en.wikipedia.org/wiki/Baguenaudier | Baguenaudier (; French for "time-waster"), also known as the Chinese rings, Cardan's suspension, Cardano's rings, Devil's needle or five pillars puzzle, is a disentanglement puzzle featuring a loop which must be disentangled from a sequence of rings on interlinked pillars. The loop can be either string or a rigid structure.
It is thought to have been invented originally in China. The origins are obscure. The American ethnographer Stewart Culin related a tradition attributing the puzzle's invention to the 2nd/3rd century Chinese general Zhuge Liang. It was used by French peasants as a locking mechanism.
Variations of this include the Devil's staircase, Devil's Halo and the impossible staircase. Another similar puzzle is the Giant's causeway which uses a separate pillar with an embedded ring.
Mathematical solution
The 19th-century French mathematician Édouard Lucas, the inventor of the Tower of Hanoi puzzle, was known to have come up with an elegant solution which used binary and Gray codes, in the same way that his puzzle can be solved. The minimum number of moves to solve an n-ringed problem has been found to be
For other formulae, see .
See also
ABACABA pattern
Disentanglement puzzle
Towers of Hanoi |
https://en.wikipedia.org/wiki/Sticky%20bead%20argument | In general relativity, the sticky bead argument is a simple thought experiment designed to show that gravitational radiation is indeed predicted by general relativity, and can have physical effects. These claims were not widely accepted prior to about 1955, but after the introduction of the bead argument, any remaining doubts soon disappeared from the research literature.
The argument is often credited to Hermann Bondi, who popularized it, but it was originally proposed by Richard Feynman.
Description
The thought experiment was first described by Feynman in 1957 at a conference at Chapel Hill, North Carolina, and later addressed in his private letter to Victor Weisskopf:
As the gravitational waves are mainly transverse, the rod has to be oriented perpendicular to the propagation direction of the wave.
History of arguments on the properties of gravitational waves
Einstein's double reversal
The creator of the theory of general relativity, Albert Einstein, argued in 1916 that gravitational radiation should be produced, according to his theory, by any mass-energy configuration that has a time-varying quadrupole moment (or higher multipole moment). Using a linearized field equation (appropriate for the study of weak gravitational fields), he derived the famous quadrupole formula quantifying the rate at which such radiation should carry away energy. Examples of systems with time varying quadrupole moments include vibrating strings, bars rotating about an axis perpendicular to the symmetry axis of the bar, and binary star systems, but not rotating disks.
In 1922, Arthur Stanley Eddington wrote a paper expressing (apparently for the first time) the view that gravitational waves are in essence ripples in coordinates, and have no physical meaning. He did not appreciate Einstein's arguments that the waves are real.
In 1936, together with Nathan Rosen, Einstein rediscovered the Beck vacuums, a family of exact gravitational wave solutions with cylindrical symmetry (so |
https://en.wikipedia.org/wiki/Burke%27s%20theorem | In queueing theory, a discipline within the mathematical theory of probability, Burke's theorem (sometimes the Burke's output theorem) is a theorem (stated and demonstrated by Paul J. Burke while working at Bell Telephone Laboratories) asserting that, for the M/M/1 queue, M/M/c queue or M/M/∞ queue in the steady state with arrivals is a Poisson process with rate parameter λ:
The departure process is a Poisson process with rate parameter λ.
At time t the number of customers in the queue is independent of the departure process prior to time t.
Proof
Burke first published this theorem along with a proof in 1956. The theorem was anticipated but not proved by O’Brien (1954) and Morse (1955). A second proof of the theorem follows from a more general result published by Reich. The proof offered by Burke shows that the time intervals between successive departures are independently and exponentially distributed with parameter equal to the arrival rate parameter, from which the result follows.
An alternative proof is possible by considering the reversed process and noting that the M/M/1 queue is a reversible stochastic process. Consider the figure. By Kolmogorov's criterion for reversibility, any birth-death process is a reversible Markov chain. Note that the arrival instants in the forward Markov chain are the departure instants of the reversed Markov chain. Thus the departure process is a Poisson process of rate λ. Moreover, in the forward process the arrival at time t is independent of the number of customers after t. Thus in the reversed process, the number of customers in the queue is independent of the departure process prior to time t.
This proof could be counter-intuitive, in the sense that the departure process of a birth-death process is independent of the service offered.
Related results and extensions
The theorem can be generalised for "only a few cases," but remains valid for M/M/c queues and Geom/Geom/1 queues.
It is thought that Burke's theorem does not |
https://en.wikipedia.org/wiki/Dimethylglycine | Dimethylglycine (DMG) is a derivative of the amino acid glycine with the structural formula (CH3)2NCH2COOH. It can be found in beans and liver, and has a sweet taste. It can be formed from trimethylglycine upon the loss of one of its methyl groups. It is also a byproduct of the metabolism of choline.
When DMG was first discovered, it was referred to as Vitamin B16, but, unlike true B vitamins, deficiency of DMG in the diet does not lead to any ill-effects and it is synthesized by the human body in the citric acid cycle meaning it does not meet the definition of a vitamin.
Uses
Dimethylglycine has been suggested for use as an athletic performance enhancer, immunostimulant, and a treatment for autism, epilepsy, or mitochondrial disease. There is no evidence that dimethylglycine is effective for treating mitochondrial disease. Published studies on the subject have shown little to no difference between DMG treatment and placebo in autism spectrum disorders.
Biological activity
Dimethylglycine has been found to act as an agonist of the glycine site of the NMDA receptor.
Preparation
This compound is commercially available as the free form amino acid, and as the hydrochloride salt []. DMG may be prepared by the alkylation of glycine via the Eschweiler–Clarke reaction. In this reaction, glycine is treated with aqueous formaldehyde in formic acid that serves as both solvent and reductant. Hydrochloric acid is added thereafter to give the hydrochloride salt. The free amino acid may have been obtained by neutralization of the acid salt, which has been performed with silver oxide.
H2NCH2COOH + 2 CH2O + 2 HCOOH → (CH3)2NCH2COOH + 2 CO2 + 2 H2O
See also
Monomethylglycine |
https://en.wikipedia.org/wiki/Grotta%20dell%27Addaura | The Addaura cave (Italian: Grotta dell'Addaura) is a complex of three natural grottoes located on the northeast side of Mount Pellegrino in Palermo, Sicily, Southern Italy. The importance of the complex is due to the presence of cave-wall engravings dated to the late Epigravettian (contemporaneous with the Magdalenian) and the Mesolithic.
On the side of Mount Pellegrino, overlooking Palermo, to the southeast of Mondello beach at above sea level, there are some open grottoes and cavities where bones and tools used for hunting have been found, attesting the presence of humans who lived in them beginning in the Paleolithic and into the Mesolithic. The finds are now conserved in Palermo's Regional Archaeological Museum. Their importance is mainly due to the presence of an extraordinary complex of rock engravings that decorate the walls, constituting a unique case in the panorama of prehistoric cave art. The name Addaura comes from , 'the circuit'.
History
The discovery of the graffiti of Addaura was recent and came about quite casually. The three grottoes that make up the Addaura complex in the massif of Mount Pellegrino had already been studied by paleoanthropologists, as the skeleton of a dwarf elephant had been discovered there.
It was after the 1943 Allied invasion of Sicily and their arrival in Palermo that the Allies, in search of a suitable site, decided to use the grottoes for storing munitions and explosives. The accidental explosion of the arsenal at the end of the war brought about the crumbling of the main grotto walls and the collapse of a rock wall, bringing to light the graffiti covered with the patina of time. The graffiti were carefully studied by the archaeologist Jole Bovio Marconi, whose studies were published in 1953.
Since 1997 the Addaura grottoes are no longer open for visitors; the site was closed because of the danger of falling boulders, due to the instability of the rocky ridge above. , the necessary measures to reinforce the ridge have |
https://en.wikipedia.org/wiki/Dalton%27s%20law | Dalton's law (also called Dalton's law of partial pressures) states that in a mixture of non-reacting gases, the total pressure exerted is equal to the sum of the partial pressures of the individual gases. This empirical law was observed by John Dalton in 1801 and published in 1802. Dalton's law is related to the ideal gas laws.
Formula
Mathematically, the pressure of a mixture of non-reactive gases can be defined as the summation:
where p1, p2, ..., pn represent the partial pressures of each component.
where xi is the mole fraction of the ith component in the total mixture of n components .
Volume-based concentration
The relationship below provides a way to determine the volume-based concentration of any individual gaseous component
where ci is the concentration of component i.
Dalton's law is not strictly followed by real gases, with the deviation increasing with pressure. Under such conditions the volume occupied by the molecules becomes significant compared to the free space between them. In particular, the short average distances between molecules increases intermolecular forces between gas molecules enough to substantially change the pressure exerted by them, an effect not included in the ideal gas model.
See also |
https://en.wikipedia.org/wiki/Train%20horn | A train horn is an air horn used as an audible warning device on diesel and electric-powered trains. Its primary purpose is to alert persons and animals to an oncoming train, especially when approaching a level crossing. They are often extremely loud, allowing them to be heard from great distances. They are also used for acknowledging signals given by railroad employees, such as during switching operations. For steam locomotives, the equivalent device is a train whistle.
History and background
Since trains move on fixed rails, they are uniquely susceptible to collision. This is exacerbated by the train's enormous weight and inertia, which make it difficult to quickly stop when encountering an obstacle. Also, trains generally do not stop at level crossings, instead relying on pedestrians and vehicles to clear the tracks when they pass. Therefore, from their beginnings, locomotives have been equipped with loud horns or bells to warn vehicles and pedestrians that they are coming. Steam locomotives had steam whistles, operated from steam produced by their boilers.
As diesel locomotives began to replace steam on most railroads during the mid-20th century, it was realized that the new locomotives were unable to utilize the steam whistles then in use. Early internal combustion locomotives were initially fitted with small truck horns or exhaust-powered whistles, but these were found to be unsuitable and hence the air horn design was scaled up and modified for railroad use. Early train horns often were tonally similar to the air horns still heard on road-going trucks today. It was found that this caused some confusion among people who were accustomed to steam locomotives and the sound of their whistles; when approaching a grade crossing, when some people heard an air horn they expected to see a truck, not a locomotive, and accidents happened. So, locomotive air horns were created that had a much higher, more musical note, tonally much more like a steam whistle. This is |
https://en.wikipedia.org/wiki/Protecode | Protecode was a private company based in Ottawa, Ontario, Canada that provided open source license and security management software used for software development license compliance.
It was acquired by Synopsys in November 2015 for undisclosed terms.
Products
Protecode Enterprise System is a suite of tools for managing open source software licensing. Protecode utilizes a proprietary database of indexed public open source projects to compare against a software codebase in order to identify open source components and determine license obligations.
The Protecode product suite produces a software inventory report, cross-referenced to licensing and copyright attributes, to indicate corresponding licensing obligations. Protecode also offers an audit service which can be used to produce an audit report, containing a software bill of materials and a license obligations report. Software audits are commonly requested prior to business transactions, like mergers and acquisitions or financing.
In 2013, Protecode announced a single seat open source management scanning tool.
Services
Protecode offers code auditing services to detect all open source and third party and associated licensing obligations before a product release or mergers and acquisitions .
Partnerships
Protecode is a member of IBM Partnerworld and System 4 received validation from IBM Rational Software for IBM Rational Team Concert and IBM Rational ClearCase.
Protecode is available on IBM SmartCloud.
Protecode Library Auditor integrates with Perforce.
Protecode is a member of the Linux Foundation and a contributor to their Open Compliance Program. Protecode is also a supporter of the Linux Foundation sponsored SPDX specification.
Protecode is a member of the Eclipse Foundation
and the GENIVI Alliance
Awards
In May 2010, Protecode was included in the “Cool Vendors in Intellectual Property, 2010” report by Gartner, Inc.
In 2009, The Branham Group included Protecode on its list of the Top 25 Canadian IT |
https://en.wikipedia.org/wiki/Hadwiger%E2%80%93Nelson%20problem | In geometric graph theory, the Hadwiger–Nelson problem, named after Hugo Hadwiger and Edward Nelson, asks for the minimum number of colors required to color the plane such that no two points at distance 1 from each other have the same color. The answer is unknown, but has been narrowed down to one of the numbers 5, 6 or 7. The correct value may depend on the choice of axioms for set theory.
Relation to finite graphs
The question can be phrased in graph theoretic terms as follows. Let G be the unit distance graph of the plane: an infinite graph with all points of the plane as vertices and with an edge between two vertices if and only if the distance between the two points is 1. The Hadwiger–Nelson problem is to find the chromatic number of G. As a consequence, the problem is often called "finding the chromatic number of the plane". By the de Bruijn–Erdős theorem, a result of , the problem is equivalent (under the assumption of the axiom of choice) to that of finding the largest possible chromatic number of a finite unit distance graph.
History
According to , the problem was first formulated by Nelson in 1950, and first published by . had earlier published a related result, showing that any cover of the plane by five congruent closed sets contains a unit distance in one of the sets, and he also mentioned the problem in a later paper . discusses the problem and its history extensively.
One application of the problem connects it to the Beckman–Quarles theorem, according to which any mapping of the Euclidean plane (or any higher dimensional space) to itself that preserves unit distances must be an isometry, preserving all distances. Finite colorings of these spaces can be used to construct mappings from them to higher-dimensional spaces that preserve distances but are not isometries. For instance, the Euclidean plane can be mapped to a six-dimensional space by coloring it with seven colors so that no two points at distance one have the same color, and then mapping |
https://en.wikipedia.org/wiki/McGruff%20the%20Crime%20Dog | McGruff the Crime Dog is an anthropomorphic animated bloodhound created by Dancer Fitzgerald Sample advertising executive Jack Keil (who also voiced the character) through the Ad Council and later the National Crime Prevention Council to increase crime awareness and personal safety in the United States. McGruff costumes are used by police outreach efforts, often with children. McGruff was created in 1979 and debuted in 1980 with a series of public service announcements educating citizens on personal security measures, such as locking doors and putting lights on timers, in order to reduce crime. His name was selected as part of a nationwide contest in July 1980.
McGruff proved to be a successful campaign with over $100 million in free air time donated in the first year reaching over 50% of adults. McGruff campaigns continued over the years to cover topics such as child abduction, robbery, anti-drug messages, and anti-bullying campaigns. From 1982 to 2012, a number of municipalities participated in the McGruff house program which offered temporary haven to children fearing immediate harm. McGruff has continued to be well-recognized, with nine out of ten people recognizing him in a 2021 survey. This is thanks partly to recent campaigns against cyber-bullying, stopping online fakes, and elder-crime.
History
Crime as a public concern
The decades prior to McGruff's creation saw an increase in U.S. public concern over crime. In the 1960s, a number of riots broke out across the U.S. and numerous public figures were assassinated, including President Kennedy, Martin Luther King Jr., and Malcolm X. Accepting the Republican nomination for president, Barry Goldwater positioned crime as one of the biggest issues facing the nation. While Goldwater lost to Lyndon Johnson, the issue of crime did not stop there. In July 1965, President Johnson formed the President's Commission on Law Enforcement and Administration of Justice to "probe ... fully and deeply into the problems of crim |
https://en.wikipedia.org/wiki/Frequency%20response | In signal processing and electronics, the frequency response of a system is the quantitative measure of the magnitude and phase of the output as a function of input frequency. The frequency response is widely used in the design and analysis of systems, such as audio and control systems, where they simplify mathematical analysis by converting governing differential equations into algebraic equations. In an audio system, it may be used to minimize audible distortion by designing components (such as microphones, amplifiers and loudspeakers) so that the overall response is as flat (uniform) as possible across the system's bandwidth. In control systems, such as a vehicle's cruise control, it may be used to assess system stability, often through the use of Bode plots. Systems with a specific frequency response can be designed using analog and digital filters.
The frequency response characterizes systems in the frequency domain, just as the impulse response characterizes systems in the time domain. In linear systems (or as an approximation to a real system neglecting second order non-linear properties), either response completely describes the system and thus have one-to-one correspondence: the frequency response is the Fourier transform of the impulse response. The frequency response allows simpler analysis of cascaded systems such as multistage amplifiers, as the response of the overall system can be found through multiplication of the individual stages' frequency responses (as opposed to convolution of the impulse response in the time domain). The frequency response is closely related to the transfer function in linear systems, which is the Laplace transform of the impulse response. They are equivalent when the real part of the transfer function's complex variable is zero.
Measurement and plotting
Measuring the frequency response typically involves exciting the system with an input signal and measuring the resulting output signal, calculating the frequency spectra |
https://en.wikipedia.org/wiki/Splenius%20capitis%20muscle | The splenius capitis () () is a broad, straplike muscle in the back of the neck. It pulls on the base of the skull from the vertebrae in the neck and upper thorax. It is involved in movements such as shaking the head.
Structure
It arises from the lower half of the nuchal ligament, from the spinous process of the seventh cervical vertebra, and from the spinous processes of the upper three or four thoracic vertebrae.
The fibers of the muscle are directed upward and laterally and are inserted, under cover of the sternocleidomastoideus, into the mastoid process of the temporal bone, and into the rough surface on the occipital bone just below the lateral third of the superior nuchal line. The splenius capitis is deep to sternocleidomastoideus at the mastoid process, and to the trapezius for its lower portion. It is one of the muscles that forms the floor of the posterior triangle of the neck.
The splenius capitis muscle is innervated by the posterior ramus of spinal nerves C3 and C4.
Function
The splenius capitis muscle is a prime mover for head extension. The splenius capitis can also allow lateral flexion and rotation of the cervical spine.
Additional images
See also |
https://en.wikipedia.org/wiki/Frederick%20Rossini | Frederick Dominic Rossini (July 18, 1899 – October 12, 1990) was an American thermodynamicist noted for his work in chemical thermodynamics.
In 1920, at the age of twenty-one, Rossini entered Carnegie-Mellon University in Pittsburgh, and soon was awarded a full-time teaching scholarship. He graduated with a B.S. in chemical engineering in 1925, followed by an M.S. degree in science in physical chemistry in 1926.
As a result of reading Lewis and Randall's classical 1923 textbook Thermodynamics and the Free Energy of Chemical Substances he wrote to Gilbert N. Lewis and as a result he was offered a teaching fellowship at the University of California at Berkeley. Among his teachers were Gilbert Lewis and William Giauque. Rossini's doctoral dissertation on the heat capacities of strong electrolytes in aqueous solution was supervised by Merle Randall. His Ph.D. degree was awarded in 1928, after only 21 months of graduate work, even though he continued to serve as a teaching fellow throughout this entire period. He worked at the National Bureau of Standards (Washington, DC) from 1928 to 1950.
In 1932, Frederick Rossini, Edward W. Washburn, and Mikkel Frandsen authored "The Calorimetric Determination of the Intrinsic Energy of Gases as a Function of the Pressure." This experiment resulted in the development of the Washburn Correction for bomb calorimetry, a decrease or correction of the results of a calorimetric procedure to normal states.
In 1950, he published his popular textbook Chemical Thermodynamics. In that year he also moved to the Carnegie Institute of Technology (Pittsburgh), where he remained until 1960. He served as dean of the Notre Dame College of Science from 1960 to 1967.
In 1973, Dr. Rossini spent the spring academic quarter at Baldwin-Wallace College, in Berea Ohio, as the first distinguished professor to occupy the Charles J. Strosacker Chair of Science. The Baldwin-Wallace College student union was named after "the late Dr. strosacker, who was vic |
https://en.wikipedia.org/wiki/Srm%20%28Unix%29 | srm (or Secure Remove) is a command line utility for Unix-like computer systems for secure file deletion. srm removes each specified file by overwriting, renaming, and truncating it before unlinking. This prevents other people from undeleting or recovering any information about the file from the command line.
Platform-specific behaviours and bugs
Filesystems with hard links
Attempting to secure delete a file with multiple hard links results in a warning from srm stating that the current access path has been unlinked, but the data itself was not overwritten or truncated. This is an undocumented feature of srm 1.2.8 on Mac OS X 10.9, and is erroneously documented in 1.2.11 as a behaviour activated by the OpenBSD rm-compatible option -P. However, in both the OS X and SourceForge srm implementations, the behaviour of unlinking but not overwriting multi-linked files is always active, as long as the platform reports hard links.
srm 1.2.8 on Mac OS X 10.9 has a -n option, which means "overwrite file, but do not rename or unlink it." However, if the file has multiple links, the multiple-link file data protection feature activates first, removing the file, even though the -n option specifies "do not rename or unlink the file". The -n option has been removed from the code and manual of srm version 1.2.11, the latest SourceForge.net version. As a consequence, this option/feature conflict does not occur.
OS X
A number of file systems support file forks (called resource forks and named forks on OS X (particularly HFS+), and alternate data streams on NTFS), or extended attributes. However, OS X is the only platform on which srm securely deletes any of this additional data in files.
On OS X, only the most common non-data fork, the resource fork, is handled in this way. This support was included in Apple’s 1.2.8 and SourceForge's 1.2.9.
srm was removed from OS X/macOS in v10.11 El Capitan, as part of the removal of the "Secure Empty Trash" feature for security reasons.
|
https://en.wikipedia.org/wiki/Fat%20Chance%3A%20Probability%20from%200%20to%201 | Fat Chance: Probability from 0 to 1 is an introductory undergraduate-level textbook on probability theory, centered on the metaphor of games of chance. It was written by Benedict Gross, Joe Harris, and Emily Riehl, based on a course for non-mathematicians taught to Harvard University undergraduates, and published by the Cambridge University Press in 2019. An associated online course has been offered to the public by Harvard.
Topics
Unusually for a probability theory book, this book does not use the phrase "random variable", instead referring to random processes as games. The first five chapters of the book concern counting problems, and include material on the exponential function, binomial coefficients, factorials, games of cards, dice, and coins, and the birthday paradox. After an interlude involving the binomial theorem, Pascal's triangle, and the Catalan numbers, the second part of the book concerns probability more directly. Its chapters concern the expected value, conditional probability and Bayes' theorem, events with unequal probabilities (biased coins and loaded dice), geometric probability, the law of large numbers, and normal distributions. The third part moves from probability to statistics, with topics including the central limit theorem and the meaning of false positives and false negatives in medical testing.
Audience and reception
Although the main purpose of the book is to be a textbook for college courses aimed at non-mathematicians, it can also be read independently by those interested in the topic. Reviewer Ludwig Paditz recommends the book to "readers without deeper knowledge in elementary statistics and probability". Reviewer Massimo Nespolo recommends as well that its readers take advantage of the associated online course offering. |
https://en.wikipedia.org/wiki/Louise%20Stokes%20Hunter | Ella Louise Stokes Hunter (died 1988) was an American mathematics educator who became the first African-American woman to earn a degree at the University of Virginia. She taught for many years at Virginia Normal and Industrial Institute and Virginia State College, two names for what is now Virginia State University.
Early life and education
Hunter was originally from Petersburg, Virginia. After studying at Peabody High School in Petersburg and the Virginia Normal and Industrial Institute, the predecessor institution to Virginia State, she went to Howard University, joined the Alpha Kappa Alpha sorority, and graduated in 1920. She earned a master's degree in education from Harvard University in 1925. Although African American women such as Alberta Virginia Scott had previously graduated from Radcliffe College, she may have been the first to earn a degree from Harvard proper.
Later in life, while working as a faculty member at Virginia State, Hunter became a doctoral student at the University of Virginia, studying mathematics education and doing her doctoral dissertation research on the transition from high school to college mathematics. She completed her Ph.D. in 1953, becoming the first African-American woman to earn a degree at the university, two months after another doctoral student in education, Walter N. Ridley, became the first African-American with a degree from the University of Virginia.
Career and later life
After graduating from Howard University, Hunter became an instructor at the Virginia Normal and Industrial Institute, where she taught for many years. In 1921, she was one of six instructors there who banded together to found the Delta Omega graduate chapter of Alpha Kappa Alpha (originally called the Nu chapter), and later she became its first historian and eighth president. On the faculty at the Virginia Normal and Industrial Institute, she met John McNeile Hunter, who began teaching electrical engineering there in 1925 and later became the third |
https://en.wikipedia.org/wiki/Monoidal%20adjunction | Suppose that and are two monoidal categories. A monoidal adjunction between two lax monoidal functors
and
is an adjunction between the underlying functors, such that the natural transformations
and
are monoidal natural transformations.
Lifting adjunctions to monoidal adjunctions
Suppose that
is a lax monoidal functor such that the underlying functor has a right adjoint . This adjunction lifts to a monoidal adjunction ⊣ if and only if the lax monoidal functor is strong.
See also
Every monoidal adjunction ⊣ defines a monoidal monad .
Adjoint functors
Monoidal categories |
https://en.wikipedia.org/wiki/Sunsite | SunSITE (Sun Software, Information & Technology Exchange) is a network of Internet servers providing archives of information, software and other publicly available resources. The project, started in the early 1990s, is run by a number of universities worldwide and was initially co-sponsored by Sun Microsystems.
The more notable SunSITEs include:
SunSITE Canada, operated by University of British Columbia = Found Without Content - 2022.04.28
SunSITE Central Europe, operated by RWTH Aachen, Germany
Sun SITE Central Europe Workshop Proceedings (CEUR-WS.org) Free Open-Access
SunSITE Poland, operated by ICM, University of Warsaw
Some former SunSITEs:
SunSITE Chile
SunSITE Czech Republic, operated by School of Computer Science, Charles University, Prague = Server Not Found - 2022.04.28
SunSITE Denmark, now running as dotsrc.org Open Source Hosting
SunSITE Mexico = Blank Page - 2022.04.28
SunSITE North Carolina, operated by University of North Carolina at Chapel Hill, now running as Ibiblio
SunSITE RedIris (Spain), operated by Spanish National Research Network = Server Not Found - 2022.04.28
SunSITE Singapore, operated by National University of Singapore = Blank Page - 2022.04.28
SunSITE Switzerland, operated by SWITCH Information Technology Services, now running as SWITCHmirror
SunSITE Tennessee operated by University of Tennessee, Knoxville = Server Not Found - 2022.04.28
SunSITE Thailand operated by Assumption University, Bangkok = Server Not Found - 2022.04.28
University of Alberta SunSITE, now running as the University of Alberta Digital Object Repository (UADORe)
No longer in operation:
SunSITE Austria , operated by University of Vienna
SunSITE Argentina, operated by Universidad de Buenos Aires.
Berkeley Digital Library SunSITE, University of California, Berkeley Libraries
SunSITE Hungary, run by Institute of Mathematics, University of Debrecen
SunSITE Indonesia, operated by Faculty of Computer Science, University of Indonesia, Jakarta
SunSIT |
https://en.wikipedia.org/wiki/Association%20Electronique%20Libre | The Association Electronique Libre (AEL, sometimes written Association électronique libre) is a Belgian non-profit digital rights advocacy and legal organization based in Belgium. Its stated mission is to "protecting the fundamental rights in the information society". It is promoting the Free Software Pact Initiative.
It has organized protests against the concept of software patents. |
https://en.wikipedia.org/wiki/Henri%20Lebesgue | Henri Léon Lebesgue (; June 28, 1875 – July 26, 1941) was a French mathematician known for his theory of integration, which was a generalization of the 17th-century concept of integration—summing the area between an axis and the curve of a function defined for that axis. His theory was published originally in his dissertation Intégrale, longueur, aire ("Integral, length, area") at the University of Nancy during 1902.
Personal life
Henri Lebesgue was born on 28 June 1875 in Beauvais, Oise. Lebesgue's father was a typesetter and his mother was a school teacher. His parents assembled at home a library that the young Henri was able to use. His father died of tuberculosis when Lebesgue was still very young and his mother had to support him by herself. As he showed a remarkable talent for mathematics in primary school, one of his instructors arranged for community support to continue his education at the Collège de Beauvais and then at Lycée Saint-Louis and Lycée Louis-le-Grand in Paris.
In 1894 Lebesgue was accepted at the École Normale Supérieure, where he continued to focus his energy on the study of mathematics, graduating in 1897. After graduation he remained at the École Normale Supérieure for two years, working in the library, where he became aware of the research on discontinuity done at that time by René-Louis Baire, a recent graduate of the school. At the same time he started his graduate studies at the Sorbonne, where he learned about Émile Borel's work on the incipient measure theory and Camille Jordan's work on the Jordan measure. In 1899 he moved to a teaching position at the Lycée Central in Nancy, while continuing work on his doctorate. In 1902 he earned his PhD from the Sorbonne with the seminal thesis on "Integral, Length, Area", submitted with Borel, four years older, as advisor.
Lebesgue married the sister of one of his fellow students, and he and his wife had two children, Suzanne and Jacques.
After publishing his thesis, Lebesgue was offered in |
https://en.wikipedia.org/wiki/Comparison%20of%20software%20for%20molecular%20mechanics%20modeling | This is a list of computer programs that are predominantly used for molecular mechanics calculations.
See also
Car–Parrinello molecular dynamics
Comparison of force-field implementations
Comparison of nucleic acid simulation software
List of molecular graphics systems
List of protein structure prediction software
List of quantum chemistry and solid-state physics software
List of software for Monte Carlo molecular modeling
List of software for nanostructures modeling
Molecular design software
Molecular dynamics
Molecular modeling on GPUs
Molecule editor
Notes and references
External links
SINCRIS
Linux4Chemistry
Collaborative Computational Project
World Index of Molecular Visualization Resources
Short list of Molecular Modeling resources
OpenScience
Biological Magnetic Resonance Data Bank
Materials modelling and computer simulation codes
A few tips on molecular dynamics
atomistic.software - atomistic simulation engines and their citation trends
Computational chemistry software
Computational chemistry
Software comparisons
Molecular dynamics software
Molecular modelling software
Science software |
https://en.wikipedia.org/wiki/Domain%20model | In software engineering, a domain model is a conceptual model of the domain that incorporates both behavior and data. In ontology engineering, a domain model is a formal representation of a knowledge domain with concepts, roles, datatypes, individuals, and rules, typically grounded in a description logic.
Overview
A domain model is a system of abstractions that describes selected aspects of a sphere of knowledge, influence or activity (a domain). The model can then be used to solve problems related to that domain.
The domain model is a representation of meaningful real-world concepts pertinent to the domain that need to be modeled in software. The concepts include the data involved in the business and rules the business uses in relation to that data. A domain model leverages natural language of the domain.
A domain model generally uses the vocabulary of the domain, thus allowing a representation of the model to be communicated to non-technical stakeholders. It should not refer to any technical implementations such as databases or software components that are being designed.
Usage
A domain model is generally implemented as an object model within a layer that uses a lower-level layer for persistence and "publishes" an API to a higher-level layer to gain access to the data and behavior of the model.
In the Unified Modeling Language (UML), a class diagram is used to represent the domain model.
See also
Domain-driven design (DDD)
Domain layer
Feature-driven development
Logical data model
OntoUML |
https://en.wikipedia.org/wiki/Organelle%20biogenesis | Organelle biogenesis is the biogenesis, or creation, of cellular organelles in cells. Organelle biogenesis includes the process by which cellular organelles are split between daughter cells during mitosis; this process is called organelle inheritance.
Discovery
Following the discovery of cellular organelles in the nineteenth century, little was known about their function and synthesis until the development of electron microscopy and subcellular fractionation in the twentieth century. This allowed experiments on the function, structure, and biogenesis of these organelles to commence.
Mechanisms of protein sorting and retrieval have been found to give organelles their characteristic composition. It is known that cellular organelles can come from preexisting organelles; however, it is a subject of controversy whether organelles can be created without a preexisting one.
Process
Several processes are known to have developed for organelle biogenesis. These can range from de novo synthesis to the copying of a template organelle; the formation of an organelle 'from scratch' and using a preexisting organelle as a template to manufacture an organelle, respectively. The distinct structures of each organelle are thought to be caused by the different mechanisms of the processes which create them and the proteins that they are made up of. Organelles may also be 'split' between two cells during the process of cellular division (known as organelle inheritance), where the organelle of the parent cell doubles in size and then splits with each half being delivered to their respective daughter cells.
The process of organelle biogenesis is known to be regulated by specialized transcription networks that modulate the expression of the genes that code for specific organellar proteins. In order for organelle biogenesis to be carried out properly, the specific genes coding for the organellar proteins must be transcribed properly and the translation of the resulting mRNA must be succes |
https://en.wikipedia.org/wiki/HL23V | HL23V was reputedly a type C RNA tumor virus first isolated in 1975 from cultured human acute myelogenous leukaemia peripheral blood leukocytes in the laboratory of Robert Gallo, which would have been the first cancer-causing retrovirus isolated from human sera. It was later shown to be a laboratory contaminant of three monkey viruses. The journal Nature, which had published the original research, later retracted the article. Investigative journalist John Crewdson, writing in 2003, described how Gallo pretended mysterious laboratory accidents such as the unplugging of a laboratory freezer to excuse his failure to share samples of "HL23V" to researchers. As "HL23V" would have been the first human retrovirus discovered, bringing attention to Gallo and thus scientific prizes, observers noted that "23" was the number of Robert Gallo's birthday.
Later in 1986, Max Essex of Harvard would announce the "discovery" of "HTLV-IV" in Senegalese women, supposedly a type C relative of HIV (then called HTLV-III) thought to be HIV-2. This similarly was determined to be a contaminant.
See also
Simian sarcoma associated virus (SSAV or SSAV-1), one of two associated viruses comprising HL23V
HTLV-1, the actual first human pathogenic retrovirus discovered in 1981 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.