source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Welbeck%20Street | Welbeck Street is a street in the West End, central London. It has historically been associated with the medical profession.
Former resident Andrew Berry was one of the men to have successfully deployed a parachute at altitude less than 3000 ft
Location
The street runs approximately north–south between New Cavendish Street at the northern end, crossing Wigmore Street near Wigmore Hall just to the east, becoming Vere Street continuing southwards. The nearest tube station is Bond Street to the south. The part south of Wigmore Street is part of the B406.
The London Welbeck Hospital, is located at 27 Welbeck Street, and the Welbeck Street Hospital for Diseases of the Nervous System was located on this street as well; the offices of the British Institute of Radiology were formerly located there. The Welbeck Clinic is located at No. 20.
There is a Russian Orthodox Chapel at 32 Welbeck Street that dates back as far as the early 19th century when the building was the residence of the Russian Embassy Chaplain. The chapel was rebuilt in 1864 and features a particularly fine iconostasis. The chapel is located behind No. 32, on the east side of the street near the northern end, and can be seen from Marylebone Mews (it is visible on Edward Stanford's 1862 map of London).
Notable people
The notorious 18th-century highwayman James MacLaine was once a grocer on Welbeck Street.
In 1799, Thomas Young established himself as a physician in this street at No 48, now recorded by a blue plaque. The street was favoured by doctors at the time and remains a leading medical location. It is close to Harley Street, now more famed for its concentration of private medical practitioners.
General John Egerton, 7th Earl of Bridgewater (b. 14 Apr 1753, d. 21 Oct 1823) married Charlotte Catherine Anne Haynes, daughter of Samuel Haynes and Elizabeth, on 14 January 1783 at 58 Welbeck Street.
Flautist Robert Sidney Pratten and his wife, the guitar virtuoso, composer and teacher Catharina Joseph |
https://en.wikipedia.org/wiki/Public%20Knowledge%20Project | The Public Knowledge Project (PKP) is a non-profit research initiative that is focused on the importance of making the results of publicly funded research freely available through open access policies, and on developing strategies for making this possible including software solutions. It is a partnership between the Faculty of Education at the University of British Columbia, the Canadian Centre for Studies in Publishing at Simon Fraser University, the University of Pittsburgh, Ontario Council of University Libraries, the California Digital Library and the School of Education at Stanford University. It seeks to improve the scholarly and public quality of academic research through the development of innovative online environments.
History
The PKP was founded in 1998 by John Willinsky in the Department of Language and Literacy Education at the Faculty of Education at the University of British Columbia, in Vancouver, British Columbia, Canada, based on his research in education and publishing. Willinsky is a leading advocate of open access publishing, and has written extensively on the value of public research.
The PKP's initial focus was on increasing access to scholarly research and output beyond the traditional academic environments. This soon led to a related interest in scholarly communication and publishing, and especially on ways to make it more cost effective and less reliant on commercial enterprises and their generally restricted access models. PKP has developed free, open source software for the management, publishing, and indexing of journals, conferences, and monographs.
The PKP has collaborated with a wide range of partners interested in making research publicly available, including the Scholarly Publishing and Academic Resources Coalition (SPARC), the Brazilian Institute for Information Science and Technology (IBICT), and the International Network for the Availability of Scientific Publications (INASP).
Together with INASP, the PKP is working with pub |
https://en.wikipedia.org/wiki/CRESTA | CRESTA (Catastrophe Risk Evaluation and Standardizing Target Accumulations) was founded as a joint project of Swiss Reinsurance Company, Gerling-Konzern Globale Reinsurance Company, and Munich Reinsurance Company. CRESTA has set itself the aim of establishing a globally uniform system for the accumulation risk control of natural hazards - particularly earthquakes, storms and floods. Those risk zones are essentially based on the observed and expected seismic activity, as well as on other natural disasters, such as droughts, floods and storms. CRESTA zones regard the distribution of insured values within a region or country for easier assessment of risks. CRESTA Zones are the essential basis for reinsurance negotiation and portfolio analysis. Nowadays, CRESTA sets widely accepted standards which apply throughout the international insurance industry. CRESTA zone information is used by most insurers for assessing the insurance catastrophe premiums they will charge.
Origin of the name
While the acronym CRESTA stands for Catastrophe Risk Evaluation and Standardizing Target Accumulations, the name was derived from the name of the hotel (Cresta Hotel) where the founding meeting occurred, and the subsequent creation of a suitable acronym to correspond to the name.
About CRESTA
CRESTA has set itself the aim of establishing a globally uniform system for the accumulation risk control of natural hazards - particularly earthquakes, storms and floods. Nowadays, widely accepted standards apply throughout the international insurance industry. CRESTA's main tasks are:
Determining country-specific zones for the uniform and detailed reporting of accumulation risk data relating to natural hazards and creating corresponding zonal maps for each country.
Drawing up standardised accumulation risk-recording forms for each country.
Working out a uniform format for the processing and electronic transfer of accumulation risk data between insurance and reinsurance companies.
In additio |
https://en.wikipedia.org/wiki/Multiprotocol%20Encapsulation%20over%20ATM | Multiprotocol Encapsulation over ATM is specified in RFC 2684. It defines two mechanisms for identifying the protocol carried in ATM Adaptation Layer 5 (AAL5) frames. It replaces RFC 1483, a standard data link access protocol supported by DSL modems.
RFC 2684 describes two encapsulation mechanisms for network traffic: Virtual Circuit Multiplexing and LLC Encapsulation. Either mechanism carries either routed or bridged protocol data units, and DSL modems often include a setting for RFC 1483 bridging. This is distinct from other "bridge modes" commonly found in combined DSL modems and routers, which turn off the router portion of the DSL modem.
In VC Multiplexing (VC-MUX), the hosts agree on the high-level protocol for a given circuit. It has the advantage of not requiring additional information in a packet, which minimises the overhead. For example, if the hosts agree to transfer IP, a sender can pass each datagram directly to AAL5 to transfer; nothing needs to be sent besides the datagram and the AAL5 trailer. The chief disadvantage of such a scheme lies in duplication of virtual circuits: a host must create a separate virtual circuit for each high-level protocol if more than one protocol is used. Because most carriers charge for each virtual circuit, customers try to avoid using multiple circuits because it adds unnecessary cost.
In LLC Encapsulation the hosts use a single virtual circuit for multiple protocols. This has the advantage of allowing all traffic over the same circuit, but the disadvantage of requiring each packet to contain octets that identify the protocol type, which adds overhead. The scheme also has the disadvantage that packets from all protocols travel with the same delay and priority.
RFC 2684 specifies that hosts can choose between the two methods of using AAL5. Both the sender and receiver must agree on how the circuit will be used, and the agreement may involve manual configuration. Furthermore, the standards suggest that when h |
https://en.wikipedia.org/wiki/Ragel | Ragel is a finite-state machine compiler and a parser generator. Initially Ragel supported output for C, C++ and Assembly source code, and was expanded to support several other languages including Objective C, D, Go, Ruby, and Java. Additional language support is also in development. It supports the generation of table or control flow driven state machines from regular expressions and/or state charts and can also build lexical analysers via the longest-match method. Ragel specifically targets text parsing and input validation.
Overview
Ragel supports the generation of table or control flow driven state machines from regular expressions and/or state charts and can also build lexical analysers via the longest-match method.
A unique feature of Ragel is that user actions can be associated with arbitrary state machine transitions using operators that are integrated into the regular expressions. Ragel also supports visualization of the generated machine via graphviz.
The above graph represents a state-machine that takes user input as a series of bytes representing ASCII characters and control codes. 48..57 is equivalent to the regular expression [0-9] (i.e. any digit), so only sequences beginning with a digit can be recognised. If 10 (line feed) is encountered, we're done. 46 is the decimal point ('.'), 43 and 45 are positive and negative signs ('+', '-') and 69/101 is uppercase/lowercase 'e' (to indicate a number in scientific format). As such it will recognize the following properly:
2
45
055
78.1
2e5
78.3e12
69.0e-3
3e+3
but not:
.3
46.
-5
3.e2
2e5.1
Syntax
Ragel's input is a regular expression only in the sense that it describes a regular language; it is usually not written in a concise regular expression, but written out into multiple parts like in Extended Backus–Naur form. For example, instead of supporting POSIX character classes in regex syntax, Ragel implements them as built-in production rules. As with usual parser generators, Ragel allows for handling |
https://en.wikipedia.org/wiki/Viropexis | Viropexis is the process by which different classes of viruses—particularly picornaviruses and papovaviruses—enter the host cell in which they will be able to replicate. The hydrophobic structures of the capsid proteins may be exposed after viral binding to the cell (see viral attachment protein). These structures help the virion or the viral genome slip through the membrane. It can be juxtaposed with viral endocytosis, which is receptor mediated, and doesn't involve direct penetration of the virion.
See also
Sources
Virology |
https://en.wikipedia.org/wiki/Food%20safety | Food safety (or food hygiene) is used as a scientific method/discipline describing handling, preparation, and storage of food in ways that prevent foodborne illness. The occurrence of two or more cases of a similar illness resulting from the ingestion of a common food is known as a food-borne disease outbreak. This includes a number of routines that should be followed to avoid potential health hazards. In this way, food safety often overlaps with food defense to prevent harm to consumers. The tracks within this line of thought are safety between industry and the market and then between the market and the consumer. In considering industry-to-market practices, food safety considerations include the origins of food including the practices relating to food labeling, food hygiene, food additives and pesticide residues, as well as policies on biotechnology and food and guidelines for the management of governmental import and export inspection and certification systems for foods. In considering market-to-consumer practices, the usual thought is that food ought to be safe in the market and the concern is safe delivery and preparation of the food for the consumer. Food safety, nutrition and food security are closely related. Unhealthy food creates a cycle of disease and malnutrition that affects infants and adults as well.
Food can transmit pathogens, which can result in the illness or death of the person or other animals. The main types of pathogens are bacteria, viruses, parasites, and fungus. The WHO Foodborne Disease Epidemiology Reference Group conducted the only study that solely and comprehensively focused on the global health burden of foodborne diseases. This study, which involved the work of over 60 experts for a decade, is the most comprehensive guide to the health burden of foodborne diseases. The first part of the study revealed that 31 foodborne hazards considered priority accounted for roughly 420,000 deaths in LMIC and posed a burden of about 33 million disa |
https://en.wikipedia.org/wiki/Kannan%20Soundararajan | Kannan Soundararajan (born December 27, 1973) is an Indian-born American mathematician and a professor of mathematics at Stanford University. Before moving to Stanford in 2006, he was a faculty member at University of Michigan, where he had also pursued his undergraduate studies. His main research interest is in analytic number theory, particularly in the subfields of automorphic L-functions, and multiplicative number theory.
Early life
Soundararajan grew up in Madras and was a student at Padma Seshadri High School in Nungambakkam in Madras. In 1989, he attended the prestigious Research Science Institute. He represented India at the International Mathematical Olympiad in 1991 and won a Silver Medal.
Education
Soundararajan joined the University of Michigan, Ann Arbor, in 1991 for undergraduate studies, and graduated with highest honours in 1995. Soundararajan won the inaugural Morgan Prize in 1995 for his work in analytic number theory while an undergraduate at the University of Michigan, where he later served as professor. He joined Princeton University in 1995 and did his Ph.D under the guidance of Professor Peter Sarnak.
Career
After his Ph.D. he received the first five-year fellowship from the American Institute of Mathematics, and held positions at Princeton University, the Institute for Advanced Study, and the University of Michigan. He moved to Stanford University in 2006 where he is, as of November 2022, the Anne T. and Robert M. Bass Professor of Mathematics.
He provided a proof of a conjecture of Ron Graham in combinatorial number theory jointly with Ramachandran Balasubramanian. He made important contributions in settling the arithmetic Quantum Unique Ergodicity conjecture for Maass wave forms and modular forms.
Awards
He received the Salem Prize in 2003 "for contributions to the area of Dirichlet L-functions and related character sums". In 2005, he won the $10,000 SASTRA Ramanujan Prize, shared with Manjul Bhargava, awarded by SASTRA in Thanjavur, |
https://en.wikipedia.org/wiki/Printed%20electronics | Printed electronics is a set of printing methods used to create electrical devices on various substrates. Printing typically uses common printing equipment suitable for defining patterns on material, such as screen printing, flexography, gravure, offset lithography, and inkjet. By electronic-industry standards, these are low-cost processes. Electrically functional electronic or optical inks are deposited on the substrate, creating active or passive devices, such as thin film transistors; capacitors; coils; resistors. Some researchers expect printed electronics to facilitate widespread, very low-cost, low-performance electronics for applications such as flexible displays, smart labels, decorative and animated posters, and active clothing that do not require high performance.
The term printed electronics is often related to organic electronics or plastic electronics, in which one or more inks are composed of carbon-based compounds. These other terms refer to the ink material, which can be deposited by solution-based, vacuum-based, or other processes. Printed electronics, in contrast, specifies the process, and, subject to the specific requirements of the printing process selected, can utilize any solution-based material. This includes organic semiconductors, inorganic semiconductors, metallic conductors, nanoparticles, and nanotubes.
For the preparation of printed electronics nearly all industrial printing methods are employed. Similar to conventional printing, printed electronics applies ink layers one atop another. So the coherent development of printing methods and ink materials are the field's essential tasks.
The most important benefit of printing is low-cost volume fabrication. The lower cost enables use in more applications. An example is RFID-systems, which enable contactless identification in trade and transport. In some domains, such as light-emitting diodes printing does not impact performance. Printing on flexible substrates allows electronics to be pla |
https://en.wikipedia.org/wiki/Background%20selection | Background selection describes the loss of genetic diversity at a non-deleterious locus due to negative selection against linked deleterious alleles. It is one form of linked selection, where the maintenance or removal of an allele from a population is dependent upon the alleles in its linkage group. The name emphasizes the fact that the genetic background, or genomic environment, of a neutral mutation has a significant impact on whether it will be preserved (genetic hitchhiking) or purged (background selection) from a population. In some cases, the term background selection is used broadly to refer to all forms of linked selection, but most often it is used only when neutral variation is reduced due to negative selection against deleterious mutations. Background selection and all forms of linked selection contradict the assumption of the neutral theory of molecular evolution that the fixation or loss of neutral alleles is entirely stochastic, the result of genetic drift. Instead, these models predict that neutral variation is correlated with the selective pressures acting on linked non-neutral genes, that neutral traits are not necessarily oblivious to selection. Because they segregate together, non-neutral mutations linked to neutral polymorphisms result in decreased levels of genetic variation relative to predictions of neutral evolution.
Relation to the neutral model
The reduction in neutral variation due to background selection can be modeled by an exponential function of the total mutation rate at the deleterious regions of the section of genome involved. The overall effect of background selection on genetic diversity resembles a reduction in effective population size. As a consequence, background selection has been used to explain many of the inconsistencies between classical models of neutral variation and observed studies of genetic diversity. For instance, the observation that genetic diversity is weakly correlated with population size, or not correlated |
https://en.wikipedia.org/wiki/Muscle%20coactivation | Muscle coactivation occurs when agonist and antagonist muscles (or synergist muscles) surrounding a joint contract simultaneously to provide joint stability, and is suggested to depend crucially on supraspinal processes involved in the control of movement. It is also known as muscle cocontraction, since two muscle groups are contracting at the same time. It is able to be measured using electromyography (EMG) from the contractions that occur. The general mechanism of it is still widely unknown. It is believed to be important in joint stabilization, as well as general motor control.
Function
Muscle coactivation allows muscle groups surrounding a joint to become more stable. This is due to both muscles (or sets of muscles) contracting at the same time, which produces compression on the joint. The joint is able to become stiffer and more stable due to this action. For example, when the biceps and the triceps coactivate, the elbow becomes more stable. This stabilization mechanism is also important for unexpected loads impeded on the joint, allowing the muscles to quickly coactivate and provide stability to the joint. This mechanism is controlled neuromuscularly, which allows the muscle(s) to contract. This occurs through a motor neuron sending a signal (through creating action potentials) to the muscle fiber to contract by releasing acetylcholine. When signals are sent to all muscle fibers in a muscle group, the muscle group will contract as a whole.
In the upper limbs, the stability of muscle coactivation allows for precise low-level physical tasks. An example of this would be picking up a small object. By protecting the muscles at the end of their range of motion, the direction of the fine movements is able to be changed. In the lower limbs, stability is important in upright standing balance. The coactivation of different muscle groups allows for proper balance and the ability to adjust weight and to stay upright on uneven ground. It is also believed to be importan |
https://en.wikipedia.org/wiki/Quantum%20potential | The quantum potential or quantum potentiality is a central concept of the de Broglie–Bohm formulation of quantum mechanics, introduced by David Bohm in 1952.
Initially presented under the name quantum-mechanical potential, subsequently quantum potential, it was later elaborated upon by Bohm and Basil Hiley in its interpretation as an information potential which acts on a quantum particle. It is also referred to as quantum potential energy, Bohm potential, quantum Bohm potential or Bohm quantum potential.
In the framework of the de Broglie–Bohm theory, the quantum potential is a term within the Schrödinger equation which acts to guide the movement of quantum particles. The quantum potential approach introduced by Bohm provides a physically less fundamental exposition of the idea presented by Louis de Broglie: de Broglie had postulated in 1925 that the relativistic wave function defined on spacetime represents a pilot wave which guides a quantum particle, represented as an oscillating peak in the wave field, but he had subsequently abandoned his approach because he was unable to derive the guidance equation for the particle from a non-linear wave equation. The seminal articles of Bohm in 1952 introduced the quantum potential and included answers to the objections which had been raised against the pilot wave theory.
The Bohm quantum potential is closely linked with the results of other approaches, in particular relating to work by Erwin Madelung of 1927 and to work by Carl Friedrich von Weizsäcker of 1935.
Building on the interpretation of the quantum theory introduced by Bohm in 1952, David Bohm and Basil Hiley in 1975 presented how the concept of a quantum potential leads to the notion of an "unbroken wholeness of the entire universe", proposing that the fundamental new quality introduced by quantum physics is nonlocality.
Quantum potential as part of the Schrödinger equation
The Schrödinger equation
is re-written using the polar form for the wave function wi |
https://en.wikipedia.org/wiki/Clivus%20%28anatomy%29 | The clivus (, Latin for "slope"), or Blumenbach clivus, is a bony part of the cranium at the base of the skull. It is a shallow depression behind the dorsum sellae of the sphenoid bone. It slopes gradually to the anterior part of the basilar occipital bone at its junction with the sphenoid bone. It extends to the foramen magnum. It is related to the pons and the abducens nerve (CN VI).
Structure
The clivus is a shallow depression behind the dorsum sellae of the sphenoid bone. It slopes gradually to the anterior part of the basilar occipital bone at its junction with the sphenoid bone. Synchondrosis of these two bones forms the clivus. The clivus extends inferiorly to the foramen magnum. On axial planes, it sits just posterior to the sphenoid sinuses. It is medial to the foramen lacerum (the internal carotid artery reaches the middle cranial fossa above the foramen lacerum), proximal to its anastomosis with the Circle of Willis. It is anterior to the basilar artery. On sagittal plane, it can be divided into two surfaces including the pharyngeal (inferior) surface and basilar (superior) surface. A small elevation known as the pharyngeal tubercle is present on the inferior surface for the fibrous raphe of pharynx to attach.
The pons sits on the basilar surface of the clivus. The abducens nerve (CN VI) also tracks along the clivus during its course.
Variations
During embryonic development, the clivus is formed by the fusion of the basiocciput and basisphenoid or also known as the sphenooccipital synchondrosis. When the fusion occurs improperly, it would give rise to gaps that are considered anatomical variations. Variations of the clivus include fossa navicularis magna, craniopharyngeal canal, canalis basilaris medianus, and transverse basilar fissure (Saucer's fissure). Ossification of the apical ligament of dens may also occur, resulting in a variant bony tubercle at the inferior end of the clivus. Condylus tertius and arcus praebasiocipitalis are the other two |
https://en.wikipedia.org/wiki/Lipocalin | The lipocalins are a family of proteins which transport small hydrophobic molecules such as steroids, bilins, retinoids, and lipids, and most lipocalins are also able to bind to complexed iron (via siderophores or flavonoids) as well as heme. They share limited regions of sequence homology and a common tertiary structure architecture. This is an eight stranded antiparallel beta barrel with a repeated + 1 topology enclosing an internal ligand binding site.
These proteins are found in gram negative bacteria, vertebrate cells, and invertebrate cells, and in plants. Lipocalins have been associated with many biological processes, among them immune response, pheromone transport, biological prostaglandin synthesis, retinoid binding, and cancer cell interactions.
Function
Immune response
Lipocalin proteins are important key players of nutritional immunity by withholding and sequestering micronutrients. They are thereby able to regulate inflammatory and detoxification processes caused by immune system activation in mammals. They are known respiratory allergens of mice, cats, dogs, horses, and other animals. Examples of lipocalin proteins involved in immune system responses include alpha-1-microglobulin, alpha-1-acid glycoprotein, and C8gamma. Structural information for many immune system influencing lipocalin proteins is available, while their exact role in biological systems is still somewhat unclear. Lipocalin allergens have been shown to evoke a Th2-deviated immune response, important for allergic sensitization, when applied in their apo-form (with an empty calyx devoid of ligands), whereas the holo-form seemed to exert immune-suppressive properties in vitro.
Pheromone transport
The lipocalin family has been connected with the transport of mammalian pheromones due to easily observable protein-pheromone interactions. Lipocalins are comparatively small in size, and are thus less complicated to study as opposed to large, bulky proteins. They can also bind to |
https://en.wikipedia.org/wiki/Colleter%20%28botany%29 | Colleters are plant structures, multicellular secretory hairs, found in groups near the base of petioles, on stipules, and on sepals. They are found in members of the Loganiaceae and Rubiaceae families. |
https://en.wikipedia.org/wiki/Hypercone | In geometry, a hypercone (or spherical cone) is the figure in the 4-dimensional Euclidean space represented by the equation
It is a quadric surface, and is one of the possible 3-manifolds which are 4-dimensional equivalents of the conical surface in 3 dimensions. It is also named "spherical cone" because its intersections with hyperplanes perpendicular to the w-axis are spheres. A four-dimensional right hypercone can be thought of as a sphere which expands with time, starting its expansion from a single point source, such that the center of the expanding sphere remains fixed. An oblique hypercone would be a sphere which expands with time, again starting its expansion from a point source, but such that the center of the expanding sphere moves with a uniform velocity.
Parametric form
A right spherical hypercone can be described by the function
with vertex at the origin and expansion speed s.
A right spherical hypercone with radius r and height h can be described by the function
An oblique spherical hypercone could then be described by the function
where is the 3-velocity of the center of the expanding sphere.
An example of such a cone would be an expanding sound wave as seen from the point of view of a moving reference frame: e.g. the sound wave of a jet aircraft as seen from the jet's own reference frame.
Note that the 3D-surfaces above enclose 4D-hypervolumes, which are the 4-cones proper.
Geometrical interpretation
The spherical cone consists of two unbounded nappes, which meet at the origin and are the analogues of the nappes of the 3-dimensional conical surface. The upper nappe corresponds with the half with positive w-coordinates, and the lower nappe corresponds with the half with negative w-coordinates.
If it is restricted between the hyperplanes w = 0 and w = r for some nonzero r, then it may be closed by a 3-ball of radius r, centered at (0,0,0,r), so that it bounds a finite 4-dimensional volume. This volume is given by the formula r4, and is the 4 |
https://en.wikipedia.org/wiki/Karava%20heraldry | Karava Heraldry includes a large number of Karava flags that have survived the ravages of time. Many are illustrated in E. W. Perera's monumental book titled 'Sinhalese Banners and Standards'.
The Symbols
The Makara
Beknopte History, 1688 describes a white flag charged with a red flag hoisted on royal ships (JRASCB XI No.38 106 & 109). And in the words of Valentine Francois, the 18th century Dutch Historian, “the Karawas displayed a white flag with the device of a particular fish in the centre".
The Makara, is a composite dragon with a curious mythical structure. It symbolizes the house of Capricorn in the Zodiac to which it has given its name Makara in the Hindu calendar. It has the head of a crocodile, horns of a goat, the body of an antelope and a snake, the tail of a fish and feet of a panther. Makara is half animal half fish and it is sometimes described as having the head of an elephant and the body of a fish. It is generally large and regarded as living in the ocean rather than in lakes or streams.
Only Varuna, the spiritual ruler of the world has power over the Makara. It is Varuna's vehicle in Hindu mythology. The Varunakulasuriya clan is the largest of the Karava clans in southern Sri Lanka. The symbolism is extremely interesting. In mythology Varuna is the chief of the Adithyas. Remnants of the name Adithya from the medieval period can still be found in Karava family names. As Adithya is a synonym for Suriya (i.e. the Sun). the Karava clan Varunakulasuriya too signifies Varuna-Adithya.
The Kokila Sandesha poem from the Kotte period refers to the Makara flag as follows:
Punsanda surindu sanda salakuna adina vara
Ban sonda telitudew tele tudeni mana hara
Min dada jaya virudu nada karana piya kara
An koda mediya tura topa sarivana pavara
See Fish and ship symbols of Sri Lanka for the significance of the Fish and Ship symbols on these flags.
The Elephant
The recurrent Elephant symbol may be from the Mahabharata, Hastinapura which means Elephant |
https://en.wikipedia.org/wiki/Quasi-Frobenius%20Lie%20algebra | In mathematics, a quasi-Frobenius Lie algebra
over a field is a Lie algebra
equipped with a nondegenerate skew-symmetric bilinear form
, which is a Lie algebra 2-cocycle of with values in . In other words,
for all , , in .
If is a coboundary, which means that there exists a linear form such that
then
is called a Frobenius Lie algebra.
Equivalence with pre-Lie algebras with nondegenerate invariant skew-symmetric bilinear form
If is a quasi-Frobenius Lie algebra, one can define on another bilinear product by the formula
.
Then one has
and
is a pre-Lie algebra.
See also
Lie coalgebra
Lie bialgebra
Lie algebra cohomology
Frobenius algebra
Quasi-Frobenius ring |
https://en.wikipedia.org/wiki/Neuronal%20calcium%20sensor-1 | Neuronal calcium sensor-1 (NCS-1) also known as frequenin homolog (Drosophila) (freq) is a protein that is encoded by the FREQ gene in humans. NCS-1 is a member of the neuronal calcium sensor family, a class of EF hand containing calcium-myristoyl-switch proteins.
Function
NCS-1 regulates synaptic transmission, helps control the dynamics of nerve terminal growth, is critical for some forms of learning and memory in C. elegans and mammals, regulates corticohippocampal plasticity; and enhancing levels of NCS-1 in the mouse dentate gyrus increases spontaneous exploration of safe environments, potentially linking NCS-1 to curiosity.
NCS-1 is a calcium sensor, not a calcium buffer (chelator); thus it is a high-affinity, low-capacity, calcium-binding protein.
Frq can substitute for calmodulin in some situations. It is thought to be associated with neuronal secretory vesicles and regulate neurosecretion.
It is the Ca2+-sensing subunit of the yeast phosphatidylinositol (PtdIns)-4-OH kinase, PIK1
It binds to many proteins, some in calcium dependent and some in calcium independent ways, and switches many of the targets "on" (some off).
Calcineurin (protein phosphatase 2B)
GRK2 (G-protein-coupled receptor kinase 2)
D2 dopamine receptor
IL1RAPL1 (interleukin-1 receptor accessory protein-like 1 protein)
PI4KIIIβ (type III phosphatidylinositol 4-kinase β)
IP3 receptor (this activity is inhibited by lithium - a drug used for the treatment of bipolar disorder)
3',5'-cyclic nucleotide phosphodiesterases
ARF1 (ADP Ribosylation factor 1)
A type (Kv4.3; Shal-related subfamily, member 3) voltage-gated potassium channels
Nitric oxide synthase
TRPC5 channel
Ric8a
Frq modulates Ca2+ entry through a functional interaction with the α1 voltage-gated Ca2+-channel subunit.
Structure
NCS-1 is a globular protein consisting of ten alpha-helices. Four pairs of alpha-helices each form independent 12-amino-acid loops containing a negatively charged calcium binding domain kno |
https://en.wikipedia.org/wiki/Neuronal%20calcium%20sensor | Neuronal Calcium Sensor is a large family of proteins which work as calcium dependent molecular switches and includes members like Frequenin (NCS1), recoverin, GCAP, neurocalcin, visinin etc. All the members carry 4 EF hand motifs (out of which only 2 or 3 bind calcium) and an N-myristoyl group.
Members of NCS family
Highly evolutionarily conserved
NCS1 (Frequenin)
VILIP-1 (Visinin-like-protein-1)
HPCAL4 (Visinin-like-protein-2)
HPCAL1 (Visinin-like-protein-3)
hippocalcin
neurocalcin
recoverin
Guanylate cyclase activator proteins (GCAPs)
Potassium Channel interacting proteins (KChIPs 1–4), including:
KCNIP1,
KCNIP2,
Calsenilin or DREAM/KChIP-3/KCNIP3 (downstream regulatory element antagonist modulator/potassium channel interacting protein),
KCNIP4 |
https://en.wikipedia.org/wiki/Six%20nines%20in%20pi | A sequence of six consecutive nines occurs in the decimal representation of the number pi (), starting at the 762nd decimal place. It has become famous because of the mathematical coincidence, and because of the idea that one could memorize the digits of up to that point, and then suggest that is rational. The earliest known mention of this idea occurs in Douglas Hofstadter's 1985 book Metamagical Themas, where Hofstadter states
This sequence of six nines is sometimes called the "Feynman point", after physicist Richard Feynman, who allegedly stated this same idea in a lecture. However it is not clear when, or even if, Feynman made such a statement. It is not mentioned in published biographies or in his autobiographies, and is unknown to his biographer, James Gleick.
Related statistics
is conjectured, but not known, to be a normal number. For a normal number sampled uniformly at random, the probability of a specific sequence of six digits occurring this early in the decimal representation is about 0.08%.
The early string of six 9's is also the first occurrence of four and five consecutive identical digits. The next sequence of six consecutive identical digits is again composed of 9's, starting at position 193,034. The next distinct sequence of six consecutive identical digits after that starts with the digit 8 at position 222,299.
The positions of the first occurrence of a string of 1, 2, 3, 4, 5, 6, 7, 8, and 9 consecutive 9's in the decimal expansion are 5; 44; 762; 762; 762; 762; 1,722,776; 36,356,642; and 564,665,206, respectively .
Decimal expansion
The first 1,001 digits of (1,000 decimal places), showing consecutive runs of three or more digits including the consecutive six 9's underlined, are as follows:
See also
0.999...
9 (number)
Mathematical coincidence
Repdigit
Ramanujan's constant |
https://en.wikipedia.org/wiki/Intelligent%20sensor | An intelligent sensor is a sensor that takes some predefined action when it senses the appropriate input (light, heat, sound, motion, touch, etc.).
Description
The sensor has to do the following tasks:
Give a digital signal.
Be able to communicate the signal.
Be able to execute logical functions and instructions.
Elements of intelligent sensors
Primary sensing element
Excitation control
Amplification
Analogue filtering
Data conversion
Compensation
Digital information processing
Digital communication processing
Technical capacities
Because the tasks are performed by microprocessors, any gadget which mixes a sensor and a microprocessor is usually called an intelligent sensor.
To qualify as an intelligent sensor, the sensor and processor must be part of the same physical unit. A sensor whose only function is to detect and send an unprocessed signal to an external system which performs some action is not considered intelligent.
Ubiquitous Sensor Networks (USN)
Ubiquitous Sensor Networks (USN) is used to describe a network of intelligent sensors that
could, one day, become ubiquitous.
See also
Real-time locating system
Smart transducer
Internet of things |
https://en.wikipedia.org/wiki/Distance%20measure | Distance measures are used in physical cosmology to give a natural notion of the distance between two objects or events in the universe. They are often used to tie some observable quantity (such as the luminosity of a distant quasar, the redshift of a distant galaxy, or the angular size of the acoustic peaks in the cosmic microwave background (CMB) power spectrum) to another quantity that is not directly observable, but is more convenient for calculations (such as the comoving coordinates of the quasar, galaxy, etc.). The distance measures discussed here all reduce to the common notion of Euclidean distance at low redshift.
In accord with our present understanding of cosmology, these measures are calculated within the context of general relativity, where the Friedmann–Lemaître–Robertson–Walker solution is used to describe the universe.
Overview
There are a few different definitions of "distance" in cosmology which are all asymptotic one to another for small redshifts. The expressions for these distances are most practical when written as functions of redshift , since redshift is always the observable. They can also be written as functions of scale factor
In the remainder of this article, the peculiar velocity is assumed to be negligible unless specified otherwise.
We first give formulas for several distance measures, and then describe them in more detail further down. Defining the "Hubble distance" as
where is the speed of light, is the Hubble parameter today, and is the dimensionless Hubble constant, all the distances are asymptotic to for small .
According to the Friedmann equations, we also define a dimensionless Hubble parameter:
Here, and are normalized values of the present radiation energy density, matter density, and "dark energy density", respectively (the latter representing the cosmological constant), and determines the curvature. The Hubble parameter at a given redshift is then .
The formula for comoving distance, which serves as the ba |
https://en.wikipedia.org/wiki/Jozef%20Schell | Jozef Stefaan "Jeff", Baron Schell (20 July 1935 – 17 April 2003) was a Belgian molecular biologist.
Schell studied zoology and microbiology at the University of Ghent, Belgium. From 1967 to 1995 he worked as a professor at the university. From 1978 to 2000 he was director and head of the Max Planck Institute for Plant Breeding Research (Institut für Züchtungsforschung) at the Max-Planck-Gesellschaft in Cologne, Germany. He received many prizes, among which were the Francqui Prize in 1979, the Wolf Prize in Agriculture in 1990, and the Japan Prize in 1998, which he shared with Marc Van Montagu. He also was appointed Professeur Honoraire, Collège de France, Paris in 1998. He was granted the title of Baron by Baudouin of Belgium.
Schell was a pioneer in genetics who focused on the interaction between plants and soil bacteria. Along with his colleague, Marc Van Montagu, Jeff Schell discovered the gene transfer mechanism between Agrobacterium and plants, which resulted in the development of methods to alter Agrobacterium into an efficient delivery system for gene engineering in plants. Besides being a prominent scientist, in 1982 he co-founded, with Marc Van Montagu, the successful biotech company Plant Genetic Systems Inc., now part of Bayer CropScience.
See also
Walter Fiers
Mary-Dell Chilton
Flanders Institute for Biotechnology (VIB)
Selected publications |
https://en.wikipedia.org/wiki/Marc%20Van%20Montagu | Marc, Baron Van Montagu (born 10 November 1933 in Ghent) is a Belgian molecular biologist. He was full professor and director of the Laboratory of Genetics at the faculty of Sciences at Ghent University (Belgium) and scientific director of the genetics department of the Flanders Interuniversity Institute for Biotechnology (VIB). Together with Jozef Schell he founded the biotech company Plant Genetic Systems Inc. (Belgium) in 1982, of which he was scientific director and member of the board of directors. Van Montagu was also involved in founding the biotech company CropDesign, of which he was a board member from 1998 to 2004. He is president of the Public Research and Regulation Initiative (PRRI).
Van Montagu and his colleagues were credited with the discovery of the Ti plasmid. They described the gene transfer mechanism between Agrobacterium and plants, which resulted in the development of methods to alter Agrobacterium into an efficient delivery system for gene engineering and to create transgenic plants. They developed plant molecular genetics, in particular molecular mechanisms for cell proliferation and differentiation and response to abiotic stresses (high light, ozone, cold, salt and drought) and constructed transgenic crops (tobacco, rape seed, corn) resistant to insect pest and tolerant to novel herbicides. His work with poplar trees resulted in engineering of trees with improved pulping qualities.
After his retirement as director of the Laboratory of Genetics at Ghent University, Marc Van Montagu created IPBO - International Plant Biotechnology Outreach, VIB-Ghent University, with the mission to foster biotechnological solutions to global agriculture. In 2015 IPBO launched the “Marc and Nora Van Montagu (MNVM) Fund” with focus on sustainable agriculture and agro-industry to the African continent.
Honors
Van Montagu has been a foreign associate of the United States National Academy of Sciences since 1986, the agricultural Academy of Russia and France, the |
https://en.wikipedia.org/wiki/Japanese%20Industrial%20Standards | are the standards used for industrial activities in Japan, coordinated by the Japanese Industrial Standards Committee (JISC) and published by the Japanese Standards Association (JSA). The JISC is composed of many nationwide committees and plays a vital role in standardizing activities across Japan.
History
In the Meiji era, private enterprises were responsible for making standards, although the Japanese government too had standards and specification documents for procurement purposes for certain articles, such as munitions.
These were summarized to form an official standard, the Japanese Engineering Standard, in 1921. During World War II, simplified standards were established to increase matériel output.
The present Japanese Standards Association was established in 1946, a year after Japan's defeat in World War II. The Japanese Industrial Standards Committee regulations were promulgated in 1946, and new standards were formed.
The Industrial Standardization Law was enacted in 1949, which forms the legal foundation for the present Japanese Industrial Standards.
New JIS mark
The Industrial Standardization Law was revised in 2004 and the JIS product certification mark was changed; since October 1, 2005, the new JIS mark has been used upon re-certification. Use of the old mark was allowed during a three-year transition period ending on September 30, 2008, and every manufacturer was able to use the new JIS mark. Therefore all JIS-certified Japanese products manufactured since October 1, 2008, have had the new JIS mark.
Standards classification and numbering
Standards are named in the format "JIS X 0208:1997", where X denotes area division, followed by four digits designating the area (five digits for ISO-corresponding standards), and four final digits designating the revision year.
Divisions of JIS and significant standards are:
A
Civil engineering and architecture
JIS A 0001 – Basic module to ISO 1006
JIS A 0002 – Glossary of terms used in building module to I |
https://en.wikipedia.org/wiki/JACKSNNZ | In international diplomacy, JACKSNNZ is the colloquial name of an informal grouping of the world's affluent non-EU countries, excluding the United States. The JACKSNNZ states are Japan, Australia, Canada, South Korea, Switzerland, Norway and New Zealand.
The term originated in the proceedings to the Sixth Review Conference of the Biological and Toxins Weapons Convention held in Geneva in 2006. In the previous review conference talks broke down over American refusals to allow for a verification mechanism be established to monitor biological weapons programs in states parties. This was against the wishes of other WEOG (Western European and Others Group) states, which also include Canada, Turkey, Australasia and Western Europe. At the 2006 Review Conference the JACKSNNZ states remain supportive of a verification protocol (although are unlikely to push for it knowing that the current US government will not accede on this point). However, the JACKSNNZ also seeks balance within the WEOG, and to protect the interests of non-EU states.
Takeshi Aoki, director of the Bioweapons and Chemical Weapons Conventions Division of the Japanese Ministry of Foreign Affairs said that JACKSNNZ is "neither a binding instrument, nor an exclusive one." The JACKSNNZ states are yet to declare a common and exclusive position in other international fora under this name.
See also
CANZUK
JUSCANZ
Middle power
Umbrella Group |
https://en.wikipedia.org/wiki/Spectral%20flatness | Spectral flatness or tonality coefficient, also known as Wiener entropy, is a measure used in digital signal processing to characterize an audio spectrum. Spectral flatness is typically measured in decibels, and provides a way to quantify how much a sound resembles a pure tone, as opposed to being noise-like.
The meaning of tonal in this context is in the sense of the amount of peaks or resonant structure in a power spectrum, as opposed to flat spectrum of a white noise. A high spectral flatness (approaching 1.0 for white noise) indicates that the spectrum has a similar amount of power in all spectral bands — this would sound similar to white noise, and the graph of the spectrum would appear relatively flat and smooth. A low spectral flatness (approaching 0.0 for a pure tone) indicates that the spectral power is concentrated in a relatively small number of bands — this would typically sound like a mixture of sine waves, and the spectrum would appear "spiky".
The spectral flatness is calculated by dividing the geometric mean of the power spectrum by the arithmetic mean of the power spectrum, i.e.:
where x(n) represents the magnitude of bin number n. Note that a single (or more) empty bin yields a flatness of 0, so this measure is most useful when bins are generally not empty.
The ratio produced by this calculation is often converted to a decibel scale for reporting, with a maximum of 0 dB and a minimum of −∞ dB.
The spectral flatness can also be measured within a specified subband, rather than across the whole band. Dubnov has shown that spectral flatness is equivalent to information theoretic concept of mutual information that is known as dual total correlation.
Applications
This measurement is one of the many audio descriptors used in the MPEG-7 standard, in which it is labelled "AudioSpectralFlatness".
In birdsong research, it has been used as one of the features measured on birdsong audio, when testing similarity between two excerpts. Spectral flatness h |
https://en.wikipedia.org/wiki/Piassava | Piassava, also piaçava (), piaçaba (), piasaba, pissaba, piassaba, and piaçá (), is a fibrous product of Brazilian palm species Attalea funifera and Leopoldinia piassaba. It is often used in making brooms and for other purposes.
Piassava was historically exported to Europe before the widespread use of synthetic materials such as plastic. Today, it is mostly used locally in South America.
See also
West African piassava palm |
https://en.wikipedia.org/wiki/Teramac | The Teramac was an experimental massively parallel computer designed by HP in the 1990s. The name reflected the project's vision to provide a programmable gate array system with capacity for a million gates running at a megahertz. Contrary to traditional systems, which are useless if there is one defect, Teramac used defective processors -- intentionally -- to demonstrate its defect-tolerant architecture. Even though the computer had 220,000 hardware defects, it was able to perform some tasks 100 times faster than a single-processor high-end workstation.
Teramac was originally developed by scientists in HP's central research lab, HP Labs, in the mid 1990s. Although it contained conventional silicon integrated circuit technology, it paved the way for some of HP's work in nanoelectronics because it provided an architecture on which a chemically assembled computer could operate.
The experience from this program was used to design the Field Programmable Nanowire Interconnect circuit.
Further reading
Computational science
Massively parallel computers |
https://en.wikipedia.org/wiki/Hitachi-LG%20Data%20Storage | Hitachi-LG Data Storage (HLDS, HL-DT-ST or H-L Data Storage), a joint venture between Hitachi, Ltd. and LG Electronics, is a manufacturer of DVD and Blu-ray optical disc drives for desktop computers and laptops. Founded in late 2000, the company began operation in January 2001. In 2006, HLDS began developing Blu-ray Disc drives. The company claims that it has led the disk drive industry in market share since its founding, with a 20% share for fiscal year 2001, 29% for fiscal year 2012, and 60% for fiscal year 2016. Building upon its core optical technology and extensive experience in production management, the company is currently expanding its product lineup to include interactive digital signage, TOF (Time of Flight) 3D sensors, unmanned store solutions, and in-vehicle air purifiers. |
https://en.wikipedia.org/wiki/Social%20medicine | Social medicine is an interdisciplinary field that focuses on the profound interplay between socio-economic factors and individual health outcomes. Rooted in the challenges of the Industrial Revolution, it seeks to:
Understand how specific social, economic, and environmental conditions directly impact health, disease, and the delivery of medical care.
Promote conditions and interventions that address these determinants, aiming for a healthier and more equitable society.
Social medicine as a scientific field gradually began in the early 19th century, the Industrial Revolution and the subsequent increase in poverty and disease among workers raised concerns about the effect of social processes on the health of the poor. The field of social medicine is most commonly addressed today by efforts to understand what are known as social determinants of health.
Scope
The major emphasis on biomedical science in medical education, health care, and medical research has resulted into a gap with our understanding and acknowledgement of far more important social determinants of health and individual disease: social-economic inequalities, war, illiteracy, detrimental life-styles (smoking, obesity), discrimination because of race, gender and religion. Farmer et al. (2006) gave the following explanation for this gap:
The holy grail of modern medicine remains the search for a molecular basis of disease. While the practical yield of such circumscribed inquiry has been enormous, exclusive focus on molecular-level phenomena has contributed to the increasing "desocialization" of scientific inquiry: a tendency to ask only biological questions about what are in fact biosocial phenomena.
They further concluded that "Biosocial understandings of medical phenomena are urgently needed".
Social medicine is a vast and evolving field, and its scope can cover a wide range of topics that touch on the intersection of society and health. The scope of social medicine medicine includes:
Social Det |
https://en.wikipedia.org/wiki/Zoia%20Ceau%C8%99escu | Zoia Ceaușescu (; 28 February 1949 – 20 November 2006) was a Romanian mathematician, the daughter of Communist leader Nicolae Ceaușescu and his wife, Elena. She was also known as Tovarășa Zoia (comrade Zoia).
Biography
Zoia Ceaușescu studied at High School nr. 24 (now Jean Monnet High School) in Bucharest and graduated in 1966. She then continued her studies at the Faculty of Mathematics, University of Bucharest. She received her Ph.D. in 1977 with thesis On Intertwining Dilations written under the direction of Ciprian Foias. Ceaușescu then worked as a researcher at the Institute of Mathematics of the Romanian Academy in Bucharest. Her field of specialization was functional analysis. Allegedly, her parents were unhappy with their daughter's choice of doing research in mathematics, so the Institute was disbanded in 1975. She moved on to work for Institutul pentru Creație Științifică și Tehnică (INCREST, Institute for Scientific and Technical Creativity), where she eventually started and headed a new department of mathematics. In 1976, Ceaușescu received the Simion Stoilow Prize for her outstanding contributions to the mathematical sciences.
She was married in 1980 to Mircea Oprean, an engineer and professor at the Polytechnic University of Bucharest.
During the Romanian Revolution, on 24 December 1989, she was arrested for "undermining the Romanian economy" and was released only eight months later, on 18 August 1990. After she was freed, she tried unsuccessfully to return to her former job at INCREST, then gave up and retired. After the revolution, some newspapers reported that she had lived a wild life, having plenty of lovers and often being drunk.
After her parents were executed, the new government confiscated the house where she and her husband lived (the house was used as proof of allegedly stolen wealth), so she had to live with friends.
After the revolution that ousted her parents, Zoia reported that during her parents' time in power her mother had asked |
https://en.wikipedia.org/wiki/Tonsil | The tonsils are a set of lymphoid organs facing into the aerodigestive tract, which is known as Waldeyer's tonsillar ring and consists of the adenoid tonsil (or pharyngeal tonsil), two tubal tonsils, two palatine tonsils, and the lingual tonsils. These organs play an important role in the immune system.
When used unqualified, the term most commonly refers specifically to the palatine tonsils, which are two lymphoid organs situated at either side of the back of the human throat. The palatine tonsils and the adenoid tonsil are organs consisting of lymphoepithelial tissue located near the oropharynx and nasopharynx (parts of the throat).
Structure
Humans are born with four types of tonsils: the pharyngeal tonsil, two tubal tonsils, two palatine tonsils and the lingual tonsils.
Development
The palatine tonsils tend to reach their largest size in puberty, and they gradually undergo atrophy thereafter. However, they are largest relative to the diameter of the throat in young children. In adults, each palatine tonsil normally measures up to 2.5 cm in length, 2.0 cm in width and 1.2 cm in thickness.
The adenoid grows until the age of 5, starts to shrink at the age of 7 and becomes small in adulthood.
Function
The tonsils are immunocompetent organs which serve as the immune system's first line of defense against ingested or inhaled foreign pathogens, and as such frequently engorge with blood to assist in immune responses to common illnesses such as the common cold. The tonsils have on their surface specialized antigen capture cells called microfold cells (M cells) that allow for the uptake of antigens produced by pathogens. These M cells then alert the B cells and T cells in the tonsil that a pathogen is present and an immune response is stimulated. B cells are activated and proliferate in areas called germinal centers in the tonsil. These germinal centers are places where B memory cells are created and secretory antibody (IgA) is produced.
Clinical significance
Th |
https://en.wikipedia.org/wiki/Nuclear%20Holocausts%3A%20Atomic%20War%20in%20Fiction | Nuclear Holocausts: Atomic War in Fiction, 1895–1984 was written by Paul Brians and published in 1987. This comprehensive study covers nuclear holocaust fiction published in English-language works between 1895 and 1984. Brians notes that 1895 marked the first appearance of an atomic weapon in fiction: Robert Cromie's Crack of Doom. The latter date marks the conclusion of this study's time span.
Brians' work examines only fiction actually depicting nuclear war and its aftermath. He does not include works of "future holocaust" where the triggering incident is not fully explained in the work. The included bibliography covers over 800 items, including novels, stories, and films from the post-holocaust genre. The included titles are arranged by author and include a brief synopsis of each work. The author has provided an online version of the published work, complete with updated sources.
Reception
Joe Dewey positively reviewed it for MFS Modern Fiction Studies in 1988, noting that "Brians' fulsome bibliography represents a valuable achievement". Paul Boyer reviewed this book for Technology and Culture in 1989. His review was also positive, concluding that Brain's work is a "thoughtful and exhaustive examination" of the topic. Alexander H. McIntire who reviewed this work the same year for the Science Fiction Studies was likewise impressed, writing that "Brians provides a treasure trove: 250 pages of extensively annotated bibliography, arranged by author’s surname, and consisting of all the works mentioned in the narrative and hundreds more. This bibliography of 800 entries, supplemented by full title and subject indexes, is reason enough to make Holocausts a part of any SF scholar’s personal reference library.".
Notes
1987 books
Books about nuclear issues
American post-apocalyptic novels
Kent State University Press books |
https://en.wikipedia.org/wiki/Flavored%20syrup | Flavored syrups typically consist of a simple syrup, that is sugar (fully mixed with water while heated), with naturally occurring or artificial (synthesized) flavorings also dissolved in them. A sugar substitute may also be used.
Flavored syrups may be used or mixed with carbonated water, coffee, pancakes, waffles, tea, cake, ice cream, and other foods. There are hundreds of flavors ranging from cherry and peach to vanilla to malt, hazelnut, coconut, almond, gingerbread, chocolate, peppermint, rootbeer, and even toasted marshmallow.
In addition to food and drink, flavored syrups are commonly used in pharmaceutical compounding.
In coffee and espresso drinks
Flavored syrups can also be used to make frappes, which when made with coffee can have other names such as frappuccino, a word created by Starbucks Corporation to trademark their version.
See also
Agave nectar
Birch syrup
Bludwine
Chocolate syrup
Italian soda
List of syrups
Maple syrup
Soda jerk
Torani |
https://en.wikipedia.org/wiki/JSGF | JSGF stands for Java Speech Grammar Format or the JSpeech Grammar Format (in a W3C Note). Developed by Sun Microsystems, it is a textual representation of grammars for use in speech recognition for technologies like XHTML+Voice. JSGF adopts the style and conventions of the Java programming language in addition to use of traditional grammar notations.
The Speech Recognition Grammar Specification was derived from this specification.
Example
The following JSGF grammar will recognize the words coffee, tea, and milk.
#JSGF V1.0;
grammar numbers;
public <drinks> = coffee | tea | milk;
See also
SRGS
XHTML+Voice
External links
JSpeech Grammar Format (W3C Note)
Interactive JSGF Generator
Speech recognition |
https://en.wikipedia.org/wiki/K-set%20%28geometry%29 | In discrete geometry, a -set of a finite point set in the Euclidean plane is a subset of elements of that can be strictly separated from the remaining points by a line. More generally, in Euclidean space of higher dimensions, a -set of a finite point set is a subset of elements that can be separated from the remaining points by a hyperplane. In particular, when (where is the size of ), the line or hyperplane that separates a -set from the rest of is a halving line or halving plane.
The -sets of a set of points in the plane are related by projective duality to the -levels in an arrangement of lines. The -level in an arrangement of lines in the plane is the curve consisting of the points that lie on one of the lines and have exactly lines below them. Discrete and computational geometers have also studied levels in arrangements of more general kinds of curves and surfaces.
Combinatorial bounds
It is of importance in the analysis of geometric algorithms to bound the number of -sets of a planar point set, or equivalently the number of -levels of a planar line arrangement, a problem first studied by Lovász and Erdős et al. The best known upper bound for this problem is , as was shown by Tamal Dey using the crossing number inequality of Ajtai, Chvátal, Newborn, and Szemerédi. However, the best known lower bound is far from Dey's upper bound: it is for some constant , as shown by Tóth.
In three dimensions, the best upper bound known is , and the best lower bound known is .
For points in three dimensions that are in convex position, that is, are the vertices of some convex polytope, the number of -sets is
, which follows from arguments used for bounding the complexity of th order Voronoi diagrams.
For the case when (halving lines), the maximum number of combinatorially distinct lines through two points of that bisect the remaining points when is
Bounds have also been proven on the number of -sets, where a -set is a -set for some . In two dimensions, the |
https://en.wikipedia.org/wiki/Indication%20%28medicine%29 | In medicine, an indication is a valid reason to use a certain test, medication, procedure, or surgery. There can be multiple indications to use a procedure or medication. An indication can commonly be confused with the term diagnosis. A diagnosis is the assessment that a particular [medical] condition is present while an indication is a reason for use. The opposite of an indication is a contraindication, a reason to withhold a certain medical treatment because the risks of treatment clearly outweigh the benefits.
In the United States, indications for prescription drugs are approved by the FDA. Indications are included in the Indications and Usage section of the Prescribing Information. The primary role of this section of labeling is to enable health care practitioners to readily identify appropriate therapies for patients by clearly communicating the drug's approved indication(s). The Indications and Usage section states the disease or condition, or manifestation or symptoms thereof, for which the drug is approved, as well as whether the drug is indicated for the treatment, prevention, mitigation, cure, relief, or diagnosis of that disease or condition. Additionally, the Indications and Usage section should contain the approved age groups as well as other information necessary to describe appropriate use (e.g., identifying the indicated patient/disease subgroups, stating if adjunctive therapy is required).
Drugs
Most countries and jurisdictions have a licensing body whose duty is to determine whether to approve a drug for a specific indication, based on the relative safety of the drug and its efficacy for the particular use. In the United States, indications for medications are regulated by the Food and Drug Administration (FDA), and are included in the package insert under the phrase "Indications and Usage". The European Medicines Agency (EMA) holds this responsibility for centrally authorized drugs in the European Union.
Label indications vs. off-label indica |
https://en.wikipedia.org/wiki/Kasner%20metric | The Kasner metric (developed by and named for the American mathematician Edward Kasner in 1921) is an exact solution to Albert Einstein's theory of general relativity. It describes an anisotropic universe without matter (i.e., it is a vacuum solution). It can be written in any spacetime dimension and has strong connections with the study of gravitational chaos.
Metric and conditions
The metric in spacetime dimensions is
,
and contains constants , called the Kasner exponents. The metric describes a spacetime whose equal-time slices are spatially flat, however space is expanding or contracting at different rates in different directions, depending on the values of the . Test particles in this metric whose comoving coordinate differs by are separated by a physical distance .
The Kasner metric is an exact solution to Einstein's equations in vacuum when the Kasner exponents satisfy the following Kasner conditions,
The first condition defines a plane, the Kasner plane, and the second describes a sphere, the Kasner sphere. The solutions (choices of ) satisfying the two conditions therefore lie on the sphere where the two intersect (sometimes confusingly also called the Kasner sphere). In spacetime dimensions, the space of solutions therefore lie on a dimensional sphere .
Features
There are several noticeable and unusual features of the Kasner solution:
The volume of the spatial slices is always . This is because their volume is proportional to , and
where we have used the first Kasner condition. Therefore can describe either a Big Bang or a Big Crunch, depending on the sense of
Isotropic expansion or contraction of space is not allowed. If the spatial slices were expanding isotropically, then all of the Kasner exponents must be equal, and therefore to satisfy the first Kasner condition. But then the second Kasner condition cannot be satisfied, for
The Friedmann–Lemaître–Robertson–Walker metric employed in cosmology, by contrast, is able to e |
https://en.wikipedia.org/wiki/List%20of%20the%20bryophytes%20of%20Canada | Parent page: Flora of Canada
This is a listing of the non-vascular plants of Canada, and includes the mosses, liverworts and hornworts.
IDD - incomplete distribution data
Anthocerotophyta (hornworts)
Anthocerotaceae
Notothyladaceae
Bryophyta (mosses)
Andreaeaceae
Andreaeobryaceae
Archidiaceae
Aulacomniaceae
Bartramiaceae
Bryaceae
Catoscopiaceae
Hypopterygiaceae
Meesiaceae
Mniaceae
Lists of plants
Non-vascular plants
Bryophytes |
https://en.wikipedia.org/wiki/EXPRESS%20%28data%20modeling%20language%29 | EXPRESS is a standard for generic data modeling language for product data. EXPRESS is formalized in the ISO Standard for the Exchange of Product model STEP (ISO 10303), and standardized as ISO 10303-11.
Overview
Data models formally define data objects and relationships among data objects for a domain of interest. Some typical applications of data models include supporting the development of databases and enabling the exchange of data for a particular area of interest. Data models are specified in a data modeling language. EXPRESS is a data modeling language defined in ISO 10303-11, the EXPRESS Language Reference Manual.
An EXPRESS data model can be defined in two ways, textually and graphically. For formal verification and as input for tools such as SDAI the textual representation within an ASCII file is the most important one. The graphical representation on the other hand is often more suitable for human use such as explanation and tutorials. The graphical representation, called EXPRESS-G, is not able to represent all details that can be formulated in the textual form.
EXPRESS is similar to programming languages such as Pascal. Within a SCHEMA various datatypes can be defined together with structural constraints and algorithmic rules. A main feature of EXPRESS is the possibility to formally validate a population of datatypes - this is to check for all the structural and algorithmic rules.
EXPRESS-G
EXPRESS-G is a standard graphical notation for information models. It is a companion to the EXPRESS language for displaying entity and type definitions, relationships and cardinality. This graphical notation supports a subset of the EXPRESS language. One of the advantages of using EXPRESS-G over EXPRESS is that the structure of a data model can be presented in a more understandable manner. A disadvantage of EXPRESS-G is that complex constraints cannot be formally specified. Figure 1 is an example. The data model presented in figure could be used to specify the r |
https://en.wikipedia.org/wiki/Ascoseira | Ascoseira is a monotypic genus of seaweed in the brown algae (class Phaeophyceae). The single and type species, Ascoseira mirabilis Skottsberg, is a large parenchymatous macroalgae, and is endemic to the Antarctic Ocean. Ascoseira is assigned to its own order. The alga grows in subtidal waters at depths of from 3 to 15 meters. |
https://en.wikipedia.org/wiki/KASH%20domain | KASH domains are conserved C-terminal protein regions less than ~30 amino acids. KASH is an acronym for Klarsicht, ANC-1, Syne Homology. KASH domains always follow a transmembrane domain. Most proteins containing KASH domains are thought to be involved in the positioning of the nucleus in the cell. KASH domains interact with proteins containing SUN domains in the space between the outer and inner nuclear membranes to bridge the nuclear envelope, and may transfer force from the nucleoskeleton to the cytoplasmic cytoskeleton and enable mechanosensory roles in cells. KASH proteins are thought to largely localize to the outer nuclear membrane, although there are reports of inner nuclear membrane localization of some KASH protein isoforms.
Examples of KASH proteins
Caenorhabditis elegans
UNC-83
ANC-1
ZYG-12
Mammals
Nesprins-1, 2, 3 and 4 (also called Synes, Mynes, Nuance, Enaptin)
Drosophila melanogaster
Klarsicht
MSP-300 |
https://en.wikipedia.org/wiki/Phoneword | Phonewords are mnemonic phrases represented as alphanumeric equivalents of a telephone number. In many countries, the digits on the telephone keypad also have letters assigned. By replacing the digits of a telephone number with the corresponding letters, it is sometimes possible to form a whole or partial word, an acronym, abbreviation, or some other alphanumeric combination.
Phonewords are the most common vanity numbers, although a few all-numeric vanity phone numbers are used. Toll-free telephone numbers are often branded using phonewords; some firms use easily memorable vanity telephone numbers like 1-800 Contacts, 1-800-Flowers, 1-866-RING-RING, or 1-800-GOT-JUNK? as brands for flagship products or names for entire companies.
Local numbers are also occasionally used, such as +1-514-AUTOBUS or STM-INFO to reach the Société de transport de Montréal, but are subject to the constraint that the first few digits are tied to a geographic location - potentially limiting the available choices based on which telephone exchanges serve a local area.
Advantages
The main advantages of phonewords over standard phone numbers include increased memorability and increased response rates to advertising. They are easier to remember than numeric phone numbers; therefore when businesses use them as a direct response tool in their advertising (radio, television, print, outdoor, etc.), they are proven to increase response rates by 30–60%.
Regular side by side testing of phonewords and phone numbers in TV and radio advertising in Australia has shown that phonewords generate up to twice as many calls as standard phone numbers. A study conducted by Roy Morgan Research in February 2006 indicated that 92% of Australians were aware of alphanumeric dialling.
Disadvantages
In the age of advanced mobile devices, there is some disadvantage in using phonewords. Devices with physical keyboards such as BlackBerry and some other smartphones do not have the apportioned letters on the keys used |
https://en.wikipedia.org/wiki/Type%20punning | In computer science, a type punning is any programming technique that subverts or circumvents the type system of a programming language in order to achieve an effect that would be difficult or impossible to achieve within the bounds of the formal language.
In C and C++, constructs such as pointer type conversion and union — C++ adds reference type conversion and reinterpret_cast to this list — are provided in order to permit many kinds of type punning, although some kinds are not actually supported by the standard language.
In the Pascal programming language, the use of records with variants may be used to treat a particular data type in more than one manner, or in a manner not normally permitted.
Sockets example
One classic example of type punning is found in the Berkeley sockets interface. The function to bind an opened but uninitialized socket to an IP address is declared as follows:
int bind(int sockfd, struct sockaddr *my_addr, socklen_t addrlen);
The bind function is usually called as follows:
struct sockaddr_in sa = {0};
int sockfd = ...;
sa.sin_family = AF_INET;
sa.sin_port = htons(port);
bind(sockfd, (struct sockaddr *)&sa, sizeof sa);
The Berkeley sockets library fundamentally relies on the fact that in C, a pointer to struct sockaddr_in is freely convertible to a pointer to struct sockaddr; and, in addition, that the two structure types share the same memory layout. Therefore, a reference to the structure field my_addr->sin_family (where my_addr is of type struct sockaddr*) will actually refer to the field sa.sin_family (where sa is of type struct sockaddr_in). In other words, the sockets library uses type punning to implement a rudimentary form of polymorphism or inheritance.
Often seen in the programming world is the use of "padded" data structures to allow for the storage of different kinds of values in what is effectively the same storage space. This is often seen when two structures are used in mutual exclusivity for optimization.
Floating- |
https://en.wikipedia.org/wiki/Single-user%20mode | Single-user mode is a mode in which a multiuser computer operating system boots into a single superuser. It is mainly used for maintenance of multi-user environments such as network servers. Some tasks may require exclusive access to shared resources, for example running fsck on a network share. This mode can also be used for security purposes network services are not run, eliminating the possibility of outside interference. On some systems a lost superuser password can be changed by switching to single-user mode, but not asking for the password in such circumstances is viewed as a security vulnerability.
Unix family
Unix-like operating systems provide single-user mode functionality either through the System V-style runlevels, BSD-style boot-loader options, or other boot-time options.
The run-level is usually changed using the init command, runlevel 1 or S will boot into single-user mode.
Boot-loader options can be changed during startup before the execution of the kernel.
In FreeBSD and DragonFly BSD it can be changed before rebooting the system with the command nextboot -o "-s" -k kernel, and its bootloader offers the option on bootup to start in single-user mode. In Solaris the command reboot -- -s will cause a reboot into single-user mode.
macOS users can accomplish this by holding down after powering the system. The user may be required to enter a password set in the firmware. In OS X El Capitan and later releases of macOS, the mode can be reversed to single-user mode with the command sudo launchctl reboot userspace -s in Terminal, and the system can be fully rebooted in single-user mode with the command sudo launchctl reboot system -s. Single-user mode is different from a safe mode boot in that the system goes directly to the console instead of starting up the core elements of macOS (items in /System/Library/, ignoring /Library/, ~/Library/, et al.). From there users are encouraged by a prompt to run fsck or other command line utilities as needed (o |
https://en.wikipedia.org/wiki/Breakthrough%20infection | A breakthrough infection is a case of illness in which a vaccinated individual becomes infected with the illness, because the vaccine has failed to provide complete immunity against the pathogen (currently only viruses). Breakthrough infections have been identified in individuals immunized against a variety of diseases including mumps, varicella (Chickenpox), influenza, and COVID-19. The characteristics of the breakthrough infection are dependent on the virus itself. Often, infection of the vaccinated individual results in milder symptoms and shorter duration than if the infection were contracted naturally.
Causes of breakthrough infections include biological factors in the recipient, improper administration or storage of vaccines, mutations in viruses, blocking antibody formation, and other factors. For these reasons, vaccines are rarely 100% effective. A 2021 study found the common flu vaccine provided immunity to the flu in 58% of recipients. The measles vaccine fails to provide immunity to 2% of children that receive the vaccine. However, if herd immunity exists, it typically prevents individuals who are ineffectively vaccinated from contracting the disease. Accordingly, herd immunity reduces the number of breakthrough infections in a population.
By disease
Varicella
The varicella vaccine is 85% effective at preventing varicella (chickenpox) infection. However, 75% of individuals that are diagnosed with breakthrough varicella exhibit milder symptoms than individuals that are not vaccinated. These individuals with mild varicella have low fevers, fewer than 50 lesions on their skin, and a maculopapular rash. In contrast, unvaccinated individuals typically have a fever of 102, 200-500 skin lesions, and macules (lesions that are not elevated) evolve to papules and vesicular lesions. Additionally, infection in unvaccinated individuals tends to last for a longer period of time than in individuals who have been vaccinated.
The majority of cases of breakthrough v |
https://en.wikipedia.org/wiki/Riemann%E2%80%93von%20Mangoldt%20formula | In mathematics, the Riemann–von Mangoldt formula, named for Bernhard Riemann and Hans Carl Friedrich von Mangoldt, describes the distribution of the zeros of the Riemann zeta function.
The formula states that the number N(T) of zeros of the zeta function with imaginary part greater than 0 and less than or equal to T satisfies
The formula was stated by Riemann in his notable paper "On the Number of Primes Less Than a Given Magnitude" (1859) and was finally proved by Mangoldt in 1905.
Backlund gives an explicit form of the error for all T > 2:
Under the Lindelöf and Riemann hypotheses the error term can be improved to and respectively.
Similarly, for any primitive Dirichlet character χ modulo q, we have
where N(T,χ) denotes the number of zeros of L(s,χ) with imaginary part between -T and T.
Notes |
https://en.wikipedia.org/wiki/Common%20iliac%20lymph%20nodes | The common iliac lymph nodes, four to six in number, are grouped behind and on the sides of the common iliac artery, one or two being placed below the bifurcation of the aorta, in front of the fifth lumbar vertebra.
They drain chiefly the hypogastric and external iliac glands, and their efferents pass to the lateral aortic glands. |
https://en.wikipedia.org/wiki/Lumbar%20lymph%20trunk | The lumbar trunks are formed by the union of the efferent vessels from the lateral aortic lymph nodes.
They receive the lymph from the lower limbs, from the walls and viscera of the pelvis, from the kidneys and suprarenal glands and the deep lymphatics of the greater part of the abdominal wall.
Ultimately, the lumbar trunks empty into the cisterna chyli, a dilatation at the beginning of the thoracic duct. |
https://en.wikipedia.org/wiki/Gemmatimonadota | The Gemmatimonadota are a phylum of bacteria established in 2003. The phylum contains two classes Gemmatimonadetes and Longimicrobia.
Species
The type species Gemmatimonas aurantiaca strain T-27T was isolated from activated sludge in a sewage treatment system in 2003. It is a Gram-negative bacterium able to grow by both aerobic and anaerobic respiration.
The second cultured species was Gemmatirosa kalamazoonensis gen. nov., sp. nov. strain KBS708, which was isolated from organically managed agricultural soil in Michigan USA.
The third cultured species Gemmatimonas phototrophica strain AP64T was isolated from a shallow freshwater desert lake Tiān é hú (Swan Lake) in North China. A unique feature of this organism is the presence of bacterial photosynthetic reaction centers. It probably acquired genes for anoxygenic photosynthesis via horizontal gene transfer. G. phototrophica is a facultative photoheterotrophic organism. It requires the supply of organic substrate for growth, but it may obtain additional energy for its metabolism from light.
Longimicrobium terrae strain CB-286315T was isolated from a soil sample from a typical Mediterranean forest ecosystem located in Granada, Spain. Due to this large phylogenetic distance from other cultured Gemmatimonades, it established a novel class named Longimicrobia.
Environmental distribution
Data from culture-independent studies indicate that Gemmatimonadota are widely distributed in many natural habitats. They make up about 2% of soil bacterial communities and has been identified as one of the top nine phyla found in soils; yet, there are currently only six cultured isolates. Gemmatimonadota have been found in a variety of arid soils, such as grassland, prairie, and pasture soil, as well as eutrophic lake sediments and alpine soils. This wide range of environments where Gemmatimonadota have been found suggests an adaptation to low soil moisture. A study conducted showed that the distribution of the Gemmatimonadota in |
https://en.wikipedia.org/wiki/Fundamental%20plane%20%28spherical%20coordinates%29 | The fundamental plane in a spherical coordinate system is a plane of reference that divides the sphere into two hemispheres. The geocentric latitude of a point is then the angle between the fundamental plane and the line joining the point to the centre of the sphere.
For a geographic coordinate system of the Earth, the fundamental plane is the Equator.
Astronomical coordinate systems have varying fundamental planes:
The horizontal coordinate system uses the observer's horizon.
The Besselian coordinate system uses Earth's terminator (day/night boundary). This is a Cartesian coordinate system (x, y, z).
The equatorial coordinate system uses the celestial equator.
The ecliptic coordinate system uses the ecliptic.
The galactic coordinate system uses the Milky Way's galactic equator.
See also
Plane of reference |
https://en.wikipedia.org/wiki/Potential%20support%20ratio | The potential support ratio (PSR) is the number of people age 15–64 per one older person aged 65 or older. This ratio describes the burden placed on the working population (unemployment and children are not considered in this measure) by the non-working elderly population.
As a population ages, the potential support ratio tends to fall. Between 1950 and 2009, the potential ratio declined from 12 to 9 potential workers per person aged 65 or over. By 2050, the potential support ratio is projected to drop further to reach 4 potential worker per older person. The reduction of potential support ratio has important implications for social security schemes, particularly for pay-as-you-go pension systems under which taxes on current workers pay the pensions of retirees.
In 2015, Japan has the lowest PSR in the world, at 1.8.
See also
Dependency ratio |
https://en.wikipedia.org/wiki/Galaxy%20effective%20radius | Galaxy effective radius or half-light radius () is the radius at which half of the total light of a galaxy is emitted. This assumes the galaxy has either intrinsic spherical symmetry or is at least circularly symmetric as viewed in the plane of the sky. Alternatively, a half-light contour, or isophote, may be used for spherically and circularly asymmetric objects.
is an important length scale in term in de Vaucouleurs law, which characterizes a specific rate at which surface brightness decreases as a function of radius:
where is the surface brightness at . At ,
Thus, the central surface brightness is approximately .
See also |
https://en.wikipedia.org/wiki/Oct-1-en-3-one | Oct-1-en-3-one (CH2=CHC(=O)(CH2)4CH3), also known as 1-octen-3-one or amyl vinyl ketone, is the odorant that is responsible for the typical "metallic" smell of metals and blood coming into contact with skin. Oct-1-en-3-one has a strong metallic mushroom-like odor with an odor detection threshold of 0.03–1.12 µg/m3 and it is the main compound responsible for the "smell of metal", followed by decanal (smell: orange skin, flowery) and nonanal (smell: tallowy, fruity). Oct-1-en-3-one is the degradative reduction product of the chemical reaction of skin lipid peroxides and Fe2+. Skin lipid peroxides are formed from skin lipid by oxidation, either enzymatically by lipoxygenases or by air oxygen. Oct-1-en-3-one is a ketone analog of the alkene 1-octene.
Natural occurrences
It is also produced by Uncinula necator, a fungus that causes powdery mildew of grape.
See also
Odorant
1-Octen-3-ol, the alcohol analog that is used by mosquitoes as an odor cue |
https://en.wikipedia.org/wiki/3D%20ultrasound | 3D ultrasound is a medical ultrasound technique, often used in fetal, cardiac, trans-rectal and intra-vascular applications. 3D ultrasound refers specifically to the volume rendering of ultrasound data. When involving a series of 3D volumes collected over time, it can also be referred to as 4D ultrasound (three spatial dimensions plus one time dimension) or real-time 3D ultrasound.
Methods
When generating a 3D volume, the ultrasound data can be collected in four common ways by a sonographer:
Freehand, which involves tilting the probe and capturing a series of ultrasound images and recording the transducer orientation for each slice.
Mechanically, where the internal linear probe tilt is handled by a motor inside the probe.
Using an endoprobe, which generates the volume by inserting a probe and then removing the transducer in a controlled manner.
A matrix array transducer, which uses beam steering to sample points throughout a pyramid shaped volume.
Risks
The general risks of ultrasound also apply to 3D ultrasound. Essentially, ultrasound is considered safe. While other imaging modalities use radioactive dye or ionizing radiation, for example, ultrasound transducers send pulses of high frequency sound into the body and then listen for the echo.
In summary, the primary risks associated with ultrasound would be the potential heating of tissue or cavitation. The mechanisms by which tissue heating and cavitation are measured are through the standards called thermal index (TI) and mechanical index (MI). Even though the FDA outlines very safe values for maximum TI and MI, it is still recommended to avoid unnecessary ultrasound imaging.
Applications
Obstetrics
3D ultrasound is useful, among other things, for facilitating the characterization of some congenital defects, such as skeletal anomalies and heart issues. With real-time 3D ultrasound, the fetal heart rate can be examined in real-time.
Cardiology
Applications of three-dimensional ultrasound in cardiac treatment |
https://en.wikipedia.org/wiki/Nonlinear%20X-wave | In physics, a nonlinear X-wave (NLX) is a multi-dimensional wave that can travel without distortion.
At variance with X-waves, a nonlinear X-wave does exist in the presence of nonlinearity, and in many cases it self-generates from a Gaussian (in any direction) wave packet.
The distinctive feature of an NLX is its "biconical" shape, (see figure) which appears as an "X" in any section plane containing the wave peak and the direction of propagation.
So far, nonlinear X-waves have been only observed in nonlinear optics experiments, and have been predicted to occur in a variety of nonlinear media including Bose–Einstein condensates.
History
Preliminary experimental results were reported CLEO/QELS conference in 2001
The first article was published in Physical Review Letters in 2003 and reported on the theoretical prediction of the existence of nonlinear X-waves.
The first experimental results also appeared in Physical Review Letters in 2003. |
https://en.wikipedia.org/wiki/Haganeyama%20Transmitter | Haganeyama Transmitter (はがね山標準電波送信所, ) is an LF-time signal transmitter at Fuji-cho, Saga-city, Saga-ken, Japan used for transmitting the time signal JJY on 60 kHz. The Haganeyama site is one of two JJY transmitters, another is the Otakadoyama site.
Summy
NAME: NICT Haganeyama LF station
Location: Summit of Mt. Hagane, Fuji-cho, Saga-city, Saga-ken
Elevation: about 900m
Latitude: 33°27'56.0"N
Longitude: 130°10'32.0"E
License: NICT
Station purpose: Transmitting the official Japanese government frequency standards and time signal
Frequency form: 250H A1B
Frequency: 60kHz
Antenna power: 50kW (Antenna efficiency: about 45%)
Antenna form: Umbrella type 200m high
Operation time: continuously
Operation start: 2001/10/01
Range: About 1,000 km
Transmission method :
See also
Otakadoyayama Transmitter |
https://en.wikipedia.org/wiki/Numbertime | Numbertime is a BBC educational numeracy television series for primary schools that was aired on BBC Two from 20 September 1993 to 3 December 2001. For its first four series, it was presented by Lolita Chakrabarti. El Nombre, an animated character used throughout the series, eventually became the concept for his own educational BBC children's television program; his name means "The Name" in Spanish, and not "The Number", which would be "El Número". The third line of his opening song and his farewell catchphrase were also changed several times during the series' run, to reflect their focus - however, the original ones ("Writing numbers in the desert sand" which was also used for the seventh series, and "Adios amigos, and keep counting" which was also used for the fourth, sixth, seventh, eighth and ninth series) remain the most famous.
For the second series, El Nombre's tagline and farewell catchphrase were changed to "Drawing shapes in the desert sand" and "Adios amigos, and keep shaping up" respectively, while for the third series, they were changed to "Righting wrongs in the desert sand" and "Adios amigos, over and out" respectively; however, for the fourth series, his tagline was changed to "Counting numbers in the desert sand" (which was also used for the ninth series), and for the third episode of the fourth series, his farewell catchphrase was changed to "Adios amigos, and fetch some water". For the fifth series, both his tagline and farewell catchphrase were changed to "Telling time in the desert sand" and "Adios amigos, 'till the next time" respectively, while for the sixth series, his tagline was changed to "Using numbers in the desert sand"; finally, for the eighth series, his tagline was changed to "Counting money in the desert sand".
Series 1: Numbers 1 to 10 (Autumn 1993)
The first series, which is aimed at four- to five-year-olds, comprised ten episodes focusing on the numbers 1-10, in order; each episode opened with Lolita introducing herself to the |
https://en.wikipedia.org/wiki/Current%20research%20information%20system | A current research information system (CRIS) is a database or other information system to store, manage and exchange contextual metadata for the research activity funded by a research funder or conducted at a research-performing organisation (or aggregation thereof).
CRIS systems are also known as Research Information Management or RIM Systems (RIMS).
Features
The data model underpinning a CRIS relies on a set of basic entities as defined by the Common European Research Information Format (CERIF ) model maintained by the non-profit organisation euroCRIS.
The links connecting these entities provide a standardised semantic layer that provides consistency to the data model. The basic CERIF entities are people, organisations, projects and outputs (publications, research data, patents). Further second-level entities in the comprehensive snapshot of research provided by CERIF are for instance funding, research facilities and equipment or skills.
System interoperability lies at the core of CRIS operation, both from an internal and an external viewpoint. Internally, information is exchanged between the multiple information-gathering systems at institutions (HR systems, project management tools, finance management systems, etc.) and the one-stop-shop CRIS where all the institutional research information is kept. From an external interoperability perspective, metadata need to be exchanged between the systems at research-performing organisations where the research is actually conducted and the systems run by research funders and governmental bodies in charge of research assessment processes. By providing a standard approach to information description, the CERIF model becomes a key feature for enabling this system interoperability.
A particularly important area of system interoperability is CRIS/IR interoperability, i.e. the information exchange workflows between Current Research Information Systems and Institutional Repositories. While these two kinds of systems were on |
https://en.wikipedia.org/wiki/EuroCRIS | euroCRIS is an international not-for-profit association founded in 2002 in order to bring together experts on research information management and research information management systems (CRIS). The euroCRIS Office is located in Nijmegen (Netherlands).
euroCRIS maintains the CERIF standard (for Common European Research Information Format) to enable CRIS system interoperability. CERIF is endorsed by the European Commission and is developed and maintained by the CERIF Task Group of euroCRIS.
The mission of euroCRIS is to foster cooperation and knowledge-sharing across the worldwide research information community and to promote interoperability of research information through the CERIF standard. Additional areas of activity also include – among others – the uptake of CRIS systems by various stakeholders, research information infrastructures on an institutional, regional, national and international level, best practices in system interoperability and the use and implementation of standards in CRIS such as identifiers, formats, semantics, (controlled) vocabularies, etc.
A key instrument for the community-driven information exchange are the international events regularly organised by euroCRIS. These include the biennial CRIS Conferences and the biannual Membership Meetings. The outputs arising from all these events are systematically archived in the euroCRIS open access repository based on a DSpace-CRIS software platform. The most recent euroCRIS event after a two-and-a-half-year hiatus due to the Covid-19 pandemic was the CRIS2022 Conference held May 12-14, 2022 in Dubrovnik (Croatia).
The euroCRIS DRIS
One of the main areas of activity of euroCRIS involves maintaining the Directory of Research Information Systems (DRIS). The DRIS is an international directory of CRIS systems currently displaying (as of June 2022) close to 1,200 entries. The vast majority of DRIS records describe institutional CRISs, but there is also a good number of instances for regional and nati |
https://en.wikipedia.org/wiki/Preferential%20looking | Preferential looking is an experimental method in developmental psychology used to gain insight into the young mind/brain. The method as used today was developed by the developmental psychologist Robert L. Fantz in the 1960s.
The Preferential Looking Technique
According to the American Psychological Association, the preferential looking technique is "an experimental method for assessing the perceptual capabilities of nonverbal individuals (e.g., human infants, nonhuman animals)". If the average infant looks longer at a novel stimulus compared to a familiar stimulus, this suggests that the infant can discriminate between the stimuli. This method has been used extensively in cognitive science and developmental psychology to assess the character of infants' perceptual systems, and, by extension, innate cognitive faculties. An investigator or examiner observes an infant's eye movements to determine which stimulus the infant fixates on.
Robert L. Fantz
Robert L. Fantz (1925-1981) was a developmental psychologist who launched several studies on infant perception including the preferential looking paradigm. Fantz introduced this paradigm in 1961 while working at the Case Western Reserve University. The preferential looking paradigm is used in studies of infants regarding cognitive development and categorization. Fantz's study showed that infants looked at patterned images longer than uniform images. He later built upon his study in 1964 to include habituation situations. These situations exhibited an infants preference for new or unusual stimuli.
Summary of Findings
Conclusions have been drawn from preferential looking experiments about the knowledge that infants possess. For example, if infants discriminate between rule-following and rule-violating stimuli—say, by looking longer, on average, at the latter than the former—then it has sometimes been concluded that infants know the rule.
Here is an example: 100 infants are shown an object that appears to teleport, viol |
https://en.wikipedia.org/wiki/65%2C537 | 65537 is the integer after 65536 and before 65538.
In mathematics
65537 is the largest known prime number of the form (). Therefore, a regular polygon with 65537 sides is constructible with compass and unmarked straightedge. Johann Gustav Hermes gave the first explicit construction of this polygon. In number
theory, primes of this form are known as Fermat primes, named after the mathematician
Pierre de Fermat. The only known prime Fermat numbers are
In 1732, Leonhard Euler found that the next Fermat number is composite:
In 1880, showed that
65537 is also the 17th Jacobsthal–Lucas number, and currently the largest known integer n for which the number is a probable prime.
Applications
65537 is commonly used as a public exponent in the RSA cryptosystem. Because it is the Fermat number with , the common shorthand is "F" or "F4". This value was used in RSA mainly for historical reasons; early raw RSA implementations (without proper padding) were vulnerable to very small exponents, while use of high exponents was computationally expensive with no advantage to security (assuming proper padding).
65537 is also used as the modulus in some Lehmer random number generators, such as the one used by ZX Spectrum, which ensures that any seed value will be coprime to it (vital to ensure the maximum period) while also allowing efficient reduction by the modulus using a bit shift and subtract. |
https://en.wikipedia.org/wiki/Covariant%20return%20type | In object-oriented programming, a covariant return type of a method is one that can be replaced by a "narrower" type when the method is overridden in a subclass. A notable language in which this is a fairly common paradigm is C++.
C# supports return type covariance as of version 9.0. Covariant return types have been (partially) allowed in the Java language since the release of JDK5.0, so the following example wouldn't compile on a previous release:// Classes used as return types:
class A {
}
class B extends A {
}
// "Class B is narrower than class A"
// Classes demonstrating method overriding:
class C {
A getFoo() {
return new A();
}
}
class D extends C {
// Overriding getFoo() in parent class C
B getFoo() {
return new B();
}
}
More specifically, covariant (wide to narrower) or contravariant (narrow to wider) return type refers to a situation where the return type of the overriding method is changed to a type related to (but different from) the return type of the original overridden method. The relationship between the two covariant return types is usually one which allows substitution of the one type with the other, following the Liskov substitution principle. This usually implies that the return types of the overriding methods will be subtypes of the return type of the overridden method. The above example specifically illustrates such a case. If substitution is not allowed, the return type is invariant and causes a compile error.
Another example of covariance with the help of built in Object and String class of Java:
class Parent {
public Object getFoo() {
return null;
}
}
class Child extends Parent {
// String is child of the greater Object class
public String getFoo() {
return "This is a string";
}
// Driver code
public static void main(String[] args) {
Child child = new Child();
System.out.println(child.getFoo());
}
}
See also
Covariance and contrava |
https://en.wikipedia.org/wiki/Cloaca%20%28embryology%29 | The cloaca (: cloacae) is a structure in the development of the urinary and reproductive organs.
The hind-gut is at first prolonged backward into the body-stalk as the tube of the allantois; but, with the growth and flexure of the tail-end of the embryo, the body-stalk, with its contained allantoic tube, is carried forward to the ventral aspect of the body, and consequently a bend is formed at the junction of the hind-gut and allantois.
This bend becomes dilated into a pouch, which constitutes the endodermal cloaca; into its dorsal part the hind-gut opens, and from its ventral part the allantois passes forward.
At a later stage the Wolffian duct and Müllerian duct open into its ventral portion.
The cloaca is, for a time, shut off from the anterior by the cloacal membrane, formed by the apposition of the ectoderm and endoderm, and reaching, at first, as far forward as the future umbilicus.
Behind the umbilicus, however, the mesoderm subsequently extends to form the lower part of the abdominal wall and pubic symphysis.
By the growth of the surrounding tissues the cloacal membrane comes to lie at the bottom of a depression, which is lined by ectoderm and named the ectodermal cloaca.
Clinical significance
A birth defect can arise known as a persistent cloaca where the rectum, vagina, and urinary tract fuse to create a common channel or cloaca.
A rare birth defect which leaves much of the abdominal organs exposed is known as cloacal exstrophy.
Additional images |
https://en.wikipedia.org/wiki/Residual%20chemical%20shift%20anisotropy | Residual chemical shift anisotropy (RCSA) is the difference between the chemical shift anisotropy (CSA) of aligned and non-aligned molecules. It is normally three orders of magnitude smaller than the static CSA, with values on the order of parts-per-billion (ppb). RCSA is useful for structural determination and it is among the new developments in NMR spectroscopy.
See also
Residual dipolar coupling |
https://en.wikipedia.org/wiki/Circular%20points%20at%20infinity | In projective geometry, the circular points at infinity (also called cyclic points or isotropic points) are two special points at infinity in the complex projective plane that are contained in the complexification of every real circle.
Coordinates
A point of the complex projective plane may be described in terms of homogeneous coordinates, being a triple of complex numbers , where two triples describe the same point of the plane when the coordinates of one triple are the same as those of the other aside from being multiplied by the same nonzero factor. In this system, the points at infinity may be chosen as those whose z-coordinate is zero. The two circular points at infinity are two of these, usually taken to be those with homogeneous coordinates
and .
Trilinear coordinates
Let A. B. C be the measures of the vertex angles of the reference triangle ABC. Then the trilinear coordinates of the circular points at infinity in the plane of the reference triangle are as given below:
or, equivalently,
or, again equivalently,
where .
Complexified circles
A real circle, defined by its center point (x0,y0) and radius r (all three of which are real numbers) may be described as the set of real solutions to the equation
Converting this into a homogeneous equation and taking the set of all complex-number solutions gives the complexification of the circle. The two circular points have their name because they lie on the complexification of every real circle. More generally, both points satisfy the homogeneous equations of the type
The case where the coefficients are all real gives the equation of a general circle (of the real projective plane). In general, an algebraic curve that passes through these two points is called circular.
Additional properties
The circular points at infinity are the points at infinity of the isotropic lines.
They are invariant under translations and rotations of the plane.
The concept of angle can be defined using the circular points, nat |
https://en.wikipedia.org/wiki/Isotropic%20line | In the geometry of quadratic forms, an isotropic line or null line is a line for which the quadratic form applied to the displacement vector between any pair of its points is zero. An isotropic line occurs only with an isotropic quadratic form, and never with a definite quadratic form.
Using complex geometry, Edmond Laguerre first suggested the existence of two isotropic lines through the point that depend on the imaginary unit :
First system:
Second system:
Laguerre then interpreted these lines as geodesics:
An essential property of isotropic lines, and which can be used to define them, is the following: the distance between any two points of an isotropic line situated at a finite distance in the plane is zero. In other terms, these lines satisfy the differential equation . On an arbitrary surface one can study curves that satisfy this differential equation; these curves are the geodesic lines of the surface, and we also call them isotropic lines.
In the complex projective plane, points are represented by homogeneous coordinates and lines by homogeneous coordinates . An isotropic line in the complex projective plane satisfies the equation:
In terms of the affine subspace , an isotropic line through the origin is
In projective geometry, the isotropic lines are the ones passing through the circular points at infinity.
In the real orthogonal geometry of Emil Artin, isotropic lines occur in pairs:
A non-singular plane which contains an isotropic vector shall be called a hyperbolic plane. It can always be spanned by a pair N, M of vectors which satisfy
We shall call any such ordered pair N, M a hyperbolic pair. If V is a non-singular plane with orthogonal geometry and N ≠ 0 is an isotropic vector of V, then there exists precisely one M in V such that N, M is a hyperbolic pair. The vectors x N and y M are then the only isotropic vectors of V.
Relativity
Isotropic lines have been used in cosmological writing to carry light. For example, in a mathematical ency |
https://en.wikipedia.org/wiki/TSMP | TSMP, an acronym for Time Synchronized Mesh Protocol, was developed by Dust Networks as a communications protocol for self-organizing networks of wireless devices called motes. TSMP devices stay synchronized to each other and communicate in time-slots, similar to other TDM (time-division multiplexing) systems. Such deterministic communication allows the devices to stay extremely low power, as the radios only turn on for the periods of scheduled communication. The protocol is designed to operate very reliably in a noisy environment. It uses channel hopping to avoid interference -- the packets between TSMP devices get sent on different radio channels depending on time of transmission. TSMP distinguishes itself from other time-slotted mesh-based protocols, in that time-slot timing is maintained continuously and enables a network to duty-cycle on a transmitter-receiver pair-wise basis, as opposed to putting the entire network to sleep for extended periods of time (as is done in a beacon-based protocol, such as DigiMesh).
Dust Networks' underlying time synchronized mesh networking technology has been standardized by the HART Communications Foundation with the WirelessHART protocol, the International Society of Automation ISA100 standard and in internet protocol standards, such as IEEE802.15.4E MAC layer, and IETF 6TiSCh.
Time synchronized mesh networking is marketed for applications that require reliability and ultra long battery life, typically measured in years.
It is intended for the industrial market for manufacturing-process monitoring and control.
See also
Wireless sensor network
External links
TSMP Whitepaper (archived webpage as of October 2006
DigiMesh
Kristofer S. J. Pister and Lance Doherty, TSMP: Time Synchronized Mesh Protocol, Proceedings of the IASTED International Symposium on Distributed Sensor Networks (DSN08), November 2008, Orlando, Florida, USA.
Wireless sensor network |
https://en.wikipedia.org/wiki/Imaginary%20line%20%28mathematics%29 | In complex geometry, an imaginary line is a straight line that only contains one real point. It can be proven that this point is the intersection point with the conjugated line.
It is a special case of an imaginary curve.
An imaginary line is found in the complex projective plane P2(C) where points are represented by three homogeneous coordinates
Boyd Patterson described the lines in this plane:
The locus of points whose coordinates satisfy a homogeneous linear equation with complex coefficients
is a straight line and the line is real or imaginary according as the coefficients of its equation are or are not proportional to three real numbers.
Felix Klein described imaginary geometrical structures: "We will characterize a geometric structure as imaginary if its coordinates are not all real.:
According to Hatton:
The locus of the double points (imaginary) of the overlapping involutions in which an overlapping involution pencil (real) is cut by real transversals is a pair of imaginary straight lines.
Hatton continues,
Hence it follows that an imaginary straight line is determined by an imaginary point, which is a double point of an involution, and a real point, the vertex of the involution pencil.
See also
Conic section
Imaginary number
Imaginary point
Real curve |
https://en.wikipedia.org/wiki/Real%20point | In geometry, a real point is a point in the complex projective plane with homogeneous coordinates for which there exists a nonzero complex number such that , , and are all real numbers.
This definition can be widened to a complex projective space of arbitrary finite dimension as follows:
are the homogeneous coordinates of a real point if there exists a nonzero complex number such that the coordinates of
are all real.
A point which is not real is called an imaginary point.
Context
Geometries that are specializations of real projective geometry, such as Euclidean geometry, elliptic geometry or conformal geometry may be complexified, thus embedding the points of the geometry in a complex projective space, but retaining the identity of the original real space as special. Lines, planes etc. are expanded to the lines, etc. of the complex projective space. As with the inclusion of points at infinity and complexification of real polynomials, this allows some theorems to be stated more simply without exceptions and for a more regular algebraic analysis of the geometry.
Viewed in terms of homogeneous coordinates, a real vector space of homogeneous coordinates of the original geometry is complexified. A point of the original geometric space is defined by an equivalence class of homogeneous vectors of the form , where is an nonzero complex value and is a real vector. A point of this form (and hence belongs to the original real space) is called a real point, whereas a point that has been added through the complexification and thus does not have this form is called an imaginary point.
Real subspace
A subspace of a projective space is real if it is spanned by real points.
Every imaginary point belongs to exactly one real line, the line through the point and its complex conjugate. |
https://en.wikipedia.org/wiki/Complex%20conjugate%20line | In complex geometry, the complex conjugate line of a straight line is the line that it becomes by taking the complex conjugate of each point on this line.
This is the same as taking the complex conjugates of the coefficients of the line. So if the equation of is , then the equation of its conjugate is .
The conjugate of a real line is the line itself.
The intersection point of two conjugated lines is always real. |
https://en.wikipedia.org/wiki/Quellung%20reaction | The quellung reaction, also called the Neufeld reaction, is a biochemical reaction in which antibodies bind to the bacterial capsule of Streptococcus pneumoniae, Klebsiella pneumoniae, Neisseria meningitidis, Bacillus anthracis, Haemophilus influenzae, Escherichia coli, and Salmonella. The antibody reaction allows these species to be visualized under a microscope. If the reaction is positive, the capsule becomes opaque and appears to enlarge.
Quellung is the German word for "swelling" and describes the microscopic appearance of pneumococcal or other bacterial capsules after their polysaccharide antigen has combined with a specific antibody. The antibody usually comes from serum taken from an immunized laboratory animal. As a result of this combination, and precipitation of the large, complex molecule formed, the capsule appears to swell, because of increased surface tension, and its outlines become demarcated.
The pneumococcal quellung reaction was first described in 1902 by the scientist Fred Neufeld, and applied only to Streptococcus pneumoniae, both as microscopic capsular swelling and macroscopic agglutination (clumping visible with the naked eye). It was initially an intellectual curiosity more than anything else, and could distinguish only the three pneumococcal serotypes known at that time. However, it acquired an important practical use with the advent of serum therapy to treat certain types of pneumococcal pneumonia in the 1920s because selection of the proper antiserum to treat an individual patient required correct identification of the infecting pneumococcal serotype, and the quellung reaction was the only method available to do this. Dr. Albert Sabin made modifications to Neufeld's technique so that it could be done more rapidly, and other scientists expanded the technique to identify 29 additional serotypes.
Application of Neufeld’s discoveries to other important areas of research came when Fred Griffith showed that pneumococci could transfer inform |
https://en.wikipedia.org/wiki/Mental%20mapping | In behavioral geography, a mental map is a person's point-of-view perception of their area of interaction. Although this kind of subject matter would seem most likely to be studied by fields in the social sciences, this particular subject is most often studied by modern-day geographers. They study it to determine subjective qualities from the public such as personal preference and practical uses of geography like driving directions.
Mass media also have a virtually direct effect on a person's mental map of the geographical world. The perceived geographical dimensions of a foreign nation (relative to one's own nation) may often be heavily influenced by the amount of time and relative news coverage that the news media may spend covering news events from that foreign region. For instance, a person might perceive a small island to be nearly the size of a continent, merely based on the amount of news coverage that he or she is exposed to on a regular basis.
In psychology, the term names the information maintained in the mind of an organism by means of which it may plan activities, select routes over previously traveled territories, etc. The rapid traversal of a familiar maze depends on this kind of mental map if scents or other markers laid down by the subject are eliminated before the maze is re-run.
Background
Mental maps are an outcome of the field of behavioral geography. The imagined maps are considered one of the first studies that intersected geographical settings with human action. The most prominent contribution and study of mental maps was in the writings of Kevin Lynch. In The Image of the City, Lynch used simple sketches of maps created from memory of an urban area to reveal five elements of the city; nodes, edges, districts, paths and landmarks. Lynch claimed that “Most often our perception of the city is not sustained, but rather partial, fragmentary, mixed with other concerns. Nearly every sense is in operation, and the image is the composite of them |
https://en.wikipedia.org/wiki/Policeman%20%28laboratory%29 | A policeman is a hand-held flexible natural-rubber or plastic scraper. The common type of it is attached to a glass rod and used in chemical laboratories to transfer residues of precipitate or solid on glass surfaces when performing gravimetric analysis. This equipment works well under gentle, delicate and precise requirement. A policeman also comes in various sizes, shapes, and types. Some of them come in one-piece flexible plastic version and some in stainless. The origin of the policeman and its name cannot be identified for sure but some clues led back to the 19th century from German chemist Carl Remigius Fresenius.
Structure
A policeman is generally a flexible natural-rubber blade attached to a glass rod, which is typically 5 mm to 6 mm diameter and 150 mm long. However, it also comes in various sizes and shapes depending on its uses. The rubber material provides chemical resistance. In some designs, there is no glass rod, but instead the whole item is made of plastic or stainless steel and is shaped into a spatula or scraper shape at the end.
Uses
A policeman can be used for cleaning the inside of glassware, or for getting the last bit of precipitate out of a vessel. Especially in chemical laboratories it is often used to transfer residues of precipitate or solid on glass surfaces when performing the gravimetric analysis. It also used in biological laboratories, to transfer tissue culture cells from a plate to a suspension. It feature is to prevent the glass rod from scratching or breaking glassware.
Origin
There is no answer on where the name "policeman" comes from, though it may be related to the function of the instrument.
It is like the police in that it protects the beaker from scratching.
It is like the police in that it gathers up any stray or escaped particles of precipitate on the beaker wall.
In chemistry, gravimetric analysis is essential. After precipitating the chemical element of interest, successfully transferring all of the precipitate |
https://en.wikipedia.org/wiki/Molecular%20cellular%20cognition | Molecular cellular cognition (MCC) is a branch of neuroscience that involves the study of cognitive processes with approaches that integrate molecular, cellular and behavioral mechanisms. Key goals of MCC studies include the derivation of molecular and cellular explanations of cognitive processes, as well as finding mechanisms and treatments for cognitive disorders.
Although closely connected with behavioral genetics, MCC emphasizes the integration of molecular and cellular explanations of behavior, instead of focusing on the connections between genes and behavior.
Unlike cognitive neuroscience, which historically has focused on the connection between human brain systems and behavior, the field of MCC has used model organisms, such as mice, to study how molecular (i.e. receptor, kinase activation, phosphatase regulation), intra-cellular (i.e. dendritic processes), and inter-cellular processes (i.e. synaptic plasticity; network representations such as place fields) modulate cognitive function.
Methods employed in MCC include (but are not limited to) transgenic organisms (i.e. mice), viral vectors, pharmacology, in vitro and in vivo electrophysiology, optogenetics, in vivo imaging, and behavioral analysis. Modeling has become an essential component of the field because of the complexity of the multilevel data generated.
Scientific roots
The field of MCC has its roots in the pioneering pharmacological studies of the role of NMDA receptor in long-term potentiation and spatial learning and in studies that used knockout mice to look at the role of the alpha calcium calmodulin kinase II and FYN kinase in hippocampal long-term potentiation and spatial learning. The field has since expanded to include a large array of molecules including CREB.
Foundation of the field
MCC became an organized field with the formation of the Molecular Cellular Cognition Society, an organization with no membership fees and meetings that emphasize the participation of junior scientists. Its |
https://en.wikipedia.org/wiki/Prostatic%20ducts | The prostatic ducts (or prostatic ductules) open into the floor of the prostatic portion of the urethra, and are lined by two layers of epithelium, the inner layer consisting of columnar and the outer of small cubical cells.
Small colloid masses, known as amyloid bodies are often found in the gland tubes.
They open onto the prostatic sinus.
See also
Prostate |
https://en.wikipedia.org/wiki/Excitation%20temperature | In statistical mechanics, the excitation temperature () is defined for a population of particles via the Boltzmann factor. It satisfies
where
is the number of particles in an upper (e.g. excited) state;
is the statistical weight of those upper-state particles;
is the number of particles in a lower (e.g. ground) state;
is the statistical weight of those lower-state particles;
is the exponential function;
is the Boltzmann constant;
is the difference in energy between the upper and lower states.
Thus the excitation temperature is the temperature at which we would expect to find a system with this ratio of level populations. However it has no actual physical meaning except when in local thermodynamic equilibrium. The excitation temperature can even be negative for a system with inverted levels (such as a maser).
In observations of the 21 cm line of hydrogen, the apparent value of the excitation temperature is often called the "spin temperature". |
https://en.wikipedia.org/wiki/Heavy-tailed%20distribution | In probability theory, heavy-tailed distributions are probability distributions whose tails are not exponentially bounded: that is, they have heavier tails than the exponential distribution. In many applications it is the right tail of the distribution that is of interest, but a distribution may have a heavy left tail, or both tails may be heavy.
There are three important subclasses of heavy-tailed distributions: the fat-tailed distributions, the long-tailed distributions, and the subexponential distributions. In practice, all commonly used heavy-tailed distributions belong to the subexponential class, introduced by Jozef Teugels.
There is still some discrepancy over the use of the term heavy-tailed. There are two other definitions in use. Some authors use the term to refer to those distributions which do not have all their power moments finite; and some others to those distributions that do not have a finite variance. The definition given in this article is the most general in use, and includes all distributions encompassed by the alternative definitions, as well as those distributions such as log-normal that possess all their power moments, yet which are generally considered to be heavy-tailed. (Occasionally, heavy-tailed is used for any distribution that has heavier tails than the normal distribution.)
Definitions
Definition of heavy-tailed distribution
The distribution of a random variable X with distribution function F is said to have a heavy (right) tail if the moment generating function of X, MX(t), is infinite for all t > 0.
That means
This is also written in terms of the tail distribution function
as
Definition of long-tailed distribution
The distribution of a random variable X with distribution function F is said to have a long right tail if for all t > 0,
or equivalently
This has the intuitive interpretation for a right-tailed long-tailed distributed quantity that if the long-tailed quantity exceeds some high level, the probability |
https://en.wikipedia.org/wiki/Euclidean%20distance%20matrix | In mathematics, a Euclidean distance matrix is an matrix representing the spacing of a set of points in Euclidean space.
For points in -dimensional space , the elements of their Euclidean distance matrix are given by squares of distances between them.
That is
where denotes the Euclidean norm on .
In the context of (not necessarily Euclidean) distance matrices, the entries are usually defined directly as distances, not their squares.
However, in the Euclidean case, squares of distances are used to avoid computing square roots and to simplify relevant theorems and algorithms.
Euclidean distance matrices are closely related to Gram matrices (matrices of dot products, describing norms of vectors and angles between them).
The latter are easily analyzed using methods of linear algebra.
This allows to characterize Euclidean distance matrices and recover the points that realize it.
A realization, if it exists, is unique up to rigid transformations, i.e. distance-preserving transformations of Euclidean space (rotations, reflections, translations).
In practical applications, distances are noisy measurements or come from arbitrary dissimilarity estimates (not necessarily metric).
The goal may be to visualize such data by points in Euclidean space whose distance matrix approximates a given dissimilarity matrix as well as possible — this is known as multidimensional scaling.
Alternatively, given two sets of data already represented by points in Euclidean space, one may ask how similar they are in shape, that is, how closely can they be related by a distance-preserving transformation — this is Procrustes analysis.
Some of the distances may also be missing or come unlabelled (as an unordered set or multiset instead of a matrix), leading to more complex algorithmic tasks, such as the graph realization problem or the turnpike problem (for points on a line).
Properties
By the fact that Euclidean distance is a metric, the matrix has the following properties.
All element |
https://en.wikipedia.org/wiki/XML%20validation | XML validation is the process of checking a document written in XML (eXtensible Markup Language) to confirm that it is both well-formed and also "valid" in that it follows a defined structure. A well-formed document follows the basic syntactic rules of XML, which are the same for all XML documents. A valid document also respects the rules dictated by a particular DTD or XML schema. Automated tools – validators – can perform well-formedness tests and many other validation tests, but not those that require human judgement, such as correct application of a schema to a data set.
Standards
OASIS CAM is a standard specification that provides contextual validation of content and structure that is more flexible than basic schema validations.
Schematron, a method for advanced XML validation.
Tools
xmllint is a command line XML tool that can perform XML validation. It can be found in UNIX / Linux environments.
XML toolkit. The XML C parser and toolkit of Gnome – libxml includes xmllint
XML Validator Online Validate your XML data.
XML Schema Validator Validate XML files against an XML Schema. |
https://en.wikipedia.org/wiki/Compton%20edge | In spectrophotometry, the Compton edge is a feature of the spectrograph that results from Compton scattering in the scintillator or detector. When a gamma-ray scatters off the scintillator but escapes, only some fraction of its energy is registered by the detector. The amount of energy deposited in the detector depends on the scattering angle of the photon, leading to a spectrum of energies each corresponding to a different scattering angle. The highest energy that can be deposited, corresponding to full backscatter, is called the Compton edge. In mathematical terms, the Compton edge is the inflection point of the high-energy side of the Compton region.
Background
In a Compton scattering process, an incident photon collides with an electron in a material. The amount of energy exchanged varies with angle, and is given by the formula:
or
E is the energy of the incident photon.
E' is the energy of the outgoing photon, which escapes the material.
is the mass of the electron.
c is the speed of light.
is the angle of deflection for the photon.
The amount of energy transferred to the material varies with the angle of deflection. As approaches zero, none of the energy is transferred. The maximum amount of energy is transferred when approaches 180 degrees.
It is impossible for the photon to transfer any more energy via this process; thus, there is a sharp cutoff at this energy, leading to the name Compton edge. If an isotope has multiple photopeaks, each inflection point will have its own Compton edge.
The region between zero energy transfer and the Compton edge is known as the Compton continuum. |
https://en.wikipedia.org/wiki/Honey%20Bee%20Genome%20Sequencing%20Consortium | The Honey Bee Genome Sequencing Consortium is an international collaborative group of genomics scientists, scientific organisations and universities trying to decipher the genome sequences of the honey bee (Apis mellifera). It was formed in 2001 by American scientists. In the US, the project is funded by the National Human Genome Research Institute (a division of the National Institutes of Health (NIH)), the United States Department of Agriculture (USDA), the Texas Agricultural Experiment Station, the University of Illinois Sociogenomics Initiative, and various beekeepers association and the bee industry.
First scientific findings show that the honey bee genome may have evolved more slowly than the genomes of the fruit fly and malaria mosquito. The bee genome contains versions of some important mammalian genes.
The complete genome of Apis mellifera has been sequenced and consists of 10,000 genes with approximately 236 million base pairs. The size of the genome is a tenth of the human genome.
The Western honey bee gene sequence showed 163 chemical receptors for smell but only 10 for taste. Besides the discovery of new genes for the use of pollen and nectar, researchers found that, in comparison with other insects, Apis mellifera has fewer genes for immunity, detoxification and the development of the cuticula.
The population genetic analysis showed Africa as the origin and hypothesized that the spread into Europe happened in at least two independent waves.
Data from the scientific collaboration was made available on BeeBase led by Texas A&M University.
BeeSpace led by the University of Illinois is an effort to complete a web navigable catalog of related information.
See also
List of sequenced eukaryotic genomes
Sources
Beekeeping organizations
Genomics
Genome projects
Beekeeping in the United States
Agricultural organizations based in the United States |
https://en.wikipedia.org/wiki/Hollow%20matrix | In mathematics, a hollow matrix may refer to one of several related classes of matrix: a sparse matrix; a matrix with a large block of zeroes; or a matrix with diagonal entries all zero.
Definitions
Sparse
A hollow matrix may be one with "few" non-zero entries: that is, a sparse matrix.
Block of zeroes
A hollow matrix may be a square n × n matrix with an r × s block of zeroes where r + s > n.
Diagonal entries all zero
A hollow matrix may be a square matrix whose diagonal elements are all equal to zero. That is, an n × n matrix A = (aij) is hollow if aij = 0 whenever i = j (i.e. aii = 0 for all i). The most obvious example is the real skew-symmetric matrix. Other examples are the adjacency matrix of a finite simple graph, and a distance matrix or Euclidean distance matrix.
In other words, any square matrix that takes the form
is a hollow matrix, where the symbol denotes an arbitrary entry.
For example,
is a hollow matrix.
Properties
The trace of a hollow matrix is zero.
If A represents a linear map with respect to a fixed basis, then it maps each basis vector e into the complement of the span of e. That is, where .
The Gershgorin circle theorem shows that the moduli of the eigenvalues of a hollow matrix are less or equal to the sum of the moduli of the non-diagonal row entries. |
https://en.wikipedia.org/wiki/Winsorized%20mean | A winsorized mean is a winsorized statistical measure of central tendency, much like the mean and median, and even more similar to the truncated mean. It involves the calculation of the mean after winsorizing — replacing given parts of a probability distribution or sample at the high and low end with the most extreme remaining values, typically doing so for an equal amount of both extremes; often 10 to 25 percent of the ends are replaced. The winsorized mean can equivalently be expressed as a weighted average of the truncated mean and the quantiles at which it is limited, which corresponds to replacing parts with the corresponding quantiles.
Advantages
The winsorized mean is a useful estimator because by retaining the outliers without taking them too literally, it is less sensitive to observations at the extremes than the straightforward mean, and will still generate a reasonable estimate of central tendency or mean for almost all statistical models. In this regard it is referred to as a robust estimator.
Drawbacks
The winsorized mean uses more information from the distribution or sample than the median. However, unless the underlying distribution is symmetric, the winsorized mean of a sample is unlikely to produce an unbiased estimator for either the mean or the median.
Example
For a sample of 10 numbers (from x(1), the smallest, to x(10) the largest; order statistic notation) the 10% winsorized mean is
The key is in the repetition of x(2) and x(9): the extras substitute for the original values x(1) and x(10) which have been discarded and replaced.
This is equivalent to a weighted average of 0.1 times the 5th percentile (x(2)), 0.8 times the 10% trimmed mean, and 0.1 times the 95th percentile (x(9)).
Notes |
https://en.wikipedia.org/wiki/Quantian | Quantian OS was a remastering of Knoppix/Debian for computational sciences. The environment was self-configuring and directly bootable CD/DVD that turns any PC or laptop (provided it can boot from cdrom/DVD) into a Linux workstation. Quantian also incorporated clusterKnoppix and added support for openMosix, including remote booting of light clients in an openMosix terminal server context permitting rapid setup of a SMP cluster computer.
Applications
Numerous software packages for usual or scientific aims come with Quantian. After the installation, total package volume is about 2.7 GB (For the detailed package list see: List of all the available packages).
The packages for "home users" include:
KDE, the default desktop environment and their components
XMMS, Kaffeine, xine media players
Internet access software, including the KPPP dialer, ISDN utilities and WLAN
The Mozilla, Mozilla Firefox and Konqueror web browsers
K3b, for CD (and DVD) management
The GIMP, an image-manipulation program
Tools for data rescue and system repair
Network analysis and administration tools
OpenOffice.org
Kile, Lyx
Additionally, some of the scientific applications/programs in Quantian are such like:
R, statistical computing software
Octave, a Matlab clone
Scilab, another Matlab clone
GSL, GNU Scientific Library
Maxima computer algebra system
Python programming language with Scipy
Fityk curve fitter
Ghemical for computational chemistry
Texmacs for wysiwyg scientific editing
Grass geographic information system
OpenDX and MayaVi data visualisation systems
Gnuplot, a command-line driven interactive data and function plotting utility
LabPlot, an application for plotting of data sets and functions |
https://en.wikipedia.org/wiki/Message%20Oriented%20Text%20Interchange%20Systems | Message-Oriented Text Interchange System (MOTIS) is an ISO messaging standard based on the ITU-T X.400 standards.
It plays a similar role to the Simple Mail Transfer Protocol (SMTP) in the TCP/IP protocol suite. |
https://en.wikipedia.org/wiki/Congenital%20nephrotic%20syndrome | Congenital nephrotic syndrome is a rare kidney disease which manifests in infants during the first 3 months of life, and is characterized by high levels of protein in the urine (proteinuria), low levels of protein in the blood, and swelling. This disease is primarily caused by genetic mutations which result in damage to components of the glomerular filtration barrier and allow for leakage of plasma proteins into the urinary space.
Signs and symptoms
Urine protein loss leads to total body swelling (generalized edema) and abdominal distension in the first several weeks to months of life. Fluid retention may lead to cough (from pulmonary edema), ascites, and widened cranial sutures and fontanelles. High urine protein loss can lead to foamy appearance of urine. Infants may be born prematurely with low birth weight, and have meconium stained amniotic fluid or a large placenta.
Complications
Frequent, severe infections: urinary loss of immunoglobulins
Malnutrition and poor growth
Blood clots (hypercoagulability): imbalance of plasma coagulation factors from urine protein loss
Hypothyroidism: urinary loss of thyroid-binding protein
Poor bone health associated with vitamin D deficiency: urinary loss of vitamin D binding protein
Acute kidney injury
Chronic kidney disease and ultimately end-stage kidney disease
Causes
Primary (genetic) causes
Mutations in the following five genes account for greater than 80% of the genetic causes of congenital nephrotic syndrome:
NPHS1 (Finnish Type): The gene NPHS1 encodes for the protein nephrin. This genetic variant is characterized by severe protein loss in the first several days to weeks of life. Fin-major and Fin-minor were the first two main genetic mutations identified in Finnish newborns, however, numerous mutations have now been identified in patients all over the world from various ethnic groups. NPHS1 mutations are the most common cause of primary congenital nephrotic syndrome, accounting for 40-80% of cases.
NPHS2 |
https://en.wikipedia.org/wiki/LRK | Long Range Kinematic (LRK) technology is a sophisticated kinematic method developed by Magellan (formerly Thales) Navigation that optimises the advantages of dual-frequency GPS operation. Other conventional methods use the dual-frequency only during initialisation. LRK makes solving ambiguities during initialisation easy and continuous dual-frequency kinematic operation possible at distances up to 40 kilometres.
Conventional dual-frequency kinematic operation is limited to about 10 kilometres, using a combined observation on GPS L1 and L2 frequencies to produce an initial wide lane solution, ambiguous to around 86 centimetres. During a second phase, the conventional kinematic method uses measurements from the L1 frequency only. This method only allows for kinematic operation as long as the de-correlation of atmospheric errors is compatible with a pure phase single-frequency solution.
Similar to the KART process, LRK is a simple and reliable method that allows any initialisation mode, from a static or fixed reference point, to On The Fly ambiguity resolution, when performing dual-frequency GPS positioning. LRK technology reduces initialisation times to a few seconds by efficiently using L2 measurements in every mode of operation. LRK maintains optimal real-time positioning accuracy to within a centimetre at a range up to 40-50 kilometres, even with a reduced number of visible satellites.
External links
https://web.archive.org/web/20060821080822/http://products.thalesnavigation.com/en/products/aboutgps/rtk.asp
Global Positioning System
Navigation
Navigational equipment |
https://en.wikipedia.org/wiki/Mosher%27s%20acid | Mosher's acid, or α-methoxy-α-trifluoromethylphenylacetic acid (MTPA) is a carboxylic acid which was first used by Harry Stone Mosher as a chiral derivatizing agent. It is a chiral molecule, consisting of R and S enantiomers.
Applications
As a chiral derivatizing agent, it reacts with an alcohol or amine of unknown stereochemistry to form an ester or amide. The absolute configuration of the ester or amide is then determined by proton and/or 19F NMR spectroscopy.
Mosher's acid chloride, the acid chloride form, is sometimes used because it has better reactivity.
See also
Pirkle's alcohol |
https://en.wikipedia.org/wiki/Stieltjes%20matrix | In mathematics, particularly matrix theory, a Stieltjes matrix, named after Thomas Joannes Stieltjes, is a real symmetric positive definite matrix with nonpositive off-diagonal entries. A Stieltjes matrix is necessarily an M-matrix. Every n×n Stieltjes matrix is invertible to a nonsingular symmetric nonnegative matrix, though the converse of this statement is not true in general for n > 2.
From the above definition, a Stieltjes matrix is a symmetric invertible Z-matrix whose eigenvalues have positive real parts. As it is a Z-matrix, its off-diagonal entries are less than or equal to zero.
See also
Hurwitz matrix
Metzler matrix |
https://en.wikipedia.org/wiki/Rainbow%20storage | Rainbow storage is a developing paper-based data storage technique first demonstrated by Indian student Sainul Abideen in November 2006. Abideen received his MCA from MES Engineering College Kuttipuram in Kerala's Malappuram District .
Initial newspaper reports of the technology were disputed by multiple technical sources, although Abideen says those reports were based on a misunderstanding of the technology. The paper meant to demonstrate the capability of storing relatively large amounts of data (and not necessarily in the gigabyte range) using textures and diagrams.
The Rainbow data storage technology claims to use geometric shapes such as triangles, circles and squares of various colors to store a large amount of data on ordinary paper or plastic surfaces. This would provide several advantages over current forms of optical- or magnetic data storage like less environmental pollution due to the biodegradability of paper, low cost and high capacity. Data could be stored on "Rainbow Versatile Disk" (RVD) or plastic/paper cards of any form factor (like SIM cards).
Criticism
Following the wide media attention this news received, some of the claims have been disputed by various experts.
Printing at 1,200 dots per inch (DPI) leads to a theoretical maximum of 1,440,000 colored dots per square inch. If a scanner can reliably distinguish between 256 unique colors (thus encoding one byte per dot), the maximum possible storage is approximately 140 megabytes for a sheet of A4 paper–much lower when the necessary error correction is employed. If the scanner were able to accurately distinguish between 16,777,216 colors (24 bits, or 3 bytes per dot), the capacity would triple, but it still falls well below the media stories' claims of several hundred gigabytes.
Printing this quantity of unique colors would require specialized equipment to generate many spot colors. The process color model used by most printers provides only four colors, with additional colors simulated |
https://en.wikipedia.org/wiki/Riesz%20space | In mathematics, a Riesz space, lattice-ordered vector space or vector lattice is a partially ordered vector space where the order structure is a lattice.
Riesz spaces are named after Frigyes Riesz who first defined them in his 1928 paper Sur la décomposition des opérations fonctionelles linéaires.
Riesz spaces have wide-ranging applications. They are important in measure theory, in that important results are special cases of results for Riesz spaces. For example, the Radon–Nikodym theorem follows as a special case of the Freudenthal spectral theorem. Riesz spaces have also seen application in mathematical economics through the work of Greek-American economist and mathematician Charalambos D. Aliprantis.
Definition
Preliminaries
If is an ordered vector space (which by definition is a vector space over the reals) and if is a subset of then an element is an upper bound (resp. lower bound) of if (resp. ) for all
An element in is the least upper bound or supremum (resp. greater lower bound or infimum) of if it is an upper bound (resp. a lower bound) of and if for any upper bound (resp. any lower bound) of (resp. ).
Definitions
Preordered vector lattice
A preordered vector lattice is a preordered vector space in which every pair of elements has a supremum.
More explicitly, a preordered vector lattice is vector space endowed with a preorder, such that for any :
Translation Invariance: implies
Positive Homogeneity: For any scalar implies
For any pair of vectors there exists a supremum (denoted ) in with respect to the order
The preorder, together with items 1 and 2, which make it "compatible with the vector space structure", make a preordered vector space.
Item 3 says that the preorder is a join semilattice.
Because the preorder is compatible with the vector space structure, one can show that any pair also have an infimum, making also a meet semilattice, hence a lattice.
A preordered vector space is a preordered vector lattice i |
https://en.wikipedia.org/wiki/Ordered%20vector%20space | In mathematics, an ordered vector space or partially ordered vector space is a vector space equipped with a partial order that is compatible with the vector space operations.
Definition
Given a vector space over the real numbers and a preorder on the set the pair is called a preordered vector space and we say that the preorder is compatible with the vector space structure of and call a vector preorder on if for all and with the following two axioms are satisfied
implies
implies
If is a partial order compatible with the vector space structure of then is called an ordered vector space and is called a vector partial order on
The two axioms imply that translations and positive homotheties are automorphisms of the order structure and the mapping is an isomorphism to the dual order structure. Ordered vector spaces are ordered groups under their addition operation.
Note that if and only if
Positive cones and their equivalence to orderings
A subset of a vector space is called a cone if for all real A cone is called pointed if it contains the origin. A cone is convex if and only if The intersection of any non-empty family of cones (resp. convex cones) is again a cone (resp. convex cone);
the same is true of the union of an increasing (under set inclusion) family of cones (resp. convex cones). A cone in a vector space is said to be generating if
Given a preordered vector space the subset of all elements in satisfying is a pointed convex cone with vertex (that is, it contains ) called the positive cone of and denoted by
The elements of the positive cone are called positive.
If and are elements of a preordered vector space then if and only if The positive cone is generating if and only if is a directed set under
Given any pointed convex cone with vertex one may define a preorder on that is compatible with the vector space structure of by declaring for all that if and only if
the positive cone of this resul |
https://en.wikipedia.org/wiki/Acoustic%20contrast%20factor | In acoustics, the acoustic contrast factor is a number that describes the relationship between the densities and the sound velocities of two media, or equivalently (because of the form of the expression), the relationship between the densities and compressibilities of two media. It is most often used in the context of biomedical ultrasonic imaging techniques using acoustic contrast agents and in the field of ultrasonic manipulation of particles (acoustophoresis) much smaller than the wavelength using ultrasonic standing waves. In the latter context, the acoustic contrast factor is the number which, depending on its sign, tells whether a given type of particle in a given medium will be attracted to the pressure nodes or anti-nodes.
Example - particle in a medium
In an ultrasonic standing wave field, a small spherical particle (, where is the particle radius, and is the wavelength) suspended in an inviscid fluid will move under the effect of an acoustic radiation force. The direction of its movement is governed by the physical properties of the particle and the surrounding medium, expressed in the form of an acoustophoretic contrast factor .
Given the compressibilities and and densities and of the medium and particle, respectively, the acoustic contrast factor can be expressed as:
For a positive value of , the particles will be attracted to the pressure nodes.
For a negative value of , the particles will be attracted to the pressure anti-nodes.
See also
Acoustic impedance
Acoustic tweezers |
https://en.wikipedia.org/wiki/Acidophile | Acidophiles or acidophilic organisms are those that thrive under highly acidic conditions (usually at pH 5.0 or below). These organisms can be found in different branches of the tree of life, including Archaea, Bacteria, and Eukarya.
Examples
A list of these organisms includes:
Archaea
Sulfolobales, an order in the Thermoproteota branch of Archaea
Thermoplasmatales, an order in the Euryarchaeota branch of Archaea
ARMAN, in the Euryarchaeota branch of Archaea
Acidianus brierleyi, A. infernus, facultatively anaerobic thermoacidophilic archaebacteria
Halarchaeum acidiphilum, acidophilic member of the Halobacteriacaeae
Metallosphaera sedula, thermoacidophilic
Bacteria
Acidobacteriota, a phylum of Bacteria
Acidithiobacillales, an order of Pseudomonadota e.g. A. ferrooxidans, A. thiooxidans
Thiobacillus prosperus, T. acidophilus, T. organovorus, T. cuprinus
Acetobacter aceti, a bacterium that produces acetic acid (vinegar) from the oxidation of ethanol.
Alicyclobacillus, a genus of bacteria that can contaminate fruit juices.
Eukarya
Mucor racemosus
Urotricha
Dunaliella acidophila
Members of the algal class Cyanidiophyceae, including Cyanidioschyzon merolae
Mechanisms of adaptation to acidic environments
Most acidophile organisms have evolved extremely efficient mechanisms to pump protons out of the intracellular space in order to keep the cytoplasm at or near neutral pH. Therefore, intracellular proteins do not need to develop acid stability through evolution. However, other acidophiles, such as Acetobacter aceti, have an acidified cytoplasm which forces nearly all proteins in the genome to evolve acid stability. For this reason, Acetobacter aceti has become a valuable resource for understanding the mechanisms by which proteins can attain acid stability.
Studies of proteins adapted to low pH have revealed a few general mechanisms by which proteins can achieve acid stability. In most acid stable proteins (such as pepsin and the soxF protein from Sulfol |
https://en.wikipedia.org/wiki/Acidophile%20%28histology%29 | Acidophile (or acidophil, or, as an adjectival form, acidophilic) is a term used by histologists to describe a particular staining pattern of cells and tissues when using haematoxylin and eosin stains. Specifically, the name refers to structures which "love" acid, and take it up readily. More specifically, acidophilia can be described by cationic groups of most often proteins in the cell readily reacting with acidic stains.
It describes the microscopic appearance of cells and tissues, as seen through a microscope, after a histological section has been stained with an acidic dye. The most common such dye is eosin, which stains acidophilic substances red and is the source of the related term eosinophilic. Note that a single cell can have both acidophilic substances/organelles and basophilic substances/organelles, albeit some have historically had so much of one stain that the cell itself is called an eosinophil.
See also
Anterior pituitary acidophil
Basophilic
Oxyphil cell
Eosinophil granulocyte
Eosinophilic
Acidophil cell
Acidophile |
https://en.wikipedia.org/wiki/Adaptive%20quadrature | Adaptive quadrature is a numerical integration method in which the integral of a function is approximated using static quadrature rules on adaptively refined subintervals of the region of integration. Generally, adaptive algorithms are just as efficient and effective as traditional algorithms for "well behaved" integrands, but are also effective for "badly behaved" integrands for which traditional algorithms may fail.
General scheme
Adaptive quadrature follows the general scheme
1. procedure integrate ( f, a, b, τ )
2.
3.
4. if ε > τ then
5. m = (a + b) / 2
6. Q = integrate(f, a, m, τ/2) + integrate(f, m, b, τ/2)
7. endif
8. return Q
An approximation to the integral of over the interval is computed (line 2), as well as an error estimate (line 3). If the estimated error is larger than the required tolerance (line 4), the interval is subdivided (line 5) and the quadrature is applied on both halves separately (line 6). Either the initial estimate or the sum of the recursively computed halves is returned (line 7).
The important components are the quadrature rule itself
the error estimator
and the logic for deciding which interval to subdivide, and when to terminate.
There are several variants of this scheme. The most common will be discussed later.
Basic rules
The quadrature rules generally have the form
where the nodes and weights are generally precomputed.
In the simplest case, Newton–Cotes formulas of even degree are used, where the nodes are evenly spaced in the interval:
When such rules are used, the points at which has been evaluated can be re-used upon recursion:
A similar strategy is used with Clenshaw–Curtis quadrature, where the nodes are chosen as
Or, when Fejér quadrature is used,
Other quadrature rules, such as Gaussian quadrature or Gauss-Kronrod quadrature, may also be used.
An algorithm may elect to use different quadrature methods on different subintervals, for example using a high-or |
https://en.wikipedia.org/wiki/Magnesium%20phosphate | Magnesium phosphate is a general term for salts of magnesium and phosphate appearing in several forms and several hydrates:
Monomagnesium phosphate (Mg(H2PO4)2).xH2O
Dimagnesium phosphate (MgHPO4).xH2O
Trimagnesium phosphate (Mg3(PO4)2).xH2O
Amorphous magnesium phosphate is also claimed.
Safety
Magnesium phosphates are listed on the U.S. FDA's Generally recognized as safe (GRAS) list of substances.
See also
Ammonium magnesium phosphate |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.