source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Cauda%20helicis | In the lower part of the helix the cartilage is prolonged downward as a tail-like process, the cauda helicis; this is separated from the antihelix by a fissure, the fissura antitragohelicina. |
https://en.wikipedia.org/wiki/Lateral%20shoot | A lateral shoot, commonly known as a branch, is a part of a plant's shoot system that develops from axillary buds on the stem's surface, extending laterally from the plant's stem.
Importance to photosynthesis
As a plant grows it requires more energy, it also is required to out-compete nearby plants for this energy. One of the ways a plant can compete for this energy is to increase its height, another is to increase its overall surface area. That is to say, the more lateral shoots a plant develops, the more foliage the plant can support increases how much photosynthesis the plant can perform as it allows for more area for the plant to uptake carbon dioxide as well as sunlight.
Genes, transcription factors, and growth
Through testing with Arabidopsis thaliana (A plant considered a model organism for plant genetic studies) genes including MAX1 and MAX2 have been found to affect growth of lateral shoots. Gene knockouts of these genes cause abnormal proliferation of the plants affected, implying they are used for repressing said growth in wild type plants. Another set of experiments with Arabidopsis thaliana testing genes in the plant hormone florigen, two genes FT and TSF (which are abbreviations for Flowering Locus T, and Twin Sister of FT) when knocked out, appear to affect lateral shoot in a negative fashion. These mutants cause slower growth and improper formation of lateral shoots, which could also mean that lateral shoots are important to florigen's function. Along with general growth there are also transcription factors that directly effect the production of additional lateral shoots like the TCP family (also known as Teosinte branched 1/cycloidea/proliferating cell factor) which are plant specific proteins that suppress lateral shoot branching. Additionally the TCP family has been found to be partially responsible for inhibiting the cell's Growth hormone–releasing hormone (GHRF) which means it also inhibits cell proliferation.
See also
Apical dominance
Sho |
https://en.wikipedia.org/wiki/Adelphogamy | Adelphogamy is a form of sexual partnership between sibling eukaryotes, especially in some species of fungi, flowering plants or ants, or in humans. In flowering plants, adelphogamy refers to sibling pollination: pollen and stigma belong to two individuals which derives from same mother plant.
In sociology, the term adelphogamy or adelphic polyandry may also refer to fraternal polyandry, or to an incestuous relationship between a brother and sister.
See also
Incest between twins
List of coupled siblings |
https://en.wikipedia.org/wiki/Life-time%20of%20correlation | In probability theory and related fields, the life-time of correlation measures the timespan over which there is appreciable autocorrelation or cross-correlation in stochastic processes.
Definition
The correlation coefficient ρ, expressed as an autocorrelation function or cross-correlation function, depends on the lag-time between the times being considered. Typically such functions, ρ(t), decay to zero with increasing lag-time, but they can assume values across all levels of correlations: strong and weak, and positive and negative as in the table.
The life-time of a correlation is defined as the length of time when the correlation coefficient is at the strong level. The durability of correlation is determined by signal (the strong level of correlation is separated from weak and negative levels). The mean life-time of correlation could measure how the durability of correlation depends on the window width size (the window is the length of time series used to calculate correlation). |
https://en.wikipedia.org/wiki/Creditcall | Creditcall (now NMI) is a payment service provider and payment gateway with offices in the United States and UK, providing credit card authorisation and settlement services to banks and processors in the United Kingdom, United States and Canada.
History
Creditcall Limited, originally Creditcall Communications Limited, was founded in 1996. Later, a North American subsidiary, Creditcall Corporation, was incorporated in 2005. The name Creditcall is derived from the name of Creditcall's first product, a telecommunications service that enabled callers to bill telephone calls to their credit or debit card. Creditcall won 3i's Business Catapult Award in 1998 along with an initial investment in the company.
In 2012 the management team completed a management buyout of the company backed by FF&P Private Equity and Bestport Ventures.
On 21 April 2014 Creditcall was awarded the Queen's Award for Enterprise.
Acquisition
In March 2018, Creditcall was acquired by US-based payments' technology company NMI. NMI is headquartered in Schaumburg, Illinois, with offices in Utah, New York, and Bristol, UK. NMI is backed by US private equity firms Insight Partners, Great Hill Partners and Francisco Partners. |
https://en.wikipedia.org/wiki/Gongronella | Gongronella is a genus of fungi belonging to the family Cunninghamellaceae.
The genus has cosmopolitan distribution.
Species
Species:
Gongronella brasiliensis
Gongronella butleri
Gongronella guangdongensis |
https://en.wikipedia.org/wiki/Reflected-wave%20switching | Reflected-wave switching is a signalling technique used in backplane computer buses such as PCI.
A backplane computer bus is a type of multilayer printed circuit board that has at least one (almost) solid layer of copper called the ground plane, and at least one layer of copper tracks that are used as wires for the signals. Each signal travels along a transmission line formed by its track and the narrow strip of ground plane directly beneath it. This structure is known in radio engineering as microstrip line.
Each signal travels from a transmitter to one or more receivers. Most computer buses use binary digital signals, which are sequences of pulses of fixed amplitude. In order to receive the correct data, the receiver must detect each pulse once, and only once. To ensure this, the designer must take the high-frequency characteristics of the microstrip into account.
When a pulse is launched into the microstrip by the transmitter, its amplitude depends on the ratio of the impedances of the transmitter and the microstrip. The impedance of the transmitter is simply its output resistance. The impedance of the microstrip is its characteristic impedance, which depends on its dimensions and on the materials used in the backplane's construction. As the leading edge of the pulse (the incident wave) passes the receiver, it may or may not have sufficient amplitude to be detected. If it does, then the system is said to use incident-wave switching. This is the system used in most computer buses predating PCI, such as the VME bus.
When the pulse reaches the end of the microstrip, its behaviour depends on the circuit conditions at this point. If the microstrip is correctly terminated (usually with a combination of resistors), the pulse is absorbed and its energy is converted to heat. This is the case in an incident-wave switching bus. If, on the other hand, there is no termination at the end of the microstrip, and the pulse encounters an open circuit, it is reflec |
https://en.wikipedia.org/wiki/Stomatorhinus | Stomatorhinus is a genus of small elephantfish in the family Mormyridae.
Species
There are currently 13 recognized species in this genus:
Stomatorhinus ater Pellegrin 1924 (Kidada mormyrid)
Stomatorhinus corneti Boulenger 1899 (Stanley Pool mormyrid)
Stomatorhinus fuliginosus Poll 1941 (Mosongolia mormyrid)
Stomatorhinus humilior Boulenger 1899 (Ibali mormyrid)
Stomatorhinus ivindoensis Sullivan & Hopkins 2005 (Biale Creek mormyrid)
Stomatorhinus kununguensis Poll 1945 (Kunungu mormyrid)
Stomatorhinus microps Boulenger 1898 (Boma mormyrid)
Stomatorhinus patrizii Vinciguerra 1928 (Buta mormyrid)
Stomatorhinus polli Matthes 1964 (Ikela mormyrid)
Stomatorhinus polylepis Boulenger 1899 (Isangila mormyrid)
Stomatorhinus puncticulatus Boulenger 1899 (Chiloango mormyrid)
Stomatorhinus schoutedeni Poll 1945 (Schouteden's mormyrid)
Stomatorhinus walkeri (Günther 1867) (Walker mormyrid) |
https://en.wikipedia.org/wiki/Cartan%27s%20lemma%20%28potential%20theory%29 | In potential theory, a branch of mathematics, Cartan's lemma, named after Henri Cartan, is a bound on the measure and complexity of the set on which a logarithmic Newtonian potential is small.
Statement of the lemma
The following statement can be found in Levin's book.
Let μ be a finite positive Borel measure on the complex plane C with μ(C) = n. Let u(z) be the logarithmic potential of μ:
Given H ∈ (0, 1), there exist discs of radii ri such that
and
for all z outside the union of these discs.
Notes
Complex analysis |
https://en.wikipedia.org/wiki/Atomic%20form%20factor | In physics, the atomic form factor, or atomic scattering factor, is a measure of the scattering amplitude of a wave by an isolated atom. The atomic form factor depends on the type of scattering, which in turn depends on the nature of the incident radiation, typically X-ray, electron or neutron. The common feature of all form factors is that they involve a Fourier transform of a spatial density distribution of the scattering object from real space to momentum space (also known as reciprocal space). For an object with spatial density distribution, , the form factor, , is defined as
,
where is the spatial density of the scatterer about its center of mass (), and is the momentum transfer. As a result of the nature of the Fourier transform, the broader the distribution of the scatterer in real space , the narrower the distribution of in ; i.e., the faster the decay of the form factor.
For crystals, atomic form factors are used to calculate the structure factor for a given Bragg peak of a crystal.
X-ray form factors
X-rays are scattered by the electron cloud of the atom and hence the scattering amplitude of X-rays increases with the atomic number, , of the atoms in a sample. As a result, X-rays are not very sensitive to light atoms, such as hydrogen and helium, and there is very little contrast between elements adjacent to each other in the periodic table. For X-ray scattering, in the above equation is the electron charge density about the nucleus, and the form factor the Fourier transform of this quantity. The assumption of a spherical distribution is usually good enough for X-ray crystallography.
In general the X-ray form factor is complex but the imaginary components only become large near an absorption edge. Anomalous X-ray scattering makes use of the variation of the form factor close to an absorption edge to vary the scattering power of specific atoms in the sample by changing the energy of the incident x-rays hence enabling the extraction of more deta |
https://en.wikipedia.org/wiki/List%20of%20version-control%20software | This is a list of notable software for version control.
Local data model
In the local-only approach, all developers must use the same file system.
Open source
Revision Control System (RCS) – stores the latest version and backward deltas for fastest access to the trunk tip compared to SCCS and an improved user interface, at the cost of slow branch tip access and missing support for included/excluded deltas.
Source Code Control System (SCCS) – part of UNIX; based on interleaved deltas, can construct versions as arbitrary sets of revisions. Extracting an arbitrary version takes essentially the same time and is thus more useful in environments that rely heavily on branching and merging with multiple "current" and identical versions.
Proprietary
The Librarian – Around since 1969, source control for IBM mainframe computers; from Applied Data Research, later acquired by Computer Associates
Panvalet – Around since the 1970s, source and object control for IBM mainframe computers.
Client–server model
In the client–server model, developers use one shared repository.
Open source
Concurrent Versions System (CVS) – originally built on RCS, licensed under the GPL.
CVSNT – cross-platform port of CVS that allows case insensitive file names among other changes
OpenCVS – unreleased CVS clone under a BSD license, emphasising security and source code correctness
Subversion (SVN) – versioning control system inspired by CVS
Vesta – build system with a versioning file system and support for distributed repositories
Proprietary
AccuRev – source configuration management tool with integrated issue tracking based on "Streams" that efficiently manages parallel and global development; replication server is also available. Now owned by Micro Focus.
Autodesk Vault – Version control tool specifically designed for Autodesk applications managing the complex relationships between design files such as AutoCAD and Autodesk Inventor.
CADES – Designer productivity and version contr |
https://en.wikipedia.org/wiki/Paris%27%20law | Paris' law (also known as the Paris–Erdogan equation) is a crack growth equation that gives the rate of growth of a fatigue crack. The stress intensity factor characterises the load around a crack tip and the rate of crack growth is experimentally shown to be a function of the range of stress intensity seen in a loading cycle. The Paris equation is
where is the crack length and is the fatigue crack growth for a load cycle . The material coefficients and are obtained experimentally and also depend on environment, frequency, temperature and stress ratio. The stress intensity factor range has been found to correlate the rate of crack growth from a variety of different conditions and is the difference between the maximum and minimum stress intensity factors in a load cycle and is defined as
Being a power law relationship between the crack growth rate during cyclic loading and the range of the stress intensity factor, the Paris–Erdogan equation can be visualized as a straight line on a log-log plot, where the x-axis is denoted by the range of the stress intensity factor and the y-axis is denoted by the crack growth rate.
The ability of ΔK to correlate crack growth rate data depends to a large extent on the fact that alternating stresses causing crack growth are small compared to the yield strength. Therefore crack tip plastic zones are small compared to crack length even in very ductile materials like stainless steels.
The equation gives the growth for a single cycle. Single cycles can be readily counted for constant-amplitude loading. Additional cycle identification techniques such as rainflow-counting algorithm need to be used to extract the equivalent constant-amplitude cycles from a variable-amplitude loading sequence.
History
In a 1961 paper, P. C. Paris introduced the idea that the rate of crack growth may depend on the stress intensity factor. Then in their 1963 paper, Paris and Erdogan indirectly suggested the equation with the aside remark " |
https://en.wikipedia.org/wiki/Canthaxanthin | Canthaxanthin is a keto-carotenoid pigment widely distributed in nature. Carotenoids belong to a larger class of phytochemicals known as terpenoids. The chemical formula of canthaxanthin is C40H52O2. It was first isolated in edible mushrooms. It has also been found in green algae, bacteria, crustaceans, and bioaccumulates in fish such as carp, golden grey mullet, seabream and trush wrasse.
Canthaxanthin is associated with E number E161g and is approved for use as a food coloring agent in different countries, including the United States and the EU; however, it is not approved for use in Australia and New Zealand. It is generally authorized for feed applications in at least the following countries: US, Canada, EU. In the EU, canthaxanthin is allowed by law to be added to trout feed, salmon feed and poultry feed. The European Union limit is 80 mg/kg of feedstuffs, 8 mg/kg in feed for egg laying hens and 25 mg/kg in feed for other poultry and salmonids.
Canthaxanthin is a potent lipid-soluble antioxidant. The biological functions of canthaxanthin are related, at least in part, to its ability to function as an antioxidant (free radical scavenging/vitamin E sparing) in animal tissues.
Biosynthesis
Due to the commercial value of carotenoids, their biosynthesis has been studied extensively in both natural producers, and non-natural (heterologous) systems such as the bacteria Escherichia coli and yeast Saccharomyces cerevisiae. Canthaxanthin biosynthesis proceeds from beta-carotene via the action of a single protein, known as a beta-carotene ketolase, that is able to add a carbonyl group to carbon 4 and 4' of the beta carotene molecule.
Although functionally identical, several distinct beta-carotene ketolase proteins are known. That is to say they differ from an evolutionary perspective in their primary amino acid/protein sequence. They are different proteins that complete the same function. Thus, bacterial (CrtW) and micro-algal beta-carotene ketolase proteins such as |
https://en.wikipedia.org/wiki/CGB5 | Chorionic gonadotropin, beta polypeptide 5 is a protein that in humans is encoded by the CGB5 gene.
Function
This gene is a member of the glycoprotein hormone beta chain family and encodes the beta 5 subunit of chorionic gonadotropin (CG). Glycoprotein hormones are heterodimers consisting of a common alpha subunit and a unique beta subunit which confers biological specificity. CG is produced by the trophoblastic cells of the placenta and stimulates the ovaries to synthesize the steroids that are essential for the maintenance of pregnancy. The beta subunit of CG is encoded by 6 genes which are arranged in tandem and inverted pairs on chromosome 19q13.3 and contiguous with the luteinizing hormone beta subunit gene. [provided by RefSeq, Jul 2008]. |
https://en.wikipedia.org/wiki/Color%20space | A color space is a specific organization of colors. In combination with color profiling supported by various physical devices, it supports reproducible representations of colorwhether such representation entails an analog or a digital representation. A color space may be arbitrary, i.e. with physically realized colors assigned to a set of physical color swatches with corresponding assigned color names (including discrete numbers infor examplethe Pantone collection), or structured with mathematical rigor (as with the NCS System, Adobe RGB and sRGB). A "color space" is a useful conceptual tool for understanding the color capabilities of a particular device or digital file. When trying to reproduce color on another device, color spaces can show whether shadow/highlight detail and color saturation can be retained, and by how much either will be compromised.
A "color model" is an abstract mathematical model describing the way colors can be represented as tuples of numbers (e.g. triples in RGB or quadruples in CMYK); however, a color model with no associated mapping function to an absolute color space is a more or less arbitrary color system with no connection to any globally understood system of color interpretation. Adding a specific mapping function between a color model and a reference color space establishes within the reference color space a definite "footprint", known as a gamut, and for a given color model, this defines a color space. For example, Adobe RGB and sRGB are two different absolute color spaces, both based on the RGB color model. When defining a color space, the usual reference standard is the CIELAB or CIEXYZ color spaces, which were specifically designed to encompass all colors the average human can see.
Since "color space" identifies a particular combination of the color model and the mapping function, the word is often used informally to identify a color model. However, even though identifying a color space automatically identifies the associated |
https://en.wikipedia.org/wiki/Bragg%27s%20law | In physics and chemistry, Bragg's law, Wulff–Bragg's condition or Laue–Bragg interference, a special case of Laue diffraction, gives the angles for coherent scattering of waves from a large crystal lattice. It encompasses the superposition of wave fronts scattered by lattice planes, leading to a strict relation between wavelength and scattering angle, or else to the wavevector transfer with respect to the crystal lattice. Such law had initially been formulated for X-rays upon crystals. However, it applies to all sorts of quantum beams, including neutron and electron waves at atomic distances if there are a large number of atoms, as well as visible light with artificial periodic microscale lattices.
History
Bragg diffraction (also referred to as the Bragg formulation of X-ray diffraction) was first proposed by Lawrence Bragg and his father, William Henry Bragg, in 1913 in response to their discovery that crystalline solids produced surprising patterns of reflected X-rays (in contrast to that of, say, a liquid). They found that these crystals, at certain specific wavelengths and incident angles, produced intense peaks of reflected radiation. The derived Bragg's law is a special interpretation of Laue diffraction, where the Braggs interpreted the constructive Laue-Bragg interference in a geometric way by reflection of waves from crystal lattice planes, such that the path-difference becomes a multiple of the incident wavelength.
Lawrence Bragg explained this result by modeling the crystal as a set of discrete parallel planes separated by a constant parameter . It was proposed that the incident X-ray radiation would produce a Bragg peak if their reflections off the various planes interfered constructively. The interference is constructive when the phase shift is a multiple of ; this condition can be expressed by Bragg's law (see Bragg condition section below) and was first presented by Lawrence Bragg on 11 November 1912 to the Cambridge Philosophical Society. Altho |
https://en.wikipedia.org/wiki/PCGamingWiki | The PCGamingWiki is a British-based collaboratively edited free wiki internet encyclopaedia focused on collecting video game behaviour data (such as save locations and startup parameters), to optimising gameplay, and fixing issues found in PC games. Intended fixes and optimisations range from simple cut-scene removals, to modifications that allow for wide-screen resolutions, and more. The wiki site runs on MediaWiki software, and was created by Andrew Tsai, a British businessman from London, England. The site was founded on . As of October 2022, the PCGamingWiki has more than 30,000 registered users, and 48,000 content pages. Since its inception, the PCGamingWiki has been featured on numerous gaming focused websites, including Kotaku, Destructoid, and Rock Paper Shotgun. It regularly receives more than 10,000 unique page views a day.
History
The PCGamingWiki was founded on , by Andrew Tsai, who is also known under the username 'Andytizer'. Tsai was motivated to create the wiki based on his experiences with the games LA Noire and Titan Quest. The wiki was mostly baren until Andrew enlisted the help of users on the website Reddit.
On 11 April 2012, Tsai attempted to Kickstart the PCGamingWiki with a goal of $60,000. The project ended on 12 May 2012, failing to complete its goal and only earning $2,736. On 19 December 2012 the project was put on Kickstarter again, this time with a goal of £500. This was much more successful, ending on 20 January 2013 with more than 400% funding.
On 24 December 2012, the PCGamingWiki forums were created with help from user JRWR. The forums are used to discuss articles, improvements, fix problems, and report bugs. As of March 2014, there have been 3374 posts to the forums.
On 26 March 2013, Andrew Tsai announced a new section of the PCGamingWiki network entitled The Port Report. The Port Report would function in a similar way to the, then recently defunct, Port Authority section of GameSpy. The Port Report would focus on vi |
https://en.wikipedia.org/wiki/Kerala%20School%20of%20Mathematics%2C%20Kozhikode | The Kerala School of Mathematics (KSoM) in Kozhikode, India is a research institute in Theoretical sciences with a focus on Mathematics. The institute is a joint venture of the Department of Atomic Energy (DAE) and the Kerala State Council for Science, Technology and Environment (KSCSTE). Kerala School of Mathematics is a center of advanced research and learning in Mathematics and is a meeting ground for leading Mathematicians from around the world.
Kerala School of Mathematics has a doctoral program to which students are admitted on an yearly basis. The institute also has an Integrated MSc-PhD program with an option for students to exit the program with an MSc degree at the end of two years.
History
Mathematics in Kerala, during the times of Madhava of Sangamagrama, majorly flourished in the Muziris region of Thrikkandiyur, Thirur, Alattiyur, and Tirunavaya in the Malabar region of Kerala. Kerala school of astronomy and mathematics flourished between the 14th and 16th centuries. Commemorating the rich heritage of Mathematics in the region, Kerala School of Mathematics was hence chosen to be set up in the scenic mountains of the Western Ghats in the city of Kozhikode.
The nascent plan to set up Kerala School of Mathematics started forming shape in around 2004. The then DAE chairman Anil Kakodkar and the then executive vice president of KSCSTE, M. S. Valiathan were instrumental in setting up the institute with the guidance of M. S. Raghunathan, Rajeeva Karandikar and Alladi Sitaram. The foundation stone of KSoM was laid by the then Chief Minister A.K. Antony in 2004. The institute was later inaugurated in 2008 by the then Chief Minister V. S. Achuthanandan and finally set up in 2009 with Parameswaran A. J. as the founding director.
External links
Official Website |
https://en.wikipedia.org/wiki/Vector%20algebra%20relations | The following are important identities in vector algebra. Identities that involve the magnitude of a vector , or the dot product (scalar product) of two vectors A·B, apply to vectors in any dimension. Identities that use the cross product (vector product) A×B are defined only in three dimensions.
Magnitudes
The magnitude of a vector A can be expressed using the dot product:
In three-dimensional Euclidean space, the magnitude of a vector is determined from its three components using Pythagoras' theorem:
Inequalities
The Cauchy–Schwarz inequality:
The triangle inequality:
The reverse triangle inequality:
Angles
The vector product and the scalar product of two vectors define the angle between them, say θ:
To satisfy the right-hand rule, for positive θ, vector B is counter-clockwise from A, and for negative θ it is clockwise.
The Pythagorean trigonometric identity then provides:
If a vector A = (Ax, Ay, Az) makes angles α, β, γ with an orthogonal set of x-, y- and z-axes, then:
and analogously for angles β, γ. Consequently:
with unit vectors along the axis directions.
Areas and volumes
The area Σ of a parallelogram with sides A and B containing the angle θ is:
which will be recognized as the magnitude of the vector cross product of the vectors A and B lying along the sides of the parallelogram. That is:
(If A, B are two-dimensional vectors, this is equal to the determinant of the 2 × 2 matrix with rows A, B.) The square of this expression is:
where Γ(A, B) is the Gram determinant of A and B defined by:
In a similar fashion, the squared volume V of a parallelepiped spanned by the three vectors A, B, C is given by the Gram determinant of the three vectors:
Since A, B, C are three-dimensional vectors, this is equal to the square of the scalar triple product below.
This process can be extended to n-dimensions.
Addition and multiplication of vectors
Commutativity of addition: .
Commutativity of scalar product: .
Anticommutativity of cross product |
https://en.wikipedia.org/wiki/Simple%20Key-Management%20for%20Internet%20Protocol | Simple Key-Management for Internet Protocol or SKIP was a protocol developed circa 1995 by the IETF Security Working Group for the sharing of encryption keys. SKIP and Photuris were evaluated as key exchange mechanisms for IPsec before the adoption of IKE in 1998.
Simple Key Management for Internet Protocols (SKIP) is similar to SSL, except that it establishes a long-term key once, and then requires no prior communication in order to establish or exchange keys on a session-by-session basis. Therefore, no connection setup overhead exists and new keys values are not continually generated. |
https://en.wikipedia.org/wiki/Game%20tree | In the context of combinatorial game theory, which typically studies sequential games with perfect information, a game tree is a graph representing all possible game states within such a game. Such games include well-known ones such as chess, checkers, Go, and tic-tac-toe. This can be used to measure the complexity of a game, as it represents all the possible ways a game can pan out. Due to the large game trees of complex games such as chess, algorithms that are designed to play this class of games will use partial game trees, which makes computation feasible on modern computers. Various methods exist to solve game trees. If a complete game tree can be generated, a deterministic algorithm, such as backward induction or retrograde analysis can be used. Randomized algorithms and minmax algorithms such as MCTS can be used in cases where a complete game tree is not feasible.
Understanding the game tree
To better understand the game tree, it can be thought of as a technique for analyzing adversarial games, which determine the actions that player takes to win the game. In game theory, a game tree is a directed graph whose nodes are positions in a game (e.g., the arrangement of the pieces in a board game) and whose edges are moves (e.g., to move pieces from one position on a board to another).
The complete game tree for a game is the game tree starting at the initial position and containing all possible moves from each position; the complete tree is the same tree as that obtained from the extensive-form game representation. To be more specific, the complete game is a norm for the game in game theory. Which can clearly express many important aspects. For example, the sequence of actions that stakeholders may take, their choices at each decision point, information about actions taken by other stakeholders when each stakeholder makes a decision, and the benefits of all possible game results.
The diagram shows the first two levels, or plies, in the game tree for tic-tac-to |
https://en.wikipedia.org/wiki/Tetrasomy | A tetrasomy is a form of aneuploidy with the presence of four copies, instead of the normal two, of a particular chromosome.
Causes
Full
Full tetrasomy of an individual occurs due to non-disjunction when the cells are dividing (meiosis I or II) to form egg and sperm cells (gametogenesis). This can result in extra chromosomes in a sperm or egg cell. After fertilization, the resulting fetus has 48 chromosomes instead of the typical 46.
Autosomal tetrasomies
Cat eye syndrome where partial tetrasomy of chromosome 22 is present
Pallister-Killian syndrome (tetrasomy 12p)
Tetrasomy 9p
Tetrasomy 18p
Tetrasomy 21, a rare form of Down syndrome
Sex-chromosome tetrasomies
Tetrasomy X
XXYY syndrome
External links
Cytogenetics
Chromosomal abnormalities |
https://en.wikipedia.org/wiki/Active%20immunotherapy | Active immunotherapy is a type of immunotherapy that aims to stimulate the host's immune system or a specific immune response to a disease or pathogen and is most commonly used in cancer treatments. Active immunotherapy is also used for treatment of neurodegenerative disorders, such as Alzheimer's disease, Parkinson's disease, prion disease, and multiple sclerosis. Active immunotherapies induce an immune response through direct immune system stimulation, while immunotherapies that administer antibodies directly to the system are classified as passive immunotherapies. Active immunotherapies can elicit generic and specific immune responses depending on the goal of the treatment. The categories of active immunotherapy divide into:
Non-specific active immunotherapy: generating a general immune system response using cytokines and other cell signaling molecules.
Specific active immunotherapy: the generation of cell-mediated and antibody immune responses focused on specific antigens expressed by the cancer cells, typically using a vaccine platform.
Active immunotherapies fall under the category of activation immunotherapies, which is a subset of immunotherapies that activate the immune system as opposed immunotherapies that suppress the immune system.
Non-specific active immunotherapy
Non-specific active immunotherapy is administered with the overall goal of generally eliminating malignant pathogens or cells from the host system. This treatment stimulates the immune system in a general sense, rather than specifically targeting a cell type e.g. cancer cells. Non-specific approaches aim to create a robust immune response that would lead to the eventual killing of malignant cells through immunomodulators such as cytokines.
Cytokines
Immunomodulating agents regulate the immune system's response and are produced by various immune cells. These agents include the following agents and markers:
BCG Vaccine
The BCG vaccine has been used against tuberculosis, mycobacteria |
https://en.wikipedia.org/wiki/ANKK1 | Ankyrin repeat and kinase domain containing 1 (ANKK1) also known as protein kinase PKK2 or sugen kinase 288 (SgK288) is an enzyme that in humans is encoded by the ANKK1 gene. The ANKK1 is a member of an extensive family of the Ser/Thr protein kinase family, and protein kinase superfamily involved in signal transduction pathways.
Clinical significance
This gene contains a single nucleotide polymorphism that causes an amino acid substitution within the 11th of 12 ankyrin repeats of ANKK1 (Glu713Lys of 765 residues). This polymorphism, which is commonly referred to Taq1A, was previously believed to be located in the promoter region of the DRD2 gene, since the polymorphism is proximal to the DRD2 gene and can influence DRD2 receptor expression. It is now known to be located in the coding region of the ANKK1 gene which controls the synthesis of dopamine in the brain. The A1 allele is associated with increased activity of striatal L-amino acid decarboxylase.
A1+ allele
Hepatitis C infection
Antisocial personality disorder
Borderline personality traits
Schizoid/avoidant behavior
Given that the A1+ allele is associated with antisocial personality disorder, one may infer that the allele is also associated with narcissistic personality disorder and histrionic personality disorder. However, these predictions have not yet been empirically verified.
A1+ genotype frequencies
European population estimates for A1+ genotype frequencies range from 20.8 to 43.4% (National Center of Biotechnology Information (NCBI), identification number rs1800497).
Addictive behaviors
The ANKK1 gene is closely linked to dopamine receptor D2 (DRD2) on chromosome band 11q23.1. The A1 allele of the Taq1A polymorphism (rs1800497T), is located ≈10kb downstream of the dopamine receptor DRD2 gene. Dopamine (DA) is a neurotransmitter in the brain, which controls feelings of wellbeing. This sensation results from the interaction of dopamine and other neurotransmitters such as serotonin, the |
https://en.wikipedia.org/wiki/Sottens%20transmitter | The Sottens Transmitter is the nationwide transmitter for French-speaking Switzerland. The transmitter is located at Sottens, Canton of Vaud, Switzerland. It is run on 765 kHz with a power of 600 kilowatts and is easily receivable during the night throughout the whole of Europe. Since 1989 the aerial used has been a centre-fed dipole fixed on the outside of a 188-metre-high grounded freestanding steel framework tower. Before 1989 a 190-metre high self-radiating, free standing steel framework tower was used as a transmission aerial.
The Sottens transmitter most recently broadcast the Option Musique radio programme from Radio Suisse Romande, up until 5 December 2010.
There is also a 125 metre tall free-standing lattice tower on the site. This tower was built in 1931 as one of a pair, which until 1958 carried between them a T-antenna for medium wave broadcasting. The second tower was dismantled in that year and rebuilt in Dole as a TV transmission tower. This tower is insulated from ground to form a tower radiator and is used as backup antenna.
After the shutdown of RSR on MW, the antenna was later used for ham radio experiments in February 2011, using both standard AM and DRM in the 80 m band.
See also
Lattice tower
External links
Sottens transmitter pictures on emetteurs.ch
http://www.skyscraperpage.com/diagrams/?b44084
http://www.skyscraperpage.com/diagrams/?b57952
The Sottens transmitter Retrieved 26 January 2006
Sottens transmitter Sottens
Lattice towers
Towers in Switzerland
Buildings and structures in the canton of Vaud
Broadcast transmitters
1931 establishments in Switzerland
Towers completed in 1931
20th-century architecture in Switzerland |
https://en.wikipedia.org/wiki/Trochoidal%20wave | In fluid dynamics, a trochoidal wave or Gerstner wave is an exact solution of the Euler equations for periodic surface gravity waves. It describes a progressive wave of permanent form on the surface of an incompressible fluid of infinite depth. The free surface of this wave solution is an inverted (upside-down) trochoid – with sharper crests and flat troughs. This wave solution was discovered by Gerstner in 1802, and rediscovered independently by Rankine in 1863.
The flow field associated with the trochoidal wave is not irrotational: it has vorticity. The vorticity is of such a specific strength and vertical distribution that the trajectories of the fluid parcels are closed circles. This is in contrast with the usual experimental observation of Stokes drift associated with the wave motion. Also the phase speed is independent of the trochoidal wave's amplitude, unlike other nonlinear wave-theories (like those of the Stokes wave and cnoidal wave) and observations. For these reasons – as well as for the fact that solutions for finite fluid depth are lacking – trochoidal waves are of limited use for engineering applications.
In computer graphics, the rendering of realistic-looking ocean waves can be done by use of so-called Gerstner waves. This is a multi-component and multi-directional extension of the traditional Gerstner wave, often using fast Fourier transforms to make (real-time) animation feasible.
Description of classical trochoidal wave
Using a Lagrangian specification of the flow field, the motion of fluid parcels is – for a periodic wave on the surface of a fluid layer of infinite depth:
where and are the positions of the fluid parcels in the plane at time , with the horizontal coordinate and the vertical coordinate (positive upward, in the direction opposing gravity). The Lagrangian coordinates label the fluid parcels, with the centres of the circular orbits – around which the corresponding fluid parcel moves with constant speed Further is the w |
https://en.wikipedia.org/wiki/Strausse | A Strausse or Strausswirtschaft (also Strauße or Straußwirtschaft) is a type of wine tavern in winegrowing areas of German-speaking countries that is only open during certain times of the year. Typically it is a pub run by winegrowers and winemakers themselves, in which they sell their own wine directly to the public. The food served needs to be simple, regional cold dishes. Other expressions like Besenwirtschaft and Besenschänke ("broom pub"), Rädlewirtschaft ("cyclists' pub") as well as Hecken- or Häckerwirtschaft are also common.
Characterization
A Strausswirtschaft is essentially understood to be a winemaker serving his own wine on his own premises. These seasonal inns are not subject to normal business laws and are thus not obliged to have a licence or to pay extra taxes. They must, however, fulfil certain conditions instead. These conditions vary from state to state but they are in general agreement on certain essential points (see "Legal aspects" below).
The kinds of locations in which a Strausswirtschaft can be found can vary considerably. Besides ones furnished like ordinary pubs there are also simple barns where benches and tables have been temporarily set up to accommodate guests. In earlier decades some winegrowers even cleared their flats or the stables to run such a tavern. In Austria this kind of pub is called a Buschenschank or Heuriger. The name is derived from a bar or post to which a so-called Föhrenbusch or a Reisigbesen (a kind of besom or broom) was attached. This helps to explain another expression associated with the Strausswirtschaft: Ausg'steckt is ("It is attached"). By attaching the bar outside, the pub owner was informing the tax collector about the pub's tax liability. The Buschenschank and the expression Ausg'steckt is can be traced back to a regulation by Empress Maria Theresia.
Origins
The enactment "Capitulare de villis vel curtis imperii" by Charlemagne is often mentioned as the historical standard for the Strausswirtschaft. I |
https://en.wikipedia.org/wiki/Ecosystem%20ecology | Ecosystem ecology is the integrated study of living (biotic) and non-living (abiotic) components of ecosystems and their interactions within an ecosystem framework. This science examines how ecosystems work and relates this to their components such as chemicals, bedrock, soil, plants, and animals.
Ecosystem ecology examines physical and biological structures and examines how these ecosystem characteristics interact with each other. Ultimately, this helps us understand how to maintain high quality water and economically viable commodity production. A major focus of ecosystem ecology is on functional processes, ecological mechanisms that maintain the structure and services produced by ecosystems. These include primary productivity (production of biomass), decomposition, and trophic interactions.
Studies of ecosystem function have greatly improved human understanding of sustainable production of forage, fiber, fuel, and provision of water. Functional processes are mediated by regional-to-local level climate, disturbance, and management. Thus ecosystem ecology provides a powerful framework for identifying ecological mechanisms that interact with global environmental problems, especially global warming and degradation of surface water.
This example demonstrates several important aspects of ecosystems:
Ecosystem boundaries are often nebulous and may fluctuate in time
Organisms within ecosystems are dependent on ecosystem level biological and physical processes
Adjacent ecosystems closely interact and often are interdependent for maintenance of community structure and functional processes that maintain productivity and biodiversity
These characteristics also introduce practical problems into natural resource management. Who will manage which ecosystem? Will timber cutting in the forest degrade recreational fishing in the stream? These questions are difficult for land managers to address while the boundary between ecosystems remains unclear; even though decisions in |
https://en.wikipedia.org/wiki/Sinogene%20Biotechnology | Sinogene Biotechnology is a biotechnology company focusing on animal cloning technology. Their services include dog, cat, and horse cloning.
Sinogene Biotechnology began by offering cloning dogs in 2017 and introduced cat cloning in 2019. In 2022, they cloned an Arctic wolf, and started horse cloning in 2023. |
https://en.wikipedia.org/wiki/Arcuate%20vein | The arcuate vein is a vessel of the renal circulation. It is located at the border of the renal cortex and renal medulla. Arcuate veins pass around the renal pyramids at the border between the renal cortex and renal medulla in an arch shape. Arcuate veins receive blood from cortical radiate veins, and in turn deliver blood into the arcuate veins. |
https://en.wikipedia.org/wiki/Pathfinding | Pathfinding or pathing is the plotting, by a computer application, of the shortest route between two points. It is a more practical variant on solving mazes. This field of research is based heavily on Dijkstra's algorithm for finding the shortest path on a weighted graph.
Pathfinding is closely related to the shortest path problem, within graph theory, which examines how to identify the path that best meets some criteria (shortest, cheapest, fastest, etc) between two points in a large network.
Algorithms
At its core, a pathfinding method searches a graph by starting at one vertex and exploring adjacent nodes until the destination node is reached, generally with the intent of finding the cheapest route. Although graph searching methods such as a breadth-first search would find a route if given enough time, other methods, which "explore" the graph, would tend to reach the destination sooner. An analogy would be a person walking across a room; rather than examining every possible route in advance, the person would generally walk in the direction of the destination and only deviate from the path to avoid an obstruction, and make deviations as minor as possible.
Two primary problems of pathfinding are (1) to find a path between two nodes in a graph; and (2) the shortest path problem—to find the optimal shortest path. Basic algorithms such as breadth-first and depth-first search address the first problem by exhausting all possibilities; starting from the given node, they iterate over all potential paths until they reach the destination node. These algorithms run in , or linear time, where V is the number of vertices, and E is the number of edges between vertices.
The more complicated problem is finding the optimal path. The exhaustive approach in this case is known as the Bellman–Ford algorithm, which yields a time complexity of , or quadratic time. However, it is not necessary to examine all possible paths to find the optimal one. Algorithms such as A* and Dijkstra' |
https://en.wikipedia.org/wiki/Rain%20%28The%20Script%20song%29 | "Rain" is a song by Irish rock band The Script. It was written by James Barry, Camille Purcell, Mark Sheehan and Danny O'Donoghue, with the latter two and Andrew Frampton, Jimbo Wallace, Max Farrar handling the song's production. The song was released on 14 July 2017, through Sony Music, as the lead single from the band's fifth studio album Freedom Child.
Background
The Script described the track as a "feel-good summer tune". "After a very long process of making 'Album 5', the song 'Rain' came right at the end. It's a summer song so we thought, only The Script can make it 'Rain in Summer'" the band stated. In an interview with Metro, the band regarded the single as a progression. "I think we have afforded ourselves a little bit of leeway. The past four records have been not the same sound but we've been progressing at a slow rate. It's been two years since we had something out so there's two years worth of progression in our music. I'm sure to a lot of people it might sound quite drastic at first but if they heard the 60 songs we've put out you'd hear a slower progression." When asked if their change in sound was deliberate, they said:"When we started this record we wanted to do something where people would hear a song and go 'oh I love that song!' and then look it up and find it's us. We wanted to change our sound a push our sound a bit. It's either adapt, change or die in this industry. It's very difficult in this industry. We wanted to revamp and reboot the sound and at the end of the day we're still the same songwriters and that's never going to go away."The band said that they always "want something different in our lives" and "hoping our songs penetrate different markets and gain new fans", They referred themselves as an ambitious band. "With Rain we just decided we needed something a bit lighter, you can't just walk into a party of people and start with a heavy topic. It's nice to have a bit of escapism that people can bob their head to and not be so serious |
https://en.wikipedia.org/wiki/Gag/pol%20translational%20readthrough%20site | Gag/pol translational readthrough site (or Retroviral readthrough element) is a cis-regulatory element found in retroviruses. The readthrough site facilitates the mechanism of translation readthrough of the stop codon at the gag-pol junction producing the gag and pol fusion protein in certain retroviruses.
Retroviruses whose gag and pol genes are in the same reading frame often depend upon approximately 5% read-through of the gag UAG termination codon to form the gag-pol polyprotein. This readthrough is usually dependent on a pseudoknot located eight nucleotides downstream of the stop codon (UAG). Sequence conservation is found in the second pseudoknot loop. |
https://en.wikipedia.org/wiki/Decussation | Decussation is used in biological contexts to describe a crossing (due to the shape of the Roman numeral for ten, an uppercase 'X' (), ). In Latin anatomical terms, the form is used, e.g. .
Similarly, the anatomical term chiasma is named after the Greek uppercase 'Χ' (chi). Whereas a decussation refers to a crossing within the central nervous system, various kinds of crossings in the peripheral nervous system are called chiasma.
Examples include:
In the brain, where nerve fibers obliquely cross from one lateral side of the brain to the other, that is to say they cross at a level other than their origin. See for examples decussation of pyramids and sensory decussation. In neuroanatomy, the term chiasma is reserved for crossing of- or within nerves such as in the optic chiasm.
In botanical leaf taxology, the word decussate describes an opposite pattern of leaves which has successive pairs at right angles to each other (i.e. rotated 90 degrees along the stem when viewed from above). In effect, successive pairs of leaves cross each other. Basil is a classic example of a decussate leaf pattern.
In tooth enamel, where bundles of rods cross each other as they travel from the enamel-dentine junction to the outer enamel surface, or near to it.
In taxonomic description where decussate markings or structures occur, names such as or or otherwise in part containing "decuss..." are common, especially in the specific epithet.
Evolutionary significance
The origin of the contralateral organization, the optic chiasm and the major decussations on the nervous system of vertebrates has been a long standing puzzle to scientists. The visual map theory of Ramón y Cajal has long been popular but has been criticized for its logical inconsistence. More recently, it has been proposed that the decussations are caused by an axial twist by which the anterior head, along with the forebrain, is turned by 180° with respect to the rest of the body.
See also
Definition of types of cro |
https://en.wikipedia.org/wiki/Phagomimicry | Phagomimicry is a defensive behaviour of sea hares, in which the animal ejects a mixture of chemicals, which mimic food, and overwhelm the senses of their predator, giving the sea hare a chance to escape. The typical defence response of the sea hare to a predator is to release two chemicals - ink from the ink gland and opaline from the opaline gland. While ink creates a dark, diffuse cloud in the water which disrupts the sensory perception of the predator by acting as a smokescreen and as a decoy, the opaline, which affects the senses dealing with feeding, causes the predator to instinctively attack the cloud of chemicals as if it were indeed food. This ink is able to mimic food by having a high concentration of amino acids and other compounds that are normally found in food, and the attack behaviour of the predator allows the sea-hares the opportunity to escape.
Inking behaviour
The inking behaviour exhibited in phagomimicry is in response to predator threat. Sea hares have many natural predators such as starfish, lobsters, and other crustaceans. When threatened by a predator, phagomimicry behaviour begins. An ink solution is released from both the opaline and ink glands individually, then the compounds mix in the mantle of the sea hare to form the ink mixture. When ink is released it creates a smoke-screen like defense mechanism allowing the sea hares time to escape while also affecting the olfactory and gustation senses of their predator. Predators are tricked into thinking that they have captured their prey due to the specific chemical composition of the ink released. This induces feeding behaviours in the predator, and again gives the sea hares a better chance of escaping predation.
Opaline gland
The opaline gland is a structure resembling a bundle of grapes attached to a central canal which is composed of epithelial cells. Synthesis of the opaline substance happens in the opaline vesicles themselves, as there are only opaline vesicles and muscle cells in t |
https://en.wikipedia.org/wiki/Lingual%20luring | Lingual luring is a form of aggressive mimicry in which a predator (typically a snake or turtle) uses its tongue to fool potential prey into approaching close to what appears to be a small wriggling worm.
Lingual lures are very well developed in young alligator snapping turtles which wait underwater with their mouths open. The tongue is wriggled like a worm and fishes attracted to it are captured by a rapid snapping of the mouth.
Lingual luring has also been noted in some species of snakes including the garter snake Thamnophis atratus and species of Nerodia (including N. sipedon, N. clarkii and N. rhombifer) – here the luring is initiated by flicking the surface of water with their tongue to attract fish. Similar behaviour has been reported in the snowy egret (Egretta thula) which uses tongue flicking when foraging. The puff adder Bitis arietans uses its tongue which is loosely wriggled like a worm to attract amphibian prey. The puff adder also uses its tail as a lure (caudal luring). |
https://en.wikipedia.org/wiki/Descendants%202 | Descendants 2 is an American musical fantasy television film. It premiered as a Disney Channel Original Movie on July 21, 2017, and was simulcast on Disney-owned networks ABC, Disney XD, Freeform, Lifetime, and Lifetime Movies. It is the second installment in the Descendants franchise and the sequel to the 2015 film Descendants. The film stars Dove Cameron, Cameron Boyce, Sofia Carson, Booboo Stewart, Mitchell Hope, and China Anne McClain. A third film, Descendants 3, premiered on August 2, 2019.
Plot
Mal struggles with her new celebrity life as the girlfriend of King Ben, putting a spell on her hair making it blonde, and uncomfortably maintaining a princess-like personality. She confides her issues in her friends Evie, Carlos, and Jay, but they are content with their new lives in Auradon. Evie scolds Mal for relying on her mother's spell book to solve her issues. Carlos, wishing to ask Jane to the upcoming Cotillion dance, turns to Mal for help. Mal gives him a sweet that will make him speak the truth, but his dog, Dude, eats it, magically speaking the blunt truth in English.
Ben eventually discovers Mal's reliance on magic, and she comes clean about her insecurities, causing a falling-out. Mal returns to the Isle of the Lost, now ruled by her former rival Uma, Ursula's daughter, along with Harry and Gil, the sons of Captain Hook and Gaston. Mal visits hairstylist Dizzy Tremaine, Drizella's daughter, who restores her signature purple hair. Harry learns of Mal's return and informs Uma.
Ben, Evie, Jay, and Carlos learn of Mal's departure and sneak onto the Isle to find her, but Gil recognizes them. Ben confronts Mal, who rejects his feelings for both his sake and Auradon's. Ben leaves dejected, only to be captured by Uma. Mal and Uma arm wrestle for Ben and Fairy Godmother's wand, which Uma wins. Uma then orders Mal and her friends to retrieve the Fairy Godmother's wand in exchange for Ben's safe return. Carlos and Jay return to Auradon Prep, where they create a |
https://en.wikipedia.org/wiki/Trusted%20Computing%20Group | The Trusted Computing Group is a group formed in 2003 as the successor to the Trusted Computing Platform Alliance which was previously formed in 1999 to implement Trusted Computing concepts across personal computers. Members include Intel, AMD, IBM, Microsoft, and Cisco.
The core idea of trusted computing is to give hardware manufacturers control over what software does and does not run on a system by refusing to run unsigned software.
History
On October 11, 1999, the Trusted Computing Platform Alliance (abbreviated as TCPA), a consortium of various technology companies including Compaq, Hewlett-Packard, IBM, Intel, and Microsoft, was formed in an effort to promote trust and security in the personal computing platform. In November 1999, the TCPA announced that over 70 leading hardware and software companies joined the alliance in the first month. On January 30, 2001, version 1.0 of the Trusted Computing Platform Specifications was released IBM was the first original equipment manufacturer to incorporate hardware features based on the specifications with the introduction of its ThinkPad T30 mobile computer in 2002.
In 2003, the TCPA was succeeded by the Trusted Computing Group, with an increased emphasis on mobile devices.
Membership fees vary by level. Promoters pay annual membership fees of $30,000, contributors pay $15,000, and depending upon company size, adopters pay annual membership fees of either $2,500 or $7,500.
Overview
TCG's most successful effort was the development of a Trusted Platform Module (TPM), a semiconductor intellectual property core or integrated circuit that conforms to the specification to enable trusted computing features in computers and mobile devices. Related efforts involved Trusted Network Connect, to bring trusted computing to network connections, and Storage Core Architecture / Security Subsystem Class, to bring trusted computing to disk drives and other storage devices. These efforts have not achieved the same level of widesp |
https://en.wikipedia.org/wiki/SYN%20cookies | SYN cookie is a technique used to resist SYN flood attacks. The technique's primary inventor Daniel J. Bernstein defines SYN cookies as "particular choices of initial TCP sequence numbers by TCP servers." In particular, the use of SYN cookies allows a server to avoid dropping connections when the SYN queue fills up. Instead of storing additional connections, a SYN queue entry is encoded into the sequence number sent in the SYN+ACK response. If the server then receives a subsequent ACK response from the client with the incremented sequence number, the server is able to reconstruct the SYN queue entry using information encoded in the TCP sequence number and proceed as usual with the connection.
Implementation
In order to initiate a TCP connection, the client sends a TCP SYN packet to the server. In response, the server sends a TCP SYN+ACK packet back to the client. One of the values in this packet is a sequence number, which is used by the TCP to reassemble the data stream. According to the TCP specification, that first sequence number sent by an endpoint can be any value as decided by that endpoint. As the sequence number is chosen by the sender, returned by the recipient, and has no otherwise-defined internal structure, it can be overloaded to carry additional data. The following describes one possible implementation, however as there is no public standard to follow, the order, length, and semantics of the fields may differ between SYN cookie implementations.
SYN cookies are initial sequence numbers that are carefully constructed according to the following rules:
let t be a slowly incrementing timestamp (typically logically right-shifted 6 positions, which gives a resolution of 64 seconds)
let m be the maximum segment size (MSS) value that the server would have stored in the SYN queue entry
let s be the result of a cryptographic hash function computed over the server IP address and port number, the client IP address and port number, and the value t. The retu |
https://en.wikipedia.org/wiki/Replication%20terminator%20Tus%20family | Tus, also known as terminus utilization substance, is a protein that binds to terminator sequences and acts as a counter-helicase when it comes in contact with an advancing helicase. The bound Tus protein effectively halts DNA polymerase movement. Tus helps end DNA replication in prokaryotes.
In E. coli, Tus binds to 10 closely related sites encoded in the chromosome. Each site is 23 base pairs. The 10 sites are called Ter sites, and are designated TerA, TerB, ..., TerJ. These binding sites are asymmetric, such that when a Tus-Ter complex (Tus protein bound to a Ter site) is encountered by a replication fork from one direction, the complex is dissociated and replication continues (permissive). But when encountered from the other direction, the Tus-Ter complex provides a much larger kinetic barrier and halts replication (non-permissive). The multiple Ter sites in the chromosome are oriented such that the two oppositely moving replication forks are both stalled in the desired termination region.
Bacillus subtilis utilize replication terminator protein (RTP) instead of Tus.
Protein domains
The Ter protein contains two domains. The N-terminal domain is composed of an alpha helices, beta sheet, and three loops. The C-terminal domain is made of two alpha helices and one beta sheet.
Further reading
"Interaction of the Escherichia coli replication terminator protein (Tus) with DNA: a model derived from DNA-binding studies of mutant proteins by surface plasmon resonance."
"Replication termination in Escherichia coli: structure and antihelicase activity of the Tus-Ter complex."
"A molecular mousetrap determines polarity of termination of DNA replication in E. coli."
"Isolation and characterization of mutants of Tus, the replication arrest protein of Escherichia coli."
"Biophysical characteristics of Tus, the replication arrest protein of Escherichia coli."
"Structure of a replication-terminator protein complexed with DNA."
Structure at protein data bank |
https://en.wikipedia.org/wiki/Biosimilar | A biosimilar (also known as follow-on biologic or subsequent entry biologic) is a biologic medical product that is almost an identical copy of an original product that is manufactured by a different company. Biosimilars are officially approved versions of original "innovator" products and can be manufactured when the original product's patent expires. Reference to the innovator product is an integral component of the approval.
Unlike with generic drugs of the more common small-molecule type, biologics generally exhibit high molecular complexity and may be quite sensitive to changes in manufacturing processes. Despite that heterogeneity, all biopharmaceuticals, including biosimilars, must maintain consistent quality and clinical performance throughout their lifecycle.
Drug-related authorities such as the European Medicines Agency (EMA) of the European Union, the United States Food and Drug Administration (FDA), and the Health Products and Food Branch of Health Canada hold their own guidance on requirements for demonstration of the similar nature of two biological products in terms of safety and efficacy. According to them, analytical studies demonstrate that the biological product is highly similar to the reference product, despite minor differences in clinically inactive components, animal studies (including the assessment of toxicity), and a clinical study or studies (including the assessment of immunogenicity and pharmacokinetics or pharmacodynamics). They are sufficient to demonstrate safety, purity, and potency in one or more appropriate conditions of use for which the reference product is licensed and is intended to be used and for which licensure is sought for the biological product.
The World Health Organization (WHO) published its "Guidelines for the evaluation of similar biotherapeutic products (SBPs)" in 2009. The purpose of this guideline is to provide an international norm for evaluating biosimilars.
The EMA has granted marketing authorizations for m |
https://en.wikipedia.org/wiki/Heat%20transfer%20coefficient | In thermodynamics, the heat transfer coefficient or film coefficient, or film effectiveness, is the proportionality constant between the heat flux and the thermodynamic driving force for the flow of heat (i.e., the temperature difference, ). It is used in calculating the heat transfer, typically by convection or phase transition between a fluid and a solid. The heat transfer coefficient has SI units in watts per square meter per kelvin (W/m2/K).
The overall heat transfer rate for combined modes is usually expressed in terms of an overall conductance or heat transfer coefficient, . In that case, the heat transfer rate is:
where (in SI units):
: surface area where the heat transfer takes place (m2)
: temperature of the surrounding fluid (K)
: temperature of the solid surface (K)
The general definition of the heat transfer coefficient is:
where:
: heat flux (W/m2); i.e., thermal power per unit area,
: difference in temperature between the solid surface and surrounding fluid area (K)
The heat transfer coefficient is the reciprocal of thermal insulance. This is used for building materials (R-value) and for clothing insulation.
There are numerous methods for calculating the heat transfer coefficient in different heat transfer modes, different fluids, flow regimes, and under different thermohydraulic conditions. Often it can be estimated by dividing the thermal conductivity of the convection fluid by a length scale. The heat transfer coefficient is often calculated from the Nusselt number (a dimensionless number). There are also online calculators available specifically for Heat-transfer fluid applications. Experimental assessment of the heat transfer coefficient poses some challenges especially when small fluxes are to be measured (e.g. ).
Composition
A simple method for determining an overall heat transfer coefficient that is useful to find the heat transfer between simple elements such as walls in buildings or across heat exchangers is shown below. Note that |
https://en.wikipedia.org/wiki/DejaGnu | DejaGnu is a software framework for testing other programs. It has a main script called runtest that goes through a directory looking at configuration files and then runs some tests with given criteria. The purpose of the DejaGnu package is to provide a single front end for all tests. It is a part of the GNU Project and is licensed under the GPL. It is based on Expect, which is in turn based on Tcl. The current maintainers are Rob Savoye and Ben Elliston.
Testing
DejaGnu has a very strong history in testing due to its Tcl base. Tcl is used extensively by companies such as Oracle and Sybase to test their products. DejaGnu allows this work to be much more structured.
The tests can be grouped according to the tool they are testing. The test is run by merely calling in the root project directory.
runtest --tool program_to_test
This will look in the directory for any folders starting with and will run all .exp files in that folder.
Embedded design
One field for which DejaGnu is particularly well suited is that of embedded system design. It allows for testing to be done remotely on development boards; separate initialization files can be created for each operating system and board. This mainly focuses on embedded targets and remote hosts. DejaGnu is thus popular with many GNU projects, at universities, and for private companies.
Files
Essential Files
Each directory in testsuite should contain tests for a specific tool. In this example, the tool being tested is the Apache webserver.
This will be the file containing tests, which in this fictional case might change configuration options, and then connect to the network and check to make sure the changes have taken effect.
This file will be run as a tool init file for the tool called toolname.
Other Files
This file is a directory specific configuration file for . Options can be placed in this file rather than retyped on each invocation; these options can include any variable passed as a command line arg |
https://en.wikipedia.org/wiki/99%20%28number%29 | 99 (ninety-nine) is the natural number following 98 and preceding 100.
In mathematics
99 is:
a composite number; a square-prime, of the form (p2, q). It is the 11th composite number of this form and the third of the form (32, q). It has an aliquot sum of 57, within an aliquot sequence of two composite numbers (99,57,23,1,0), to the Prime in the 23-aliquot tree.
a Kaprekar number
a lucky number
a palindromic number
the ninth repdigit
the sum of the cubes of three consecutive integers: 99 = 23 + 33 + 43
the sum of the sums of the divisors of the first 11 positive integers.
the highest two digit number in decimal.
In music
"99", a song by the band Toto on the Hydra album.
"99 Bottles of Beer", a counting song.
"99 Luftballons", a German-language song by the band Nena.
"99 Problems", a song by Jay-Z on The Black Album.
"99 Ways to Die", a song by Megadeth on the Hidden Treasures EP.
In other fields
The atomic number of einsteinium, an actinide.
".99" is frequently used as a price ender in pricing.
99, the jersey number of Wayne Gretzky, since permanently retired |
https://en.wikipedia.org/wiki/Refraction%20microtremor | Refraction microtremor (ReMi) is a surface-performed geophysical survey developed by Dr. John Louie (and others) based on previously existing principles of evaluating surface waves and in particular Rayleigh waves. The refraction microtremor technology was developed at the University of Nevada and is owned by the State of Nevada. Optim of Reno, Nevada has the exclusive license to develop the technology, and SeisOpt® ReMi™ has been available commercially from Optim since 2004. Since Rayleigh waves are dispersive, the propagating waves are measured along a linear seismic array and evaluated relative to wave frequency and slowness (or the inverse of the velocity). Due to the dispersive characteristics of higher frequency waves travelling through the more shallow conditions and lower frequency waves passing through deeper materials, a 1-D subsurface profile can be generated based on the velocity with depth.
Equipment and Field Procedures
The method utilizes equipment typically employed in seismic refraction surveys. This equipment consists of a seismograph, geophones placed in an array, and a seismic source. An array consists of 6 to 48+ geophones placed at some interval along the ground surface to measure the propagating waves and are connected to a seismograph that records the data. Rayleigh waves are generated from seismic sources described to be active, passive, or a combination of both. For example, active sources can be generated sources from a sledge hammer striking a plate or other sources that are created near the traverse. Passive sources can be nearby highway traffic, construction equipment working in the distance, etc. Several recordings (typically 15 to 60+ seconds long) are captured and stored for later analysis. As multiple traverses are performed along the ground surface, the 1-D profiles can be compiled to generate a 2-D profile of the subsurface conditions. It is worth noting that since seismic traverses typically "average" conditions along the traver |
https://en.wikipedia.org/wiki/Logarithmic%20norm | In mathematics, the logarithmic norm is a real-valued functional on operators, and is derived from either an inner product, a vector norm, or its induced operator norm. The logarithmic norm was independently introduced by Germund Dahlquist and Sergei Lozinskiĭ in 1958, for square matrices. It has since been extended to nonlinear operators and unbounded operators as well. The logarithmic norm has a wide range of applications, in particular in matrix theory, differential equations and numerical analysis. In the finite-dimensional setting, it is also referred to as the matrix measure or the Lozinskiĭ measure.
Original definition
Let be a square matrix and be an induced matrix norm. The associated logarithmic norm of is defined
Here is the identity matrix of the same dimension as , and is a real, positive number. The limit as equals , and is in general different from the logarithmic norm , as for all matrices.
The matrix norm is always positive if , but the logarithmic norm may also take negative values, e.g. when is negative definite. Therefore, the logarithmic norm does not satisfy the axioms of a norm. The name logarithmic norm, which does not appear in the original reference, seems to originate from estimating the logarithm of the norm of solutions to the differential equation
The maximal growth rate of is . This is expressed by the differential inequality
where is the upper right Dini derivative. Using logarithmic differentiation the differential inequality can also be written
showing its direct relation to Grönwall's lemma. In fact, it can be shown that the norm of the state transition matrix associated to the differential equation is bounded by
for all .
Alternative definitions
If the vector norm is an inner product norm, as in a Hilbert space, then the logarithmic norm is the smallest number such that for all
Unlike the original definition, the latter expression also allows to be unbounded. Thus differential operators too can have |
https://en.wikipedia.org/wiki/Volcano%20warning%20schemes%20of%20the%20United%20States | In October 2006, the United States Geological Survey (USGS) adopted a nationwide alert system for characterizing the level of unrest and eruptive activity at volcanoes. The system is now used by the Alaska Volcano Observatory, the California Volcano Observatory (California and Nevada), the Cascades Volcano Observatory (Washington, Oregon and Idaho), the Hawaiian Volcano Observatory and the Yellowstone Volcano Observatory (Montana, Wyoming, Colorado, Utah, New Mexico and Arizona).
Under this system, the USGS ranks the level of activity at a U.S. volcano using the terms "normal", for typical volcanic activity in a non-eruptive phase; "advisory", for elevated unrest; "watch", for escalating unrest or an eruption underway that poses limited hazards; and, "warning", if a highly hazardous eruption is underway or imminent. These levels reflect conditions at a volcano and the expected or ongoing hazardous volcanic phenomena. When an alert level is assigned by an observatory, accompanying text will give a fuller explanation of the observed phenomena and clarify hazard implications to affected groups.
Summary of Volcanic Activity Alert Notification System
Aviation color codes
Earlier volcano warning schemes for the United States
Prior to October 2006, three parallel Volcano warning schemes were used by the United States Geological Survey and the volcano observatories for different volcano ranges in the United States. They each have a base level for dormant-quiescent states and three grades of alert.
Color Code Conditions, Long Valley Caldera and Mono-Inyo Craters Region, California
Developed in 1997 to replace a previous 5-level system devised in 1991.
Level of Concern Color Codes for volcanoes in Alaska
The Alaska Volcano Observatory (AVO) used the following color-coded system to rate volcanic activity. It was originally established during the 1989-90 eruption of Redoubt Volcano.
All five classifications are spelled as proper nouns, i.e., Level of Concern Color Cod |
https://en.wikipedia.org/wiki/Banner%20of%20arms | A banner of arms is a type of heraldic flag, characterised by sharing its imagery with that of the coat of arms (i.e. the shield of a full heraldic achievement, rendered in a square or rectangular shape of the flag).
The term is derived from the terminology of heraldry but mostly used in vexillology. Examples of modern national flags which are banners of arms are the flags of Austria, Iraq, and Switzerland.
The banner of arms is sometimes simply called a banner, but a banner is in a more strict sense a one of a kind personal flag of a nobleman held in battle.
Examples
National flags
Subdivision flags
County flags
City flags
Organization flags
Notes |
https://en.wikipedia.org/wiki/Sensorimotor%20network | The sensorimotor network (SMN), also known as somatomotor network, is a large-scale brain network that primarily includes somatosensory (postcentral gyrus) and motor (precentral gyrus) regions and extends to the supplementary motor areas (SMA). The auditory cortex may also be included. The SMN is activated during motor tasks, such as finger tapping, indicating that the network readies the brain when performing and coordinating motor tasks.
Clinical significance
Dysfunction in the SMN has been implicated in various neuropsychiatric disorders.
Bipolar Disorder: The psychomotor disturbances that characterize the depressive and manic phases of bipolar disorder may be related to dysfunction in the sensorimotor network (SMN) and its balance with other large-scale networks such as the default mode network.
Amyotrophic Lateral Sclerosis: Altered functional connectivity patterns in the SMN may contribute to various symptoms in the neurodegenerative disease .
Nomenclature
In 2019, Uddin et al. proposed that pericentral network (PN) be used as a standard anatomical name for the network. |
https://en.wikipedia.org/wiki/Quantum%20foundations | Quantum foundations is a discipline of science that seeks to understand the most counter-intuitive aspects of quantum theory, reformulate it and even propose new generalizations thereof. Contrary to other physical theories, such as general relativity, the defining axioms of quantum theory are quite ad hoc, with no obvious physical intuition. While they lead to the right experimental predictions, they do not come with a mental picture of the world where they fit.
There exist different approaches to resolve this conceptual gap:
First, one can put quantum physics in contraposition with classical physics: by identifying scenarios, such as Bell experiments, where quantum theory radically deviates from classical predictions, one hopes to gain physical insights on the structure of quantum physics.
Second, one can attempt to find a re-derivation of the quantum formalism in terms of operational axioms.
Third, one can search for a full correspondence between the mathematical elements of the quantum framework and physical phenomena: any such correspondence is called an interpretation.
Fourth, one can renounce quantum theory altogether and propose a different model of the world.
Research in quantum foundations is structured along these roads.
Non-classical features of quantum theory
Quantum nonlocality
Two or more separate parties conducting measurements over a quantum state can observe correlations which cannot be explained with any local hidden variable theory. Whether this should be regarded as proving that the physical world itself is "nonlocal" is a topic of debate, but the terminology of "quantum nonlocality" is commonplace. Nonlocality research efforts in quantum foundations focus on determining the exact limits that classical or quantum physics enforces on the correlations observed in a Bell experiment or more complex causal scenarios. This research program has so far provided a generalization of Bell's theorem that allows falsifying all classical theories wi |
https://en.wikipedia.org/wiki/Sexual%20anomalies | Sexual anomalies, also known as sexual abnormalities, are a set of clinical conditions due to chromosomal, gonadal and/or genitalia variation. Individuals with congenital (inborn) discrepancy between sex chromosome, gonadal, and their internal and external genitalia are categorised as individuals with a disorder of sex development (DSD). Afterwards, if the family or individual wishes, they can partake in different management and treatment options for their conditions (e.g. hormone therapy).
Infants born with atypical genitalia often cause confusion and distress for the family. Psychosexual development is influenced by numerous factors that include, but are not limited to, gender differences in brain structure, genes associated with sexual development, prenatal androgen exposure, interactions with family, and cultural and societal factors. Because of the complex and multifaceted factors involved, communication and psychosexual support are all important.
A team of experts, or patient support groups, are usually recommended for cases related to sexual anomalies. This team of experts are usually derived from a variety of disciplines including pediatricians, neonatologists, pediatric urologists, pediatric general surgeons, endocrinologists, geneticists, radiologists, psychologists and social workers. These professionals are capable of providing first line (prenatal) and second line diagnostic (postnatal) tests to examine and diagnose sexual anomalies.
Overview
In the normal prenatal stages of fetal development, the fetus is exposed to testosterone - albeit more in male fetuses than female ones. Upon the presence of the 5α-reductase enzyme, testosterone is converted to dihydrotestosterone (i.e. DHT). If DHT is present, the male external genitalia will develop.
Development of male external genitalia:
Genital tubercle forms the penis
Urethral folds forms the urethra
Genital swelling forms the scrotum
On the other hand, if maternal placenta estrogen is present wit |
https://en.wikipedia.org/wiki/GC-content | In molecular biology and genetics, GC-content (or guanine-cytosine content) is the percentage of nitrogenous bases in a DNA or RNA molecule that are either guanine (G) or cytosine (C). This measure indicates the proportion of G and C bases out of an implied four total bases, also including adenine and thymine in DNA and adenine and uracil in RNA.
GC-content may be given for a certain fragment of DNA or RNA or for an entire genome. When it refers to a fragment, it may denote the GC-content of an individual gene or section of a gene (domain), a group of genes or gene clusters, a non-coding region, or a synthetic oligonucleotide such as a primer.
Structure
Qualitatively, guanine (G) and cytosine (C) undergo a specific hydrogen bonding with each other, whereas adenine (A) bonds specifically with thymine (T) in DNA and with uracil (U) in RNA. Quantitatively, each GC base pair is held together by three hydrogen bonds, while AT and AU base pairs are held together by two hydrogen bonds. To emphasize this difference, the base pairings are often represented as "G≡C" versus "A=T" or "A=U".
DNA with low GC-content is less stable than DNA with high GC-content; however, the hydrogen bonds themselves do not have a particularly significant impact on molecular stability, which is instead caused mainly by molecular interactions of base stacking. In spite of the higher thermostability conferred to a nucleic acid with high GC-content, it has been observed that at least some species of bacteria with DNA of high GC-content undergo autolysis more readily, thereby reducing the longevity of the cell per se. Because of the thermostability of GC pairs, it was once presumed that high GC-content was a necessary adaptation to high temperatures, but this hypothesis was refuted in 2001. Even so, it has been shown that there is a strong correlation between the optimal growth of prokaryotes at higher temperatures and the GC-content of structural RNAs such as ribosomal RNA, transfer RNA, and many |
https://en.wikipedia.org/wiki/Rhizophora%20%C3%97%20lamarckii | Rhizophora × lamarckii is a hybrid of Rhizophora apiculata and Rhizophora stylosa. Found in the Indo-West Pacific region within the Indomalaya biome in the Sunda Shelf mangroves ecoregion, the hybrid is widespread and shares many characters of its parents.
Description
The tree grows up to 25 meters in height. The bark is gray with fissures. Leaf blade is obovate elliptic with light green midrib and dense black dots at under surface. Leaf apex has a clear mucronate spike and leaf stalk is 2–3 cm long. Inflorescence is branched 1-2 times with 2-4 buds borne on short peduncle. Petals are white and slightly hairy with style 2–3 mm long.
Chemistry
Analysis using reversed phase high performance liquid chromatography showed that there was great similarity in chemical constituents of leaves of the hybrid compared to those of its parents (Chan & Wong, 2009). Peak areas were intermediate in R. × lamarckii with higher values in R. stylosa and lower values in R. apiculata. One of the compounds identified in the leaves of the hybrid and its parents was rutin. |
https://en.wikipedia.org/wiki/Hilbert%20series%20and%20Hilbert%20polynomial | In commutative algebra, the Hilbert function, the Hilbert polynomial, and the Hilbert series of a graded commutative algebra finitely generated over a field are three strongly related notions which measure the growth of the dimension of the homogeneous components of the algebra.
These notions have been extended to filtered algebras, and graded or filtered modules over these algebras, as well as to coherent sheaves over projective schemes.
The typical situations where these notions are used are the following:
The quotient by a homogeneous ideal of a multivariate polynomial ring, graded by the total degree.
The quotient by an ideal of a multivariate polynomial ring, filtered by the total degree.
The filtration of a local ring by the powers of its maximal ideal. In this case the Hilbert polynomial is called the Hilbert–Samuel polynomial.
The Hilbert series of an algebra or a module is a special case of the Hilbert–Poincaré series of a graded vector space.
The Hilbert polynomial and Hilbert series are important in computational algebraic geometry, as they are the easiest known way for computing the dimension and the degree of an algebraic variety defined by explicit polynomial equations. In addition, they provide useful invariants for families of algebraic varieties because a flat family has the same Hilbert polynomial over any closed point . This is used in the construction of the Hilbert scheme and Quot scheme.
Definitions and main properties
Consider a finitely generated graded commutative algebra over a field , which is finitely generated by elements of positive degree. This means that
and that .
The Hilbert function
maps the integer to the dimension of the -vector space . The Hilbert series, which is called Hilbert–Poincaré series in the more general setting of graded vector spaces, is the formal series
If is generated by homogeneous elements of positive degrees , then the sum of the Hilbert series is a rational fraction
where is a polynomial |
https://en.wikipedia.org/wiki/Dependency%20need | Dependency need is "the vital, originally infantile needs for mothering, love, affection, shelter, protection, security, food, and warmth." (Segen, 1992)
A dependency need is thought to be characterized by two components: (1) It is a real need of an organism, something that must be present in order for the organism to be able to thrive, (2) It is something that an individual cannot provide for him or herself. It is well known that infants have many dependency needs; some of these needs are obvious, others have only come to the attention of researchers as the result of epidemiological studies. The more obvious needs of infants include: adequate feeding, adequate watering, adequate cleaning, adequate shelter, and more specifically, keeping the infant's body temperature within the narrow range of normalcy.
On the other hand, it was not well known until the middle of the 20th century that infants also required the presence of warmth and affection, known as "maternal warmth". The greatest number of dependency needs seem to be encompassed in infancy, but dependency needs begin to change and decrease with age and maturity. This marked decrease in dependency needs as an individual gets older can be largely attributed to the notion that, as an individual gets older, he or she becomes capable of providing these things for him or herself. To some extent, these needs remain present even into adulthood. Even as adults, people have certain universal dependency needs that remain constant throughout the lifespan that they are not able to provide for themselves; these include: the need to belong, need for affection, as well as the need for emotional support. These needs can usually be met by partnership, in which both partners get used to depending on one another. If adults lack partnership, their needs can usually be met by family and/or friend relationships.
Importance
Dependency need is an important psychological concept, encompassing the fields of psychological, evolutiona |
https://en.wikipedia.org/wiki/Compressed%20sensing%20in%20speech%20signals | In communications technology, the technique of compressed sensing (CS) may be applied to the processing of speech signals under certain conditions. In particular, CS can be used to reconstruct a sparse vector from a smaller number of measurements, provided the signal can be represented in sparse domain. "Sparse domain" refers to a domain in which only a few measurements have non-zero values.
Theory
Suppose a signal can be represented in a domain where only coefficients out of (where ) are non-zero, then the signal is said to be sparse in that domain.
This reconstructed sparse vector can be used to construct back the original signal if the sparse domain of signal is known. CS can be applied to speech signal only if sparse domain of speech signal is known.
Consider a speech signal , which can be represented in a domain such that , where speech signal , dictionary matrix and the sparse coefficient vector . This speech signal is said to be sparse in domain , if the number of significant (non zero) coefficients in sparse vector is , where .
The observed signal is of dimension . To reduce the complexity for solving using CS speech signal is observed using a measurement matrix such that
where , and measurement matrix such that .
Sparse decomposition problem for eq. 1 can be solved as standard minimization as
If measurement matrix satisfies the restricted isometric property (RIP) and is incoherent with dictionary matrix . then the reconstructed signal is much closer to the original speech signal.
Different types of measurement matrices like random matrices can be used for speech signals.
Estimating the sparsity of a speech signal is a problem since the speech signal varies greatly over time and thus sparsity of speech signal also varies highly over time. If sparsity of speech signal can be calculated over time without much complexity that will be best. If this is not possible then worst-case scenario for sparsity can be considered for |
https://en.wikipedia.org/wiki/Red-eared%20slider%20%C3%97%20yellow-bellied%20slider | The red-eared slider × yellow-bellied slider (Trachemys scripta elegans × Trachemys scripta scripta) is an intergradation of a red-eared slider and yellow-bellied slider subspecies. On the side of its head where a normal yellow-bellied slider would have solid yellow, there will be red interposed. The bottom of the shell is a bright yellow with varying numbers of black dots. In the wild it eats minnows, plants, carrion, snails and insects that fall in the water. The average lifespan is usually over 30 years.
The yellow facial markings of the yellowbelly are more distinct in this species but the red ear can still be seen
This species often behaves similarly to the red eared slider |
https://en.wikipedia.org/wiki/Boreal%20ecosystem | A boreal ecosystem is an ecosystem with a subarctic climate located in the Northern Hemisphere, approximately between 50° to 70°N latitude. These ecosystems are commonly known as taiga and are located in parts of North America, Europe, and Asia. The ecosystems that lie immediately to the south of boreal zones are often called hemiboreal. There are a variety of processes and species that occur in these areas as well.
The Köppen symbols of boreal ecosystems are Dfc, Dwc, Dfd, and Dwd.
Boreal ecosystems are some of the most vulnerable to climate change. Both loss of permafrost, reductions in cold weather and increases in summer heat cause significant changes to ecosystems, displacing cold-adapted species, increasing forest fires, and making ecosystems vulnerable to changing to other ecosystem types. These changes can cause Climate change feedback cycles, where thawing permafrost and changing ecosystems release more greenhouse gas emissions into the atmosphere causing more climate change.
Boreal Species
The species within boreal ecosystems varies as it consists of both terrestrial and aquatic habitats. The species composition include many generalized and less specialized feeders. From the equator to the poles, species richness decreases, and there is a negative relationship with species richness changes as climate changes.
However, despite not being as biodiverse as tropical systems, this area has a variety of species. Boreal ecosystems are filled with a multitude of flora species from black and white spruce, to willows, wildflowers, and alders. Caribou, although not there year round, come down and into these regions during the winter to forage for lichen. A few fish species include salmonids, smelts, sticklebacks, lamprey and sculpins. For salmon these systems are vital: relying on the riparian systems within boreal ecosystems for multiple life stages in both the beginning and the end of their life cycle, sockeye rely on the provided freshwater environments as |
https://en.wikipedia.org/wiki/Alsys | Alsys, SA. (founded 1980, merged 1995) was a software development company created to support initial work on the Ada programming language.
In July 1995, Alsys merged to become Thomson Software Products (TSP), which merged into Aonix in 1996.
History
Alsys SA. the French company was founded in 1980 by Jean Ichbiah (1940–2007). Also in 1980 the American subsidiary Alsys Inc was formed with Ben Brosgol (from Intermetrics), and Pascal Clève.
In 1985 a British subsidiary, Alsys ltd was formed with John Barnes as the MD.
During the merger mania of the 1990s, Alsys was repositioned via a series of mergers.
In 1991 Alsys was acquired by Thomson-CSF.
In November 1992, Thomson-CSF acquired TeleSoft and merged it with Alsys.
In July 1995, Thomson-CSF merged two of their subsidiaries, Alsys and MUST Software, a software development corporation based in Norwalk, Connecticut, to form Thomson Software Products (TSP).
In November 1996, TSP merged with IDE (Interactive Development Environments, Inc.) to form Aonix.
Thomson-CSF (now known as Thales), sold Aonix to Gore Technology Group (GTG) in the late 1990s.
Aonix acquired Select Software in 2001.
In January 2003, GTG sold the Critical Development Solutions (CDS) division of Aonix, which included the Alsys, Telesoft and IDE product lines, to a group of French investors. The name Aonix was kept for this new company, while Select Business Solutions was the name given to the part under Gore control.
In 2003, Aonix acquired NewMonics of Tucson, Arizona, a supplier of Java-compliant virtual machines for embedded and real-time systems.
In January, 2010 Aonix merged with Artisan Software Tools to form Atego.
Alsys was one of the few companies that developed products that unleashed the protected mode of the 80286 processor. At the time, most applications were limited to using only 640K of memory. With the Ada compiler, applications could be built using up to 16MB of memory.
Notes |
https://en.wikipedia.org/wiki/Copenhagen%20interpretation | The Copenhagen interpretation is a collection of views about the meaning of quantum mechanics, stemming from the work of Niels Bohr, Werner Heisenberg, Max Born, and others. The term "Copenhagen interpretation" was apparently coined by Heisenberg during the 1950s to refer to ideas developed in the 1925–1927 period, glossing over his disagreements with Bohr. Consequently, there is no definitive historical statement of what the interpretation entails. Features common across versions of the Copenhagen interpretation include the idea that quantum mechanics is intrinsically indeterministic, with probabilities calculated using the Born rule, and the principle of complementarity, which states that objects have certain pairs of complementary properties that cannot all be observed or measured simultaneously. Moreover, the act of "observing" or "measuring" an object is irreversible, and no truth can be attributed to an object except according to the results of its measurement (that is, the Copenhagen interpretation rejects counterfactual definiteness). Copenhagen-type interpretations hold that quantum descriptions are objective, in that they are independent of physicists' personal beliefs and other arbitrary mental factors.
Over the years, there have been many objections to aspects of Copenhagen-type interpretations, including the discontinuous and stochastic nature of the "observation" or "measurement" process, the apparent subjectivity of requiring an observer, the difficulty of defining what might count as a measuring device, and the seeming reliance upon classical physics in describing such devices. Still, including all the variations, the interpretation remains one of the most commonly taught.
Background
Starting in 1900, investigations into atomic and subatomic phenomena forced a revision to the basic concepts of classical physics. However, it was not until a quarter-century had elapsed that the revision reached the status of a coherent theory. During the intervening |
https://en.wikipedia.org/wiki/Immune%20receptor | An immune receptor (or immunologic receptor) is a receptor, usually on a cell membrane, which binds to a ligand (usually another protein, such as cytokine) and causes a response in the immune system.
Types
The main receptors in the immune system are pattern recognition receptors (PRRs), Toll-like receptors (TLRs), killer activated and killer inhibitor receptors (KARs and KIRs), complement receptors, Fc receptors, B cell receptors and T cell receptors.
See also
Antigen |
https://en.wikipedia.org/wiki/PYGO1 | Pygopus homolog 1 (Drosophila) is a protein in humans that is encoded by the PYGO1 gene. |
https://en.wikipedia.org/wiki/Georges%20Giraud | Georges Julien Giraud (22 July 1889 – 16 March 1943) was a French mathematician, working in potential theory, partial differential equations, singular integrals and singular integral equations: he is mainly known for his solution of the regular oblique derivative problem and also for his extension to –dimensional () singular integral equations of the concept of symbol of a singular integral, previously introduced by Solomon Mikhlin.
Biography
Honors
The scientific work of Georges Giraud was widely acknowledged and earned him several prizes, mainly, but not exclusively, awarded him by the French Academy of Sciences: he was seven times recipient of academy prizes.
In 1919, he was awarded the "Prix Francœur" for his work on the theory of automorphic functions: the members of the commission who examined his work and nominated him were Camille Jordan, Paul Appell, Marie Georges Humbert, Jacques Hadamard, Édouard Goursat, Joseph Boussinesq, Léon Lecornu and Emile Picard (the relator). For the same motivation, On 17 December 1923 he was awarded the "Gustave Roux" prize.
In 1924 he won the Hirn Foundation Prize, for his whole scientific work: he won again the same prize in 1935, for his work on singularities of boundary value problems in the theory of partial differential equations.
In 1928 Giraud won the "Grand Prix des sciences Mathématiques" for his work in the theory of partial differential equations: for the same motivation, in 1930 he was also awarded the "Prix Houllevigue". In the same year, he was also awarded the prize of the Lasserre foundation.
In 1933 he was recipient of the Prix Saintour, for his work on partial differential and integral equations.
Finally, in 1935, apart from winning the Hirn foundation prize for a second time, he was awarded the prize of the Annali della Reale Scuola Normale Superiore di Pisa, equally divided between him, Guido Ascoli and Pietro Buzano: the members of the jury who awarded ex-aequo the prize were Guido Fubini, Mauro Pi |
https://en.wikipedia.org/wiki/Nonspecific%20immune%20cell | A non-specific immune cell is an immune cell (such as a macrophage, neutrophil, or dendritic cell) that responds to many antigens, not just one antigen. Non-specific immune cells function in the first line of defense against infection or injury. The innate immune system is always present at the site of infection and ready to fight the bacteria; it can also be referred to as the "natural" immune system. The cells of the innate immune system do not have specific responses and respond to each foreign invader using the same mechanism.
The innate immune system
There are two categories to which parts of the immune system are assigned: the non-specific, or innate immune system and the adaptive immune system. The non-specific response is a generalized response to pathogen infections involving the use of several white blood cells and plasma proteins. Non-specific immunity, or innate immunity, is the immune system with which you were born, made up of phagocytes and barriers. Phagocytosis, derived from the Greek words , meaning to eat, or cell, and “osis” meaning process, was first described by Élie Metchnikoff, who won the Nobel Prize 100 years ago. Phagocytosis involves the internalization of solids, such as bacteria, by an organism.
Macrophages, neutrophils, and dendritic cells are all cells of the innate immune system that utilize phagocytosis and are equipped with Toll-like receptors (TLR). Toll-like receptors are present on each of these cells and recognize a variety of microbial products resulting in the induction of more specific immune responses. When a phagocytic cell engulfs bacteria, a phagosome is formed around it and the entire complex is ultimately trafficked to the lysosome for degradation. These cells that participate in the non-specific immune system response do not differentiate between types of microorganisms but do have the ability to discern between what is self and what is non-self. The cells of this system are known as non-specific immune cell |
https://en.wikipedia.org/wiki/Quantum%20crystallography | Quantum crystallography is a branch of crystallography that investigates crystalline materials within the framework of quantum mechanics, with analysis and representation, in position or in momentum space, of quantities like wave function, electron charge and spin density, density matrices and all properties related to them (like electric potential, electric or magnetic moments, energy densities, electron localization function, one electron potential, etc.).
Like the quantum chemistry, Quantum crystallography involves both experimental and computational work. The theoretical part of quantum crystallography is based on quantum mechanical calculations of atomic/molecular/crystal wave functions, density matrices or density models, used to simulate the electronic structure of a crystalline material. While in quantum chemistry, the experimental works mainly rely on spectroscopy, in quantum crystallography the scattering techniques (X-rays, neutrons, γ-Rays, electrons) play the central role, although spectroscopy as well as atomic microscopy are also sources of information.
The connection between crystallography and quantum chemistry has always been very tight, after X-ray diffraction techniques became available in crystallography. In fact, the scattering of radiation enables mapping the one-electron distribution or the elements of a density matrix.
The kind of radiation and scattering determines the quantity which is represented (electron charge or spin) and the space in which it is represented (position or momentum space).
Although the wave function is typically assumed not to be directly measurable, recent advances enable also to compute wave functions that are restrained to some experimentally measurable observable (like the scattering of a radiation).
The term Quantum Crystallography was first introduced in revisitation articles by L. Huang, L. Massa and Nobel Prize winner Jerome Karle, who associated it with two mainstreams: a) crystallographic information that |
https://en.wikipedia.org/wiki/Chmod | In Unix and Unix-like operating systems, is the command and system call used to change the access permissions and the special mode flags (the setuid, setgid, and sticky flags) of file system objects (files and directories). Collectively these were originally called its modes, and the name was chosen as an abbreviation of change mode.
History
A command first appeared in AT&T UNIX version 1, along with the system call.
As systems grew in number and types of users, access-control lists were added to many file systems in addition to these most basic modes to increase flexibility.
The version of bundled in GNU coreutils was written by David MacKenzie and Jim Meyering. The command is available as a separate package for Microsoft Windows as part of the UnxUtils collection of native Win32 ports of common GNU Unix-like utilities. The command has also been ported to the IBM i operating system.
Command syntax
Throughout this section, ser refers to the owner of the file, as a reminder that the symbolic form of the command uses "u".
chmod [options] mode[,mode] file1 [file2 ...]
Usually implemented options include:
Recursive, i.e. include objects in subdirectories.
verbose, show objects changed (unchanged objects are not shown).
If a symbolic link is specified, the target object is affected. File modes directly associated with symbolic links themselves are typically not used.
To view the file mode, the or commands may be used:
$ ls -l findPhoneNumbers.sh
-rwxr-xr-- 1 dgerman staff 823 Dec 16 15:03 findPhoneNumbers.sh
$ stat -c %a findPhoneNumbers.sh
754
The , , and specify the read, write, and execute access (the first character of the display denotes the object type; a hyphen represents a plain file). The script can be read, written to, and executed by the user ; read and executed by members of the group; and only read by any other users.
The main parts of the permissions:
For example:
Each group of three characters define permissions for each c |
https://en.wikipedia.org/wiki/Penrose%20interpretation | The Penrose interpretation is a speculation by Roger Penrose about the relationship between quantum mechanics and general relativity. Penrose proposes that a quantum state remains in superposition until the difference of space-time curvature attains a significant level.
Overview
Penrose's idea is inspired by quantum gravity, because it uses both the physical constants and . It is an alternative to the Copenhagen interpretation, which posits that superposition fails when an observation is made (but that it is non-objective in nature), and the many-worlds interpretation, which states that alternative outcomes of a superposition are equally "real", while their mutual decoherence precludes subsequent observable interactions.
Penrose's idea is a type of objective collapse theory. For these theories, the wavefunction is a physical wave, which experiences wave function collapse as a physical process, with observers not having any special role. Penrose theorises that the wave function cannot be sustained in superposition beyond a certain energy difference between the quantum states. He gives an approximate value for this difference: a Planck mass worth of matter, which he calls the "'one-graviton' level". He then hypothesizes that this energy difference causes the wave function to collapse to a single state, with a probability based on its amplitude in the original wave function, a procedure derived from standard quantum mechanics. Penrose's "'one-graviton' level" criterion forms the basis of his prediction, providing an objective criterion for wave function collapse. Despite the difficulties of specifying this in a rigorous way, he proposes that the basis states into which the collapse takes place are mathematically described by the stationary solutions of the Schrödinger–Newton equation.
Recent work indicates an increasingly deep inter-relation between quantum mechanics and gravitation.
Physical consequences
Accepting that wavefunctions are physically real, Penrose |
https://en.wikipedia.org/wiki/Borromean%20nucleus | A Borromean nucleus is an atomic nucleus comprising three bound components in which any subsystem of two components is unbound. This has the consequence that if one component is removed, the remaining two comprise an unbound resonance, so that the original nucleus is split into three parts.
The name is derived from the Borromean rings, a system of three linked rings in which no pair of rings is linked.
Examples of Borromean nuclei
Many Borromean nuclei are light nuclei near the nuclear drip lines that have a nuclear halo and low nuclear binding energy. For example, the nuclei , , and each possess a two-neutron halo surrounding a core containing the remaining nucleons. These are Borromean nuclei because the removal of either neutron from the halo will result in a resonance unbound to one-neutron emission, whereas the dineutron (the particles in the halo) is itself an unbound system. Similarly, is a Borromean nucleus with a two-proton halo; both the diproton and are unbound.
Additionally, is a Borromean nucleus comprising two alpha particles and a neutron; the removal of any one component would produce one of the unbound resonances , , or .
Several Borromean nuclei such as and the Hoyle state (an excited resonance in ) play an important role in nuclear astrophysics. Namely, these are three-body systems whose unbound components (formed from ) are intermediate steps in the triple-alpha process; this limits the rate of production of heavier elements, for three bodies must react nearly simultaneously.
Borromean nuclei consisting of more than three components can also exist. These also lie along the drip lines; for instance, is a five-body Borromean system with a four-neutron halo. It is also possible that nuclides produced in the alpha process (such as and ) may be clusters of alpha particles, having a similar structure to Borromean nuclei.
, the heaviest known Borromean nucleus is . Heavier species along the neutron drip line have since been observed; these |
https://en.wikipedia.org/wiki/ModelSim | ModelSim is a multi-language environment by Siemens (previously developed by Mentor Graphics,) for simulation of hardware description languages such as VHDL, Verilog and SystemC, and includes a built-in C debugger. ModelSim can be used independently, or in conjunction with Intel Quartus Prime, PSIM, Xilinx ISE or Xilinx Vivado. Simulation is performed using the graphical user interface (GUI), or automatically using scripts.
Editions
Mentor HDL simulation products are offered in multiple editions, such as ModelSim PE and Questa Sim.
Questa Sim offers high-performance and advanced debugging capabilities, while ModelSim PE is the entry-level simulator for hobbyists and students. Questa Sim is used in large multi-million gate designs, and is supported on Microsoft Windows and Linux, in 32-bit and 64-bit architectures.
ModelSim can also be used with MATLAB/Simulink, using Link for ModelSim. Link for ModelSim is a fast bidirectional co-simulation interface between Simulink and ModelSim. For such designs, MATLAB provides a numerical simulation toolset, while ModelSim provides tools to verify the hardware implementation & timing characteristics of the design.
Language support
ModelSim uses a unified kernel for simulation of all supported languages, and the method of debugging embedded C code is the same as VHDL or Verilog.
ModelSim and Questa Sim products enable simulation, verification and debugging for the following languages:
VHDL
Verilog
Verilog 2001
SystemVerilog
PSL
SystemC
See also
Intel Quartus Prime
Icarus Verilog
List of HDL simulators
NCSim
Verilator
Xilinx ISE
Xilinx Vivado |
https://en.wikipedia.org/wiki/List%20of%20ATSC%20standards | Below are the published ATSC standards for ATSC digital television service, issued by the Advanced Television Systems Committee.
A/49: Ghost Canceling Reference Signal for NTSC (for adjacent-channel interference or co-channel interference with analog NTSC stations nearby)
A/52B: audio data compression (Dolby AC-3 and E-AC-3)
A/53E: "ATSC Digital Television Standard" (the primary document governing the standard)
A/55: "Program Guide for Digital Television" (now deprecated in favor of A/65 PSIP)
A/56: "System Information for Digital Television" (now deprecated in favor of A/65 PSIP)
A/57A: "Content Identification and Labeling for ATSC Transport" (for assigning a unique digital number to each episode of each TV show, to assist DVRs)
A/63: "Standard for Coding 25/50 Hz Video" (for use with PAL and SECAM-originated programming)
A/64A "Transmission Measurement and Compliance for Digital Television"
A/65C: "Program and System Information Protocol for Terrestrial Broadcast and Cable" (PSIP includes virtual channels, electronic program guides, and content ratings)
A/68: "PSIP Standard for Taiwan" (defines use of Chinese characters via Unicode 3.0)
A/69: recommended practices for implementing PSIP at a TV station
A/70A: "Conditional Access System for Terrestrial Broadcast"
A/71: "ATSC Parameterized Services Standard"
A/72: "Video System Characteristics of AVC in the ATSC Digital Television System" (implementing H.264/MPEG-4 as well as MVC for 3D television)
A/76: "Programming Metadata Communication Protocol" (XML-based PMCP maintains PSIP metadata though a TV station's airchain)
A/79: "Conversion of ATSC Signals for Distribution to NTSC Viewers" (recommended practice, issued February 2009)
A/80: "Modulation and Coding Requirements for Digital TV (DTV) Applications Over Satellite" (ATSC-S)
A/81: "Direct-to-Home Satellite Broadcast Standard" (not yet implemented by any services)
A/82: "Automatic Transmitter Power Control (ATPC) Data Return Link (DRL) Standard"
A/85: "Techniqu |
https://en.wikipedia.org/wiki/Azure%20Virtual%20Desktop | Azure Virtual Desktop (AVD), formerly known as Windows Virtual Desktop (WVD), is a Microsoft Azure-based system for virtualizing its Windows operating systems, providing virtualized desktops and applications securely in the cloud (over the Internet). It is aimed at enterprise customers rather than at individual users.
WVD was first announced by Microsoft in September 2018, available as a public preview in March 2019, and generally available at the end of September 2019.
Azure Virtual Desktop with Windows 10/11 Enterprise Multi-Session is a cloud-based alternative to an on-premise Remote Desktop Server (RDS). AVD is deployed in Azure Cloud as a virtual machine. License costs are already included in several Microsoft 365 subscriptions, including Microsoft 365 Business Premium or Microsoft 365 E3.
Availability / compatibility
Azure Virtual Desktop supports Windows 10/11 multi-session, Windows 10/11 single-session, Windows Server 2012 R2 and newer operating systems.
See also
Windows 365
Remote Desktop Services
Windows Virtual PC |
https://en.wikipedia.org/wiki/Bird%20flight | Bird flight is the primary mode of locomotion used by most bird species in which birds take off and fly. Flight assists birds with feeding, breeding, avoiding predators, and migrating.
Bird flight is one of the most complex forms of locomotion in the animal kingdom. Each facet of this type of motion, including hovering, taking off, and landing, involves many complex movements. As different bird species adapted over millions of years through evolution for specific environments, prey, predators, and other needs, they developed specializations in their wings, and acquired different forms of flight.
Various theories exist about how bird flight evolved, including flight from falling or gliding (the trees down hypothesis), from running or leaping (the ground up hypothesis), from wing-assisted incline running or from proavis (pouncing) behavior.
Basic mechanics of bird flight
Lift, drag and thrust
The fundamentals of bird flight are similar to those of aircraft, in which the aerodynamic forces sustaining flight are lift, drag, and thrust. Lift force is produced by the action of air flow on the wing, which is an airfoil. The airfoil is shaped such that the air provides a net upward force on the wing, while the movement of air is directed downward. Additional net lift may come from airflow around the bird's body in some species, especially during intermittent flight while the wings are folded or semi-folded (cf. lifting body).
Aerodynamic drag is the force opposite to the direction of motion, and hence the source of energy loss in flight. The drag force can be separated into two portions, lift-induced drag, which is the inherent cost of the wing producing lift (this energy ends up primarily in the wingtip vortices), and parasitic drag, including skin friction drag from the friction of air and body surfaces and form drag from the bird's frontal area. The streamlining of bird's body and wings reduces these forces. Unlike aircraft, which have engines to produce thrust, bi |
https://en.wikipedia.org/wiki/Frequency%20offset | In radio engineering, a frequency offset is an intentional slight shift of broadcast radio frequency (RF), to reduce interference with other transmitters.
Interference
The most important problem encountered in broadcasting via terrestrial transmitters is the interference from other broadcasters. In principle, each broadcaster has a different radio frequency (planned by the public authority) in a common reception area to avoid interference from each other. However still there are two problems: spurious radiation of adjacent channels and fringe reception.
Fringe reception is unintended reception under certain weather conditions. The exceptionally long-range reception means that the receiver may be tuned to more than one transmitter (transmitting at same frequency) at the same time. These transmitters may transmit programs of different broadcasters as well as the programs of the same broadcaster. In analogue transmission, even the transmitters transmitting the very same program interfere each other because of phase differences of the incoming signal, but in digital transmission the transmitters transmitting the same program in the same channel may reinforce each other.
The shift in RF
In order to reduce the interference from the fringe area transmitters transmitting in the same channels, a method named frequency offset is often used. By this method, a slightly shifted RF is assigned for a transmitter which may experience interference from other transmitters.
In TV transmission, the shifted RF is calculated by the formula;
where,
is the offset RF,
is the standard channel frequency,
is an integer such that or
is the line frequency. ( 15625 Hz. for 625-line systems, 15750 Hz. for System M and System J and 20475 Hz. for system E)
Precision offset is the same as frequency offset, except that in this case, the offset frequency is produced by a cesium controlled oscillator.
Example
8p offset of channel 9 in system B can be found readily
The standard |
https://en.wikipedia.org/wiki/Photovoltaic%20effect | The photovoltaic effect is the generation of voltage and electric current in a material upon exposure to light. It is a physical and chemical phenomenon.
The photovoltaic effect is closely related to the photoelectric effect. For both phenomena, light is absorbed, causing excitation of an electron or other charge carrier to a higher-energy state. The main distinction is that the term photoelectric effect is now usually used when the electron is ejected out of the material (usually into a vacuum) and photovoltaic effect used when the excited charge carrier is still contained within the material. In either case, an electric potential (or voltage) is produced by the separation of charges, and the light has to have a sufficient energy to overcome the potential barrier for excitation. The physical essence of the difference is usually that photoelectric emission separates the charges by ballistic conduction and photovoltaic emission separates them by diffusion, but some "hot carrier" photovoltaic devices concepts blur this distinction.
History
The first demonstration of the photovoltaic effect, by Edmond Becquerel in 1839, used an electrochemical cell. He explained his discovery in Comptes rendus de l'Académie des sciences, "the production of an electric current when two plates of platinum or gold immersed in an acid, neutral, or alkaline solution are exposed in an uneven way to solar radiation."
The first solar cell, consisting of a layer of selenium covered with a thin film of gold, was experimented by Charles Fritts in 1884, but it had a very poor efficiency. However, the most familiar form of the photovoltaic effect uses solid-state devices, mainly in photodiodes. When sunlight or other sufficiently energetic light is incident upon the photodiode, the electrons present in the valence band absorb energy and, being excited, jump to the conduction band and become free. These excited electrons diffuse, and some reach the rectifying junction (usually a diode p–n juncti |
https://en.wikipedia.org/wiki/Weibel%20instability | The Weibel instability is a plasma instability present in homogeneous or nearly homogeneous electromagnetic plasmas which possess an anisotropy in momentum (velocity) space. This anisotropy is most generally understood as two temperatures in different directions. Burton Fried showed that this instability can be understood more simply as the superposition of many counter-streaming beams. In this sense, it is like the two-stream instability except that the perturbations are electromagnetic and result in filamentation as opposed to electrostatic perturbations which would result in charge bunching. In the linear limit the instability causes exponential growth of electromagnetic fields in the plasma which help restore momentum space isotropy. In very extreme cases, the Weibel instability is related to one- or two-dimensional stream instabilities.
Consider an electron-ion plasma in which the ions are fixed and the electrons are hotter in the y-direction than in x or z-direction.
To see how magnetic field perturbation would grow, suppose a field B = B cos kx spontaneously arises from noise. The Lorentz force then bends the electron trajectories with the result that upward-moving-ev x B electrons congregate at B and downward-moving ones at A. The resulting current sheets generate magnetic field that enhances the original field and thus
perturbation grows.
Weibel instability is also common in astrophysical plasmas, such as collisionless shock formation in supernova remnants and -ray bursts.
A Simple Example of Weibel Instability
As a simple example of Weibel instability, consider an electron beam with density and initial velocity propagating in a plasma of density with velocity . The analysis below will show how an electromagnetic perturbation in the form of a plane wave gives rise to a Weibel instability in this simple anisotropic plasma system. We assume a non-relativistic plasma for simplicity.
We assume there is no background electric or magnetic field i.e |
https://en.wikipedia.org/wiki/Little%20gastrin%20I | Little gastrin I is a form of gastrin commonly called as gastrin-17. This is a protein hormone, secreted by the intestine.
Gastrin II has identical amino acid composition to Gastrin I, the only difference is that the single tyrosine residue is sulfated in Gastrin II. |
https://en.wikipedia.org/wiki/Random%20Hacks%20of%20Kindness | Random Hacks of Kindness (RHoK) is a global community of technologists dedicated to solving problems for charities, non-profits and social enterprises by organising recurring Hackathons that has existed since 2009. The organisation currently has a presence in over 20 cities throughout 5 continents, and had 2000 participants in 2017.
History
Origins
Random Hacks of Kindness grew out of an industry panel discussion at the first Crisis camp Bar Camp in Washington, D.C. in June 2009. Panel attendees included Patrick Svenburg of Microsoft, Phil Dixon and Jeff Martin of Google and Jeremy Johnstone of Yahoo!. They agreed to use their developer communities to create solutions that will affect disaster response, risk reduction and recovery. The idea was for a "hackathon" with developers producing open source solutions. The World Bank's Disaster Risk Reduction Unit (Stuart Gill) and NASA's Open Government team (Robbie Schingler) joined the partnership and these "founding partners" (Microsoft, Yahoo, Google, NASA and the World Bank) decided on the name "Random Hacks of Kindness" for their first event.
An innovation incubator in the area of sustainable development, SecondMuse acts as "operational lead" for Random Hacks of Kindness, coordinating global volunteer efforts, facilitating collaborative partnerships, and managing communications and branding.
2009: RHoK 0
The first Random Hacks of Kindness (RHoK 0) was held at the Hacker Dojo in Mountain View, California on November 12–14, 2009. FEMA Administrator Craig Fugate gave the keynote and made a call to action to the developers to apply their creativity to the challenges and featured hacks. The first RHoK event is known as RHoK 0 after 0-based array indexing in computer programming.
Featured projects were
I'm OK
Tweak the Tweet (not a code "hack", but an edit/republish "hack")
Break Glass
Tweak the Tweet was used during the Haiti earthquake response in January 2010
2010: RHoK 1.0 and 2.0
The second RHoK event w |
https://en.wikipedia.org/wiki/Augmented%20Reality%20Sandtable | The Augmented Reality Sandtable (ARES) is an interactive, digital sand table that uses augmented reality (AR) technology to create a 3D battlespace map. It was developed by the Human Research and Engineering Directorate (HRED) at the Army Research Laboratory (ARL) to combine the positive aspects of traditional military sand tables with the latest digital technologies to better support soldier training and offer new possibilities of learning. It uses a projector to display a topographical map on top of the sand in a regular sandbox as well as a motion sensor that keeps track of changes in the layout of the sand to appropriately adjust the computer-generated terrain display.
An ARL study conducted in 2017 with 52 active duty military personnel (36 males and 16 females) found that the participants who used ARES spent less time setting up the table compared to participants who used a traditional sand table. In addition, ARES demonstrated a lower perceived workload score, as measured using the NASA Task Load Index (NASA-TLX) ratings, compared to the traditional sand table. However, there was no significant difference in post-knowledge test scores in recreating the visual map.
Development
The ARES project was one of the 25 ARL initiatives in development from 1995 to 2015 that focused on visualizing spatial data on virtual or sand table interfaces. It was developed by HRED's Simulation and Training Technology Center (STTC) with Charles Amburn as the principal investigator. Collaborations involved with ARES included Dignitas Technologies, Design Interactive (DI), the University of Central Florida's Institute for Simulation and Training, and the U.S. Military Academy at West Point.
ARES was largely designed to be a tangible user interface (TUI), in which digital information can be manipulated using physical objects such as a person's hand. It was constructed using commercial off-the-shelf components, including a projector, a laptop, an LCD monitor, Microsoft's Xbox Kinec |
https://en.wikipedia.org/wiki/Schl%C3%BCsselger%C3%A4t%2039 | The Schlüsselgerät 39 (SG-39) was an electrically operated rotor cipher machine, invented by the German Fritz Menzer during World War II. The device was the evolution of the Enigma rotors coupled with three Hagelin pin wheels to provide variable stepping of the rotors. All three wheels stepped once with each encipherment. Rotors stepped according to normal Enigma rules, except that an active pin at the reading station for a pin wheel prevented the coupled rotor from stepping. The cycle for a normal Enigma was 17,576 characters. When the Schlüsselgerät 39 was correctly configured, its cycle length was characters, which was more than 15,000 times longer than a standard Enigma. The Schlüsselgerät 39 was fully automatic, in that when a key was pressed, the plain and cipher letters were printed on separate paper tapes, divided into five-digit groups. The Schlüsselgerät 39 was abandoned by German forces in favour of the Schlüsselgerät 41.
Technical description
Note: Otto Buggisch gave the technical description of the cipher unit as part of TICOM homework.
Gerät 39 is an electrically operated cipher machine. The cipher technique is derived from the Enigma cipher machine. A direct current passes through 3 or 4 wheels, with 26 positions, I, II, II, a reflector wheel U, and the again through the 3 wheels in reverse order, III II and I. Unlike the Enigma, the wheels here do not control their own movement: this is done through 3 independent pin-wheels N1 N2 and N3 with periods 21,23 and 25. The figures were distributed among N1 N2 and N2 in possibly two different configurations.
The pin wheels have a uniform motion, i.e. they move one position for every letter keyed. As for the movement of the key wheels and other details, the machine passed through different stages of development in the course of time, for which there were no specific names and which will be denoted here by a,b,c and d.
Each of three wheels moves on one place when there is an active pin at the sensing |
https://en.wikipedia.org/wiki/Follicle-stimulating%20hormone%20receptor | The follicle-stimulating hormone receptor or FSH receptor (FSHR) is a transmembrane receptor that interacts with the follicle-stimulating hormone (FSH) and represents a G protein-coupled receptor (GPCR). Its activation is necessary for the hormonal functioning of FSH. FSHRs are found in the ovary, testis, and uterus.
FSHR gene
The gene for the FSHR is found on chromosome 2 p21 in humans. The gene sequence of the FSHR consists of about 2,080 nucleotides.
Receptor structure
The FSHR consists of 695 amino acids and has a molecular mass of about 76 kDa. Like other GPCRs, the FSH-receptor possesses seven membrane-spanning domains or transmembrane helices.
The extracellular domain of the receptor contains 11 leucine-rich repeats and is glycosylated. It has two subdomains, a hormone-binding subdomain followed by a signal-specificity subdomain. The hormone-binding subdomain is responsible for the high-affinity hormone binding, and the signal-specificity subdomain, containing a sulfated tyrosine at position 335 (sTyr) in a hinge loop, is required for the hormone activity.
The transmembrane domain contains two highly conserved cysteine residues that build disulfide bonds to stabilize the receptor structure. A highly conserved Asp-Arg-Tyr triplet motif is present in GPCR family members in general and may be of importance to transmit the signal. In FSHR and its closely related other glycoprotein hormone receptor members (LHR and TSHR), this conserved triplet motif is a variation Glu-Arg-Trp sequence.
The C-terminal domain is intracellular and brief, rich in serine and threonine residues for possible phosphorylation.
Ligand binding and signal transduction
Upon initial binding to the LRR region of FSHR, FSH reshapes its conformation to form a new pocket. FSHR then inserts its sulfotyrosine from the hinge loop into the pockets and activates the 7-helical transmembrane domain. This event leads to a transduction of the signal that activates the G protein that is bo |
https://en.wikipedia.org/wiki/Table%20of%20Gaussian%20integer%20factorizations | A Gaussian integer is either the zero, one of the four units (±1, ±i), a Gaussian prime or composite. The article is a table of Gaussian Integers followed either by an explicit factorization or followed by the label (p) if the integer is a Gaussian prime. The factorizations take the form of an optional unit multiplied by integer powers of Gaussian primes.
Note that there are rational primes which are not Gaussian primes. A simple example is the rational prime 5, which is factored as in the table, and therefore not a Gaussian prime.
Conventions
The second column of the table contains only integers in the first quadrant, which means the real part x is positive and the imaginary part y is non-negative. The table might have been further reduced to the integers in the first octant of the
complex plane using the symmetry
.
The factorizations are often not unique in the sense that the unit could be absorbed into any other factor with exponent equal to one. The entry , for example, could also be written as . The entries in the table resolve this ambiguity by the following convention: the factors are primes in the right complex half plane with absolute value of the real part larger than or equal to the absolute value of the imaginary part.
The entries are sorted according to increasing norm . The table is complete up to the maximum norm at the end of the table in the sense that
each composite or prime in the first quadrant appears in the second column.
Gaussian primes occur only for a subset of norms, detailed in sequence . This here is a human-readable version of sequences and .
Factorizations
See also
Gaussian integer
Table of divisors
Integer factorization |
https://en.wikipedia.org/wiki/Spatial%20ecology | Spatial ecology studies the ultimate distributional or spatial unit occupied by a species. In a particular habitat shared by several species, each of the species is usually confined to its own microhabitat or spatial niche because two species in the same general territory cannot usually occupy the same ecological niche for any significant length of time.
Overview
In nature, organisms are neither distributed uniformly nor at random, forming instead some sort of spatial pattern. This is due to various energy inputs, disturbances, and species interactions that result in spatially patchy structures or gradients. This spatial variance in the environment creates diversity in communities of organisms, as well as in the variety of the observed biological and ecological events. The type of spatial arrangement present may suggest certain interactions within and between species, such as competition, predation, and reproduction. On the other hand, certain spatial patterns may also rule out specific ecological theories previously thought to be true.
Although spatial ecology deals with spatial patterns, it is usually based on observational data rather than on an existing model. This is because nature rarely follows set expected order. To properly research a spatial pattern or population, the spatial extent to which it occurs must be detected. Ideally, this would be accomplished beforehand via a benchmark spatial survey, which would determine whether the pattern or process is on a local, regional, or global scale. This is rare in actual field research, however, due to the lack of time and funding, as well as the ever-changing nature of such widely-studied organisms such as insects and wildlife. With detailed information about a species' life-stages, dynamics, demography, movement, behavior, etc., models of spatial pattern may be developed to estimate and predict events in unsampled locations.
History
Most mathematical studies in ecology in the nineteenth century assumed a un |
https://en.wikipedia.org/wiki/Dominating%20set | In graph theory, a dominating set for a graph is a subset of its vertices, such that any vertex of is either in , or has a neighbor in . The domination number is the number of vertices in a smallest dominating set for .
The dominating set problem concerns testing whether for a given graph and input ; it is a classical NP-complete decision problem in computational complexity theory. Therefore it is believed that there may be no efficient algorithm that can compute for all graphs . However, there are efficient approximation algorithms, as well as efficient exact algorithms for certain graph classes.
Dominating sets are of practical interest in several areas. In wireless networking, dominating sets are used to find efficient routes within ad-hoc mobile networks. They have also been used in document summarization, and in designing secure systems for electrical grids.
Formal definition
Given an undirected graph , a subset of vertices is called a dominating set if for every vertex , there is a vertex such that .
Every graph has at least one dominating set: if the set of all vertices, then by definition D is a dominating set, since there is no vertex . A more interesting challenge is to find small dominating sets. The domination number of is defined as: .
Variants
A connected dominating set is a dominating set that is also connected. If S is a connected dominating set, one can form a spanning tree of G in which S forms the set of non-leaf vertices of the tree; conversely, if T is any spanning tree in a graph with more than two vertices, the non-leaf vertices of T form a connected dominating set. Therefore, finding minimum connected dominating sets is equivalent to finding spanning trees with the maximum possible number of leaves.
A total dominating set (or strongly-dominating set) is a set of vertices such that all vertices in the graph, including the vertices in the dominating set themselves, have a neighbor in the dominating set. That is: for every |
https://en.wikipedia.org/wiki/Tolman%20surface%20brightness%20test | The Tolman surface brightness test is one out of six cosmological tests that were conceived in the 1930s to check the viability of and compare new cosmological models. Tolman's test compares the surface brightness of galaxies as a function of their redshift (measured as z). Such a comparison was first proposed in 1930 by Richard C. Tolman as a test of whether the universe is expanding or static. It is a unique test of cosmology, as it is independent of dark energy, dark matter and Hubble constant parameters, testing purely for whether Cosmological Redshift is caused by an expanding universe or not.
In a simple (static and flat) universe, the light received from an object drops proportional to the square of its distance and the apparent area of the object also drops proportional to the square of the distance, so the surface brightness (light received per surface area) would be constant, independent of the distance. In an expanding universe, however, there are two effects that change this relation. First, the rate at which photons are received is reduced because each photon has to travel a little farther than the one before. Second, the energy of each photon observed is reduced by the redshift. At the same time, distant objects appear larger than they really are because the photons observed were emitted at a time when the object was closer. Adding these effects together, the surface brightness in a simple expanding universe (flat geometry and uniform expansion over the range of redshifts observed) should decrease with the fourth power of .
One of the earliest and most comprehensive studies was published in 1996, as observational requirements limited the practicality of the test till then. This test found consistency with an expanding universe. However, therein, the authors note that:
A later paper that reviewed this one removed their assumed expansion cosmology for calculating SB, to make for a fair test, and found that the 1996 results, once the correction was |
https://en.wikipedia.org/wiki/Conflict%20continuum | A conflict continuum is a model or concept various social science researchers use when modeling conflict on a continuum from low to high-intensity, such as from aggression to irritation to explosiveness.
The mathematical model of game theory originally posited only a winner and a loser (a zero-sum game) in a conflict, but was extended to cooperation (a win-win situation and a non-zero sum game), and lets users specify any point between cooperation, peace, rivalry, contest, crisis, and conflict among stakeholders.
Overview
By the decade of the 2010s, military planners realized that additional capabilities in communications, sensors and weapons countermeasures made it possible for competitors to react to a contestant's moves in the "gray zone" just short of conflict. In 2018 Kelly McCoy identified a continuum within competition itself, as explored in the United States Joint Staff's Joint Concept for Integrated Campaigning (JCIC), up to the point just short of armed conflict, while noting Perkins' connection to deterrence in the continuum. In 2020, Donald Stoker and Craig Whiteside cautioned that for strategists, the "gray zone" must not blur peace and war; they offered an analysis of the need for strategists to clearly distinguish peace, competition, contest, conflict, and war.
Standoff is the condition of deadlock between antagonists, sometimes measured by the distance between them (standoff distance). For antagonists in a non-zero sum game, von Neumann and Morgenstern showed in 1944 that this condition is equivalent to a zero-sum game with +1 antagonists, where the +1st player ("the fictitious player") is not an entity. Rather the fictitious player represents the global profit (or loss) of the players in the non-zero sum game.
If we reduce this game to a zero-sum 3-player game by the introduction of a fictitious player 3, then the characteristic function becomes the one given In Tibor Scitovsky's terminology (more commonly known as the Kaldor–Hicks criterion), |
https://en.wikipedia.org/wiki/EcosimPro | EcosimPro is a simulation tool developed by Empresarios Agrupados A.I.E for modelling simple and complex physical processes that can be expressed in terms of Differential algebraic equations or Ordinary differential equations and Discrete event simulation.
The application runs on the various Microsoft Windows platforms and uses its own graphic environment for model design.
The modelling of physical components is based on the EcosimPro language (EL) which is very similar to other conventional Object-oriented programming languages but is powerful enough to model continuous and discrete processes.
This tool employs a set of libraries containing various types of components (mechanical, electrical, pneumatic, hydraulic, etc.) that can be reused to model any type of system.
It is used within ESA for propulsion systems analysis and is the recommended ESA analysis tool for ECLS systems.
Origins
The EcosimPro Tool Project began in 1989 with funds from the European Space Agency (ESA) and with the goal of simulating environmental control and life support systems for crewed spacecraft, such as the Hermes shuttle.
The multidisciplinary nature of this modelling tool led to its use in many other disciplines, including fluid mechanics, chemical processing, control, energy, propulsion and flight dynamics. These complex applications have demonstrated that EcosimPro is very robust and ready for use in many other fields.
The modelling language
Code examples
Differential equation
To familiarize yourself with the use of EcosimPro, first create a simple component to solve a differential equation. Although EcosimPro is designed to simulate complex systems, it can also be used independently of a physical system as if it were a pure equation solver. The example in this section illustrates this type of use. It solves the following differential equation to introduce a delay to variable x:
which is equivalent to
where x and y have a time dependence that will be defined in the |
https://en.wikipedia.org/wiki/Engineering%20informatics | The term Engineering Informatics may be related to information engineering (differently understood information processing), computer engineering (development of computer hardware-software systems), or computational engineering (development of software for engineering purposes), among other meanings. This word is used with different context in different countries. In general, some people assume that the central area of interest in informatics is information processing within man-made artificial (engineering) systems, called also computational or computer systems. The focus on artificial systems separates informatics from psychology and cognitive science, which focus on information processing within natural systems (primarily people). However, nowadays these fields have areas where they overlap, e.g. in field of affective computing.
Computer Engineering as a discipline of field study
Computer-aided design (CAD), intelligent CAD, engineering analysis, collaborative design support, computer-aided engineering, and product life-cycle management are some of the terms that have emerged over the past decades of computing in engineering. Codification and automation of engineering knowledge and methods have had major impact on engineering practice. The use of computers by engineers has consistently tracked advancements in computer and information sciences. Computing, algorithms, computational methods, and engineering have increasingly intertwined themselves as developments in theory and practice in both disciplines influence each other. Therefore, it is now time to begin using the term “engineering informatics” to cover the science of the information that flows through these processes.
Informatics, with origins in the German word "Informatik" referring to automated information processing, has evolved to its current broad definition. The rise of the term informatics can be attributed to the breadth of disciplines that are now accepted and envisioned as contributing to the fi |
https://en.wikipedia.org/wiki/Path-based%20strong%20component%20algorithm | In graph theory, the strongly connected components of a directed graph may be found using an algorithm that uses depth-first search in combination with two stacks, one to keep track of the vertices in the current component and the second to keep track of the current search path. Versions of this algorithm have been proposed by , , , , and ; of these, Dijkstra's version was the first to achieve linear time.
Description
The algorithm performs a depth-first search of the given graph G, maintaining as it does two stacks S and P (in addition to the normal call stack for a recursive function).
Stack S contains all the vertices that have not yet been assigned to a strongly connected component, in the order in which the depth-first search reaches the vertices.
Stack P contains vertices that have not yet been determined to belong to different strongly connected components from each other. It also uses a counter C of the number of vertices reached so far, which it uses to compute the preorder numbers of the vertices.
When the depth-first search reaches a vertex v, the algorithm performs the following steps:
Set the preorder number of v to C, and increment C.
Push v onto S and also onto P.
For each edge from v to a neighboring vertex w:
If the preorder number of w has not yet been assigned (the edge is a tree edge), recursively search w;
Otherwise, if w has not yet been assigned to a strongly connected component (the edge is a forward/back/cross edge):
Repeatedly pop vertices from P until the top element of P has a preorder number less than or equal to the preorder number of w.
If v is the top element of P:
Pop vertices from S until v has been popped, and assign the popped vertices to a new component.
Pop v from P.
The overall algorithm consists of a loop through the vertices of the graph, calling this recursive search on each vertex that does not yet have a preorder number assigned to it.
Related algorithms
Like this algorithm, Tarjan's strongly connected components algo |
https://en.wikipedia.org/wiki/Torticollis | Torticollis, also known as wry neck, is an extremely painful, dystonic condition defined by an abnormal, asymmetrical head or neck position, which may be due to a variety of causes. The term torticollis is derived from the Latin words tortus, meaning "twisted", and collum, meaning "neck".
The most common case has no obvious cause, and the pain and difficulty with turning the head usually goes away after a few days, even without treatment in adults.
Signs and symptoms
Torticollis is a fixed or dynamic tilt, rotation, with flexion or extension of the head and/or neck.
The type of torticollis can be described depending on the positions of the head and neck.
laterocollis: the head is tipped toward the shoulder
rotational torticollis: the head rotates along the longitudinal axis towards the shoulder
anterocollis: forward flexion of the head and neck and brings the chin towards the chest
retrocollis: hyperextension of head and neck backward bringing the back of the head towards the back
A combination of these movements may often be observed. Torticollis can be a disorder in itself as well as a symptom in other conditions.
Other signs and symptoms include:
Neck pain
Occasional formation of a mass
Thickened or tight sternocleidomastoid muscle
Tenderness on the cervical spine
Tremor in head
Unequal shoulder heights
Decreased neck movement
Causes
A multitude of conditions may lead to the development of torticollis including: muscular fibrosis, congenital spine abnormalities, or toxic or traumatic brain injury.
A rough categorization discerns between congenital torticollis and acquired torticollis.
Other categories include:
Osseous
Traumatic
CNS/PNS
Ocular
Non-muscular soft tissue
Spasmodic
Drug induced
Oral ties (lip and tongue ties)
Congenital muscular torticollis
Congenital muscular torticollis is the most common torticollis that is present at birth. Congenital muscular torticollis is the third most common congenital musculoskeletal deformity in chi |
https://en.wikipedia.org/wiki/Glycoside%20hydrolase%20family%204 | In molecular biology, glycoside hydrolase family 4 is a family of glycoside hydrolases , which are a widespread group of enzymes that hydrolyse the glycosidic bond between two or more carbohydrates, or between a carbohydrate and a non-carbohydrate moiety. A classification system for glycoside hydrolases, based on sequence similarity, has led to the definition of >100 different families. This classification is available on the CAZy web site, and also discussed at CAZypedia, an online encyclopedia of carbohydrate active enzymes.
Glycoside hydrolase family 4 CAZY GH_4 comprises enzymes with several known activities; 6-phospho-beta-glucosidase (); 6-phospho-alpha-glucosidase (); alpha-galactosidase (); alpha-D-glucuronidase (EC 3.2.1.139). 6-phospho-alpha-glucosidase requires both NAD(H) and divalent metal (Mn2+, Fe2+, Co2+, or Ni2+) for activity.
External references
GH4 in CAZypedia |
https://en.wikipedia.org/wiki/Working%20set | Working set is a concept in computer science which defines the amount of memory that a process requires in a given time interval.
Definition
Peter Denning (1968) defines "the working set of information of a process at time to be the collection of information referenced by the process during the process time interval ".
Typically the units of information in question are considered to be memory pages. This is suggested to be an approximation of the set of pages that the process will access in the future (say during the next time units), and more specifically is suggested to be an indication of what pages ought to be kept in main memory to allow most progress to be made in the execution of that process.
Rationale
The effect of the choice of what pages to be kept in main memory (as distinct from being paged out to auxiliary storage) is important: if too many pages of a process are kept in main memory, then fewer other processes can be ready at any one time. If too few pages of a process are kept in main memory, then its page fault frequency is greatly increased and the number of active (non-suspended) processes currently executing in the system approaches zero.
The working set model states that a process can be in RAM if and only if all of the pages that it is currently using (often approximated by the most recently used pages) can be in RAM. The model is an all or nothing model, meaning if the pages it needs to use increases, and there is no room in RAM, the process is swapped out of memory to free the memory for other processes to use.
Often a heavily loaded computer has so many processes queued up that, if all the processes were allowed to run for one scheduling time slice, they would refer to more pages than there is RAM, causing the computer to "thrash".
By swapping some processes from memory, the result is that processes—even processes that were temporarily removed from memory—finish much sooner than they would if the computer attempted to run them all at |
https://en.wikipedia.org/wiki/Comparison%20of%20text%20editors | This article provides basic comparisons for notable text editors. More feature details for text editors are available from the Category of text editor features and from the individual products' articles. This article may not be up-to-date or necessarily all-inclusive.
Feature comparisons are made between stable versions of software, not the upcoming versions or beta releases – and are exclusive of any add-ons, extensions or external programs (unless specified in footnotes).
Overview
Operating system support
This section lists the operating systems that different editors can run on. Some editors run on additional operating systems that are not listed.
Cross-platform
Natural language (localization)
Document interface
Notes
Multiple instances: multiple instances of the program can be opened simultaneously for editing multiple files. Applies both for single document interface (SDI) and multiple document interface (MDI) programs. Also applies for program that has a user interface that looks like multiple instances of the same program (such as some versions of Microsoft Word).
Single document window splitting: window can be split to simultaneously view different areas of a file.
MDI: Overlappable windows: each opened document gets its own fully movable window inside the editor environment.
MDI: Tabbed document interface: multiple documents can be viewed as tabs in a single window.
MDI: Window splitting: splitting application window to show multiple documents (non-overlapping windows).
Basic features
Programming features
Notes
Syntax highlighting: Displays text in different colors and fonts according to the category of terms.
Function list: Lists all functions from current file in a window or sidebar and allows user to jump directly to the definition of that function for example by double-clicking on the function name in the list. More or less realtime (does not require creating a symbol database, see below).
Symbol database: Database of functions, variable |
https://en.wikipedia.org/wiki/MIPOL1 | MIPOL1 (Mirror Image Polydactyly 1), also known as CCDC193 (Coiled-coil domain containing 193), is a protein that in humans is encoded by the MIPOL1 gene. Mutation of this gene is associated with mirror-image polydactyly (also known as Laurin-Sandrow syndrome.) in humans, which is a rare genetic condition characterized by mirror-image duplication of digits.
Gene
MIPOL1 is also known as CCDC193 (Coiled-coil domain containing 193).
Locus
The MIPOL1 gene is located at 14q13.3-q21.1 on the plus strand, spanning base pairs 37,197,888 to 37,579,207 (in the human GRCh38 primary assembly, length: 381,320 base pairs), consisting of 15 exons and 11 introns. Some notable genes in its neighborhood include SLC25A21 (mutation of this gene causes synpolydactyly) and FOXA1.
mRNA
MIPOL1 has at least 15 known splice isoforms produced by alternative splicing.
Protein
Properties
The unmodified MIPOL1 protein isoform 1 in humans has an isoelectric point of 5.6 and molecular weight 51.5 kDa. Relative to other human proteins, MIPOL1 consists of unusually low amounts of Proline and Glycine and higher amounts of Glutamic acid and Glutamine.
Isoforms
There are at least three known isoforms of this protein in humans produced by alternative splicing: isoform 1, of length 442 amino acids, isoform 2 of length 261 amino acids and isoform 3 of length 169 amino acids.
Domains and motifs
MIPOL1 contains two coiled-coil domains in its C-terminus at positions 107 – 212 and 253 – 435 (shown in Fig.1). A bipartite nuclear localization signal is predicted at position 128 – 143.
Post-translational modifications
The following post-translational modifications are predicted using bioinformatics tools for MIPOL1. Multiple phosphorylation sites are predicted for this protein, that are conserved in close orthologs, including a Casein kinase 1 (CK1) site, three Casein kinase 2 (CK2) sites, and three NEK2 sites.
Structure
The exact structure of the MIPOL1 has not yet been characterized. Homology-based |
https://en.wikipedia.org/wiki/Search%20for%20Hidden%20Particles | The Search for Hidden Particle (SHiP) is a fixed-target experiment at CERN's Super Proton Synchrotron (SPS) with the goal of searching for the interactions and measurements of the weakly interacting particles. In October 2013, the Expression of Interest letter for SHiP was submitted to the SPS Council (SPSC). Following which the Technical Proposal was submitted in April 2015, describing the experimental and detector facility. The Comprehensive Design Study was completed during 2016-19.
SHiP Collaboration intends to search for the weakly interacting particles whose masses are below the Fermi energy scale. Such particles cannot be detected at Large Hadron Collider yet, though the High Luminosity LHC may open some possibilities. Alongside, the SHiP detector will also search for weakly-interacting sub-GeV dark matter particles.
SHiP also plans to add information to the domain of tau neutrino physics. Out of the three neutrino flavors, the tau neutrino is the least studied. The experiment will aim to make the first direct observation of anti-tau neutrino, as well as measurements of the tau-neutrino and anti-tau neutrino cross-sections. Another goal is to study lepton flavor non-conservation, by observing the decays of the tau-leptons. |
https://en.wikipedia.org/wiki/Black%20and%20white%20hat%20symbolism%20in%20film | In American films of the Western genre between the 1920s and the 1940s, white hats were often worn by heroes and black hats by villains to symbolize the contrast in good versus evil. The 1903 short film The Great Train Robbery was the first to apply this convention. Two exceptions to the convention were portrayals by William Boyd (active 1918–1954), who wore dark clothing as Hopalong Cassidy, and Robert Taylor's portrayal in the film The Law and Jake Wade (1958).
The book Investigating Information Society said the convention was arbitrarily imposed by filmmakers in the genre with the expectation that audiences would understand the categorizations. It said whiteness was associated with "purity, cleanliness, and moral righteousness", which is reminiscent of a woman's wedding dress traditionally being white. The book said, "The difference, of course, has to do with particular cultural conceptions of gender and sexuality and the context within which white is worn." The convention also carried a practical benefit -- it helped audiences identify heroes and villains during fast-paced fight scenes in black and white films, even when one actor was filmed from behind.
In the 21st century, Western films referenced and spun the convention in different ways. In the 2005 film Brokeback Mountain, one of the two starring cowboys wears black while the other wears white. The film does not disclose any standard conventions for the symbolism other than the wearer of the black hat being shot like in early films. In the 2007 film 3:10 to Yuma, a remake of the 1957 film, a henchman hiring local gunmen to free his boss from jail, tells them not to shoot at "the black hat", a light reference to the convention. The black and white cowboy hats play an important role in characterization in "Westworld", where the protagonist chooses to wear a white cowboy hat while an antagonist wears a black hat. The series re-uses the trope with another character, Logan, who dons a black cowboy hat before s |
https://en.wikipedia.org/wiki/Banach%20space | In mathematics, more specifically in functional analysis, a Banach space (pronounced ) is a complete normed vector space. Thus, a Banach space is a vector space with a metric that allows the computation of vector length and distance between vectors and is complete in the sense that a Cauchy sequence of vectors always converges to a well-defined limit that is within the space.
Banach spaces are named after the Polish mathematician Stefan Banach, who introduced this concept and studied it systematically in 1920–1922 along with Hans Hahn and Eduard Helly.
Maurice René Fréchet was the first to use the term "Banach space" and Banach in turn then coined the term "Fréchet space".
Banach spaces originally grew out of the study of function spaces by Hilbert, Fréchet, and Riesz earlier in the century. Banach spaces play a central role in functional analysis. In other areas of analysis, the spaces under study are often Banach spaces.
Definition
A Banach space is a complete normed space
A normed space is a pair
consisting of a vector space over a scalar field (where is commonly or ) together with a distinguished
norm Like all norms, this norm induces a translation invariant
distance function, called the canonical or (norm) induced metric, defined for all vectors by
This makes into a metric space
A sequence is called or or if for every real there exists some index such that
whenever and are greater than
The normed space is called a and the canonical metric is called a if is a , which by definition means for every Cauchy sequence in there exists some such that
where because this sequence's convergence to can equivalently be expressed as:
The norm of a normed space is called a if is a Banach space.
L-semi-inner product
For any normed space there exists an L-semi-inner product on such that for all ; in general, there may be infinitely many L-semi-inner products that satisfy this condition. L-semi-inner products are a generalizati |
https://en.wikipedia.org/wiki/Joanne%20Moldenhauer | Joanne K. Moldenhauer (née Gatz, March 15, 1928 – February 14, 2016) was an American high school mathematics teacher and two-time winner of the Edyth May Sliffe Award.
Education
Moldenhauer's father served in the United States Army. She was born in Omaha, Nebraska, and attended Benson High School in Omaha. She graduated in 1949 from the Iowa State College, with a degree in physics, hoping to go on to a research career in physics was but blocked from that goal because there were few places for women in physics at the time.
Career
After graduating from Iowa State, Moldenhauer became a high school mathematics teacher in Omaha, and two years later became a student again at the University of Minnesota. She completed a master's degree in mathematics at Minnesota in 1952, and started work as an electrical engineer at General Electric. Her job there involved the development of fire-control systems for military aircraft. However, bored with her work, she soon returned to high school teaching. In 1955 she became a high school teacher in Schenectady, New York, where she had been working for General Electric, and in 1956 she moved to Davis Senior High School in Davis, California.
After 50 years as a teacher in Davis, Moldenhauer retired in 2006.
Contributions and recognition
Moldenhauer won the Edyth May Sliffe Award for Distinguished High School Mathematics Teaching of the Mathematical Association of America twice, in 1990 and 2001. She was also a winner of Stanford University's Frederick Emmons Terman Engineering Award, given annually by the graduating engineering students at Stanford to a distinguished high school teacher. She was a two-time winner of Harvey Mudd College's Distinguished Teaching Award. Unusually for a high school mathematics teacher, Moldenhauer has an Erdős number of 2, from her collaboration with mathematician Sherman K. Stein and mechanical engineer Anthony S. Wexler on "Trigonometry and a Wood Bowl". |
https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93Woods%20number | In number theory, a positive integer is said to be an Erdős–Woods number if it has the following property:
there exists a positive integer such that in the sequence of consecutive integers, each of the elements has a non-trivial common factor with one of the endpoints. In other words, is an Erdős–Woods number if there exists a positive integer such that for each integer between and , at least one of the greatest common divisors or is greater than .
Examples
The first Erdős–Woods numbers are
16, 22, 34, 36, 46, 56, 64, 66, 70, 76, 78, 86, 88, 92, 94, 96, 100, 106, 112, 116 … .
History
Investigation of such numbers stemmed from the following prior conjecture by Paul Erdős:
There exists a positive integer such that every integer is uniquely determined by the list of prime divisors of .
Alan R. Woods investigated this question for his 1981 thesis. Woods conjectured that whenever , the interval always includes a number coprime to both endpoints. It was only later that he found the first counterexample, , with . The existence of this counterexample shows that 16 is an Erdős–Woods number.
proved that there are infinitely many Erdős–Woods numbers, and showed that the set of Erdős–Woods numbers is recursive. |
https://en.wikipedia.org/wiki/National%20Research%20Ethics%20Service | The National Research Ethics Service (NRES) is a UK medical quango which deals with research ethics. Principal Investigators must describe the experiment they intend to pursue to the NRES for its approval, failing which the study is prohibited.
History
The NRES was launched on 1 April 2007.
The adjective "National" was omitted from the name at some unknown point in time.
Purpose
In 2009, the NRES issued a leaflet in which it described its purpose:
The National Research Ethics Service (NRES) reviews research proposals to protect the rights and safety of research participants and enables ethical research which is of potential benefit to science and society.
Substance of reports
The substance of the NRES reports can be gleaned from an approval obtained in 2011 by Stephanie Taylor, who was then Professor of Public Health and Primary Care at Queen Mary University of London. |
https://en.wikipedia.org/wiki/Biaxial%20tensile%20testing | In materials science and solid mechanics, biaxial tensile testing is a versatile technique to address the mechanical characterization of planar materials. It is a generalized form of tensile testing in which the material sample is simultaneously stressed along two perpendicular axes. Typical materials tested in biaxial configuration include
metal sheets,
silicone elastomers,
composites,
thin films,
textiles
and biological soft tissues.
Purposes of biaxial tensile testing
A biaxial tensile test generally allows the assessment of the mechanical properties
and a complete characterization for uncompressible isotropic materials, which can be obtained through a fewer amount of specimens with respect to uniaxial tensile tests.
Biaxial tensile testing is particularly suitable for understanding the mechanical properties of biomaterials, due to their directionally oriented microstructures.
If the testing aims at the material characterization of the post elastic behaviour, the uniaxial results become inadequate, and a biaxial test is required in order to examine the plastic behaviour.
In addition to this, using uniaxial test results to predict rupture under biaxial stress states seems to be inadequate.
Even if a biaxial tensile test is performed in a planar configuration, it may be equivalent to the stress state applied on three-dimensional geometries, such as cylinders with an inner pressure and an axial stretching.
The relationship between the inner pressure and the circumferential stress is given by the Mariotte formula:
where is the circumferential stress, P the inner pressure, D the inner diameter and t the wall thickness of the tube.
Equipment
Typically, a biaxial tensile machine is equipped with motor stages, two load cells and a gripping system.
Motor stages
Through the movement of the motor stages a certain displacement is applied on the material sample. If the motor stage is one, the displacement is the same in the two direction and only the equi- |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.