source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Neutral%20axis | The neutral axis is an axis in the cross section of a beam (a member resisting bending) or shaft along which there are no longitudinal stresses or strains.
Theory
If the section is symmetric, isotropic and is not curved before a bend occurs, then the neutral axis is at the geometric centroid of a beam or shaft. All fibers on one side of the neutral axis are in a state of tension, while those on the opposite side are in compression.
Since the beam is undergoing uniform bending, a plane on the beam remains plane. That is:
Where is the shear strain and is the shear stress
There is a compressive (negative) strain at the top of the beam, and a tensile (positive) strain at the bottom of the beam. Therefore by the Intermediate Value Theorem, there must be some point in between the top and the bottom that has no strain, since the strain in a beam is a continuous function.
Let L be the original length of the beam (span)
ε(y) is the strain as a function of coordinate on the face of the beam.
σ(y) is the stress as a function of coordinate on the face of the beam.
ρ is the radius of curvature of the beam at its neutral axis.
θ is the bend angle
Since the bending is uniform and pure, there is therefore at a distance y from the neutral axis with the inherent property of having no strain:
Therefore the longitudinal normal strain varies linearly with the distance y from the neutral surface. Denoting as the maximum strain in the beam (at a distance c from the neutral axis), it becomes clear that:
Therefore, we can solve for ρ, and find that:
Substituting this back into the original expression, we find that:
Due to Hooke's Law, the stress in the beam is proportional to the strain by E, the modulus of elasticity:
Therefore:
From statics, a moment (i.e. pure bending) consists of equal and opposite forces. Therefore, the total amount of force across the cross section must be 0.
Therefore:
Since y denotes the distance from the neutral axis to any point on the face, it |
https://en.wikipedia.org/wiki/Laboratory%20oven | Laboratory ovens are a common piece of equipment that can be found in electronics, materials processing, forensic, and research laboratories. These ovens generally provide pinpoint temperature control and uniform temperatures throughout the heating process. The following applications are some of the common uses for laboratory ovens: annealing, die-bond curing, drying or dehydrating, Polyimide baking, sterilizing, evaporating. Typical sizes are from one cubic foot to . Some ovens can reach temperatures that are higher than 300 degrees Celsius. These temperatures are then applied from all sides of the oven to provide constant heat to sample.
Laboratory ovens can be used in numerous different applications and configurations, including clean rooms, forced convection, horizontal airflow, inert atmosphere, natural convection, and pass through.
There are many types of laboratory ovens that are used throughout laboratories. Standard digital ovens are mainly used for drying and heating processes while providing temperature control and safety. Heavy duty ovens are used more in the industrial laboratories and provide testing and drying for biological samples. High temperature ovens are custom built and have additional insulation lining. This is needed for the oven due to its high temperatures that can reach up to 500 degrees Celsius. Other forms of the laboratory oven include vacuum ovens, forced air convection ovens, and gravity convection ovens.
Forensic labs use vacuum ovens that have been configured in specific ways to assist in developing fingerprints. Gravity convection ovens are used for biological purposes such as removing biological contaminants from samples. Along with forced-air ovens, they are also used in environmental studies to dry out samples that have been taken. These samples are weighed before and after to calculate the amount of moisture in the sample.
Laboratory Oven Safety
Laboratory ovens contain many components and other procedures that can be ha |
https://en.wikipedia.org/wiki/Piggybacking%20%28Internet%20access%29 | Piggybacking on Internet access is the practice of establishing a wireless Internet connection by using another subscriber's wireless Internet access service without the subscriber's explicit permission or knowledge. It is a legally and ethically controversial practice, with laws that vary by jurisdiction around the world. While completely outlawed or regulated in some places, it is permitted in others.
A customer of a business providing hotspot service, such as a hotel or café, is generally not considered to be piggybacking, though non-customers or those outside the premises who are simply in reach may be. Many such locations provide wireless Internet access as a free or paid-for courtesy to their patrons or simply to draw people to the area. Others near the premises may be able to gain access.
Piggybacking is distinct from wardriving, which involves only the logging or mapping of the existence of access points.
Background
Piggybacking has become a widespread practice in the 21st century due to the advent of wireless Internet connections and wireless access points. Computer users who either do not have their own connections or who are outside the range of their own might find someone else's by wardriving or luck and use that one.
However, those residing near a hotspot or another residence with the service have been found to have the ability to piggyback off such connections without patronizing these businesses, which has led to more controversy. While some may be in reach from their own home or nearby, others may be able to do so from the parking lot of such an establishment, from another business that generally tolerates the user's presence, or from the public domain. Others, especially those living in apartments or town houses, may find themselves able to use a neighbour's connection.
Wi-Fi hotspots, unsecured and secured, have been recorded to some degree with GPS-coordinates. Some sites host searchable databases or maps of the locations of user-submitted a |
https://en.wikipedia.org/wiki/SMPTE%20344M | SMPTE 344M is a standard published by SMPTE which expands upon SMPTE 259M allowing for bit-rates of 540 Mbit/s, allowing EDTV resolutions of 480p and 576p.
This standard is part of a family of standards that define a serial digital interface. |
https://en.wikipedia.org/wiki/SMPTE%20424M | SMPTE 424M is a standard published by SMPTE which expands upon SMPTE 259M, SMPTE 344M, and SMPTE 292M allowing for bit-rates of 2.970 Gbit/s and 2.970/1.001 Gbit/s over a single-link coaxial cable. These bit-rates are sufficient for 1080p video at 50 or 60 frames per second. The initial 424M standard was published in 2006, with a revision published in 2012 (SMPTE ST 424:2012). This standard is part of a family of standards that define a serial digital interface (SDI); it is commonly known as 3G-SDI.
Formats
Within this standard, there are three formats:
Level A format is the direct mapping of uncompressed 1080p (up to 60 fps) video into a serial digital interface at the nominal 3 Gbit/s. That is, one video signal, one video stream, in one cable.
Level B-DL format is the mapping of dual-link HD-SDI/SMPTE 372M (i.e.: 1080p up to 60 fps) in a single serial digital interface at the nominal 3 Gbit/s. That is, one video signal, two streams, in one cable.
Level B-DS format is the dual-stream carriage of two independent HD-SDI/SMPTE 292M signals (720p up to 60 fps or 1080i/1080p up to 30 fps) in a single serial digital interface at the nominal 3 Gbit/s. That is, two video signals, two video streams, in one cable. |
https://en.wikipedia.org/wiki/SMPTE%20372M | SMPTE 372M is a standard published by SMPTE which expands upon SMPTE 259M, SMPTE 344M, and SMPTE 292M allowing for bit-rates of 2.970 Gbit/s, and 2.970/1.001 Gbit/s over two wires. These bit-rates are sufficient for 1080p video at 50 or 60 frames per second.
This standard is informally known as dual-link HD-SDI and is part of a family of standards that define a Serial Digital Interface. |
https://en.wikipedia.org/wiki/Roskamp%20Institute | The Roskamp Institute, was co-founded by Robert and Diane Roskamp, and Fiona Crawford and Michael Mullan in Sarasota, Florida in 2003. It is a nonprofit biomedical research facility specializing neurological research including Alzheimer's disease, traumatic brain injury, Gulf War syndrome, and posttraumatic stress disorder. It also operates an onsite neurology clinic. The institute is focused on finding the causes and treatments for neuropsychiatric and neurodegenerative diseases.
The institute's lead researchers, Michael Mullan and Fiona Crawford, were members of a team of scientists who discovered the first genetic errors causing Alzheimer's disease in 1991 in the APP gene in early onset familial cases. Mullan and Crawford also discovered the Swedish mutation which has been incorporated into transgenic mice which are widely used to understand the disease and test new treatments.
The institute is particularly focused on translational research that can lead to novel drug or other therapeutic interventions in neurodegenerative disorders. In this regard, Institute scientists discovered that certain members of a class of drugs called dihydropyridines [DHPs] can lower the levels of amyloid beta in the brains of transgenic models of the disease and decided to take one of them, nilvadipine, forward into clinical trials for Alzheimer's disease. This work was conducted by Archer Pharmaceuticals, a for-profit spin off of the institute, headed by Mullan. In partnership with colleagues at Trinity College, Dublin led by Brian Lawlor, Archer and Institute scientists conducted an open label phase I/II trial of nilvadipine in mild to moderate Alzheimer's disease subjects. More recently, in collaboration with multiple partners at academic institutes in Europe, and again led by Lawlor, Archer and Roskamp Institute scientists partnered to conduct a phase III clinical trial of nilvadipine in mild to moderate Alzheimer's disease.
The institute is currently housed in a scientific r |
https://en.wikipedia.org/wiki/Olericulture | Olericulture is the science of vegetable growing, dealing with the culture of non-woody (herbaceous) plants for food.
Olericulture is the production of plants for use of the edible parts. Vegetable crops can be classified into nine major categories:
Potherbs and greens – spinach and collards
Salad crops – lettuce, celery
Cole crops – cabbage and cauliflower
Root crops (tubers) – potatoes, beets, carrots, radishes
Bulb crops – onions, leeks
Legumes – beans, peas
Cucurbits – melons, squash, cucumber
Solanaceous crops – tomatoes, peppers, potatoes
Sweet corn
Olericulture deals with the production, storage, processing and marketing of vegetables. It encompasses crop establishment, including cultivar selection, seedbed preparation and establishment of vegetable crops by seed and transplants.
It also includes maintenance and care of vegetable crops as well commercial and non-traditional vegetable crop production including organic gardening and organic farming; sustainable agriculture and horticulture; hydroponics; and biotechnology.
See also
Agriculture – the cultivation of animals, plants, fungi and other life forms for food, fiber, and other products used to sustain life.
Horticulture – the industry and science of plant cultivation including the process of preparing soil for the planting of seeds, tubers, or cuttings.
Pomology – a branch of botany that studies and cultivates pome fruit, and sometimes applied more broadly, to the cultivation of any type of fruit.
Tropical horticulture – a branch of horticulture that studies and cultivates garden plants in the tropics, i.e., the equatorial regions of the world. |
https://en.wikipedia.org/wiki/General%20set%20theory | General set theory (GST) is George Boolos's (1998) name for a fragment of the axiomatic set theory Z. GST is sufficient for all mathematics not requiring infinite sets, and is the weakest known set theory whose theorems include the Peano axioms.
Ontology
The ontology of GST is identical to that of ZFC, and hence is thoroughly canonical. GST features a single primitive ontological notion, that of set, and a single ontological assumption, namely that all individuals in the universe of discourse (hence all mathematical objects) are sets. There is a single primitive binary relation, set membership; that set a is a member of set b is written a ∈ b (usually read "a is an element of b").
Axioms
The symbolic axioms below are from Boolos (1998: 196), and govern how sets behave and interact.
As with Z, the background logic for GST is first order logic with identity. Indeed, GST is the fragment of Z obtained by omitting the axioms Union, Power Set, Elementary Sets (essentially Pairing) and Infinity and then taking a theorem of Z, Adjunction, as an axiom.
The natural language versions of the axioms are intended to aid the intuition.
1) Axiom of Extensionality: The sets x and y are the same set if they have the same members.
The converse of this axiom follows from the substitution property of equality.
2) Axiom Schema of Specification (or Separation or Restricted Comprehension): If z is a set and is any property which may be satisfied by all, some, or no elements of z, then there exists a subset y of z containing just those elements x in z which satisfy the property . The restriction to z is necessary to avoid Russell's paradox and its variants. More formally, let be any formula in the language of GST in which x may occur freely and y does not. Then all instances of the following schema are axioms:
3) Axiom of Adjunction: If x and y are sets, then there exists a set w, the adjunction of x and y, whose members are just y and the members of x.
Adjunction refers to an |
https://en.wikipedia.org/wiki/Protected%20Media%20Path | The Protected Media Path is a set of technologies creating a "Protected Environment," first included in Microsoft's Windows Vista operating system, that is used to enforce digital rights management (or DRM) protections on content.
Its subsets are Protected Video Path (PVP) and Protected User Mode Audio (PUMA). Any application that uses Protected Media Path in Windows uses Media Foundation.
Overview
The protected environment in which DRM content is played contains the media components that play DRM content, so the application only needs to provide remote control (play, rewind, pause, and so on), rather than having to handle unprotected content data. The protected environment also provides all the necessary support for Microsoft-approved (signed) third-party software modules to be added. It provides a "wall" against outside copying, where within the walls, content can be processed without making the content available to unapproved software.
In order to prevent users from copying DRM content, Windows Vista provides process isolation and continually monitors what kernel-mode software is loaded. If an unverified component is detected, then Vista will stop playing DRM content, rather than risk having the content copied. The protected environment is implemented completely in software, so software-based attacks such as patching the Windows kernel are possible.
These restrictions concern the various outputs from the PC. For DRM content, digital outputs such as Digital Visual Interface (DVI) and High Definition Multimedia Interface (HDMI) will have High-bandwidth Digital Content Protection (HDCP) enabled, to prevent someone from recording the digital stream. Even analog TV-style outputs typically require some restrictions, provided by mechanisms such as Macrovision and CGMS-A. These restrictions only apply to DRM-restricted content, such as HD DVD or Blu-ray that are encrypted with AACS, and also apply in Windows XP using supported playback applications. Users' standard u |
https://en.wikipedia.org/wiki/Schild%20equation | In pharmacology, Schild regression analysis, based upon the Schild equation, both named for Heinz Otto Schild, are tools for studying the effects of agonists and antagonists on the response caused by the receptor or on ligand-receptor binding.
Concept
Dose-response curves can be constructed to describe response or ligand-receptor complex formation as a function of the ligand concentration. Antagonists make it harder to form these complexes by inhibiting interactions of the ligand with its receptor. This is seen as a change in the dose response curve: typically a rightward shift or a lowered maximum. A reversible competitive antagonist should cause a rightward shift in the dose response curve, such that the new curve is parallel to the old one and the maximum is unchanged. This is because reversible competitive antagonists are surmountable antagonists. The magnitude of the rightward shift can be quantified with the dose ratio, r. The dose ratio r is the ratio of the dose of agonist required for half maximal response with the antagonist present divided by the agonist required for half maximal response without antagonist ("control"). In other words, the ratio of the EC50s of the inhibited and un-inhibited curves. Thus, r represents both the strength of an antagonist and the concentration of the antagonist that was applied. An equation derived from the Gaddum equation can be used to relate r to , as follows:
where
r is the dose ratio
is the concentration of the antagonist
is the equilibrium constant of the binding of the antagonist to the receptor
A Schild plot is a double logarithmic plot, typically as the ordinate and as the abscissa. This is done by taking the base-10 logarithm of both sides of the previous equation after subtracting 1:
This equation is linear with respect to , allowing for easy construction of graphs without computations. This was particular valuable before the use of computers in pharmacology became widespread. The y-intercept of the equat |
https://en.wikipedia.org/wiki/Mental%20protuberance | The symphysis of the external surface of the mandible divides below and encloses a triangular eminence, the mental protuberance, the base of which is depressed in the center but raised on either side to form the mental tubercle. The size and shape of the bones making up this structure are responsible for the size and shape of a person's chin. Synonyms of mental protuberance include mental process and protuberantia mentalis. Mental in this sense derives from Latin mentum (chin), not mens (mind), source of the more common meaning of mental. |
https://en.wikipedia.org/wiki/CD19 | B-lymphocyte antigen CD19, also known as CD19 molecule (Cluster of Differentiation 19), B-Lymphocyte Surface Antigen B4, T-Cell Surface Antigen Leu-12 and CVID3 is a transmembrane protein that in humans is encoded by the gene CD19. In humans, CD19 is expressed in all B lineage cells. Contrary to some early doubts, human plasma cells do express CD19, as confirmed by others. CD19 plays two major roles in human B cells: on the one hand, it acts as an adaptor protein to recruit cytoplasmic signaling proteins to the membrane; on the other, it works within the CD19/CD21 complex to decrease the threshold for B cell receptor signaling pathways. Due to its presence on all B cells, it is a biomarker for B lymphocyte development, lymphoma diagnosis and can be utilized as a target for leukemia immunotherapies.
Structure
In humans, CD19 is encoded by the 7.41 kilobase CD19 gene located on the short arm of chromosome 16. It contains at least fifteen exons, four that encode extracellular domain and nine that encode cytoplasmic domains, with a total of 556 amino acids. Experiments show that there are multiple mRNA transcripts; however, only two have been isolated in vivo.
CD19 is a 95 kd Type I transmembrane glycoprotein in the immunoglobulin superfamily (IgSF) with two extracellular C2-set Ig-like domains and a relatively large, 240 amino acid, cytoplasmic tail that is highly conserved among mammalian species. The extracellular C2-type Ig-like domains are divided by a potential disulfide linked non-Ig-like domain and N-linked carbohydrate addition sites. The cytoplasmic tail contains at least nine tyrosine residues near the C-terminus. Within these residues, Y391, Y482, and Y513 have been shown to be essential to the biological functions of CD19. Phenylalanine substitution for tyrosine at Y482 and Y513 leads to the inhibition of phosphorylation at the other tyrosines.
Expression
CD19 is widely expressed during all phases of B cell development until terminal differentiation i |
https://en.wikipedia.org/wiki/Chromosome%20combing | Chromosome combing (also known as molecular combing or DNA combing) is a technique used to produce an array of uniformly stretched DNA that is then highly suitable for nucleic acid hybridization studies such as fluorescent in situ hybridisation (FISH) which benefit from the uniformity of stretching, the easy access to the hybridisation target sequences, and the resolution offered by the large distance between two probes, which is due to the stretching of the DNA by a factor of 1.5 times the crystallographic length of DNA.
DNA in solution (i.e. with a randomly-coiled structure) is stretched by retracting the meniscus of the solution at a constant rate (typically 300 µm/s). The ends of DNA strands, which are thought to be frayed (i.e. open and exposing polar groups) bind to ionisable groups coating a silanized glass plate at a pH below the pKa of the ionizable groups (ensuring they are charged enough to interact with the ends of DNA). The rest of the DNA, which is mostly dsDNA, cannot form these interactions (aside from a few ‘touch down’ segments along the length of the DNA strand) so is available for hybridisation to probes. As the meniscus retracts, surface retention creates a force that acts on DNA to retain it in the liquid phase; however this force is inferior to the strength of the DNA’s attachment; the result is that the DNA is stretched as it enters the air phase; as the force acts in the locality of the air/liquid phase, it is invariant to different lengths or conformations of the DNA in solution, so DNA of any length will be stretched the same as the meniscus retracts. As this stretching is constant along the length of a DNA, distance along the strand can be related to base content; 1 µm is approximately equivalent to 2 kb.
DNA regions of interest are observed by hybridising them with probes labelled by haptens like biotin; this can then be bound by one or more layers of fluorochrome-associated ligands (such as immunofluorescence antibodies). Multicolour |
https://en.wikipedia.org/wiki/Support%20function | In mathematics, the support function hA of a non-empty closed convex set A in
describes the (signed) distances of supporting hyperplanes of A from the origin. The support function is a convex function on .
Any non-empty closed convex set A is uniquely determined by hA. Furthermore, the support function, as a function of the set A, is compatible with many natural geometric operations, like scaling, translation, rotation and Minkowski addition.
Due to these properties, the support function is one of the most central basic concepts in convex geometry.
Definition
The support function
of a non-empty closed convex set A in is given by
; see
. Its interpretation is most intuitive when x is a unit vector:
by definition, A is contained in the closed half space
and there is at least one point of A in the boundary
of this half space. The hyperplane H(x) is therefore called a supporting hyperplane
with exterior (or outer) unit normal vector x.
The word exterior is important here, as
the orientation of x plays a role, the set H(x) is in general different from H(-x).
Now hA is the (signed) distance of H(x) from the origin.
Examples
The support function of a singleton A={a} is .
The support function of the Euclidean unit ball is where is the 2-norm.
If A is a line segment through the origin with endpoints -a and a then .
Properties
As a function of x
The support function of a compact nonempty convex set is real valued and continuous, but if the
set is closed and unbounded, its support function is extended real valued (it takes the value
). As any nonempty closed convex set is the intersection of
its supporting half spaces, the function hA determines A uniquely.
This can be used to describe certain geometric properties of convex sets analytically.
For instance, a set A is point symmetric with respect to the origin if and only if hA
is an even function.
In general, the support function is not differentiable.
However, directional derivatives |
https://en.wikipedia.org/wiki/Codex%20Mexicanus | The Codex Mexicanus is an early colonial Mexican pictorial manuscript.
The Codex can be divided into several sections:
The saints, the European calendar and zodiac.
The Aztec calendar.
Accounts in the Aztec pictographic writing system.
A family tree of the rulers of Mexico.
The history of the Mexica from their departure from Aztlan.
Colonial history.
Two Christian scenes: the Temptation of Christ and the Adoration.
A tonalamatl. This last section is incomplete.
It is currently held in the Bibliothèque Nationale, Paris.
See also
Aztec codices
Codex Vaticanus B |
https://en.wikipedia.org/wiki/De%20novo%20protein%20structure%20prediction | In computational biology, de novo protein structure prediction refers to an algorithmic process by which protein tertiary structure is predicted from its amino acid primary sequence. The problem itself has occupied leading scientists for decades while still remaining unsolved. According to Science, the problem remains one of the top 125 outstanding issues in modern science. At present, some of the most successful methods have a reasonable probability of predicting the folds of small, single-domain proteins within 1.5 angstroms over the entire structure.
De novo methods tend to require vast computational resources, and have thus only been carried out for relatively small proteins. De novo protein structure modeling is distinguished from Template-based modeling (TBM) by the fact that no solved homologue to the protein of interest is used, making efforts to predict protein structure from amino acid sequence exceedingly difficult. Prediction of protein structure de novo for larger proteins will require better algorithms and larger computational resources such as those afforded by either powerful supercomputers (such as Blue Gene or MDGRAPE-3) or distributed computing projects (such as Folding@home, Rosetta@home, the Human Proteome Folding Project, or Nutritious Rice for the World). Although computational barriers are vast, the potential benefits of structural genomics (by predicted or experimental methods) to fields such as medicine and drug design make de novo structure prediction an active research field.
Background
Currently, the gap between known protein sequences and confirmed protein structures is immense. At the beginning of 2008, only about 1% of the sequences listed in the UniProtKB database corresponded to structures in the Protein Data Bank (PDB), leaving a gap between sequence and structure of approximately five million. Experimental techniques for determining tertiary structure have faced serious bottlenecks in their ability to determine structures for p |
https://en.wikipedia.org/wiki/Bitfrost | Bitfrost is the security design specification for the OLPC XO, a low cost laptop intended for children in developing countries and developed by the One Laptop Per Child (OLPC) project. Bitfrost's main architect is Ivan Krstić. The first public specification was made available in February 2007.
Bitfrost architecture
Passwords
No passwords are required to access or use the computer.
System of rights
Every program, when first installed, requests certain bundles of rights, for instance "accessing the camera", or "accessing the internet". The system keeps track of these rights, and the program is later executed in an environment which makes only the requested resources available. The implementation is not specified by Bitfrost, but dynamic creation of security contexts is required. The first implementation was based on vserver, the second and current implementation is based on user IDs and group IDs (/etc/password is edited when an activity is started), and a future implementation might involve SE Linux or some other technology.
By default, the system denies certain combinations of rights; for instance, a program would not be granted both the right to access the camera and to access the internet. Anybody can write and distribute programs that request allowable right combinations. Programs that require normally unapproved right combinations need a cryptographic signature by some authority. The laptop's user can use the built-in security panel to grant additional rights to any application.
Modifying the system
The users can modify the laptop's operating system, a special version of Fedora Linux running the new Sugar graphical user interface and operating on top of Open Firmware. The original system remains available in the background and can be restored.
By acquiring a developer key from a central location, a user may even modify the background copy of the system and many aspects of the BIOS. Such a developer key is only given out after a waiting period (so that thef |
https://en.wikipedia.org/wiki/Radical%20of%20a%20module | In mathematics, in the theory of modules, the radical of a module is a component in the theory of structure and classification. It is a generalization of the Jacobson radical for rings. In many ways, it is the dual notion to that of the socle soc(M) of M.
Definition
Let R be a ring and M a left R-module. A submodule N of M is called maximal or cosimple if the quotient M/N is a simple module. The radical of the module M is the intersection of all maximal submodules of M,
Equivalently,
These definitions have direct dual analogues for soc(M).
Properties
In addition to the fact rad(M) is the sum of superfluous submodules, in a Noetherian module rad(M) itself is a superfluous submodule.
A ring for which rad(M) = {0} for every right R-module M is called a right V-ring.
For any module M, rad(M/rad(M)) is zero.
M is a finitely generated module if and only if the cosocle M/rad(M) is finitely generated and rad(M) is a superfluous submodule of M.
See also
Socle (mathematics)
Jacobson radical |
https://en.wikipedia.org/wiki/Cayley%27s%20mousetrap | Mousetrap is the name of a game introduced by the English mathematician Arthur Cayley. In the game, cards numbered through ("say thirteen" in Cayley's original article) are shuffled to place them in some random permutation and are arranged in a circle with their faces up. Then, starting with the first card, the player begins counting and moving to the next card as the count is incremented. If at any point the player's current count matches the number on the card currently being pointed to, that card is removed from the circle and the player starts all over at on the next card. If the player ever removes all of the cards from the permutation in this manner, then the player wins. If the player reaches the count and cards still remain, then the game is lost.
In order for at least one card to be removed, the initial permutation of the cards must not be a derangement. However, this is not a sufficient condition for winning, because it does not take into account subsequent removals. The number of ways the cards can be arranged such that the entire game is won, for n = 1, 2, ..., are
1, 1, 2, 6, 15, 84, 330, 1812, 9978, 65503, ... .
For example with four cards, the probability of winning is 0.25, but this reduces as the number of cards increases, and with thirteen cards it is about 0.0046. |
https://en.wikipedia.org/wiki/Pauli%20equation | In quantum mechanics, the Pauli equation or Schrödinger–Pauli equation is the formulation of the Schrödinger equation for spin-½ particles, which takes into account the interaction of the particle's spin with an external electromagnetic field. It is the non-relativistic limit of the Dirac equation and can be used where particles are moving at speeds much less than the speed of light, so that relativistic effects can be neglected. It was formulated by Wolfgang Pauli in 1927.
Equation
For a particle of mass and electric charge , in an electromagnetic field described by the magnetic vector potential and the electric scalar potential , the Pauli equation reads:
Here are the Pauli operators collected into a vector for convenience, and is the momentum operator in position representation. The state of the system, (written in Dirac notation), can be considered as a two-component spinor wavefunction, or a column vector (after choice of basis):
.
The Hamiltonian operator is a 2 × 2 matrix because of the Pauli operators.
Substitution into the Schrödinger equation gives the Pauli equation. This Hamiltonian is similar to the classical Hamiltonian for a charged particle interacting with an electromagnetic field. See Lorentz force for details of this classical case. The kinetic energy term for a free particle in the absence of an electromagnetic field is just where is the kinetic momentum, while in the presence of an electromagnetic field it involves the minimal coupling , where now is the kinetic momentum and is the canonical momentum.
The Pauli operators can be removed from the kinetic energy term using the Pauli vector identity:
Note that unlike a vector, the differential operator has non-zero cross product with itself. This can be seen by considering the cross product applied to a scalar function :
where is the magnetic field.
For the full Pauli equation, one then obtains
for which only a few analytic results are known, e.g., in the context of Landau |
https://en.wikipedia.org/wiki/Carboxyfluorescein%20succinimidyl%20ester | Carboxyfluorescein succinimidyl ester (CFSE) is a fluorescent cell staining dye. CFSE is cell permeable and covalently couples, via its succinimidyl group, to intracellular molecules, notably, to intracellular lysine residues and other amine sources. Due to this covalent coupling reaction, fluorescent CFSE can be retained within cells for extremely long periods. Also, due to this stable linkage, once incorporated within cells, the dye is not transferred to adjacent cells.
CFSE is commonly confused with carboxyfluorescein diacetate succinimidyl ester (CFDA-SE), although they are not strictly the same molecule; CFDA-SE, due to its acetate groups, is highly cell permeable, while CFSE is much less so. As CFDA-SE, which is non-fluorescent, enters the cytoplasm of cells, intracellular esterases remove the acetate groups and convert the molecule to the fluorescent ester.
CFSE was originally developed as a fluorescent dye that could be used to stably label lymphocytes and track their migration within animals for many months. Subsequent studies revealed that the dye can be used to monitor lymphocyte proliferation, both in vitro and in vivo, due to the progressive halving of CFSE fluorescence within daughter cells following each cell division. The only limitation is that CFSE at high concentrations can be toxic for cells. However, when CFSE labelling is performed optimally, approximately 7-8 cell divisions can be identified before the CFSE fluorescence is too low to be distinguished above the autofluorescence background. Thus CFSE represents an extremely valuable fluorescent dye for immunological studies, allowing lymphocyte proliferation, migration and positioning to be simultaneously monitored. By the use of fluorescent antibodies against different lymphocyte cell surface markers it is also possible to follow the proliferation behaviour of different lymphocyte subsets. In addition, unlike other methods, CFSE-labeled viable cells can be recovered for further analysis.
Sin |
https://en.wikipedia.org/wiki/Revision%20tag | A revision tag is a textual label that can be associated with a specific revision of a project maintained by a version control system. This allows the user to define a meaningful name to be given to a particular state of a project that is under version control. This label can then be used in place of the revision identifier for commands supported by the version control system.
For example, in software development, a tag may be used to identify a specific release of the software such as "version 1.2".
See also
Trunk (software)
Branching (version control)
Tag (metadata) |
https://en.wikipedia.org/wiki/Retired%20number | Retiring the number of an athlete is an honor a team bestows upon a player, usually after the player has left the team, retires from the sport, or dies, by taking his or her former number out of circulation. Once a number is retired, no future player from the team may wear that number on their uniform, unless the player so-honored permits it; however, in many cases the number cannot be used at all. Such an honor may also be bestowed on players who had highly memorable careers, died prematurely under tragic circumstances, or have had their promising careers ended by serious injury. Some sports that retire team numbers include baseball, cricket, ice hockey, basketball, American football, and association football. Retired jerseys are often referred to as "hanging from the rafters" as they are, literally, put to hang in the team's home venue.
The first number officially retired by a team in a professional sport was that of ice hockey player Ace Bailey, whose number 6 was retired by the Toronto Maple Leafs in 1934. Some teams have also retired number 12 in honor of their fans, or the "twelfth man". Similarly, the Sacramento Kings and Orlando Magic retired number 6 in honor of their fans, the "sixth man". In some cases, a team may decide to retire a number in honor of tragedies involving the team's city or state. For example, the number 58 was retired in 2018 by the Vegas Golden Knights hockey team in honor of the 58 victims killed in the 2017 Las Vegas shooting.
North American sports leagues
If a jersey is retired and an active player is still wearing it, the player is usually permitted to wear the number for his entire career as a player. If in the sport, managers and coaches wear uniform numbers, and the player later becomes a coach for the same team, he is also permitted to wear it as a coach.
However, in some cases the player may still elect to change their number. For instance, in 1987 the Boston Bruins of the National Hockey League decided to retire jersey numb |
https://en.wikipedia.org/wiki/BSI%20Group | The British Standards Institution (BSI) is the national standards body of the United Kingdom. BSI produces technical standards on a wide range of products and services and also supplies certification and standards-related services to businesses.
History
BSI was founded as the Engineering Standards Committee in London in 1901. It subsequently extended its standardization work and became the British Engineering Standards Association in 1918, adopting the name British Standards Institution in 1931 after receiving a Royal Charter in 1929. In 1998 a revision of the Charter enabled the organization to diversify and acquire other businesses, and the trading name was changed to BSI Group.
The Group now operates in 195 countries. The core business remains standards and standards related services, although the majority of the Group's revenue comes from management systems assessment and certification work.
In 2021, BSI appointed its first female chief executive officer, Susan Taylor Martin.
Activities
BSI produces British Standards, and, as the UK's National Standards Body, is also responsible for the UK publication, in English, of international and European standards. BSI is obliged to adopt and publish all European Standards as identical British Standards (prefixed BS EN) and to withdraw pre-existing British Standards that are in conflict. However, it has the option to adopt and publish international standards (prefixed BS ISO or BS IEC).
In response to commercial demands, BSI also produces commissioned standards products such as Publicly Available Specifications, (PASs), Private Standards and Business Information Publications. These products are commissioned by individual organizations and trade associations to meet their needs for standardized specifications, guidelines, codes of practice etc. Because they are not subject to the same consultation and consensus requirements as formal standards, the lead time is shorter.
BSI also publishes standards-related books, CD |
https://en.wikipedia.org/wiki/Dent%27s%20disease | Dent's disease (or Dent disease) is a rare X-linked recessive inherited condition that affects the proximal renal tubules of the kidney. It is one cause of Fanconi syndrome, and is characterized by tubular proteinuria, excess calcium in the urine, formation of calcium kidney stones, nephrocalcinosis, and chronic kidney failure.
"Dent's disease" is often used to describe an entire group of familial disorders, including X-linked recessive nephrolithiasis with kidney failure, X-linked recessive hypophosphatemic rickets, and both Japanese and idiopathic low-molecular-weight proteinuria. About 60% of patients have mutations in the CLCN5 gene (Dent 1), which encodes a kidney-specific chloride/proton antiporter, and 15% of patients have mutations in the OCRL1 gene (Dent 2).
Signs and symptoms
Dent's disease often produces the following signs and symptoms:
Extreme thirst combined with dehydration, which leads to frequent urination
Nephrolithiasis (kidney stones)
Hypercalciuria (high urine calcium - >300 mg/d or >4 mg/kg per d with normal levels blood/serum calcium)
Aminoaciduria (amino acids in urine)
Phosphaturia (phosphate in urine)
Glycosuria (glucose in urine)
Kaliuresis (potassium in urine)
Hyperuricosuria (excessive amounts of uric acid in the urine)
Impaired urinary acidification
Rickets
In a study of 25 patients with Dent's disease, 9 of 15 men, and one of 10 women had end-stage kidney disease by the age of 47.
Genetics
Dent disease 1
Dent's disease is a X-linked recessive disorder. The males are prone to manifesting symptoms in early adulthood with symptoms of calculi, rickets or even with kidney failure in more severe cases.
In humans, gene CLCN5 is located on chromosome Xp11.22, and has a 2238-bp coding sequence that consists of 11 exons that span 25 to 30 kb of genomic DNA and encode a 746-amino-acid protein. CLCN5 belongs to the family of voltage-gated chloride channel genes (CLCN1-CLCN7, CLCKa and CLCKb) that have about 12 transmembrane domai |
https://en.wikipedia.org/wiki/Spatial%20data%20infrastructure | A Spatial Data Infrastructure (SDI), also called geospatial data infrastructure, is a data infrastructure implementing a framework of geographic data, metadata, users and tools that are interactively connected in order to use spatial data in an efficient and flexible way. Another definition is "the technology, policies, standards, human resources, and related activities necessary to acquire, process, distribute, use, maintain, and preserve spatial data".
A further definition is given in Kuhn (2005): "An SDI is a coordinated series of agreements on technology standards, institutional arrangements, and policies that enable the discovery and use of geospatial information by users and for purposes other than those it was created for."
General
Some of the main principles are that data and metadata should not be managed centrally, but by the data originator and/or owner, and that tools and services connect via computer networks to the various sources. A GIS is often the platform for deploying an individual node within an SDI. To achieve these objectives, good coordination between all the actors is necessary and the definition of standards is very important.
Due to its nature (size, cost, number of t-related. An example of an existing SDI, since 2002, is the National Spatial Data Infrastructure (NSDI) created by the OMB Circular A-16 in the United States. At the European side, since 2007, the INSPIRE is a European Commission initiative to build a European SDI beyond national boundaries and ultimately the United Nations Spatial Data Infrastructure (UNSDI) will do the same for over 30 UN Funds, Programmes, Specialized Agencies and member countries.
Software components
An SDI should enable the discovery and delivery of spatial data from a data repository, via a spatial service provider, to a user. As mentioned earlier it is often wished that the data provider is able to update spatial data stored in a repository. Hence, the basic software components of an SDI are:
|
https://en.wikipedia.org/wiki/Landau%E2%80%93Pomeranchuk%E2%80%93Migdal%20effect | In high-energy physics, the Landau–Pomeranchuk–Migdal effect, also known as the Landau–Pomeranchuk effect and the Pomeranchuk effect, or simply LPM effect, is a reduction of the bremsstrahlung and pair production cross sections at high energies or high matter densities. It is named in honor to Lev Landau, Isaak Pomeranchuk and Arkady Migdal.
Overview
A high energy particle undergoing multiple soft scatterings from a medium will experience interference effects between adjacent scattering sites. From uncertainty as the longitudinal momentum transfer gets small the particles wavelength will increase, if the wavelength becomes longer than the mean free path in the medium (the average distance between scattering sites) then the scatterings can no longer be treated as independent events, this is the LPM effect. The Bethe–Heitler spectrum for multiple scattering induced radiation assumes that the scatterings are independent, the quantum interference between successive scatterings caused by the LPM effect leads to suppression of the radiation spectrum relative to that predicted by Bethe–Heitler.
The suppression occurs in different parts of the emission spectrum, for quantum electrodynamics (QED) small photon energies are suppressed, and for quantum chromodynamics (QCD) large gluon energies are suppressed. In QED the rescattering of the high energy electron dominates the process, in QCD the emitted gluons carry color charge and interact with the medium also. Since the gluons are soft their rescattering will provide the dominant modification to the spectrum.
Lev Landau and Isaak Pomeranchuk showed that the formulas for bremsstrahlung and pair creation in matter which had been formulated by Hans Bethe and Walter Heitler (the Bethe–Heitler formula) were inapplicable at high energy or high matter density. The effect of multiple Coulomb scattering by neighboring atoms reduces the cross sections for pair production and bremsstrahlung. Arkady Migdal developed a formula applicabl |
https://en.wikipedia.org/wiki/Directed%20infinity | A directed infinity is a type of infinity in the complex plane that has a defined complex argument θ but an infinite absolute value r. For example, the limit of 1/x where x is a positive real number approaching zero is a directed infinity with argument 0; however, 1/0 is not a directed infinity, but a complex infinity. Some rules for manipulation of directed infinities (with all variables finite) are:
Here, sgn(z) = is the complex signum function.
See also
Point at infinity |
https://en.wikipedia.org/wiki/The%20Ultimate%20Resource | The Ultimate Resource is a 1981 book written by Julian Lincoln Simon challenging the notion that humanity was running out of natural resources. It was revised in 1996 as The Ultimate Resource 2.
Overview
The overarching thesis on why there is no resource crisis is that as a particular resource becomes more scarce, its price rises. This price rise creates an incentive for people to discover more of the resource, ration and recycle it, and eventually, develop substitutes. The "ultimate resource" is not any particular physical object but the capacity for humans to invent and adapt.
Scarcity
The work opens with an explanation of scarcity, noting its relation to price; high prices denote relative scarcity and low prices indicate abundance. Simon usually measures prices in wage-adjusted terms, since this is a measure of how much labor is required to purchase a fixed amount of a particular resource. Since prices for most raw materials (e.g., copper) have fallen between 1800 and 1990 (adjusting for wages and adjusting for inflation), Simon argues that this indicates that those materials have become less scarce.
Forecasting
Simon makes a distinction between "engineering” and "economic" forecasting. Engineering forecasting consists of estimating the amount of known physical amount of resources, extrapolates the rate of use from current use and subtracts one from the other. Simon argues that these simple analyses are often wrong. While focusing only on proven resources is helpful in a business context, it is not appropriate for economy-wide forecasting. There exist undiscovered sources, sources not yet economically feasible to extract, sources not yet technologically feasible to extract, and ignored resources that could prove useful but are not yet worth trying to discover.
To counter the problems of engineering forecasting, Simon proposes economic forecasting, which proceeds in three steps in order to capture, in part, the unknowns the engineering method leaves out (p 27) |
https://en.wikipedia.org/wiki/Stacked%20Volumetric%20Optical%20Disc | The Stacked Volumetric Optical Disc (or SVOD) is an optical disc format developed by Hitachi Maxell, which uses an array of wafer-thin optical discs to allow data storage.
Each "layer" (a thin polycarbonate disc) holds around 9.4 GB of information, and the wafers are stacked in layers of 20, 25, 100, or more, giving a substantially larger overall data capacity; for example, 100× cartridges could hold 940 GB using the system as announced.
Hitachi Maxell announced the creation of the SVOD standard in 2006, intending to launch it the next year. Aimed primarily at commercial users, the target price was ¥40,000 for a cartridge of 100 thin discs, with the potential to expand into the home user market. When they announced the system, Hitachi Maxell publicly recognized the possibility that the system could be eventually modified for use with a blue-violet laser, similar to Blu-ray discs, which could have expanded the capacity of the system to 3-5 TB. It is possible that they in fact developed this "second generation" SVOD for use with standard Blu-ray lasers, with each thin disc having a storage capacity of 25 GB, or a 100-disc cartridge having a storage of 5 TB. Hitachi Maxell developed systems both for burning to the media using standard DVD optical heads, and pre-recording to the media using a special heat imprint technique they called "nanoimprinting." Though nanoimprinting initially required 6 minutes per disc for pressing, they had improved it to 8 seconds, and intended to achieve a comparable throughput to standard DVD pressing. The primary application of the SVOD system seemed to be business data archival, replacing digital tape archives.
In 2007, Japanese broadcaster NHK announced a similar system, based on Blu-ray discs, of stacked optical storage media specifically designed to rotate at high speeds, up to 15,000 RPM.
SVOD was anticipated to be a likely be a candidate, along with Holographic Versatile Discs (HVDs), to be a next-generation optical disc standa |
https://en.wikipedia.org/wiki/Foreign%20body%20reaction | A foreign body reaction (FBR) is a typical tissue response to a foreign body within biological tissue. It usually includes the formation of a foreign body granuloma. Tissue-encapsulation of an implant is an example, as is inflammation around a splinter. Foreign body granuloma formation consists of protein adsorption, macrophages, multinucleated foreign body giant cells (macrophage fusion), fibroblasts, and angiogenesis. It has also been proposed that the mechanical property of the interface between an implant and its surrounding tissues is critical for the host response.
In the long term, the foreign body reaction results in encapsulation of the foreign body within a calcified shell. For example, a lithopedion is a rare phenomenon which occurs most commonly when a fetus dies during an abdominal pregnancy, is too large to be reabsorbed by the body, and calcifies.
Foreign body reaction to biomaterial implantation
Following biomaterial implantation, blood and body fluids contact the implant surface. Host blood proteins adsorb onto the implant surface and a fibrin matrix forms. Acute and chronic inflammation follow the initial blood protein deposition and matrix formation. Macrophages at the implant site fuse to form foreign body giant cells. Following the inflammatory response, granulation tissue form. The end stage of the foreign body reaction is the fibrous capsule formation around the implanted biomaterial. The biocompatibility of the device affects the severity of the foreign body reaction. The foreign body reaction can lead to device failure.
Protein adsorption
During blood-biomaterial interaction, blood proteins spontaneously adsorb to the biomaterial surface. The biomaterial surface properties affect the types, concentrations, and conformation of proteins that adsorb to the surface. The Vroman effect can describe the time-dependent behavior of this protein adsorption. Surface-adsorbed proteins regulate inflammatory cell interaction and adhesion. The deposit |
https://en.wikipedia.org/wiki/Champalimaud%20Foundation | The Champalimaud Foundation () is a private biomedical research foundation. It was created according to the will of the late entrepreneur António de Sommer Champalimaud, in 2004. The complete name of the foundation honors the mother and father of the founder and is Fundação Anna de Sommer Champalimaud e Dr. Carlos Montez Champalimaud. It is located in Lisbon, Portugal.
Overview
The mission of the Foundation is "to develop programmes of advanced biomedical research and provide clinical care of excellence, with a focus on translating pioneering scientific discoveries into solutions which can improve the quality of life of individuals around the world."
The foundation undertakes research in the fields of neuroscience and oncology at the modernistic Champalimaud Centre for the Unknown in Lisbon, opened in 2011. Research into visual impairment is undertaken via an outreach program.
The Champalimaud Clinical Center (CCC) is a modern scientific, medical and technological institution providing specialized clinical treatment for oncology. The Center develops advanced programs for research of diseases. The CCC tries to customize the therapies in order to achieve more effectiveness in controlling and treating the diseases. It was designed by Indian architect Charles Correa.
Management
The management of the Foundation consists of Board of Directors, General Council, Scientific Committee, Ethics Committee and Vision Award Jury. The acting President is Leonor Beleza appointed by António Champalimaud in his will.
António Champalimaud Vision Award
The award was established in 2007 to recognise contributions to research into vision. In even numbered years it is awarded for contributions to overall vision research and in odd numbered years for contributions to the alleviation of visual problems, primarily in developing countries.
Recipients
Source: Champalimaud Foundation
2018: Jean Bennett, Albert Maguire, Robin Ali, James Bainbridge, Samuel Jacobson, William W. Hauswirth a |
https://en.wikipedia.org/wiki/K%C3%A1rm%C3%A1n%E2%80%93Howarth%20equation | In isotropic turbulence the Kármán–Howarth equation (after Theodore von Kármán and Leslie Howarth 1938), which is derived from the Navier–Stokes equations, is used to describe the evolution of non-dimensional longitudinal autocorrelation.
Mathematical description
Consider a two-point velocity correlation tensor for homogeneous turbulence
For isotropic turbulence, this correlation tensor can be expressed in terms of two scalar functions, using the invariant theory of full rotation group, first derived by Howard P. Robertson in 1940,
where is the root mean square turbulent velocity and are turbulent velocity in all three directions. Here, is the longitudinal correlation and is the lateral correlation of velocity at two different points. From continuity equation, we have
Thus uniquely determines the two-point correlation function. Theodore von Kármán and Leslie Howarth derived the evolution equation for from Navier–Stokes equation as
where uniquely determines the triple correlation tensor
Loitsianskii's invariant
L.G. Loitsianskii derived an integral invariant for the decay of the turbulence by taking the fourth moment of the Kármán–Howarth equation in 1939, i.e.,
If decays faster than as and also in this limit, if we assume that vanishes, we have the quantity,
which is invariant. Lev Landau and Evgeny Lifshitz showed that this invariant is equivalent to conservation of angular momentum. However, Ian Proudman and W.H. Reid showed that this invariant does not hold always since is not in general zero, at least, in the initial period of the decay. In 1967, Philip Saffman showed that this integral depends on the initial conditions and the integral can diverge under certain conditions.
Decay of turbulence
For the viscosity dominated flows, during the decay of turbulence, the Kármán–Howarth equation reduces to a heat equation once the triple correlation tensor is neglected, i.e.,
With suitable boundary conditions, the solution to above equation is g |
https://en.wikipedia.org/wiki/Human%20nose | The human nose is the most protruding part of the face. It bears the nostrils and is the first organ of the respiratory system. It is also the principal organ in the olfactory system. The shape of the nose is determined by the nasal bones and the nasal cartilages, including the nasal septum which separates the nostrils and divides the nasal cavity into two. On average, the nose of a male is larger than that of a female.
The nose has an important function in breathing. The nasal mucosa lining the nasal cavity and the paranasal sinuses carries out the necessary conditioning of inhaled air by warming and moistening it. Nasal conchae, shell-like bones in the walls of the cavities, play a major part in this process. Filtering of the air by nasal hair in the nostrils prevents large particles from entering the lungs. Sneezing is a reflex to expel unwanted particles from the nose that irritate the mucosal lining. Sneezing can transmit infections, because aerosols are created in which the droplets can harbour pathogens.
Another major function of the nose is olfaction, the sense of smell. The area of olfactory epithelium, in the upper nasal cavity, contains specialised olfactory cells responsible for this function.
The nose is also involved in the function of speech. Nasal vowels and nasal consonants are produced in the process of nasalisation. The hollow cavities of the paranasal sinuses act as sound chambers that modify and amplify speech and other vocal sounds.
There are several plastic surgery procedures that can be done on the nose, known as rhinoplasties available to correct various structural defects or to change the shape of the nose. Defects may be congenital, or result from nasal disorders or from trauma. These procedures are a type of reconstructive surgery. Elective procedures to change a nose shape are a type of cosmetic surgery.
Structure
Several bones and cartilages make up the bony-cartilaginous framework of the nose, and the internal structure. The no |
https://en.wikipedia.org/wiki/Nashville%20Number%20System | The Nashville Number System is a method of transcribing music by denoting the scale degree on which a chord is built. It was developed by Neal Matthews in the late 1950s as a simplified system for the Jordanaires to use in the studio and further developed by Charlie McCoy. It resembles the Roman numeral and figured bass systems traditionally used to transcribe a chord progression since the 1700s. The Nashville Number System was compiled and published in a book by Chas. Williams in 1988.
The Nashville Number System can be used by anyone, including someone with only a rudimentary background in music theory. Improvisation structures can be explained using numbers, and chord changes can be communicated mid-song by holding up the corresponding number of fingers. The system is flexible and can be embellished to include more information (such as chord color or to denote a bass note in an inverted chord). The system makes it easy for bandleaders, the record producer, or the lead vocalist to change the key of songs when recording in the studio or playing live since the new key has to be stated before the song is started. The rhythm section members can then use their knowledge of harmony to perform the song in a new key.
Scale degrees and major chords
The Nashville Number System (also referred to as NNS) is similar to (movable-do) Solfège, which uses "Do Ré Mi Fa Sol La Ti" to represent the seven scale degrees of the Major scale. It is also similar to roman numeral analysis; however, the NNS instead uses Arabic numerals to represent each of the scale degrees.
In the key of C, the numbers would correspond as follows:
C=1, D=2, E=3, F=4, G=5, A=6, B=7.
In the key of B, the numbers would be
B=1, C=2, D=3, E=4, F=5, G=6, A=7.
The key may be specified at the top of the written chord chart or given orally by the bandleader, record producer, or lead singer. The numbers do not change when transposing the composition into another key. They are relative to the new root note. |
https://en.wikipedia.org/wiki/Worm%20cast | A worm cast is a structure created by worms, typically on soils such as those on beaches that gives the appearance of multiple worms. They are also used to trace the location of one or more worms. |
https://en.wikipedia.org/wiki/Ectodomain | An ectodomain is the domain of a membrane protein that extends into the extracellular space (the space outside a cell). Ectodomains are usually the parts of proteins that initiate contact with surfaces, which leads to signal transduction. A notable example of an ectodomain is the S protein, commonly known as the spike protein, of the viral particle responsible for the COVID-19 pandemic. The ectodomain region of the spike protein (S) is essential for attachment and eventual entry of the viral protein into the host cell.
Ectodomains play a crucial part in the signaling pathways of viruses. Recent findings have indicated that certain antibodies including the anti-receptor binding domain (anti-RBD) or anti-spike ectodomain (anti-ECD) IgG titers can act as virus neutralization titers (VN titers) which can be identified in individuals with diseases, dyspnea and hospitalizations. In perspective of severe acute respiratory syndrome corona virus 2 (SARS-Cov-2) these specific ectodomains may detect antibody efficacy against SARS-Cov-2, in which VN titers can classify eligible plasma donors. Protective measures against diseases and respiratory conditions can further be advanced through ongoing research on ectodomains. Ectodomain's play a crucial part in the signaling pathways of viruses. In perspective of severe acute respiratory syndrome corona virus 2 (SARS-Cov-2) these specific ectodomains may detect antibody efficacy against SARS-Cov-2, in which VN titers can classify eligible plasma donors. Protective measures against diseases and respiratory conditions can further be advanced through ongoing research on ectodomains.
Ectodomains also interact with membrane systems inducing vesicle aggregation, lipid mixing and liposome leakage which provides information as to how certain viruses spread infection throughout the cellular domain. Specifically, the hepatitis C virus (HCV) utilize a fusion process in which the ectodomain of HCV E2 envelope protein confers fusogenic prope |
https://en.wikipedia.org/wiki/Windkessel%20effect | Windkessel effect is a term used in medicine to account for the shape of the arterial blood pressure waveform in terms of the interaction between the stroke volume and the compliance of the aorta and large elastic arteries (Windkessel vessels) and the resistance of the smaller arteries and arterioles. Windkessel when loosely translated from German to English means 'air chamber', but is generally taken to imply an elastic reservoir. The walls of large elastic arteries (e.g. aorta, common carotid, subclavian, and pulmonary arteries and their larger branches) contain elastic fibers, formed of elastin. These arteries distend when the blood pressure rises during systole and recoil when the blood pressure falls during diastole. Since the rate of blood entering these elastic arteries exceeds that leaving them via the peripheral resistance, there is a net storage of blood in the aorta and large arteries during systole, which discharges during diastole. The compliance (or distensibility) of the aorta and large elastic arteries is therefore analogous to a capacitor (employing the hydraulic analogy); to put it another way, these arteries collectively act as a hydraulic accumulator.
The Windkessel effect helps in damping the fluctuation in blood pressure (pulse pressure) over the cardiac cycle and assists in the maintenance of organ perfusion during diastole when cardiac ejection ceases. The idea of the Windkessel was alluded to by Giovanni Borelli, although Stephen Hales articulated the concept more clearly and drew the analogy with an air chamber used in fire engines in the 18th century. Otto Frank, an influential German physiologist, developed the concept and provided a firm mathematical foundation. Frank's model is sometimes called a two-element Windkessel to distinguish it from more recent and more elaborate Windkessel models (e.g. three- or four-element and non-linear Windkessel models).
Model types
Modeling of a Windkessel
Windkessel physiology remains a relevant y |
https://en.wikipedia.org/wiki/IESR | The JISC Information Environment Service Registry (IESR) provides a registry of all "significant research collections in the UK" and supports "teaching, learning and research". It is funded by the Jisc under its Shared Infrastructure Services programme.
The IESR aims to provide a “Yellow Pages for the academic internet” accessible through web and machine interfaces. It provides a reliable source of information that other applications, such as portals, can freely access through machine-to-machine protocols, in order to help their end users discover resources of assistance to them.
IESR contains information about the resources themselves, technical details about how to access the resources, and contact details for the resource providers.
IESR is a Mimas service based at the University of Manchester. |
https://en.wikipedia.org/wiki/Stochastically%20stable%20equilibrium | In game theory, a stochastically stable equilibrium is a refinement of the evolutionarily stable state in evolutionary game theory, proposed by Dean Foster and Peyton Young. An evolutionary stable state S is also stochastically stable if under vanishing noise, the probability that the population is in the vicinity of state S does not go to zero.
The concept is extensively used in models of learning in populations, where "noise" is used to model experimentation or replacement of unsuccessful players with new players (random mutation). Over time, as the need for experimentation dies down or the population becomes stable, the population will converge towards a subset of evolutionarily stable states. Foster and Young have shown that this subset is the set of states with the highest potential. |
https://en.wikipedia.org/wiki/Common%20bunt | Common bunt, also known as hill bunt, Indian bunt European bunt, stinking smut or covered smut, is a disease of both spring and winter wheats. It is caused by two very closely related fungi, Tilletia tritici (syn. Tilletia caries) and T. laevis (syn. T. foetida).
Symptoms
Plants with common bunt may be moderately stunted but infected plants cannot be easily recognized until near maturity and even then it is seldom conspicuous. After initial infection, the entire kernel is converted into a sorus consisting of a dark brown to black mass of teliospores covered by a modified periderm, which is thin and papery. The sorus is light to dark brown and is called a bunt ball. The bunt balls resemble wheat kernels but tend to be more spherical. The bunted heads are slender, bluish-green and may stay greener longer than healthy heads. The bunt balls change to a dull gray-brown at maturity, at which they become conspicuous. The fragile covering of the bunt balls are ruptured at harvest, producing clouds of spores. The spores have a fishy odor. Intact sori can also be found among harvested grain.
Disease cycle
Millions of spores are released at harvest and contaminate healthy kernels or land on other plant parts or the soil. The spores persist on the contaminated kernels or in the soil. The disease is initiated when soil-borne, or in particular seed-borne, teliospores germinate in response to moisture and produce hyphae that infect germinating seeds by penetrating the coleoptile before plants emerge. Cool soil temperatures (5° to 10 °C) favor infection. The intercellular hyphae become established in the apical meristem and are maintained systemically within the plant. After initial infection, hyphae are sparse in plants. The fungus proliferates in the spikes when ovaries begin to form. Sporulation occurs in endosperm tissue until the entire kernel is converted into a sorus consisting of a dark brown to black mass of teliospores covered by a modified periderm, which is |
https://en.wikipedia.org/wiki/GeForce%209%20series | The GeForce 9 series is the ninth generation of Nvidia's GeForce series of graphics processing units, the first of which was released on February 21, 2008. Products are based on a slightly repolished Tesla microarchitecture, adding PCIe 2.0 support, improved color and z-compression, and built on a 65 nm process, later using 55 nm process to reduce power consumption and die size (GeForce 8 G8x GPUs only supported PCIe 1.1 and were built on 90 nm process or 80 nm process).
GeForce 9100 Series
Geforce 9100 G
65 nm G98 GPU
PCI-E x16
64-bit bus
4 raster operations pipelines (ROP), 8 unified shaders
540 MHz core clock
256 MB DDR2, 400 MHz memory clock
1300 MHz shader clock
5.1 G texels/s fill rate
7.6 GB/s memory bandwidth
Supports DirectX 10, SM 4.0
OpenGL 2.1 compliance
Supports 1st generation PureVideo HD technology with partial VC1 decoding
GeForce 9300 Series
Geforce 9300 GS
On May 1, 2008, the GeForce 9300 GS was officially launched.
80 nm G86 GPU
PCI-E x16
64-bit bus
8 ROP, 16 unified shaders
450 MHz core clock
512 MB DDR2, 400 MHz memory clock
900 MHz shader clock
3.6 Gtexels/s fill rate
6.4 GB/s memory bandwidth
Supports DirectX 10, SM 4.0
OpenGL 2.1 compliance
GeForce 9400 Series
GeForce 9400 GT
On August 27, 2008, the GeForce 9400 GT was officially launched.
65 nm G96 GPU
16 stream processors
550 MHz core, with a 1350 MHz unified shader clock
4.4 Gtexels/s fillrate
256/512/1024 MB 800 MHz DDR2 or 256 MB 1600 MHz GDDR3, both with a 128-bit memory bus
12.8 GB/s memory bandwidth for boards configured with DDR2 800 MHz memory
Supports DirectX 10, Shader Model 4.0, OpenGL 3.3, and PCI-Express 2.0
Supports 2nd generation PureVideo HD technology with partial VC1 decoding and HybridPower technology.
Minimum of 300 watt power supply
GeForce 9500 Series
GeForce 9500 GT
On July 29, 2008, the GeForce 9500 GT was officially launched.
65 nm G96 GPU
32 stream processors (32 CUDA cores)
4 multi processors (each multi processor has 8 cores) |
https://en.wikipedia.org/wiki/Temperate%20deciduous%20forest | Temperate deciduous or temperate broad-leaf forests are a variety of temperate forest 'dominated' by deciduous trees that lose their leaves each winter. They represent one of Earth's major biomes, making up 9.69% of global land area. These forests are found in areas with distinct seasonal variation that ranges from warm, moist summers to cool winters. The six major areas of this forest type occur in the Northern Hemisphere: North America, East Asia, Central and Western Europe (except Brittany, Cornwall, Wales, Ireland and western Scotland), northern Spain, Denmark, southern Sweden, southern Norway and in the southern hemisphere in Patagonia (Chile and Argentina). Examples of typical trees in the Northern Hemisphere's deciduous forests include oak, maple, basswood, beech and elm, while in the Southern Hemisphere, trees of the genus Nothofagus dominate this type of forest.
Resources
Humans have often colonised areas in the temperate deciduous forest. They have harvested wood for timber and charcoal. During the European colonization of North America, potash made from tree ashes was exported back to Europe as fertilizer. This left less than one-quarter of original forests remaining. Many forests are now small fragments dissected by fields and roads; these islands of green often differ substantially from the original forests, particularly along the edges.
The introduction of exotic diseases continues to be a threat to forest trees and, hence, the forest; examples include the loss of chestnut and elm. At the same time, species such as deer, which are clearing rather than true forest animals, have expanded their range and proliferated in these altered landscapes.
Large deer populations have deleterious effects on tree regeneration overall, but particularly for edible species including yew, yellow birch, and hemlock. Deer grazing also has significant negative effects on the number and kind of herbaceous flowering plants. The continuing pressure to increase deer popu |
https://en.wikipedia.org/wiki/Barnes%E2%80%93Hut%20simulation | The Barnes–Hut simulation (named after Josh Barnes and Piet Hut) is an approximation algorithm for performing an n-body simulation. It is notable for having order O(n log n) compared to a direct-sum algorithm which would be O(n2).
The simulation volume is usually divided up into cubic cells via an octree (in a three-dimensional space), so that only particles from nearby cells need to be treated individually, and particles in distant cells can be treated as a single large particle centered at the cell's center of mass (or as a low-order multipole expansion). This can dramatically reduce the number of particle pair interactions that must be computed.
Some of the most demanding high-performance computing projects do computational astrophysics using the Barnes–Hut treecode algorithm,
such as DEGIMA.
Algorithm
The Barnes–Hut tree
In a three-dimensional n-body simulation, the Barnes–Hut algorithm recursively divides the n bodies into groups by storing them in an octree (or a quad-tree in a 2D simulation). Each node in this tree represents a region of the three-dimensional space.
The topmost node represents the whole space, and its eight children represent the eight octants of the space. The space is recursively subdivided into octants until each subdivision contains 0 or 1 bodies (some regions do not have bodies in all of their octants).
There are two types of nodes in the octree: internal and external nodes. An external node has no children and is either empty or represents a single body. Each internal node represents the group of bodies beneath it, and stores the center of mass and the total mass of all its children bodies.
Calculating the force acting on a body
To calculate the net force on a particular body, the nodes of the tree are traversed, starting from the root. If the center of mass of an internal node is sufficiently far from the body, the bodies contained in that part of the tree are treated as a single particle whose position and mass is respectivel |
https://en.wikipedia.org/wiki/Grothendieck%20inequality | In mathematics, the Grothendieck inequality states that there is a universal constant with the following property. If Mij is an n × n (real or complex) matrix with
for all (real or complex) numbers si, tj of absolute value at most 1, then
for all vectors Si, Tj in the unit ball B(H) of a (real or complex) Hilbert space H, the constant being independent of n. For a fixed Hilbert space of dimension d, the smallest constant that satisfies this property for all n × n matrices is called a Grothendieck constant and denoted . In fact, there are two Grothendieck constants, and , depending on whether one works with real or complex numbers, respectively.
The Grothendieck inequality and Grothendieck constants are named after Alexander Grothendieck, who proved the existence of the constants in a paper published in 1953.
Motivation and the operator formulation
Let be an matrix. Then defines a linear operator between the normed spaces and for . The -norm of is the quantity
If , we denote the norm by .
One can consider the following question: For what value of and is maximized? Since is linear, then it suffices to consider such that contains as many points as possible, and also such that is as large as possible. By comparing for , one sees that for all .
One way to compute is by solving the following quadratic integer program:
To see this, note that , and taking the maximum over gives . Then taking the maximum over gives by the convexity of and by the triangle inequality. This quadratic integer program can be relaxed to the following semidefinite program:
It is known that exactly computing for is NP-hard, while exacting computing is NP-hard for .
One can then ask the following natural question: How well does an optimal solution to the semidefinite program approximate ? The Grothendieck inequality provides an answer to this question: There exists a fixed constant such that, for any , for any matrix , and for any Hilbert space ,
Bounds on |
https://en.wikipedia.org/wiki/Stefan%20Bergman | Stefan Bergman (5 May 1895 – 6 June 1977) was a Congress Poland-born American mathematician whose primary work was in complex analysis. His name is also written Bergmann; he dropped the second "n" when he came to the U. S. He is best known for the kernel function he discovered while at University of Berlin in 1922. This function is known today as the Bergman kernel. Bergman taught for many years at Stanford University, and served as an advisor to several students.
Biography
Born in Częstochowa, Congress Poland, Russian Empire, to a German Jewish family, Bergman received his Ph.D. at University of Berlin in 1921 for a dissertation on Fourier analysis. His advisor, Richard von Mises, had a strong influence on him, lasting for the rest of his career. In 1933, Bergman was forced to leave his post at the Berlin University because he was a Jew. He fled first to Russia, where he stayed until 1939, and then to Paris. In 1939, he emigrated to the United States, where he would remain for the rest of life. He was elected a Fellow of the American Academy of Arts and Sciences in 1951. He was a professor at Stanford University from 1952 until his retirement in 1972. He was an invited speaker at the International Congress of Mathematicians in 1950 in Cambridge, Massachusetts and in 1962 in Stockholm (On meromorphic functions of several complex variables). He died in Palo Alto, California, aged 82.
The Bergman Prize
The Stefan Bergman Prize in mathematics was initiated by Bergman's wife in her will, in memory of her husband's work. The American Mathematical Society supports the prize and selects the committee of judges. The prize is awarded for:
the theory of the kernel function and its applications in real and complex analysis; or
function-theoretic methods in the theory of partial differential equations of elliptic type with a special attention to Bergman's and related operator methods.
Selected publications
.
.
.
. The original edition was published in 1941 by Inters |
https://en.wikipedia.org/wiki/Coded%20aperture | Coded apertures or coded-aperture masks are grids, gratings, or other patterns of materials opaque to various wavelengths of electromagnetic radiation. The wavelengths are usually high-energy radiation such as X-rays and gamma rays. A coded "shadow" is cast upon a plane by blocking radiation in a known pattern. The properties of the original radiation sources can then be mathematically reconstructed from this shadow. Coded apertures are used in X- and gamma ray imaging systems, because these high-energy rays cannot be focused with lenses or mirrors that work for visible light.
Rationale
Imaging is usually done at optical wavelengths using lenses and mirrors. However, the energy of hard X-rays and γ-rays is too high to be reflected or refracted, and simply passes through the lenses and mirrors of optical telescopes. Image modulation by apertures is, therefore, often used instead. The pinhole camera is the most basic form of such a modulation imager, but its disadvantage is low throughput, as its small aperture allows through little radiation. Only a tiny fraction of the light passes through the pinhole, which causes a low signal-to-noise ratio. To solve this problem, the mask can contain many holes, in one of several particular patterns, for example. Multiple masks, at varying distances from a detector, add flexibility to this tool. Specifically the modulation collimator, invented by Minoru Oda, was used to identify the first cosmic X-ray source and thereby to launch the new field of X-ray astronomy in 1965. Many other applications in other fields, such as tomography, have since appeared.
In a coded aperture more complicated than a pinhole camera, images from multiple apertures will overlap at the detector array. It is thus necessary to use a computational algorithm (which depends on the precise configuration of the aperture arrays) to reconstruct the original image. In this way a sharp image can be achieved without a lens. The image is formed from the whole array |
https://en.wikipedia.org/wiki/A%20Declaration%20of%20the%20Independence%20of%20Cyberspace | "A Declaration of the Independence of Cyberspace" is a widely distributed early paper on the applicability (or lack thereof) of government to the rapidly growing Internet. Commissioned for the pioneering Internet project 24 Hours in Cyberspace, it was written by John Perry Barlow, a founder of the Electronic Frontier Foundation, and published online on February 8, 1996, from Davos, Switzerland. It was written primarily in response to the passing into law of the Telecommunications Act of 1996 in the United States. In 2014, the Department of Records recorded and released audio and video content of Barlow reading the Declaration.
Content
The declaration sets out, in sixteen short paragraphs, a rebuttal to government of the Internet by any outside force, specifically the United States. It states that the United States did not have the consent of the governed to apply laws to the Internet, and that the Internet was outside any country's borders. Instead, the Internet was developing its own social contracts to determine how to handle its problems, based on the golden rule. It does this in language evocative of the United States Declaration of Independence and obliquely cites it in its final paragraphs. Although the paper mentions the Telecommunications Act, it also accuses China, Germany, France, Russia, Singapore, and Italy of stifling the Internet.
Background
At the time the paper was written, Barlow had already written extensively on the Internet and its social and legal phenomena, as well as being a founding member of the Electronic Frontier Foundation. The work he was known best for previously, "The Economy of Ideas", published March 1994 in Wired magazine, also made allusions to Thomas Jefferson and some of the ideas he would write about in his declaration.
Critical response
Because of its subject matter, Barlow's work quickly became famous and widely distributed on the Internet. Within three months, an estimated 5,000 websites had copies of the declaration. At |
https://en.wikipedia.org/wiki/Regina%20Tyshkevich | Regina Iosifovna Tyshkevich (; 20 October 1929 – 17 November 2019) was a Belarusian mathematician, an expert in graph theory, Doctor of Physical and Mathematical Sciences, professor of the Belarusian State University.
Her main scientific interests included Intersection graphs, degree sequences, and the reconstruction conjecture. She was also known for an independent introduction and investigation of the class of split graphs and for her contributions to line graphs of hypergraphs.
In 1998, she was awarded the Belarus State Prize for her book Lectures in Graph Theory. Of note is her textbook An Introduction into Mathematics, written together with her two colleagues.
In October 2009 an international conference "Discrete Mathematics, Algebra, and their Applications", sponsored by the Central European Initiative, was held in Minsk, Belarus in honor of her 80th anniversary.
Regina Tyshkevich was a direct descendant of the Tyszkiewicz
magnate family, therefore her colleagues sometimes called her "the countess of graph theory", which is a pun in the Russian language: the Russian word "граф" (graf) is a homonym for two words meaning "count" and "graph".
Books and selected publications
(With ) "Commutative Matrices", 1968, Academic Press
Russian original: "Perestanovochnye matritsy" 1966, 2nd edition: 2003,
(With Emilichev, V. A., Melnikov, O. I., Sarvanov, V. I.) "Lectures on Graph Theory", B. I. Wissenschaftsverlag, 1994
Russian original: "Lektsii po teorii grafov", 1990
(With O. Melnikov and V. Sarvanov, etc.) "Exercises in Graph Theory", Kluwer Academic Publishers, 1998,
"Linear Algebra and Analytical Geometry (Линейная алгебра и аналитическая геометрия)
Кононов С.Г., Тышкевич Р.И., Янчевский В.И. "Введение в математику" ("An Introduction into Mathematics") 3 volumes, Minsk, Belarusian State University, 2003
R.I. Tyshkevich. Decomposition of graphical sequences and unigraphs // Discrete Math., 2000, Vol. 220, p. 201 - 238.
Yury Metelsky, Regina Tyshkevich |
https://en.wikipedia.org/wiki/Objective-collapse%20theory | Objective-collapse theories, also known as models of spontaneous wave function collapse or dynamical reduction models, are proposed solutions to the measurement problem in quantum mechanics. As with other theories called interpretations of quantum mechanics, they are possible explanations of why and how quantum measurements always give definite outcomes, not a superposition of them as predicted by the Schrödinger equation, and more generally how the classical world emerges from quantum theory. The fundamental idea is that the unitary evolution of the wave function describing the state of a quantum system is approximate. It works well for microscopic systems, but progressively loses its validity when the mass / complexity of the system increases.
In collapse theories, the Schrödinger equation is supplemented with additional nonlinear and stochastic terms (spontaneous collapses) which localize the wave function in space. The resulting dynamics is such that for microscopic isolated systems, the new terms have a negligible effect; therefore, the usual quantum properties are recovered, apart from very tiny deviations. Such deviations can potentially be detected in dedicated experiments, and efforts are increasing worldwide towards testing them.
An inbuilt amplification mechanism makes sure that for macroscopic systems consisting of many particles, the collapse becomes stronger than the quantum dynamics. Then their wave function is always well-localized in space, so well-localized that it behaves, for all practical purposes, like a point moving in space according to Newton's laws.
In this sense, collapse models provide a unified description of microscopic and macroscopic systems, avoiding the conceptual problems associated to measurements in quantum theory.
The most well-known examples of such theories are:
Ghirardi–Rimini–Weber (GRW) model
Continuous spontaneous localization (CSL) model
Diósi–Penrose (DP) model
Collapse theories stand in opposition to many-world |
https://en.wikipedia.org/wiki/Gauss%E2%80%93Lucas%20theorem | In complex analysis, a branch of mathematics, the Gauss–Lucas theorem gives a geometric relation between the roots of a polynomial and the roots of its derivative . The set of roots of a real or complex polynomial is a set of points in the complex plane. The theorem states that the roots of all lie within the convex hull of the roots of , that is the smallest convex polygon containing the roots of . When has a single root then this convex hull is a single point and when the roots lie on a line then the convex hull is a segment of this line. The Gauss–Lucas theorem, named after Carl Friedrich Gauss and Félix Lucas, is similar in spirit to Rolle's theorem.
Formal statement
If is a (nonconstant) polynomial with complex coefficients, all zeros of belong to the convex hull of the set of zeros of .
Special cases
It is easy to see that if is a second degree polynomial, the zero of is the average of the roots of . In that case, the convex hull is the line segment with the two roots as endpoints and it is clear that the average of the roots is the middle point of the segment.
For a third degree complex polynomial (cubic function) with three distinct zeros, Marden's theorem states that the zeros of are the foci of the Steiner inellipse which is the unique ellipse tangent to the midpoints of the triangle formed by the zeros of .
For a fourth degree complex polynomial (quartic function) with four distinct zeros forming a concave quadrilateral, one of the zeros of lies within the convex hull of the other three; all three zeros of lie in two of the three triangles formed by the interior zero of and two others zeros of .
In addition, if a polynomial of degree of real coefficients has distinct real zeros we see, using Rolle's theorem, that the zeros of the derivative polynomial are in the interval which is the convex hull of the set of roots.
The convex hull of the roots of the polynomial
particularly includes the point
Proof
See also
Marden's theorem
|
https://en.wikipedia.org/wiki/2007%20Bernard%20Matthews%20H5N1%20outbreak | The 2007 Bernard Matthews H5N1 outbreak was an occurrence of avian influenza in England caused by the H5N1 subtype of Influenza virus A that began on 30 January 2007. The infection affected poultry at one of Bernard Matthews' farms in Holton in Suffolk. It was the third instance of H5N1-subtype detected in the United Kingdom and a range of precautions were instituted to prevent spread of the disease including a large cull of turkeys, the imposition of segregation zones, and a disinfection programme for the plant.
The cause of the outbreak was not determined. However, it was considered significant that Bernard Matthews regularly transports turkeys and turkey products between the UK and its plant in Hungary, and that the H5N1 strains previously found in Hungary, and those found at Suffolk, were effectively genetically identical.
Background
H5N1
H5N1 is a subtype of the Influenza A virus, the viruses responsible for influenza in humans and many other animal species. A bird-adapted strain of H5N1, called HPAI A(H5N1) for "highly pathogenic avian influenza virus of type A of subtype H5N1", is the causative agent of H5N1 flu.
HPAI A(H5N1) is considered an avian disease, although there is some evidence of limited human-to-human transmission of the virus. A risk factor for contracting the virus is handling of infected poultry, but transmission of the virus from infected birds to humans is inefficient. Poultry farming practices have changed due to H5N1. The cost of poultry farming has increased, while the cost to consumers has gone down, due to fears from H5N1 driving demand below supply.
Recent outbreaks in the UK and rest of EU
The outbreak was the third instance of H5N1 detected in the United Kingdom. The first outbreak occurred in October 2005 among exotic birds imported from Taiwan and South America at a privately owned quarantine facility in Essex, England. The second instance involved a dead whooper swan found to have the virus in Cellardyke, Scotland in April 2 |
https://en.wikipedia.org/wiki/List%20of%20nuclear%20holocaust%20fiction | This list of nuclear holocaust fiction lists the many works of speculative fiction that attempt to describe a world during or after a massive nuclear war, nuclear holocaust, or crash of civilization due to a nuclear electromagnetic pulse.
Films
Television programs
A Carol for Another Christmas (ABC, 1964), Rod Serling TV film
A Day Called 'X' (CBS, 1957)
Adventure Time (Cartoon Network, 2010–2018)
American Horror Story: Apocalypse (FX, 2018)
Battlestar Galactica (Sci-Fi, 2003, 2004–2009)
Buck Rogers in the 25th Century (NBC, 1979)
By Dawn's Early Light (HBO, 1990)
Countdown to Looking Glass (HBO, 1984)
Dark Angel (Fox, 2000–2002)
Der Dritte Weltkrieg (ZDF, 1998)
Fail Safe (CBS, 2000)
Genesis II (CBS, 1973)
Jericho (CBS, 2006–2008)
Level Seven (BBC, 1966), adapted by J. B. Priestley for Out of the Unknown
On the Beach (Showtime, 2000)
Planet Earth (ABC, 1974)
Special Bulletin (NBC, 1983)
Terminator: The Sarah Connor Chronicles (Fox, 2008–2009)
Testament (PBS, 1983)
The 100 (The CW, 2014–2020)
The Day After (ABC, 1983)
The Martian Chronicles miniseries (NBC, 1980)
The War Game (BBC, 1965)
Threads (BBC, 1984)
Whoops Apocalypse (ITV, 1982)
Woops! (Fox, 1992)
World War III miniseries (NBC, 1982)
Television episodes
The Motorola Television Hour: "Atomic Attack" (1954 ABC-TV series Season 1, Episode 15) – A family living 50 miles away try to flee from the fallout of an hydrogen bomb that fell on New York City. Based on the novel Shadow on the Hearth (1950) by Judith Merrill.
The Twilight Zone: "Time Enough at Last" (1959)
Playhouse 90: "Alas, Babylon" (1960)
The Twilight Zone: "The Old Man in the Cave" (1963)
Star Trek: "Space Seed" (1967)
Star Trek: "Assignment: Earth" (1968) – The crew goes back in time to find out how the human race was able to survive the Cold War.
The Twilight Zone: "A Little Peace and Quiet" (1985)
The Twilight Zone: "Quarantine" (1986)
The Twilight Zone: "Shelter Skelter" (1987)
The Outer Limits: "Bits of Lov |
https://en.wikipedia.org/wiki/Vector%20measure | In mathematics, a vector measure is a function defined on a family of sets and taking vector values satisfying certain properties. It is a generalization of the concept of finite measure, which takes nonnegative real values only.
Definitions and first consequences
Given a field of sets and a Banach space a finitely additive vector measure (or measure, for short) is a function such that for any two disjoint sets and in one has
A vector measure is called countably additive if for any sequence of disjoint sets in such that their union is in it holds that
with the series on the right-hand side convergent in the norm of the Banach space
It can be proved that an additive vector measure is countably additive if and only if for any sequence as above one has
where is the norm on
Countably additive vector measures defined on sigma-algebras are more general than finite measures, finite signed measures, and complex measures, which are countably additive functions taking values respectively on the real interval the set of real numbers, and the set of complex numbers.
Examples
Consider the field of sets made up of the interval together with the family of all Lebesgue measurable sets contained in this interval. For any such set define
where is the indicator function of Depending on where is declared to take values, two different outcomes are observed.
viewed as a function from to the -space is a vector measure which is not countably-additive.
viewed as a function from to the -space is a countably-additive vector measure.
Both of these statements follow quite easily from the criterion () stated above.
The variation of a vector measure
Given a vector measure the variation of is defined as
where the supremum is taken over all the partitions
of into a finite number of disjoint sets, for all in Here, is the norm on
The variation of is a finitely additive function taking values in It holds that
for any in If is finite, the |
https://en.wikipedia.org/wiki/Culinology | Culinology, according to Jeff Cousminer in Food Product Design Magazine, is a term that was coined by the first president and founder of the Research Chefs Association, Winston Riley. The original meaning of the word was quite different from what it has come to mean today. Originally, the word was designed to be a combination of two words: "culinary" and "technology". So the first meaning of the word was the convergence of culinary arts and all technology, which includes communications, chemistry, physiology, economics and many others.
There are accredited culinology educational programs offered by many institutions. The curriculums of such courses combine the disciplines of cooking and food science. According to industry professionals, such as Harry Crane, culinology should "help jump-start product development."
Culinologists work in diverse aspects of food—from experimental chefs and menu planners to food manufacturing to fine dining. The word is protected by the professional association, the Research Chefs Association, which owns the registered trademark.
See also
Molecular gastronomy |
https://en.wikipedia.org/wiki/Weighted%20matroid | In combinatorics, a branch of mathematics, a weighted matroid is a matroid endowed with function with respect to which one can perform a greedy algorithm.
A weight function for a matroid assigns a strictly positive weight to each element of . We extend the function to subsets of by summation; is the sum of over in . A matroid with an associated weight function is called a weighted matroid.
Spanning forest algorithms
As a simple example, say we wish to find the maximum spanning forest of a graph. That is, given a graph and a weight for each edge, find a forest containing every vertex and maximizing the total weight of the edges in the tree. This problem arises in some clustering applications. If we look at the definition of the forest matroid above, we see that the maximum spanning forest is simply the independent set with largest total weight — such a set must span the graph, for otherwise we can add edges without creating cycles. But how do we find it?
Finding a basis
There is a simple algorithm for finding a basis:
Initially let be the empty set.
For each in
if is independent, then set to .
The result is clearly an independent set. It is a maximal independent set because if is not independent for some subset of , then is not independent either (the contrapositive follows from the hereditary property). Thus if we pass up an element, we'll never have an opportunity to use it later. We will generalize this algorithm to solve a harder problem.
Extension to optimal
An independent set of largest total weight is called an optimal set. Optimal sets are always bases, because if an edge can be added, it should be; this only increases the total weight. As it turns out, there is a trivial greedy algorithm for computing an optimal set of a weighted matroid. It works as follows:
Initially let be the empty set.
For each in , taken in (monotonically) decreasing order by weight
if is independent, then set to .
This algorithm finds a basis, since it |
https://en.wikipedia.org/wiki/Pinwheel%20tiling | In geometry, pinwheel tilings are non-periodic tilings defined by Charles Radin and based on a construction due to John Conway.
They are the first known non-periodic tilings to each have the property that their tiles appear in infinitely many orientations.
Conway's tessellation
Let be the right triangle with side length , and .
Conway noticed that can be divided in five isometric copies of its image by the dilation of factor .
By suitably rescaling and translating/rotating, this operation can be iterated to obtain an infinite increasing sequence of growing triangles all made of isometric copies of .
The union of all these triangles yields a tiling of the whole plane by isometric copies of .
In this tiling, isometric copies of appear in infinitely many orientations (this is due to the angles and of each being algebraically independent to over the reals.).
Despite this, all the vertices have rational coordinates.
The pinwheel tilings
Radin relied on the above construction of Conway to define pinwheel tilings.
Formally, the pinwheel tilings are the tilings whose tiles are isometric copies of , in which a tile may intersect another tile only either on a whole side or on half the length side, and such that the following property holds.
Given any pinwheel tiling , there is a pinwheel tiling which, once each tile is divided in five following the Conway construction and the result is dilated by a factor , is equal to .
In other words, the tiles of any pinwheel tilings can be grouped in sets of five into homothetic tiles, so that these homothetic tiles form (up to rescaling) a new pinwheel tiling.
The tiling constructed by Conway is a pinwheel tiling, but there are uncountably many other different pinwheel tilings.
They are all locally undistinguishable (i.e., they have the same finite patches).
They all share with the Conway tiling the property that tiles appear in infinitely many orientations (and vertices have rational coordinates).
The main result prove |
https://en.wikipedia.org/wiki/Ali%20Moustafa%20Mosharafa | Dr. Ali Moustafa Mosharafa () (11 July 1898 – 16 January 1950) was an Egyptian theoretical physicist. He was professor of applied mathematics in the Faculty of Science at Cairo University, and also served as its first dean. He contributed to the development of quantum theory as well as the theory of relativity.
Biography
Birth and early life
Mosharafa obtained his primary certificate in 1910 ranking second nationwide. He obtained his Baccalaureate at the age of 16, becoming the youngest student at that time to be awarded such a certificate, and again ranking second. He preferred to enroll in the Teachers' College rather than the faculties of Medicine or Engineering due to his deep interest in mathematics.
He graduated in 1917. Due to his excellence in mathematics, the Egyptian Ministry of Education sent him to England where he obtained a BSc (Honors) from the University of Nottingham in 1920. The Egyptian University consented to grant Mosharafa another scholarship to complete his doctoral thesis. During his stay in London, he was published many times in prominent science magazines. He obtained a PhD in 1923 from King's College London in the shortest possible time permissible according to the regulations there. In 1924 Mosharafa was awarded the degree of Doctor of Science, the first Egyptian and 11th scientist in the entire world to obtain such a degree.
Academic career
He became a teacher in the Higher Teachers' college in Cairo University, he became an associate professor of mathematics in the Faculty of Science because he was under the age of 30, the minimum age required for fulfilling the post of a professor. In 1926 his promotion to professor was raised in the Parliament, then chaired by Saad Zaghloul. The Parliament lauded his qualifications and merits which surpassed those of the English dean of the faculty and he was promoted to professor.
He was the first Egyptian professor of applied mathematics in the Faculty of Science. He became dean of the facu |
https://en.wikipedia.org/wiki/List%20of%20thermal%20conductivities | In heat transfer, the thermal conductivity of a substance, k, is an intensive property that indicates its ability to conduct heat. For most materials, the amount of heat conducted varies (usually non-linearly) with temperature.
Thermal conductivity is often measured with laser flash analysis. Alternative measurements are also established.
Mixtures may have variable thermal conductivities due to composition. Note that for gases in usual conditions, heat transfer by advection (caused by convection or turbulence for instance) is the dominant mechanism compared to conduction.
This table shows thermal conductivity in SI units of watts per metre-kelvin (W·m−1·K−1). Some measurements use the imperial unit BTUs per foot per hour per degree Fahrenheit ( =
Sortable list
This concerns materials at atmospheric pressure and around .
Analytical list
Thermal conductivities have been measured with longitudinal heat flow methods where the experimental arrangement is so designed to accommodate heat flow in only the axial direction, temperatures are constant, and radial heat loss is prevented or minimized. For the sake of simplicity the conductivities that are found by that method in all of its variations are noted as L conductivities, those that are found by radial measurements of the sort are noted as R conductivities, and those that are found from periodic or transient heat flow are distinguished as P conductivities. Numerous variations of all of the above and various other methods have been discussed by some G. K. White, M. J. Laubits, D. R. Flynn, B. O. Peirce and R. W. Wilson and various other theorists who are noted in an international Data Series from Purdue University, Volume I pages 14a–38a.
This concerns materials at various temperatures and pressures.
See also
Laser flash analysis
List of insulation materials
R-value (insulation)
Thermal transmittance
Specific heat capacity
Thermal conductivity
Thermal conductivities of the elements (data page)
Thermal |
https://en.wikipedia.org/wiki/Tricorn%20%28mathematics%29 | In mathematics, the tricorn, sometimes called the Mandelbar set, is a fractal defined in a similar way to the Mandelbrot set, but using the mapping instead of used for the Mandelbrot set. It was introduced by W. D. Crowe, R. Hasson, P. J. Rippon, and P. E. D. Strain-Clark. John Milnor found tricorn-like sets as a prototypical configuration in the parameter space of real cubic polynomials, and in various other families of rational maps.
The characteristic three-cornered shape created by this fractal repeats with variations at different scales, showing the same sort of self-similarity as the Mandelbrot set. In addition to smaller tricorns, smaller versions of the Mandelbrot set are also contained within the tricorn fractal.
Formal definition
The tricorn is defined by a family of quadratic antiholomorphic polynomials
given by
where is a complex parameter. For each , one looks at the forward orbit
of the critical point of the antiholomorphic polynomial . In analogy with the Mandelbrot set, the tricorn is defined as the set of all parameters for which the forward orbit of the critical point is bounded. This is equivalent to saying that the tricorn is the connectedness locus of the family of quadratic antiholomorphic polynomials; i.e. the set of all parameters for which the Julia set is connected.
The higher degree analogues of the tricorn are known as the multicorns. These are the connectedness loci of the family of antiholomorphic polynomials .
Basic properties
The tricorn is compact, and connected. In fact, Nakane modified Douady and Hubbard's proof of the connectedness of the Mandelbrot set to construct a dynamically defined real-analytic diffeomorphism from the exterior of the tricorn onto the exterior of the closed unit disc in the complex plane. One can define external parameter rays of the tricorn as the inverse images of radial lines under this diffeomorphism.
Every hyperbolic component of the tricorn is simply connected.
The boundary of every |
https://en.wikipedia.org/wiki/Self-testing%20code | Self-testing code is software that incorporates built-in tests (see test-first development).
In Java, to execute a unit test from the command line, a class can have methods like the following.
// Executing <code>main</code> runs the unit test.
public static void main(String[] args) {
test();
}
static void test() {
assert foo == bar;
}
To invoke a full system test, a class can incorporate a method call.
public static void main(String[] args) {
test();
TestSuite.test(); // invokes full system test
}In addition, Java has some Jupiter API libraries for self-testing code. assert can be used in various ways such as assert equals, which checks if the given variable is equal to the value given.@Test
void checkplayer() {
Board board = new Board(10);
board.addplayer(1);
int check = board.getCurrentPlayer(1);
assertEquals(1, check);
}
See also
Software development
Extreme programming |
https://en.wikipedia.org/wiki/The%20Lady%20Vanishes | The Lady Vanishes is a 1938 British mystery thriller film directed by Alfred Hitchcock, starring Margaret Lockwood and Michael Redgrave. Written by Sidney Gilliat and Frank Launder, based on the 1936 novel The Wheel Spins by Ethel Lina White, the film is about a beautiful English tourist travelling by train in continental Europe who discovers that her elderly travelling companion seems to have disappeared from the train. After her fellow passengers deny ever having seen the elderly lady, the young woman is helped by a young musicologist, the two proceeding to search the train for clues to the old lady's disappearance.
The Lady Vanishes was filmed at the Gainsborough Studios in Islington, London. Hitchcock caught Hollywood's attention with the film and moved to Hollywood soon after its release. Although the director's three previous efforts had done poorly at the box office, The Lady Vanishes was widely successful, and confirmed American producer David O. Selznick's belief that Hitchcock indeed had a future in Hollywood cinema.
The British Film Institute ranked The Lady Vanishes the 35th best British film of the 20th century. In 2017, a poll of 150 actors, directors, writers, producers and critics for Time Out magazine saw it ranked the 31st best British film ever. It is one of Hitchcock's most renowned British films, and the first of three screen versions of White's novel to date.
Plot
After visiting the fictional country of Bandrika, English tourist Iris Henderson is returning home to get married, but an avalanche blocks the railway line. The stranded passengers are forced to spend the night at a hotel. In the same predicament are Charters and Caldicott, cricket enthusiasts anxious to see the last days of a Test match in Manchester, and Miss Froy, a governess and music teacher. Miss Froy listens to a folk singer in the street, but he is strangled to death by an unseen murderer.
That evening, Iris is bothered by a loud noise from the room above hers. It is cause |
https://en.wikipedia.org/wiki/Laryngeal%20ventricle | The laryngeal ventricle, (also called the ventricle of the larynx, laryngeal sinus, or Morgagni's sinus) is a fusiform fossa, situated between the vestibular and vocal folds on either side, and extending nearly their entire length. There is also a sinus of Morgagni in the pharynx.
The fossa is bounded, above, by the free crescentic edge of the vestibular ligament; below, by the straight margin of the vocal fold and laterally, by the mucous membrane covering the corresponding thyroarytenoid muscle.
The anterior part of the ventricle leads up by a narrow opening into a pouch-like diverticulum, a mucous membranous sac of variable size called the appendix of the laryngeal ventricle. The appendix (also called the laryngeal saccule, pouch or Hilton's pouch) extends vertically from the laryngeal ventricle. It runs between the vestibular fold, thyroarytenoid muscle, and thyroid cartilage, and is conical, bending slightly backward. It is covered in roughly seventy mucous glands. The muscles surrounding the appendix compress it until mucus is secreted to lubricate the vocal folds.
Additional images |
https://en.wikipedia.org/wiki/Symbols%20of%20Alberta | Alberta is one of Canada's provinces, and has established several official emblems that reflect the province's history, its natural and diverse landscapes, and its people.
Official symbols of Alberta
De facto symbols
While not officially adopted through legislation as emblems by the government of Alberta, these places and things are popularly associated with (hence could be considered symbols of) the province.
See also
List of Canadian provincial and territorial symbols
Canadian royal symbols |
https://en.wikipedia.org/wiki/Ficus%20pumila | Ficus pumila, commonly known as the creeping fig or climbing fig, is a species of flowering plant in the mulberry family, native to East Asia (China, Japan, Vietnam) and naturalized in parts of the southeastern and south-central United States. It is also found in cultivation as a houseplant. The Latin specific epithet pumila means "dwarf", and refers to the very small leaves of the plant.
Description
Ficus pumila is a woody evergreen liana, growing to . It can grow up to tall if it isn't regularly pruned. The juvenile foliage is much smaller and thinner than mature leaves produced as the plant ages. The leaves are oval, cordate, asymmetrical, with opposite veins. It is creeping or can behave like a liana and also climb trees, rocks, etc. up to 4 m in height or more. The aerial roots secrete a translucent latex that hardens on drying, allowing the sticks to adhere to their support.
Cultivation
As the common name, "creeping fig" indicates, the plant has a creeping/vining habit and is often used in gardens and landscapes where it covers the ground and climbs up trees and walls. It is hardy down to and does not tolerate frost. Therefore in temperate regions it is often seen as a houseplant. It is fast-growing and requires little in the way of care. It can be invasive when environmental conditions are favorable. Its secondary roots or tendrils can cause structural damage to certain buildings with fragile mortar or structures made of fragile materials.
It has gained the Royal Horticultural Society's Award of Garden Merit.
The plant requires the fig wasp Blastophaga pumilae for pollination, and is fed upon by larvae of the butterfly Marpesia petreus.
Varieties and cultivars
Ficus pumila var. awkeotsang — awkeotsang creeping fig
Ficus pumila var. quercifolia — oak leaf creeping fig
Ficus pumila 'Curly' — curly creeping fig; crinkled leaf form
Ficus pumila 'Variegata' and Ficus pumila 'Snowflake' — variegated creeping fig; variegated foliage
Cuisine
The fruit of |
https://en.wikipedia.org/wiki/Symbols%20of%20British%20Columbia | British Columbia is Canada's westernmost province, and has established several provincial symbols.
Official symbols
Other Symbols |
https://en.wikipedia.org/wiki/Symbols%20of%20Manitoba | There are several symbols of Manitoba, one of the ten provinces of Canada. These symbols are designated by The Coat of Arms, Emblems and the Manitoba Tartan Act, which came into force on Feb 1, 1988.
Symbols |
https://en.wikipedia.org/wiki/Symbols%20of%20Newfoundland%20and%20Labrador | Newfoundland and Labrador is one of Canada's provinces, and has established several official symbols.
Labrador, the mainland portion of the province, has its own distinct cultural identity and has established several unofficial symbols for itself.
Official symbols of Newfoundland and Labrador
Unofficial symbols of Labrador |
https://en.wikipedia.org/wiki/Symbols%20of%20New%20Brunswick | New Brunswick is one of Canada's provinces, and has established several provincial symbols.
Official symbols |
https://en.wikipedia.org/wiki/Symbols%20of%20Nova%20Scotia | Nova Scotia is one of Canada's provinces, and has established several provincial symbols.
Symbols |
https://en.wikipedia.org/wiki/Frame-bursting | Frame-bursting is a communication protocol feature used at the link layer in communication networks to alter the transmission characteristics in order to benefit from higher throughput. It is a technique sometimes used in communication protocols for shared mediums to achieve higher throughput by allowing the transmitter to send a series of frames in succession without relinquishing control of the transmission medium. Related techniques used to achieve the same goal include fast frames wherein the inter-frame wait interval is reduced, and jumbo frames wherein the size of the frame is increased. Frame bursting may also benefit from packet aggregation. Communication protocols for shared mediums are designed to relinquish the medium and wait for a while after the transmission of a MAC layer frame in order to facilitate the fair use of the medium by multiple users. Frame bursting may be permissible in certain scenarios such as when the link is point-to-point or when the signal from other users is indistinguishable from noise. Frame bursting allows for more data packets per time interval at the cost of wait time for other users.
In the case of wireless technology, the draft 802.11e quality of service specification allows frame bursting under some situations. Frame bursting may increase the throughput of any (point-to-point) 802.11a, b, g or n link connection under certain conditions. This is done by reducing the overhead associated with the wireless session in either of the following two modes:
Access point to client and vice versa
Client to client in ad hoc mode
Frame bursting and fast framing allow a wireless client to upload data at higher throughputs by using the inter-frame wait intervals to "burst" a sequence of up to three packets before waiting the required period. This allows more data to be sent with less waiting. However, their use can also result in unbalanced allocation of airtime where there are a mix of clients with and without Frame-Bursting. In such |
https://en.wikipedia.org/wiki/Symbols%20of%20Ontario | Ontario is a province of Canada that has established several official emblems and symbols to reflect the province's history, natural resources, and its people. In addition to official symbols, several other emblems and symbols exist that are commonly associated with the province.
Official symbols
Several emblems and symbols are used to officially represent the province, established through royal warrant or through the Legislative Assembly of Ontario. They include:
Other symbols
Several emblems and symbols exist that are commonly associated with province. They include:
Symbols of the lieutenant governor of Ontario
There exists several official emblems and symbols to represent the lieutenant governor of Ontario. |
https://en.wikipedia.org/wiki/Symbols%20of%20Quebec | The people and province of Quebec have created and established several symbols throughout Quebec's history to represent the collective identity of its residents. Many of Quebec's symbols are related to its history, to catholicism, to Quebec's winters and/or the fauna and flora of Quebec. The motif most commonly seen in Quebec's various symbols is the fleur de lys, which is associated with the French language and New France.
Symbols
The fleur-de-lis, one of Quebec's most common symbols, is an ancient symbol of the French monarchy and was first shown in Quebec on the shores of Gaspésie in 1534 when Jacques Cartier arrived in Quebec for the first time. Saint-Jean-Baptiste, the patron saint of Canadiens, is honoured every 24 June during Saint-Jean-Baptiste Day. The expression La belle province is still used as a nickname for the province. Finally, the Great Seal of Quebec is used to authenticate documents issued by the government of Quebec.
Coat of arms
The coat of arms of Quebec dates back to 1868, shortly after the creation of Quebec as a province of Canada. The arms were granted by a royal warrant issued by Queen Victoria.
The arms were adopted in their current form by the government of Quebec in 1939 to reflect Quebec's political history: the French regime is symbolised by the gold fleur-de-lis on a blue background; the British regime is symbolised by a gold lion on a red background; the pre-Confederation period is symbolised by three green maple leaves on a gold background.
Flag
The government of Quebec adopted the Fleurdelisé flag in 1948. The cross represents the faith of the province's founders, while the fleur-de-lys and blue colour recall Quebec's French origins.
When Samuel de Champlain founded Québec City in 1608, his ship hoisted the French merchant flag, which consisted of a white cross on a blue background. Later on, at the Battle of Carillon, in 1758, the Flag of Carillon was flown. This flag inspired the first members of the Saint-Jean-Baptist |
https://en.wikipedia.org/wiki/Symbols%20of%20Prince%20Edward%20Island | Prince Edward Island is one of Canada's provinces, and has established several provincial symbols.
Symbols |
https://en.wikipedia.org/wiki/Symbols%20of%20Saskatchewan | Saskatchewan is one of Canada's provinces, and has established several provincial symbols.
Symbols |
https://en.wikipedia.org/wiki/Symbols%20of%20the%20Northwest%20Territories | The Northwest Territories, one of Canada's territories, has established several territorial symbols.
Symbols |
https://en.wikipedia.org/wiki/Symbols%20of%20Nunavut | Nunavut is one of Canada's territories, and has established several territorial symbols.
Symbols of Nunavut
"Qimmiq" or "qimmik" is the Inuit language word for "dog"
Great Seal
Like Yukon, Nunavut does not have an official Great Seal. |
https://en.wikipedia.org/wiki/Symbols%20of%20Yukon | Yukon is one of Canada's territories, and has established several territorial symbols.
Official symbols
Great Seal
Like Nunavut, Yukon does not have an official Great Seal. |
https://en.wikipedia.org/wiki/G%20protein-coupled%20receptor%20kinase%202 | G-protein-coupled receptor kinase 2 (GRK2) is an enzyme that in humans is encoded by the ADRBK1 gene. GRK2 was initially called Beta-adrenergic receptor kinase (βARK or βARK1), and is a member of the G protein-coupled receptor kinase subfamily of the Ser/Thr protein kinases that is most highly similar to GRK3(βARK2).
Functions
G protein-coupled receptor kinases phosphorylate activated G protein-coupled receptors, which promotes the binding of an arrestin protein to the receptor. Arrestin binding to phosphorylated, active receptor prevents receptor stimulation of heterotrimeric G protein transducer proteins, blocking their cellular signaling and resulting in receptor desensitization. Arrestin binding also directs receptors to specific cellular internalization pathways, removing the receptors from the cell surface and also preventing additional activation. Arrestin binding to phosphorylated, active receptor also enables receptor signaling through arrestin partner proteins. Thus the GRK/arrestin system serves as a complex signaling switch for G protein-coupled receptors.
GRK2 and the closely related GRK3 phosphorylate receptors at sites that encourage arrestin-mediated receptor desensitization, internalization and trafficking rather than arrestin-mediated signaling (in contrast to GRK5 and GRK6, which have the opposite effect). This difference is one basis for pharmacological biased agonism (also called functional selectivity), where a drug binding to a receptor may bias that receptor’s signaling toward a particular subset of the actions stimulated by that receptor.
GRK2 is expressed broadly in tissues, but generally at higher levels than the related GRK3. GRK2 was originally identified as a protein kinase that phosphorylated the β2-adrenergic receptor, and has been most extensively studied as a regulator of adrenergic receptors (and other GPCRs) in the heart, where it has been proposed as a drug target to treat heart failure. Strategies to inhibit GRK2 include us |
https://en.wikipedia.org/wiki/Command%20Data%20Buffer | Command Data Buffer (CDB) was a system used by the United States Air Force's Minuteman ICBM force. CDB was a method to transfer targeting information from a Minuteman Launch Control Center to an individual missile by communications lines. Prior to CDB, new missile guidance would have to be physically loaded at the launch facility; the process usually took hours.
History
The surviving remnant of the Minuteman Command Control System (MICCS), CDB permitted the rapid, remote, retargeting of the Minuteman III fleet. CDB was operational at all Minuteman III wings by 15 Aug 1977. Minuteman II wings had a similar install, designated Improved Launch Control System, providing the older system the potential for remote retargeting.
Phaseout
CDB was replaced in the late 1990s by the Rapid Execution and Combat Targeting system, currently in use by United States ICBM forces.
See also
LGM-30 Minuteman
Launch control center (ICBM)
Improved Launch Control System - Minuteman II upgrade similar to CDB
Rapid Execution and Combat Targeting System (REACT) |
https://en.wikipedia.org/wiki/Complement%20control%20protein | Complement control protein are proteins that interact with components of the complement system.
The complement system is tightly regulated by a network of proteins known as "regulators of complement activation (RCA)" that help distinguish target cells as "self" or "non-self." A subset of this family of proteins, complement control proteins (CCP), are characterized by domains of conserved repeats that direct interaction with components of the complement system. These "Sushi" domains have been used to identify other putative members of the CCP family. There are many other RCA proteins that do not fall into this family.
Most CCPs prevent activation of the complement system on the surface of host cells and protect host tissues against damage caused by autoimmunity. Because of this, these proteins play important roles in autoimmune disorders and cancers.
Members
Most of the well-studied proteins within this family can be categorized in two classes:
Membrane-bound complement regulators
Membrane Cofactor Protein, MCP (CD46)
Decay Accelerating Factor, DAF (CD55)
Protectin (CD59)
Complement C3b/C4b Receptor 1, CR1 (CD35)
Complement Regulator of the Immunoglobulin Superfamily, CRIg
Soluble complement regulators
Factor H
C4-Binding Protein (C4bp)
Other proteins with characteristic CCP domains have been identified including members of the sushi domain containing (SUSD) protein family and Human CUB and sushi multiple domains family (CSMD).
Mechanisms of protection
Every cell in the human body is protected by one or more of the membrane-associated RCA proteins, CR1, DAF or MCP. Factor H and C4BP circulate in the plasma and are recruited to self-surfaces through binding to host-specific polysaccharides such as the glycosaminoglycans.
Most CCPs function by preventing convertase activity. Convertases, specifically the C3 convertases C3b.Bb and C4b.2a, are the enzymes that drive complement activation by activating C3b, a central component of the complement syst |
https://en.wikipedia.org/wiki/Lapel%20pin | A lapel pin, also known as an enamel pin, is a small pin worn on clothing, often on the lapel of a jacket, attached to a bag, or displayed on a piece of fabric. Lapel pins can be ornamental or can indicate the wearer's affiliation with an organization or cause. Before the popularity of wearing lapel pins, boutonnières were worn.
Popular usage
Lapel pins are frequently used as symbols of achievement and belonging in different organizations. Lapel pins from the organization are often collected by members and non-members alike.
Businesses, corporates, & political parties also use lapel pins to designate achievement and membership. Lapel pins are a common element of employee recognition programs, and they are presented to individuals as a symbol of an accomplishment. Like fraternity and sorority pins, these lapel pins instill a sense of belonging to an elite group of performers at the organization. Businesses also award lapel pins to employees more frequently to boost employee morale, productivity, and employee engagement.
The Soviet Union had great production of these. Besides pins showing political figures and as souvenirs for tourist spots, there were pins for various sports, cultural, and political gatherings and for technical achievements of the Soviet Union.
In recent years, pin collecting and trading has also become a popular hobby. Demand for pin designs based on popular cartoon characters and themes such as Disney, Betty Boop, and Hard Rock Cafe has surged and led to the creation of pin trading events and other social activities. Disney pin trading is a prime example of this.
Cultural significance
In the USSR and the People's Republic of China, the prominent lapel pins with portraits of Lenin and Mao Zedong, respectively, were worn by youth as well as by Communist party members or people who felt like showing their official political credo. In Czechoslovakia, the Mao badges/pins were worn in the late 1960s and early 1970s by non-conformist youth as a pr |
https://en.wikipedia.org/wiki/Benzisothiazolinone | Benzisothiazolinone (BIT) is an organic compound with the formula C6H4SN(H)CO. A white solid, it is structurally related to isothiazole, and is part of a class of molecules called isothiazolinones. BIT is widely used as a preservative and antimicrobial.
Usage
Benzisothiazolinone has a microbicide and a fungicide mode of action. It is widely used as a preservative, for example in:
emulsion paints, caulks, varnishes, adhesives, inks, and photographic processing solutions
home cleaning and car care products; laundry detergents, stain removers and fabric softeners;
industrial settings, for example in textile spin-finish solutions, leather processing solutions, preservation of fresh animal hides and skins
agriculture in pesticide formulations
gas and oil drilling in muds and packer fluids preservation.
In paints, it is commonly used alone or as a mixture with methylisothiazolinone. Typical concentrations in products are 200–400 ppm depending on the application area and the combination with other biocides. According to a study in Switzerland, 19% of the paints, varnishes and coatings contained BIT in 2000. The fraction in adhesives, sealants, plasters and fillers was shown at that time as 25%. A later study in 2014 shows a dramatic rise in usage, to 95.8% of house paints.
Home cleaning and other care products that are high in water are easily contaminated by microorganisms, so isothiazolinones are often used as a preservatives in these products because they are good at combatting a broad array of bacteria, fungi, and yeasts.
A Swiss investigation found that BIT is used in concentrations between 50 and 500 ppm in tattooing ink. According to regulations in the EU and Switzerland, BIT cannot be used in cosmetics. However, it is allowed in the United States and Canada.
Health hazards
Given sufficient dose and duration, dermal exposure can produce skin sensitization and allergic contact dermatitis, and is classified as an irritant for skin and eyes. BIT's low mol |
https://en.wikipedia.org/wiki/Natural%20exponential%20family | In probability and statistics, a natural exponential family (NEF) is a class of probability distributions that is a special case of an exponential family (EF).
Definition
Univariate case
The natural exponential families (NEF) are a subset of the exponential families. A NEF is an exponential family in which the natural parameter η and the natural statistic T(x) are both the identity. A distribution in an exponential family with parameter θ can be written with probability density function (PDF)
where and are known functions.
A distribution in a natural exponential family with parameter θ can thus be written with PDF
[Note that slightly different notation is used by the originator of the NEF, Carl Morris. Morris uses ω instead of η and ψ instead of A.]
General multivariate case
Suppose that , then a natural exponential family of order p has density or mass function of the form:
where in this case the parameter
Moment and cumulant generating functions
A member of a natural exponential family has moment generating function (MGF) of the form
The cumulant generating function is by definition the logarithm of the MGF, so it is
Examples
The five most important univariate cases are:
normal distribution with known variance
Poisson distribution
gamma distribution with known shape parameter α (or k depending on notation set used)
binomial distribution with known number of trials, n
negative binomial distribution with known
These five examples – Poisson, binomial, negative binomial, normal, and gamma – are a special subset of NEF, called NEF with quadratic variance function (NEF-QVF) because the variance can be written as a quadratic function of the mean. NEF-QVF are discussed below.
Distributions such as the exponential, Bernoulli, and geometric distributions are special cases of the above five distributions. For example, the Bernoulli distribution is a binomial distribution with n = 1 trial, the exponential distribution is a gamma distribution with s |
https://en.wikipedia.org/wiki/Pullback | In mathematics, a pullback is either of two different, but related processes: precomposition and fiber-product. Its dual is a pushforward.
Precomposition
Precomposition with a function probably provides the most elementary notion of pullback: in simple terms, a function of a variable where itself is a function of another variable may be written as a function of This is the pullback of by the function
It is such a fundamental process that it is often passed over without mention.
However, it is not just functions that can be "pulled back" in this sense. Pullbacks can be applied to many other objects such as differential forms and their cohomology classes; see
Pullback (differential geometry)
Pullback (cohomology)
Fiber-product
The pullback bundle is an example that bridges the notion of a pullback as precomposition, and the notion of a pullback as a Cartesian square. In that example, the base space of a fiber bundle is pulled back, in the sense of precomposition, above. The fibers then travel along with the points in the base space at which they are anchored: the resulting new pullback bundle looks locally like a Cartesian product of the new base space, and the (unchanged) fiber. The pullback bundle then has two projections: one to the base space, the other to the fiber; the product of the two becomes coherent when treated as a fiber product.
Generalizations and category theory
The notion of pullback as a fiber-product ultimately leads to the very general idea of a categorical pullback, but it has important special cases: inverse image (and pullback) sheaves in algebraic geometry, and pullback bundles in algebraic topology and differential geometry.
See also:
Pullback (category theory)
Fibred category
Inverse image sheaf
Functional analysis
When the pullback is studied as an operator acting on function spaces, it becomes a linear operator, and is known as the transpose or composition operator. Its adjoint is the push-forward, or, in the conte |
https://en.wikipedia.org/wiki/Salt%20substitute | A salt substitute, also known as low-sodium salt, is a low-sodium alternative to edible salt (table salt) marketed to reduce the risk of high blood pressure and cardiovascular disease associated with a high intake of sodium chloride while maintaining a similar taste.
The leading salt substitutes are non-sodium table salts, which have their tastes as a result of compounds other than sodium chloride. Non-sodium salts reduce daily sodium intake and reduce the health effects of this element.
Low sodium diet
According to current WHO guidelines, adults should consume less than 2,000 mg of sodium per day (i.e. about 5 grams of traditional table salt), and at least 3,510 mg of potassium per day. In Europe, adults and children consume about twice as much sodium as recommended by experts.
Research
In 2021, a large randomised controlled trial of 20,995 older people in China found that use of a potassium salt substitute in home cooking over a five-year period reduced the risk of stroke by 14%, major cardiovascular events by 13% and all-cause mortality by 12% compared to use of regular table salt.
The study found no significant difference in hyperkalaemia between the two groups, though people with serious kidney disease were excluded from the trial. The salt substitute used was 25% potassium chloride and 75% sodium chloride.
A 2022 Cochrane review of 26 trials involving salt substitutes found their use probably slightly reduces blood pressure, non-fatal stroke, non-fatal acute coronary syndrome and heart disease death in adults compared to use of regular table salt. A separate systematic review and meta-analysis published in the same year of 21 trials involving salt substitutes found protective effects of salt substitute on total mortality, cardiovascular mortality and cardiovascular events.
Examples
Potassium
Potassium closely resembles the saltiness of sodium. In practice, potassium chloride (also known as potassium salt) is the most commonly used salt substitute. |
https://en.wikipedia.org/wiki/Venue%20%28sound%20system%29 | VENUE is a brand of live sound digital mixing consoles introduced by Digidesign in February 2005. The family now includes 5 different consoles and a number of ways they can be configured. They can all be connected to Pro Tools, the audio editing software also created by Avid/Digidesign, to provide recording and 'Virtual Soundcheck' facilities. One of the system's key marketing points is its use of the same AAX DSP/TDM plugins as Pro Tools, an industry standard digital audio workstation (DAW). This is designed to enable the sounds recorded by the artist in the studio to be easily recreated on stage, and to allow for greater flexibility in signal processing without heavy and mechanical-shock-sensitive racks of external processors. There is also a PC-based offline editor for creation and editing of show files, although there is no audio processing in the editor.
Digidesign was acquired by Avid Technology in 1995, but its products continued to be branded Digidesign until the brand was phased out in 2010, and VENUE systems are now branded Avid.
Console products
D-Show
This was the first and largest control surface in the VENUE range, and was discontinued in 2015. The main unit, which includes a trackball, master controls, faders and meters, can be expanded with up to three sidecars to give a maximum of 56 faders. One sidecar was included in the standard package.
The console can either be connected to the Mix Rack, which contains all FOH I/O, DSP and stage I/O in one rack, or to the combination of an FOH rack and a Stage Rack. The FOH rack contains the DSP and a small amount of inputs and outputs, and the Stage Rack contains the main stage I/O. They are linked using a digital snake cable, with an option to use a redundant cable in parallel. There can be a maximum of two Stage Racks linked to one FOH rack.
Profile
This smaller console contains all of the features of the larger console in a smaller form factor, and quickly became the industry's standard touring and fe |
https://en.wikipedia.org/wiki/Phragmosome | The phragmosome is a sheet of cytoplasm forming in highly vacuolated plant cells in preparation for mitosis. In contrast to animal cells, plant cells often contain large central vacuoles occupying up to 90% of the total cell volume and pushing the nucleus against the cell wall. In order for mitosis to occur, the nucleus has to move into the center of the cell. This happens during G2 phase of the cell cycle.
Initially, cytoplasmic strands form that penetrate the central vacuole and provide pathways for nuclear migration. Actin filaments along these cytoplasmic strands pull the nucleus into the center of the cell. These cytoplasmic strands fuse into a transverse sheet of cytoplasm along the plane of future cell division, forming the phragmosome. Phragmosome formation is only clearly visible in dividing plant cells that are highly vacuolated.
Just before mitosis, a dense band of microtubules appears around the phragmosome and the future division plane just below the plasma membrane. This preprophase band marks the equatorial plane of the future mitotic spindle as well as the future fusion sites for the new cell plate with the existing cell wall. It disappears as soon as the nuclear envelope breaks down and the mitotic spindle forms.
When mitosis is completed, the cell plate and new cell wall form starting from the center along the plane occupied by the phragmosome. The cell plate grows outwards until it fuses with the cell wall of the dividing cell at exactly the spots predicted by the preprophase band. |
https://en.wikipedia.org/wiki/Piwi | Piwi (or PIWI) genes were identified as regulatory proteins responsible for stem cell and germ cell differentiation. Piwi is an abbreviation of P-element Induced WImpy testis in Drosophila. Piwi proteins are highly conserved RNA-binding proteins and are present in both plants and animals. Piwi proteins belong to the Argonaute/Piwi family and have been classified as nuclear proteins. Studies on Drosophila have also indicated that Piwi proteins have no slicer activity conferred by the presence of the Piwi domain. In addition, Piwi associates with heterochromatin protein 1, an epigenetic modifier, and piRNA-complementary sequences. These are indications of the role Piwi plays in epigenetic regulation. Piwi proteins are also thought to control the biogenesis of piRNA as many Piwi-like proteins contain slicer activity which would allow Piwi proteins to process precursor piRNA into mature piRNA.
Protein structure and function
The structure of several Piwi and Argonaute proteins (Ago) have been solved. Piwi proteins are RNA-binding proteins with 2 or 3 domains: The N-terminal PAZ domain binds the 3'-end of the guide RNA; the middle MID domain binds the 5'-phosphate of RNA; and the C-terminal PIWI domain acts as an RNase H endonuclease that can cleave RNA. The small RNA partners of Ago proteins are microRNAs (miRNAs). Ago proteins utilize miRNAs to silence genes post-transcriptionally or use small-interfering RNAs (siRNAs) in both transcription and post-transcription silencing mechanisms. Piwi proteins interact with piRNAs (28–33 nucleotides) that are longer than miRNAs and siRNAs (~20 nucleotides), suggesting that their functions are distinct from those of Ago proteins.
Human Piwi proteins
Presently there are four known human Piwi proteins—PIWI-like protein 1, PIWI-like protein 2, PIWI-like protein 3 and PIWI-like protein 4. Human Piwi proteins all contain two RNA binding domains, PAZ and Piwi. The four PIWI-like proteins have a spacious binding site within the PAZ dom |
https://en.wikipedia.org/wiki/Spongin | Spongin, a modified type of collagen protein, forms the fibrous skeleton of most organisms among the phylum Porifera, the sponges. It is secreted by sponge cells known as spongocytes.
Spongin gives a sponge its flexibility. True spongin is found only in members of the class Demospongiae.
Research directions
Use in the removal of phenolic compounds from wastewater
Researchers have found spongin to be useful in the photocatalytic degradation and removal of bisphenols (such as BPA) in wastewater. A heterogeneous catalyst consisting of a spongin scaffold for iron phthalocyanine (SFe) in conjunction with peroxide and UV radiation has been shown to remove phenolic wastes more quickly and efficiently than conventional methods. Other research using spongin scaffolds for the immobilization of Trametes versicolor Laccase has shown similar results in phenol degradation. |
https://en.wikipedia.org/wiki/Diagnostic%20board | In electronic systems a diagnostic board is a specialized device with diagnostic circuitry on a printed circuit board that connects to a computer or other electronic equipment replacing an existing module, or plugging into an expansion card slot.
A multi-board electronic system such as a computer comprises multiple printed circuit boards or cards connected via connectors. When a fault occurs in the system, it is sometimes possible to isolate or identify the fault by replacing one of the boards with a diagnostic board. A diagnostic board can range from extremely simple to extremely sophisticated.
Simple standard diagnostic plug-in boards for computers are available that display numeric codes to assist in identifying issues detected during the power-on self-test executed automatically during system startup.
Dummy board
A dummy board provides a minimal interface. This type of diagnostic board in intended to confirm that the interface is correctly implemented. For example, a PC motherboard manufacturer can test PCI functionality of a PC motherboard by connecting a dummy PCI board into each PCI slot on the motherboard
Extender board
An extender board (or board extender, card extender, extender card) is a simple circuit board that interposes between a card cage backplane and the circuit board of interest to physically 'extend' the circuit board of interest out from the card cage allowing access to both sides of the circuit board to connect diagnostic equipment such as an oscilloscope or systems analyzer. For example, a PCI extender board can be plugged into a PCI slot on a computer motherboard, and then a PCI card connected to the extender board to 'extend' the board into free space for access. This approach was common in the 1970s and 1980s particularly on S-100 bus systems.
The concept can become unworkable when signal timing is affected by the length of the signal paths on the diagnostic board, as well as introducing Radio Frequency Interference (RFI) into the ci |
https://en.wikipedia.org/wiki/Roland%20GR-500 | The Roland GR-500 is a guitar synthesizer. Manufactured by the Roland Corporation and FujiGen in 1977, it was one of the first guitar synthesizers.
Overview
The synthesizer module included Polyensemble, Bass, Solo Synth, and External synthesizer control. Much of the voltage-controller filter and voltage controlled amplifier sections were based on previous analog mono-synths from Roland. Sliders adjust the VCO, VCF, VCA, and LFO sections; there is no memory to store settings.
The synth module is controlled by a highly modified guitar: the GS-500 guitar controller was built in a partnership between Roland and the large Japanese guitar builder FujiGen. The result was the Fuji Roland Corporation, established in 1977 through a joint capital investment.
The GS-500 controller used a special pickup system that connected to the synth module via Roland's own 24-pin interface. Controls on the guitar primarily were for adjusting the relative volume of the different sections: guitar, polyensemble, bass, solo section, and external synth. The GS-500 guitar does not have a standard 1/4" guitar output and cannot be used without the GR-500 synthesizer module. It has a particular infinite sustain system: the frets in the GR-500 are connected to its electrical ground, and when a player frets a string, an electric current passed through the string. The electric signal passing through the string is a greatly amplified version of the string signal detected by the divided hexaphonic pickup. Large magnets replaced the traditional "neck" pickup. As a result of Fleming's Law, the alternating electric current in the string passing through the strong magnetic field caused the string to vibrate and create a feedback loop and infinite sustain. The GS-500 used a bridge with plastic saddles to electrically isolate each string.
Here is a brief description of each section:
Guitar: this is the output from a humbucking pickup. A three position EQ switch, plus tone control, provided variation to t |
https://en.wikipedia.org/wiki/Implication%20graph | In mathematical logic and graph theory, an implication graph is a skew-symmetric, directed graph composed of vertex set and directed edge set . Each vertex in represents the truth status of a Boolean literal, and each directed edge from vertex to vertex represents the material implication "If the literal is true then the literal is also true". Implication graphs were originally used for analyzing complex Boolean expressions.
Applications
A 2-satisfiability instance in conjunctive normal form can be transformed into an implication graph by replacing each of its disjunctions by a pair of implications. For example, the statement can be rewritten as the pair . An instance is satisfiable if and only if no literal and its negation belong to the same strongly connected component of its implication graph; this characterization can be used to solve 2-satisfiability instances in linear time.
In CDCL SAT-solvers, unit propagation can be naturally associated with an implication graph that captures all possible ways of deriving all implied literals from decision literals, which is then used for clause learning. |
https://en.wikipedia.org/wiki/Hierarchical%20file%20system | In computing, "a hierarchical file system is a file system that uses directories to organize files into a tree structure."
In a hierarchical file system, directories contain information about both files and other directories, called subdirectories which, in turn, can point to other subdirectories, and so on. This is organized as a tree structure, or hierarchy, generally portrayed with the root at the top. The root directory is the base of the hierarchy, and is usually stored at some fixed location on disk.
A hierarchical file system contrasts with a flat file system, where information about all files is stored in a single directory, and there are no subdirectories.
Almost all file systems today are hierarchical. What is referred to as a file system is a specific instance of a hierarchical system. For example, NTFS, HPFS, and ext4, all implement a hierarchical system with different features for buffering, file allocation, and file recovery.
Concepts
Path
"A file path describes the location of a file in a web site's folder [directory] structure." That is, it represents the directory nodes visited from the root directory to the file as a list of node names, with the items in the list separated by path separators. The path separator is > on Multics, / on Unix-like systems, and \ on MS-DOS 2.0 and later, Windows, and OS/2 systems.
An absolute path begins at the root directory; that is, begins with a path separator character, which, at the beginning of a path, represents the root directory. A path consisting only of a path separator character refers to the root directory.
Working directory
The working directory of a process is a directory dynamically associated with each process. Files are searched relative to the working directory, rather than from the root directory. At logon, the user's working directory is set to their home directory; it can be set afterwards by using a command.
A relative path represents the directory nodes visited from the working director |
https://en.wikipedia.org/wiki/495%20%28number%29 | 495 (four hundred [and] ninety-five) is the natural number following 494 and preceding 496. It is a pentatope number (and so a binomial coefficient ). The maximal number of pieces that can be obtained by cutting an annulus with 30 cuts.
Kaprekar transformation
The Kaprekar's routine algorithm is defined as follows for three-digit numbers:
Take any three-digit number, other than repdigits such as 111. Leading zeros are allowed.
Arrange the digits in descending and then in ascending order to get two three-digit numbers, adding leading zeros if necessary.
Subtract the smaller number from the bigger number.
Go back to step 2 and repeat.
Repeating this process will always reach 495 in a few steps. Once 495 is reached, the process stops because 954 – 459 = 495.
Example
For example, choose 495:
495
The only three-digit numbers for which this function does not work are repdigits such as 111, which give the answer 0 after a single iteration. All other three-digit numbers work if leading zeros are used to keep the number of digits at 3:
211 – 112 = 099
990 – 099 = 891 (rather than 99 – 99 = 0)
981 – 189 = 792
972 – 279 = 693
963 – 369 = 594
954 − 459 = 495
The number 6174 has the same property for the four-digit numbers, albeit has a much greater percentage of workable numbers.
See also
Collatz conjecture — sequence of unarranged-digit numbers always ends with the number 1. |
https://en.wikipedia.org/wiki/Stream%20thrust%20averaging | In fluid dynamics, stream thrust averaging is a process used to convert three-dimensional flow through a duct into one-dimensional uniform flow. It makes the assumptions that the flow is mixed adiabatically and without friction. However, due to the mixing process, there is a net increase in the entropy of the system. Although there is an increase in entropy, the stream thrust averaged values are more representative of the flow than a simple average as a simple average would violate the second Law of Thermodynamics.
Equations for a perfect gas
Stream thrust:
Mass flow:
Stagnation enthalpy:
Solutions
Solving for yields two solutions. They must both be analyzed to determine which is the physical solution. One will usually be a subsonic root and the other a supersonic root. If it is not clear which value of velocity is correct, the second law of thermodynamics may be applied.
Second law of thermodynamics:
The values and are unknown and may be dropped from the formulation. The value of entropy is not necessary, only that the value is positive.
One possible unreal solution for the stream thrust averaged velocity yields a negative entropy. Another method of determining the proper solution is to take a simple average of the velocity and determining which value is closer to the stream thrust averaged velocity. |
https://en.wikipedia.org/wiki/World%20War%20I%20cryptography | With the rise of easily-intercepted wireless telegraphy, codes and ciphers were used extensively in World War I. The decoding by British Naval intelligence of the Zimmermann telegram helped bring the United States into the war.
Trench codes were used by field armies of most of the combatants (Americans, British, French, German) in World War I.
The most commonly used codes were simple substitution ciphers. More important messages generally used mathematical encryption for extra security. The use of these codes required the distribution of codebooks to military personnel, which proved to be a security liability since these books could be stolen by enemy forces.
Britain
British decrypting was carried out in Room 40 by the Royal Navy and in MI1 by British Military (Army) Intelligence.
Zimmermann telegram
Arthur Zimmermann
MI1 British Military (Army) Intelligence
Room 40 Royal Navy (Britain)
Alastair Denniston Room 40
James Alfred Ewing Room 40, first head
Nigel de Grey Room 40
William R. Hall ‘Blinker’ Hall, Room 40, second head
Malcolm Vivian Hay, MI1(b), head from 1915
Dilly Knox Room 40
Oliver Strachey MI1
William Montgomery (cryptographer) Room 40
Playfair cipher
Russia
In the 1914 Battle of Tannenberg, different corps of the Russian Imperial army were unable to decipher each others messages, so they sent them in plain text. They were easily intercepted. Meanwhile, German cryptanalysts were also able to read the enciphered ones.
Ernst Fetterlein was in the Tsarist Russian Ministry of Foreign Affairs from 1896 and solved (among others) German, Austrian and British codes. He became chief cryptographer with the rank of admiral. With the Russian Revolution in 1917 he fled to Britain and was recruited to Room 40 in June 1918 to work on Austrian, Bolshevik and Georgian codes.
The Russians used an overcomplicated version of the Vigenère Cipher. It was broken within three days by Austro-Hungarian cryptanalyst Hermann Pokorny.
France
The French Army |
https://en.wikipedia.org/wiki/Rubroboletus%20pulcherrimus | Rubroboletus pulcherrimus—known as Boletus pulcherrimus until 2015—is a species of mushroom in the family Boletaceae. It is a large bolete from Western North America with distinguishing features that include a netted surface on the stem, a red to brown cap and stem color, and red pores that stain blue upon injury. Until 2005 this was the only bolete that has been implicated in the death of someone consuming it; a couple developed gastrointestinal symptoms in 1994 after eating this fungus with the husband succumbing. Autopsy revealed infarction of the midgut.
Taxonomy
American mycologists Harry D. Thiers and Roy E. Halling were aware of confusion on the west coast of North America over red-pored boletes; two species were traditionally recognised—Boletus satanas and Boletus eastwoodiae. However, they strongly suspected the type specimen of the latter species was in fact the former. In reviewing material they published a new name for the taxon, which Thiers had written about in local guidebooks as B. eastwoodiae, as they felt that name to be invalid. Hence in 1976 they formally described Boletus pulcherrimus, from the Latin pulcherrimus, meaning "very pretty". It was transferred to the genus Rubroboletus in 2015 along with several other allied reddish colored, blue-staining bolete species such as B. eastwoodiae and B. satanas.
Description
Colored various shades of olive- to reddish-brown, the cap may sometimes reach in diameter and is convex in shape before flattening at maturity. The cap surface may be smooth or velvety when young, but may be scaled in older specimens; the margin of the cap is curved inwards in young specimens but rolls out and flattens as it matures.
The cap may reach a thickness of when mature. The adnate (attached squarely to the stem) poroid surface is bright red to dark red or red-brown and bruise dark blue or black; there are 2 to 3 pores per mm in young specimens, and in maturity they expand to about 1 or 2 per mm. In cross section, the tu |
https://en.wikipedia.org/wiki/Cervical%20spinal%20stenosis | Cervical spinal stenosis is a bone disease involving the narrowing of the spinal canal at the level of the neck. It is frequently due to chronic degeneration, but may also be congenital. Treatment is frequently surgical.
Cervical spinal stenosis is one of the most common forms of spinal stenosis, along with lumbar spinal stenosis (which occurs at the level of the lower back instead of in the neck). Thoracic spinal stenosis, at the level of the mid-back, is much less common. Cervical spinal stenosis can be far more dangerous by compressing the spinal cord. Cervical canal stenosis may lead to serious symptoms such as major body weakness and paralysis. Such severe spinal stenosis symptoms are virtually absent in lumbar stenosis, however, as the spinal cord terminates at the top end of the adult lumbar spine, with only nerve roots (cauda equina) continuing further down.
Diagnosis
Treatments
Nonsurgical treatment
Potential nonsurgical treatments include:
Education about the course of the condition and how to relieve symptoms
Medicines to relieve pain and inflammation, such as acetaminophen or nonsteroidal anti-inflammatory drugs (NSAIDs)
Exercise, to maintain or achieve overall good health, aerobic exercise, strength training muscles that support the neck and spine, core exercises, and posture correction.
Weight loss, to relieve symptoms and slow the progression of the stenosis
Physical therapy, to provide education, instruction, and support for self-care; physical therapy instructs on stretching and strength exercises that may lead to a decrease in pain and other symptoms
Surgery
Potential surgical treatments include:
Anterior cervical discectomy and fusion – A surgical treatment of nerve root or spinal cord compression by decompressing the spinal cord and nerve roots of the cervical spine with a discectomy to stabilize the corresponding vertebrae.
Laminoplasty – A surgical procedure to relieve pressure on the spinal cord by cutting the lamina on both si |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.