source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Canadian%20National%20Calibration%20Reference%20Centre | The National Calibration Reference Centre for Bioassay and In Vivo Monitoring (NCRC) is administered by the Radiation Protection Bureau of the Canadian Federal Department of Health. It was created in 1982 through a Memorandum of Understanding (MOU) signed between the regulator of the nuclear industry, the Canadian Nuclear Safety Commission (CNSC), formerly the Atomic Energy Control Board, and the Department of Health with the specific mission of providing "practical reference standards" for measurements used for internal dosimetry. Two other Reference Centres were created at the same time. These were to have equivalent roles for (1) external dosimetry and (2) radon and radioactive atmospheres, and were administered, respectively, by the Canadian National Research Council and the federal Department of Energy, Mines and Resources. The choice of the three agencies to act as Reference Centres was based on their expertise acquired over years of work in their respective fields and the fact that they operate independently of both the CNSC and the nuclear industry.
The role of the NCRC is in keeping with Health Canada's mandate to protect and preserve the health of Canadians. Specifically, its focus is to provide:
Independent external quality control on a regular basis through regularly scheduled intercomparisons;
Improvement of measurement techniques through advice and training; and
Information on improving calibrations by supplying practical reference standards and the necessary techniques.
While the provision of intercomparison programs is its principal function, the NCRC also provides the following:
Advice and assistance;
Bioassay measurements and internal dose assessments of suspected radionuclide intakes;
Research in collaboration with other agencies on radionuclide metabolism and biokinetics; and
Methods development in support of its intercomparison programs and research studies.
The NCRC provides a national calibration reference service to universities, ho |
https://en.wikipedia.org/wiki/Healing%20of%20periapical%20lesions | Apical periodontitis is typically the body's defense response to the threat of microbial invasion from the root canal. Primary among the members of the host defense mechanism is the polymorphonuclear leukocyte, otherwise known as the neutrophil. The task of the neutrophil is to locate and destroy microbes that intrude into the body – anywhere in the body – and they represent the hallmark of acute inflammation.
The body's response to microbial invasion
In response to tissue injury, neutrophils leave the circulatory system in great numbers and gather at the site of tissue injury. They are drawn to the site by chemotaxis, following a concentration gradient of chemotactic molecules until they reach the site of greatest concentration: the site of injury and microbial presence. Once there, the antimicrobial action of superoxide and hydrogen peroxide, derived from the metabolic processes of the neutrophils, act to combat the microbial invasion. While primarily mobilized to kill the invading microorganisms, the neutrophils actually cause a significant amount of host tissue damage as well. Although the neutrophils themselves rarely remain alive for more than a few days, the excessive accumulation of dead neutrophils and the enzymes they released is a major cause of tissue breakdown in the acute phases of apical periodontitis.
Soon after inflammation has been initiated, macrophages enter the scene and, if not controlled by the initial ambush of neutrophils and their tactics, the microbial invasion is faced with a second strike consisting of these leukocytes, along with lymphocytes. Together, the cells of this second strike compose the bulk of the apical periodontitis lesion and serve an important role in the subsequent chronic phase of inflammation of apical periodontitis, as they can live for many months. Some researchers posit that it must not be macrophages that are involved, as they could not appropriately discriminate between the varied array of opsonized entities a |
https://en.wikipedia.org/wiki/Soluble%20low-density%20lipoprotein%20receptor-related%20protein | Soluble low-density lipoprotein receptor-related protein (sLRP, LRP-515) is a biological substance naturally produced by the human body. This protein has been found to bind to and neutralize anywhere from 70 to 90 percent of the amyloid-beta peptide that also naturally circulates in healthy human or mouse plasma. Impairment of this function is strongly associated with, and may soon be shown definitively to be the principal cause of, Alzheimer's disease. |
https://en.wikipedia.org/wiki/List%20of%20Intel%20Itanium%20processors | The Itanium from Intel is a high-end server and supercomputer microprocessor.
Itanium (2001)
Merced (180 nm)
Steppings: C0, C1 and C2. CPUID: 0007000604h (stepping C0), 0007000704h (stepping C1) or 0007000804h (stepping C2). Transistor count: 25.4 million for CPU, 295 million for the external L3 cache. The FSB data bus is 64 bits wide, not 128 like in Itanium 2.
Itanium 2 (2002-2007)
Itanium 2 uses socket PAC611 with a 128 bit wide FSB. The 90 nm CPUs (9000 and 9100 series) bring dual-core chips and an updated microarchitecture adding multithreading and splitting the L2 cache into a 256 KB data cache and 1 MB instruction cache per core (the pre-9000 series L2 cache being a 256 KB common cache). All Itaniums except some 130 nm models are capable of >2-socket SMP.
McKinley (180 nm)
Stepping: B3. Die size: 421 mm². Transistor count: 221 million. CPUID: 001F000704h
Madison (130 nm)
Stepping: B1. Die size: 374 mm². Transistor count: 410 million. CPUID: 001F010504h. The Madison 9M table contains the 4MB and 6MB successors of the first Madisons.
Deerfield
The same chip as Madison, but at a lower voltage.
Madison 9M (130 nm)
Steppings: A1 and A2. Die size: 432 mm². Transistor count: 592 million. CPUID: 001F020104h (stepping A1) or 001F020204h (stepping A2). 9M is the chip of all the third generation Itanium 2s, irrespective of the amount of enabled cache.
Fanwood
The same chip as Madison 9M, but restricted to 2-socket and uniprocessor systems.
HP mx2 MCM (130 nm)
This multi-chip module codenamed Hondo is not an Intel product, but a separate project of Hewlett-Packard to pack two CPUs onto one PAC611 socket. The S-Spec SL75Z was assigned to the chips that Intel sent to HP for use in mx2.
Montecito (90 nm)
Steppings: C1 and C2. Die size: 596 mm². Transistor count: 1720 million. CPUID: 0020000504h (stepping C1) or 0020000704h (stepping C2). All processors can support the legacy 400 MT/s FSB. From Montecito onwards all Itaniums are MP-capable.
Montvale (90 nm)
The c |
https://en.wikipedia.org/wiki/Self-propelled%20modular%20transporter | A self-propelled modular transporter or sometimes self-propelled modular trailer (SPMT) is a platform heavy hauler with a large array of wheels which is an upgraded version of a hydraulic modular trailer. SPMTs are used for transporting massive objects, such as large bridge sections, oil refining equipment, cranes, motors, spacecraft and other objects that are too big or heavy for trucks. Ballast tractors can however provide traction and braking for the SPMTs on inclines and descents.
SPMTs are used in many industry sectors worldwide such as the construction and oil industries, in the shipyard and offshore industry, for road transportation, on plant construction sites and even for moving oil platforms. They have begun to be used to replace bridge spans in the United States, Europe, Asia and more recently Canada.
A typical SPMT has a grid of computer-controlled axles, usually 2 axles across and 4–8 axles along. When two (or more) axles are placed in series, this is called an axle line. All axles are individually controllable, in order to evenly distribute weight and to steer accurately. Each axle can swivel through 270°, with some manufacturers offering up to a full 360° of motion. The axles are coordinated by the control system to allow the SPMT to turn, move sideways or even rotate in place. Some SPMTs allow the axles to telescope independently of each other so that the load can be kept flat and evenly distributed while moving over uneven terrain. Each axle can also contain a hydrostatic drive unit.
A hydraulic power pack can be attached to the SPMT to provide power for steering, suspension and drive functions. This power pack is driven by an internal combustion engine. A single power pack can drive a string of SPMTs. As SPMTs often carry the world's heaviest loads on wheeled vehicles, they are very slow, often moving at under while fully loaded. Some SPMTs are controlled by a worker with a hand-held control panel, while others have a driver cabin. Multiple SP |
https://en.wikipedia.org/wiki/Edge-matching%20puzzle | An edge-matching puzzle is a type of tiling puzzle involving tiling an area with (typically regular) polygons whose edges are distinguished with colours or patterns, in such a way that the edges of adjacent tiles match.
Edge-matching puzzles are known to be NP-complete, and capable of conversion to and from equivalent jigsaw puzzles and polyomino packing puzzle.
The first edge-matching puzzles were patented in the U.S. by E. L. Thurston in 1892. Current examples of commercial edge-matching puzzles include the Eternity II puzzle, Tantrix, Kadon Enterprises' range of edge-matching puzzles, and the Edge Match Puzzles iPhone app.
Notable variations
MacMahon Squares
MacMahon Squares is the name given to a recreational math puzzle suggested by British mathematician Percy MacMahon, who published a treatise on edge-colouring of a variety of shapes in 1921. This particular puzzle uses 24 tiles consisting of all permutations of 3 colors for the edges of a square. The tiles must be arranged into a 6×4 rectangular area such that all edges match and, furthermore, only one color is used for the outside edge of the rectangle.
This puzzle can be extended to tiles with permutations of 4 colors, arranged in 10×7. In either case, the squares are a subset of the Wang tiles, reducing tiles that are similar under rotation. Solutions number well into the thousands.
MacMahon Squares, along with variations on the idea, was commercialized as Multimatch.
TetraVex
TetraVex is a computer game that presents the player with a square grid and a collection of tiles, by default nine square tiles for a 3×3 grid. Each tile has four single-digit numbers, one on each edge. The objective of the game is to place the tiles into the grid in the proper position, completing this puzzle as quickly as possible. The tiles cannot be rotated, and two can be placed next to each other only if the numbers on adjacent edges match.
TetraVex was inspired by "the problem of tiling the plane" as described by Don |
https://en.wikipedia.org/wiki/Strictly%20convex%20space | In mathematics, a strictly convex space is a normed vector space (X, || ||) for which the closed unit ball is a strictly convex set. Put another way, a strictly convex space is one for which, given any two distinct points x and y on the unit sphere ∂B (i.e. the boundary of the unit ball B of X), the segment joining x and y meets ∂B only at x and y. Strict convexity is somewhere between an inner product space (all inner product spaces being strictly convex) and a general normed space in terms of structure. It also guarantees the uniqueness of a best approximation to an element in X (strictly convex) out of a convex subspace Y, provided that such an approximation exists.
If the normed space X is complete and satisfies the slightly stronger property of being uniformly convex (which implies strict convexity), then it is also reflexive by Milman–Pettis theorem.
Properties
The following properties are equivalent to strict convexity.
A normed vector space (X, || ||) is strictly convex if and only if x ≠ y and || x || = || y || = 1 together imply that || x + y || < 2.
A normed vector space (X, || ||) is strictly convex if and only if x ≠ y and || x || = || y || = 1 together imply that || αx + (1 − α)y || < 1 for all 0 < α < 1.
A normed vector space (X, || ||) is strictly convex if and only if x ≠ 0 and y ≠ 0 and || x + y || = || x || + || y || together imply that x = cy for some constant c > 0;
A normed vector space (X, || ||) is strictly convex if and only if the modulus of convexity δ for (X, || ||) satisfies δ(2) = 1.
See also
Uniformly convex space
Modulus and characteristic of convexity |
https://en.wikipedia.org/wiki/Frequency-locked%20loop | A frequency-lock, or frequency-locked loop (FLL), is an electronic control system that generates a signal that is locked to the frequency of an input or "reference" signal. This circuit compares the frequency of a controlled oscillator to the reference, automatically raising or lowering the frequency of the oscillator until its frequency (but not necessarily its phase) is matched to that of the reference.
A frequency-locked loop is an example of a control system using negative feedback. Frequency-lock loops are used in radio, telecommunications, computers and other electronic applications to generate stable frequencies, or to recover a signal from a noisy communication channel.
See also
Phase-locked loop |
https://en.wikipedia.org/wiki/Steffensen%27s%20inequality | Steffensen's inequality is an equation in mathematics named after Johan Frederik Steffensen.
It is an integral inequality in real analysis, stating:
If ƒ : [a, b] → R is a non-negative, monotonically decreasing, integrable function
and g : [a, b] → [0, 1] is another integrable function, then
where |
https://en.wikipedia.org/wiki/Instruments%20%28software%29 |
Instruments (formerly Xray) is an application performance analyzer and visualizer, integrated in Xcode 3.0 and later versions of Xcode. It is built on top of the DTrace tracing framework from OpenSolaris, which was ported to Mac OS X v10.5 and which is available in all following versions of macOS.
Instruments shows a time line displaying any event occurring in the application, such as CPU activity variation, memory allocation, and network and file activity, together with graphs and statistics. Group of events are monitored via customizable "instruments", which have the ability to record user generated events and replay (emulate) them exactly as many times as needed, so a developer can see the effect of code changes without actually doing the repetitive work. The Instrument Builder feature allows the creation of custom analysis instruments.
Features
Built-in instruments can track
CPU activity of processes and threads.
Memory allocation and release, garbage collection and memory leaks.
File reads, writes, locks.
Network activity and traffic. This instrument works like Activity Monitor but also stores the data for future reference.
Graphics and inner workings of OpenGL and Metal.
Energy diagnostics and "dead" objects.
UI automation and Core animation.
User events, such as keyboard keys pressed and mouse moves and clicks with exact time.
See also
List of performance analysis tools
Dashcode
Shark (application) |
https://en.wikipedia.org/wiki/Strassmann%27s%20theorem | In mathematics, Strassmann's theorem is a result in field theory. It states that, for suitable fields, suitable formal power series with coefficients in the valuation ring of the field have only finitely many zeroes.
History
It was introduced by .
Statement of the theorem
Let K be a field with a non-Archimedean absolute value | · | and let R be the valuation ring of K. Let f(x) be a formal power series with coefficients in R other than the zero series, with coefficients an converging to zero with respect to | · |. Then f(x) has only finitely many zeroes in R. More precisely, the number of zeros is at most N, where N is the largest index with |aN| = max |an|.
As a corollary, there is no analogue of Euler's identity, e2πi = 1, in Cp, the field of p-adic complex numbers.
See also
p-adic exponential function |
https://en.wikipedia.org/wiki/Marula%20oil | Marula oil is extracted from the kernels (nuts) of the fruits of the Marula tree (Sclerocarya birrea), from the family Anacardiaceae. There are two types of marula oil, the oil extracted from the seeds and the oil extracted from the nut's hard shell. Marula oil is traditionally used in cosmetics, in food as a cooking oil and meat preservative and to treat leather. Marula oil can also be used as body lotion. In Namibia Marula fruit is processed into a range of juices, jellies and jams.
Chemical composition
Marula oil contains a large proportion of monounsaturated fatty acids which make the oil very stable. The fatty acid composition of marula oil includes:
Monounsaturated fatty acids:
Oleic acid (70–78%)
Polyunsaturated fatty acids:
Linoleic acid (4.0–7.0%)
Alpha-linolenic acid (0.1–0.7%)
Saturated fatty acids:
Palmitic acid (9–12%)
Stearic acid (5.0–8.0%)
Arachidonic acid (0.3–0.7%)
Tocopherols, sterols, flavonoids, procyanidin, gallotannin and catechins are also found in marula oil.
Physical properties
Marula oil has a clear, light yellow colour and a nutty aroma. It has a saponification value of approximately 188–199 and a specific gravity of 0.91–0.92 (at 15 °C).
Traditional uses
The Tsonga people of South Africa and Mozambique have used the oil as a moisturising body lotion for women and also as a massage oil for babies. In the past, Namibian women used marula oil rather than water to clean themselves.
Marula oil is used in diets, especially for people of the Inhambane Province in Mozambique, Owamboland in north central Namibia, Northern KwaZulu-Natal in South Africa and the Zvishavane district of Zimbabwe. Furthermore, marula plays an important role in the diet of Bushmen and Bantus. The Venda use the oil from the kernels to preserve meat, which enables it to last up to a year. Marula oil is considered a delicacy by local people, and is added to many traditional and modern recipes. |
https://en.wikipedia.org/wiki/Spijker%27s%20lemma | In mathematics, Spijker's lemma is a result in the theory of rational mappings of the Riemann sphere. It states that the image of a circle under a complex rational map with numerator and denominator having degree at most n has length at most 2nπ.
Applications
Spijker's lemma can be used to derive a sharp bound version of Kreiss matrix theorem.
See also
Buffon's needle
External links |
https://en.wikipedia.org/wiki/W-box | The W box is a deoxyribonucleic acid (DNA) cis-regulatory element sequence, (T)TGAC(C/T), which is recognized by the family of WRKY transcription factors.
Functionality and conservation of the W-box element across plant species has been shown by gel shift experiments, random binding site selection, yeast one-hybrid screens and co-transfection assays performed with many different WRKY proteins. In silico-based studies together with functional studies of plant promoters have identified clusters of W-boxes in stress-inducible promoters. The binding of WRKY proteins to W-boxes is a feature of both biotic and abiotic stress responses, together with other plant processes such as germination. It has also been shown that multiple W-boxes have a synergistic effect on transcription.
Almost all WRKY transcription factors bind preferentially to W-boxes, and since their discovery, this has raised the question as to how they show specificity for the promoters of their target genes. Ciolkowski et al. (2008) showed that although the W-box core is required, adjacent sequences also play a role in determining binding-site preference. Recent evidence suggests that the TGAC core is more degenerate, composed of a guanine adenine cytosine (GAC) core, and the upstream thymine and downstream pyrimidine flanking sequences help dictate recognition by specific WRKY factors. Basic residues of the WRKY protein domain also are believed to recognize the phosphate backbone of the cis-element.
Recently, Yamasaki et al. have determined the solution structure of the C-terminal WRKY domain of Arabidopsis WRKY4 in complex with the W-box DNA by NMR. They found that a four-stranded β-sheet enters the major groove of DNA in a structure they called the β-wedge, where the sheet is nearly perpendicular to the DNA helical ais. As predicted amino acids in the conserved WRKYGQK signature motif contact the W-box DNA.
External Links and Useful Resources
WRKY Transcription Factor Family at The Arabidopsis |
https://en.wikipedia.org/wiki/Rema%20Lapouse%20Award | Rema Lapouse Award is granted to an outstanding scientist in the area of psychiatric epidemiology in recognition of "significant contributions to the scientific understanding of the epidemiology and control of mental disorders. It is sponsored by the Mental Health, Epidemiology, and Applied Public Health Statistics Sections of the American Public Health Association. It was established in 1972 by the American physician Milton Terris in honor of his wife, Dr. Rema Lapouse, who was a founding member of the Mental Health Section.
Recipients
Other recipients
No other known recipients.
See also
List of medicine awards |
https://en.wikipedia.org/wiki/Lagrange%20bracket | Lagrange brackets are certain expressions closely related to Poisson brackets that were introduced by Joseph Louis Lagrange in 1808–1810 for the purposes of mathematical formulation of classical mechanics, but unlike the Poisson brackets, have fallen out of use.
Definition
Suppose that (q1, …, qn, p1, …, pn) is a system of canonical coordinates on a phase space. If each of them is expressed as a function of two variables, u and v, then the Lagrange bracket of u and v is defined by the formula
Properties
Lagrange brackets do not depend on the system of canonical coordinates (q, p). If (Q,P) = (Q1, …, Qn, P1, …, Pn) is another system of canonical coordinates, so that
is a canonical transformation, then the Lagrange bracket is an invariant of the transformation, in the sense that
Therefore, the subscripts indicating the canonical coordinates are often omitted.
If Ω is the symplectic form on the 2n-dimensional phase space W and u1,…,u2n form a system of coordinates on W, the symplectic form can be written as
where the matrix
::
represents the components of , viewed as a tensor, in the coordinates u. This matrix is the inverse of the matrix formed by the Poisson brackets
of the coordinates u.
As a corollary of the preceding properties, coordinates (Q1, ..., Qn, P1, …, Pn) on a phase space are canonical if and only if the Lagrange brackets between them have the form
See also
Lagrangian mechanics
Hamiltonian mechanics |
https://en.wikipedia.org/wiki/Right%20half-plane | In complex analysis, the (open) right half-plane is the set of all points in the complex plane whose real part is strictly positive, that is, the set . |
https://en.wikipedia.org/wiki/Permutohedron | In mathematics, the permutohedron (also spelled permutahedron) of order n is an (n − 1)-dimensional polytope embedded in an n-dimensional space. Its vertex coordinates (labels) are the permutations of the first n natural numbers. The edges identify the shortest possible paths (sets of transpositions) that connect two vertices (permutations). Two permutations connected by an edge differ in only two places (one transposition), and the numbers on these places are neighbors (differ in value by 1).
The image on the right shows the permutohedron of order 4, which is the truncated octahedron. Its vertices are the 24 permutations of (1, 2, 3, 4). Parallel edges have the same edge color. The 6 edge colors correspond to the 6 possible transpositions of 4 elements, i.e. they indicate in which two places the connected permutations differ. (E.g. red edges connect permutations that differ in the last two places.)
History
According to , permutohedra were first studied by . The name permutoèdre was coined by . They describe the word as barbaric, but easy to remember, and submit it to the criticism of their readers.
The alternative spelling permutahedron is sometimes also used. Permutohedra are sometimes called permutation polytopes, but this terminology is also used for the related Birkhoff polytope, defined as the convex hull of permutation matrices. More generally, uses that term for any polytope whose vertices have a bijection with the permutations of some set.
Vertices, edges, and facets
The permutohedron of order has vertices, each of which is adjacent to others.
The number of edges is , and their length is .
Two connected vertices differ by swapping two coordinates, whose values differ by 1. The pair of swapped places corresponds to the direction of the edge.
(In the example image the vertices and are connected by a blue edge and differ by swapping 2 and 3 on the first two places. The values 2 and 3 differ by 1. All blue edges correspond to swaps of coordinates |
https://en.wikipedia.org/wiki/Doubly%20linked%20face%20list | In applied mathematics, a doubly linked face list (DLFL) is an efficient data structure for storing 2-manifold mesh data. The structure stores linked lists for a 3D mesh's faces, edges, vertices, and corners. The structure guarantees the preservation of the manifold property. |
https://en.wikipedia.org/wiki/Barm | Barm, also called ale yeast, is the foam or scum formed on the top of a fermenting liquid, such as beer, wine, or feedstock for spirits or industrial ethanol distillation. It is used to leaven bread, or set up fermentation in a new batch of liquor. Barm, as a leaven, has also been made from ground millet combined with must out of wine-tubs and is sometimes used in English baking as a synonym for a natural leaven (sourdough). Various cultures derived from barm, usually Saccharomyces cerevisiae, became ancestral to most forms of brewer's yeast and baker's yeast currently on the market.
A barm cake is a soft, round, flattish bread roll from North West England, traditionally leavened with barm. In Ireland, barm is used in the traditional production of barmbrack, a fruited bread.
Emptins, a homemade product similar to barm and usually made from hops or potatoes and the dregs of cider or ale casks, was a common leavener for those living in rural areas far from a brewery, distillery, or bakery from which they could source barm or yeast.
See also
Kaiser roll
Sour mash
Yeast in winemaking |
https://en.wikipedia.org/wiki/Cocompact%20group%20action | In mathematics, an action of a group G on a topological space X is cocompact if the quotient space X/G is a compact space. If X is locally compact, then an equivalent condition is that there is a compact subset K of X such that the image of K under the action of G covers X. It is sometimes referred to as mpact, a tongue-in-cheek reference to dual notions where prefixing with "co-" twice would "cancel out". |
https://en.wikipedia.org/wiki/Quasi-isometry | In mathematics, a quasi-isometry is a function between two metric spaces that respects large-scale geometry of these spaces and ignores their small-scale details. Two metric spaces are quasi-isometric if there exists a quasi-isometry between them. The property of being quasi-isometric behaves like an equivalence relation on the class of metric spaces.
The concept of quasi-isometry is especially important in geometric group theory, following the work of Gromov.
Definition
Suppose that is a (not necessarily continuous) function from one metric space to a second metric space . Then is called a quasi-isometry from to if there exist constants , , and such that the following two properties both hold:
For every two points and in , the distance between their images is up to the additive constant within a factor of of their original distance. More formally:
Every point of is within the constant distance of an image point. More formally:
The two metric spaces and are called quasi-isometric if there exists a quasi-isometry from to .
A map is called a quasi-isometric embedding if it satisfies the first condition but not necessarily the second (i.e. it is coarsely Lipschitz but may fail to be coarsely surjective). In other words, if through the map, is quasi-isometric to a subspace of .
Two metric spaces M1 and M2 are said to be quasi-isometric, denoted , if there exists a quasi-isometry .
Examples
The map between the Euclidean plane and the plane with the Manhattan distance that sends every point to itself is a quasi-isometry: in it, distances are multiplied by a factor of at most . Note that there can be no isometry, since, for example, the points are of equal distance to each other in Manhattan distance, but in the Euclidean plane, there are no 4 points that are of equal distance to each other.
The map (both with the Euclidean metric) that sends every -tuple of integers to itself is a quasi-isometry: distances are preserved exactly, and every real t |
https://en.wikipedia.org/wiki/Glossary%20of%20plant%20morphology | This page provides a glossary of plant morphology. Botanists and other biologists who study plant morphology use a number of different terms to classify and identify plant organs and parts that can be observed using no more than a handheld magnifying lens. This page provides help in understanding the numerous other pages describing plants by their various taxa. The accompanying page—Plant morphology—provides an overview of the science of the external form of plants. There is also an alphabetical list: Glossary of botanical terms. In contrast, this page deals with botanical terms in a systematic manner, with some illustrations, and organized by plant anatomy and function in plant physiology.
This glossary primarily includes terms that deal with vascular plants (ferns, gymnosperms and angiosperms), particularly flowering plants (angiosperms). Non-vascular plants (bryophytes), with their different evolutionary background, tend to have separate terminology. Although plant morphology (the external form) is integrated with plant anatomy (the internal form), the former became the basis of the taxonomic description of plants that exists today, due to the few tools required to observe.
Many of these terms date back to the earliest herbalists and botanists, including Theophrastus. Thus, they usually have Greek or Latin roots. These terms have been modified and added to over the years, and different authorities may not always use them the same way.
This page has two parts: The first deals with general plant terms, and the second with specific plant structures or parts.
General plant terms
Abaxial – located on the side facing away from the axis.
Adaxial – located on the side facing towards the axis.
Dehiscent – opening at maturity
Gall – outgrowth on the surface caused by invasion by other lifeforms, such as parasites
Indehiscent – not opening at maturity
Reticulate – web-like or network-like
Striated – marked by a series of lines, grooves, or ridges
Tesselate – |
https://en.wikipedia.org/wiki/Pseudoelementary%20class | In logic, a pseudoelementary class is a class of structures derived from an elementary class (one definable in first-order logic) by omitting some of its sorts and relations. It is the mathematical logic counterpart of the notion in category theory of (the codomain of) a forgetful functor, and in physics of (hypothesized) hidden variable theories purporting to explain quantum mechanics. Elementary classes are (vacuously) pseudoelementary but the converse is not always true; nevertheless pseudoelementary classes share some of the properties of elementary classes such as being closed under ultraproducts.
Definition
A pseudoelementary class is a reduct of an elementary class. That is, it is obtained by omitting some of the sorts and relations of a (many-sorted) elementary class.
Examples
The theory with equality of sets under union and intersection, whose structures are of the form (W, ∪, ∩), can be understood naively as the pseudoelementary class formed from the two-sorted elementary class of structures of the form (A, W, ∪, ∩, ∈) where ∈ ⊆ A×W and ∪ and ∩ are binary operations (qua ternary relations) on W. The theory of the latter class is axiomatized by
∀X,Y∈W.∀a∈A.[ a ∈ X∪Y ⇔ a ∈ X ∨ a ∈ Y]
∀X,Y∈W.∀a∈A.[ a ∈ X∩Y ⇔ a ∈ X ∧ a ∈ Y]
∀X,Y∈W.[ (∀a∈A.[a ∈ X ⇔ a ∈ Y]) → X = Y]
In the intended interpretation A is a set of atoms a,b,..., W is a set of sets of atoms X,Y,... and ∈ is the membership relation between atoms and sets. The consequences of these axioms include all the laws of distributive lattices. Since the latter laws make no mention of atoms they remain meaningful for the structures obtained from the models of the above theory by omitting the sort A of atoms and the membership relation ∈. All distributive lattices are representable as sets of sets under union and intersection, whence this pseudoelementary class is in fact an elementary class, namely the variety of distributive lattices.
In this example both classes (respectively before and |
https://en.wikipedia.org/wiki/Reduct | In universal algebra and in model theory, a reduct of an algebraic structure is obtained by omitting some of the operations and relations of that structure. The opposite of "reduct" is "expansion".
Definition
Let A be an algebraic structure (in the sense of universal algebra) or a structure in the sense of model theory, organized as a set X together with an indexed family of operations and relations φi on that set, with index set I. Then the reduct of A defined by a subset J of I is the structure consisting of the set X and J-indexed family of operations and relations whose j-th operation or relation for j ∈ J is the j-th operation or relation of A. That is, this reduct is the structure A with the omission of those operations and relations φi for which i is not in J.
A structure A is an expansion of B just when B is a reduct of A. That is, reduct and expansion are mutual converses.
Examples
The monoid (Z, +, 0) of integers under addition is a reduct of the group (Z, +, −, 0) of integers under addition and negation, obtained by omitting negation. By contrast, the monoid (N, +, 0) of natural numbers under addition is not the reduct of any group.
Conversely the group (Z, +, −, 0) is the expansion of the monoid (Z, +, 0), expanding it with the operation of negation. |
https://en.wikipedia.org/wiki/CandyFab | The CandyFab is a method of producing physical objects out of a computer representation of the structure. It differs from some other 3D printing methods in the following aspects:
It is optimized for relatively large pieces using low to medium print resolution.
To reduce the hazards of working with large amounts of media, non-toxic materials (ideally, food-grade) are used. The prototype CandyFab 4000 unit uses granulated sugar as its print medium, giving rise to its name, but other materials with low melting temperatures and low toxicities are still under consideration.
It uses low cost parts and construction to make it easier for others to design or build their own, with plans available as open source.
Technology
The CandyFab uses a heat source mounted on a computer-controlled X-Y positioning head to fuse the surface of a granular bed of the print media. The only thing which comes into contact with the media is heated air, which is turned on and off by the software synchronously with the motion of the positioning head. Fabrication of the shape of the part being produced progresses in layers; after each complete pass, the bed is lowered and a fresh layer of granular media is applied on top. The unfused media serves to support overhangs and thin walls in the part being produced, reducing the need for auxiliary temporary supports for the workpiece. The movable bed is of a size suitable for producing finished parts several kilograms in weight.
The resolution of features produced correspond to a smallest volume element of 2.5 x 2.5 x 2.7 mm or less. Pieces produced from ordinary granular sugar have fairly good strength and feature an amber to brown surface color owing to caramelization of the sugar. Special attention has been paid to the selection of all materials coming into contact with the sugar bed or with the hot air stream to make it possible to fabricate food-grade pieces if desired.
There is an effort to encourage further work on improving the technology |
https://en.wikipedia.org/wiki/Reaction%20dynamics | Reaction dynamics is a field within physical chemistry, studying why chemical reactions occur, how to predict their behavior, and how to control them. It is closely related to chemical kinetics, but is concerned with individual chemical events on atomic length scales and over very brief time periods. It considers state-to-state kinetics between reactant and product molecules in specific quantum states, and how energy is distributed between translational, vibrational, rotational, and electronic modes.
Experimental methods of reaction dynamics probe the chemical physics associated with molecular collisions. They include crossed molecular beam and infrared chemiluminescence experiments, both recognized by the 1986 Nobel Prize in Chemistry awarded to Dudley Herschbach, Yuan T. Lee, and John C. Polanyi "for their contributions concerning the dynamics of chemical elementary processes", In the crossed beam method used by Herschbach and Lee, narrow beams of reactant molecules in selected quantum states are allowed to react in order to determine the reaction probability as a function of such variables as the translational, vibrational and rotational energy of the reactant molecules and their angle of approach. In contrast the method of Polanyi measures vibrational energy of the products by detecting the infrared chemiluminescence emitted by vibrationally excited molecules, in some cases for reactants in defined energy states.
Spectroscopic observation of reaction dynamics on the shortest time scales is known as femtochemistry, since the typical times studied are of the order of 1 femtosecond = 10−15 s. This subject has been recognized by the award of the 1999 Nobel Prize in Chemistry to Ahmed Zewail.
In addition, theoretical studies of reaction dynamics involve calculating the potential energy surface for a reaction as a function of nuclear positions, and then calculating the trajectory of a point on this surface representing the state of the system. A correction can be |
https://en.wikipedia.org/wiki/Lymph%20heart | A lymph heart is an organ which pumps lymph in lungfishes, amphibians, reptiles, and flightless birds back into the circulatory system. In some amphibian species, lymph hearts are in pairs, and may number as many as 200 in one animal the size of a worm, while newts and salamanders have as many as 16 to 23 pairs of lymph hearts.
Lymph hearts are thought to have evolved in Rhipidistia. Mammals have lost the lymph heart as a centralized organ, instead having the lymph vessel themselves contract to pump lymph.
and other amphibians
The lymphatic system of a frog consists of lymph, lymph vessels, lymph heart, lymph spaces and spleen.
Lymphatics and lymph
As lymph is a filtrate of blood, it closely resembles the plasma in its water content. Lymph also contains a small amount of metabolic waste and a much smaller amount of protein than that of blood. Lymph vessels carry the lymph and, in the frog, open into the four lymph hearts. These lymph hearts are located on the dorsal side of frog's body. The front pair is situated below the shoulder blades. The posterior pair is on either side of a long, rod-like bone called a urostyle, formed by the fusion of the last few vertebrae. The anterior pair opens into the subclavian vein and the posterior pair into the femoral vein. The pair near the third vertebra pumps lymph into the jugular vein. The other pair at the end of the vertebral column pump lymph into the iliac vein in the legs.
The position of mammalian jugular lymph sacs coincide with that of amphibian anterior lymph hearts.
Mechanism of the lymph hearts
The lymph hearts rhythmically and slowly pump to drive the lymph into the veins. It is possible to see the lymph hearts beat by looking on the dorsal surface on either side of the urostyle. In the toad, the normal lymph heart rate is about 50 beats per minute. Thus the lymph emerging out of blood ultimately merges into the blood. It returns the proteins back to blood.
Amphibian lymph hearts are made up from three tissu |
https://en.wikipedia.org/wiki/Application-specific%20instruction%20set%20processor | An application-specific instruction set processor (ASIP) is a component used in system on a chip design. The instruction set architecture of an ASIP is tailored to benefit a specific application. This specialization of the core provides a tradeoff between the flexibility of a general purpose central processing unit (CPU) and the performance of an application-specific integrated circuit (ASIC).
Some ASIPs have a configurable instruction set. Usually, these cores are divided into two parts: static logic which defines a minimum ISA (instruction-set architecture) and configurable logic which can be used to design new instructions. The configurable logic can be programmed either in the field in a similar fashion to a field-programmable gate array (FPGA) or during the chip synthesis. ASIPs have two ways of generating code: either through a retargetable code generator or through a retargetable compiler generator. The retargetable code generator uses the application, ISA, and Architecture Template to create the code generator for the object code. The retargetable compiler generator uses only the ISA and Architecture Template as the basis for creating the compiler. The application code will then be used by the compiler to create the object code.
ASIPs can be used as an alternative of hardware accelerators for baseband signal processing or video coding. Traditional hardware accelerators for these applications suffer from inflexibility. It is very difficult to reuse the hardware datapath with handwritten finite-state machines (FSM). The retargetable compilers of ASIPs help the designer to update the program and reuse the datapath. Typically, the ASIP design is more or less dependent on the tool flow because designing a processor from scratch can be very complicated. One approach is to describe the processor using a high level language and then to automatically generate the ASIP's software toolset.
Examples
RISC-V Instruction Set Architecture (ISA) provides minimum base ins |
https://en.wikipedia.org/wiki/God%20Created%20the%20Integers | God Created the Integers: The Mathematical Breakthroughs That Changed History is a 2005 anthology, edited by Stephen Hawking, of "excerpts from thirty-one of the most important works in the history of mathematics."
The title of the book is a reference to a quotation attributed to mathematician Leopold Kronecker, who once wrote that "God made the integers; all else is the work of man."
Content
The works are grouped by author and ordered chronologically. Each section is prefaced by notes on the mathematician's life and work. The anthology includes works by the following mathematicians:
Selections from the works of Euler, Bolyai, Lobachevsky and Galois, which are included in the second edition of the book (published in 2007), were not included in the first edition.
Editions |
https://en.wikipedia.org/wiki/Island%20of%20inversion | An island of inversion is a region of the chart of nuclides where isotopes have enhanced stability in a sea of mostly fleeting and unstable nuclei at the edge of the nuclear map. Each island contains isotopes with a non-standard ordering of single particle levels in the nuclear shell model. Such an area was first described in 1975 by French physicists carrying out spectroscopic mass measurements of exotic isotopes of lithium and sodium. Since then further studies have shown that five such regions exist within the known table of nuclides. These are centered at neutron-rich isotopes of five elements, namely 11Li, 20C, 31Na, 42Si, and 64Cr. Because there are five known islands of inversion, physicists have suggested renaming the phenomenon as an "archipelago of islands of shell breaking". Studies with the purpose of defining the edges of this region are still ongoing.
See also
Table of nuclides
Periodic table and Periodic table (extended)
Island of stability |
https://en.wikipedia.org/wiki/Alloenzyme | Alloenzymes (or also called allozymes) are variant forms of an enzyme which differ structurally but not functionally from other allozymes coded for by different alleles at the same locus. These are opposed to isozymes, which are enzymes that perform the same function, but which are coded by genes located at different loci.
Alloenzymes are common biological enzymes that exhibit high levels of functional evolutionary conservation throughout specific phyla and kingdoms. They are used by phylogeneticists as molecular markers to gauge evolutionary histories and relationships between different species. This can be done because allozymes do not have the same structure. They can be separated by capillary electrophoresis. However, some species are monomorphic for many of their allozymes which would make it difficult for phylogeneticists to assess the evolutionary histories of these species. In these instances, phylogeneticists would have to use another method to determine the evolutionary history of a species.
These enzymes generally perform very basic functions found commonly throughout all lifeforms, such as DNA polymerase, the enzyme that repairs and copies DNA. Significant changes in this enzyme reflect significant events in evolutionary history of organisms. As expected DNA polymerase shows relatively small differences in its amino acid sequence between phyla and even kingdoms.
The key to choosing which alloenzyme to use in a comparison between multiple species is to choose one that is as variable as possible while still being present in all the organisms. By comparing the amino acid sequence of the enzyme in the species, more amino acid similarities should be seen in species that are more closely related, and fewer between those that are more distantly related. The less well conserved the enzyme is, the more amino acid differences will be present in even closely related species. |
https://en.wikipedia.org/wiki/ExaGrid | ExaGrid Systems, Inc. is a tiered backup storage company headquartered in Marlborough, Massachusetts.
ExaGrid History
The company was founded in 2002 by Dave Therrien to solve the challenge of integrated primary storage with data backup. The company was funded by venture backed funding from Sigma Partners and Highland Capital Partners.
The name ExaGrid was derived from two words: “Exa” from exabytes of data and , “Grid,” an approach to grid computing scale-out storage which ExaGrid was based on.
In 2005, Bill Andrews joined the company as President and CEO. Bill currently remains in the position as of 2022. Under Bill’s leadership the company pivoted to backup storage with a tiered backup storage architecture.
In 2006, ExaGrid shipped its first backup storage appliances.
Over the years from 2006 to current, ExaGrid has added many features/functions such as zone level deduplication, adaptive deduplication, global deduplication, scale to 2.7PB full backups, 32 appliances in a single scale-out system, the ability to deduplicate Commvault data, the ability to support Veritas NetBackup Accelerator, support of the Veeam Data Mover and Veeam SOBR (scale-out backup repository) and many other features and functions.
ExaGrid appliances are used by companies to store their backup data. The appliances are used in IT data centers in commercial accounts, public education, local government, federal government and universities.
In 2007 ExaGrid raised capital from Lehman Brothers (now Tenaya Capital)
From 2007 to current, ExaGrid has opened offices in over 20 countries including: Argentina/Chile, Australia, Benelux, Brazil, Canada, Colombia, Czech Republic, France, Germany, Hong Kong, Iberia, Israel, Malaysia, Mexico, Nordics, Poland, Russia, Saudi, Singapore, South Africa, Turkey, United Kingdom, United States.
In 2009 ExaGrid raised capital from Investor Growth Capital (Investor AB)
In 2009, ExaGrid’s largest competitor, Data Domain was acquired by EMC, a leader storage |
https://en.wikipedia.org/wiki/Perm%20%28unit%29 | A perm is a unit of permeance or "water vapor transmission" given a certain differential in partial pressures on either side of a material or membrane.
Definitions
U.S. perm
The U.S. perm is defined as 1 grain of water vapor per hour, per square foot, per inch of mercury.
{|
|-
|1 U.S. perm
|= 0.659045 metric perms
|-
|||≈ 57.2135 ng·s−1·m−2·Pa−1
|}
Metric perm
The metric perm (not an SI unit) is defined as 1 gram of water vapor per day, per square meter, per millimeter of mercury.
{|
|-
|1 metric perm
|= 1.51735 US perms
|-
|||≈ 86.8127 ng·s−1·m−2·Pa−1
|}
Equivalent SI unit
The equivalent SI measure is the nanogram per second per square meter per pascal.
{|
|-
|1 ng·s−1·m−2·Pa−1
|≈ 0.0174784 US perms
|-
|||≈ 0.0115191 metric perms
|}
The base normal SI unit for permeance is the kilogram per second per square meter per pascal.
{|
|-
|1 kg·s−1·m−2·Pa−1
|≈ 1.74784x1010 US perms
|-
|||≈ 1.15191x1010 metric perms
|}
German Institute for Standardization unit
A variant of the metric perm is used in DIN Standard 53122, where permeance is also expressed in grams per square meter per day, but at a fixed, "standard" vapor-pressure difference of 17.918 mmHg. This unit is thus 17.918 times smaller than a metric perm, corresponding to about 0.084683 of a U.S. perm. |
https://en.wikipedia.org/wiki/Gisiro%20Maruyama | was a Japanese mathematician, noted for his contributions to the study of stochastic processes. The Euler–Maruyama method for the numerical solution of stochastic differential equations bears his name.
Maruyama was born in 1916 and graduated from Tohoku University, where he studied Fourier analysis and physics. He began his mathematical work with a paper on Fourier analysis in 1939.
He became interested in probability theory through the study of Norbert Wiener's work. He was appointed Assistant professor at the Kyushu University in 1941.
When Kiyosi Itô published his papers on stochastic differential equations in 1942, Maruyama immediately recognized the importance of this work and soon published a series of papers on stochastic differential equations and Markov processes.
Maruyama is known in particular for his 1955 study of the convergence properties of the finite-difference approximations for the numerical solution of stochastic differential equations, now known as the Euler–Maruyama method.
In harmonic analysis, he studied the ergodicity and mixing properties of stationary stochastic processes in terms of their spectral properties.
Maruyama also studied quasi-invariance properties of the Wiener measure, extending previous work by Cameron and Martin to diffusion processes. |
https://en.wikipedia.org/wiki/Feit%E2%80%93Thompson%20conjecture | In mathematics, the Feit–Thompson conjecture is a conjecture in number theory, suggested by . The conjecture states that there are no distinct prime numbers p and q such that
divides .
If the conjecture were true, it would greatly simplify the final chapter of the proof of the Feit–Thompson theorem that every finite group of odd order is solvable. A stronger conjecture that the two numbers are always coprime was disproved by with the counterexample p = 17 and q = 3313 with common factor 2pq + 1 = 112643.
It is known that the conjecture is true for q = 3 .
Informal probability arguments suggest that the "expected" number of counterexamples to the Feit–Thompson conjecture is very close to 0, suggesting that the Feit–Thompson conjecture is likely to be true.
See also
Cyclotomic polynomials
Goormaghtigh conjecture |
https://en.wikipedia.org/wiki/New%20Guinea%20Highlands | The New Guinea Highlands, also known as the Central Range or Central Cordillera, is a long chain of mountain ranges on the island of New Guinea, including the island's tallest peak, Puncak Jaya, Indonesia, , the highest mountain in Oceania. The range is home to many intermountain river valleys, many of which support thriving agricultural communities. The highlands run generally east-west the length of the island, which is divided politically between Indonesia in the west and Papua New Guinea in the east.
Geography
The Central Cordillera, some peaks of which are capped with ice, consists of (from east to west): the Central Highlands and Eastern Highlands of Papua New Guinea including the Owen Stanley Range in the southeast, whose highest peak is Mount Victoria at 4,038 metres (13,248 feet), the Albert Victor Mountains, the Sir Arthur Gordon Range, and the Bismarck Range, whose highest peak is Mount Wilhelm at 4,509 metres (14,793 feet), which is an extinct volcano with a crater lake; the Star Mountains on the Papua New Guinea–Indonesia border; and the Maoke Mountains or Snow Range in Indonesia, where perpetual snow was found by H. A. Lorentz in 1909 at 4,461 m (14,635 ft), and whose highest peaks are Puncak Jaya (Mt. Carstensz) at 4,884 m (16,024 feet), Puncak Mandala (Mt. Juliana) at 4,760 m (15,610 ft) and Puncak Trikora (Mt. Wilhelmina) at 4,750 m (15,580 ft).
Although some valleys such as the Waghi Valley in the Western Highlands, Papua New Guinea are heavily cultivated and support urban settlements most of the mountains have traditional tribal village communities in the grassy mountain valleys. The PNG highland provinces are: Eastern Highlands Province; Simbu Province (or Chimbu) whose centre is the small coffee-growing town of Kundiawa on the Wahgi River near Mount Wilhelm; Jiwaka Province; the Western Highlands; the rugged Enga Province the home of the Enga people with its administration in the very small town of Wabag on the Lai River, and containing the la |
https://en.wikipedia.org/wiki/Rackspace%20Technology | Rackspace Technology, Inc. is an American cloud computing company based in Windcrest, Texas, an inner suburb of San Antonio, Texas. The company also has offices in Blacksburg, Virginia, and Austin, Texas, as well as in Australia, Canada, United Kingdom, India, Dubai, Switzerland, the Netherlands, Germany, Singapore, Mexico, and Hong Kong. Its data centers are located in Amsterdam (Netherlands), Virginia (USA), Chicago (USA), Dallas (USA), London (UK), Frankfurt (Germany), Hong Kong (China), Kansas City (USA), New York City (USA), San Jose (USA), Shanghai (China), Queenstown (Singapore), and Sydney (Australia).
History
Although the founders began as application developers for end-users, they found that most companies did not either know how or want to host their applications. The founders wanted to focus on application development–not hosting–but they were unable to find an opportunity to outsource the hosting work. Eventually, the founders realized that it would be better to create a product to serve the hosting need and launch it as a company. Rackspace was launched in October 1998 with Richard Yoo as its CEO. Although most hosting companies focused on the technology end of hosting, Rackspace created its "Fanatical Support" offering to focus on service and support. On March 28, 2000, Rackspace received funding through lead investor Norwest Venture Partners and Sequoia Capital. George J. Still, Jr., Managing Partner at Norwest, subsequently joined the Board of Directors.
In 2008, Rackspace moved its headquarters from a building once occupied by Datapoint Corporation to the then-unoccupied Windsor Park Mall in Windcrest, Texas. Rackspace's Chairman, Graham Weston, owned the Montgomery Ward building in the mall until 2006 when it was sold to a developer. The city of Windcrest purchased south of the mall to create a residential and retail complex. The facility is located next to Roosevelt High School, and many Roosevelt students intern at Rackspace.
Fortunes "T |
https://en.wikipedia.org/wiki/Lung%20counter | A lung counter is a system consisting of a radiation detector, or detectors, and associated electronics that is used to measure radiation emitted from radioactive material that has been inhaled by a person and is sufficiently insoluble as to remain in the lung for weeks, months, or years. They are frequently used in occupations where workers may be exposed to radiation.
The lung counter may be placed on or near the body. These systems are also often housed in a low background counting chamber. Such a chamber may have thick walls made of low-background steel (~20-25 cm thick) and lined with lead, cadmium, tin, or polypropylene, with a final layer of copper. The purpose of the lead, cadmium (or tin), and copper is to reduce the background in the low energy region of a gamma spectrum (typically less than 200 keV).
Calibration
As a lung counter is primarily measuring radioactive materials that emit low energy gamma rays or x-rays, the phantom used to calibrate the system must be anthropometric. An example of such a phantom is the Lawrence Livermore National Laboratory Torso Phantom.
See also
Bomab |
https://en.wikipedia.org/wiki/Raining%20on%20Sunday | "Raining on Sunday" is a song co-written by country music artist Radney Foster and Darrell Brown. It was initially recorded on Foster's 1999 Arista Records album See What You Want to See. Foster's version of the song features a backing vocal from Darius Rucker of the rock band Hootie & the Blowfish.
Keith Urban covered the song for his 2002 album Golden Road. His rendition was released as the album's second single in January 2003.
Personnel
The following musicians perform on Urban's version:
Keith Urban — lead vocals, lead guitar
Tom Bukovac — rhythm guitar
Matt Chamberlain — drums
Eric Darken — percussion
Dann Huff — rhythm guitar
Scotty Huff — background vocals
Steve Nathan — keyboards
Jimmie Lee Sloas — bass guitar
Chart performance
Weekly charts
Year-end charts |
https://en.wikipedia.org/wiki/Stack%20buffer%20overflow | In software, a stack buffer overflow or stack buffer overrun occurs when a program writes to a memory address on the program's call stack outside of the intended data structure, which is usually a fixed-length buffer.
Stack buffer overflow bugs are caused when a program writes more data to a buffer located on the stack than what is actually allocated for that buffer. This almost always results in corruption of adjacent data on the stack, and in cases where the overflow was triggered by mistake, will often cause the program to crash or operate incorrectly. Stack buffer overflow is a type of the more general programming malfunction known as buffer overflow (or buffer overrun). Overfilling a buffer on the stack is more likely to derail program execution than overfilling a buffer on the heap because the stack contains the return addresses for all active function calls.
A stack buffer overflow can be caused deliberately as part of an attack known as stack smashing. If the affected program is running with special privileges, or accepts data from untrusted network hosts (e.g. a webserver) then the bug is a potential security vulnerability. If the stack buffer is filled with data supplied from an untrusted user then that user can corrupt the stack in such a way as to inject executable code into the running program and take control of the process. This is one of the oldest and more reliable methods for attackers to gain unauthorized access to a computer.
Exploiting stack buffer overflows
The canonical method for exploiting a stack-based buffer overflow is to overwrite the function return address with a pointer to attacker-controlled data (usually on the stack itself). This is illustrated with strcpy() in the following example:
#include <string.h>
void foo(char *bar)
{
char c[12];
strcpy(c, bar); // no bounds checking
}
int main(int argc, char **argv)
{
foo(argv[1]);
return 0;
}
This code takes an argument from the command line and copies it to a local sta |
https://en.wikipedia.org/wiki/Recreation%20ecology | Recreation ecology is the scientific study of environmental impacts resulting from recreational activity in protected natural areas. This field of study includes research and monitoring assessments of biophysical changes, analyses to identify causal and influential factors or support carrying capacity planning and management, and investigations of the efficacy of educational, regulatory, and site management actions designed to minimize recreation impacts. These ecological understandings of environmental impacts of outdoor recreation is critical to the management of recreation, ecotourism and visitation to natural spaces. Recreation ecology research has looked at the ecological impacts of hiking, camping and other outdoor recreation activities where the use and visitation is concentrated. As outdoor recreation shows increasing participation globally, questions and concerns are raised to which these can be managed sustainably with minimal impact to the environment.
History
While scientific studies of human trampling can be traced back to the late 1920s, a substantial body of recreation ecology literature did not accumulate until the 1970s when visitation to the outdoors soared, threatening the ecology of natural and semi-natural areas.
Recreation ecology as a field of study more officially began in the early 1960s and was addressed in depth by J. Alan Wagar in his work titled The Carrying Capacity of Wild Lands For Recreation, published in 1964 in the Society of American Foresters. In this publication, Wagar poses the question: do wild lands have carrying capacities for recreation use? Wagar addresses this question in terms of: (1) impacts of outdoor recreation on people (2) impacts of people in these outdoor spaces and (3) management procedures to address issues of overcrowding in wild lands for recreation.
In the past few decades, more than 1000 articles on recreation ecology have been published. As it is projected that the amount of time spent and the numbers o |
https://en.wikipedia.org/wiki/John%20Lawson%20Stoddard | John Lawson Stoddard (April 24, 1850 – June 5, 1931) was an American lecturer, author and photographer. He was a pioneer in the use of the stereopticon or magic lantern, adding photographs to his popular lectures about his travels around the world. Because he published books related to his travels, he is credited with developing the genre of travelogues.
In 1935, Daniel Crane Taylor wrote, "Stoddard's rise to fame was spectacular and unprecedented in the annals of American entertainers. No American lecturer, musician or actor has ever won so large a following in so short a time. From his second season, almost every lecture was sold out…He filled Daly's Theatre, one of the largest in New York, fifty times a season for ten years. …This would mean that Stoddard alone drew approximately one hundred thousand persons in New York each year."
Early life
Stoddard was born in Brookline, Massachusetts to a wealthy family. He was the son of Sarah Lothrop and Lewis Tappan Stoddard.
He was educated at private schools in Boston. He attended Williams College, graduating with an A.B. in 1871. At Williams he was a member of the fraternity Delta Psi (aka St. Anthony Hall). He studied theology at Yale Divinity School for two years, but left before he graduated.
Career
During the 1873–1874 academic year, Stoddard taught the classics at Boston Latin School. Between 1874 and 1876, Stoddard began traveling around the world, mostly to Constantinople, Egypt, Greece, and Palestine. After two years of traveling, he returned to teaching.
In 1879, Stoddard turned his travel experiences into a series of popular lectures delivered throughout North America. He pioneered the used of the stereopticon, also known as the magic lantern, which gave his lectures the "gimmick" of a visual component—the black and white photographs Stoddard took on his travels. His lectures were so popular that he soon became a household name. Stoddard also continued to travel and gather new content for his programs, g |
https://en.wikipedia.org/wiki/Bob%27s%20Red%20Mill | Bob's Red Mill is an American brand of whole-grain foods marketed by employee-owned American company Bob's Red Mill Natural Foods of Milwaukie, Oregon. The company was established in 1978 by Bob and Charlee Moore.
Bob's Red Mill Natural Foods is a producer of natural, with some certified organic choices, and gluten-free milled grain products, billing itself as the "nation's leading miller of diverse whole-grain foods." Its products are distributed throughout the United States, Canada, and a number of other locations such as the Caribbean. It produces over 400 products, primarily whole grains that are ground with quartz millstones which come from several 120-year-old mills, as well as baking mixes, beans, seeds, nuts, dried fruits, spices, and herbs. They are sold through seventy natural food and specialty grocery distributors in the United States and Canada, their online store, and the company's factory store and restaurant.
History
Moore's beginning as a business owner was in gasoline, not grains. In the 1950s, he briefly owned a filling station in Los Angeles. The smog in the city influenced Bob and his wife Charlee to sell the station, and move to Mammoth Lakes, a small resort town in the mountains just to the north of Los Angeles where he opened a second gas station. It failed after a year and the family was forced to move temporarily into a rental owned by their minister. Moore got a job working as a manager at a Firestone Tires store and got back on his feet. He bought a five-acre goat farm where he and Charlee raised their boys. He and his sons sold milk and eggs locally. Charlee began experimenting with baking whole grain bread.
Moore's drive for healthier foods started with his father's death of a heart attack at age 49, and his grandmother's healthy eating obsession.
He began experimenting with stone-ground flours in the mid-1960s after reading "John Goffe's Mill," a book about an archeologist who rebuilt a flour mill and went into business with no p |
https://en.wikipedia.org/wiki/Evolutionarily%20stable%20set | In game theory an evolutionarily stable set (ES set), sometimes referred to as evolutionary stable sets, is a set of strategies, which score equally against each other, each of which would be an evolutionarily stable strategy (ESS) were it not for the other members of the set. ES sets were first defined by Thomas (1985ab). |
https://en.wikipedia.org/wiki/Robinson%27s%20joint%20consistency%20theorem | Robinson's joint consistency theorem is an important theorem of mathematical logic. It is related to Craig interpolation and Beth definability.
The classical formulation of Robinson's joint consistency theorem is as follows:
Let and be first-order theories. If and are consistent and the intersection is complete (in the common language of and ), then the union is consistent. A theory is called complete if it decides every formula, meaning that for every sentence the theory contains the sentence or its negation but not both (that is, either or ).
Since the completeness assumption is quite hard to fulfill, there is a variant of the theorem:
Let and be first-order theories. If and are consistent and if there is no formula in the common language of and such that and then the union is consistent.
See also |
https://en.wikipedia.org/wiki/Anti%E2%80%93Saccharomyces%20cerevisiae%20antibody | Anti-Saccharomyces cerevisiae antibodies (ASCAs) are antibodies against antigens presented by the cell wall of the yeast Saccharomyces cerevisiae. These antibodies are directed against oligomannose sequences α-1,3 Man (α-1,2 Man α-1,2 Man)n (n = 1 or 2). ASCAs and perinuclear antineutrophil cytoplasmic antibodies (pANCAs) are the two most useful and often discriminating biomarkers for colitis. ASCA tends to recognize Crohn's disease more frequently, whereas pANCA tend to recognize ulcerative colitis.
ASCA antibodies react to a yeast protein with mannans, a 200-kDa glycoprotein.
Diseases
Diseases in which ASCA are found include the following:
Behçet's disease - The association with ASCA is not generally strong, but increased in patients with gastrointestinal symptoms.
Coeliac disease
Colitis
Ulcerative colitis-familial.
Microscopic colitis
Collagenous colitis
Crohn's disease
Intestinal yeast and ASCA positive
Intestinal yeast infections are seen in malabsorptive diseases like coeliac disease. In Crohn's disease and ulcerative colitis the presence of intestinal S. cerevisiae is rare, but the association with irritable bowel in coeliac disease remains unstudied.
Anti-mannans
Mannan (oligomannan) is a component of the yeast cell wall. Antibodies to yeast mannans are found at increased frequency in Crohn's disease and ASCA positive Crohn's tend to have lower low levels of mannan-binding lectin. Experimentally, antibodies to mannans from yeast can also crossreact to mannans of other types of yeast. Study of the sugars indicated that a mannotetraose (4-mer) was responsible for highest response. Studies of the 200 kDa glycoprotein antibodies found them commonly in healthy people, suggesting that the disease associated antibodies are to their carbohydrate moieties. Mannans from other yeast, for example candida albicans, have found to cross react with ASCA which suggests that other yeast may induce ASCA associated diseases. ASCA are serological markers of candid |
https://en.wikipedia.org/wiki/Electrostatic%20spray-assisted%20vapour%20deposition | Electrostatic spray-assisted vapour deposition (ESAVD) is a technique (developed by a company called IMPT) to deposit both thin and thick layers of a coating onto various substrates. In simple terms chemical precursors are sprayed across an electrostatic field towards a heated substrate, the chemicals undergo a controlled chemical reaction and are deposited on the substrate as the required coating. Electrostatic spraying techniques were developed in the 1950s for the spraying of ionised particles on to charged or heated substrates.
ESAVD (branded by IMPT as Layatec) is used for many applications in many markets including:
Thermal barrier coatings for jet engine turbine blades
Various thin layers in the manufacture of flat panel displays and photovoltaic panels, CIGS and CZTS-based thin-film solar cells.
Electronic components
Biomedical coatings
Glass coatings (such as self-cleaning)
Corrosion protection coatings
The process has advantages over other techniques for layer deposition (plasma, electron-beam) in that it does not require the use of any vacuum, electron beam or plasma so reduces the manufacturing costs. It also uses less power and raw materials making it more environmentally friendly. Also the use of the electrostatic field means that the process can coat complex 3D parts easily. |
https://en.wikipedia.org/wiki/Sinus%20%28botany%29 | In botany, a sinus is a space or indentation between two lobes or teeth, usually on a leaf. The term is also used in mycology. For example, one of the defining characteristics of North American species in the Morchella elata clade of morels is the presence of a sinus where the cap attaches to the stipe.
See also
Leaf shape
Sulcus (morphology) |
https://en.wikipedia.org/wiki/WikiScanner | WikiScanner (also known as Wikipedia Scanner) was a publicly searchable database that linked anonymous edits on Wikipedia to the organizations where those edits apparently originated. It did this by cross-referencing the edits with data on the owners of the associated block of IP addresses, though it did not investigate edits made under a username. It was created by Virgil Griffith and released on August 13, 2007.
In his "WikiScanner FAQ" Griffith stated his belief that WikiScanner could help make Wikipedia more reliable for controversial topics. He also indicated that he had never been employed by the Wikimedia Foundation and claimed his work on WikiScanner was "100% noncommercial".
On December 21, 2012, a research group from released an open-source clone of WikiScanner called WikiWatchdog.
By April 2013, attempts to run "WikiScanner Classic" from wikiscanner.virgil.gr returned to the WikiScanner home page, which identified itself as "WIKIWATCHER.COM"; and invoking "WikiScanner2 PreviewNew!" led to a "failure to load the page due to timeout" error.
In 2007, Virgil Griffith said he had to take WikiScanner down, as it was costing him "several thousand USD per month." He added below this on his WikiScanner webpage that as a grad student at Caltech in 2008 he developed with the aid of several undergraduates "a suite of Wikipedia-related tools known collectively as "WikiWatcher" which included: WikiScanner2 (Daniel), Wikiganda (Rishi), Poor Man's Checkuser, and BeaverScope," which he launched at the Hackers on Planet Earth (HOPE) conference that year. They used used "high-quality data" from Quova, and among them WikiWatcher "had some media successes, but when the summer was over there was no one to maintain the tools and they fell into disrepair."
Design
The tool's database contained 34 million entries on anonymous edits (those by users who were not logged in to Wikipedia) between February 7, 2002, and August 4, 2007. Griffith stated that the database was constr |
https://en.wikipedia.org/wiki/Metabolic%20imprinting | Metabolic imprinting refers to the long-term physiological and metabolic effects that an offspring's prenatal and postnatal environments have on them. Perinatal nutrition has been identified as a significant factor in determining an offspring's likelihood of it being predisposed to developing cardiovascular disease, obesity, and type 2 diabetes amongst other conditions.
During pregnancy, maternal glucose can cross the blood-placental barrier meaning maternal hyperglycaemia is associated with foetal hyperglycaemia. Despite maternal glucose being able to cross the blood-placental barrier, maternal insulin is not able and the foetus has to make its own. As a result, if a mother is hyperglycaemic the foetus is likely to be hyperinsulinaemic which leads to it having increased levels of growth and adiposity.
Maternal undernutrition
Maternal undernutrition has been linked with low birth weight and also a number of diseases, including Cardiovascular disease, stroke, hypertension and diabetes. When a foetus is in the womb and is not receiving sufficient nutrition, it can adapt to prioritize organ growth and increased metabolic efficiency to prepare itself for life in an energy deficient environment. Postnatally, when given the correct nutrition, babies exhibit ‘catch up growth’, potentially leading to obesity and other related complications. Studies based around restricting animals food intake throughout gestation have discovered that a reduction of just 30% of normal intake can cause low birth weight and increase sensitivity to high-fat-diet induced obesity.
In animal models, intrauterine undernutrition has been shown to be associated with hypertension later in life. This is because the formation of the kidneys is inhibited, which decreases filtration and flow rate through the nephrons, leading to increased blood pressure.
More extreme prenatal conditions such as famine have been shown to have effects on the neurodevelopment of a foetus. After the Dutch Famine of the |
https://en.wikipedia.org/wiki/Research%20Computing%20Services | Research Computing Services (separated in August 2007 from the former Manchester Computing at the University of Manchester), provides the focus for the University of Manchester's activities in supercomputing or high-performance computing, grid computing or e-science and computational science. Research Computing Services activities include services, training and research & development.
Supercomputers
The University of Manchester has been home to many supercomputers, starting from the 1948 Manchester Baby - the world's first stored program computer. Others have included CDC7600 (1972, and a second in 1977), a CDC Cyber 205, VP1200, VPX and 240/10. The CSAR service (see below) supercomputers included a 576 PE Cray T3E-1200E (1998, upgraded to 816PE in 2000), and SGI Origin 3000 (2001) and Altix (2003) systems. More recently some large clusters (e.g., the 200 processor Dell EM64T cluster) have been installed.
National Computing Services
Research Computing Services and its predecessors (Manchester Computing etc.) have been providing (high performance) computing services nationally in the UK since the 1970s. Manchester Computing operated the UK's 1998-2006 national supercomputer service CSAR with SGI and CSC Ltd. It currently operates other national computer services in the UK, including the Access Grid Support Centre (AGSC) and, as part of consortia, the UK National Grid Service (NGS) and North West Grid.
Research Centres
Research Computing Services is a part of several research centres including E-Science North West (ESNW), and the UK's National Centre for e-Social Science (NCeSS). |
https://en.wikipedia.org/wiki/Force10 | Dell Force10 (formerly nCore Networks, Force10 Networks), was a United States company that developed and marketed 10 Gigabit and 40 Gigabit Ethernet switches for computer networking to corporate, educational, and governmental customers. It had offices in North America, Europe, and the Asia Pacific region.
In August 2011, Dell completed the acquisition of Force10 and changed the name to Dell Force10.
In mid 2013, the Force10 designation was dropped from the products in favor of the data center networking line of the Dell Networking brand, and some of the other product lines were sold.
History
Founding
The company was founded by PK Dubey, Naresh Nigam and Som Sikdar. It was named by founder Som Sikdar, an avid sailor, after Beaufort Force 10 (Storm, Whole gale) on the Beaufort scale for wind speeds, indicating a storm with high speed winds, and matched their focus on 10 Gigabit Ethernet switching and routing products.
Acquisition
In January 2009, Force10 was acquired by Turin Networks (Founded by Philip Yim), which had previously purchased Carrier Access Corporation and White Rock Networks. Carrier Access Corporation itself had previously purchased Mangrove Systems and White Rock Networks had previously purchased Seranoa Networks.
On July 20, 2011 Dell announced it intended to fully acquire Force10 for an undisclosed amount. With the acquisition, Dell offered products for the data center where Dell focuses on the Ethernet switches. Dell Force10 continued to offer their non-Ethernet backhaul and metro-access platforms as well.
Telmar Network Technology of Plano, Texas, announced the acquisition of the Force10 Turin transport product lines from Dell in May, 2013, and has resumed support and development of the Traverse, TraverseEdge, TransAccess, TransNav, MasterSeries, Adit, Wide Bank, and Broadmore products. Telmar Network Technology, Inc. is a wholly owned subsidiary of Jabil Circuit, Inc. of St. Petersburg, FL.
iQor of St. Petersburg, FL, announced, in Dece |
https://en.wikipedia.org/wiki/VRBD%20agar | VRDBA, VRDB Agar, or Violet Red Bile Dextrose Agar, is a microbiological growth medium. It can be used in agar plates to monitor or assess bacterial growth in the laboratory, particularly the growth of Enterobacteriaceae, for which it is selective. |
https://en.wikipedia.org/wiki/Plate%20count%20agar | Plate count agar (PCA), also called standard methods agar (SMA), is a microbiological growth medium commonly used to assess or to monitor "total" or viable bacterial growth of a sample. PCA is not a selective medium.
The total number of living aerobic bacteria can be determined using a plate count agar (PCA) which is a substrate for bacteria to grow on. The medium contains casein which provides nitrogen, carbon, amino acids, vitamins and minerals to aid in the growth of the organism. Yeast extract is the source for vitamins, particularly of B-group. Glucose is the fermentable carbohydrate and agar is the solidifying agent. This is a non-selective medium and the bacteria is counted as colony forming units per gram (CFU/g) in solid samples and (CFU/ml) in liquid samples.
Pour Plate Technique
The pour plate technique is the typical technique used to prepare PCAs. Here, the inoculum is added to the molten agar before pouring the plate. The molten agar is cooled to about 45 degrees Celsius and is poured using a sterile method into a petri dish containing a specific diluted sample. From here, the plates are rotated to ensure the samples are uniformly mixing with the agar. Incubation of the plates is the next step and is carried out for about 3 days at 20 to 30 degrees Celsius.
Composition of Plate Count Agar
Ingredients Gms/L
Enzymatic Digest of Casein/tryptone 5.0
Yeast Extract 2.5
Glucose 1.0
Agar 15.0
Benefits to using PCA:
- Easy to perform
- There is a larger sample volume than the surface spread method allowing for detection of lower microbiological concentrations
- The agar surface does not have to be pre-dried
- The number of microbes/ mL in a specimen can be determined
- You do not need previously prepared plates |
https://en.wikipedia.org/wiki/Digital%20differential%20analyzer%20%28graphics%20algorithm%29 | In computer graphics, a digital differential analyzer (DDA) is hardware or software used for interpolation of variables over an interval between start and end point. DDAs are used for rasterization of lines, triangles and polygons. They can be extended to non linear functions, such as perspective correct texture mapping, quadratic curves, and traversing voxels.
In its simplest implementation for linear cases such as lines, the DDA algorithm interpolates values in interval by computing for each xi the equations xi = xi−1 + 1, yi = yi−1 + m, where m is the slope of the line. This slope can be expressed in DDA as follows:
In fact any two consecutive points lying on this line segment should satisfy the equation.
Performance
The DDA method can be implemented using floating-point or integer arithmetic. The native floating-point implementation requires one addition and one rounding operation per interpolated value (e.g. coordinate x, y, depth, color component etc.) and output result. This process is only efficient when an FPU with fast add and rounding operation will be available.
The fixed-point integer operation requires two additions per output cycle, and in case of fractional part overflow, one additional increment and subtraction. The probability of fractional part overflows is proportional to the ratio m of the interpolated start/end values.
DDAs are well suited for hardware implementation and can be pipelined for maximized throughput.
Algorithm
A linear DDA starts by calculating the smaller of dy or dx for a unit increment of the other. A line is then sampled at unit intervals in one coordinate and corresponding integer values nearest the line path are determined for the other coordinate.
Considering a line with positive slope, if the slope is less than or equal to 1, we sample at unit x intervals (dx=1) and compute successive y values as
Subscript k takes integer values starting from 0, for the 1st point and increases by 1 until endpoint is reache |
https://en.wikipedia.org/wiki/Triclocarban | Triclocarban (sometimes abbreviated as TCC) is an antibacterial chemical once common in, but now phased out of, personal care products like soaps and lotions. It was originally developed for the medical field. Although the mode of action is unknown, TCC can be effective in fighting infections by targeting the growth of bacteria such as Staphylococcus aureus. Additional research seeks to understand its potential for causing antibacterial resistance and its effects on organismal and environmental health.
Usage
Triclocarban has been used as an antimicrobial and antifungal compound since the 1960s. It was commonly found in personal care products as an antimicrobial in soaps, lotions, deodorants, toothpaste, and plastic. about 80% of all antimicrobial bar soap sold in the United States contained triclocarban. In 2011 United States consumers were spending nearly 1 billion dollars annually on products containing triclocarban and triclosan.
In December 2013, the Food and Drug Administration (FDA) required all companies to prove within the next year, that triclocarban is not harmful to consumers. Companies like Johnson & Johnson, Procter & Gamble, Colgate-Palmolive, and Avon began phasing out antibacterial ingredients due to health concerns.
By 2016 usage of triclocarban in soaps had declined to 40%, and that September the FDA banned triclocarban, triclosan and 17 other common antibacterial chemicals by September 2017, for their failure to be proven safe, or more effective than plain soap and water.
Chemical structure and properties
Triclocarban, 3-(4-chlorophenyl)-1-(3,4-dichlorophenyl)urea, is a white powder that is insoluble in water. While triclocarban has two chlorinated phenyl rings, it is structurally similar to carbanilide compounds often found in pesticides (such as diuron) and some drugs. Chlorination of ring structures is often associated with hydrophobicity, persistence in the environment, and bioaccumulation in fatty tissues of living organisms. For this re |
https://en.wikipedia.org/wiki/Apeirotope | In geometry, an apeirotope or infinite polytope is a generalized polytope which has infinitely many facets.
Definition
Abstract apeirotope
An abstract n-polytope is a partially ordered set P (whose elements are called faces) such that P contains a least face and a greatest face, each maximal totally ordered subset (called a flag) contains exactly n + 2 faces, P is strongly connected, and there are exactly two faces that lie strictly between a and b are two faces whose ranks differ by two. An abstract polytope is called an abstract apeirotope if it has infinitely many faces.
An abstract polytope is called regular if its automorphism group Γ(P) acts transitively on all of the flags of P.
Classification
There are two main geometric classes of apeirotope:
honeycombs in n dimensions, which completely fill an n-dimensional space.
skew apeirotopes, comprising an n-dimensional manifold in a higher space
Honeycombs
In general, a honeycomb in n dimensions is an infinite example of a polytope in n + 1 dimensions.
Tilings of the plane and close-packed space-fillings of polyhedra are examples of honeycombs in two and three dimensions respectively.
A line divided into infinitely many finite segments is an example of an apeirogon.
Skew apeirotopes
Skew apeirogons
A skew apeirogon in two dimensions forms a zig-zag line in the plane. If the zig-zag is even and symmetrical, then the apeirogon is regular.
Skew apeirogons can be constructed in any number of dimensions. In three dimensions, a regular skew apeirogon traces out a helical spiral and may be either left- or right-handed.
Infinite skew polyhedra
There are three regular skew apeirohedra, which look rather like polyhedral sponges:
6 squares around each vertex, Coxeter symbol {4,6|4}
4 hexagons around each vertex, Coxeter symbol {6,4|4}
6 hexagons around each vertex, Coxeter symbol {6,6|3}
There are thirty regular apeirohedra in Euclidean space. These include those listed above, as well as (in the plane) pol |
https://en.wikipedia.org/wiki/Permissive%20dialing | In North America, permissive dialing is the ability to make phone calls in an area subject to a newly introduced area code by using both the new and preexisting dialing methods.
When an area is given a new area code under a split plan, the area's previous area code would no longer be valid for calls in the area, so calls to numbers using the old area code will not work. To alleviate misdialing frustration, the local routing can be set up such that both the old and new area codes will work for the same telephone exchange. During this period, the local numbering authority must not reassign the area's existing exchanges to the remaining area of the old area code, nor vice versa. At the end of the permissive dialing period, the old area code is no longer valid for numbers in the affected area.
Under an overlay plan, permissive dialing refers to the ability to continue to connect calls via 7-digit dialing while also making 10-digit dialing valid. Again, the affected area must not introduce any new ambiguous telephone exchanges. At the end of the period, 10-digit dialing becomes mandatory. |
https://en.wikipedia.org/wiki/Town%20sign | A town sign or city limit sign is a road sign placed at the side of the road or street at the boundary of the territory of a city, town, or village. Town signs may be placed for reading both by drivers entering the town and, in a different format, by those exiting it. Signs give the name of the town in the local official languages, and sometimes in other languages. In some countries, town signs are also an essential part of the traffic law, for example by defining (explicitly or implicitly) the speed limit within the town's territory. In some countries, such as Germany and Austria, signs aren't placed at the exact boundary of a city or town, but rather at the location where the settlement's built-up area (continuous string of buildings/development) begins and ends.
In the UK town sign may refer to a prominent, decorative, often carved sign, commemorating the values and history of the town. Synonymous with Village sign; Commonly in the centre of the town.
In much of the United States, there is a similar county sign at the boundary between each county (or independent city not part of one), indicating the county being entered and often the one being left. Even if not done within a given U.S. state, there is also nearly always a welcome sign at the state line on every major highway, and most any other road. The welcome signs on Interstate highways are usually very large and have graphics, and may have an attached text-only sign directing motorists to the welcome center at a rest area. On smaller roads, they are usually more similar to town signs, showing the state and county, often with other signs indicating speed limit, a state law (such as "burn headlights during rain"), and/or a change in time zone.
Gallery
Street furniture
Traffic signs
Signage
Road safety |
https://en.wikipedia.org/wiki/Unary%20language | In computational complexity theory, a unary language or tally language is a formal language (a set of strings) where all strings have the form 1k, where "1" can be any fixed symbol. For example, the language {1, 111, 1111} is unary, as is the language {1k | k is prime}. The complexity class of all such languages is sometimes called TALLY.
The name "unary" comes from the fact that a unary language is the encoding of a set of natural numbers in the unary numeral system. Since the universe of strings over any finite alphabet is a countable set, every language can be mapped to a unique set A of natural numbers; thus, every language has a unary version {1k | k in A}. Conversely, every unary language has a more compact binary version, the set of binary encodings of natural numbers k such that 1k is in the language.
Since complexity is usually measured in terms of the length of the input string, the unary version of a language can be "easier" than the original language. For example, if a language can be recognized in O(2n) time, its unary version can be recognized in O(n) time, because n has become exponentially larger. More generally, if a language can be recognized in O(f(n)) time and O(g(n)) space, its unary version can be recognized in O(n + f(log n)) time and O(g(log n)) space (we require O(n) time just to read the input string). However, if membership in a language is undecidable, then membership in its unary version is also undecidable.
Relationships to other complexity classes
TALLY is contained in P/poly—the class of languages that can be recognized in polynomial time given an advice function that depends only on the input length. In this case, the required advice function is very simple—it returns a single bit for each input length k specifying whether 1k is in the language or not.
A unary language is necessarily a sparse language, since for each n it contains at most one value of length n and at most n values of length at most n, but not all sparse languag |
https://en.wikipedia.org/wiki/Sparse%20language | In computational complexity theory, a sparse language is a formal language (a set of strings) such that the complexity function, counting the number of strings of length n in the language, is bounded by a polynomial function of n. They are used primarily in the study of the relationship of the complexity class NP with other classes. The complexity class of all sparse languages is called SPARSE.
Sparse languages are called sparse because there are a total of 2n strings of length n, and if a language only contains polynomially many of these, then the proportion of strings of length n that it contains rapidly goes to zero as n grows. All unary languages are sparse. An example of a nontrivial sparse language is the set of binary strings containing exactly k 1 bits for some fixed k; for each n, there are only strings in the language, which is bounded by nk.
Relationships to other complexity classes
SPARSE contains TALLY, the class of unary languages, since these have at most one string of any one length. Although not all languages in P/poly are sparse, there is a polynomial-time Turing reduction from any language in P/poly to a sparse language.
Fortune showed in 1979 that if any sparse language is co-NP-complete, then P = NP;
Mahaney used this to show in 1982 that if any sparse language is NP-complete, then P = NP (this is Mahaney's theorem).
A simpler proof of this based on left-sets was given by Ogihara and Watanabe in 1991.
Mahaney's argument does not actually require the sparse language to be in NP (because the existence of an NP-hard sparse set implies the existence of an NP-complete sparse set), so there is a sparse NP-hard set if and only if P = NP.
Further, E ≠ NE if and only if there exist sparse languages in NP that are not in P.
There is a Turing reduction (as opposed to the Karp reduction from Mahaney's theorem) from an NP-complete language to a sparse language if and only if .
In 1999, Jin-Yi Cai and D. Sivakumar, building on work by Ogihara, showed tha |
https://en.wikipedia.org/wiki/State%20quality%20mark%20of%20the%20USSR | The State quality mark of the USSR (, transliteration ) was the official Soviet mark for the certification of quality established in 1967.
Symbol
The sign was a pentagonal shield with a rotated letter K (from Russian word – quality) stylized as scales below the Cyrillic abbreviation for USSR (, ).
History
It was used to mark consumer, production, and technical goods to certify that they met quality standards and, in general, to increase the effectiveness of the production system in the USSR.
The goods themselves or their packaging were marked, as was the accompanying documentation, labels or tags. Rules of its use were defined by GOST, an acronym for "state standard" (), section 1.9-67 (April 7, 1967).
The right to use the sign was leased to the enterprises for 2–3 years based on the examination of the goods by the State Attestation Commission (, ) that should certify that the goods are of the "higher quality category". That is:
their quality "meets or exceeds the quality of the best international analogs",
parameters of quality are stable,
goods fully satisfy Soviet state standards,
goods are compatible with international standards,
production of goods is economically effective, and
they satisfy the demands of the state economy and the population.
Obtaining the sign allowed the enterprises to increase the state controlled price for the goods by ten percent. When the sign was introduced it indeed suggested high quality of the goods but after some time a lot of Soviet-made goods were certified for the sign while their quality often remained below expectations of customers.
After dissolution of the Soviet Union, the Russian government introduced its own sign for certification of quality, known as the Rostest mark (or R mark).
See also
Certification mark
State Emblem of the Soviet Union
Rostest – organization responsible for the newer Rostest mark |
https://en.wikipedia.org/wiki/Aggregate%20Level%20Simulation%20Protocol | The Aggregate Level Simulation Protocol (ALSP) is a protocol and supporting software that enables simulations to interoperate with one another. Replaced by the High Level Architecture (simulation) (HLA), it was used by the US military to link analytic and training simulations.
ALSP consists of:
ALSP Infrastructure Software (AIS) that provides distributed runtime simulation support and management;
A reusable ALSP Interface consisting of generic data exchange message protocols; and
Participating simulations adapted for use with ALSP.
History
In 1990, the Defense Advanced Research Projects Agency (DARPA) employed The MITRE Corporation to study the application of distributed interactive simulation principles employed in SIMNET to aggregate-level constructive training simulations. Based on prototype efforts, a community-based experiment was conducted in 1991 to extend SIMNET to link the US Army's Corps Battle Simulation (CBS) and the US Air Force's Air Warfare Simulation (AWSIM). The success of the prototype and users' recognition of the value of this technology to the training community led to development of production software. The first ALSP confederation, providing air-ground interactions between CBS and AWSIM, supported three major exercises in 1992.
By 1995, ALSP had transitioned to a multi-Service program with simulations representing the US Army (CBS), the US Air Force (AWSIM), the US Navy (RESA), the US Marine Corps (MTWS), electronic warfare (JECEWSI), logistics (CSSTSS), and intelligence (TACSIM). The program had also transitioned from DARPA's research and development emphasis to mainstream management by the US Army's Program Executive Office for Simulation, Training, and Instrumentation (PEO STRI)
Contributions
ALSP developed and demonstrated key aspects of distributed simulation, many of which were applied in the development of HLA.
No central node so that simulations can join and depart from the confederation at will
Geographic distribution wher |
https://en.wikipedia.org/wiki/Cooling%20flow | A cooling flow occurs when the intracluster medium (ICM) in the centres of galaxy clusters should be rapidly cooling at the rate of tens to thousands of solar masses per year. This should happen as the ICM (a plasma) is quickly losing its energy by the emission of X-rays. The X-ray brightness of the ICM is proportional to the square of its density, which rises steeply towards the centres of many clusters. Also the temperature falls to typically a third or a half of the temperature in the outskirts of the cluster. The typical [predicted] timescale for the ICM to cool is relatively short, less than a billion years. As material in the centre of the cluster cools out, the pressure of the overlying ICM should cause more material to flow inwards (the cooling flow).
In a steady state, the rate of mass deposition, i.e. the rate at which the plasma cools, is given by
where L is the bolometric (i.e. over the entire spectrum) luminosity of the cooling region, T is its temperature, k is the Boltzmann constant and μm is the mean molecular mass.
Cooling flow problem
It is currently thought that the very large amounts of expected cooling are in reality much smaller, as there is little evidence for cool X-ray emitting gas in many of these systems. This is the cooling flow problem. Theories for why there is little evidence of cooling include
Heating by the central Active galactic nucleus (AGN) in clusters, possibly via sound waves (seen in the Perseus and Virgo clusters)
Thermal conduction of heat from the outer parts of clusters
Cosmic ray heating
Hiding cool gas by absorbing material
Mixing of cool gas with hotter material
Heating by AGN is the most popular explanation, as they emit a lot of energy over their lifetimes, and some of the alternatives listed have theoretical problems.
See also
List of plasma physics articles |
https://en.wikipedia.org/wiki/Doubly%20ionized%20oxygen | In astronomy and atomic physics, doubly ionized oxygen is the ion O2+ (O III in spectroscopic notation). Its emission forbidden lines in the visible spectrum fall primarily at the wavelength 500.7 nm, and secondarily at 495.9 nm. Before spectra of oxygen ions became known, these lines once led to a spurious identification of the substance as a new chemical element. Concentrated levels of O III are found in diffuse and planetary nebulae. Consequently, narrow band-pass filters that isolate the 500.7 nm and 495.9 nm wavelengths of light, that correspond to green-turquoise-cyan spectral colors, are useful in observing these objects, causing them to appear at higher contrast against the filtered and consequently blacker background of space (and possibly light-polluted terrestrial atmosphere) where the frequencies of [O III] are much less pronounced.
These emission lines were first discovered in the spectra of planetary nebulae in the 1860s. At that time, they were thought to be due to a new element which was named nebulium. In 1927, Ira Sprague Bowen published the current explanation identifying their source as doubly ionized oxygen.
Other transitions include the forbidden 88.4 μm and 51.8 μm transitions in the far infrared region.
Permitted lines of O III lie in the middle ultraviolet band and are hence inaccessible to terrestrial astronomy.
See also
H II region
Emission nebula |
https://en.wikipedia.org/wiki/Banach%E2%80%93Stone%20theorem | In mathematics, the Banach–Stone theorem is a classical result in the theory of continuous functions on topological spaces, named after the mathematicians Stefan Banach and Marshall Stone.
In brief, the Banach–Stone theorem allows one to recover a compact Hausdorff space X from the Banach space structure of the space C(X) of continuous real- or complex-valued functions on X. If one is allowed to invoke the algebra structure of C(X) this is easy – we can identify X with the spectrum of C(X), the set of algebra homomorphisms into the scalar field, equipped with the weak*-topology inherited from the dual space C(X)*. The Banach-Stone theorem avoids reference to multiplicative structure by recovering X from the extreme points of the unit ball of C(X)*.
Statement
For a compact Hausdorff space X, let C(X) denote the Banach space of continuous real- or complex-valued functions on X, equipped with the supremum norm ‖·‖∞.
Given compact Hausdorff spaces X and Y, suppose T : C(X) → C(Y) is a surjective linear isometry. Then there exists a homeomorphism φ : Y → X and a function g ∈ C(Y) with
such that
The case where X and Y are compact metric spaces is due to Banach, while the extension to compact Hausdorff spaces is due to Stone. In fact, they both prove a slight generalization—they do not assume that T is linear, only that it is an isometry in the sense of metric spaces, and use the Mazur–Ulam theorem to show that T is affine, and so is a linear isometry.
Generalizations
The Banach–Stone theorem has some generalizations for vector-valued continuous functions on compact, Hausdorff topological spaces. For example, if E is a Banach space with trivial centralizer and X and Y are compact, then every linear isometry of C(X; E) onto C(Y; E) is a strong Banach–Stone map.
A similar technique has also been used to recover a space X from the extreme points of the duals of some other spaces of functions on X.
The noncommutative analog of the Banach-Stone theorem is the folklore |
https://en.wikipedia.org/wiki/Multipliers%20and%20centralizers%20%28Banach%20spaces%29 | In mathematics, multipliers and centralizers are algebraic objects in the study of Banach spaces. They are used, for example, in generalizations of the Banach–Stone theorem.
Definitions
Let (X, ‖·‖) be a Banach space over a field K (either the real or complex numbers), and let Ext(X) be the set of extreme points of the closed unit ball of the continuous dual space X∗.
A continuous linear operator T : X → X is said to be a multiplier if every point p in Ext(X) is an eigenvector for the adjoint operator T∗ : X∗ → X∗. That is, there exists a function aT : Ext(X) → K such that
making the eigenvalue corresponding to p. Given two multipliers S and T on X, S is said to be an adjoint for T if
i.e. aS agrees with aT in the real case, and with the complex conjugate of aT in the complex case.
The centralizer (or commutant) of X, denoted Z(X), is the set of all multipliers on X for which an adjoint exists.
Properties
The multiplier adjoint of a multiplier T, if it exists, is unique; the unique adjoint of T is denoted T∗.
If the field K is the real numbers, then every multiplier on X lies in the centralizer of X.
See also
Centralizer and normalizer |
https://en.wikipedia.org/wiki/Radioecology | Radioecology is the branch of ecology concerning the presence of radioactivity in Earth’s ecosystems. Investigations in radioecology include field sampling, experimental field and laboratory procedures, and the development of environmentally predictive simulation models in an attempt to understand the migration methods of radioactive material throughout the environment.
The practice consists of techniques from the general sciences of physics, chemistry, mathematics, biology, and ecology, coupled with applications in radiation protection. Radioecological studies provide the necessary data for dose estimation and risk assessment regarding radioactive pollution and its effects on human and environmental health.
Radioecologists detect and evaluate the effects of ionizing radiation and radionuclides on ecosystems, and then assess their risks and dangers. Interest and studies in the area of radioecology significantly increased in order to ascertain and manage the risks involved as a result of the Chernobyl disaster. Radioecology arose in line with increasing nuclear activities, particularly following the Second World War in response to nuclear atomic weapons testing and the use of nuclear reactors to produce electricity.
History
Artificial radioactive affliction to Earth’s environment began with nuclear weapon testing during World War II, but did not become a prominent topic of public discussion until the 1980s. The Journal of Environmental Radioactivity (JER) was the first collection of literature on the subject, and its inception was not until 1984. As demand for construction of nuclear power plants increased, it became necessary for humankind to understand how radioactive material interacts with various ecosystems in order to prevent or minimize potential damage. The aftermath of Chernobyl was the first major employment of radioecological techniques to combat radioactive pollution from a nuclear power plant.
Collection of radioecological data from the Chernobyl |
https://en.wikipedia.org/wiki/Psarolepis | Psarolepis (; psārolepis, from Greek ψαρός 'speckled' and λεπίς 'scale') is a genus of extinct bony fish which lived around 397 to 418 million years ago (Pridoli to Lochkovian stages). Fossils of Psarolepis have been found mainly in South China and described by paleontologist Xiaobo Yu in 1998. It is not known certainly in which group Psarolepis belongs, but paleontologists agree that it probably is a basal genus and seems to be close to the common ancestor of lobe-finned and ray-finned fishes. In 2001, paleontologist John A. Long compared Psarolepis with onychodontiform fishes and refer to their relationships.
Description
Psarolepis had a pair of 'parasymphysical tooth whorls', teeth which extend up at the front of the lower jaw. The head was made of several thick dermal plates and covered with deep pock-marks and large pores. Another trait is a large pectoral spine, just in front of the pectoral fin, extending back from the shoulder girdle, and a dorsal spine located in front of a median fin behind the head, which gives the fish a shark-like form.
The pock-marked head of Psarolepis was made of plates containing a layer of porcelain-like cosmine. Because the cosmine layer obscures the suture lines of the skull, it is difficult to study the exact bone structure. The snout was strangely humped and the nostrils were located above the eyes, which were just above the upper jaw.
The most spectacular findings were the fin spines. Two are known: one extending back from the shoulder girdle and another which is associated with the dorsal fin. These fin spines are found only in primitive jawed fishes and are apparently absent from the most primitive sharks, but present in abundance in more derived forms.
Psarolepis had teeth at the very front of the snout with large fangs on the tooth plate. Outstanding feature are the 'parasymphysical tooth whorls' which place the fish in the order of Onychodontida. The premaxilla and the dentary had large inner teeth and irregular array |
https://en.wikipedia.org/wiki/Arboreal%20locomotion | Arboreal locomotion is the locomotion of animals in trees. In habitats in which trees are present, animals have evolved to move in them. Some animals may scale trees only occasionally, but others are exclusively arboreal. The habitats pose numerous mechanical challenges to animals moving through them and lead to a variety of anatomical, behavioral and ecological consequences as well as variations throughout different species. Furthermore, many of these same principles may be applied to climbing without trees, such as on rock piles or mountains.
Some animals are exclusively arboreal in habitat, such as the tree snail.
Biomechanics
Arboreal habitats pose numerous mechanical challenges to animals moving in them, which have been solved in diverse ways. These challenges include moving on narrow branches, moving up and down inclines, balancing, crossing gaps, and dealing with obstructions.
Diameter
Moving along narrow surfaces, such as a branch of a tree, can create special difficulties for animals who are not adapted to deal with balancing on small diameter substrates. During locomotion on the ground, the location of the center of mass may swing from side to side. But during arboreal locomotion, this would result in the center of mass moving beyond the edge of the branch, resulting in a tendency to topple over and fall. Not only do some arboreal animals have to be able to move on branches of varying diameter, but they also have to eat on these branches, resulting in the need for the ability to balance while using their hands to feed themselves. This resulted in various types of grasping such as pedal grasping in order to clamp themselves onto small branches for better balance.
Incline
Branches are frequently oriented at an angle to gravity in arboreal habitats, including being vertical, which poses special problems. As an animal moves up an inclined branch, it must fight the force of gravity to raise its body, making the movement more difficult. To get past thi |
https://en.wikipedia.org/wiki/Aronszajn%20tree | In set theory, an Aronszajn tree is a tree of uncountable height with no uncountable branches and no uncountable levels. For example, every Suslin tree is an Aronszajn tree. More generally, for a cardinal κ, a κ-Aronszajn tree is a tree of height κ in which all levels have size less than κ and all branches have height less than κ (so Aronszajn trees are the same as -Aronszajn trees). They are named for Nachman Aronszajn, who constructed an Aronszajn tree in 1934; his construction was described by .
A cardinal κ for which no κ-Aronszajn trees exist is said to have the tree property
(sometimes the condition that κ is regular and uncountable is included).
Existence of κ-Aronszajn trees
Kőnig's lemma states that -Aronszajn trees do not exist.
The existence of Aronszajn trees (-Aronszajn trees) was proven by Nachman Aronszajn, and implies that the analogue of Kőnig's lemma does not hold for uncountable trees.
The existence of -Aronszajn trees is undecidable in ZFC: more precisely, the continuum hypothesis implies the existence of an -Aronszajn tree, and Mitchell and Silver showed that it is consistent (relative to the existence of a weakly compact cardinal) that no -Aronszajn trees exist.
Jensen proved that V = L implies that there is a κ-Aronszajn tree (in fact a κ-Suslin tree) for every infinite successor cardinal κ.
showed (using a large cardinal axiom) that it is consistent that no -Aronszajn trees exist for any finite n other than 1.
If κ is weakly compact then no κ-Aronszajn trees exist. Conversely, if κ is inaccessible and no κ-Aronszajn trees exist, then κ is weakly compact.
Special Aronszajn trees
An Aronszajn tree is called special if there is a function f from the tree to the rationals so that
f(x) < f(y) whenever x < y. Martin's axiom MA() implies that all Aronszajn trees are special, a proposition sometimes abbreviated by EATS. The stronger proper forcing axiom implies the stronger statement that for any two Aronszajn trees there is a club set of |
https://en.wikipedia.org/wiki/Parent%E2%80%93child%20interaction%20therapy | Parent–child interaction therapy (PCIT) is an intervention developed by Sheila Eyberg (1988) to treat children between ages 2 and 7 with disruptive behavior problems. PCIT is an evidence-based treatment (EBT) for young children with behavioral and emotional disorders that places emphasis on improving the quality of the parent-child relationship and changing parent-child interaction patterns.
Disruptive behavior is the most common reason for referral of young children for mental health services and can vary from relatively minor infractions such as talking back to significant acts of aggression. The most commonly treated Disruptive Behavior Disorders may be classified as Oppositional Defiant Disorder (ODD) or Conduct Disorder (CD), depending on the severity of the behavior and the nature of the presenting problems. The disorders often co-occur with Attention-Deficit Hyperactivity Disorder (ADHD). It uses a unique combination of behavioral therapy, play therapy, and parent training to teach more effective discipline techniques and improve the parent–child relationship.
PCIT is typically administered once a week, with 1-hour sessions, for 10-14 sessions total and consists of two treatment phases: Child-Directed Interaction (CDI) and Parent-Directed Interaction (PDI). The CDI component focuses on improving the quality of the parent-child relationship, which will help promote changes in behavior. This sets the foundation for the PDI stage, which continues to encourage appropriate play while also focusing on a structured and consistent approach to discipline.
History
PCIT was derived from several theories, including attachment theory, social learning theory, and parenting styles theory.
Attachment theory
According to attachment theory by Ainsworth, “sensitive and responsive parenting” during infancy and toddlerhood leads the child to develop an expectation that their needs can be met by the parent. Thus, parents who show their young children greater warmth and are mor |
https://en.wikipedia.org/wiki/Netgear%20Digital%20Entertainer | Netgear's Digital Entertainer line of products are digital media players that can pull multimedia content from home computers to the typical audio/video entertainment center. There are three products in the line, the EVA700, the HD EVA8000 and the current EVA9150 Digital Entertainer Elite. All support high definition video, the EVA700 via component output up to 1080i and the EVA8000/EVA9000 up to 1080p with both component and HDMI connectors. All models support audio, video, image and streaming audio and video formats and can be networked via wired and wireless Ethernet. The EVA700 is Intel Viiv certified.
Description
Common features of both EVA models include being able to access digital media files on a network attached home computer running Windows XP, stream audio and video from the PC, and stream Internet radio (streaming MP3). USB storage devices such as iPods, thumb drives, and some digital cameras with USB interfaces can be attached directly to the units for playback of media on those USB devices.
As the Digital Entertainer products are Intel Viiv compliant, setup and media access with an Intel Viiv PC is more automated than without such a PC. However, an Intel Viiv PC is not required in order to operate an EVA unit.
The systems come with a remote control with which to manipulate the on-screen display on your television along with a large selection of suitable connecting cables.
Inputs/Outputs: At a minimum, EVAs have RCA connectors for composite video output, as well as for component video output, S-Video output, a digital S/PDIF audio output, stereo RCA audio outputs, at least 1 USB 2.0 port and 1 8P8C 10/100 Ethernet port.
History: MP101 and MP115
NETGEAR started off in the media streamer arena with the MP101, an audio-only small form factor media streamer. It had a four-line fluorescent display with line stereo RCA and 3.5" headphone outputs. Connectivity provided by built-in Ethernet and 802.11b wireless with WEP support. Controlled by a |
https://en.wikipedia.org/wiki/Sleep%20diary | A sleep diary is a record of an individual's sleeping and waking times with related information, usually over a period of several weeks. It is self-reported or can be recorded by a caregiver.
The sleep diary, or sleep log, is a tool used by doctors and patients. It is a useful resource in the diagnosis and treatment of especially circadian rhythm sleep disorders, and in monitoring whether treatment of those and other sleep disorders is successful.
Sleep diaries may be used in conjunction with actigraphy.
In addition to being a useful tool for medical professionals in the diagnosis of sleep problems, a sleep diary can help make individuals more aware of the parameters affecting their sleep. This data alone can help people self-diagnose what helps them get a good sleep.
Components
The information contained in a sleep diary includes some or all of the following points:
The time the person had wanted or intended to wake up
The time the person woke up
Whether the person woke up spontaneously, by an alarm clock, or because of another (specified) disturbance
The time the person got out of bed
A few words about how the person felt during the day (mood, drowsiness, etc.), often on a scale from 1 to 5 and the major cause
The start and end times of any daytime naps and exercises
The name, dosage and time of any drugs used including medication, sleep aids, caffeine and alcohol
The time and type/ heaviness of evening meal
Activities the last hour before bedtime, such as meditation, watching TV, playing PC-games
Stress level before bedtime, often on a scale from 1 to 5 and the major cause
The time the person tried to fall asleep
The time the person thinks sleep onset occurred
Activity during aforementioned two moments (remaining eyes closed, meditating, etc.)
The presumed cause, number, time, and length of any nighttime awakenings and activities during these moments
Quality of sleep
Level of comfort of any recalled good or bad dreams
Data collection
Sleep l |
https://en.wikipedia.org/wiki/Hanseatic%20flags | Hanseatic flags are the banners of Hanseatic cities that were flown by cogs and other ships of the Hanseatic League from 13th to 17th centuries.
History
Originally, Hanseatic ships displayed red gonfalones on their masts, which had a cross at its peak to denote the protection of the sovereign. Red was also the colour used by Danish and English shipping, the English later adopting the St George's Cross. From the second half of the 13th century, the individual Hanseatic cities created various banners to distinguish themselves from other member cities. The red gonfalone remained in use in addition to these flags. The oldest Hanseatic flag is the plain red banner used by Hamburg. Hanseatic flags were mostly red-white and some featured symbols, such as crosses.
Many cities that were members of the Hanseatic league continue to use red and white as their city colours today.
Hanseatic pennant
In addition to these banners, ships also flew a Hanseatic pennant (Hanseatenwimpel) where the upper half is white (silver) and the lower half is red.
Banners
13th century
14th century
15th century
Other seals and coins
Flags of Hanseatic cities today
External links
International Civic Heraldry
Flags
Lists and galleries of flags
Historical flags
Flags of Germany
Flags of Poland
Flags of Russia
Flags of Latvia |
https://en.wikipedia.org/wiki/Multiplicative%20cascade | In mathematics, a multiplicative cascade is a fractal/multifractal distribution of points produced via an iterative and multiplicative random process.
Definition
The plots above are examples of multiplicative cascade multifractals.
To create these distributions there are a few steps to take. Firstly, we must create a lattice of cells which will be our underlying probability density field.
Secondly, an iterative process is followed to create multiple levels of the lattice: at each iteration the cells are split into four equal parts (cells). Each new cell is then assigned a probability randomly from the set without replacement, where . This process is continued to the Nth level. For example, in constructing such a model down to level 8 we produce a 48 array of cells.
Thirdly, the cells are filled as follows: We take the probability of a cell being occupied as the product of the cell's own pi and those of all its parents (up to level 1). A Monte Carlo rejection scheme is used repeatedly until the desired cell population is obtained, as follows: x and y cell coordinates are chosen randomly, and a random number between 0 and 1 is assigned; the (x, y) cell is then populated depending on whether the assigned number is lesser than (outcome: not populated) or greater or equal to (outcome: populated) the cell's occupation probability.
Examples
To produce the plots above we filled the probability density field with 5,000 points in a space of 256 × 256.
An example of the probability density field:
The fractals are generally not scale-invariant and therefore cannot be considered standard fractals. They can however be considered multifractals. The Rényi (generalized) dimensions can be theoretically predicted. It can be shown that as ,
where N is the level of the grid refinement and,
See also
Fractal dimension
Hausdorff dimension
Scale invariance |
https://en.wikipedia.org/wiki/Tasman%20Outflow | The Tasman Outflow is a water pathway connecting water from the Pacific Ocean and the Indian Ocean. The existence of the outflow was published by scientists of the Australian CSIRO's Division of Marine and Atmospheric Research team in August 2007, interpreting salinity and temperature data captured from 1950 to 2002. The Tasman Outflow is seen as the missing link in the supergyre of the Southern Hemisphere and an important part of the thermohaline circulation.
Features
The source of the water of the Tasman Outflow is the East Australian Current. Until 2007, it was assumed that the water of this current moved in a southeastern direction towards New Zealand. However, this eastward turn toward New Zealand only occurred close to the surface, as was confirmed by the use of Argo floats at the sea surface and at a depth of 1000 dbar. At intermediate depth -around 300 to 1000 meter- the water actually turns south and westward, moving around the south of Tasmania. This water, which escapes from the East Australian Current and moves past Tasmania, is called the Tasman Outflow. The current moves further westward past the Great Australian Bight and into the Indian Ocean. In this way, the Tasman Outflow links the South Pacific Ocean to the Indian Ocean. Due to its depth, the current mainly transports Subantarctic Mode Water and Antarctic Intermediate Water with a volume transport of 4.2 ± 4.3 Sv. Here Sv stands for Sverdrup, a measure for volumetransport in the ocean. The current is limited to a narrow path between Tasmania and the Antarctic circumpolar current, due to the strong eastward Antarctic Circumpolar Current to the south of Tasmania.
Role in the thermohaline circulation
Before the discovery of the Tasman Outflow, research on the thermohaline circulation in the Southern Hemisphere was mainly focused on two other routes. One of them is known as the cold route, which moves through the Drake Passage and transports cold water deep in the ocean around Antarctica into the |
https://en.wikipedia.org/wiki/Adenylylation | Adenylylation, more commonly known as AMPylation, is a process in which an adenosine monophosphate (AMP) molecule is covalently attached to the amino acid side chain of a protein. This covalent addition of AMP to a hydroxyl side chain of the protein is a post-translational modification. Adenylylation involves a phosphodiester bond between a hydroxyl group of the molecule undergoing adenylylation, and the phosphate group of the adenosine monophosphate nucleotide (i.e. adenylic acid). Enzymes that are capable of catalyzing this process are called AMPylators.
The known amino acids to be targeted in the protein are tyrosine and threonine, and sometimes serine. When charges on a protein undergo a change, it affects the characteristics of the protein, normally by altering its shape via interactions of the amino acids which make up the protein. AMPylation can have various effects on the protein. These are properties of the protein like, stability, enzymatic activity, co-factor binding, and many other functional capabilities of a protein. Another function of adenylylation is amino acids activation, which is catalyzed by tRNA aminoacyl synthetase. The most commonly identified protein to receive AMPylation are GTPases, and glutamine synthetase.
Adenylylators
Enzymes responsible for AMPylation, called AMPylators or Adenylyltransferase, fall into two different families, all depending on their structural properties and mechanism used. AMPylator is created by two catalytic homologous halves. One half is responsible for catalyzing the adenylylation reaction, while the other half catalyzes the phosphorolytic deadenylylation reaction. These two families are the DNA-β-polymerase-like and the Fic family.
DNA-β-polymerase-like, is a family of Nucleotidyltransferase. It more specifically is known as the GlnE family. There is a specific motif that is used to clarify this particular family. The motif consists of a three stranded β-sheet which is part of magnesium ion coordination and p |
https://en.wikipedia.org/wiki/The%20Oracle%20J2EE%20Companion | The Oracle J2EE Companion is a book by Narendra M Thumbhekodige, Jai Krishna, Padmanabhan Manavazhi, Rajesh Sundararaghavan, Ravi Shankar, Reghu Krishna Pillai, V Srinivasan, Umesh Kulkarni and Vikas Mathur. The book is about Oracle Java EE technology stack and deals with different steps of learning concepts and technologies in Internet programming. It leverages the Oracle Technology Network (OTN) extensively to showcase Java EE technology.
The book was published by Tata McGraw-Hill in 2004 and was written as a textbook for the e-Pulse course conducted by Oracle India Pvt Ltd at R V Engineering college in Bangalore. The book is available in English and Chinese language.
External links
Software engineering books
Java platform |
https://en.wikipedia.org/wiki/Hoosier%20cabinet | A Hoosier cabinet (also known as a "Hoosier") is a type of cupboard or free-standing kitchen cabinet that also serves as a workstation. It was popular in the first few decades of the 20th century in the United States, since most houses did not have built-in kitchen cabinetry. The Hoosier Manufacturing Co. of New Castle, Indiana, was one of the earliest and largest manufacturers of this product, causing the term "Hoosier cabinet" to become a generic term for that type of furniture. By 1920, the Hoosier Manufacturing Company had sold two million cabinets.
Hoosier-style cabinets were also made by dozens of other companies, and most were in the Hoosier State or located nearby. Some of the larger manufacturers were Campbell-Smith-Ritchie (Boone); Coppes Brothers and Zook (the Napanee); McDougall Company; and G. I. Sellers and Sons. Hoosier cabinets evolved over the years to include more accessories and innovations that made life easier for cooks in the kitchen. They peaked in popularity in the 1920s, then declined as homes began to be constructed with built-in kitchen cabinets and counter tops. The Hoosier Manufacturing Company was sold in 1942 and liquidated. Today, Hoosier cabinets are valued by antique collectors.
Background
From 1890 to 1930, more houses were built in the United States than all of the country's prior years combined. Very few homes had built-in kitchen cabinets during the 19th century, and it was not until the late 1920s that built-in cabinets became a standard kitchen furnishing. Around the 1890s, several furniture manufacturers in Indiana discovered that a stand-alone kitchen cabinet with storage and a workspace (essentially a baker's cabinet with extra storage) was easy to sell. It was a kitchen workstation with ingredient and equipment storage where the cook could complete all food preparation and not move until it was time to cook the food. The Hoosier Manufacturing Company was one of those early manufacturers. In 1900, the company moved a sho |
https://en.wikipedia.org/wiki/Mandevilla%20sanderi | Mandevilla sanderi, the Brazilian jasmine, is a vine belonging to the genus Mandevilla. Grown as an ornamental plant, the species is endemic to the State of Rio de Janeiro in Brazil. It is a rapidly growing, creeping, perennial plant, pruning shoots about 60 cm per year.
Despite its common name, the species is not a "true jasmine" and not of the genus Jasminum.
Etymology
The genus name Mandevilla was awarded by John Lindley, a botanist, in memory of Henri Mandeville (1773-1861), one of his fellow British gardening enthusiasts who was a diplomat in Buenos Aires (Argentina). The sanderi species name refers to Henry Frederick Conrad Sander (1847-1920), a horticulturist and collector from Hertfordshire (in the UK) who brought the plant back from Brazil.
In 1896 WB Hemsley of Kew Gardens gave the first botanical description of the plant, which he named Dipladenia sanderi Hemsl. However, in 1933 Robert E. Woodson, who had undertaken a large taxonomic study of the Apocynaceae, made significant changes in the Mandevilla constituency. By including several genres such as Dipladenia inside Mandevilla, the plant ended up with the name Mandevilla sanderi.
Description
Mandevilla sanderi is a shrub with a naturally bushy habit, 2–3 meters high, or 4.5 meters (15 feet) if the climate is warm. It is able to develop long, woody stems based on lignin and climbs by twining around some support. This twining growth is characterized by long internodes, small leaves and a stem rarely carrying flowers. The plant contains a white latex, which is viscous, toxic, and can be irritating. In addition to fine roots, it has large tuberous roots that contain starch and a reserve of water, allowing it to withstand drought. The evergreen, petiolate, thick, leathery, dark green leaves are opposite, and grow to 6 cm (2.5 in) long. The blade is ovate-elliptical, 5–6 cm long, with a glossy upper surface and a thick epidermis. The apex is shortly acuminate.
The inflorescences are simple racemes, usu |
https://en.wikipedia.org/wiki/Syntrophin | The syntrophins are a family of five 60-kiloDalton proteins that are associated with dystrophin, the protein associated with Duchenne muscular dystrophy and Becker muscular dystrophy. The name comes from the Greek word syntrophos, meaning "companion." The five syntrophins are encoded by separate genes and are termed α, β1, β2, γ1, and γ2. Syntrophin was first identified as a dystrophin-associated protein present in the Torpedo electric organ (originally called "58K protein"). Subsequently, α-syntrophin was shown to be the predominant isoform in skeletal muscle where it is localized on the sarcolemma and enriched at the neuromuscular junction. The β-syntrophins and γ2-syntrophin are also present in skeletal muscle but also are in most other tissues. The expression of γ1-syntrophin is mostly confined to brain. The syntrophins are adaptor proteins that use their multiple protein interaction domains (two pleckstrin homology domains and a PDZ domain) to localize a variety of signaling proteins (kinases, ion channels, water channels, nitric oxide synthase) to specific intracellular locations. α-Syntrophin binds to nNOS in the dystrophin-associated glycoprotein complex in skeletal muscle cells. There it produces NO upon muscle contraction leading to dilation of the arteries in the local area. |
https://en.wikipedia.org/wiki/Teledotcom | A teledotcom is a domain name that not only spells a memorable word but also has a matching toll-free telephone number. This means the word that spells something will have a toll-free prefix and a top level domain extension after it. A teledotcom can be accessible by either phone or internet address.
Most owners of a toll-free number that spells something (see Phoneword), will also try to register the matching domain name and therefore create a teledotcom; a valuable asset should they wish to sell it. It is significantly more difficult for an owner of a valuable domain names, because the matching toll-free numbers are much rarer. As a consequence of the difficulties in obtaining the domain name and the matching toll-free telephone number, the number of teledotcoms is quite low.
Marketing
An example teledotcom is 1-800-Flowers.com. Customers can either visit the website or telephone 1-800-FLOWERS (1-800-356-9377). Others examples are 1-800-Mattress.com and 1-800 Contacts.com. To extend their marketing reach, these companies will also advertise on television across the United States.
Buyer Jim McCann stated that 1-800-flowers purchased the toll free number from a flower company that was seven million dollars in debt. The toll-free numbers are considered so valuable to the company that the company now owns the same number under the 888, and 877 prefix. When the 866 toll free prefix was launched, competition for the flowers number was stiff, with 1-800-flowers eventually losing to a rival company.
Unlike domain names where one person can own eLoans, and another can own eLoan and all can be allowed to use the names as they are considered fair use, toll-free numbers that are first to market and part of a brand, are usually protected. For example, when a New York-based business tried to set up 1-718-Mattress, it was forced to relinquish the use of the number as it confused people by too closely copying the brand of 1-800-Mattress.
Using a toll-free number as a bran |
https://en.wikipedia.org/wiki/Yamabe%20invariant | In mathematics, in the field of differential geometry, the Yamabe invariant, also referred to as the sigma constant, is a real number invariant associated to a smooth manifold that is preserved under diffeomorphisms. It was first written down independently by O. Kobayashi and R. Schoen and takes its name from H. Yamabe. Used by Vincent Moncrief and Arthur Fischer to study reduced Hamiltonian for Einstein's equations.
Definition
Let be a compact smooth manifold (without boundary) of dimension . The normalized Einstein–Hilbert functional assigns to each Riemannian metric on a real number as follows:
where is the scalar curvature of and is the volume density associated to the metric . The exponent in the denominator is chosen so that the functional is scale-invariant: for every positive real constant , it satisfies . We may think of as measuring the average scalar curvature of over . It was conjectured by Yamabe that every conformal class of metrics contains a metric of constant scalar curvature (the so-called Yamabe problem); it was proven by Yamabe, Trudinger, Aubin, and Schoen that a minimum value of is attained in each conformal class of metrics, and in particular this minimum is achieved by a metric of constant scalar curvature.
We define
where the infimum is taken over the smooth real-valued functions on . This infimum is finite (not ): Hölder's inequality implies . The number is sometimes called the conformal Yamabe energy of (and is constant on conformal classes).
A comparison argument due to Aubin shows that for any metric , is bounded above by , where
is the standard metric on the -sphere . It follows that if we define
where the supremum is taken over all metrics on , then (and is in particular finite). The
real number is called the Yamabe invariant of .
The Yamabe invariant in two dimensions
In the case that , (so that M is a closed surface) the Einstein–Hilbert functional is given by
where is the Gauss curvat |
https://en.wikipedia.org/wiki/DORIS%20%28satellite%20system%29 | DORIS is a French satellite system used for the determination of satellite orbits (e.g. TOPEX/Poseidon) and for positioning.
The name is an acronym of "Doppler Orbitography and Radiopositioning Integrated by Satellite" or, in French, Détermination d'Orbite et Radiopositionnement Intégré par Satellite.
Principle
Ground-based radio beacons emit a signal which is picked up by receiving satellites. This is in reverse configuration to other GNSS, in which the transmitters are space-borne and receivers are in majority near the surface of the Earth. A frequency shift of the signal occurs that is caused by the movement of the satellite (Doppler effect). From this observation satellite orbits, ground positions, as well as other parameters can be derived.
Organization
DORIS is a French system which was initiated and is maintained by the French Space Agency (CNES). It is operated from Toulouse.
Ground segment
The ground segment includes about 50-60 ground stations, equally distributed over the Earth and ensure a good coverage for orbit determination. For the installation of a beacon only electricity is required because the station only emits a signal but does not receive any information. DORIS beacons transmit to the satellites on two UHF frequencies, 401.25 MHz and 2036.25 MHz.
Australian ground segments
There are two active DORIS stations in Australia:
Yatharagga - active
Orroral Valley Tracking Station - no longer active
Mount Stromlo Observatory - currently active, replaced Orroral Valley Tracking station installation
Space segment
The best known satellites equipped with DORIS receivers are the altimetry satellites TOPEX/Poseidon, Jason-1, OSTM/Jason-2, Jason-3, and Sentinel-6 Michael Freilich. They are used to observe the ocean surface as well as currents or wave heights. DORIS contributes to their orbit accuracy of about 2 cm.
Other DORIS satellites are the Envisat, SPOT, HY-2A and CryoSat-2 satellites.
Positioning
Apart from orbit determination, the DOR |
https://en.wikipedia.org/wiki/Bomab | The BOttle MAnnequin ABsorber phantom was developed by Bush in 1949 (Bush 1949) and has since been accepted in North America as the industry standard (ANSI 1995) for calibrating whole body counting systems.
The phantom consists of 10 polyethylene bottles, either cylinders or elliptical cylinders, that represent the head, neck chest, abdomen, thighs, calves, and arms. Each section is filled with a radioactive solution, in water, that has the amount of radioactivity proportional to the volume of each section. This simulates a homogeneous distribution of material throughout the body. The solution will also be acidified and contain stable element carrier so that the radioactivity does not plate out on the container walls.
The phantom, which contains a known amount of radioactivity can be used to calibrate the whole body counter by relating the observed response to the known amount of radioactivity. As different radioactive materials emit different energies of gamma photons, the calibration has to be repeated to cover the expected energy range: usually 120 to 2,000 keV.
Examples of radioactive isotopes that are used for efficiency calibration include 57Co, 60Co, 88Y, 137Cs and 152Eu.
Although the phantom was designed to be used lying down, it is used in any orientation.
Other uses
Performance testing: BOMAB phantoms are sometimes used by performance testing organizations to test operating assay facilities. Phantoms, containing known quantities of radioactive material, are sent to assay facilities as blind samples.
Design characteristics: Phantoms can be used to evaluate the relative effect of size, shape and positioning on the performance of in vivo measurement equipment.
Background: A water filled BOMAB is often used to estimate the (blank) background for in vivo assay systems.
Detection Limits: A BOMAB filled with approximately 140 g of K-40, which is the nominal content in a 70 kg man, is sometimes used to estimate detection sensitivity of in vivo pers |
https://en.wikipedia.org/wiki/Buffon%27s%20noodle | In geometric probability, the problem of Buffon's noodle is a variation on the well-known problem of Buffon's needle, named after Georges-Louis Leclerc, Comte de Buffon who lived in the 18th century. This approach to the problem was published by Joseph-Émile Barbier in 1860.
Buffon's needle
Suppose there exist infinitely many equally spaced parallel lines, and we were to randomly toss a needle whose length is less than or equal to the distance between adjacent lines. What is the probability that the needle will lie across a line upon landing?
To solve this problem, let be the length of the needle and be the distance between two adjacent lines. Then, let be the acute angle the needle makes with the horizontal, and let be the distance from the center of the needle to the nearest line.
The needle lies across the nearest line if and only if . We see this condition from the right triangle formed by the needle, the nearest line, and the line of length when the needle lies across the nearest line.
Now, we assume that the values of are randomly determined when they land, where , since , and . The sample space for is thus a rectangle of side lengths and .
The probability of the event that the needle lies across the nearest line is the fraction of the sample space that intersects with . Since , the area of this intersection is given by
Now, the area of the sample space is
Hence, the probability of the event is
Bending the needle
The formula stays the same even when the needle is bent in any way (subject to the constraint that it must lie in a plane), making it a "noodle"—a rigid plane curve. We drop the assumption that the length of the noodle is no more than the distance between the parallel lines.
The probability distribution of the number of crossings depends on the shape of the noodle, but the expected number of crossings does not; it depends only on the length L of the noodle and the distance D between the parallel lines (observe that a curved n |
https://en.wikipedia.org/wiki/Life%20on%20Titan | Whether there is life on Titan, the largest moon of Saturn, is currently an open question and a topic of scientific assessment and research. Titan is far colder than Earth, but of all the places in the Solar System, Titan is the only place besides Earth known to have liquids in the form of rivers, lakes, and seas on its surface. Its thick atmosphere is chemically active and rich in carbon compounds. On the surface there are small and large bodies of both liquid methane and ethane, and it is likely that there is a layer of liquid water under its ice shell. Some scientists speculate that these liquid mixes may provide prebiotic chemistry for living cells different from those on Earth.
In June 2010, scientists analyzing data from the Cassini–Huygens mission reported anomalies in the atmosphere near the surface which could be consistent with the presence of methane-producing organisms, but may alternatively be due to non-living chemical or meteorological processes. The Cassini–Huygens mission was not equipped to look directly for micro-organisms or to provide a thorough inventory of complex organic compounds.
Chemistry
Titan's consideration as an environment for the study of prebiotic chemistry or potentially exotic life stems in large part due to the diversity of the organic chemistry that occurs in its atmosphere, driven by photochemical reactions in its outer layers. The following chemicals have been detected in Titan's upper atmosphere by Cassinis mass spectrometer:
As mass spectrometry identifies the atomic mass of a compound but not its structure, additional research is required to identify the exact compound that has been detected. Where the compounds have been identified in the literature, their chemical formula has been replaced by their name above. The figures in Magee (2009) involve corrections for high pressure background. Other compounds believed to be indicated by the data and associated models include ammonia, polyynes, amines, ethylenimine, deuter |
https://en.wikipedia.org/wiki/The%20Portopia%20Serial%20Murder%20Case | is an 1983 adventure game designed by Yuji Horii and published by Enix. It was first released on the NEC PC-6001 and has since been ported to other personal computers, the Nintendo Famicom, mobile phone services and most recently, Windows as Square Enix showing off their natural language processing technology.
In the game, the player must resolve a murder mystery by searching for clues, exploring different areas, interacting with characters, and solving item-based puzzles. The game features first-person graphics, nonlinear gameplay, an open world, conversations with non-player characters, branching dialogue choices, suspect interrogations, nonlinear storytelling, and plot twists. The Famicom version also features a command menu system, point-and-click interface, and 3D dungeon maze. Upon its release, The Portopia Serial Murder Case was well received in Japan. It became an influential title, helping to define the visual novel genre as well as inspiring Japanese game designers such as Hideo Kojima and Nintendo's Eiji Aonuma.
Gameplay
The Portopia Serial Murder Case follows a first-person perspective and narrative. The various events are described with still pictures and text messages. The player interacts with the game using a verb-noun parser which requires typing precise commands with the keyboard. Finding the exact words to type is considered part of the riddles that must be solved. While sound effects are present, the game lacks music and a save function. It features a conversation system with branching dialogue choices, where the story develops through entering commands and receiving answers to them from the player's sidekick or non-player characters.
The game features nonlinear gameplay, allowing multiple different ways to achieve objectives. This includes travelling between different areas in an open world and making choices that determine the dialogues and order of events as well as alternative endings depending on who the player identifies as the culprit. |
https://en.wikipedia.org/wiki/Power%20Management%20Bus | The Power Management Bus (PMBus) is a variant of the System Management Bus (SMBus) which is targeted at digital management of power supplies. Like SMBus, it is a relatively slow speed two wire communications protocol based on I²C. Unlike either of those standards, it defines a substantial number of domain-specific commands rather than just saying how to communicate using commands defined by the reader.
Overview
The first part gives an overview with particular reference to SMBus, while the second part goes into detail about all the commands defined for PMBus devices. There are both standardized commands and manufacturer specific commands. Conformance requirements for PMBus are minimal, and are described in Part I of the specification. See the PMBus 1.1 specification for full details.
Comparison to SMBus
At the lowest level, PMBus follows SMBus 1.1 with a few differences. This information is presented in more detail in Part I of the PMBus specification:
400 kHz bus speeds are allowed (vs. the 100 kHz limit of SMBus)
In PMBus, blocks may include up to 255 bytes (vs. the 32 byte limit of SMbus).
As in SMBus 2.0, only seven bit addressing is used.
Some commands use the SMBus 2.0 block process calls.
Either the SMBALERT# mechanism or the SMBus 2.0 host notify protocol may be used to notify the host about faults.
PMBus devices are required to support a Group Protocol, where devices defer acting on commands until they receive a terminating STOP. Since commands can be issued to many different devices before that STOP, this lets the PMBus master synchronize their actions.
An "extended command" protocol is defined, using a second command byte to add 256 more codes each for both standard and manufacturer-specific commands.
PMBus commands
The PMBus command space can be seen as exposing a variety of readable, and often writable, device attributes such as measured voltage and current levels, temperatures, fan speeds, and more. Different devices will expose different a |
https://en.wikipedia.org/wiki/7JP4 | The 7JP4 is an early black and white or monochrome cathode ray tube (also called picture tube and kinescope). It was a popular type used in late 1940s low cost and small table model televisions. The 7JP4 has a 7" diameter round screen which was often partially masked. Unlike later electromagnetically deflected TV tubes, the 7JP4 is electrostatically deflected like an oscilloscope tube.
Development
The 7JP4 is part of the 7JPx series of circular face electrostatic cathode ray tubes (CRT). Originally developed for radar applications as a display device for radar display A scopes around 1944. After World War 2 the CRT was adapted for television applications. There are three versions. The 7JP4 (P4 represents the phosphor that glows white and has medium persistence) for television. For oscilloscope applications the 7JP1 was used (P1 phosphor has a green trace and short persistence). Radar applications the 7JP7 was used (P7 phosphor has a blue-white trace with a long persistence). This CRT was produced by multiple manufacturers (RCA, General Electric, Sylvania Electric Products and Tung-sol). Except for the type of phosphor used all three are identical in operation and connection. The screen diagonal is 7 inches (17.8 cm) for 7JP1 and 7JP4, but only 5.5 inches (14 cm) for the 7JP7.
7JP4 Electrical Characteristics
Some General Electrical Characteristics are shown below. Second Anode + Grid 2 (Pin 9) and Plates (Pin 10 & 11 and Pin 7 & 8) have a maximum value limit of 6000 volts dc. Internal arcing can be expected when this voltage is exceeded. Actual values are typically in the 4000 to 5500 volt range, and some sets were operated as low as 2000 volts dc. Grid 1 (Pin 3) can receive either a Negative Bias or a Positive Bias. Pin-2 is the brightness voltage, and pin-3 is the video signal which rides on top of DC, and pin-2 is a DC Level which varies with the Brightness Control. Some sets are backwards and have pin-3 on ground and video on pin-2 along with brightness ad |
https://en.wikipedia.org/wiki/List%20of%20router%20and%20firewall%20distributions | This is a list of router and firewall distributions, which are operating systems designed for use as routers and/or firewalls.
See also
List of router firmware projects
Comparison of router software projects |
https://en.wikipedia.org/wiki/Singular%20integral | In mathematics, singular integrals are central to harmonic analysis and are intimately connected with the study of partial differential equations. Broadly speaking a singular integral is an integral operator
whose kernel function K : Rn×Rn → R is singular along the diagonal x = y. Specifically, the singularity is such that |K(x, y)| is of size |x − y|−n asymptotically as |x − y| → 0. Since such integrals may not in general be absolutely integrable, a rigorous definition must define them as the limit of the integral over |y − x| > ε as ε → 0, but in practice this is a technicality. Usually further assumptions are required to obtain results such as their boundedness on Lp(Rn).
The Hilbert transform
The archetypal singular integral operator is the Hilbert transform H. It is given by convolution against the kernel K(x) = 1/(πx) for x in R. More precisely,
The most straightforward higher dimension analogues of these are the Riesz transforms, which replace K(x) = 1/x with
where i = 1, ..., n and is the i-th component of x in Rn. All of these operators are bounded on Lp and satisfy weak-type (1, 1) estimates.
Singular integrals of convolution type
A singular integral of convolution type is an operator T defined by convolution with a kernel K that is locally integrable on Rn\{0}, in the sense that
Suppose that the kernel satisfies:
The size condition on the Fourier transform of K
The smoothness condition: for some C > 0,
Then it can be shown that T is bounded on Lp(Rn) and satisfies a weak-type (1, 1) estimate.
Property 1. is needed to ensure that convolution () with the tempered distribution p.v. K given by the principal value integral
is a well-defined Fourier multiplier on L2. Neither of the properties 1. or 2. is necessarily easy to verify, and a variety of sufficient conditions exist. Typically in applications, one also has a cancellation condition
which is quite easy to check. It is automatic, for instance, if K is an odd function. If |
https://en.wikipedia.org/wiki/Package%20on%20a%20package | Package on a package (PoP) is an integrated circuit packaging method to vertically combine discrete logic and memory ball grid array (BGA) packages. Two or more packages are installed atop each other, i.e. stacked, with a standard interface to route signals between them. This allows higher component density in devices, such as mobile phones, personal digital assistants (PDA), and digital cameras, at the cost of slightly higher height requirements. Stacks with more than 2 packages are uncommon, due to heat dissipation considerations.
Configuration
Two widely used configurations exist for PoP:
Pure memory stacking: two or more memory only packages are stacked on each other
Mixed logic-memory stacking: logic (CPU) package on the bottom, memory package on top. For example, the bottom could be a system on a chip (SoC) for a mobile phone. The logic package is on the bottom because it needs many more BGA connections to the motherboard.
During PCB assembly, the bottom package of a PoP stack is placed directly on the PCB, and the other package(s) of the stack are stacked on top.
The packages of a PoP stack become attached to each other (and to the PCB) during reflow soldering.
Benefits
The package on a package technique tries to combine the benefits of traditional packaging with the benefits of die-stacking techniques, while avoiding their drawbacks.
Traditional packaging places each die in its own package, a package designed for normal PCB assembly techniques that place each package directly on the PCB side-by-side.
The 3D die-stacking system in package (SiP) techniques stacks multiple die in a single package, which has several advantages and also some disadvantages compared to traditional PCB assembly.
In embedded PoP techniques, chips are embedded in a substrate on the bottom of the package. This PoP technology enables smaller packages with shorter electrical connections and is supported by companies such as Advanced Semiconductor Engineering (ASE).
Advantages ove |
https://en.wikipedia.org/wiki/Word%20%28group%20theory%29 | In group theory, a word is any written product of group elements and their inverses. For example, if x, y and z are elements of a group G, then xy, z−1xzz and y−1zxx−1yz−1 are words in the set {x, y, z}. Two different words may evaluate to the same value in G, or even in every group. Words play an important role in the theory of free groups and presentations, and are central objects of study in combinatorial group theory.
Definitions
Let G be a group, and let S be a subset of G. A word in S is any expression of the form
where s1,...,sn are elements of S, called generators, and each εi is ±1. The number n is known as the length of the word.
Each word in S represents an element of G, namely the product of the expression. By convention, the unique identity element can be represented by the empty word, which is the unique word of length zero.
Notation
When writing words, it is common to use exponential notation as an abbreviation. For example, the word
could be written as
This latter expression is not a word itself—it is simply a shorter notation for the original.
When dealing with long words, it can be helpful to use an overline to denote inverses of elements of S. Using overline notation, the above word would be written as follows:
Reduced words
Any word in which a generator appears next to its own inverse (xx−1 or x−1x) can be simplified by omitting the redundant pair:
This operation is known as reduction, and it does not change the group element represented by the word. Reductions can be thought of as relations (defined below) that follow from the group axioms.
A reduced word is a word that contains no redundant pairs. Any word can be simplified to a reduced word by performing a sequence of reductions:
The result does not depend on the order in which the reductions are performed.
A word is cyclically reduced if and only if every cyclic permutation of the word is reduced.
Operations on words
The product of two words is obtained by concatenation |
https://en.wikipedia.org/wiki/Proof%20that%20%CF%80%20is%20irrational | In the 1760s, Johann Heinrich Lambert was the first to prove that the number is irrational, meaning it cannot be expressed as a fraction , where and are both integers. In the 19th century, Charles Hermite found a proof that requires no prerequisite knowledge beyond basic calculus. Three simplifications of Hermite's proof are due to Mary Cartwright, Ivan Niven, and Nicolas Bourbaki. Another proof, which is a simplification of Lambert's proof, is due to Miklós Laczkovich. Many of these are proofs by contradiction.
In 1882, Ferdinand von Lindemann proved that is not just irrational, but transcendental as well.
Lambert's proof
In 1761, Lambert proved that is irrational by first showing that this continued fraction expansion holds:
Then Lambert proved that if is non-zero and rational, then this expression must be irrational. Since , it follows that is irrational, and thus is also irrational. A simplification of Lambert's proof is given below.
Hermite's proof
Written in 1873, this proof uses the characterization of as the smallest positive number whose half is a zero of the cosine function and it actually proves that is irrational. As in many proofs of irrationality, it is a proof by contradiction.
Consider the sequences of real functions and for defined by:
Using induction we can prove that
and therefore we have:
So
which is equivalent to
Using the definition of the sequence and employing induction we can show that
where and are polynomial functions with integer coefficients and the degree of is smaller than or equal to In particular,
Hermite also gave a closed expression for the function namely
He did not justify this assertion, but it can be proved easily. First of all, this assertion is equivalent to
Proceeding by induction, take
and, for the inductive step, consider any natural number If
then, using integration by parts and Leibniz's rule, one gets
If with and in , then, since the coefficients of are integers and its de |
https://en.wikipedia.org/wiki/Philosophy%20of%20Mathematics%20Education%20Journal | The Philosophy of Mathematics Education Journal is a peer-reviewed open-access academic journal published and edited by Paul Ernest (University of Exeter). It publishes articles relevant to the philosophy of mathematics education, a subfield of mathematics education that often draws in issues from the philosophy of mathematics. The journal includes articles, graduate student assignments, theses, and other pertinent resources.
Special issues of the journal have focussed on
social justice issues in mathematics education, part 1 (issue no. 20, 2007)
semiotics of mathematics education (issue no. 10, 1997)
See also
List of scientific journals in mathematics education
External links
Philosophy journals
Open access journals
Academic journals established in 1990
English-language journals
Mathematics education journals
Mathematics education in the United Kingdom |
https://en.wikipedia.org/wiki/Traveling%20microscope | A travelling microscope is an instrument for measuring length with a resolution typically in the order of 0.01mm. The precision is such that better-quality instruments have measuring scales made from Invar to avoid misreadings due to thermal effects. The instrument comprises a microscope mounted on two rails fixed to, or part of a very rigid bed. The position of the microscope can be varied coarsely by sliding along the rails, or finely by turning a screw. The eyepiece is fitted with fine cross-hairs to fix a precise position, which is then read off the vernier scale. Some instruments, such as that produced in the 1960s by the Precision Tool and Instrument Company of Thornton Heath, Surrey, England, also measure vertically. The purpose of the microscope is to aim at reference marks with much higher accuracy than is possible using the naked eye. It is used in laboratories to measure the refractive index of flat specimens using the geometrical concepts of ray optics (Duc de Chaulnes’ method). It is also used to measure very short distances precisely, for example the diameter of a capillary tube. This mechanical instrument has now largely been superseded by electronic- and optically based measuring devices that are both very much more accurate and considerably cheaper to produce.
Travelling microscope consists of a cast iron base with machined-Vee-top surface and is fitted with three levelling screws. A metallic carriage, clamped to a spring-loaded bar slides with its attached vernier and reading lens along an inlaid strip of metal scale. The scale is divided in half millimeters. Fine adjustments are made by means of a micrometer screw for taking accurate reading. Both vernier reading to 0.01mm or 0.02mm. Microscope tube consists of 10x Eyepice and 15mm or 50mm or 75mm objectives. The Microscope, with its rack and pinion attachment is mounted on a vertical slide, which too, runs with an attached vernier along the vertical scale. The microscope is free to rotate n vert |
https://en.wikipedia.org/wiki/Darwin%20%28programming%20language%29 | Darwin is a closed source programming language developed by Gaston Gonnet and colleagues at ETH Zurich. It is used to develop the OMA orthology inference software, which was also initially developed by Gonnet. The language backend consists of the kernel, responsible for performing simple mathematical calculations, for transporting and storing data and for interpreting the user's commands, and the library, a set of programs which can perform more complicated calculations. The target audience for the language is the biosciences, so the library consisted of routines such as those to compute pairwise alignments, phylogenetic trees, multiple sequence alignments, and to make secondary structure predictions.
Example Code
One would write the Hello World program as:
printf('Hello, world!\n');
The following procedure calculates the factorial of a number:
factorial := proc ( n )
if (n=0) then
return(1);
else
return(n * factorial(n-1));
fi;
end:
See also
List of programming languages |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.