source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/XACML |
The eXtensible Access Control Markup Language (XACML) is an XML-based standard markup language for specifying access control policies. The standard, published by OASIS, defines a declarative fine-grained, attribute-based access control policy language, an architecture, and a processing model describing how to evaluate access requests according to the rules defined in policies.
XACML is primarily an attribute-based access control system. In XACML, attributes – information about the subject accessing a resource, the resource to be addressed, and the environment – act as inputs for the decision of whether access is granted or not. XACML can also be used to implement role-based access control.
In XACML, access control decisions to be taken are expressed as Rules. Each Rule comprises a series of conditions which decide whether a given request is approved or not. If a Rule is applicable to a request but the conditions within the Rule fail to evaluate, the result is Indeterminate. Rules are grouped together in Policies, and a PolicySet contains Policies and possibly other PolicySets. Each of these also includes a Target, a simple condition that determines whether it should be evaluated for a given request. Combining algorithms can be used to combine Rules and Policies with potentially differing results in various ways. XACML also supports obligations and advice expressions. Obligations specify actions which must be executed during the processing of a request, for example for logging. Advice expressions are similar, but may be ignored.
XACML separates access control functionality into several components. Each operating environment in which access control is used has a Policy Enforcement Point (PEP) which implements the functionality to demand authorization and to grant or deny access to resources. These refer to an environment-independent and central Policy Decision Point (PDP) which actually makes the decision on whether access is granted. The PDP refers to policies s |
https://en.wikipedia.org/wiki/KRCW-TV | KRCW-TV (channel 32) is a television station licensed to Salem, Oregon, United States, serving as the CW outlet for the Portland area. It is owned and operated by network majority owner Nexstar Media Group alongside CBS affiliate KOIN (channel 6). Both stations share studios in the basement of the KOIN Center skyscraper on Southwest Columbia Street in downtown Portland, while KRCW-TV's transmitter is located in the Sylvan-Highlands neighborhood of the city.
Previously, KRCW-TV maintained separate studios on Southwest Arctic Drive in Beaverton, while KOIN's facilities only housed KRCW-TV's master control and some internal operations. Despite Salem being KRCW-TV's city of license, the station maintains no physical presence there.
History
Early history
The station was launched on May 8, 1989, under the call sign KUTF (standing for "Keep Up the Faith"), its original transmitter was located outside Molalla. The station's original programming format almost entirely consisted of religious programs. It was originally operated by Dove Broadcasting, owner of Christian television station WGGS-TV in Greenville, South Carolina; local productions included a version of WGGS's popular Nite Line talk program.
Despite its long legacy in Christian television (its flagship has been on the air since 1972), Dove struggled to build a support base for KUTF. In May 1990, the station went dark. According to station insiders, the Jim Bakker and Jimmy Swaggart scandals gave potential supporters pause. It did not help matters that the station had received competition a few months after signing on from KNMT, with wealthier ownership (Trinity Broadcasting Network, through subsidiary National Minority Television) and a stronger signal.
KUTF resumed broadcasting a month later. Dove sold KUTF to Eagle Broadcasting on July 17, 1991. The call sign was changed to KEBN on February 11, 1992; the new owners then proceeded to relaunch the station as "Oregon's New Eagle 32", becoming a general entertai |
https://en.wikipedia.org/wiki/Instant%20mashed%20potatoes | Instant mashed potatoes are potatoes that have been through an industrial process of cooking, mashing and dehydrating to yield a packaged convenience food that can be reconstituted by adding hot water or milk, producing an approximation of mashed potatoes. They are available in many different flavors.
Mashed potatoes can be reconstituted from potato flour, but the process is made more difficult by lumping; a key characteristic of instant mashed potatoes is that it is in the form of flakes or granules, eliminating the chunkiness. Analogous to instant mashed potatoes are instant poi made from taro and instant fufu made from yams or yam substitutes including cereals. Poha, an instant rice mush, is also much in the same spirit, as more broadly are other instant porridges, formed from flakes, granules, or pearls to avoid lumping. Brands include Smash and Idahoan Foods.
Flaked instant mashed potatoes are most commonly found in stores in the United States, Mexico and Canada. Granulated forms are generally reserved more for institutional or restaurant use.
History
The practice of drying starchy root vegetables for preservation and portability is widely attested around the world, and likely dates back to before the advent of agriculture. Potatoes in particular have been freeze dried since at least the time of the Inca empire, in the form of chuño; another example is in Japanese Ainu cuisine, where potatoes are collected frozen from the ground in spring, then dried.
U.S. Patent 1025373, developed by Ernest William Cooke, titled "Dehydrate Potatoes and Process of Preparing the Same", and describing a product that was to be reconstituted in hot water, was applied for in 1905 and granted in 1912.
Flake-form instant mashed potatoes date back at least to 1954, when two United States Department of Agriculture researchers were issued a patent for "Drum drying of cooked mashed potatoes" (U.S. Patent 2,759,832), which describes the end product specifically being "as a thin sh |
https://en.wikipedia.org/wiki/Infraspinatus%20muscle | In human anatomy, the infraspinatus muscle is a thick triangular muscle, which occupies the chief part of the infraspinatous fossa. As one of the four muscles of the rotator cuff, the main function of the infraspinatus is to externally rotate the humerus and stabilize the shoulder joint.
Structure
It attaches medially to the infraspinous fossa of the scapula and laterally to the middle facet of the greater tubercle of the humerus.
The muscle arises by fleshy fibers from the medial two-thirds of the infraspinatous fossa, and by tendinous fibers from the ridges on its surface; it also arises from the infraspinatous fascia which covers it, and separates it from the teres major and teres minor.
The fibers converge to a tendon, which glides over the lateral border of the spine of the scapula and passing across the posterior part of the capsule of the shoulder-joint, is inserted into the middle impression on the greater tubercle of the humerus.
The trapezoidal insertion of the infraspinatus onto the humerus is much larger than the equivalent insertion of the supraspinatus, the reason why the infraspinatus is involved in rotator cuff tears about as frequently as the supraspinatus.
Relations
The tendon of this muscle is sometimes separated from the capsule of the shoulder-joint by a bursa, which may communicate with the joint cavity.
Innervation
The suprascapular nerve innervates the supraspinatus and infraspinatus muscles. These muscles function to abduct and laterally rotate the arm, respectively.
Variation
The infraspinatus is frequently fused with the teres minor.
Function
The infraspinatus is the main external rotator of the shoulder. When the arm is fixed, it adducts the inferior angle of the scapula. Its synergists are teres minor and the deltoid. The infraspinatus and teres minor rotate the head of the humerus outward (external, or lateral, rotation); they also assist in carrying the arm backward. Additionally, the infraspinatus reinforces the capsule of th |
https://en.wikipedia.org/wiki/210%20%28number%29 | 210 (two hundred [and] ten) is the natural number following 209 and preceding 211.
In mathematics
210 is a composite number, an abundant number, Harshad number, and the product of the first four prime numbers (2, 3, 5, and 7), and thus a primorial. It is also the least common multiple of these four prime numbers. It is the sum of eight consecutive prime numbers (13 + 17 + 19 + 23 + 29 + 31 + 37 + 41 = 210).
It is a triangular number (following 190 and preceding 231), a pentagonal number (following 176 and preceding 247), and the second smallest to be both triangular and pentagonal (the third is 40755).
It is also an idoneal number, a pentatope number, a pronic number, and an untouchable number. 210 is also the third 71-gonal number, preceding 418. It is the first primorial number greater than 2 which is not adjacent to 2 primes (211 is prime, but 209 is not).
It is the largest number n such that all primes between n/2 and n yield a representation as a sum of two primes.
Integers between 211 and 219
211
212
213
214
215
216
217
218
219
See also
210 BC
AD 210
North American telephone area code area code 210 |
https://en.wikipedia.org/wiki/High-fructose%20corn%20syrup | High-fructose corn syrup (HFCS), also known as glucose–fructose, isoglucose and glucose–fructose syrup, is a sweetener made from corn starch. As in the production of conventional corn syrup, the starch is broken down into glucose by enzymes. To make HFCS, the corn syrup is further processed by D-xylose isomerase to convert some of its glucose into fructose. HFCS was first marketed in the early 1970s by the Clinton Corn Processing Company, together with the Japanese Agency of Industrial Science and Technology, where the enzyme was discovered in 1965.
As a sweetener, HFCS is often compared to granulated sugar, but manufacturing advantages of HFCS over sugar include that it is cheaper. "HFCS 42" and "HFCS 55" refer to dry weight fructose compositions of 42% and 55% respectively, the rest being glucose. HFCS 42 is mainly used for processed foods and breakfast cereals, whereas HFCS 55 is used mostly for production of soft drinks.
The United States Food and Drug Administration (FDA) states that it is not aware of evidence showing that HFCS is less safe than traditional sweeteners such as sucrose and honey. Uses and exports of HFCS from American producers have grown steadily during the early 21st century.
Food
In the U.S., HFCS is among the sweeteners that mostly replaced sucrose (table sugar) in the food industry. Factors contributing to the increased use of HFCS in food manufacturing include production quotas of domestic sugar, import tariffs on foreign sugar, and subsidies of U.S. corn, raising the price of sucrose and reducing that of HFCS, making it a lower cost for manufacturing among sweetener applications. In spite of having a 10% greater fructose content, the relative sweetness of HFCS 55, used most commonly in soft drinks, is comparable to that of sucrose. HFCS provides advantages in food and beverage manufacturing, such as simplicity for formulation and stability, enabling processing efficiencies.
HFCS (or standard corn syrup) is the primary ingredient i |
https://en.wikipedia.org/wiki/Stafford%20knot | The Stafford knot, more commonly known as the Staffordshire knot, is a distinctive three-looped knot that is the traditional symbol of the English county of Staffordshire and of its county town, Stafford. It is a particular representation of the simple overhand knot, the most basic knot of all.
Origins
One legend of its origin, generally considered mythical, is that three convicted criminals who had committed a crime together were due to be executed in Stafford gaol. There was argument over who should be hanged first but the hangman solved the problem by devising this knot and hanging the three simultaneously. The knot can be seen on a carved Anglo-Saxon cross in a churchyard in Stoke-upon-Trent, giving its name as the Staffordshire Knot and also on a seventh-century Anglo-Saxon object from the Staffordshire hoard. This strongly suggests it pre-dates the Norman and medieval period, being probably either a heraldic symbol of early Mercia or a Celtic Christian symbol brought to Staffordshire by missionary monks from Lindisfarne.
The earliest known appearance of the knot in association with the Stafford family was on the 15th-century seal of Lady Joan de Stafford, Lady Wake (daughter of Hugh de Stafford, 2nd Earl of Stafford), who adapted the Wake knot for her use. The seal, on which four knots appear tied on a string around her coat of arms, is now in the British Museum. After Lady Wake died childless in 1443, her nephew Humphrey de Stafford, 6th Earl of Stafford inherited her personal property. Humphrey, who was made the Duke of Buckingham the following year, incorporated his aunt's adaptation of the Wake knot as a cordon around his seal, although he used three knots and not four.
The knot appears on a drawing of the standard of Sir Henry de Stafford, which was flown . In a visitation of Stafford in 1583, the town's arms were recorded for the first time as or a chevron gules, a Stafford knot argent. These arms, which represent the traditional Stafford arms with |
https://en.wikipedia.org/wiki/Full%20configuration%20interaction | Full configuration interaction (or full CI) is a linear variational approach which provides numerically exact solutions (within the infinitely flexible complete basis set) to the electronic time-independent, non-relativistic Schrödinger equation.
Explanation
It is a special case of the configuration interaction method in which all Slater determinants (or configuration state functions, CSFs) of the proper symmetry are included in the variational procedure (i.e., all Slater determinants obtained by exciting all possible electrons to all possible virtual orbitals, orbitals which are unoccupied in the electronic ground state configuration). This method is equivalent to computing the eigenvalues of the electronic molecular Hamiltonian within the basis set of the above-mentioned configuration state functions.
In a minimal basis set a full CI computation is very easy. But in larger basis sets this is usually just a limiting case which is not often attained. This is because exact solution of the full CI determinant is NP-complete, so the existence of a polynomial time algorithm is unlikely. The Davidson correction is a simple correction which allows one to estimate the value of the full CI energy from a limited configuration interaction expansion result.
Because the number of determinants required in the full CI expansion grows factorially with the number of electrons and orbitals, full CI is only possible for atoms or very small molecules with about a dozen or fewer electrons. Full CI problems including several million up to a few billion determinants are possible using current algorithms. Because full CI results are exact within the space spanned by the orbital basis set, they are invaluable in benchmarking approximate quantum chemical methods. This is particularly important in cases such as bond-breaking reactions, diradicals, and first-row transition metals, where electronic near-degeneracies can invalidate the approximations inherent in many standard methods su |
https://en.wikipedia.org/wiki/Tin%28IV%29%20oxide | Tin(IV) oxide, also known as stannic oxide, is the inorganic compound with the formula SnO2. The mineral form of SnO2 is called cassiterite, and this is the main ore of tin. With many other names, this oxide of tin is an important material in tin chemistry. It is a colourless, diamagnetic, amphoteric solid.
Structure
Tin(IV) oxide crystallises with the rutile structure. As such the tin atoms are six coordinate and the oxygen atoms three coordinate. SnO2 is usually regarded as an oxygen-deficient n-type semiconductor.
Hydrous forms of SnO2 have been described as stannic acid. Such materials appear to be hydrated particles of SnO2 where the composition reflects the particle size.
Preparation
Tin(IV) oxide occurs naturally. Synthetic tin(IV) oxide is produced by burning tin metal in air. Annual production is in the range of 10 kilotons. SnO2 is reduced industrially to the metal with carbon in a reverberatory furnace at 1200–1300 °C.
Amphoterism
Although SnO2 is insoluble in water, it is amphoteric, dissolving in base and acid. "Stannic acid" refers to hydrated tin (IV) oxide, SnO2, which is also called "stannic oxide."
Tin oxides dissolve in acids. Halogen acids attack SnO2 to give hexahalostannates, such as [SnI6]2−. One report describes reacting a sample in refluxing HI for many hours.
SnO2 + 6 HI → H2SnI6 + 2 H2O
Similarly, SnO2 dissolves in sulfuric acid to give the sulfate:
SnO2 + 2 H2SO4 → Sn(SO4)2 + 2 H2O
SnO2 dissolves in strong bases to give "stannates," with the nominal formula Na2SnO3. Dissolving the solidified SnO2/NaOH melt in water gives Na2[Sn(OH)6], "preparing salt," which is used in the dye industry.
Uses
In conjunction with vanadium oxide, it is used as a catalyst for the oxidation of aromatic compounds in the synthesis of carboxylic acids and acid anhydrides.
Ceramic glazes
Tin(IV) oxide has long been used as an opacifier and as a white colorant in ceramic glazes.’The Glazer’s Book’ – 2nd edition. A.B.Searle.The Technical Press Limited. Lon |
https://en.wikipedia.org/wiki/Obstetrical%20dilemma | The obstetrical dilemma is a hypothesis to explain why humans often require assistance from other humans during childbirth to avoid complications, whereas most non-human primates give birth unassisted with relatively little difficulty. This occurs due to the tight fit of the fetal head to the maternal birth canal, which is additionally convoluted, meaning the head and therefore body of the infant must rotate during childbirth in order to fit, unlike in other, non-upright walking mammals. Consequently, there is a usually high incidence of cephalopelvic disproportion and obstructed labor in humans.
The obstetrical dilemma claims that this difference is due to the biological trade-off imposed by two opposing evolutionary pressures in the development of the human pelvis: smaller birth canals in the mothers, and larger brains, and therefore skulls in the babies. Proponents believe bipedal locomotion (the ability to walk upright) decreased the size of the bony parts of the birth canal. They also believe that as hominids' and humans' skull and brain sizes increased over the millennia, that women needed wider hips to give birth, that these wider hips made women inherently less able to walk or run than men, and that babies had to be born earlier to fit through the birth canal, resulting in the so-called fourth trimester period for newborns (being born when the baby seems less developed than in other animals). Recent evidence has suggested bipedal locomotion is only a part of the strong evolutionary pressure constraining the expansion of the maternal birth canal. In addition to bipedal locomotion, the reduced strength of the pelvic floor due to a wider maternal pelvis also leads to fitness detriments in the mother pressuring the birth canal to remain relatively narrow.
This idea was widely accepted when first published in 1960, but has since been criticized by other scientists.
History
The term, obstetrical dilemma, was coined in 1960, by Sherwood Larned Washburn, a pr |
https://en.wikipedia.org/wiki/Cutting%20%28plant%29 | A plant cutting is a piece of a plant that is used in horticulture for vegetative (asexual) propagation. A piece of the stem or root of the source plant is placed in a suitable medium such as moist soil. If the conditions are suitable, the plant piece will begin to grow as a new plant independent of the parent, a process known as striking. A stem cutting produces new roots, and a root cutting produces new stems. Some plants can be grown from leaf pieces, called leaf cuttings, which produce both stems and roots. The scions used in grafting are also called cuttings.
Propagating plants from cuttings is an ancient form of cloning. There are several advantages of cuttings, mainly that the produced offspring are practically clones of their parent plants. If a plant has favorable traits, it can continue to pass down its advantageous genetic information to its offspring. This is especially economically advantageous as it allows commercial growers to clone a certain plant to ensure consistency throughout their crops.
Evolutionary advantage: Succulents
Cuttings are used as a method of asexual reproduction in succulent horticulture, commonly referred to as vegetative reproduction. A cutting can also be referred to as a propagule. Succulents have evolved with the ability to use adventitious root formation in reproduction to increase fitness in stressful environments. Succulents grow in shallow soils, rocky soils, and desert soils. Seedlings from sexual reproduction have a low survival rate; however, plantlets from the excised stem cuttings and leaf cuttings, broken off in the natural environment, are more successful.
Cuttings have both water and carbon stored and available, which are resources needed for plant establishment. The detached part of the plant remains physiologically active, allowing mitotic activity and new root structures to form for water and nutrient uptake. Asexual reproduction of plants is also evolutionarily advantageous as it allows plantlets to be bette |
https://en.wikipedia.org/wiki/Radio%20Link%20Protocol | Radio Link Protocol (RLP) is an automatic repeat request (ARQ) fragmentation protocol used over a wireless (typically cellular) air interface. Most wireless air interfaces are tuned to provide 1% packet loss, and most Vocoders are mutually tuned to sacrifice very little voice quality at 1% packet loss. However, 1% packet loss is intolerable to all variants of TCP, and so something must be done to improve reliability for voice networks carrying TCP/IP data.
A RLP detects packet losses and performs retransmissions to bring packet loss down to .01%, or even .0001%, which is suitable for TCP/IP applications. RLP also implements stream fragmentation and reassembly, and sometimes, in-order delivery. Newer forms of RLP also provide framing and compression, while older forms of RLP rely upon a higher-layer PPP protocols to provide these functions.
A RLP transport cannot ask the air interface to provide a certain payload size. Instead, the air interface scheduler determines the packet size, based upon constantly changing channel conditions, and upcalls RLP with the chosen packet payload size, right before transmission. Most other fragmentation protocols, such as those of 802.11b and IP, used payload sizes determined by the upper layers, and call upon the MAC to create a payload of a certain size. These other protocols are not as flexible as RLP, and can sometimes fail to transmit during a deep fade in a wireless environment.
Because a RLP payload size can be as little as 11 bytes, based upon a CDMA IS-95 network's smallest voice packet size, RLP headers must be very small, to minimize overhead. This is typically achieved by allowing both ends to negotiate a variable 'sequence number space', which is used to number each byte in the transmission stream. In some variants of RLP, this sequence counter can be as small as 6 bits.
A RLP protocol can be ACK-based or NAK-based. Most RLPs are NAK-based, meaning that forward-link sender assumes that each transmission got through, a |
https://en.wikipedia.org/wiki/Microsoft%20Forefront%20Threat%20Management%20Gateway | Microsoft Forefront Threat Management Gateway (Forefront TMG), formerly known as Microsoft Internet Security and Acceleration Server (ISA Server), is a discontinued network router, firewall, antivirus program, VPN server and web cache from Microsoft Corporation. It ran on Windows Server and works by inspecting all network traffic that passes through it.
Features
Microsoft Forefront TMG offers a set of features which include:
Routing and remote access features: Microsoft Forefront TMG can act as a router, an Internet gateway, a virtual private network (VPN) server, a network address translation (NAT) server and a proxy server.
Security features: Microsoft Forefront TMG is a firewall which can inspect network traffic (including web content, secure web content and emails) and filter out malware, attempts to exploit security vulnerabilities and content that does not match a predefined security policy. In technical sense, Microsoft Forefront TMG offers application layer protection, stateful filtering, content filtering and anti-malware protection.
Network performance features: Microsoft Forefront TMG can also improve network performance: It can compress web traffic to improve communication speed. It also offers web caching: It can cache frequently-accessed web content so that users can access them faster from the local network cache. Microsoft Forefront TMG 2010 can also cache data received through Background Intelligent Transfer Service, such as updates of software published on Microsoft Update website.
History
Microsoft Proxy Server
The Microsoft Forefront Threat Management Gateway product line originated with Microsoft Proxy Server. Developed under the code-name "Catapult", Microsoft Proxy Server v1.0 was first launched in January 1997, and was designed to run on Windows NT 4.0. Microsoft Proxy Server v1.0 was a basic product designed to provide Internet Access for clients in a LAN Environment via TCP/IP. Support was also provided for IPX/SPX networks (primarily |
https://en.wikipedia.org/wiki/Air%20handler | An air handler, or air handling unit (often abbreviated to AHU), is a device used to regulate and circulate air as part of a heating, ventilating, and air-conditioning (HVAC) system. An air handler is usually a large metal box containing a blower, furnace or A/C elements, filter racks or chambers, sound attenuators, and dampers. Air handlers usually connect to a ductwork ventilation system that distributes the conditioned air through the building and returns it to the AHU, sometimes exhausting air to the atmosphere and bringing in fresh air. Sometimes AHUs discharge (supply) and admit (return) air directly to and from the space served without ductwork
Small air handlers, for local use, are called terminal units, and may only include an air filter, coil, and blower; these simple terminal units are called blower coils or fan coil units. A larger air handler that conditions 100% outside air, and no recirculated air, is known as a makeup air unit (MAU) or fresh air handling unit (FAHU). An air handler designed for outdoor use, typically on roofs, is known as a packaged unit (PU), heating and air conditioning unit (HCU), or rooftop unit (RTU).
Construction
The air handler is normally constructed around a framing system with metal infill panels as required to suit the configuration of the components. In its simplest form the frame may be made from metal channels or sections, with single skin metal infill panels. The metalwork is normally galvanized for long term protection. For outdoor units some form of weatherproof lid and additional sealing around joints is provided.
Larger air handlers will be manufactured from a square section steel framing system with double skinned and insulated infill panels. Such constructions reduce heat loss or heat gain from the air handler, as well as providing acoustic attenuation. Larger air handlers may be several meters long and are manufactured in a sectional manner and therefore, for strength and rigidity, steel section base rails |
https://en.wikipedia.org/wiki/Hadamard%20factorization%20theorem | In mathematics, and particularly in the field of complex analysis, the Hadamard factorization theorem asserts that every entire function with finite order can be represented as a product involving its zeroes and an exponential of a polynomial. It is named for Jacques Hadamard.
The theorem may be viewed as an extension of the fundamental theorem of algebra, which asserts that every polynomial may be factored into linear factors, one for each root. It is closely related to Weierstrass factorization theorem, which does not restrict to entire functions with finite orders.
Formal statement
Define the Hadamard canonical factors Entire functions of finite order have Hadamard's canonical representation:where are those roots of that are not zero (), is the order of the zero of at (the case being taken to mean ), a polynomial (whose degree we shall call ), and is the smallest non-negative integer such that the seriesconverges. The non-negative integer is called the genus of the entire function . In this notation,In other words: If the order is not an integer, then is the integer part of . If the order is a positive integer, then there are two possibilities: or .
Furthermore, Jensen's inequality implies that its roots are distributed sparsely, with critical exponent .
For example, , and are entire functions of genus .
Critical exponent
Define the critical exponent of the roots of as the following:where is the number of roots with modulus . In other words, we have an asymptotic bound on the growth behavior of the number of roots of the function:It's clear that .
Theorem: If is an entire function with infinitely many roots, thenNote: These two equalities are purely about the limit behaviors of a real number sequence that diverges to infinity. It does not involve complex analysis.
Proposition: , by Jensen's formula.
Proof
Since is also an entire function with the same order and genus, we can wlog assume .
If has only finitely many roots, then |
https://en.wikipedia.org/wiki/Bochner%27s%20theorem | In mathematics, Bochner's theorem (named for Salomon Bochner) characterizes the Fourier transform of a positive finite Borel measure on the real line. More generally in harmonic analysis, Bochner's theorem asserts that under Fourier transform a continuous positive-definite function on a locally compact abelian group corresponds to a finite positive measure on the Pontryagin dual group. The case of sequences was first established by Gustav Herglotz (see also the related Herglotz representation theorem.)
The theorem for locally compact abelian groups
Bochner's theorem for a locally compact abelian group G, with dual group , says the following:
Theorem For any normalized continuous positive-definite function f on G (normalization here means that f is 1 at the unit of G), there exists a unique probability measure μ on such that
i.e. f is the Fourier transform of a unique probability measure μ on . Conversely, the Fourier transform of a probability measure on is necessarily a normalized continuous positive-definite function f on G. This is in fact a one-to-one correspondence.
The Gelfand–Fourier transform is an isomorphism between the group C*-algebra C*(G) and C0(Ĝ). The theorem is essentially the dual statement for states of the two abelian C*-algebras.
The proof of the theorem passes through vector states on strongly continuous unitary representations of G (the proof in fact shows that every normalized continuous positive-definite function must be of this form).
Given a normalized continuous positive-definite function f on G, one can construct a strongly continuous unitary representation of G in a natural way: Let F0(G) be the family of complex-valued functions on G with finite support, i.e. h(g) = 0 for all but finitely many g. The positive-definite kernel K(g1, g2) = f(g1 − g2) induces a (possibly degenerate) inner product on F0(G). Quotiening out degeneracy and taking the completion gives a Hilbert space
whose typical element is an equivalence cl |
https://en.wikipedia.org/wiki/Michelson%E2%80%93Gale%E2%80%93Pearson%20experiment | The Michelson–Gale–Pearson experiment (1925) is a modified version of the Michelson–Morley experiment and the Sagnac-Interferometer. It measured the Sagnac effect due to Earth's rotation, and thus tests the theories of special relativity and luminiferous ether along the rotating frame of Earth.
Experiment
The aim, as it was first proposed by Albert A. Michelson in 1904 and then executed in 1925 by Michelson and Henry G. Gale, was to find out whether the rotation of the Earth has an effect on the propagation of light in the vicinity of the Earth.
The Michelson-Gale experiment was a very large ring interferometer, (a perimeter of 1.9 kilometers), large enough to detect the angular velocity of the Earth. Like the original Michelson-Morley experiment, the Michelson-Gale-Pearson version compared the light from a single source (carbon arc) after travelling in two directions. The major change was to replace the two "arms" of the original MM version with two rectangles, one much larger than the other. Light was sent into the rectangles, reflecting off mirrors at the corners, and returned to the starting point. Light exiting the two rectangles was compared on a screen just as the light returning from the two arms would be in a standard MM experiment. The expected fringe shift in accordance with the stationary aether and special relativity was given by Michelson as:
where is the displacement in fringes, the area in square kilometers, the latitude of the experiment site in Clearing, Illinois (41° 46'), the speed of light, the angular velocity of Earth, the effective wavelength used. In other words, this experiment was aimed to detect the Sagnac effect due to Earth's rotation.
Result
The outcome of the experiment was that the angular velocity of the Earth as measured by astronomy was confirmed to within measuring accuracy. The ring interferometer of the Michelson-Gale experiment was not calibrated by comparison with an outside reference (which was not possible, becau |
https://en.wikipedia.org/wiki/SYNOP | SYNOP (surface synoptic observations) is a numerical code (called FM-12 by WMO) used for reporting weather observations made by staffed and automated weather stations. SYNOP reports are typically sent every six hours by Deutscher Wetterdienst on shortwave and low frequency using RTTY. A report consists of groups of numbers (and slashes where data is not available) describing general weather information, such as the temperature, barometric pressure and visibility at a weather station. It can be decoded by open-source software such as seaTTY, metaf2xml or Fldigi.
SYNOP information is collected by more than 7600 manned and unmanned meteorological stations and more than 2500 mobile stations around the world and is used for weather forecasting and climatic statistics. The format of the original messages is abbreviated, some items are coded.
Message format
Following is the general structure of a SYNOP message. The message consists of a sequence of numeric groups, which may also contain slashes (indicating missing data) in addition to numeric digits. Leading numbers are fixed group indicators that indicate the type of observation following, and letters are replaced with numbers giving the weather data. Messages from shipboard weather stations, and in different regions of the world, use variations on this scheme.
YYGGiw IIiii iRiXhVV Nddff (00fff) 1snTTT 2snTdTdTd 3PoPoPoPo 4PPPP 5appp 6RRRtR 7wwW1W2 8NhCLCMCH (9GGgg)
YYGGiw: the date and time of the observation; YY for the day of the month, GG for the hour of the observation in UTC; iw for the manner of wind observation (a code number: 0 for estimated wind speed in meters per second, 1 for measured wind speed in meters per second, 2 and 3 likewise but in knots, or slash for no wind speed observations).
IIiii: weather station identification code; II for a block number allocated (by the WMO) to a country or a region of the world, for example 02 for Scandinavia or 72 and 74 for the continental US; iii is the code of a |
https://en.wikipedia.org/wiki/Chu%C3%B1o | () is a freeze-dried potato product traditionally made by Quechua and Aymara communities of Bolivia and Peru, and is known in various countries of South America, including Bolivia, Peru, Chile and Northwest Argentina. It is a five-day process, obtained by exposing a bitter, frost-resistant variety of potatoes to the very low night temperatures of the Andean Altiplano, freezing them, and subsequently exposing them to the intense sunlight of the day (this being the traditional process). The word comes from Quechua , meaning 'frozen potato' ('wrinkled' in the dialects of the Junín Region).
Origins
The existence of dates back to before the time of the Inca Empire in the 13th century, based on findings that have been made of the product at various archaeological sites. Specifically, they have been found at Tiwanaku, site of a culture which developed in the Collao Plateau, a geographic zone which includes territories of Bolivia and Peru.
It had been described in 1590 by Spanish chronicler José de Acosta. Due to its portability, long shelf life, and nutritional value, was eaten by Inca soldiers on marches. Indeed Carl Troll argued that the nighttime sub-freezing temperatures of southern Peruvian highlands that allowed for production favoured the rise of the Inca Empire.
A form of ( or ) is made from the starchy storage roots of Alstroemeria ligtu.
Production
is made at the beginning of winter during June and July, during which time the temperatures reach around at elevations of over . After fall harvest (April–May), potatoes are selected for the production of , typically small ones for ease of processing. These small potatoes are spread closely on flat ground, and allowed to freeze with the low night temperatures and dehydrate in the daytime, for about three nights. This process results in natural freeze-drying.
By the end of this process, the potatoes are taken to – flat areas where the potatoes can be laid out. The term is Aymara in origin and translates |
https://en.wikipedia.org/wiki/Roger%20J.%20Williams | Roger John Williams (August 14, 1893 – February 20, 1988), was an American biochemist. He is known for is work on vitamins and human nutrition. He had leading roles in the discovery of folic acid, pantothenic acid, vitamin B6, lipoic acid, and avidin. He was elected to the National Academy of Sciences in 1946, and served as the president of the American Chemical Society in 1957. In his later career he spent time writing for a popular audience on the importance of nutrition.
Early life and education
Roger John Williams was born in Ootacamund, India of American parents on August 14, 1893. His family returned to the US when he was two years old, and he grew up in Kansas and California. He attributed his early interest in chemistry to the influence of his brother Robert R. Williams, eight years his senior, who was also a chemist. Robert is remembered for being the first to synthesize thiamine (vitamin B1). Roger was an undergraduate at the University of Redlands and received his bachelor's degree in 1914. He received a teaching certificate from the University of California, Berkeley the following year and began work as a science teacher in California. After a year of teaching, he decided to return to school and began graduate work at the University of Chicago, from which he received his Ph.D. in 1919. His Ph.D Thesis was entitled The Vitamine Requirements of Yeasts. Julius Stieglitz was the chairman of the department of chemistry at U Chicago when Roger Williams was a student there, and Williams later described Stieglitz as a major influence upon him in organic chemistry.
Academic career
Williams began his academic research career by joining the faculty at the University of Oregon in 1920. During the following twelve years he spent there, he discovered pantothenic acid. In 1932 he moved to Oregon State College and in 1939 he moved again to the University of Texas at Austin. He founded and became the founding director of the Biochemical Institute (later the Clayton Fo |
https://en.wikipedia.org/wiki/Normal%20bundle | In differential geometry, a field of mathematics, a normal bundle is a particular kind of vector bundle, complementary to the tangent bundle, and coming from an embedding (or immersion).
Definition
Riemannian manifold
Let be a Riemannian manifold, and a Riemannian submanifold. Define, for a given , a vector to be normal to whenever for all (so that is orthogonal to ). The set of all such is then called the normal space to at .
Just as the total space of the tangent bundle to a manifold is constructed from all tangent spaces to the manifold, the total space of the normal bundle to is defined as
.
The conormal bundle is defined as the dual bundle to the normal bundle. It can be realised naturally as a sub-bundle of the cotangent bundle.
General definition
More abstractly, given an immersion (for instance an embedding), one can define a normal bundle of N in M, by at each point of N, taking the quotient space of the tangent space on M by the tangent space on N. For a Riemannian manifold one can identify this quotient with the orthogonal complement, but in general one cannot (such a choice is equivalent to a section of the projection ).
Thus the normal bundle is in general a quotient of the tangent bundle of the ambient space restricted to the subspace.
Formally, the normal bundle to N in M is a quotient bundle of the tangent bundle on M: one has the short exact sequence of vector bundles on N:
where is the restriction of the tangent bundle on M to N (properly, the pullback of the tangent bundle on M to a vector bundle on N via the map ). The fiber of the normal bundle in is referred to as the normal space at (of in ).
Conormal bundle
If is a smooth submanifold of a manifold , we can pick local coordinates around such that is locally defined by ; then with this choice of coordinates
and the ideal sheaf is locally generated by . Therefore we can define a non-degenerate pairing
that induces an isomorphism of sheaves . We can rephrase thi |
https://en.wikipedia.org/wiki/Glucagon-like%20peptide-2 | Glucagon-like peptide-2 (GLP-2) is a 33 amino acid peptide with the sequence HADGSFSDEMNTILDNLAARDFINWLIQTKITD (see Proteinogenic amino acid) in humans. GLP-2 is created by specific post-translational proteolytic cleavage of proglucagon in a process that also liberates the related glucagon-like peptide-1 (GLP-1). GLP-2 is produced by the intestinal endocrine L cell and by various neurons in the central nervous system. Intestinal GLP-2 is co-secreted along with GLP-1 upon nutrient ingestion.
When externally administered, GLP-2 produces a number of effects in humans and rodents, including intestinal growth, enhancement of intestinal function, reduction in bone breakdown and neuroprotection. GLP-2 may act in an endocrine fashion to link intestinal growth and metabolism with nutrient intake. GLP-2 and related analogs may be treatments for short bowel syndrome, Crohn's disease, osteoporosis and as adjuvant therapy during cancer chemotherapy.
GLP-2 has an antidepressant effect in a mouse model of depression when delivered via intracerebroventricular injection. However, a GLP-2 derivative (PAS-CPP-GLP-2) was shown to be efficiently delivered to the brain intranasally, with similar efficacy.
See also
Glucagon-like peptide 2 receptor |
https://en.wikipedia.org/wiki/MPS%20%28format%29 | MPS (Mathematical Programming System) is a file format for presenting and archiving linear programming (LP) and mixed integer programming problems.
Overview
The format was named after an early IBM LP product and has emerged as a de facto standard ASCII medium among most of the commercial LP solvers. Essentially all commercial LP solvers accept this format, and it is also accepted by the open-source COIN-OR system. Other software may require a customized reader routine in order to read MPS files. However, with the acceptance of algebraic modeling languages MPS usage has declined. For example, according to the NEOS server statistics in January 2011 less than 1% of submissions were in MPS form compared to 59.4% of AMPL and 29.7% of GAMS submissions.
MPS is column-oriented (as opposed to entering the model as equations), and all model components (variables, rows, etc.) receive names. MPS is an old format, so it is set up for punch cards: Fields start in column 2, 5, 15, 25, 40 and 50. Sections of an MPS file are marked by so-called header cards, which are distinguished by their starting in column 1. Although it is typical to use upper-case throughout the file for historical reasons, many MPS-readers will accept mixed-case for anything except the header cards, and some allow mixed-case anywhere. The names that you choose for the individual entities (constraints or variables) are not important to the solver; one should pick meaningful names, or easy names for a post-processing code to read.
MPS format
Here is a little sample model written in MPS format (explained in more detail below):
NAME TESTPROB
ROWS
N COST
L LIM1
G LIM2
E MYEQN
COLUMNS
XONE COST 1 LIM1 1
XONE LIM2 1
YTWO COST 4 LIM1 1
YTWO MYEQN -1
ZTHREE COST 9 LIM2 1
ZTHREE MYEQN 1
R |
https://en.wikipedia.org/wiki/FreeRADIUS | FreeRADIUS is a modular, high performance free RADIUS suite developed and distributed under the GNU General Public License, version 2, and is free for download and use. The FreeRADIUS Suite includes a RADIUS server, a BSD-licensed RADIUS client library, a PAM library, an Apache module, and numerous additional RADIUS related utilities and development libraries.
In most cases, the word "FreeRADIUS" refers to the free open-source RADIUS server from this suite.
FreeRADIUS is the most popular open source RADIUS server and the most widely deployed RADIUS server in the world. It supports all common authentication protocols, and the server comes with a PHP-based web user administration tool called dialupadmin. It is the basis for many commercial RADIUS products and services, such as embedded systems, RADIUS appliances that support Network Access Control, and WiMAX. It supplies the AAA needs of many Fortune-500 companies, telcos, and Tier 1 ISPs. It is also widely used in the academic community, including eduroam. The server is fast, feature-rich, modular, and scalable.
History
FreeRADIUS was started in August 1999 by Alan DeKok and Miquel van Smoorenburg. Miquel had previously written the Cistron RADIUS server, which had gained widespread usage once the Livingston server was no longer being maintained. FreeRADIUS was started to create a new RADIUS server, using a modular design that would encourage more active community involvement.
As of November 2014, the FreeRADIUS Project has three Core Team members: Alan DeKok (Project Leader), Arran Cudbard-Bell (Principal Architect), and Matthew Newton.
The latest major release is FreeRADIUS 3. FreeRADIUS 3 includes support for RADIUS over TLS, including RadSec, a completely rewritten rlm_ldap module, and hundreds of other minor consistency and usability enhancements. The latest mature version is maintained for stability rather than features.
The previous major release v2.2.x has entered the final phase of its lifecycle, and w |
https://en.wikipedia.org/wiki/Ion%20beam%20analysis | Ion beam analysis (IBA) is an important family of modern analytical techniques involving the use of MeV ion beams to probe the composition and obtain elemental depth profiles in the near-surface layer of solids. All IBA methods are highly sensitive and allow the detection of elements in the sub-monolayer range. The depth resolution is typically in the range of a few
nanometers to a few ten nanometers. Atomic depth resolution can be achieved, but requires special equipment. The analyzed depth ranges from a few ten nanometers to a few ten micrometers. IBA methods are always quantitative with an accuracy of a few percent.
Channeling allows to determine the depth profile of damage in single crystals.
RBS: Rutherford backscattering is sensitive to heavy elements in a light matrix
EBS: Elastic (non-Rutherford) backscattering spectrometry can be sensitive even to light elements in a heavy matrix. The term EBS is used when the incident particle is going so fast that it exceeds the "Coulomb barrier" of the target nucleus, which therefore cannot be treated by Rutherford's approximation of a point charge. In this case Schrödinger's equation should be solved to obtain the scattering cross-section (see http://www-nds.iaea.org/sigmacalc/ ).
ERD: Elastic recoil detection is sensitive to light elements in a heavy matrix
PIXE: Particle-induced X-ray emission gives the trace and minor elemental composition
NRA: Nuclear reaction analysis is sensitive to particular isotopes
Channelling: The fast ion beam can be aligned accurately with major axes of single crystals; then the strings of atoms "shadow" each other and the backscattering yield falls dramatically. Any atoms off their lattice sites will give visible extra scattering. Thus damage to the crystal is visible, and point defects (interstitials) can even be distinguished from dislocations.
The quantitative evaluation of IBA methods requires the use of specialized simulation and data analysis software. SIMNRA and DataFur |
https://en.wikipedia.org/wiki/Bleed%20screw | A bleed screw is a device used to create a temporary opening in an otherwise closed hydraulic system, which facilitates the removal of air or another substance from the system by way of pressure and density differences.
Applications
Domestic heating radiators
On a home radiator unit, the bleed screw can be opened, usually by means of a key, to allow unwanted air to escape from the unit. Bleed screws are also found on some pump types fulfilling a similar purpose.
They are most often located at the top of the radiator on the side of the inflow pipe. The screw itself, usually a hexagonal or square knob, is inside a small round protrusion.
The key looks similar to that used to wind a clock. It is inserted into the protrusion, mates with the bleed screw and turns it. Opening the bleed screw then allows air which has risen to the top of the radiator system (the top of the radiator itself) to escape and new water to take its place. Removing the air and allowing water to displace it makes the radiator work more efficiently since water conducts heat better than air.
A domestic radiator may need bleeding in this way either during a heating upgrade, power flushing or simply when the radiator feels colder at the top can indicate there is air in the radiator that needs to be released.
Engine cooling
Engine cooling systems can also have bleed screws. They can take the form of a bolt with a hole through the middle that is threaded into a hole on the engine's cylinder head. This hole goes into the water jacket of the cylinder head. In other designs, the bleed screw is placed in the uppermost hose which leads to the heater core, i.e. at the highest point of the cooling system.
When the bleed screw is loosened, antifreeze is added to the engine's cooling system and the increase in fluid pressure displaces air through the opened bleed screw. When liquid begins to flow out, all air has been removed from the system and the bleed screw is closed.
Bleed screws are not common o |
https://en.wikipedia.org/wiki/Ambient%20intelligence | Ambient intelligence (AmI) is a term used in computing to refer to electronic environments that are sensitive and responsive to the presence of people. The term is generally applied to consumer electronics, telecommunications, and computing.
Ambient intelligence is intended to enable devices to work in concert with people in carrying out their everyday life activities in an intuitive way by using information and intelligence hidden in the network connecting these devices. An example of ambient intelligence is the Internet of Things. A typical context of the ambient intelligence environment is home, but it may also be used in workspaces (offices, co-working), public spaces (based on technologies such as smart streetlights), and hospital environments.
The concept of ambient intelligence was originally developed in the late 1990s by Eli Zelkha and his team at Palo Alto Ventures for the time frame 2010–2020. Developers theorize that as devices grow smaller, more connected, and more integrated into our environment, the technological framework behind them will disappear into our surroundings until only the user interface remains perceivable by people.
Overview
The ambient intelligence concept builds upon pervasive computing, ubiquitous computing, profiling, context awareness, and human-centric computer interaction design. It is characterized by systems and technologies that are:
Embedded: Many networked devices are integrated into the environment.
Context aware: These devices can recognize you and your situational context.
Personalized: They can be tailored to your needs.
Adaptive: They can change in response to you.
Anticipatory: They can anticipate your desires without conscious mediation.
Successful implementation of ambient intelligence requires several vital technologies to exist. These include hidden, user-friendly hardware such as miniaturization, nanotechnology, and smart devices, as well as human-centric computer interfaces (intelligent agents, multim |
https://en.wikipedia.org/wiki/Pentagram%20%28video%20game%29 | Pentagram is a ZX Spectrum and MSX video game released by Ultimate Play the Game in 1986. It is the fourth in the Sabreman series, following on from his adventures in Sabre Wulf, Underwurlde and Knight Lore. Similarly to Knight Lore it uses the isometric Filmation engine. The game was written by either Tim and Chris Stamper or a U.S. Gold programming team.
Introduction
Typically, for an Ultimate release, the inlay card provides little actual instruction for playing the game, but includes a cryptic short story as an introduction.
This was Ultimate's way of describing the object of the game, which is to recover the lost Pentagram, an artifact of magical power. Firstly, Sabreman must locate one of the wells located in the maze of screens, shoot it several times with his spell and take the resultant bucket of water to one of the broken obelisks. When dropped on these, the water will "heal" the stone. This must be done with each of the four obelisks to make the titular Pentagram appear in one of the rooms. Once this is done, five magic runestones must be found and placed on the Pentagram itself.
Gameplay
Though the objective in Pentagram is more complex and obscure than the simple "find and fetch" gameplay of the two previous Filmation games Knight Lore and Alien 8, the gameplay is similar to those two titles. The main differences in this final revision of the Filmation engine are the new ability to shoot enemies with a projectile magic spell, and the ability of the enemies to respawn. The "directional control" system of the previous games was also removed because the Spectrum's single joystick button was now needed to fire Sabreman's spell, so it could no longer be used to jump (instead, "down" on the joystick is used to jump).
The basic gameplay is the same as that of Sabreman's previous outing Knight Lore (without that game's day/night shapeshifting cycle), as he wanders a mazelike system of screens filled with enemies, pieces of movable scenery (often forming o |
https://en.wikipedia.org/wiki/Econometric%20Society | The Econometric Society is an international society of academic economists interested in applying statistical tools in the practiced of econometrics. It is an independent organization with no connections to societies of professional mathematicians or statisticians.
It was founded on December 29, 1930, at the Statler Hotel in Cleveland, Ohio. Its first president was Irving Fisher.
As of 2014, there are about 700 Elected Fellows of the Econometric Society, making it one of the most prevalent research affiliations. New fellows are elected each year by the current fellows.
The sixteen founding members were Ragnar Frisch, Charles F. Roos, Joseph A. Schumpeter, Harold Hotelling, Henry Schultz, Karl Menger, Edwin B. Wilson, Frederick C. Mills, William F. Ogburn, J. Harvey Rogers, Malcolm C. Rorty, Carl Snyder, Walter A. Shewhart, Øystein Ore, Ingvar Wedervang and Norbert Wiener.
The Econometric Society sponsors the Economics academic journal Econometrica and publishes the journals Theoretical Economics and Quantitative Economics.
Officers
The Econometric Society is led by a president, who serves a one-year term. Election as a Fellow of the Econometric Society is considered by much of the economics profession to be an honor.
List of presidents of the Econometric Society
List of fellows of the Econometric Society
Honorary lectures
The Econometric Society sponsors several annual awards, in which the honored member delivers a lecture:
Frisch Medal
Walras–Bowley Lecture
Fisher–Schultz Lecture
Jacob Marschak Lecture |
https://en.wikipedia.org/wiki/Surroundings | Surroundings are the area around a given physical or geographical point or place. The exact definition depends on the field. Surroundings can also be used in geography (when it is more precisely known as vicinity, or vicinage) and mathematics, as well as philosophy, with the literal or metaphorically extended definition.
In thermodynamics, the term (and its synonym, environment) is used in a more restricted sense, meaning everything outside the thermodynamic system. Often, the simplifying assumptions are that energy and matter may move freely within the surroundings, and that the surroundings have a uniform composition.
See also
Distance
Environment (biophysical)
Environment (systems)
Neighbourhood (mathematics)
Social environment
Proxemics
Geography
Thermodynamics |
https://en.wikipedia.org/wiki/Different%20ideal | In algebraic number theory, the different ideal (sometimes simply the different) is defined to measure the (possible) lack of duality in the ring of integers of an algebraic number field K, with respect to the field trace. It then encodes the ramification data for prime ideals of the ring of integers. It was introduced by Richard Dedekind in 1882.
Definition
If OK is the ring of integers of K, and tr denotes the field trace from K to the rational number field Q, then
is an integral quadratic form on OK. Its discriminant as quadratic form need not be +1 (in fact this happens only for the case K = Q). Define the inverse different or codifferent or Dedekind's complementary module as the set I of x ∈ K such that tr(xy) is an integer for all y in OK, then I is a fractional ideal of K containing OK. By definition, the different ideal δK is the inverse fractional ideal I−1: it is an ideal of OK.
The ideal norm of δK is equal to the ideal of Z generated by the field discriminant DK of K.
The different of an element α of K with minimal polynomial f is defined to be δ(α) = f′(α) if α generates the field K (and zero otherwise): we may write
where the α(i) run over all the roots of the characteristic polynomial of α other than α itself. The different ideal is generated by the differents of all integers α in OK. This is Dedekind's original definition.
The different is also defined for a finite degree extension of local fields. It plays a basic role in Pontryagin duality for p-adic fields.
Relative different
The relative different δL / K is defined in a similar manner for an extension of number fields L / K. The relative norm of the relative different is then equal to the relative discriminant ΔL / K. In a tower of fields L / K / F the relative differents are related by δL / F = δL / KδK / F.
The relative different equals the annihilator of the relative Kähler differential module :
The ideal class of the relative different δL / K is always a square in the class gr |
https://en.wikipedia.org/wiki/Takeuti%27s%20conjecture | In mathematics, Takeuti's conjecture is the conjecture of Gaisi Takeuti that a sequent formalisation of second-order logic has cut-elimination (Takeuti 1953). It was settled positively:
By Tait, using a semantic technique for proving cut-elimination, based on work by Schütte (Tait 1966);
Independently by Prawitz (Prawitz 1968) and Takahashi (Takahashi 1967) by a similar technique (Takahashi 1967) - although Prawitz's and Takahashi's proofs are not limited to second-order logic, but concern higher-order logics in general;
It is a corollary of Jean-Yves Girard's syntactic proof of strong normalization for System F.
Takeuti's conjecture is equivalent to the 1-consistency of second-order arithmetic in the sense that each of the statements can be derived from each other in the weak system PRA. It is also equivalent to the strong normalization of the Girard/Reynold's System F.
See also
Hilbert's second problem |
https://en.wikipedia.org/wiki/ISO/IEC%2011179 | The ISO/IEC 11179 Metadata Registry (MDR) standard is an international ISO/IEC standard for representing metadata for an organization in a metadata registry. It documents the standardization and registration of metadata to make data understandable and shareable.
Intended purpose
Organizations exchange data between computer systems precisely using enterprise application integration technologies. Completed transactions are often transferred to separate data warehouse and business rules systems with structures designed to support data for analysis. A de facto standard model for data integration platforms is the Common Warehouse Metamodel (CWM). Data integration is often also solved as a problem of data, rather than metadata, with the use of so-called master data. ISO/IEC 11179 claims that it is a standard for metadata-driven exchange of data in an heterogeneous environment, based on exact definitions of data.
Structure of an ISO/IEC 11179 metadata registry
The ISO/IEC 11179 model is a result of two principles of semantic theory, combined with basic principles of data modelling.
The first principle from semantic theory is the thesaurus type relation between wider and more narrow (or specific) concepts, e.g. the wide concept "income" has a relation to the more narrow concept "net income".
The second principle from semantic theory is the relation between a concept and its representation, e.g., "buy" and "purchase" are the same concept although different terms are used.
A basic principle of data modelling is the combination of an object class and a characteristic. For example, "Person - hair color".
When applied to data modelling, ISO/IEC 11179 combines a wide "concept" with an "object class" to form a more specific "data element concept". For example, the high-level concept "income" is combined with the object class "person" to form the data element concept "net income of person". Note that "net income" is more specific than "income".
The different possible repre |
https://en.wikipedia.org/wiki/Roy%20Kerr | Roy Patrick Kerr (; born 16 May 1934) is a New Zealand mathematician who discovered the Kerr geometry, an exact solution to the Einstein field equation of general relativity. His solution models the gravitational field outside an uncharged rotating massive object, including a rotating black hole. His solution to Einstein's equations predicted spinning black holes before they were discovered.
Early life and education
Kerr was born in 1934 in Kurow, New Zealand. He was born into a dysfunctional family, and his mother was forced to leave when he was three. When his father went to war, he was sent to a farm. After his father's return from war, they moved to Christchurch. He was accepted to St Andrew's College, a private school, as his father had served under a former headmaster. Kerr's mathematical talent was first recognised while he was still a student at St Andrew's College. Although there was no mathematics teacher there at the time, he was able in 1951 to go straight into the third year of mathematics at Canterbury University College, a constituent of the University of New Zealand and the precursor to the University of Canterbury. Their regulations did not permit him to graduate until 1954 and so it was not until September 1955 that he moved to the University of Cambridge, where he earned his PhD in 1959. His dissertation concerned the equations of motion in general relativity.
Career and research
After a postdoctoral fellowship at Syracuse University, where Einstein's collaborator Peter Bergmann was a professor, he spent some time working for the United States Air Force at Wright-Patterson Air Force Base. Kerr speculated that the "main reason why the US Air Force had created a General Relativity section was probably to show the U.S. Navy that they could also do pure research."
Work at Texas and Canterbury
In 1962, Kerr joined Alfred Schild and his Relativity Group at the University of Texas at Austin. As Kerr wrote in 2009:
By the summer of 1963, Maarten Schmi |
https://en.wikipedia.org/wiki/Bioorthogonal%20chemical%20reporter | In chemical biology, bioorthogonal chemical reporter is a non-native chemical functionality that is introduced into the naturally occurring biomolecules of a living system, generally through metabolic or protein engineering. These functional groups are subsequently utilized for tagging and visualizing biomolecules. Jennifer Prescher and Carolyn R. Bertozzi, the developers of bioorthogonal chemistry, defined bioorthogonal chemical reporters as "non-native, non-perturbing chemical handles that can be modified in living systems through highly selective reactions with exogenously delivered probes." It has been used to enrich proteins and to conduct proteomic analysis.
In the early development of the technique, chemical motifs have to fulfill criteria of biocompatibility and selective reactivity in order to qualify as bioorthogonal chemical reporters. Some combinations of proteinogenic amino acid side chains meet the criteria, as do ketone and aldehyde tags. Azides and alkynes are other examples of chemical reporters.
A bioorthogonal chemical reporter must be incorporated into a biomolecule. This occurs via metabolism. The chemical reporter is linked to a substrate, which a cell can metabolize. |
https://en.wikipedia.org/wiki/Polymer%20science | Polymer science or macromolecular science is a subfield of materials science concerned with polymers, primarily synthetic polymers such as plastics and elastomers. The field of polymer science includes researchers in multiple disciplines including chemistry, physics, and engineering.
Subdisciplines
This science comprises three main sub-disciplines:
Polymer chemistry or macromolecular chemistry is concerned with the chemical synthesis and chemical properties of polymers.
Polymer physics is concerned with the physical properties of polymer materials and engineering applications. Specifically, it seeks to present the mechanical, thermal, electronic and optical properties of polymers with respect to the underlying physics governing a polymer microstructure. Despite originating as an application of statistical physics to chain structures, polymer physics has now evolved into a discipline in its own right.
Polymer characterization is concerned with the analysis of chemical structure, morphology, and the determination of physical properties in relation to compositional and structural parameters.
History of polymer science
The first modern example of polymer science is Henri Braconnot's work in the 1830s. Henri, along with Christian Schönbein and others, developed derivatives of the natural polymer cellulose, producing new, semi-synthetic materials, such as celluloid and cellulose acetate. The term "polymer" was coined in 1833 by Jöns Jakob Berzelius, though Berzelius did little that would be considered polymer science in the modern sense. In the 1840s, Friedrich Ludersdorf and Nathaniel Hayward independently discovered that adding sulfur to raw natural rubber (polyisoprene) helped prevent the material from becoming sticky. In 1844 Charles Goodyear received a U.S. patent for vulcanizing natural rubber with sulfur and heat. Thomas Hancock had received a patent for the same process in the UK the year before. This process strengthened natural rubber and prevented it |
https://en.wikipedia.org/wiki/Gabriel%20Carroll | Gabriel Drew Carroll (born December 24, 1982) is a Professor of Economics at the University of Toronto. He was born to tech industry worker parents in Oakland. He graduated from Harvard University with B.A. in mathematics and linguistics in 2005 and received his doctorate in economics from MIT in 2012. He was recognized as a child prodigy and received numerous awards in mathematics while a student.
Carroll won two gold medals (1998, 2001) and a silver medal (1999) at the International Mathematical Olympiad, earning a perfect score at the 2001 International Mathematical Olympiad held in Washington, D.C., shared only with American teammate Reid W. Barton and Chinese teammates Liang Xiao and Zhiqiang Zhang.
Gabriel earned a place among the top five ranked competitors (who are themselves not ranked against each other) in the William Lowell Putnam Competition all four years that he was eligible (2000–2003), a feat matched by only seven others (Don Coppersmith (1968–1971), Arthur Rubin (1970–1973), Bjorn Poonen (1985–1988), Ravi Vakil (1988–1991), Reid W. Barton (2001–2004), Daniel Kane (2003–2006), and Brian R. Lawrence (2007–08, 2010–11). His top-5 performance in 2000 was particularly notable, as he was officially taking the exam in spite of only being a high school senior, thus forfeiting one of his years of eligibility in college. He was on the first place Putnam team twice (2001–02) and the second place team once (2003).
He has earned awards in science and math, including the Intel Science Talent Search, has taught mathematics classes and tutorials, and plays the piano. He was a Research Science Institute scholar in 2000.
Carroll proposed Problem 3 of IMO 2009 and Problem 3 of IMO 2010. He also proposes problems to the USAMO such as problem 3 in 2007, 2008, 2010 and problem 6 in 2009.
During the 2005–06 academic year, he taught English in Chaling, Hunan, China. He worked at the National Bureau of Economic Research from 2006 to 2007 and was an Assistant Professor |
https://en.wikipedia.org/wiki/E-plane%20and%20H-plane | The E-plane and H-plane are reference planes for linearly polarized waveguides, antennas and other microwave devices.
In waveguide systems, as in the electric circuits, it is often desirable to be able to split the circuit power into two or more fractions. In a waveguide system, an element called a junction is used for power division.
In a low frequency electrical network, it is possible to combine circuit elements in series or in parallel, thereby dividing the source power among several circuit components.
In microwave circuits, a waveguide with three independent ports is called a TEE junction. The output of E-Plane Tee is 180° out of phase where the output of H-plane Tee is in phase.
E-Plane
For a linearly-polarized antenna, this is the plane containing the electric field vector (sometimes called the E aperture) and the direction of maximum radiation. The electric field or "E" plane determines the polarization or orientation of the radio wave. For a vertically polarized antenna, the E-plane usually coincides with the vertical/elevation plane. For a horizontally polarized antenna, the E-Plane usually coincides with the horizontal/azimuth plane.
E- plane and H-plane should be 90 degrees apart.
H-plane
In the case of the same linearly polarized antenna, this is the plane containing the magnetic field vector (sometimes called the H aperture) and the direction of maximum radiation. The magnetizing field or "H" plane lies at a right angle to the "E" plane. For a vertically polarized antenna, the H-plane usually coincides with the horizontal/azimuth plane. For a horizontally polarized antenna, the H-plane usually coincides with the vertical/elevation plane.
Illustrations
Co- and cross-polarizations
Co-polarization (co-pol) and cross-polarization (cross-pol) are defined for the radiating E and H planes. These directions are defined in spherical coordinates corresponding to the spherical wavefronts of the propagating wave. By convention, the co-pol direction |
https://en.wikipedia.org/wiki/CYPRIS%20%28microchip%29 | CYPRIS (cryptographic RISC microprocessor) was a cryptographic processor developed by the Lockheed Martin Advanced Technology Laboratories. The device was designed to implement NSA encryption algorithms and had a similar intent to the AIM and Sierra crypto modules. However, the principal references date back to the late 1990s and it does not appear that the CYPRIS ever earned NSA's Type 1 certification, without which it could not be used to protect classified government traffic.
According to a manufacturer presentation, |
https://en.wikipedia.org/wiki/Modulated%20continuous%20wave | Modulated continuous wave (MCW) is Morse code telegraphy, transmitted using an audio tone to modulate a carrier wave.
The Federal Communications Commission defines modulated continuous wave in 47 CFR §97.3(c)(4) as "Tone-modulated international Morse code telegraphy emissions having designators with A, C, D, F, G, H or R as the first symbol; 2 as the second symbol; A or B as the third symbol." See Types of radio emissions for a general explanation of these symbols.
Types of Morse code radio transmissions (CW and MCW) discussed in this article include:
A1A and A2A — Double-sideband amplitude modulation (AM); One channel containing digital information, no subcarrier (A1A) or using a subcarrier (A2A); Aural telegraphy (intended to be decoded by ear)
F2A — Frequency modulation (FM); One channel containing digital information, using a subcarrier; Aural telegraphy
J2A and J2B — Single-sideband with suppressed carrier; One channel containing digital information, using a subcarrier; Aural telegraphy (J2A) or Electronic telegraphy (intended to be decoded by machine) (J2B)
Unlike A1A CW transmissions, A2A MCW will produce an audible audio tone from an AM radio receiver that is not equipped with a beat oscillator. MCW is commonly used by RDF beacons to transmit the station identifier.
F2A MCW Morse can be heard on a normal FM radio receiver, and it is commonly used by both commercial and amateur repeater stations for identification. Also, F2A is sometimes used by other types of stations operating under automatic control, such as a telemetry transmitter or a remote base station.
MCW can be generated by any AM or FM radio transmitter with audio input from an audio oscillator or equivalent audio source. When an SSB transmitter is modulated by Morse code of only a single audio frequency, the resulting radio frequency emission is J2A or J2B and therefore is CW by definition, not MCW.
Within the United States, MCW transmission is not permitted to amateur radio operators in |
https://en.wikipedia.org/wiki/Evolutionary%20capacitance | Evolutionary capacitance is the storage and release of variation, just as electric capacitors store and release charge. Living systems are robust to mutations. This means that living systems accumulate genetic variation without the variation having a phenotypic effect. But when the system is disturbed (perhaps by stress), robustness breaks down, and the variation has phenotypic effects and is subject to the full force of natural selection. An evolutionary capacitor is a molecular switch mechanism that can "toggle" genetic variation between hidden and revealed states. If some subset of newly revealed variation is adaptive, it becomes fixed by genetic assimilation. After that, the rest of variation, most of which is presumably deleterious, can be switched off, leaving the population with a newly evolved advantageous trait, but no long-term handicap. For evolutionary capacitance to increase evolvability in this way, the switching rate should not be faster than the timescale of genetic assimilation.
This mechanism would allow for rapid adaptation to new environmental conditions. Switching rates may be a function of stress, making genetic variation more likely to affect the phenotype at times when it is most likely to be useful for adaptation. In addition, strongly deleterious variation may be purged while in a partially cryptic state, so cryptic variation that remains is more likely to be adaptive than random mutations are. Capacitance can help cross "valleys" in the fitness landscape, where a combination of two mutations would be beneficial, even though each is deleterious on its own.
There is currently no consensus about the extent to which capacitance might contribute to evolution in natural populations. The possibility of evolutionary capacitance is considered to be part of the extended evolutionary synthesis.
Switches that turn robustness to phenotypic rather than genetic variation on and off do not fit the capacitance analogy, as their presence does not cause |
https://en.wikipedia.org/wiki/KASY-TV | KASY-TV (channel 50) is a television station in Albuquerque, New Mexico, United States, affiliated with MyNetworkTV. It is owned by Mission Broadcasting alongside Santa Fe–licensed CW affiliate KWBQ (channel 19) and its Roswell-based satellite, KRWB-TV (channel 21). The two stations share studios with dual CBS/Fox affiliate KRQE (channel 13) on Broadcast Plaza in Albuquerque; KASY-TV's transmitter is located atop Sandia Crest.
Nexstar Media Group, which owns KRQE and holds a majority stake in The CW, provides master control, technical, engineering and accounting services for KASY-TV and KWBQ through a shared services agreement (SSA), though the two stations are otherwise operated separately from KRQE as Mission handles programming, advertising sales and retransmission consent negotiations.
History
KASY-TV first signed on the air on October 6, 1995, owned by Ramar Communications and managed by Lee Enterprises (then-owners of CBS affiliate KRQE) under a local marketing agreement (LMA). The station was primarily a UPN affiliate, but had a secondary affiliation with The WB; this was easy to do as neither network had more than a couple nights a week of programming at that time. Initially, KASY ran cartoons, old movies, talk shows, and classic and recent off-network sitcoms. In fall 1997, KASY dropped WB programming and became an exclusive UPN affiliate; The WB would return to the market when upstart KWBQ signed on in March 1999 with a similar general entertainment format. In the interim, WB programming was brought in out-of-market from KTLA in Los Angeles or Chicago-based superstation WGN on Albuquerque area cable providers.
In June 1999, ACME Communications, KWBQ's owner, bought KASY from Ramar and terminated the local marketing agreement with Lee Enterprises, resulting in the creation of the first major television duopoly in the Albuquerque market. Most of the programming inventory airing on KASY was also acquired by ACME, while some of the shows that aired on KASY |
https://en.wikipedia.org/wiki/IBM%20Global%20Mirror | Global Mirror is an IBM technology that provides data replication over extended distances between two sites for business continuity and disaster recovery. If adequate bandwidth exists, Global Mirror provides a recovery point objective (RPO) of as low as 3–5 seconds between the two sites at extended distances with no performance impact on the application at the primary site. It replicates the data asynchronously and also forms a consistency group at a regular interval allowing a clean recovery of the application.
The two sites can be on separate continents or simply on different utility grids. IBM also provides a synchronous data replication called Metro Mirror, which is designed to support replication at "Metropolitan" distances of (normally) less than 300 km.
Global Mirror is based on IBM Copy Services functions: Global Copy and FlashCopy. Global mirror periodically pauses updates of the primary volumes and swaps change recording bitmaps. It then uses the previous bitmap to drain updates from the primary volumes to the secondaries. After all primary updates have been drained, the secondary volumes are used as the source for a FlashCopy to tertiary volumes at the recovery site. This ensures that the tertiary copy of the volumes has point-in-time consistency. By grouping many volumes into one Global Mirror session multiple volumes may be copied to the recovery site simultaneously while maintaining point-in-time consistency across those volumes.
Global Mirror can be combined with a wide area network clustering product like Geographically Dispersed Parallel Sysplex (GDPS), HACMP/XD, or IBM TotalStorage Continuous Availability for Windows to provide for automated failover between sites. This combined solution provides lower recovery time objective (RTO), because it allows most applications to automatically resume productive operation in 30–600 seconds.
The Global Mirror function is available on IBM Storage devices including the DS8000 series (DS8100, DS8300, DS8700 |
https://en.wikipedia.org/wiki/Neurofibrillary%20tangle | Neurofibrillary tangles (NFTs) are aggregates of hyperphosphorylated tau protein that are most commonly known as a primary biomarker of Alzheimer's disease. Their presence is also found in numerous other diseases known as tauopathies. Little is known about their exact relationship to the different pathologies.
Formation
Neurofibrillary tangles are formed by hyperphosphorylation of a microtubule-associated protein known as tau, causing it to aggregate, or group, in an insoluble form. (These aggregations of hyperphosphorylated tau protein are also referred to as PHF, or "paired helical filaments"). The precise mechanism of tangle formation is not completely understood, though it is typically recognized that tangles are a primary causative factor in neurodegenerative disease.
Cytoskeletal changes
Three different maturation states of NFT have been defined using anti-tau and anti-ubiquitin immunostaining. At stage 0 there are morphologically normal pyramidal cells showing diffuse or fine granular cytoplasmic staining with anti-tau. In other words, cells are healthy with minimal tau presence; at stage 1 some delicate elongate inclusions are stained by tau antibodies (these are early tangles); stage 2 is represented by the classic NFT demonstration with anti-tau staining; stage 3 is exemplified by ghost tangles (tangles outside of cells where the host neuron has died), which are characterized by a reduced anti-tau but marked anti-ubiquitin immunostaining.
Causes
Mutated tau
The traditional understanding is that tau binds to microtubules and assists with their self-assembly, formation and stabilization. However, when tau is hyperphosphorylated, it is unable to bind and the microtubules become unstable and begin disintegrating. The unbound tau clumps together in formations called neurofibrillary tangles. More explicitly, intracellular lesions known as pretangles develop when tau is phosphorylated excessively and on improper amino acid residues. These lesions, over time, |
https://en.wikipedia.org/wiki/Minimalism%20%28computing%29 | In computing, minimalism refers to the application of minimalist philosophies and principles in the design and use of hardware and software. Minimalism, in this sense, means designing systems that use the least hardware and software resources possible.
History
In the late 1970s and early 1980s, programmers worked within the confines of relatively expensive and limited resources of common platforms. Eight or sixteen kilobytes of RAM was common; 64 kilobytes was considered a vast amount and was the entire address space accessible to the 8-bit CPUs predominant during the earliest generations of personal computers. The most common storage medium was the 5.25 inch floppy disk holding from 88 to 170 kilobytes. Hard drives with capacities from five to ten megabytes cost thousands of dollars.
Over time, personal-computer memory capacities expanded by orders of magnitude and mainstream programmers took advantage of the added storage to increase their software's capabilities and to make development easier by using higher-level languages. By contrast, system requirements for legacy software remained the same. As a result, even the most elaborate, feature-rich programs of yesteryear seem minimalist in comparison with current software.
One example of a program whose system requirements once gave it a heavyweight reputation is the GNU Emacs text editor, which gained the backronym "Eight Megabytes And Constantly Swapping" in an era when 8 megabytes was a lot of RAM. Today, Emacs' mainly textual buffer-based paradigm uses far fewer resources than desktop metaphor GUI IDEs with comparable features such as Eclipse or Netbeans. In a speech at the 2002 International Lisp Conference, Richard Stallman indicated that minimalism was a concern in his development of GNU and Emacs, based on his experiences with Lisp and system specifications of low-end minicomputers at the time.
As the capabilities and system requirements of common desktop software and operating systems grew throughout th |
https://en.wikipedia.org/wiki/Hydraulic%20manifold | A hydraulic manifold is a component that regulates fluid flow between pumps and actuators and other components in a hydraulic system. It is like a switchboard in an electrical circuit because it lets the operator control how much fluid flows between which components of a hydraulic machinery. For example, in a backhoe loader a manifold turns on or shuts off or diverts flow to the telescopic arms of the front bucket and the back bucket. The manifold is connected to the levers in the operator's cabin which the operator uses to achieve the desired manifold behaviour.
A manifold is composed of assorted hydraulic valves connected to each other. It is the various combinations of states of these valves that allow complex control behaviour in a manifold.
A hydraulic manifold is a block of metal with flow paths drilled through it, connecting various ports. Hydraulic manifolds consist of one or more relative large pipes called a "barrel" or "main", with numerous junctions connecting smaller pipes and ports.
See also
Block and bleed manifold |
https://en.wikipedia.org/wiki/KSCW-DT | KSCW-DT (channel 33) is a television station in Wichita, Kansas, United States, affiliated with The CW. It is owned by Gray Television alongside Hutchinson-licensed CBS affiliate KWCH-DT (channel 12). Both stations share studios on 37th Street in northeast Wichita, while KSCW-DT's transmitter is located in rural northeastern Reno County (east of Hutchinson).
KSCW-DT also operates a digital replacement translator on UHF channel 33 (its previous analog signal allotment) from a transmitter in North Wichita, just north of the station's studio facility.
History
The station was first licensed on June 8, 1988, under permits from LIN TV filing an application with the Federal Communications Commission (FCC), under the call letters KWCV. It was temporarily licensed as DKWCV on November 5, 1998, but on February 11, 1999, it was changed back to KWCV.
The station first signed on the air on August 5, 1999, along with LIN TV (which also briefly owned KAKE and its satellites in 2000 before selling KAKE to Benedek Broadcasting) forming and owning 50% of Banks Broadcasting, which would become the station's owner. Originally operating as a WB affiliate, it was branded on-air as "Kansas' WB". Prior to the station's launch, The WB's programming could only be viewed in the Wichita market through Chicago-based cable superstation WGN, which carried the network's programming nationwide from The WB's January 1995 launch until October 1999, or Denver's KWGN-TV on cable or satellite. The station's original transmitter was located on a tower near Colwich.
On January 24, 2006, the Warner Bros. unit of Time Warner and CBS Corporation announced that the two companies would shut down The WB and UPN and combine the networks' respective programming onto a newly created "fifth" network called The CW. One month later on February 22, 2006, News Corporation announced that it would launch another new network called MyNetworkTV. On March 21, not long after it was announced that the station would becom |
https://en.wikipedia.org/wiki/KMTW | KMTW (channel 36) is a television station licensed to Hutchinson, Kansas, United States, serving the Wichita area as an affiliate of the digital multicast network Dabl. It is owned by the Mercury Broadcasting Company, which maintains a local marketing agreement (LMA) with Sinclair Broadcast Group, owner of dual Fox/MyNetworkTV affiliate KSAS-TV (channel 24), for the provision of certain services. Both stations share studios on North West Street in northwestern Wichita, while KMTW's transmitter is located in rural southwestern Harvey County (on the town limits of Halstead).
History
On June 27, 1997, Clear Channel Communications (owner of Fox affiliate KSAS-TV (channel 24)) entered into a local marketing agreement with Goddard-based Three Feathers Communications, Inc. to form a new television station in Hutchinson, Kansas. Initially bearing the name KAWJ, the construction permit took the KSCC ("Kansas Clear Channel") call letters on October 9, 1998.
On July 30, 1999, Three Feathers filed an application to sell the license of KSCC to Viacom's Paramount Stations Group, with the application being granted by the Federal Communications Commission (FCC) on October 1 the same year.
The station first signed on the air on January 5, 2001 (the station first appeared on Cox Cable starting in August 2000) affiliating with UPN as an owned-and-operated station, a rarity for a market of Wichita's size. However, just prior to the station's sign-on, its license assets were sold to San Antonio–based Mercury Broadcasting Company. In June 2001, Mercury Broadcasting would take over ownership of KSCC. Prior to the station's sign-on, UPN programming was seen on a secondary basis on sister station KSAS-TV.
In 2003, Clear Channel attempted to buy the station outright, but was denied a "failing station" waiver by the FCC. This special approval for the sale was necessary because the Wichita–Hutchinson designated market area has only seven "unique" full-power television stations. The full-po |
https://en.wikipedia.org/wiki/Rupert%20Riedl | Rupert Riedl (22 February 1925 – 18 September 2005) was an Austrian zoologist.
Biography
Riedl was a scientist with broad interests, whose influence in epistemology grounded in evolutionary theory was notable, although less in English-speaking circles than in German or even Spanish speaking ones. His 1984 work, Biology of Knowledge: The evolutionary basis of reason examined cognitive abilities and the increasing complexity of biological diversification over the immense periods of evolutionary time.
Riedl built upon the work of the Viennese school of thought initially typified by Konrad Lorenz, and continued in Vienna by Gerhard Vollmer, Franz Wuketits, and in Spain by Nicanor Ursura. Riedl was skeptical of German idealism, and nourished by the tradition that produced the scientists and philosophers of science Ernst Mach, Ludwig Boltzmann, Erwin Schrödinger, Karl Popper, Hans Reichenbach and Sigmund Freud.
Lorenz believed that the Kantian framework of cognitive concepts such as three-dimensional space and time were not fixed but built up over phylogenetic history, potentially subject to further developments. Lorenz’s position, as expanded by Riedl, attempted to make it easier to assimilate non-common sense areas of physics such as quantum field theory and string theory.
Riedl drew clear distinctions between the deductive and inductive (non conscious) cognitive processes characteristic of the left and right cerebral hemispheres. His analysis of what he called "the pitfalls of reason" deserves special attention. He, like Lorenz, was concerned with cognitive processes that might endanger the future of civilization.
Riedl had less direct influence on academic philosophy than his profound influence on the thinking of investigators in neuroscience such as Michael Gazzaniga, Antonio Damasio, and Vilayanur S. Ramachandran, whose investigations combine synergistically with those of more physiologically-oriented scientists such as Eric Kandel and Rodolfo Llinás, as well a |
https://en.wikipedia.org/wiki/Cookie%20dough | Cookie dough is an uncooked blend of cookie ingredients. While cookie dough is normally intended to be baked into individual cookies before eating, edible cookie dough is made to be eaten as is, and usually is made without eggs to make it safer for human consumption.
Cookie dough can be made at home or bought pre-made in packs (frozen logs, buckets, etc.). Dessert products containing cookie dough include ice cream and candy. In addition, pre-made cookie dough is sold in different flavors.
When made at home, common ingredients include flour, butter, white sugar, salt, vanilla extract, and eggs. If the dough is made with the intention of baking, then leavening agents such as baking soda or baking powder are added. However, these are often excluded in cookie doughs that are designed to be eaten raw. Chocolate chip cookie dough is a popular variation that can be made by adding chocolate chips to the mix.
History
Cookie dough is derived from the creation of cookies that dates back as far as 7th century Persia, where they were used as test cakes. Persia was one of the first countries to use sugar and soon became known for luxurious cakes and pastries. The early cookie was first labelled as a test cake before it was referred to as a "cookie" because the Persians would bake a small amount of cake batter in the oven to test the oven temperature, and it would come out looking like a mini cake. The concept of cookies spread and became known worldwide. They evolved into Biscuits for convenience as they were easier to keep fresh for a longer period and were simple to carry for travel.
Cookies became established in Europe sometime between the 17th and 18th century, as baking gained popularity. At that time the word "cookie" was first used. The term comes from the Dutch language where Koekje means "small or little cake". During the ensuing Industrial Revolution, more cookie recipes became available. New forms and flavors of cookies continue to be created, one of which is th |
https://en.wikipedia.org/wiki/Nuclear%20lightbulb | A nuclear lightbulb is a hypothetical type of spacecraft engine using a gaseous fission reactor to achieve nuclear propulsion. Specifically it would be a type of gas core reactor rocket that uses a quartz wall to separate nuclear fuel from coolant/propellant. It would be operated at temperatures of up to 22,000°C where the vast majority of the electromagnetic emissions would be in the hard ultraviolet range. Fused silica is almost completely transparent to this light, so it would be used to contain the uranium hexafluoride and allow the light to heat reaction mass in a rocket or to generate electricity using a heat engine or photovoltaics.
This type of reactor shows great promise in both of these roles.
Rocket engine
As a rocket engine it, like all nuclear rocket designs, can greatly exceed the exhaust speed and specific impulse of a chemical rocket. However, it also does not involve the release of any radioactive material from the rocket, unlike open cycle designs which would cause nuclear fallout if used in a planetary atmosphere (e.g. Project Orion). The theoretical specific impulse (Isp) range from 1500 to 3000 seconds.
Electrical power generation
As a method to generate electricity, nuclear lightbulbs are extremely efficient because higher-temperature heat contains more Gibbs free energy than the low-temperature heat produced in current fossil-fuel plants and water-cooled nuclear reactors. |
https://en.wikipedia.org/wiki/Personal%20Handy-phone%20System | The Personal Handy-phone System (PHS), also marketed as the Personal Communication Telephone (PCT) in Thailand, and the Personal Access System (PAS) and commercially branded as Xiaolingtong () in Mainland China, was a mobile network system operating in the 1880–1930 MHz frequency band, used mainly in Japan, China, Taiwan, and some other Asian countries and regions.
Outline
Technology
PHS is essentially a cordless telephone like DECT, with the capability to handover from one cell to another. PHS cells are small, with transmission power of base station a maximum of 500 mW and range typically measures in tens or at most hundreds of metres (some can range up to about 2 kilometres in line-of-sight), contrary to the multi-kilometre ranges of CDMA and GSM. This makes PHS suitable for dense urban areas, but impractical for rural areas, and the small cell size also makes it difficult if not impossible to make calls from rapidly moving vehicles.
PHS uses TDMA/TDD for its radio channel access method, and 32 kbit/s ADPCM for its voice codec. Modern PHS phone can also support many value-added services such as high speed wireless data/Internet connection (64 kbit/s and higher), WWW access, e-mailing, and text messaging.
PHS technology is also a popular option for providing a wireless local loop, where it is used for bridging the "last mile" gap between the POTS network and the subscriber's home. It was developed under the concept of providing a wireless front-end of an ISDN network. Thus a PHS base station is compatible with ISDN and is often connected directly to ISDN telephone exchange equipment e.g. a digital switch.
In spite of its low-cost base station, micro-cellular system and "Dynamic Cell Assignment" system, PHS offers higher number-of-digits frequency use efficiency with lower cost (throughput per area basis), compared with typical 3G cellular telephone systems. It enables flat-rate wireless service such as AIR-EDGE, throughout Japan.
The speed of an AIR-EDGE |
https://en.wikipedia.org/wiki/Mathematics%20education%20in%20New%20York | Mathematics education in New York in regard to both content and teaching method can vary depending on the type of school a person attends. Private school math education varies between schools whereas New York has statewide public school requirements where standardized tests are used to determine if the teaching method and educator are effective in transmitting content to the students. While an individual private school can choose the content and educational method to use, New York State mandates content and methods statewide. Some public schools have and continue to use established methods, such as Montessori for teaching such required content. New York State has used various foci of content and methods of teaching math including New Math (1960s), 'back to the basics' (1970s), Whole Math (1990s), Integrated Math, and Everyday Mathematics.
How to teach math, what to teach, and its effectiveness has been a topic of debate in New York State and nationally since the "Math Wars" started in the 1960s. Often, current political events influence how and what is taught. The politics in turn influence state legislation. California, New York, and several other states have influenced textbook content produced by publishers.
The state of New York has implemented a novel curriculum for high school mathematics.
The courses Algebra I, Geometry, and Algebra II/Trigonometry are required courses mandated by the New York State Department of Education for high school graduation.
2007-present
Algebra
This is the first course in the new three-year curriculum. It was originally "Math A," but was replaced with "Integrated Algebra." In 2009 when Common Core was adopted, "Algebra I" replaced "Integrated Algebra" and is still in use today.
Students learn to how write, solve, and graph equations and inequalities. They will also learn how to solve systems of equations, as well as how to simplify exponents, quadratic equations, exponential functions, polynomials, radicals, and rational expr |
https://en.wikipedia.org/wiki/Species%E2%80%93area%20relationship | The species–area relationship or species–area curve describes the relationship between the area of a habitat, or of part of a habitat, and the number of species found within that area. Larger areas tend to contain larger numbers of species, and empirically, the relative numbers seem to follow systematic mathematical relationships. The species–area relationship is usually constructed for a single type of organism, such as all vascular plants or all species of a specific trophic level within a particular site. It is rarely if ever, constructed for all types of organisms if simply because of the prodigious data requirements. It is related but not identical to the species discovery curve.
Ecologists have proposed a wide range of factors determining the slope and elevation of the species–area relationship. These factors include the relative balance between immigration and extinction, rate and magnitude of disturbance on small vs. large areas, predator-prey dynamics, and clustering of individuals of the same species as a result of dispersal limitation or habitat heterogeneity. The species–area relationship has been reputed to follow from the 2nd law of thermodynamics. In contrast to these "mechanistic" explanations, others assert the need to test whether the pattern is simply the result of a random sampling process. Species–area relationships are often evaluated in conservation science in order to predict extinction rates in the case of habitat loss and habitat fragmentation.
Authors have classified the species–area relationship according to the type of habitats being sampled and the census design used. Frank W. Preston, an early investigator of the theory of the species–area relationship, divided it into two types: samples (a census of a contiguous habitat that grows in the census area, also called "mainland" species–area relationships), and isolates (a census of discontiguous habitats, such as islands, also called "island" species–area relationships). Michael Rosenzwe |
https://en.wikipedia.org/wiki/Functional%20requirement | In software engineering and systems engineering, a functional requirement defines a function of a system or its component, where a function is described as a summary (or specification or statement) of behavior between inputs and outputs.
Functional requirements may involve calculations, technical details, data manipulation and processing, and other specific functionality that define what a system is supposed to accomplish. Behavioral requirements describe all the cases where the system uses the functional requirements, these are captured in use cases. Functional requirements are supported by non-functional requirements (also known as "quality requirements"), which impose constraints on the design or implementation (such as performance requirements, security, or reliability). Generally, functional requirements are expressed in the form "system must do <requirement>," while non-functional requirements take the form "system shall be <requirement>." The plan for implementing functional requirements is detailed in the system design, whereas non-functional requirements are detailed in the system architecture.
As defined in requirements engineering, functional requirements specify particular results of a system. This should be contrasted with non-functional requirements, which specify overall characteristics such as cost and reliability. Functional requirements drive the application architecture of a system, while non-functional requirements drive the technical architecture of a system.
In some cases a requirements analyst generates use cases after gathering and validating a set of functional requirements. The hierarchy of functional requirements collection and change, broadly speaking, is: user/stakeholder request → analyze → use case → incorporate. Stakeholders make a request; systems engineers attempt to discuss, observe, and understand the aspects of the requirement; use cases, entity relationship diagrams, and other models are built to validate the requirement; |
https://en.wikipedia.org/wiki/Fuse%20%28hydraulic%29 | In hydraulic systems, a fuse (or velocity fuse) is a component which prevents the sudden loss of hydraulic fluid pressure. It is a safety feature, designed to allow systems to continue operating, or at least to not fail catastrophically, in the event of a system breach. It does this by stopping or greatly restricting the flow of hydraulic fluid through the fuse if the flow exceeds a threshold.
The term "fuse" is used here in analogy with electrical fuses which perform a similar function.
Hydraulic systems rely on high pressures (usually over 7000 kPa) to work properly. If a hydraulic system loses fluid pressure, such as due to a burst hydraulic hose, it will become inoperative and components such as actuators may collapse. This is an undesirable condition in life-critical systems such as aircraft or heavy machinery, such as forklifts. Hydraulic fuses help guard against catastrophic failure of a hydraulic system by automatically isolating the defective branch.
When a hydraulic system is damaged, there is generally a rapid flow of hydraulic fluid towards the breach. Most hydraulic fuses detect this flow and seal themselves (or restrict flow) if the flow exceeds a predetermined limit. There are many different fuse designs but most involve a passive spring-controlled mechanism which closes when the pressure differential across the fuse becomes excessive.
Many gas station pumps are equipped with a velocity fuse to limit gasoline flow. The fuse can be heard to engage with a "click" on some pumps if the nozzle trigger is depressed fully. A slight reduction in fuel flow can be observed. The fuse resets instantly upon releasing the trigger.
Types
There are two types of hydraulic fuses. The first one acts like a pressure relief valve, venting in case of a pressure surge. The second is more or less like a check valve. The only difference is a check valve is in place to prevent upstream fluid from coming back and venting out. A fuse is in place before the ventin |
https://en.wikipedia.org/wiki/Cauliflory | Cauliflory is a botanical term referring to plants that flower and fruit from their main stems or woody trunks, rather than from new growth and shoots. It is rare in temperate regions but common in tropical forests.
There have been several strategies to distinguish among types of cauliflory historically, including the location or age of branch where inflorescences grow, whether inflorescences attach to stolons or branches, and whether axillary nodes or adventitious nodes develop into reproductive tissues. Cauliflory is a non-homologous phenomenon with several different sources of development and evolutionary value.
The development of buds in axillary cauliflorous species occurs through either the re-use of the same position or old tissue over seasons of growth or release from dormancy. In both cases, vascularization of the bud must occur from pre-existing tissue, such as the pith. In Cercis canadensis, dormant buds break annually in a sympodial pattern. If flowers develop adventitiously, they form similarly to epicormic tissues and may be reactive to immediate environmental conditions. In certain species of Ficus, flowers may be produced from axillary buds in young plants and change to adventitious buds later.
One frequently suggested hypothesis for the evolution of cauliflory is to allow trees to be pollinated or have their seeds dispersed by animals, especially bats, that climb on trunks and sturdy limbs to feed on the nectar and fruits. Some species may instead have fruit which drops from the canopy and ripen only after they reach the ground, an alternative strategy termed nonfunctionally caulicarpic fruits. In Ficus, there is not an association between the evolution of cauliflory as an apomorphy and ecological associations. Alternative hypotheses have focused on competition for sugar and minerals between flowers and young leaves, mechanical support for larger flowers and fruits particularly in Atrocarpus and Durio, and evolutionary theory built on the plant a |
https://en.wikipedia.org/wiki/List%20of%20mesons | This list is of all known and predicted scalar, pseudoscalar and vector mesons. See list of particles for a more detailed list of particles found in particle physics.
This article contains a list of mesons, unstable subatomic particles composed of one quark and one antiquark. They are part of the hadron particle family—particles made of quarks. The other members of the hadron family are the baryons—subatomic particles composed of three quarks. The main difference between mesons and baryons is that mesons have integer spin (thus are bosons) while baryons are fermions (half-integer spin). Because mesons are bosons, the Pauli exclusion principle does not apply to them. Because of this, they can act as force mediating particles on short distances, and thus play a part in processes such as the nuclear interaction.
Since mesons are composed of quarks, they participate in both the weak and strong interactions. Mesons with net electric charge also participate in the electromagnetic interaction. They are classified according to their quark content, total angular momentum, parity, and various other properties such as C-parity and G-parity. While no meson is stable, those of lower mass are nonetheless more stable than the most massive mesons, and are easier to observe and study in particle accelerators or in cosmic ray experiments. They are also typically less massive than baryons, meaning that they are more easily produced in experiments, and will exhibit higher-energy phenomena sooner than baryons would. For example, the charm quark was first seen in the J/Psi meson () in 1974, and the bottom quark in the upsilon meson () in 1977. The top quark (the last and heaviest quark to be discovered to date) was first observed at Fermilab in 1995.
Each meson has a corresponding antiparticle (antimeson) where quarks are replaced by their corresponding antiquarks and vice versa. For example, a positive pion () is made of one up quark and one down antiquark; and its corresponding anti |
https://en.wikipedia.org/wiki/Oakley%20protocol | The Oakley Key Determination Protocol is a key-agreement protocol that allows authenticated parties to exchange keying material across an insecure connection using the Diffie–Hellman key exchange algorithm. The protocol was proposed by Hilarie K. Orman in 1998, and formed the basis for the more widely used Internet Key Exchange protocol.
The Oakley protocol has also been implemented in Cisco Systems' ISAKMP daemon. |
https://en.wikipedia.org/wiki/Ervin%20L%C3%A1szl%C3%B3 | Ervin László (; born 12 June 1932) is a Hungarian philosopher of science, systems theorist, integral theorist, originally a classical pianist. He is an advocate of the theory of quantum consciousness.
Early life and education
László was born in Budapest, Hungary, the son of a shoe manufacturer and a mother who played the piano; László himself started playing the piano when he was five years old, and gave his first piano concert with the Budapest Symphony Orchestra at the age of nine. After World War II, he moved to the United States.
Career
László is a visiting faculty member at the Graduate Institute Bethany. He has published about 75 books and over 400 papers, and is editor of World Futures: The Journal of General Evolution.
László participated in the Stock Exchange of Visions project in 2006. In 2010, he was elected an external member of the Hungarian Academy of Sciences.
In Hungary, the minister of environment appointed Laszlo as one of the leaders of the ministry's campaign concerning global warming.
Awards and honors
In 2002, László received an honorary doctorate from the University of Pécs.
Personal life
László married Carita Jägerhorn af Spurila 16 November 1956. One of their two sons is Alexander Laszlo.
Work
Systems theory
László became a leading exponent of Ludwig von Bertalanffy’s general systems theory. László viewed systems theory not only as scientifically important; he also saw in it the potential to establish an objective basis for humanist values, deriving from a consideration of a natural systems hierarchy and its evolution. In his opinion, “The ethics and natural philosophy of this new world view can help explicate and justify an emerging supranational social ethos: ‘reverence for natural systems’.”
General Evolutionary Research Group
In 1984, László was co-founder with Béla H. Bánáthy, Riane Eisler, John Corliss, Francisco Varela, Vilmos Csanyi, Gyorgy Kampis, David Loye, Jonathan Schull and Eric Chaisson of the initially secret Ge |
https://en.wikipedia.org/wiki/Liquid%20chromatography%E2%80%93mass%20spectrometry | Liquid chromatography–mass spectrometry (LC–MS) is an analytical chemistry technique that combines the physical separation capabilities of liquid chromatography (or HPLC) with the mass analysis capabilities of mass spectrometry (MS). Coupled chromatography - MS systems are popular in chemical analysis because the individual capabilities of each technique are enhanced synergistically. While liquid chromatography separates mixtures with multiple components, mass spectrometry provides spectral information that may help to identify (or confirm the suspected identity of) each separated component. MS is not only sensitive, but provides selective detection, relieving the need for complete chromatographic separation. LC–MS is also appropriate for metabolomics because of its good coverage of a wide range of chemicals. This tandem technique can be used to analyze biochemical, organic, and inorganic compounds commonly found in complex samples of environmental and biological origin. Therefore, LC–MS may be applied in a wide range of sectors including biotechnology, environment monitoring, food processing, and pharmaceutical, agrochemical, and cosmetic industries. Since the early 2000s, LC–MS (or more specifically LC–MS–MS) has also begun to be used in clinical applications.
In addition to the liquid chromatography and mass spectrometry devices, an LC–MS system contains an interface that efficiently transfers the separated components from the LC column into the MS ion source. The interface is necessary because the LC and MS devices are fundamentally incompatible. While the mobile phase in a LC system is a pressurized liquid, the MS analyzers commonly operate under high vacuum. Thus, it is not possible to directly pump the eluate from the LC column into the MS source. Overall, the interface is a mechanically simple part of the LC–MS system that transfers the maximum amount of analyte, removes a significant portion of the mobile phase used in LC and preserves the chemical identit |
https://en.wikipedia.org/wiki/Pierre%20Berthelot | Pierre Berthelot (; born 1943) is a mathematician at the University of Rennes. He developed crystalline cohomology and rigid cohomology.
Publications
Berthelot, Pierre Cohomologie cristalline des schémas de caractéristique p>0. Lecture Notes in Mathematics, Vol. 407. Springer-Verlag, Berlin-New York, 1974. 604 pp.
Berthelot, Pierre; Ogus, Arthur Notes on crystalline cohomology. Princeton University Press, Princeton, N.J.; University of Tokyo Press, Tokyo, 1978. vi+243 pp. |
https://en.wikipedia.org/wiki/Sky%2B | Sky+ (pronounced Sky Plus) is a discontinued personal video recorder (PVR) and subscription service from the satellite television provider Sky in the UK and Ireland. Launched in September 2001, it allows customers to record, pause and instantly rewind live TV. The system performs these functions using an internal hard drive inside the Sky+ set top box, an upgrade over the standard Digibox.
On 25 August 2001, the Sky+ demonstration was added to the Sky Guide demonstration, and was shown on Sky Welcome (Channel 998), which lasted for 15 minutes. A demo of the Sky+ box was shown on the Sky Customer Channel (Channel 999).
Originally a Sky+ subscription cost £10 per month - this fee was discontinued for subscribers from 1 July 2007. By July 2002 the service attracted 25,000 subscribers, and by 30 September 2009, there were 5.9 million customers with Sky+. Sky+ was also released in Italy, Germany and Austria.
During its lifetime its chief competitors in the UK market were Freeview+, Freesat+, BT Vision, and Virgin Media's V+ and TiVo. In the Republic of Ireland, Sky+ competed with Virgin Media Horizon TV and Saorview. In October 2016, Sky stopped selling the Sky+ subscription service, replacing it with Sky Q, although users of Sky+ can continue using their legacy box.
Technical information
Combined digital satellite receiver/decoder and personal video recorder (PVR).
Twin digital satellite tuners – for connection to identical independent feeds from Astra 28.2°E. Allows simultaneous recording/viewing or recording of 2 channels at once.
The set-top box middleware is provided by OpenTV, but the EPG and all the software extensions that manage the PVR functions are produced by NDS under the name of XTV PVR.
Sky+ has its own electronic programme guide made by Sky. From here, users can see what programmes are on in the next seven days. The current EPG software version (as of July 2010) is Sky+ 5.08.6.
Versions
There have been various versions of Sky+:
Sky+ 40 GB (discont |
https://en.wikipedia.org/wiki/Devolution%20%28biology%29 | Devolution, de-evolution, or backward evolution (not to be confused with dysgenics) is the notion that species can revert to supposedly more primitive forms over time. The concept relates to the idea that evolution has a purpose (teleology) and is progressive (orthogenesis), for example that feet might be better than hooves or lungs than gills. However, evolutionary biology makes no such assumptions, and natural selection shapes adaptations with no foreknowledge of any kind. It is possible for small changes (such as in the frequency of a single gene) to be reversed by chance or selection, but this is no different from the normal course of evolution and as such de-evolution is not compatible with a proper understanding of evolution due to natural selection.
In the 19th century, when belief in orthogenesis was widespread, zoologists (such as Ray Lankester and Anton Dohrn) and the palaeontologists Alpheus Hyatt and Carl H. Eigenmann advocated the idea of devolution. The concept appears in Kurt Vonnegut's 1985 novel Galápagos, which portrays a society that has evolved backwards to have small brains.
Dollo's law of irreversibility, first stated in 1893 by the palaeontologist Louis Dollo, denies the possibility of devolution. The evolutionary biologist Richard Dawkins explains Dollo's law as being simply a statement about the improbability of evolution's following precisely the same path twice.
Context
The idea of devolution is based on the presumption of orthogenesis, the view that evolution has a purposeful direction towards increasing complexity. Modern evolutionary theory, beginning with Darwin at least, poses no such presumption, and the concept of evolutionary change is independent of either any increase in complexity of organisms sharing a gene pool, or any decrease, such as in vestigiality or in loss of genes. Earlier views that species are subject to "cultural decay", "drives to perfection", or "devolution" are practically meaningless in terms of current (neo |
https://en.wikipedia.org/wiki/Calcification | Calcification is the accumulation of calcium salts in a body tissue. It normally occurs in the formation of bone, but calcium can be deposited abnormally in soft tissue, causing it to harden. Calcifications may be classified on whether there is mineral balance or not, and the location of the calcification. Calcification may also refer to the processes of normal mineral deposition in biological systems, such as the formation of stromatolites or mollusc shells (see Biomineralization).
Signs and symptoms
Calcification can manifest itself in many ways in the body depending on the location.
In the pulpal structure of a tooth, calcification often presents asymptomatically, and is diagnosed as an incidental finding during radiographic interpretation. Individual teeth with calcified pulp will typically respond negatively to vitality testing; teeth with calcified pulp often lack sensation of pain, pressure, and temperature.
Causes of soft tissue calcification
Calcification of soft tissue (arteries, cartilage, heart valves, etc.) can be caused by vitamin K2 deficiency or by poor calcium absorption due to a high calcium/vitamin D ratio. This can occur with or without a mineral imbalance.
A common misconception is that calcification is caused by excess amount of calcium in diet. Dietary calcium intake is not associated with accumulation of calcium in soft tissue, and calcification occurs irrespective of the amount of calcium intake.
Intake of excessive vitamin D can cause vitamin D poisoning and excessive intake of calcium from the intestine which, when accompanied by a deficiency of vitamin K (perhaps induced by an anticoagulant), can result in calcification of arteries and other soft tissue. Such metastatic soft tissue calcification is mainly in tissues containing "calcium catchers" such as elastic fibres or mucopolysaccharides. These tissues especially include the lungs (pumice lung) and the aorta.
Mineral balance
Dystrophic calcification, without a systemic m |
https://en.wikipedia.org/wiki/3D%20Life | 3D Life is a cellular automaton. It is a three-dimensional extension of Game of Life, investigated by Carter Bays. A number of different semitotalistic rules for the 3D rectangular Moore neighborhood were investigated.
It was popularized by A. K. Dewdney in his "Computer Recreations" column in Scientific American magazine. |
https://en.wikipedia.org/wiki/Population%20health | Population health has been defined as "the health outcomes of a group of individuals, including the distribution of such outcomes within the group". It is an approach to health that aims to improve the health of an entire human population. It has been described as consisting of three components. These are "health outcomes, patterns of health determinants, and policies and interventions".
A priority considered important in achieving the aim of population health is to reduce health inequities or disparities among different population groups due to, among other factors, the social determinants of health (SDOH). The SDOH include all the factors (social, environmental, cultural and physical) that the different populations are born into, grow up and function with throughout their lifetimes which potentially have a measurable impact on the health of human populations. The population health concept represents a change in the focus from the individual-level, characteristic of most mainstream medicine. It also seeks to complement the classic efforts of public health agencies by addressing a broader range of factors shown to impact the health of different populations. The World Health Organization's Commission on Social Determinants of Health, reported in 2008, that the SDOH factors were responsible for the bulk of diseases and injuries and these were the major causes of health inequities in all countries. In the US, SDOH were estimated to account for 70% of avoidable mortality.
From a population health perspective, health has been defined not simply as a state free from disease but as "the capacity of people to adapt to, respond to, or control life's challenges and changes". The World Health Organization (WHO) defined health in its broader sense in 1946 as "a state of complete physical, mental, and social well-being and not merely the absence of disease or infirmity."
Healthy People 2020
Healthy People 2020 is a web site sponsored by the US Department of Health and Human |
https://en.wikipedia.org/wiki/Carboxysome | Carboxysomes are bacterial microcompartments (BMCs) consisting of polyhedral protein shells filled with the enzymes ribulose-1,5-bisphosphate carboxylase/oxygenase (RuBisCO)—the predominant enzyme in carbon fixation and the rate limiting enzyme in the Calvin cycle—and carbonic anhydrase.
Carboxysomes are thought to have evolved as a consequence of the increase in oxygen concentration in the ancient atmosphere; this is because oxygen is a competing substrate to carbon dioxide in the RuBisCO reaction. To overcome the inefficiency of RuBisCO, carboxysomes concentrate carbon dioxide inside the shell by means of co-localized carbonic anhydrase activity, which produces carbon dioxide from the bicarbonate that diffuses into the carboxysome. The resulting concentration of carbon dioxide near RuBisCO decreases the proportion of ribulose-1,5-bisphosphate oxygenation and thereby avoids costly photorespiratory reactions. The surrounding shell provides a barrier to carbon dioxide loss, helping to increase its concentration around RuBisCO.
Carboxysomes are an essential part of the broader metabolic network called the Carbon dioxide-Concentrating Mechanism (CCM), which functions in two parts: (1) Membrane transporters concentrate inorganic carbon (Ci) in the cell cytosol which is devoid of carbonic anhydrases. Carbon is primarily stored in the form of HCO3- which cannot re-cross the lipid membrane, as opposed to neutral CO2 which can easily escape the cell. This stockpiles carbon in the cell, creating a disequilibrium between the intracellular and extracellular environments of about 30x the Ci concentration in water. (2) Cytosolic HCO3- diffuses into the carboxysome, where carboxysomal carbonic anhydrases dehydrate it back to CO2 in the vicinity of Rubisco, allowing Rubisco to operate at its maximal rate.
Carboxysomes are the best studied example of bacterial microcompartments, the term for functionally diverse organelles that are alike in having a protein shell.
Discovery
Pol |
https://en.wikipedia.org/wiki/Let%20It%20Snow%21%20Let%20It%20Snow%21%20Let%20It%20Snow%21 | "Let It Snow! Let It Snow! Let It Snow!", also known as simply "Let It Snow", is a song written by lyricist Sammy Cahn and composer Jule Styne in July 1945 in Hollywood, California, during a heat wave as Cahn and Styne imagined cooler conditions. The song was first recorded that fall by Vaughn Monroe, was released just after Thanksgiving, and became a hit by Christmas.
Despite the lyrics making no mention of any holiday, the song has come to be regarded as a Christmas song worldwide due to its winter theme, being played on radio stations during the Christmas and holiday season, and having often been covered by various artists on Christmas-themed albums. In the Southern Hemisphere, it can be played during the winter months of June, July, and August; and in New Zealand, some play it at Matariki.
Frank Sinatra version
American singer Frank Sinatra released a version as a single in 1950 that featured The B. Swanson Quartet.
Certifications
Dean Martin version
American singer Dean Martin released a version of the song in 1959, as part of his album A Winter Romance, and a re-recorded version in 1966, as part of The Dean Martin Christmas Album. The song entered the Billboard Hot 100 for the first time in 2018.
Certifications
Jessica Simpson version
American singer Jessica Simpson released a version of the song in 2004, as part of her album Rejoyce: The Christmas Album. Her version was produced by Billy Mann. The song has a music video. Simpson's version reached No. 20 on the Billboard Adult Contemporary chart.
Other charting recordings
Glee Cast version
Michael Bublé version
Other notable versions
Widely heard recordings of the song include:
1945 (first recording) – Vaughn Monroe for RCA Victor, which became a popular hit, reaching No. 1 on the Billboard "Best Sellers" music chart for five weeks from late December into early 1946. Vaughn later re-recorded the song in stereo for his 1958 album There I Sing/Swing It Again.
1946 – Woody Herman for Columbia Recor |
https://en.wikipedia.org/wiki/KFOX-TV | KFOX-TV (channel 14) is a television station in El Paso, Texas, United States, affiliated with the Fox network. It is owned by Sinclair Broadcast Group alongside dual CBS/MyNetworkTV affiliate KDBC-TV (channel 4). Both stations share studios on South Alto Mesa Drive in northwest El Paso, while KFOX-TV's transmitter is located atop the Franklin Mountains on the El Paso city limits.
Established as El Paso's first non-network TV station in 1979 after years of telecasting Christian programs on cable, the station as KCIK struggled financially and introduced secular entertainment programs. While it was owned in turn by two Christian groups, it continued this orientation and affiliated with Fox in 1986. It prospered with the new affiliation and introduced local news in 1997 after being sold to Cox Television. Sinclair acquired KFOX and KDBC in separate transactions in 2013, combining their operations.
History
Launch and early years
Six years before a signal was broadcast on channel 14 in El Paso, the foundation was laid for the station that would occupy it with the launch of a Christian television station, known as International Christian Television (ICT), on El Paso's cable system in 1973. The station was operated by a company known as Missionary Radio Evangelism, Inc. (MRE), led by Pete Warren and Alex Blomerth, and began to telecast seven days a week on cable channel 8 in 1974. That year, it purchased its first mobile production van. As early as mid-1974, the group had its sights set on building UHF channel 14 in El Paso: its club of donors was the "1400 Club", and it was soliciting donations with an eye to building capacity to make the leap. Pledge drives were also held to raise funds.
On May 24, 1976, Missionary Radio Evangelism filed a formal application with the Federal Communications Commission (FCC) for a channel 14 construction permit, which was granted on December 23. While ICT/MRE promised an Easter 1977 launch after getting the permit, viewers would have t |
https://en.wikipedia.org/wiki/Marko%20Poga%C4%8Dnik | Marko Pogačnik (born 11 August 1944) is a Slovenian artist and author.
Background
Pogačnik studied at the Academy of Arts in Ljubljana, the capital of Slovenia, where he graduated in 1967. He was a co-founder of the neo-avantgarde artistic movement OHO, members of which were also Tomaž Šalamun and Slavoj Žižek. From the 1980s, he embraced a holistic vision of art. He claims to have developed a method of earth healing similar to acupuncture by using pillars and positioning them on so called 'lithopuncture' points of the landscape.
In 1991, he designed the official coat of arms of the newly constituted Republic of Slovenia. In the year 2006, he joined the Movement for Justice and Development led by Janez Drnovšek, the President of Slovenia at the time.
In 1998, together with his daughter Ana, he founded the Lifenet movement, which has been described by scholars as a "typical New Age" group.
He lives and works in the village of Šempas in the lower Vipava Valley. In the last decade, the town of Nova Gorica, in which municipal territory he resides, has commissioned a number of public monuments from Pogačnik, most notably the monument to the 1000 years of the first mention of Gorica and Solkan, which stands in the town's central public square.
Since 2008, a group of his monuments and birch trees, titled the Hologram of Europe (), stands at the crossroad of Tivoli Street () and Slovene Street () in Ljubljana.
Awards
In 1991, Marko Pogačnik received the Prešeren Fund Award for his work. In 2008, he received the Jakopič Award, the central Slovenian fine arts award. He was nominated two times for the Right Livelihood Award.
Publications
Nature Spirits and Elemental Beings - Working with the Intelligence in Nature (Findhorn Press, Scotland, 1996). |
https://en.wikipedia.org/wiki/Discharge%20pressure | Discharge pressure (also called high side pressure or head pressure) is the pressure generated on the output side of a gas compressor in a refrigeration or air conditioning system. The discharge pressure is affected by several factors: size and speed of the condenser fan, condition and cleanliness of the condenser coil, and the size of the discharge line. An extremely high discharge pressure coupled with an extremely low suction pressure is an indicator of a refrigerant restriction.
Cooling technology
Hydraulics
Hydrostatics
Pressure |
https://en.wikipedia.org/wiki/Capacity%20optimization | Capacity optimization is a general term for technologies used to improve storage use by shrinking stored data. Primary technologies used for capacity optimization are data deduplication and data compression. These are delivered as software or hardware, integrated with storage systems or delivered as standalone products. Deduplication algorithms look for redundancy in sequences of bytes across comparison windows. Typically using cryptographic hash functions as identifiers of unique sequences, sequences are compared to the history of other such sequences, and where possible, the first uniquely stored version of a sequence is referenced rather than stored again. Different methods for selecting data windows include 4KB blocks to full-file comparisons known as single-instance storage (SIS).
Capacity optimization generally refers to the use of this kind of technology in a storage system. An example of this kind of system is the Venti file system in the Plan9 open source OS. There are also implementations in networking (especially wide-area networking), where they are sometimes called bandwidth optimization or WAN optimization.
Commercial implementations of capacity optimization are most often found in backup/recovery storage, where storage of iterating versions of backups day to day creates an opportunity for reduction in space using this approach. The term was first used widely in 2005. |
https://en.wikipedia.org/wiki/MTD%28f%29 | MTD(f) is an alpha-beta game tree search algorithm modified to use ‘zero-window’ initial search bounds, and memory (usually a transposition table) to reuse intermediate search results. MTD(f) is a shortened form of MTD(n,f) which stands for Memory-enhanced Test Driver with node ‘n’ and value ‘f’. The efficacy of this paradigm depends on a good initial guess, and the supposition that the final minimax value lies in a narrow window around the guess (which becomes an upper/lower bound for the search from root). The memory structure is used to save an initial guess determined elsewhere.
MTD(f) was introduced in 1994 and largely supplanted NegaScout (PVS), the previously dominant search paradigm for chess, checkers, othello and other game automatons.
Origin
MTD(f) was first described in a University of Alberta Technical Report authored by Aske Plaat, Jonathan Schaeffer, Wim Pijls, and Arie de Bruin, which would later receive the ICCA Novag Best Computer Chess Publication award for 1994/1995. The algorithm MTD(f) was created out of a research effort to understand the SSS* algorithm, a best-first search algorithm invented by George Stockman in 1979. SSS* was found to be equivalent to a series of Alpha-beta pruning|alpha-beta calls, provided that alpha-beta used storage, such as a transposition table.
The name MTD(f) stands for Memory-enhanced Test Driver, referencing Judea Pearl's Test algorithm, which performs Zero-Window Searches. MTD(f) is described in depth in Aske Plaat's 1996 PhD thesis.
Zero-window searches
A "zero-window" search is an alpha-beta search whose upper and lower bounds are identical, or differ by one unit, so that the return value is guaranteed to fall outside the bound(s) (or in an exceptionally lucky case, be equal to the bound).
MTD(f) derives its efficiency by only performing zero-window alpha-beta searches, with a previously determined "good" bound (i.e. beta). In MTD(f), AlphaBeta fails high or low, returning a lower bound or an upper bound |
https://en.wikipedia.org/wiki/Sodium%20aluminosilicate | Sodium aluminosilicate refers to compounds which contain sodium, aluminium, silicon and oxygen, and which may also contain water. These include synthetic amorphous sodium aluminosilicate, a few naturally occurring minerals and synthetic zeolites. Synthetic amorphous sodium aluminosilicate is widely used as a food additive, E 554.
Amorphous sodium aluminosilicate
This substance is produced with a wide range of compositions and has many different applications. It is encountered as an additive E 554 in food where it acts as an anticaking (free flow) agent. As it is manufactured with a range of compositions it is not strictly a chemical compound with a fixed stoichiometry. One supplier quotes a typical analysis for one of their products as 14SiO2·Al2O3·Na2O·3H2O,(Na2Al2Si14O32·3H2O).
The US FDA has as of April 1, 2012 approved sodium aluminosilicate (sodium silicoaluminate) for direct contact with consumable items under 21 CFR 182.2727. Sodium aluminosilicate is used as molecular sieve in medicinal containers to keep contents dry.
Sodium aluminosilicate may also be listed as:
aluminium sodium salt
sodium silicoaluminate
aluminosilicic acid, sodium salt
sodium aluminium silicate
aluminum sodium silicate
sodium silico aluminate
sasil
As a problem in industrial processes
The formation of sodium aluminosilicate makes the Bayer process uneconomical for bauxites high in silica.
Minerals sometimes called sodium aluminosilicate
Naturally occurring minerals that are sometimes given the chemical name, sodium aluminosilicate include albite (NaAlSi3O8, an end-member of the plagioclase series) and jadeite (NaAlSi2O6).
Synthetic zeolites sometimes called sodium aluminosilicate
Synthetic zeolites have complex structures and examples (with structural formulae) are:
Na12Al12Si12O48·27H2O, zeolite A (Linde type A sodium form, NaA), used in laundry detergents
Na16Al16Si32O96·16H2O, Analcime, IUPAC code ANA
Na12Al12Si12O48·q H2O, Losod
Na384Al384Si384O1536·518H2O, Lind |
https://en.wikipedia.org/wiki/Inclusion%20bodies | Inclusion bodies are aggregates of specific types of protein found in neurons, a number of tissue cells including red blood cells, bacteria, viruses, and plants. Inclusion bodies of aggregations of multiple proteins are also found in muscle cells affected by inclusion body myositis and hereditary inclusion body myopathy.
Inclusion bodies in neurons may be accumulated in the cytoplasm or nucleus, and are associated with many neurodegenerative diseases.
Inclusion bodies in neurodegenerative diseases are aggregates of misfolded proteins (aggresomes) and are hallmarks of many of these diseases, including Lewy bodies in Lewy body dementias, and Parkinson's disease, neuroserpin inclusion bodies called Collins bodies in familial encephalopathy with neuroserpin inclusion bodies, inclusion bodies in Huntington's disease, Papp–Lantos bodies in multiple system atrophy, and various inclusion bodies in frontotemporal dementia including Pick bodies. Bunina bodies in motor neurons are a core feature of amyotrophic lateral sclerosis.
Other usual cell inclusions are often temporary inclusions of accumulated proteins, fats, secretory granules or other insoluble components.
Inclusion bodies are found in bacteria as particles of aggregated protein. They have a higher density than many other cell components but are porous. They typically represent sites of viral multiplication in a bacterium or a eukaryotic cell and usually consist of viral capsid proteins.
Inclusion bodies contain very little host protein, ribosomal components or DNA/RNA fragments. They often almost exclusively contain the over-expressed protein and aggregation and has been reported to be reversible. It has been suggested that inclusion bodies are dynamic structures formed by an unbalanced equilibrium between aggregated and soluble proteins of Escherichia coli. There is a growing body of information indicating that formation of inclusion bodies occurs as a result of intracellular accumulation of partially folded ex |
https://en.wikipedia.org/wiki/Tanner%20graph | In coding theory, a Tanner graph, named after Michael Tanner, is a bipartite graph used to state constraints or equations which specify error correcting codes. In coding theory, Tanner graphs are used to construct longer codes from smaller ones. Both encoders and decoders employ these graphs extensively.
Origins
Tanner graphs were proposed by Michael Tanner as a means to create larger error correcting codes from smaller ones using recursive techniques. He generalized the techniques of Elias for product codes.
Tanner discussed lower bounds on the codes obtained from these graphs irrespective of the specific characteristics of the codes which were being used to construct larger codes.
Tanner graphs for linear block codes
Tanner graphs are partitioned into subcode nodes and digit nodes. For linear block codes, the subcode nodes denote rows of the parity-check matrix H. The digit nodes represent the columns of the matrix H. An edge connects a subcode node to a digit node if a nonzero entry exists in the intersection of the corresponding row and column.
Bounds proven by Tanner
Tanner proved the following bounds
Let be the rate of the resulting linear code, let the degree of the digit nodes be and the degree of the subcode nodes be . If each subcode node is associated with a linear code (n,k) with rate r = k/n, then the rate of the code is bounded by
Computational complexity of Tanner graph based methods
The advantage of these recursive techniques is that they are computationally tractable. The coding
algorithm for Tanner graphs is extremely efficient in practice, although it is not
guaranteed to converge except for cycle-free graphs, which are known not to admit asymptotically
good codes.
Applications of Tanner graph
Zemor's decoding algorithm, which is a recursive low-complexity approach to code construction, is based on Tanner graphs.
Notes
Michael Tanner's Original paper
Michael Tanner's page
Coding theory
Application-specific graphs |
https://en.wikipedia.org/wiki/Yungas | The Yungas (Aymara yunka warm or temperate Andes or earth, Quechua yunka warm area on the slopes of the Andes) is a bioregion of a narrow band of forest along the eastern slope of the Andes Mountains from Peru and Bolivia, and extends into Northwest Argentina at the slope of the Andes pre-cordillera. It is a transitional zone between the Andean highlands and the eastern forests. Like the surrounding areas, the Yungas belong to the Neotropical realm; the climate is rainy, humid, and warm.
Setting
The Yungas forests are extremely diverse, ranging from moist lowland forest to evergreen montane forest and cloud forests. The terrain, formed by valleys, fluvial mountain trails and streams, is extremely rugged and varied, contributing to the ecological diversity and richness. A complex mosaic of habitats occur with changing latitude as well as elevation. There are high levels of biodiversity and species endemism throughout the Yungas regions. Many of the forests are evergreen, and the South Andean Yungas contains what may be the last evergreen forests resulting from Quaternary glaciations.
World Wildlife ecoregions
The World Wide Fund for Nature has delineated three yungas ecoregions along the eastern side of the Andes:
The northernmost is the Peruvian Yungas, located entirely within Peru and stretching nearly the whole length of the country.
The Bolivian Yungas lies to the south, entirely within Bolivia. The Cordillera Apolobamba marks the boundary between the Peruvian Yungas and Bolivian Yungas.
The Southern Andean Yungas begins in southern Bolivia and continues to the north of Argentina. It is a humid forest region between the drier Gran Chaco region to the east and the dry, high altitude Puna region to the west.
Yungas are transitional zones between the Andean highlands and the eastern forests. The yungas forests are extremely diverse, ranging from moist lowland forest to evergreen montane forest and cloud forests. The terrain is extremely rugged and varied, co |
https://en.wikipedia.org/wiki/Sexual%20jealousy | Sexual jealousy is a special form of jealousy in sexual relationships, based on suspected or imminent sexual infidelity. The concept is studied in the field of evolutionary psychology.
Basis
Evolutionary psychologists have suggested that there is a gender difference in sexual jealousy, driven by men and women's different reproductive biology. The theory proposes that a man perceives a threat to his relationship's future because he could be fooled into raising children that are not his own. In contrast, a woman risks losing to another the relationship and all the benefits that entails. Research has shown that men are impacted more by sexual infidelity, while women are more impacted by emotional infidelity.
An alternative explanation is from a social-cognitive perspective. Typically, men place importance on their masculinity and sexual dominance. When the male's partner commits sexual infidelity, these two components of his ego become severely threatened. Women are more emotionally invested in a relationship, and therefore experience a threat to their self-perception when a partner commits infidelity, more concerned with risk to the emotional content than the sexual.
Some research has suggested that there are no gender differences in sexual jealousy, concluding that males and females both equally experience distress over emotional and sexual infidelity. Sexual jealousy is cross-culturally universal but how it manifests itself may differ across cultures.
Gender-specific behaviors
Female
Psychologists have found that males react very strongly to sexual infidelity, whereas females are more likely to forgive a one-time sexual adventure if it does not threaten the male parental investment. Therefore, jealousy is likely to be evoked in females if they feel that their partner may leave them for another woman; this has been shown to be more likely to occur if the male commits emotional infidelity. Emotional infidelity occurs when one partner develops a meaningful, emot |
https://en.wikipedia.org/wiki/Matrix%20congruence | In mathematics, two square matrices A and B over a field are called congruent if there exists an invertible matrix P over the same field such that
PTAP = B
where "T" denotes the matrix transpose. Matrix congruence is an equivalence relation.
Matrix congruence arises when considering the effect of change of basis on the Gram matrix attached to a bilinear form or quadratic form on a finite-dimensional vector space: two matrices are congruent if and only if they represent the same bilinear form with respect to different bases.
Note that Halmos defines congruence in terms of conjugate transpose (with respect to a complex inner product space) rather than transpose, but this definition has not been adopted by most other authors.
Congruence over the reals
Sylvester's law of inertia states that two congruent symmetric matrices with real entries have the same numbers of positive, negative, and zero eigenvalues. That is, the number of eigenvalues of each sign is an invariant of the associated quadratic form.
See also
Congruence relation
Matrix similarity
Matrix equivalence |
https://en.wikipedia.org/wiki/Apo2.7 | Apo2.7 is a protein confined to the mitochondrial membrane. It can be detected during early stages of apoptosis. It can be used to detect apoptosis via flow cytometry. |
https://en.wikipedia.org/wiki/4B3T | 4B3T, which stands for 4 (four) binary 3 (three) ternary, is a line encoding scheme used for ISDN PRI interface. 4B3T represents four binary bits using three pulses.
Description
It uses three states:
+ (positive pulse),
0 (no pulse), and
− (negative pulse).
This means we have 24 = 16 input combinations to represent, using 33 = 27 output combinations. 000 is not used to avoid long periods without a transition. 4B3T uses a paired disparity code to achieve an overall zero DC bias: six triplets are used which have no DC component (0+−, 0−+, +0−, −0+, +−0, −+0), and the remaining 20 are grouped into 10 pairs with differing disparity (e.g. ++− and −−+). When transmitting, the DC bias is tracked and a combination chosen that has a DC component of the opposite sign to the running total.
This mapping from 4 bits to three ternary states is given in a table known as Modified Monitoring State 43 (MMS43).
A competing encoding technique, used for the ISDN basic rate interface where 4B3T is not used, is 2B1Q.
The sync sequence used is the 11-symbol Barker code, +++−−−+−−+− or its reverse, −+−−+−−−+++.
Encoding table
Each 4-bit input group is encoded as a 3-symbol group (transmitted left to right) from the following table.
Encoding requires keeping track of the accumulated DC offset, the number of + pulses minus the number of − pulses in all preceding groups. The starting value is arbitrary; here we use the values 1 through 4, although −1.5, −0.5, +0.5 and +1.5 is another possibility.
This code forces a transition after at most five consecutive identical non-zero symbols, or four consecutive zero symbols.
Decoding table
Decoding is simpler, as the decoder does not need to keep track of the encoder state, although doing so allows greater error detection. The 000 triplet is not a legal encoded sequence, but is typically decoded as binary 0000.
See also
Other line codes that have 3 states:
hybrid ternary code
bipolar encoding
MLT-3 encoding
B3ZS |
https://en.wikipedia.org/wiki/Neopterin | Neopterin is an organic compound belonging to the pteridine class of heterocyclic compounds.
Neopterin belongs to the chemical group known as pteridines. It is synthesised by human macrophages upon stimulation with the cytokine interferon-gamma and is indicative of a pro-inflammatory immune status. Neopterin serves as a marker of cellular immune system activation. In humans neopterin follows a circadian and circaseptan rhythm.
Biosynthesis
The biosynthesis of neopterin occurs in two steps from guanosine triphosphate (GTP). The first being catalyzed by GTP cyclohydrolase, which opens the ribose group. Phosphatases next catalyze the hydrolysis of the phosphate ester group.
Neopterin as disease marker
Measurement of neopterin concentrations in body fluids like blood serum, cerebrospinal fluid or urine provides information about activation of cellular immune activation in humans under the control of T helper cells type 1. High neopterin production is associated with increased production of reactive oxygen species, neopterin concentrations also allow to estimate the extent of oxidative stress elicited by the immune system.
Increased neopterin production is found in, but not limited to, the following diseases:
Viral infections including human immunodeficiency virus (HIV), hepatitis B and hepatitis C, SARS-CoV-1, SARS-CoV-2.
Bacterial infections by intracellular living bacteria such as Borrelia (Lyme disease), Mycobacterium tuberculosis, and Helicobacter pylori.
parasites such as Plasmodium (malaria)
Autoimmune diseases such as rheumatoid arthritis (RA) and systemic lupus erythematosus (SLE)
Malignant tumor diseases
Allograft rejection episodes.
A leukodystrophy called Aicardi-Goutieres syndrome
Depression and somatization.
Neopterin concentrations usually correlate with the extent and activity of the disease, and are also useful to monitor during therapy in these patients. Elevated neopterin concentrations are among the best predictors of adverse outcome in pati |
https://en.wikipedia.org/wiki/Neo-Hookean%20solid | A neo-Hookean solid is a hyperelastic material model, similar to Hooke's law, that can be used for predicting the nonlinear stress-strain behavior of materials undergoing large deformations. The model was proposed by Ronald Rivlin in 1948. In contrast to linear elastic materials, the stress-strain curve of a neo-Hookean material is not linear. Instead, the relationship between applied stress and strain is initially linear, but at a certain point the stress-strain curve will plateau. The neo-Hookean model does not account for the dissipative release of energy as heat while straining the material and perfect elasticity is assumed at all stages of deformation.
The neo-Hookean model is based on the statistical thermodynamics of cross-linked polymer chains and is usable for plastics and rubber-like substances. Cross-linked polymers will act in a neo-Hookean manner because initially the polymer chains can move relative to each other when a stress is applied. However, at a certain point the polymer chains will be stretched to the maximum point that the covalent cross links will allow, and this will cause a dramatic increase in the elastic modulus of the material. The neo-Hookean material model does not predict that increase in modulus at large strains and is typically accurate only for strains less than 20%. The model is also inadequate for biaxial states of stress and has been superseded by the Mooney-Rivlin model.
The strain energy density function for an incompressible neo-Hookean material in a three-dimensional description is
where is a material constant, and is the first invariant (trace), of the right Cauchy-Green deformation tensor, i.e.,
where are the principal stretches.
For a compressible neo-Hookean material the strain energy density function is given by
where is a material constant and is the deformation gradient. It can be shown that in 2D, the strain energy density function is
Several alternative formulations exist for compressible neo-Hoo |
https://en.wikipedia.org/wiki/Stokes%20phenomenon | In complex analysis the Stokes phenomenon, discovered by , is where the asymptotic behavior of functions can differ in different regions of the complex plane. This seemingly gives rise to a paradox when looking at the asymptotic expansion of an analytic function. Since an analytic function is continuous you would expect the asymptotic expansion to be continuous. This paradox is the subject of Stokes' early research and is known as Stokes phenomenon. The regions in the complex plane with different asymptotic behaviour are bounded by possibly one or two types of curves known as Stokes curves and Anti-Stokes Curves. This apparent paradox has since been resolved and the supposed discontinuous jump in the asymptotic expansions has been shown to be smooth and continuous. In order to resolve this paradox the asymptotic expansion needs to be handled in a careful manner. More specifically the asymptotic expansion must include additional exponentially small terms relative to the usual algebraic terms included in a usual asymptotic expansion. What happens in Stokes phenomenon is that an asymptotic expansion in one region may contain an exponentially small contribution (neglecting this contribution still gives a correct asymptotic expansion for that region). However, this exponentially small term can become exponentially large in another region of the complex plane, this change occurs across the Anti-Stokes curves. Furthermore the exponentially small term may switch on or off other exponentially small terms, this change occurs across a Stokes curve. Including these exponentially small terms allows the asymptotic expansion to be written as a continuous expansion for the entire complex domain which resolves the Stokes Phenomenon paradox.
Stokes Curves and anti-Stokes Curves
Across a Stokes curve, an exponentially small term can switch on or off another exponentially small term.
Across an anti-Stokes curve, a subdominant exponentially small term can switch to a dominant exponen |
https://en.wikipedia.org/wiki/Veterinarian%27s%20Oath | The Veterinarian's Oath was adopted by the American Veterinary Medical Association's House of Delegates July 1969, and amended by the AVMA Executive Board, November 1999 and December 2010.
The Veterinarian Oath taken by Canadian veterinarians, established by the Canadian Veterinary Medical Association in 2004, has some minor deviations from that of the American Veterinary Medical Association. It reads as follows:
See also
Veterinary ethics |
https://en.wikipedia.org/wiki/Pick-and-place%20machine | Surface-mount technology (SMT) component placement systems, commonly called pick-and-place machines or P&Ps, are robotic machines which are used to place surface-mount devices (SMDs) onto a printed circuit board (PCB). They are used for high speed, high precision placing of a broad range of electronic components, for example capacitors, resistors, integrated circuits onto the PCBs which are in turn used in computers, consumer electronics as well as industrial, medical, automotive, military and telecommunications equipment. Similar equipment exists for through-hole components.
This type of equipment is sometimes used to package microchips using the flip chip method.
History
1980s and 1990s
During this time, a typical SMT assembly line employed two different types of pick-and-place (P&P) machines arranged in sequence.
The unpopulated board was fed into a rapid placement machine. These machines, sometimes called chip shooters, place mainly low-precision, simple package components such as resistors and capacitors. These high-speed P&P machines were built around a single turret design capable of mounting up to two dozen stations. As the turret spins, the stations passing the back of the machine pick up parts from tape feeders mounted on a moving carriage. As the station proceeds around the turret, it passes an optical station that calculates the angle at which the part was picked up, allowing the machine to compensate for drift. Then, as the station reaches the front of the turret, the board is moved into the proper position, the nozzle is spun to put the part in proper angular orientation, and the part is placed on the board. Typical chip shooters can, under optimal conditions, place up to 53,000 parts per hour, or almost 15 parts per second.
Because the PCB is moved rather than the turret, only lightweight parts that will not be shaken loose by the violent motion of the PCB can be placed this way.
From the high speed machine, the board transits to a precision pla |
https://en.wikipedia.org/wiki/Storage%20clamp | A clamp is a compact heap, mound or pile of materials. A storage clamp is used in the agricultural industry for temporary storage of root crops such as potato, turnip, rutabaga, mangelwurzel, and sugar beet.
A clamp is formed by excavating a shallow rectangular depression in a field to make a base for the clamp. Root crops are then stacked onto the base up to a height of about . When the clamp is full, the earth scraped from the field to make the base is then used to cover the root crops to a depth of several inches. Straw or old hay may be used to protect the upper surface from rain erosion.
A well-made clamp will keep the vegetables cool and dry for many months. Most clamps are relatively long and narrow, allowing the crops to be progressively removed from one end without disturbing the remaining vegetables. The use of a clamp allows a farmer to feed vegetables into market over many months.
See also
Food preservation
Root cellar
Brick clamp
Charcoal clamp
CLAMP, an artist collective named after potato clamps |
https://en.wikipedia.org/wiki/Input%20queue | In computer science, an input queue is a collection of processes in storage that are waiting to be brought into memory to run a program. Input queues are mainly used in Operating System Scheduling which is a technique for distributing resources among processes. Input queues not only apply to operating systems (OS), but may also be applied to scheduling inside networking devices. The purpose of scheduling is to ensure resources are being distributed fairly and effectively; therefore, it improves the performance of the system.
Essentially, a queue is a collection which has data added in the rear position and removed from the front position. There are many different types of queues, and the ways they operate may be totally different.
Operating systems use First-Come, First-Served queues, Shortest remaining time, Fixed priority pre-emptive scheduling, round-robin scheduling and multilevel queue scheduling.
Network devices use First-In-First-Out queue, Weighted fair queue, Priority queue and Custom queue.
Operating system
In operating systems, processes are loaded into memory, and wait for their turn to be executed by the central processing unit (CPU). CPU scheduling manages process states and decides when a process will be executed next by using the input queue.
First-Come, First-out
First-Come, First-out processes are taken out from the queue in consecutive order that they are put into the queue. With this method, every process is treated equally. If there are two processes of different priority and the lower priority process enters the queue first, it will be executed first. This approach may not be ideal if different processes have different priorities, especially if the processes are long running.
Shortest remaining time
The shortest remaining time method tries to predict the processing time of developments and places them into the queue from the smallest to largest processing time. This method estimates and predicts based on prior history records. In ter |
https://en.wikipedia.org/wiki/Shooting%20ratio | The shooting ratio or "Bertolo code" in filmmaking and television production is the ratio between the total duration of its footage created for possible use in a project and that which appears in its final cut.
A film with a shooting ratio of 2:1 would have shot twice the amount of footage than was used in the film. In real terms this means that 120 minutes of footage would have been shot to produce a film of 60 minutes in length.
While shooting ratios can vary greatly between productions, a typical shooting ratio for a production using film stock will be between 6:1 and 10:1, whereas a similar production using video is likely to be much higher. This is a direct result of the significant difference in price between video tape stock and film stock and the necessary processing. Although the decisions, styles and preferences of the filmmakers can affect the shooting ratio of a project greatly, the nature of the production (genre, form, single camera, multi-camera, etc.) greatly affects the typical range of the ratios seen – documentary films typically have the highest (often exceeding 100:1 following the rise of video and digital media) and animated films have the lowest (typically as close to 1:1 as possible, since the creation of footage frame by frame makes the time costs of animation extremely high compared to live action). Animated productions will often shoot acting reference (by animators of themselves and or others), location reference, and performance reference (taken of voice actors), but these pieces of reference footage are not regarded as counting towards the shooting ratio, as they were never intended to appear in the projects they were created for. Audition footage, screen tests, and location reference are similarly not counted towards a narrative film's shooting ratio, live action or animated, for the same reason. Since a documentary may potentially use any footage that is shot at any point for any reason, documentary productions do not have similar e |
https://en.wikipedia.org/wiki/Espacenet | Espacenet (formerly stylized as esp@cenet) is a free online service for searching patents and patent applications. Espacenet was developed by the European Patent Office (EPO) together with the member states of the European Patent Organisation. Most member states have an Espacenet service in their national language, and access to the EPO's worldwide database, most of which is in English. In 2022, the Espacenet worldwide service claimed to have records on more than 140 million patent publications.
History
By launching Espacenet in 1998, the EPO is said to have "revolutionized public access to international patent information, releasing patent data from its paper prisons and changing forever how patents are disseminated, organized, searched, and retrieved."
In 2004, i.e. in the early years of Espacenet, Nancy Lambert considered that, although free, Espacenet, like the United States Patent and Trademark Office (USPTO) database of US patents, "still tend[ed] to have primitive search engines and in some cases rather cumbersome mechanisms to download patents." She reported it as being deliberate, on the part of the USPTO and EPO, "who have said they do not wish to compete unfairly with commercial vendors". In 2009, Espacenet offered the so-called SmartSearch which allows a query to be composed using a subset of Contextual Query Language (CQL).
In 2012, the EPO launched "Patent Translate", a free online automatic translation service for patents. Created in partnership with Google, the translation engine was "specifically built to handle complex and technical patent vocabulary", using "millions of official, human-translated patent documents" to train the translation engine. It covers translations between English and 31 other languages. According to the Patent Information News' magazine published by the EPO, a 2013 independent study compared Espacenet with DepatisNet, Freepatentsonline, Google Patent and the public search facility at the USPTO. In that study, Espacenet |
https://en.wikipedia.org/wiki/Comparison%20of%20source-code-hosting%20facilities | A source-code-hosting facility (also known as forge) is a file archive and web hosting facility for source code of software, documentation, web pages, and other works, accessible either publicly or privately. They are often used by open-source software projects and other multi-developer projects to maintain revision and version history, or version control. Many repositories provide a bug tracking system, and offer release management, mailing lists, and wiki-based project documentation. Software authors generally retain their copyright when software is posted to a code hosting facilities.
General information
Features
Version control systems
Popularity
Discontinued: CodePlex, Gna!, Google Code.
Specialized hosting facilities
The following are open-source software hosting facilities that only serve a specific narrowly focused community or technology.
Former hosting facilities
Alioth (Debian) – In 2018, Alioth has been replaced by a GitLab based solution hosted on salsa.debian.org. Alioth has been finally switched off in June 2018.
BerliOS – abandoned in April 2014
Betavine – abandoned somewhere in 2015.
CodeHaus – shut down in May 2015
CodePlex – shut down in December 2017.
Fedora Hosted – closed in March 2017
Gitorious – shut down in June 2015.
Gna! – shut down in 2017.
Google Code – closed in January 2016, all projects archived. See http://code.google.com/archive/.
java.net – Java.net and kenai.com hosting closed April 2017.
Phabricator – wound down operations 1 June 2021, all projects continued to be hosted with very limited support after 31 August 2021.
Tigris.org – shut down in July 2020.
Mozdev.org - shut down in July 2020.
See also
Comparison of version-control software
Distributed version control
Forge (software)
List of free software project directories
List of version-control software
Source code escrow for closed-source software
Version control (source-code-management systems)
Notes |
https://en.wikipedia.org/wiki/Timeline%20of%20entomology | Entomology, the scientific study of insects and closely related terrestrial arthropods, has been impelled by the necessity of societies to protect themselves from insect-borne diseases, crop losses to pest insects, and insect-related discomfort, as well as by people's natural curiosity. This timeline article traces the history of entomology.
Timelines of entomology
Timeline of entomology – prior to 1800
Timeline of entomology – 1800–1850
Timeline of entomology – 1850–1900
Timeline of entomology – post 1900
History of classification
Many different classifications were proposed by early entomologists. It is important to realise that whilst many early names survive, they may be at different levels in the phylogenetic hierarchy. For instance, many families were first published as genera, as for example the genus Mymar, proposed by Alexander Henry Haliday in 1829, is now represented by the family Mymaridae.
History of forensic entomology
See also
European and American voyages of scientific exploration
List of natural history dealers |
https://en.wikipedia.org/wiki/Alston%20Scott%20Householder | Alston Scott Householder (5 May 1904 – 4 July 1993) was an American mathematician who specialized in mathematical biology and numerical analysis.
He is the inventor of the Householder transformation and of Householder's method.
Career
Householder was born in Rockford, Illinois, USA. He received a BA in philosophy from the Northwestern University of Evanston, Illinois in 1925, and an MA, also in philosophy, from Cornell University in 1927. He taught mathematics while preparing for his PhD, which was awarded at the University of Chicago in 1937. His thesis dealt with the topic of the calculus of variations.
After receiving his doctorate, Householder concentrated on the field of mathematical biology, working with several other researchers with Nicolas Rashevsky at the University of Chicago.
In 1946, Householder joined the Mathematics Division of the Oak Ridge National Laboratory, where he was appointed chair in 1948; it is during this period that his interests shift toward numerical analysis. In 1969 he left ORNL to become Professor of Mathematics at the University of Tennessee, where he eventually became chairman. In 1974 he retired and went to live in Malibu, California.
Householder contributed in different ways to the organisation of research. He was president of the Society for Industrial and Applied Mathematics (SIAM) and of the Association for Computing Machinery (ACM). He was a member of the redactional committees for Psychometrika, Numerische Mathematik, Linear Algebra and Its Applications, and was editor in chief of the SIAM Journal on Numerical Analysis. He opened up his wide personal bibliography on numerical linear algebra in form of a KWIC index. He also organized the important Gatlinburg Conferences, which are still held under the name Householder Symposia.
Personal life
Householder spent his youth in Alabama. He was first married to Belle Householder (died 1975, children: John and Jackie) and remarried 1984 to Heidi Householder (née Vogg). He d |
https://en.wikipedia.org/wiki/Cyberun | Cyberun is a ZX Spectrum video game by Ultimate Play the Game and published by U.S. Gold in 1986. Although not part of the Jetman series, it has similarities to Jetpac in that the player must construct their spaceship from parts, then seek out resources and power-ups.
Gameplay
The player controls a spaceship trapped on a planet inhabited by hostile aliens. The goal is to upgrade the spaceship with parts scattered around the planet and mine a valuable element called "Cybernite". The atmosphere above ground is populated by flying aliens and clouds that drip acid, damaging the ship's shields. The ship requires fuel to fly, and once exhausted will bounce along the ground of the planet unable to climb. A similar enemy ship is also on the planet attempting to mine the Cybernite before the player. Fuel can be replenished by tankers on the planet surface, but damaged shields cannot be repaired. The player must venture into caverns below the surface in order to mine the Cybernite, which can only be done once the ship has been upgraded to include a mining laser. Once sufficient Cybernite has been collected, the player can escape to the next planet in the Zebarema system.
Reception
The game was well received by critics, with Crash awarding it a 90% Crash Smash, and Your Spectrum giving it 8/10, describing the game as "a classic pick up the pieces and shoot em up with brilliant graphics". |
https://en.wikipedia.org/wiki/Beetle%20%28ASIC%29 | The Beetle ASIC is an analog readout chip. It is developed for the LHCb experiment at CERN.
Overview
The chip integrates 128 channels with low-noise charge-sensitive pre-amplifiers and shapers. The pulse shape can be chosen such that it complies with LHCb specifications: a peaking time of 25 ns with a remainder of the peak voltage after 25 ns of less than 30%. A comparator per channel with configurable polarity provides a binary signal. Four adjacent comparator channels are being ORed and brought off chip via LVDS drivers.
Either the shaper or comparator output is sampled with the LHC bunch-crossing frequency of 40 MHz into an analog pipeline. This ring buffer has a programmable latency of a maximum of 160 sampling intervals and an integrated derandomising buffer of 16 stages. For analogue readout data is multiplexed with up to 40 MHz onto one or four ports. A binary readout mode operates at up to 80 MHz output rate on two ports. Current drivers bring the serialised data off chip.
The chip can accept trigger rates up to 1.1 MHz to perform a dead-timeless readout within 900 ns per trigger. For testability and calibration purposes, a charge injector with adjustable pulse height is implemented. The bias settings and various other parameters can be controlled via a standard I²C-interface. The chip is radiation hardened to an accumulated dose of more than 100 Mrad. Robustness against single event upset is achieved by redundant logic.
External links
Beetle - a readout chip for LHCb
The Large Hadron Collider beauty experiment
Application-specific integrated circuits
CERN |
https://en.wikipedia.org/wiki/Bubbler%20%28video%20game%29 | Bubbler is a ZX Spectrum video game developed and published by Ultimate Play the Game in 1987. It was Ultimate's final release for 8-bit home computers before evolving into Rare. The game is an isometric platform game in the style of Marble Madness (1984).
Development
A Commodore 64 version was outsourced to Lynsoft but the release was cancelled as Ultimate thought the game was running too slowly.
Reception
Crash magazine reviewer Ricky disliked the impreciseness of the controls. Sinclair User were more impressed by the game; they did not consider it to be one of Ultimate's most original game or particularly well presented but thought it was very addictive. It was awarded a 5 star rating. |
https://en.wikipedia.org/wiki/Jerzy%20Neyman | Jerzy Neyman (April 16, 1894 – August 5, 1981; born Jerzy Spława-Neyman; ) was a Polish mathematician and statistician who spent the first part of his professional career at various institutions in Warsaw, Poland and then at University College London, and the second part at the University of California, Berkeley. Neyman first introduced the modern concept of a confidence interval into statistical hypothesis testing and co-revised Ronald Fisher's null hypothesis testing (in collaboration with Egon Pearson).
Life and career
He was born into a Polish family in Bendery, in the Bessarabia Governorate of the Russian Empire, the fourth of four children of Czesław Spława-Neyman and Kazimiera Lutosławska. His family was Roman Catholic, and Neyman served as an altar boy during his early childhood. Later, Neyman would become an agnostic. Neyman's family descended from a long line of Polish nobles and military heroes. He graduated from the Kamieniec Podolski gubernial gymnasium for boys in 1909 under the name Yuri Cheslavovich Neyman. He began studies at Kharkiv University in 1912, where he was taught by Ukrainian probabilist Sergei Natanovich Bernstein. After he read 'Lessons on the integration and the research of the primitive functions' by Henri Lebesgue, he was fascinated with measure and integration.
In 1921, he returned to Poland in a program of repatriation of POWs after the Polish-Soviet War.
He earned his Doctor of Philosophy degree at University of Warsaw in 1924 for a dissertation titled "On the Applications of the Theory of Probability to Agricultural Experiments". He was examined by Wacław Sierpiński and Stefan Mazurkiewicz, among others. He spent a couple of years in London and Paris on a fellowship to study statistics with Karl Pearson and Émile Borel. After his return to Poland, he established the Biometric Laboratory at the Nencki Institute of Experimental Biology in Warsaw.
He published many books dealing with experiments and statistics, and devised the way |
https://en.wikipedia.org/wiki/Carlos%20Westendorp | Carlos Westendorp y Cabeza (born 7 January 1937) is a Spanish diplomat and former politician. He is the former Minister of Foreign Affairs and also served as High Representative for Bosnia and Herzegovina succeeding Carl Bildt and was powered with upholding the Dayton Peace Agreement.
Career
Born in Madrid on 7 January 1937, Westendorp joined the Spanish Diplomatic Service in 1966. Following several assignments abroad (from 1966 to 1969: Deputy Consul General in São Paulo, Brazil; from 1975 to 1979: Commercial and Economic Counselor at the Spanish Embassy in the Hague, the Netherlands) and in Spain (1969–1975: Head of Economic Studies at the Diplomatic School; Director of Technological Agreements in the Ministry of Foreign Affairs; Chief of Cabinet of the Minister of Industry) he dedicated a great part of his professional career to the process of integration of Spain into the European Communities.
Between 1979 and 1985 at the Ministry of European Affairs, he successively served as Adviser to the Minister, as Head of the Minister’s Private Office and as Secretary General, presiding over the technical team in charge of the accession negotiations. In 1986, when Spain joined the European Communities, he was appointed its first Ambassador Permanent Representative. He chaired the Committee of Permanent Representatives (COREPER) during the first Spanish Presidency of the EEC in 1989.
From 1991 to 1995 he was Spain’s Secretary of State for the European Union. He was centrally involved in the Spanish Presidency of the EU in 1995, which coincided with the adoption of the Euro, the launching of the Barcelona process and the signing of the transatlantic agenda. In this last capacity, he chaired the Reflection group set up to prepare the negotiations on treaty change which led to the Treaties of Amsterdam and subsequently, Nice.
In December 1995, he was appointed Minister of Foreign Affairs and served in that capacity until the end of the last government presided by Felipe G |
https://en.wikipedia.org/wiki/Avid%20DNxHD | Avid DNxHD ("Digital Nonlinear Extensible High Definition") is a lossy high-definition video post-production codec developed by Avid for multi-generation compositing with reduced storage and bandwidth requirements. It is an implementation of SMPTE VC-3 standard.
Overview
DNxHD is a video codec intended to be usable as both an intermediate format suitable for use while editing and as a presentation format. DNxHD data is typically stored in an MXF container, although it can also be stored in a QuickTime container.
On February 13, 2008, Avid reported that DNxHD was approved as compliant with the SMPTE VC-3 standard.
DNxHD is intended to be an open standard, but as of March 2008, has remained effectively a proprietary Avid format. The source code for the Avid DNxHD codec is freely available from Avid for internal evaluation and review, although commercial use requires Avid licensing approval. It has been commercially licensed to a number of companies including Ikegami, FilmLight, Harris Corporation, JVC, Seachange, EVS Broadcast Equipment.
On September 14, 2014, at the Avid Connect event in Amsterdam, Netherlands, Avid announced the DNxHR codec to support resolutions greater than 1080p, such as 2K and 4K.
On December 22, 2014, Avid Technology released an update for Media Composer that added support for 4K resolution, the Rec. 2020 color space, and a bit rate of up to 3,730 Mbit/s with the DNxHR codec.
Implementations
DNxHD was first supported in Avid DS Nitris (Sept 2004), then Avid Media Composer Adrenaline with the DNxcel option (Dec 2004) and finally by Avid Symphony Nitris (Dec 2005). Xpress Pro is limited to using DNxHD 8-bit compression, which is either imported from file or captured using a Media Composer with Adrenaline hardware. Media Composer 2.5 also allows editing of fully uncompressed HD material that was either imported or captured on a Symphony Nitris or DS Nitris system. Ikegami's Editcam camera system is unique in its support for DNxHD, and rec |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.