source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/De%20Branges%20space
In mathematics, a de Branges space (sometimes written De Branges space) is a concept in functional analysis and is constructed from a de Branges function. The concept is named after Louis de Branges who proved numerous results regarding these spaces, especially as Hilbert spaces, and used those results to prove the Bieberbach conjecture. De Branges functions A Hermite-Biehler function, also known as de Branges function is an entire function E from to that satisfies the inequality , for all z in the upper half of the complex plane . Definition 1 Given a Hermite-Biehler function , the de Branges space is defined as the set of all entire functions F such that where: is the open upper half of the complex plane. . is the usual Hardy space on the open upper half plane. Definition 2 A de Branges space can also be defined as all entire functions satisfying all of the following conditions: Definition 3 There exists also an axiomatic description, useful in operator theory. As Hilbert spaces Given a de Branges space . Define the scalar product: A de Branges space with such a scalar product can be proven to be a Hilbert space.
https://en.wikipedia.org/wiki/General%20feature%20format
In bioinformatics, the general feature format (gene-finding format, generic feature format, GFF) is a file format used for describing genes and other features of DNA, RNA and protein sequences. GFF Versions The following versions of GFF exist: General Feature Format Version 2, generally deprecated Gene Transfer Format 2.2, a derivative used by Ensembl Generic Feature Format Version 3 Genome Variation Format, with additional pragmas and attributes for sequence_alteration features GFF2/GTF had a number of deficiencies, notably that it can only represent two-level feature hierarchies and thus cannot handle the three-level hierarchy of gene → transcript → exon. GFF3 addresses this and other deficiencies. For example, it supports arbitrarily many hierarchical levels, and gives specific meanings to certain tags in the attributes field. The GTF is identical to GFF, version 2. GFF general structure All GFF formats (GFF2, GFF3 and GTF) are tab delimited with 9 fields per line. They all share the same structure for the first 7 fields, while differing in the content and format of the ninth field. Some field names have been changed in GFF3 to avoid confusion. For example, the "seqid" field was formerly referred to as "sequence", which may be confused with a nucleotide or amino acid chain. The general structure is as follows: The 8th field: phase of CDS features Simply put, CDS means "CoDing Sequence". The exact meaning of the term is defined by Sequence Ontology (SO). According to the GFF3 specification: Meta Directives In GFF files, additional meta information can be included and follows after the ## directive. This meta information can detail GFF version, sequence region, or species (full list of meta data types can be found at Sequence Ontology specifications). GFF software Servers Servers that generate this format: Clients Clients that use this format: Validation The modENCODE project hosts an online GFF3 validation tool with generous limits of 286.10 MB a
https://en.wikipedia.org/wiki/Obstacle%20avoidance
In robotics, obstacle avoidance is the task of satisfying some control objective subject to non-intersection or non-collision position constraints. What is critical about obstacle avoidance concept in this area is the growing need of usage of unmanned aerial vehicles in urban areas for especially military applications where it can be very useful in city wars. Normally obstacle avoidance is considered to be distinct from path planning in that one is usually implemented as a reactive control law while the other involves the pre-computation of an obstacle-free path which a controller will then guide a robot along. With recent advanced in the autonomous vehicles sector, a good and dependable obstacle avoidance feature of a driverless platform is also required to have a robust obstacle detection module. Reactive obstacle avoidance is a behavior based control strategy in a robot. It is a task similar to the navigation problem and produces a collision free motion. See also D* dynamic pathfinding algorithm Robotics Robot control
https://en.wikipedia.org/wiki/Tissue%20tropism
Tissue tropism is the range of cells and tissues of a host that support growth of a particular pathogen, such as a virus, bacterium or parasite. Some bacteria and viruses have a broad tissue tropism and can infect many types of cells and tissues. Other viruses may infect primarily a single tissue. For example, rabies virus affects primarily neuronal tissue. Influencing factors Factors influencing viral tissue tropism include: The presence of cellular receptors permitting viral entry. Availability of transcription factors involved in viral replication. The molecular nature of the viral tropogen or virus surface, such as the glycoprotein, which interacts with the corresponding cell receptor. The cellular receptors are the proteins found on a cell or viral surface. These receptors are like keys, allowing the viral cell to fuse with or attach itself to a cell. The way that these proteins are acquired is through a similar process to that of an infection cycle. How 'tropic' tissue is acquired Tissue tropism develops in the following stages: Virus with GPX enters body (where GP - glycoprotein and X is the numeric value given to the GP) Viral cell "targets" cell with a GPX receptors Viral cell fuses with the host cell and inserts its contents into the host cell Reverse transcription occurs Viral DNA is incorporated with host DNA via viral enzyme Production of RNA and viral protein Viral particle is assembled Viral particle buds out of the cell, taking a chunk of the cell membrane with it and acquiring a new tissue with all the receptors it needs to continue tissue tropism Example: HIV has a gp120, which is precisely what the CD4 marker is on the surface of the macrophages and T cells. Thus HIV can enter T cells and macrophages See also Host tropism Endothelial Cell Tropism
https://en.wikipedia.org/wiki/Rough%20number
A k-rough number, as defined by Finch in 2001 and 2003, is a positive integer whose prime factors are all greater than or equal to k. k-roughness has alternately been defined as requiring all prime factors to strictly exceed k. Examples (after Finch) Every odd positive integer is 3-rough. Every positive integer that is congruent to 1 or 5 mod 6 is 5-rough. Every positive integer is 2-rough, since all its prime factors, being prime numbers, exceed 1. See also Buchstab function, used to count rough numbers Smooth number Notes
https://en.wikipedia.org/wiki/Synaptotropic%20hypothesis
The synaptotropic hypothesis, also called the synaptotrophic hypothesis, is a neurobiological hypothesis of neuronal growth and synapse formation. The hypothesis was first formulated by J.E. Vaughn in 1988, and remains a focus of current research efforts. The synaptotropic hypothesis proposes that input from a presynaptic to a postsynaptic cell (and maturation of excitatory synaptic inputs) eventually can change the course of synapse formation at dendritic and axonal arbors. This synapse formation is required for the development of neuronal structure in the functioning brain. Dendritic Arbor Development Growth Dendrites of central nervous system neurons grow by addition and retraction of thin branches. This process is highly dynamic. Only a small fraction of newly added branches are actually maintained to become long-lasting components of the arbor. This process suggests that the branches sample the environment to detect the appropriate cells with which to form synapses. As a result, the hypothesis predicts that growth will be directed into regions containing more presynaptic elements. This morphology can be stabilized by creating microtubule nucleation at the microtubules. Synaptogenesis The formation of new synapses begins with initial contact between cells via cell-cell adhesion. This contact often occurs between either axonal or dendritic filopodia, which are highly dynamic and rarely stabilize. Next, the adhesive contact is converted to a nascent synapse, which contains glutamatergic NMDA receptors, but not AMPA receptors. However, the activation of NMDARs by glutamate can trigger the recruitment of AMPARs from the postsynaptic density. They also have a relatively high concentration of dense-core vesicles, which are thought to deliver structural proteins to the presynaptic site. Synapse Maturation Maturation of glutamatergic synapses involves changes in the amplitude of AMPA receptor-mediated synaptic transmission, as well as in the NMDAR subunit composit
https://en.wikipedia.org/wiki/Phene
A phene is an individual genetically determined characteristic or trait which can be possessed by an organism, such as eye colour, height, behavior, tooth shape or any other observable characteristic. Phene - phenotype - phenome distinction The term 'phene' was evidently coined as an obvious parallel construct to 'gene'. Phene is to Phenotype as Gene is to Genotype, and Similarly Phene is to Phenome as Gene is to Genome. More specifically, a Phene is an abstract concept describing a particular characteristic which can be possessed by an organism. Whereas Phenotype refers to a collection of Phenes possessed by a particular organism, and Phenome refers to the entire set of Phenes that exist within an organism or species. Genome wide association studies use "phenes" or "traits" (symptoms) to distinguish groups in the human population. These groups are then employed to identify associations with genetic alleles that are more common in the symptomatic group than in the asymptomatic control group. Allen et al. report that with respect to Schizophrenia "Research in molecular genetics has focused on detecting multiple genes of small effect" This indicates the importance of discovering individual traits or "phenes" that are governed by single genes. Schizophrenia or bipolar disorder may be described as a phenotype but how many individual traits or "phenes" contribute to these phenotypes? Very large genome wide association studies have not found many significant gene linkages. On the contrary the results of these studies implicate a large number of gene alleles that have a very small effect (phene). It is important to note that the word phenotype was originally used to refer to both the trait/character itself (e.g. the blue eyes phenotype) and the set of traits/characteristics possessed by the organism (clair's eye-colour phenotype is blue). While this definition is still used in many places, the lack of distinction can make indepth explanations confusing and thus use of
https://en.wikipedia.org/wiki/Electron%20pair
In chemistry, an electron pair or Lewis pair consists of two electrons that occupy the same molecular orbital but have opposite spins. Gilbert N. Lewis introduced the concepts of both the electron pair and the covalent bond in a landmark paper he published in 1916. Because electrons are fermions, the Pauli exclusion principle forbids these particles from having the same quantum numbers. Therefore, for two electrons to occupy the same orbital, and thereby have the same orbital quantum number, they must have different spin quantum number. This also limits the number of electrons in the same orbital to two. The pairing of spins is often energetically favorable, and electron pairs therefore play a large role in chemistry. They can form a chemical bond between two atoms, or they can occur as a lone pair of valence electrons. They also fill the core levels of an atom. Because the spins are paired, the magnetic moment of the electrons cancel one another, and the pair's contribution to magnetic properties is generally diamagnetic. Although a strong tendency to pair off electrons can be observed in chemistry, it is also possible that electrons occur as unpaired electrons. In the case of metallic bonding the magnetic moments also compensate to a large extent, but the bonding is more communal so that individual pairs of electrons cannot be distinguished and it is better to consider the electrons as a collective 'sea'. A very special case of electron pair formation occurs in superconductivity: the formation of Cooper pairs. In unconventional superconductors, whose crystal structure contains copper anions, the electron pair bond is due to antiferromagnetic spin fluctuations. See also Electron pair production Frustrated Lewis pair Jemmis mno rules Lewis acids and bases Nucleophile Polyhedral skeletal electron pair theory
https://en.wikipedia.org/wiki/Transponder%20%28aeronautics%29
A transponder (short for transmitter-responder and sometimes abbreviated to XPDR, XPNDR, TPDR or TP) is an electronic device that produces a response when it receives a radio-frequency interrogation. Aircraft have transponders to assist in identifying them on air traffic control radar. Collision avoidance systems have been developed to use transponder transmissions as a means of detecting aircraft at risk of colliding with each other. Air traffic control units use the term "squawk" when they are assigning an aircraft a transponder code, e.g., "Squawk 7421". Squawk thus can be said to mean "select transponder code" or "squawking xxxx" to mean "I have selected transponder code xxxx". The transponder receives interrogation from the Secondary Surveillance Radar on 1030 MHz and replies on 1090 MHz. Secondary surveillance radar Secondary surveillance radar (SSR) is referred to as "secondary", to distinguish it from the "primary radar" that works by passively reflecting a radio signal off the skin of the aircraft. Primary radar determines range and bearing to a target with reasonably high fidelity, but it cannot determine target elevation (altitude) reliably except at close range. SSR uses an active transponder (beacon) to transmit a response to an interrogation by a secondary radar. This response most often includes the aircraft's pressure altitude and a 4-digit octal identifier. Operation A pilot may be requested to squawk a given code by an air traffic controller, via the radio, using a phrase such as "Cessna 123AB, squawk 0363". The pilot then selects the 0363 code on their transponder and the track on the air traffic controller's radar screen will become correctly associated with their identity. Because primary radar generally gives bearing and range position information, but lacks altitude information, mode C and mode S transponders also report pressure altitude. Mode C altitude information conventionally comes from the pilot's altimeter, and is transmitted
https://en.wikipedia.org/wiki/Heterotachy
Heterotachy refers to variations in lineage-specific evolutionary rates over time. In the field of molecular evolution, the principle of heterotachy states that the substitution rate of sites in a gene can change through time. It has been proposed that the positions that show switches in substitution rate over time (that is, heterotachous sites) are good indicators of functional divergence. However, it appears that heterotachy is a much more general process, since most variable sites of homologous proteins with no evidence of functional shift are heterotachous. The covarion hypothesis is a specific form of heterotachy. Some studies have proposed functional divergence models that are also heterotachous. Additionally, some mixture models that do not explicitly account for rate-shift, but site-partitions evolving at different relative substitution rates across lineages are mathematically heterotachous. Failure to take heterotachy into account in phylogenetic reconstructions may lead to incorrect phylogenetic trees. Thus Zhong et al. (2011) say that heterotachy is one of the reasons for variability in reconstructions of the origin of gnetophytes.
https://en.wikipedia.org/wiki/World%20Ocean%20Circulation%20Experiment
The World Ocean Circulation Experiment (WOCE) was a component of the international World Climate Research Program, and aimed to establish the role of the World Ocean in the Earth's climate system. WOCE's field phase ran between 1990 and 1998, and was followed by an analysis and modeling phase that ran until 2002. When the WOCE was conceived, there were three main motivations for its creation. The first of these is the inadequate coverage of the World Ocean, specifically in the Southern Hemisphere. Data was also much more sparse during the winter months than the summer months, and there was—and still to some extent—a critical need for data covering all seasons. Secondly, the data that did exist was not initially collected for studying ocean circulation and was not well suited for model comparison. Lastly, there were concerns involving the accuracy and reliability of some measurements. The WOCE was meant to address these problems by providing new data collected in ways designed to "meet the needs of global circulation models for climate prediction." Goals Two major goals were set for the campaign. 1. Develop ocean models that can be used in climate models and collect the data necessary for testing them Specifically, understand: Large scale fluxes of heat and fresh water Dynamical balance of World Ocean circulation Components of ocean variability on months to years The rates and nature of formation, ventilation and circulation of water masses that influence the climate system on time scales from ten to one hundred years In order to achieve Goal 1, the WCRP outlined and established Core Projects that would receive priority. The first of these was the "Global Description" project, which was meant to obtain data on the circulation of heat, fresh water and chemicals, as well as the statistics of eddies. The second project—"Southern Ocean"—placed particular emphasis on studying the Antarctic Circumpolar Current and the Southern Ocean’s interaction with the W
https://en.wikipedia.org/wiki/Hering%E2%80%93Breuer%20reflex
The Hering–Breuer inflation reflex, named for Josef Breuer and Ewald Hering, is a reflex triggered to prevent the over-inflation of the lung. Pulmonary stretch receptors present on the wall of bronchi and bronchioles of the airways respond to excessive stretching of the lung during large inspirations. Once activated, they send action potentials through large myelinated fibers of the vagus nerve to the inspiratory area in the medulla and apneustic center of the pons. In response, the inspiratory area is inhibited directly and the apneustic center is inhibited from activating the inspiratory area. This inhibits inspiration, allowing expiration to occur. The Hering–Breuer inflation reflex should not be confused with the deflation reflex discovered by the same individuals, Hering and Breuer. The majority of this page discusses the inflation reflex; the deflation reflex is considered separately at the end. History Josef Breuer and Ewald Hering reported in 1868 that a maintained distention of the lungs of anesthetized animals decreased the frequency of the inspiratory effort or caused a transient apnea. The stimulus was therefore pulmonary inflation. Anatomy and physiology The Hering-Breuer reflex, put simply, is what keeps the lungs from over-inflating with inspired air. The neural circuit that controls the Hering–Breuer inflation reflex involves several regions of the central nervous system, and both sensory and motor components of the vagus nerve. Increased sensory activity of the pulmonary-stretch lung afferents (via the vagus nerve) results in inhibition of the central inspiratory drive and thus inhibition of inspiration and initiation of expiration. The lung afferents also send inhibitory projections to the cardiac vagal motor neurones (CVM) in the nucleus ambiguus (NA) and dorsal motor vagal nucleus (DMVN). The CVMs, which send motor fibers to the heart via the vagus nerve, are responsible for tonic inhibitory control of heart rate. Thus, an increase in p
https://en.wikipedia.org/wiki/Bor%20S.%20Luh%20International%20Award
The Bor S. Luh International Award has been awarded every year since 1956. Before 2005, this award was named the International Award. It is given to an individual or institution that had outstanding efforts in one of the following areas in food technology: 1) International exchange of ideas, 2) better international understanding, and/or 3) practical successful technology transfer to an economically depressed area in a developed or developing area. The award was renamed for Bor S. Luh (1916-2001), who was born and educated in China before completing his education in the United States. Luh was the first president of the Chinese American Food Society in 1974-5 and received its Professional Achievement Award in 1984. Award winners receive a plaque from the Bor S. Luh Endowment Fund of the Institute of Food Technologists Foundation and a USD 3000 honorarium. Winners
https://en.wikipedia.org/wiki/Kim%20Tu-bong
Kim Tu-bong (; 16 February 1889 – March 1958 or later) was the first Chairman of the Workers' Party of North Korea (one of predecessor of today's Workers' Party of Korea, the other is Workers' Party of South Korea) from 1946 to 1949. He was known in Korean history as a linguist, scholar, revolutionary and politician. His most famous work was under Ju Sigyeong; later, after participating in the March 1st Movement, he with other Korean leaders of the time established a provisional government-in-exile in China, and because of his communist beliefs he played an important role in the early North Korean communist government. He and other members of the Yan'an faction formed the New People's Party when they returned from exile. After the New People's Party merged into the Workers Party of North Korea (WPNK) in 1946 at the 1st WPNK Congress, he became WPNK Chairman. He was the first head of state (Chairman of the Presidium of the Supreme People's Assembly) of North Korea from 1948 to 1957. He is most remembered in South Korea for his efforts in establishing the Korean linguistic field and especially that of Hangul. Much of his work both political and linguistically was done while living in China with the exiled government of Korea. He is also known by his pen name Baekyeon. He was purged by Kim Il-Sung in 1957. Early childhood and education Born on 2 February 1889, in South Korea's South Gyeongsang Province, near modern-day Pusan, he spent his early years being homeschooled during the time of imperial rule. He would move to Seoul at the age of 20 (1908) to attend both Geho School and Baechae School and in that same year graduate from Bogo High School. While he was in Seoul he would join the Korea Youth organization in 1913 and the following year (1914) leave Baechae School. He was also an editor for the magazine. Early linguistic work After graduating from Bosungkobo (Bosung College) in 1908, Kim Tu-bong worked closely with a linguistics professor from Bosungkobo named
https://en.wikipedia.org/wiki/Laser%20TV
Laser color television (laser TV), or laser color video display, is a type of television that utilizes two or more individually modulated optical (laser) rays of different colors to produce a combined spot that is scanned and projected across the image plane by a polygon-mirror system or less effectively by optoelectronic means to produce a color-television display. The systems work either by scanning the entire picture a dot at a time and modulating the laser directly at high frequency, much like the electron beams in a cathode ray tube, or by optically spreading and then modulating the laser and scanning a line at a time, the line itself being modulated in much the same way as with digital light processing (DLP). The special case of one ray reduces the system to a monochrome display as, for example, in black and white television. This principle applies to a direct view display as well as to a (front or rear) laser projector system. Laser TV technology began to appear in the 1990s. In the 21st century, the rapid development and maturity of semiconductor lasers and other technologies gave it new advantages. History The laser source for television or video display was originally proposed by Helmut K.V. Lotsch in the German Patent 1 193 844. In December 1977 H.K.V. Lotsch and F. Schroeter explained laser color television for conventional as well as projection-type systems and gave examples of potential applications. 18 years later the German-based company Schneider AG presented a functional laser-TV prototype at IFA'95 in Berlin/Germany. Due to the bankruptcy of Schneider AG, however, the prototype was never developed further to a market-ready product. Proposed in 1966, laser illumination technology remained too costly to be used in commercially viable consumer products. At the Las Vegas Consumer Electronics Show in 2006, Novalux Inc., developer of Necsel semiconductor laser technology, demonstrated their laser illumination source for projection displays and a
https://en.wikipedia.org/wiki/Computer%20programming%20in%20the%20punched%20card%20era
From the invention of computer programming languages up to the mid-1970s, most computer programmers created, edited and stored their programs line by line on punch cards. Punched cards A punched card is a flexible write-once medium that encodes data, most commonly 80 characters. Groups or "decks" of cards form programs and collections of data. The term is often used interchangeably with punch card, the difference being that an unused card is a "punch card," but once information had been encoded by punching holes in the card, it was now a "punched card." For simplicity, this article will use the term punched card to refer to either. Often programmers first wrote their program out on special forms called coding sheets, taking care to distinguish the digit zero from the letter O, the digit one from the letter I, eight from B, two from Z, and so on using local conventions such as the "slashed zero". These forms were then taken by keypunch operators, who using a keypunch machine such as the IBM 026 (later IBM 029) punched the deck. Often another keypunch operator would then take that deck and re-punch from the coding sheets – but using a "verifier" such as the IBM 059 that checked that the original punching had no errors. A typing error generally necessitated re-punching an entire card. The editing of programs was facilitated by reorganizing the cards, and removing or replacing the lines that had changed; programs were backed up by duplicating the deck, or writing it to magnetic tape. In smaller organizations programmers might do their own punching, and in all cases would often have access to a keypunch to make small changes to a deck. Work environment The description below describes an all-IBM shop (a "shop" is programmer jargon for a programming site) but shops using other brands of mainframes (or minicomputers) would have similar equipment although because of cost or availability might have different manufacturer's equipment, e.g. an NCR, ICL, Hewlett-Packard (H
https://en.wikipedia.org/wiki/Cluster%20manager
Within cluster and parallel computing, a cluster manager is usually backend graphical user interface (GUI) or command-line interface (CLI) software that runs on a set of cluster nodes that it manages (in some cases it runs on a different server or cluster of management servers). The cluster manager works together with a cluster management agent. These agents run on each node of the cluster to manage and configure services, a set of services, or to manage and configure the complete cluster server itself (see super computing.) In some cases the cluster manager is mostly used to dispatch work for the cluster (or cloud) to perform. In this last case a subset of the cluster manager can be a remote desktop application that is used not for configuration but just to send work and get back work results from a cluster. In other cases the cluster is more related to availability and load balancing than to computational or specific service clusters. See also List of cluster management software Grid network Further reading Cluster management Adaptive Control of Extreme-scale Stream Processing Systems Proceedings of the 26th IEEE International Conference on Distributed Computing Systems. Design, implementation, and evaluation of the linear road benchmark on the stream processing core Proceedings of the 2006 ACM SIGMOD international conference on Management of data. Parallel Job Scheduling A Status Report (2004) 10th Workshop on Job Scheduling Strategies for Parallel Processing, New-York, NY, June 2004. Condor-G: A Computation Management Agent for Multi-Institutional Grids Springer Journal Cluster Computing Volume 5, Number 3 / July, 2002 From clusters to the fabric: the job management perspective Cluster Computing, 2003. Proceedings. 2003 IEEE International Conference on An Overview of the Galaxy Management Framework for Scalable Enterprise Cluster Computing IEEE International Conference on Cluster Computing (Cluster'00), 2000. Performance and Interoperability Issues
https://en.wikipedia.org/wiki/Avrami%20equation
The Avrami equation describes how solids transform from one phase to another at constant temperature. It can specifically describe the kinetics of crystallisation, can be applied generally to other changes of phase in materials, like chemical reaction rates, and can even be meaningful in analyses of ecological systems. The equation is also known as the Johnson–Mehl–Avrami–Kolmogorov (JMAK) equation. The equation was first derived by Johnson, Mehl, Avrami and Kolmogorov (in Russian) in a series of articles published in the Journal of Chemical Physics between 1939 and 1941. Moreover, Kolmogorov treated statistically the crystallization of a solid in 1937 (in Russian, Kolmogorov, A. N., Izv. Akad. Nauk. SSSR., 1937, 3, 355). Transformation kinetics Transformations are often seen to follow a characteristic s-shaped, or sigmoidal, profile where the transformation rates are low at the beginning and the end of the transformation but rapid in between. The initial slow rate can be attributed to the time required for a significant number of nuclei of the new phase to form and begin growing. During the intermediate period the transformation is rapid as the nuclei grow into particles and consume the old phase while nuclei continue to form in the remaining parent phase. Once the transformation approaches completion, there remains little untransformed material for further nucleation, and the production of new particles begins to slow. Additionally, the previously formed particles begin to touch one another, forming a boundary where growth stops. Derivation The simplest derivation of the Avrami equation makes a number of significant assumptions and simplifications: Nucleation occurs randomly and homogeneously over the entire untransformed portion of the material. The growth rate does not depend on the extent of transformation. Growth occurs at the same rate in all directions. Growth stops as a result of impingement of the growing grains. If these conditions are me
https://en.wikipedia.org/wiki/Diorama%20%28Efteling%29
The Diorama is a miniature world in Efteling amusement park in the Netherlands. The highly detailed mountainous world, or Diorama, was designed by Anton Pieck and opened in 1971, in honour of the 20th birthday of Efteling. The visitors can walk around a 60 metre long show-case with mountains, little villages, castles and churches, moving trains and automobiles and flowing water. Most of the Diorama is set in day-time, but a smaller part is devoted to night-time The landscape has been built entirely out of styrofoam. The Diorama is the first attraction with contributions of Ton van de Ven, the creative director of Efteling at that time. He made some sketches for it, but they weren’t used for the Diorama; years later they were used for one of the scenes in the dark-ride Dreamflight. The attraction was completely renovated in 2007, all rail tracks were replaced, switching mechanisms for the signals are removed and there are seven small attractions from the "real efteling" where added. Trivia One of the wooden bridges over the railroad tracks in the first scene has collapsed and a new wooden bridge has been built next to it. It seems to be a creative idea, but it was actually stepped on by one of the builders by accident. Exceptionally, Märklin for many years manufactured the Minex steam trains specially for Efteling.
https://en.wikipedia.org/wiki/IRE%20%28unit%29
The IRE unit is used in the measurement of composite video signals. Its name is derived from the initials of the Institute of Radio Engineers. A value of 100 IRE is defined to be +714 mV in an analog NTSC video signal. A value of 0 IRE corresponds to the voltage value of 0 mV, the signal value during the blanking period. The sync pulse is normally 40 IRE below this 0 IRE value, so the total range covered from peak to trough of an all white signal would be 140 IRE. Video signals use the "IRE" unit instead of DC voltages to describe levels and amplitudes. Based on a standard 1 Vpp NTSC composite-video signal that swings from -286 mV (sync tip) to +714 mV (peak video), a 140 IRE peak-to-peak convention is established. Thus, one NTSC IRE unit is 7.143 mV ( V or mV), where -40 IRE is equivalent to -285.7 mV, and +100 IRE is equivalent to +714.3 mV. 0 IRE is equivalent to 0 V. The black level is equivalent to 53.57 mV (7.5 IRE). The PAL video signal is slightly different in that it swings from -300 mV to +700 mV, instead. Thus, one PAL IRE unit is 7 mV, where -43 IRE is equivalent to -300 mV at the sync tip, and +100 IRE is equivalent to +700 mV at the peak video level. Black level is the same as the blanking level 0 mV (0 IRE). The reason IRE is a relative measurement (percent) is because a video signal may be any amplitude. This unit is used in the ITU-R BT.470 which defines PAL, NTSC and SECAM:
https://en.wikipedia.org/wiki/Astrological%20symbols
Historically, astrological and astronomical symbols have overlapped. Frequently used symbols include signs of the zodiac and classical planets. These originate from medieval Byzantine codices. Their current form is a product of the European Renaissance. Other symbols for astrological aspects are used in various astrological traditions. History and origin Symbols for the classical planets, zodiac signs, aspects, lots, and the lunar nodes appear in the medieval Byzantine codices in which many ancient horoscopes were preserved. In the original papyri of these Greek horoscopes, there was a circle with the glyph representing shine () for the Sun; and a crescent for the Moon. Classical planets The written symbols for Mercury, Venus, Jupiter, and Saturn have been traced to forms found in late Classical Greek papyri. The symbols for Jupiter and Saturn are monograms of the initial letters of the corresponding Greek names, and the symbol for Mercury is a stylized caduceus. A.S.D. Maunder finds antecedents of the planetary symbols in earlier sources, used to represent the gods associated with the classical planets. Bianchini's planisphere, produced in the 2nd century, shows Greek personifications of planetary gods charged with early versions of the planetary symbols: Mercury has a caduceus; Venus has, attached to her necklace, a cord connected to another necklace; Mars, a spear; Jupiter, a staff; Saturn, a scythe; the Sun, a circlet with rays radiating from it; and the Moon, a headdress with a crescent attached. A diagram in Johannes Kamateros' 12th-century Compendium of Astrology shows the Sun represented by the circle with a ray, Jupiter by the letter zeta (the initial of Zeus, Jupiter's counterpart in Greek mythology), Mars by a shield crossed by a spear, and the remaining classical planets by symbols resembling the modern ones, without the cross-mark seen in modern versions of the symbols. The modern sun symbol, pictured as a circle with a dot (), first appeared in the
https://en.wikipedia.org/wiki/Oospore
An oospore is a thick-walled sexual spore that develops from a fertilized oosphere in some algae, fungi, and oomycetes. They are believed to have evolved either through the fusion of two species or the chemically induced stimulation of mycelia, leading to oospore formation. In Oomycetes, oospores can also result from asexual reproduction, by apomixis. These are found in fungi as sexual spores which help the sexual reproduction of fungi. These haploid, non-motile spores are the site of meiosis and karyogamy in oomycetes. A dormant oospore, when observed under an electron microscope, has led researchers to draw conclusion that there is only a single central globule with other storage bodies surrounding it.
https://en.wikipedia.org/wiki/Router%20alert%20label
In MPLS, a label with the value of 1 represents the router alert label. This label value is legal anywhere in the label stack except at the bottom. When a received packet contains this label value at the top of the label stack, it is delivered to a local software module for processing. The actual forwarding of the packet is determined by the label beneath it in the stack. However, if the packet is forwarded further, the Router Alert Label should be pushed back onto the label stack before forwarding. The use of this label is analogous to the use of the "Router Alert" option in IPv4 packets. Since this label cannot occur at the bottom of the stack, it is not associated with a particular network layer protocol. External links http://www.cisco.com/warp/public/105/mpls_faq_4649.shtml MPLS Label Stack Encoding RFC IP Router Alert Option RFC MPLS networking Internet Standards Network protocols Tunneling protocols
https://en.wikipedia.org/wiki/Thue%20number
In the mathematical area of graph theory, the Thue number of a graph is a variation of the chromatic index, defined by and named after mathematician Axel Thue, who studied the squarefree words used to define this number. Alon et al. define a nonrepetitive coloring of a graph to be an assignment of colors to the edges of the graph, such that there does not exist any even-length simple path in the graph in which the colors of the edges in the first half of the path form the same sequence as the colors of the edges in the second half of the path. The Thue number of a graph is the minimum number of colors needed in any nonrepetitive coloring. Variations on this concept involving vertex colorings or more general walks on a graph have been studied by several authors including Barát and Varjú, Barát and Wood (2005), Brešar and Klavžar (2004), and Kündgen and Pelsmajer. Example Consider a pentagon, that is, a cycle of five vertices. If we color the edges with two colors, some two adjacent edges will have the same color x; the path formed by those two edges will have the repetitive color sequence xx. If we color the edges with three colors, one of the three colors will be used only once; the path of four edges formed by the other two colors will either have two consecutive edges or will form the repetitive color sequence xyxy. However, with four colors it is not difficult to avoid all repetitions. Therefore, the Thue number of C5 is four. Results Alon et al. use the Lovász local lemma to prove that the Thue number of any graph is at most quadratic in its maximum degree; they provide an example showing that for some graphs this quadratic dependence is necessary. In addition they show that the Thue number of a path of four or more vertices is exactly three, that the Thue number of any cycle is at most four, and that the Thue number of the Petersen graph is exactly five. The known cycles with Thue number four are C5, C7, C9, C10, C14, and C17. Alon et al. conj
https://en.wikipedia.org/wiki/Type%20erasure
In programming languages, type erasure is the load-time process by which explicit type annotations are removed from a program, before it is executed at run-time. Operational semantics not requiring programs to be accompanied by types are named type-erasure semantics, in contrast with type-passing semantics. Type-erasure semantics is an abstraction principle, ensuring that the run-time execution of a program doesn't depend on type information. In the context of generic programming, the opposite of type erasure is named reification. Type inference The reverse operation is named type inference. Though type erasure can be an easy way to define typing over implicitly typed languages (an implicitly typed term is well-typed if and only if it is the erasure of a well-typed explicitly typed lambda term), it doesn't require an algorithm to check implicitly typed terms. See also Template (C++) Problems with type erasure (in Generics in Java) Type polymorphism
https://en.wikipedia.org/wiki/Hash%20chain
A hash chain is the successive application of a cryptographic hash function to a piece of data. In computer security, a hash chain is a method used to produce many one-time keys from a single key or password. For non-repudiation, a hash function can be applied successively to additional pieces of data in order to record the chronology of data's existence. Definition A hash chain is a successive application of a cryptographic hash function to a string . For example, gives a hash chain of length 4, often denoted Applications Leslie Lamport suggested the use of hash chains as a password protection scheme in an insecure environment. A server which needs to provide authentication may store a hash chain rather than a plain text password and prevent theft of the password in transmission or theft from the server. For example, a server begins by storing which is provided by the user. When the user wishes to authenticate, they supply to the server. The server computes and verifies this matches the hash chain it has stored. It then stores for the next time the user wishes to authenticate. An eavesdropper seeing communicated to the server will be unable to re-transmit the same hash chain to the server for authentication since the server now expects . Due to the one-way property of cryptographically secure hash functions, it is infeasible for the eavesdropper to reverse the hash function and obtain an earlier piece of the hash chain. In this example, the user could authenticate 1000 times before the hash chain were exhausted. Each time the hash value is different, and thus cannot be duplicated by an attacker. Binary hash chains Binary hash chains are commonly used in association with a hash tree. A binary hash chain takes two hash values as inputs, concatenates them and applies a hash function to the result, thereby producing a third hash value. The above diagram shows a hash tree consisting of eight leaf nodes and the hash chain for the third leaf node. In ad
https://en.wikipedia.org/wiki/Lymphocytic%20colitis
Lymphocytic colitis is a subtype of microscopic colitis, a condition characterized by chronic non-bloody watery diarrhea. Causes No definite cause has been determined. The peak incidence of lymphocytic colitis is in persons over age 50; the disease affects women and men equally. Some reports have implicated long-term usage of NSAIDs, proton pump inhibitors, and selective serotonin reuptake inhibitors, and other drugs. Associations with other autoimmune disorders suggests that overactive immune responses occur. Diagnosis The colonoscopy is normal but histology of the mucosal biopsy reveals an accumulation of lymphocytes in the colonic epithelium and connective tissue (lamina propria). Collagenous colitis shares this feature but additionally shows a distinctive thickening of the subepithelial collagen table. Treatment Budesonide, in colonic release preparations, has been shown in randomized controlled trials to be effective in treating this disorder. Over-the-counter antidiarrheal drugs may be effective for some people with lymphocytic colitis. Anti-inflammatory drugs, such as salicylates, mesalazine, and systemic corticosteroids may be prescribed for people who do not respond to other drug treatment. The long-term prognosis for this disease is good with a proportion of people suffering relapses which respond to treatment. History Lymphocytic colitis was first described in 1989. See also Colitis
https://en.wikipedia.org/wiki/K-line%20%28x-ray%29
The K-line is a spectral peak in astronomical spectrometry used, along with the L-line, to observe and describe the light spectrum of stars. The K-line is associated with iron (Fe) and is described as being from emissions at ~6.4keV (thousands of electron volts). On 5 October 2006 NASA announced the results of research using the Japanese JAXA Suzaku satellite, after earlier work with the XMM-Newton satellite. "The observations include clocking the speed of a black hole's spin rate and measuring the angle at which matter pours into the void, as well as evidence for a wall of X-ray light pulled back and flattened by gravity." The study teams observed X-ray emissions from the "broad iron K line" near the event horizon of several super-massive black holes of galaxies called MCG-6-30-15 and MCG-5-23-16. The normally narrow K-line is broadened by the doppler shift (red shift or blue shift) of the X-ray light emitted by matter being affected by the gravity of the black hole. The results coincide with predictions Albert Einstein's theory of general relativity. The teams were led by Andrew Fabian of Cambridge University, England, and James Reeves of NASA's Goddard Space Flight Center, Greenbelt, Maryland, United States. See also K- and L- Electron shell Siegbahn notation
https://en.wikipedia.org/wiki/World%20Meteorological%20Organization%20squares
World Meteorological Organization (WMO) squares is a system of geocodes that divides a world map with latitude-longitude gridlines into grid cells of 10° latitude by 10° longitude, each with a unique, 4-digit numeric identifier (refer chart at NODC World Ocean Database 2005 page). On the plate carrée projection, the grid cells appear square; however, if the Mercator projection is used, the grid cells appear 'stretched' vertically nearer the tops and bottoms of the map. On the actual surface of the Globe, the cells are approximately "square" only adjacent to the Equator, and become progressively narrower and tapered (also with curved northern and southern boundaries) as they approach the poles, and cells adjoining the poles are unique in possessing three faces rather than four. Each 10°x10° square is allocated a number between 1000 and 7817. The numbering system is based first on "global quadrant" numbers where 1=NE, 3=SE, 5=SW, 7=NW which gives the initial digit of any square code (1xxx, 3xxx, 5xxx, 7xxx). The second digit (x0xx through x8xx) indicates the number of tens of degrees latitude (north in global quadrants 1 and 7, south in global quadrants 3 and 5) of the 'minimum' square boundary (nearest to the Equator), i.e. a cell extending between 10°N and 20°N (or 10°S and 20°S) has this digit = 1, a cell extending between 20°N and 30°N has this digit = 2, etc. The third and fourth digits (xx00 through xx17) similarly indicate the number of tens of degrees of longitude of the 'minimum' square boundary, nearest to the Prime Meridian. By way of illustration, the square 1000 thus extends from 0°N to 10°N, 0°E to 10°E, and the square 7817 from 80°N to 90°N, 170°W to 180°W, adjacent to the major portion of the International Date Line. In this manner, reverse-engineering (decoding) the relevant square boundaries from any particular WMO Square identifier is straightforward, in contrast to some other similar systems e.g. Marsden squares. WMO squares are also used as the
https://en.wikipedia.org/wiki/Caruso%20sauce
Caruso sauce or salsa Caruso is a warm sauce made of cream, ham, cheese, beef extract, and mushrooms. Recipes may also include nuts or onions. It is served with pasta (typically cappelletti). History Caruso sauce was first created in the 1950s in Uruguay, by Raymundo Monti of the restaurant 'Mario and Alberto', located at the intersection of Constituyente and Tacuarembó Streets in Montevideo. Monti wanted to create a new recipe following the current traditions of Italian cuisine. The dish was named in honor of the famous Neapolitan tenor Enrico Caruso (1873–1921) who was a popular figure in South America during his tours of the 1910s. The sauce was originally thought to be a variant of bechamel but its flavor is distinctly different. Several culinary seminars referred to Caruso sauce as "the new invention" and it gained international culinary recognition. In recent decades, the sauce has become increasingly popular in most South American and Western European countries. Due to the shared cultural background existing between Uruguay and Argentina, it is not unusual to encounter Caruso sauce on restaurant menus in Buenos Aires. It can even be found in some Brazilian restaurants. See also List of sauces
https://en.wikipedia.org/wiki/HMG-CoA
β-Hydroxy β-methylglutaryl-CoA (HMG-CoA), also known as 3-hydroxy-3-methylglutaryl coenzyme A, is an intermediate in the mevalonate and ketogenesis pathways. It is formed from acetyl CoA and acetoacetyl CoA by HMG-CoA synthase. The research of Minor J. Coon and Bimal Kumar Bachhawat in the 1950s at University of Illinois led to its discovery. HMG-CoA is a metabolic intermediate in the metabolism of the branched-chain amino acids, which include leucine, isoleucine, and valine. Its immediate precursors are β-methylglutaconyl-CoA (MG-CoA) and β-hydroxy β-methylbutyryl-CoA (HMB-CoA). HMG-CoA reductase catalyzes the conversion of HMG-CoA to mevalonic acid, a necessary step in the biosynthesis of cholesterol. Biosynthesis Mevalonate pathway Mevalonate synthesis begins with the beta-ketothiolase-catalyzed Claisen condensation of two molecules of acetyl-CoA to produce acetoacetyl CoA. The following reaction involves the joining of acetyl-CoA and acetoacetyl-CoA to form HMG-CoA, a process catalyzed by HMG-CoA synthase. In the final step of mevalonate biosynthesis, HMG-CoA reductase, an NADPH-dependent oxidoreductase, catalyzes the conversion of HMG-CoA into mevalonate, which is the primary regulatory point in this pathway. Mevalonate serves as the precursor to isoprenoid groups that are incorporated into a wide variety of end-products, including cholesterol in humans. Ketogenesis pathway HMG-CoA lyase breaks it into acetyl CoA and acetoacetate. See also Steroidogenic enzyme
https://en.wikipedia.org/wiki/Ictal-Interictal%20SPECT%20Analysis%20by%20SPM
ISAS (Ictal-Interictal SPECT Analysis by SPM) is an objective tool for analyzing ictal vs. interictal SPECT scans. The goal of ictal SPECT is to localize the region of seizure onset for epilepsy surgery planning. ISAS was introduced and validated in two recent studies (Chang et al., 2002; McNally et al., 2005). This site is a technical supplement to (McNally et al., 2005), which should enable ISAS to be implemented at any center for further study and analysis. Analysis The basic idea of ISAS is to compute the difference between an ictal and interictal SPECT scan for a single patient. The differences of the ictal/inter-ictal comparison are checked against a healthy normal database to determine the normal expected variation. Significant increases and decreases in CBF (cerebral blood flow) between the interictal and ictal SPECT can then be detected. The analysis is conducted using SPM (statistical parametric mapping). Cerebral blood flow is known to increase during seizures at the site of seizure onset. Since SPECT is an indicator of CBF, increases in the SPECT during seizures can be useful for seizure localization. CBF decreases are more complicated, and occur both during and following seizures in multiple locations. Details of ISAS interpretation can be found in McNally et al., 2005, but in summary: 1. For true ictal SPECT (patient injected before end of seizure), seizure onset can be reliably localized based on SPECT increases. 2. For post-ictal SPECT, seizure onset cannot be reliably localized to a single lobe based on SPECT increase or decreases. However, the side (L or R) of seizure onset can be reliably determined based on which hemisphere has greater overall SPECT decreases (hypoperfusion asymmetry index). The requirements for implementing ISAS are relatively simple. All that is needed is a computer running MATLAB and an operator with sufficient imaging experience to download and implement the SPM (statistical parametric mapping) analysis. NOTE: ISAS
https://en.wikipedia.org/wiki/Performance%20Application%20Programming%20Interface
In computer science, Performance Application Programming Interface (PAPI) is a portable interface (in the form of a library) to hardware performance counters on modern microprocessors. It is being widely used to collect low level performance metrics (e.g. instruction counts, clock cycles, cache misses) of computer systems running UNIX/Linux operating systems. PAPI provides predefined high level hardware events summarized from popular processors and direct access to low level native events of one particular processor. Counter multiplexing and overflow handling are also supported. Operating system support for accessing hardware counters is needed to use PAPI. For example, prior to 2010, a Linux/x86 kernel had to be patched with a performance monitoring counters driver (perfctr link) to support PAPI. Since Linux version 2.6.32, and PAPI 2010 releases, PAPI can leverage the existing perf subsystem in Linux, and thus does not need any out of tree driver to be functional anymore. Supported Operating Systems and requirements are listed in the official repository's documentation INSTALL.txt. See also Performance analysis Further reading A Portable Programming Interface for Performance Evaluation on Modern Processors / International Journal of High Performance Computing Applications archive Volume 14 Issue 3, August 2000, Pages 189-204 doi:10.1177/109434200001400303 Dongarra, Jack, et al. "Using PAPI for hardware performance monitoring on Linux systems" // Conference on Linux Clusters: The HPC Revolution. Vol. 5. Linux Clusters Institute, 2001. External links Official site Philip Mucci, Performance Monitoring with PAPI / Dr.Dobbs, June 01, 2005 Development of a PAPI Backend for the Sun Niagara 2 Processor, 2009 Profilers Software optimization
https://en.wikipedia.org/wiki/Census%20of%20Marine%20Life
The Census of Marine Life was a 10-year, US $650 million scientific initiative, involving a global network of researchers in more than 80 nations, engaged to assess and explain the diversity, distribution, and abundance of life in the oceans. The world's first comprehensive Census of Marine Life — past, present, and future — was released in 2010 in London. Initially supported by funding from the Alfred P. Sloan Foundation, the project was successful in generating many times that initial investment in additional support and substantially increased the baselines of knowledge in often underexplored ocean realms, as well as engaging over 2,700 different researchers for the first time in a global collaborative community united in a common goal, and has been described as "one of the largest scientific collaborations ever conducted". Project history According to Jesse Ausubel, Senior Research Associate of the Program for the Human Environment of Rockefeller University and science advisor to the Alfred P. Sloan Foundation, the idea for a "Census of Marine Life" originated in conversations between himself and Dr. J. Frederick Grassle, an oceanographer and benthic ecology professor at Rutgers University, in 1996. Grassle had been urged to talk with Ausubel by former colleagues at the Woods Hole Oceanographic Institution and was at that time unaware that Ausubel was also a program manager at the Alfred P. Sloan Foundation, funders of a number of other large scale "public good" science-based projects such as the Sloan Digital Sky Survey. Ausubel was instrumental in persuading the Foundation to fund a series of "feasibility workshops" over the period 1997-1998 into how the project might be conducted, one result of these workshops being the broadening of the initial concept from a "Census of the Fishes" into a comprehensive "Census of Marine Life". Results from these workshops, plus associated invited contributions, formed the basis of a special issue of Oceanography magazine i
https://en.wikipedia.org/wiki/Boletus%20aereus
Boletus aereus, the dark cep or bronze bolete, is a highly prized and much sought-after edible mushroom in the family Boletaceae. The bolete is widely consumed in Spain (Basque Country and Navarre), France, Italy, Greece, and generally throughout the Mediterranean. Described in 1789 by French mycologist Pierre Bulliard, it is closely related to several other European boletes, including B. reticulatus, B. pinophilus, and the popular B. edulis. Some populations in North Africa have in the past been classified as a separate species, B. mamorensis, but have been shown to be phylogenetically conspecific to B. aereus and this taxon is now regarded as a synonym. The fungus predominantly grows in habitats with broad-leaved trees and shrubs, forming symbiotic ectomycorrhizal associations in which the underground roots of these plants are enveloped with sheaths of fungal tissue (hyphae). The cork oak (Quercus suber) is a key host. The fungus produces spore-bearing fruit bodies above ground in summer and autumn. The fruit body has a large dark brown cap, which can reach in diameter. Like other boletes, B. aereus has tubes extending downward from the underside of the cap, rather than gills; spores escape at maturity through the tube openings, or pores. The pore surface of the fruit body is whitish when young, but ages to a greenish-yellow. The squat brown stipe, or stem, is up to 15 cm (6 in) tall and thick and partially covered with a raised network pattern, or reticulation. Taxonomy and phylogeny French mycologist Pierre Bulliard described Boletus aereus in 1789. The species epithet is the Latin adjective aerěus, meaning "made with bronze or copper". His countryman Lucien Quélet transferred the species to the now-obsolete genus Dictyopus in 1886, which resulted in the synonym Dictyopus aereus, while René Maire reclassified it as a subspecies of B. edulis in 1937. In 1940, Manuel Cabral de Rezende-Pinto published the variety B. aereus var. squarrosus from collections made
https://en.wikipedia.org/wiki/Ensemble%20interpretation
The ensemble interpretation of quantum mechanics considers the quantum state description to apply only to an ensemble of similarly prepared systems, rather than supposing that it exhaustively represents an individual physical system. The advocates of the ensemble interpretation of quantum mechanics claim that it is minimalist, making the fewest physical assumptions about the meaning of the standard mathematical formalism. It proposes to take to the fullest extent the statistical interpretation of Max Born, for which he won the Nobel Prize in Physics in 1954. On the face of it, the ensemble interpretation might appear to contradict the doctrine proposed by Niels Bohr, that the wave function describes an individual system or particle, not an ensemble, though he accepted Born's statistical interpretation of quantum mechanics. It is not quite clear exactly what kind of ensemble Bohr intended to exclude, since he did not describe probability in terms of ensembles. The ensemble interpretation is sometimes, especially by its proponents, called "the statistical interpretation", but it seems perhaps different from Born's statistical interpretation. As is the case for "the" Copenhagen interpretation, "the" ensemble interpretation might not be uniquely defined. In one view, the ensemble interpretation may be defined as that advocated by Leslie E. Ballentine, Professor at Simon Fraser University. His interpretation does not attempt to justify, or otherwise derive, or explain quantum mechanics from any deterministic process, or make any other statement about the real nature of quantum phenomena; it intends simply to interpret the wave function. It does not propose to lead to actual results that differ from orthodox interpretations. It makes the statistical operator primary in reading the wave function, deriving the notion of a pure state from that. In the opinion of Ballentine, perhaps the most notable supporter of such an interpretation was Albert Einstein: History In his 1
https://en.wikipedia.org/wiki/Graduate%20Texts%20in%20Mathematics
Graduate Texts in Mathematics (GTM) () is a series of graduate-level textbooks in mathematics published by Springer-Verlag. The books in this series, like the other Springer-Verlag mathematics series, are yellow books of a standard size (with variable numbers of pages). The GTM series is easily identified by a white band at the top of the book. The books in this series tend to be written at a more advanced level than the similar Undergraduate Texts in Mathematics series, although there is a fair amount of overlap between the two series in terms of material covered and difficulty level. List of books Introduction to Axiomatic Set Theory, Gaisi Takeuti, Wilson M. Zaring (1982, 2nd ed., ) Measure and Category – A Survey of the Analogies between Topological and Measure Spaces, John C. Oxtoby (1980, 2nd ed., ) Topological Vector Spaces, H. H. Schaefer, M. P. Wolff (1999, 2nd ed., ) A Course in Homological Algebra, Peter Hilton, Urs Stammbach (1997, 2nd ed., ) Categories for the Working Mathematician, Saunders Mac Lane (1998, 2nd ed., ) Projective Planes, Daniel R. Hughes, Fred C. Piper, (1982, ) A Course in Arithmetic, Jean-Pierre Serre (1996, ) Axiomatic Set Theory, Gaisi Takeuti, Wilson M. Zaring, (1973, ) Introduction to Lie Algebras and Representation Theory, James E. Humphreys (1997, ) A Course in Simple-Homotopy Theory, Marshall. M. Cohen, (1973, ) Functions of One Complex Variable I, John B. Conway (1978, 2nd ed., ) Advanced Mathematical Analysis, Richard Beals (1973, ) Rings and Categories of Modules, Frank W. Anderson, Kent R. Fuller (1992, 2nd ed., ) Stable Mappings and Their Singularities, Martin Golubitsky, Victor Guillemin, (1974, ) Lectures in Functional Analysis and Operator Theory, Sterling K. Berberian, (1974, ) The Structure of Fields, David J. Winter, (1974, ) Random Processes, Murray Rosenblatt, (1974, ) Measure Theory, Paul R. Halmos (1974, ) A Hilbert Space Problem Book, Paul R. Halmos (1982, 2nd ed., ) Fibre Bundles, Dale Husemoller (1994,
https://en.wikipedia.org/wiki/Cumulative%20hierarchy
In mathematics, specifically set theory, a cumulative hierarchy is a family of sets indexed by ordinals such that If is a limit ordinal, then Some authors additionally require that or that . The union of the sets of a cumulative hierarchy is often used as a model of set theory. The phrase "the cumulative hierarchy" usually refers to the standard cumulative hierarchy of the von Neumann universe with introduced by . Reflection principle A cumulative hierarchy satisfies a form of the reflection principle: any formula in the language of set theory that holds in the union of the hierarchy also holds in some stages . Examples The von Neumann universe is built from a cumulative hierarchy . The sets of the constructible universe form a cumulative hierarchy. The Boolean-valued models constructed by forcing are built using a cumulative hierarchy. The well founded sets in a model of set theory (possibly not satisfying the axiom of foundation) form a cumulative hierarchy whose union satisfies the axiom of foundation.
https://en.wikipedia.org/wiki/Australian%20Game%20Developers%20Conference
The Australian Game Developers Conference (AGDC) was an annual conference held from 1999 to 2005 that brought together Australian and overseas game developers, publishers, programmers, artists, production staff, computer graphics companies, audio companies, software tool developers, buyers and suppliers to the game development industry. It was owned by the Academy of Interactive Entertainment (AIE) and was run by Interactive Entertainment Events, a subsidiary of AIE. History The inaugural conference was held in Sydney, New South Wales, in 1999. Due to enormous financial support provided by Multimedia Victoria, the conference was held in Melbourne, Victoria, from 2000 to 2005. The conference program regularly featured: Trade Exhibition Schools & Computer Games Academic Summit Pitch to Publishers Day VIP and Delegate networking functions Up to seven concurrent conference sessions Keynote presentations The Australian Game Developer Awards Speakers International keynote speakers included: Ian Livingstone - Eidos (UK); Ray Muzyka - Bioware Corp. (Canada); Rob Pardo - Blizzard Entertainment (USA) and Aaron Lieberman - Bungie (USA); Jason Rubin, ex-President, Naughty Dog, Inc. (USA); Bill Roper – Flagship Studios (USA), Laura Fryer, Microsoft Xbox (USA); Ian Fischer, Ensemble Studios; Seamus Blackley, Capital Entertainment Group Inc.; Phil Harrison, Sony Computer Entertainment Europe; Dave Campbell, Discreet; and Lars Gustavsson, DICE. 2005: Final AGDC On the last day of conference proceedings in December 2005, Founder of the Australian Game Developers Conference, John De Margheriti, issued a press release announcing that AIE would stop running the premier industry conference event to allow the industry association, the Game Developers Association of Australia (GDAA), to build a brand new industry conference. Other Australian industry conferences Other Australian Industry conferences include: GO3 Electronic & Entertainment Expo, which started in 2006 Australian E
https://en.wikipedia.org/wiki/Lecithinase
Lecithinase is a type of phospholipase that acts upon lecithin. It can be produced by Clostridium perfringens, Staphylococcus aureus, Pseudomonas aeruginosa or Listeria monocytogenes. C. perfringens alpha toxin (lecithinase) causes myonecrosis and hemolysis. The lecithinase of S. aureus is used in detection of coagulase-positive strains, because of high link between lecithinase activity and coagulase activity.
https://en.wikipedia.org/wiki/Gun%20data%20computer
The gun data computer was a series of artillery computers used by the U.S. Army for coastal artillery, field artillery and anti-aircraft artillery applications. For antiaircraft applications they were used in conjunction with a director computer. Variations M1: This was used by seacoast artillery for major-caliber seacoast guns. It computed continuous firing data for a battery of two guns that were separated by not more than . It utilised the same type of input data furnished by a range section with the then-current (1940) types of position-finding and fire-control equipment. M3: This was used in conjunction with the M9 and M10 directors to compute all required firing data, i.e. azimuth, elevation and fuze time. The computations were made continuously, so that the gun was at all times correctly pointed and the fuze correctly timed for firing at any instant. The computer was mounted in the M13 or M14 director trailer. M4: This was identical to the M3 except for some mechanisms and parts which were altered to allow for different ammunition being used. M8: This was an electronic computer (using vacuum tube technology) built by Bell Labs and used by coast artillery with medium-caliber guns (up to ). It made the following corrections: wind, drift, earth's rotation, muzzle velocity, air density, height of site and spot corrections. M9: This was identical to the M8 except for some mechanisms and parts which were altered to accommodate anti-aircraft ammunition and guns. M10: A ballistics computer, part of the M38 fire control system, for Skysweeper anti-aircraft guns. M13: A ballistics computer for M48 tanks. M14: A ballistics computer for M103 heavy tanks. M15: A part of the M35 field artillery fire-control system, which included the M1 gunnery officer console and M27 power supply. M16: A ballistics computer for M60A1 tanks. M18: FADAC (Field Artillery Digital Automatic Computer), an all-transistorized general-purpose digital computer manufactured by Amelco (T
https://en.wikipedia.org/wiki/Phenethyl%20alcohol
Phenethyl alcohol, or 2-phenylethanol, is an organic compound with the chemical formula . It is a colourless liquid with a pleasant floral odor. It occurs widely in nature, being found in a variety of essential oils. It is slightly soluble in water (2 ml per 100 ml of ), but miscible with most organic solvents. The molecule of phenethyl alcohol consists of a phenethyl group () attached to a hydroxyl group (). Synthesis Phenethyl alcohol is prepared commercially via two routes. Most common is the Friedel-Crafts reaction between benzene and ethylene oxide in the presence of aluminium trichloride. The reaction affords the aluminium alkoxide that is subsequently hydrolyzed to the desired product. The main side product is diphenylethane, which can be avoided by use of excess benzene. Hydrogenation of styrene oxide also affords phenethyl alcohol. Laboratory methods Phenethyl alcohol can also be prepared by the reaction between phenylmagnesium bromide and ethylene oxide: Phenethyl alcohol can also be produced by biotransformation from L-phenylalanine using immobilized yeast Saccharomyces cerevisiae. It is also possible to produce phenethyl alcohol by the reduction of phenylacetic acid using sodium borohydride and iodine in THF. Occurrence and uses Phenethyl alcohol is found in extract of rose, carnation, hyacinth, Aleppo pine, orange blossom, ylang-ylang, geranium, neroli, and champaca. It is also an autoantibiotic produced by the fungus Candida albicans. It is therefore a common ingredient in flavors and perfumery, particularly when the odor of rose is desired. It is used as an additive in cigarettes. It is also used as a preservative in soaps due to its stability in basic conditions. It is of interest due to its antimicrobial properties. See also 1-Phenylethanol Congener Wine chemistry Phenethylamine
https://en.wikipedia.org/wiki/Food%20Technology%20Industrial%20Achievement%20Award
The Food Technology Industrial Achievement Award has been awarded by the Institute of Food Technologists since 1959. It is awarded for the development of an outstanding food process or product that represents a significant advance in the application of food technology to food production. The process or product must have been successfully applied in actual commercial operations between six months and seven years before December 1 in the year of the nomination. Sponsored by Food Technology magazine, award winners receive a plaque from IFT. Winners
https://en.wikipedia.org/wiki/Chinese%20American%20Food%20Society
The Chinese American Food Society (CAFS) is an American-based organization founded in 1974 to develop relationships among Chinese-born food scientists in academia, government, and industry. Its current president is Yao Olive Li, a professor at California State Polytechnic University, Pomona. Founded at the 1974 Institute of Food Technologists (IFT) annual meeting in New Orleans, Louisiana, the organization was first called the "IFT-Chinese Association," then changed to "The Association of Chinese Food Scientists in America" before changing to its current name in 1983. The first award was issued in 1983 and it was awarded to Stephen S. Chang who received the Professional Achievement Award. A Distinguished Service Award was established in 1987, and awards for graduate students and high school students were also created in 1983 and 2002, respectively. The CAFS is partnered with the IFT and its International division, and also publishes an online/electronically issued newsletter two to four times a year. In 2004, CAFS was a main supporter of the Food Summit in China event held jointly with the IFT and the Chinese Institute of Food Science and Technology (CIFST) that was held in Beijing on November 7–10 of that year. At the 2007 IFT Annual Meeting in Chicago, Illinois, they hosted a symposium on Chinese health food involving food science and technology issues on July 27–28. Over the past 45 years, CAFS has been established as a platform for Food Science and Technology professionals to exchange information, collaborate on various taskforce activities, and to promote knowledge advancement, technology innovation and transfer across several continents. Each year there are always one or more CAFS lifetime members who are recognized as IFT Fellows, Annual Achievement Awardees, and recipients of various award and recognition from a broad range of national and international organizations.
https://en.wikipedia.org/wiki/Asset%20turnover
Asset turnover (ATO), total asset turnover, or asset turns is a financial ratio that measures the efficiency of a company's use of its assets in generating sales revenue or sales income to the company. Asset turnover is considered to be a Profitability Ratio, which is a group of financial ratios that measure how efficiently a company uses assets. Asset turnover can be further sub-divided into fixed asset turnover, which measures a company's use of its fixed assets to generate revenue, and working capital turnover, which measures a company's use of its current assets minus liabilities to generate revenue. Total asset turnover ratios can be used to calculate Return On Equity (ROE) figures as part of DuPont analysis. As a financial and activity ratio, and as part of DuPont analysis, asset turnover is a part of company fundamental analysis. Companies with low profit margins tend to have high asset turnover, while those with high profit margins have low asset turnover. Companies in the retail industry tend to have a very high turnover ratio due mainly to cutthroat and competitive pricing. "Sales" is the value of "Net Sales" or "Sales" from the company's income statement "Average Total Assets" is the average of the values of "Total assets" from the company's balance sheet in the beginning and the end of the fiscal period. It is calculated by adding up the assets at the beginning of the period and the assets at the end of the period, then dividing that number by two. This method can produce unreliable results for businesses that experience significant intra-year fluctuations. For such businesses it is advisable to use some other formula for Average Total Assets. Alternatively, "Average Total Assets" can be ending total assets.
https://en.wikipedia.org/wiki/Subsurface%20lithoautotrophic%20microbial%20ecosystem
Subsurface lithoautotrophic microbial ecosystems, or "SLIMEs" (also abbreviated "SLMEs" or "SLiMEs"), are a type of endolithic ecosystems. They are defined by Edward O. Wilson as "unique assemblages of bacteria and fungi that occupy pores in the interlocking mineral grains of igneous rock beneath Earth's surface." Endolithic systems are still at an early stage of exploration. In some cases its biota can support simple invertebrates, most organisms are unicellular. Near-surface layers of rock may contain blue-green algae but most energy comes from chemical synthesis of minerals. The limited supply of energy limits the rates of growth and reproduction. In deeper rock layers microbes are exposed to high pressures and temperatures.
https://en.wikipedia.org/wiki/PoxNora
PoxNora: Battlefield of the Immortals is a multiplayer online game that combines a digital collectible card game with a turn-based strategy game in a fantasy setting. PoxNora was originally launched via Java Web Start through a browser and can be played on Microsoft Windows, Mac OS X, and Linux. The game is free to play with "Sample Battlegroups", and players can purchase additional game pieces, called "runes", and build their own strategies. The game currently includes more than 1600 runes. The game was originally designed and developed by Octopi Media Design Lab, which as of January 16, 2009 became owned by and operated by Sony Online Entertainment. This division of SOE was formerly referred to as SOETucson (because Octopi's headquarters was located in Tucson, Arizona). On April 1, 2011, SOE closed down the Tucson studio along with its Seattle and Denver studios, laying off over 200 employees in the process. Some members of the PoxNora team were moved to the San Diego HQ to continue development. Gameplay Players build decks, called "Battle-groups", on PoxNora'''s website. A Battlegroup consists of thirty runes in any combination of spells, relics, equipment and champions. The game is launched via Java Web Start and players enter one of four game lobbies where they can chat or enter games. A fifth game lobby is reserved solely for trading discussion. Games consist of two players, though matches against AI "bots" are possible now with the release of single-player campaigns, first introduced in the "Path to Conquest" Expansion. Each player's Battlegroup is shuffled (like a deck of cards) and (in standard settings) players reveal two runes per turn (except the first turn, when four are revealed and the second player's first turn, in which five runes are revealed). The game is played on one of 8 different square, "tiled" maps with various obstacles, fonts, and a Shrine representing each player. The first to destroy their opponent's Shrine wins. Players a
https://en.wikipedia.org/wiki/Kwasha%20Lipton
Kwasha Lipton was an employee benefits consulting firm located in Fort Lee, New Jersey. History It was founded in 1944 by H. Charles "Chick" Kwasha and Maurice Lipton. Kwasha Lipton is best known for creating a special type of defined benefit pension plan called a cash balance plan, which it designed for the employees of Bank of America in 1985. Kwasha Lipton implemented dozens of cash balance plans for its other clients, and these plans were ultimately adopted by various companies throughout the United States. Some employees allege that the plans discriminate against older workers in violation of federal age discrimination law because of the way in which younger and older employees earn benefits under cash balance plans. Many cases found plan sponsors liable most were later reversed by appellate courts. However to retroactively approve the plan Congress enacted the Pension Protection Act of 2006. This settled whether the cash balance plan created ab initio violated age discrimination, however plan conversions which created cash balance plans out of existing traditional pensions by freezing accruals of older workers is still an unsettled issue. Kwasha Lipton was acquired in 1996 by Coopers & Lybrand and became the Kwasha Lipton Group of Coopers & Lybrand. Coopers & Lybrand then merged with Price Waterhouse on July 1, 1998 to form PricewaterhouseCoopers. The Kwasha Lipton Group was combined with PricewaterhouseCoopers' other benefit consulting groups to create Unifi, which shortly thereafter was sold to Mellon Financial. Mellon combined Unifi with its Buck Consultants subsidiary to create Mellon HR Solutions, which was sold to Affiliated Computer Services (ACS) for $405 million in 2005 and renamed Buck Consultants. In 2010, ACS was sold to Xerox for $6.4 billion. Buck Consultants was later rebranded as "Buck Consultants at Xerox". Kwasha Lipton owned and occupied a building in Fort Lee, New Jersey. It is the first office building visible to the North after
https://en.wikipedia.org/wiki/Trypticase%20soy%20agar
Trypticase soy agar or tryptone soya agar (TSA) and Trypticase soy broth or tryptone soya broth (TSB) with agar are growth media for the culturing of bacteria. They are general-purpose, nonselective media providing enough nutrients to allow for a wide variety of microorganisms to grow. They are used for a wide range of applications, including culture storage, enumeration of cells (counting), isolation of pure cultures, or simply general culture. TSA contains enzymatic digests of casein and soybean meal, which provide amino acids and other nitrogenous substances, making it a nutritious medium for a variety of organisms. Glucose is the energy source. Sodium chloride maintains the osmotic equilibrium, while dipotassium phosphate acts as buffer to maintain pH. Agar extracted from any number of organisms is used as a gelling agent. The medium may be supplemented with blood to facilitate the growth of more fastidious bacteria or antimicrobial agents to permit the selection of various microbial groups from pure microbiota. As with any media, minor changes may be made to suit specific circumstances. TSA is frequently the base medium of other agar plate types. For example, blood agar plates (BAP) are made by enriching TSA plates with defibrinated sheep blood, and chocolate agar is made through additional cooking of BAP. Nutrient agar is also similar to TSA. One liter of the agar contains: 15 g tryptone 5 g "soytone" – enzymatic digest of soybean meal 5 g sodium chloride 15 g agar
https://en.wikipedia.org/wiki/Chiral%20resolution
Chiral resolution, or enantiomeric resolution, is a process in stereochemistry for the separation of racemic mixture into their enantiomers. It is an important tool in the production of optically active compounds, including drugs. Another term with the same meaning is optical resolution. The use of chiral resolution to obtain enantiomerically pure compounds has the disadvantage of necessarily discarding at least half of the starting racemic mixture. Asymmetric synthesis of one of the enantiomers is one means of avoiding this waste. Crystallization of diastereomeric salts The most common method for chiral resolution involves conversion of the racemic mixture to a pair of diastereomeric derivatives by reacting them with chiral derivatizing agents, also known as chiral resolving agents. The derivatives which are then separated by conventional crystallization, and converted back to the enantiomers by removal of the resolving agent. The process can be laborious and depends on the divergent solubilities of the diastereomers, which is difficult to predict. Often the less soluble diastereomer is targeted and the other is discarded or racemized for reuse. It is common to test several resolving agents. Typical derivatization involves salt formation between an amine and a carboxylic acid. Simple deprotonation then yields back the pure enantiomer. Examples of chiral derivatizing agents are tartaric acid and the amine brucine. The method was introduced (again) by Louis Pasteur in 1853 by resolving racemic tartaric acid with optically active (+)-cinchotoxine. Case study One modern-day method of chiral resolution is used in the organic synthesis of the drug duloxetine: In one of its steps the racemic alcohol 1 is dissolved in a mixture of toluene and methanol to which solution is added optically active (S)-mandelic acid 3. The alcohol (S)-enantiomer forms an insoluble diastereomeric salt with the mandelic acid and can be filtered from the solution. Simple deprotonation with s
https://en.wikipedia.org/wiki/Syntactic%20predicate
A syntactic predicate specifies the syntactic validity of applying a production in a formal grammar and is analogous to a semantic predicate that specifies the semantic validity of applying a production. It is a simple and effective means of dramatically improving the recognition strength of an LL parser by providing arbitrary lookahead. In their original implementation, syntactic predicates had the form “( α )?” and could only appear on the left edge of a production. The required syntactic condition α could be any valid context-free grammar fragment. More formally, a syntactic predicate is a form of production intersection, used in parser specifications or in formal grammars. In this sense, the term predicate has the meaning of a mathematical indicator function. If p1 and p2, are production rules, the language generated by both p1 and p2 is their set intersection. As typically defined or implemented, syntactic predicates implicitly order the productions so that predicated productions specified earlier have higher precedence than predicated productions specified later within the same decision. This conveys an ability to disambiguate ambiguous productions because the programmer can simply specify which production should match. Parsing expression grammars (PEGs), invented by Bryan Ford, extend these simple predicates by allowing "not predicates" and permitting a predicate to appear anywhere within a production. Moreover, Ford invented packrat parsing to handle these grammars in linear time by employing memoization, at the cost of heap space. It is possible to support linear-time parsing of predicates as general as those allowed by PEGs, but reduce the memory cost associated with memoization by avoiding backtracking where some more efficient implementation of lookahead suffices. This approach is implemented by ANTLR version 3, which uses Deterministic finite automata for lookahead; this may require testing a predicate in order to choose between transitions of th
https://en.wikipedia.org/wiki/Bioastronautics
Bioastronautics is a specialty area of biological and astronautical research which encompasses numerous aspects of biological, behavioral, and medical concern governing humans and other living organisms in a space flight environment; and includes design of payloads, space habitats, and life-support systems. In short, it spans the study and support of life in space. Bioastronautics includes many similarities with its sister discipline astronautical hygiene; they both study the hazards that humans may encounter during a space flight. However, astronautical hygiene differs in many respects e.g. in this discipline, once a hazard is identified, the exposure risks are then assessed and the most effective measures determined to prevent or control exposure and thereby protect the health of the astronaut. Astronautical hygiene is an applied scientific discipline that requires knowledge and experience of many fields including bioastronautics, space medicine, ergonomics etc. The skills of astronautical hygiene are already being applied for example, to characterise Moon dust and design the measures to mitigate exposure during lunar exploration, to develop accurate chemical monitoring techniques and use the results in the setting SMACs. Of particular interest from a biological perspective are the effects of reduced gravitational force felt by inhabitants of spacecraft. Often referred to as "microgravity", the lack of sedimentation, buoyancy, or convective flows in fluids results in a more quiescent cellular and intercellular environment primarily driven by chemical gradients. Certain functions of organisms are mediated by gravity, such as gravitropism in plant roots and negative gravitropism in plant stems, and without this stimulus growth patterns of organisms onboard spacecraft often diverge from their terrestrial counterparts. Additionally, metabolic energy normally expended in overcoming the force of gravity remains available for other functions. This may take
https://en.wikipedia.org/wiki/Anophthalmia
Anophthalmia (Greek: ἀνόφθαλμος, "without eye") is the medical term for the absence of one or both eyes. Both the globe (human eye) and the ocular tissue are missing from the orbit. The absence of the eye will cause a small bony orbit, a constricted mucosal socket, short eyelids, reduced palpebral fissure and malar prominence. Genetic mutations, chromosomal abnormalities, and prenatal environment can all cause anophthalmia. Anophthalmia is an extremely rare disease and is mostly rooted in genetic abnormalities. It can also be associated with other syndromes. Causes SOX2 The most common genetic cause for anophthalmia is mutated SOX2 gene. Sox2 anophthalmia syndrome is caused by a mutation in the Sox2 gene that does not allow it to produce the Sox2 protein that regulates the activity of other genes by binding to certain regions of DNA. Without this Sox2 protein, the activity of genes that is important for the development of the eye is disrupted. Sox2 anophthalmia syndrome is an autosomal dominant inheritance, but the majority of patients who have Sox2 anophthalmia are the first in their family history to have this mutation. In certain cases, one parent will possess the mutated gene only in their egg or sperm cell and the offspring will inherit it through that. This is called germline mosaicism. There are at least 33 mutations in the Sox2 gene that have been known to cause anophthalmia. Some of these gene mutations will cause the Sox2 protein not to be formed, while other mutations will yield a non-functional version of this protein. RBP4 RBP4 has recently been linked to autosomal dominant form of anophthalmia. This form of anophthalmia has variable penetrance and a unique maternal inheritance effect that is rooted in pregnancy. Specifically, the disease only occurs when a mother and fetus both carry a RBP4 mutation which predisposes the fetus to vitamin A deficiency (a known environmental risk factor for anophthalmia) during pregnancy. If Vitamin A deficiency occur
https://en.wikipedia.org/wiki/A-MAC
In television electronics, A-MAC carries digital information: sound, and data-teletext on an FM subcarrier at 7 MHz. Since the vision bandwidth of a standard MAC signal is 8.4 MHz, the horizontal resolution on A-MAC has to be reduced to make room for the 7 MHz carrier. A-MAC has not been used in service. Technical details MAC transmits luminance and chrominance data separately in time rather than separately in frequency (as other analog television formats do, such as composite video). Audio and Scrambling (selective access) Audio, in a format similar to NICAM was transmitted digitally rather than as an FM subcarrier. The MAC standard included a standard scrambling system, EuroCrypt, a precursor to the standard DVB-CSA encryption system. TV transmission systems Analog high-definition television systems PAL, what MAC technology tried to replace SECAM, what MAC technology tried to replace A-MAC B-MAC C-MAC D-MAC E-MAC S-MAC D2-MAC HD-MAC, an early high-definition television standard allowing for 2048x1152 resolution. DVB-S, MAC technology was replaced by this standard DVB-T, MAC technology was replaced by this standard External links Multiplexed Analogue Components in "Analog TV Broadcast Systems" by Paul Schlyter Video formats Television technology
https://en.wikipedia.org/wiki/B-MAC
B-MAC is a form of analog video encoding, specifically a type of Multiplexed Analogue Components (MAC) encoding. MAC encoding was designed in the mid 80s for use with Direct Broadcast Satellite systems. Other analog video encoding systems include NTSC, PAL and SECAM. Unlike the FDM method used in those, MAC encoding uses a TDM method. B-MAC was a proprietary MAC encoding used by Scientific-Atlanta for encrypting broadcast video services; the full name was "Multiple Analogue Component, Type B". B-MAC uses teletext-style non-return-to-zero (NRZ) signaling with a capacity of 1.625 Mbit/s. The video and audio/data signals are therefore combined at baseband. Both PAL (626/50) and NTSC (525/60) versions of B-MAC were developed and used. User base (PAL/NTSC zones) This system was used in South Africa and Australia (for TVRO until 2000). B-MAC was used for satellite broadcasts of the American Forces Radio and Television Service from the early 1980s until 1996-1997 when the analogue standard was replaced by the digital PowerVu system. B-MAC has not been used for DTH applications since Primestar switched to an all-digital delivery system in the mid-1990s. Technical details MAC transmits luminance and chrominance data separately in time rather than separately in frequency (as other analog television formats do, such as composite video). Audio and Scrambling (selective access) Audio, in a format similar to NICAM was transmitted digitally rather than as an FM subcarrier. The MAC standard included a standard scrambling system, EuroCrypt, a precursor to the standard DVB-CSA encryption system. See also Analog high-definition television systems PAL, what MAC technology tried to replace SECAM, what MAC technology tried to replace A-MAC B-MAC C-MAC D-MAC E-MAC S-MAC D2-MAC HD-MAC, an early high-definition television standard allowing for 2048x1152 resolution. DVB-S, MAC technology was replaced by this standard DVB-T, MAC technology was replaced by th
https://en.wikipedia.org/wiki/C-MAC
C-MAC is the television technology variant approved by the European Broadcasting Union (EBU) for satellite transmissions. The digital information is modulated using 2-4PSK (phase-shift keying), a variation of quadrature PSK where only two of the phaser angles (±90°) are used. The data capacity for C-MAC is 3 Mbit/s. C-MAC data has to be sent to the transmitter separately from the vision. The transmitter switches between FM (vision) and PSK (sound/data) modulation during each television line period. C-MAC variants : E-MAC E-MAC (Extended MAC) is 16:9 version of C-MAC. Originally E-MAC was designed for 15:9 pictures, it later adopted the 16:9 aspect ratio. In E-MAC all the 4:3 information is transmitted exactly as in C-MAC so that C-MAC receivers are still compatible. E-MAC hides extra luminance and chrominance information in the field blanking interval and parts of the line blanking interval. E-MAC has a lower data capacity because luminance is hidden where data would usually be located. A 'steering' signal is transmitted to indicate to the 16:9 receiver whereabouts the 4:3 picture information. E-MAC receivers stitch the 4:3 and helper wide-screen data into a seamless 16:9 picture. Technical details MAC transmits luminance and chrominance data separately in time rather than separately in frequency (as other analog television formats do, such as composite video). Audio and Scrambling (selective access) Audio, in a format similar to NICAM was transmitted digitally rather than as an FM sub-carrier. The MAC standard included a standard scrambling system, Euro-Crypt, a precursor to the standard DVB-CSA encryption system. See also TV transmission systems Analog high-definition television systems DVB-S DVB-T Multiplexed Analogue Components PAL SECAM
https://en.wikipedia.org/wiki/D-MAC
Among the family of MAC or Multiplexed Analogue Components systems for television broadcasting, D-MAC is a reduced bandwidth variant designed for transmission down cable. The data is duobinary coded with a data burst rate of 20.25Mbit/s so that 0° as well as ±90° phasors are used. D-MAC has a bandwidth of 8.4 MHz versus 27 MHz for C-MAC. Most cable systems work on EBU 7 MHz channel spacing, so this approach did not work universally. D-MAC's bandwidth problems were later fixed by D2-MAC. D2-MAC: A fix for D-MAC D-MAC consumed too much bandwidth for many applications, so D2-MAC was designed for European cable TV systems. Luminance and chrominance MAC transmits luminance and chrominance data separately in time rather than separately in frequency (as other analog television formats do, such as composite video). Audio and scrambling (selective access) Audio, in a format similar to NICAM was transmitted digitally rather than as an FM subcarrier. The MAC standard included a standard scrambling system, EuroCrypt, a precursor to the standard DVB-CSA encryption system. History and politics MAC was developed by the UK's Independent Broadcasting Authority (IBA) and in 1982 was adopted as the transmission format for the UK's forthcoming direct broadcast satellite (DBS) television services (eventually provided by British Satellite Broadcasting). The following year MAC was adopted by the European Broadcasting Union (EBU) as the standard for all DBS. By 1986, despite there being two standards, D-MAC and D2-MAC, favoured by different countries in Europe, an EU Directive imposed MAC on the national DBS broadcasters, to provide a stepping stone from analogue PAL and Secam formats to the eventual high definition and digital television of the future, with European TV manufacturers in a privileged position to provide the equipment required. However, the Astra satellite system was also starting up at this time (the first satellite, Astra 1A was launched in 1989) and that ope
https://en.wikipedia.org/wiki/D2-MAC
D2-MAC is a satellite television transmission standard, a member of Multiplexed Analogue Components family. It was created to solve D-MAC's bandwidth usage by further reducing it, allowing usage of the system on cable and satellite broadcast. It could carry four high quality (15 kHz bandwidth) sound channels or eight lower quality audio channels. It was adopted by Scandinavian, German and French satellite broadcasts (CNBC Europe, TV3 (Sweden), TV3 (Denmark), EuroSport, NRK 1, TV-Sat 2, TDF 1, TDF 2, etc). The system was used until July 2006 in Scandinavia and until the mid-1990s for German and French sound channels. Technical details MAC transmits luminance and chrominance data separately in time rather than separately in frequency (as other analog television formats do, such as composite video). Audio, in a format similar to NICAM was transmitted digitally rather than as an FM sub-carrier. The MAC standard included a standard scrambling system, EuroCrypt, a precursor to the standard DVB-CSA encryption system. D2-MAC uses half the data rate of D-MAC D2-MAC has a reduced vision bandwidth, about 1/2 that of D-MAC. D2-MAC retains most of the quality of a D-MAC signal—but consumes only 5 MHz of bandwidth. History and politics MAC was developed by the UK's Independent Broadcasting Authority (IBA) and in 1982 was adopted as the transmission format for the UK's forthcoming direct broadcast satellite (DBS) television services (eventually provided by British Satellite Broadcasting). The following year MAC was adopted by the European Broadcasting Union (EBU) as the standard for all DBS. By 1986, despite there being two standards, D-MAC and D2-MAC, favoured by different countries in Europe, an EU Directive imposed MAC on the national DBS broadcasters, to provide a stepping stone from analogue PAL and SECAM formats to the eventual high definition and digital television of the future, with European TV manufacturers in a privileged position to provide the equipment req
https://en.wikipedia.org/wiki/C-squares
C-squares (acronym for the Concise Spatial QUery And REpresentation System) is a system of spatially unique, location-based identifiers (geocodes) for areas on the surface of the earth, represented as cells from a latitude- and longitude-based Discrete Global Grid at a hierarchical set of resolution steps, obtained by progressively subdividing 10×10 degree World Meteorological Organization squares; the term "c-square" is also available for use to designate any component cell of the grid. Individual cell identifiers incorporate literal values of latitude and longitude in an interleaved notation (producing grid resolutions of 10, 1, 0.1 degrees, etc.), together with additional digits that support intermediate grid resolutions of 5, 0.5, 0.05 degrees, etc. The system was initially designed to represent data "footprints" or spatial extents in a more flexible manner than a standard minimum bounding rectangle, and to support "lightweight", text-based spatial querying; it can also provide a set of identifiers for grid cells used for assembly, storage and analysis of spatially organised data, in a unified notation that transcends national or jurisdictional boundaries. Dataset extents expressed in c-squares notation can be visualised using a web-based utility, the c-squares mapper, an online instance of which is currently provided by CSIRO Oceans and Atmosphere in Australia. C-squares codes and associated published software are free to use and the software is released under version 2 of the GNU General Public License (GPL), a licence of the Free Software Foundation. History The c-squares method was developed by Tony Rees at CSIRO Oceans and Atmosphere in Australia (then "CSIRO Marine Research") in 2001–2, initially as a method for spatial indexing, rapid query, and compact storage and visualization of dataset spatial "footprints" in an agency-specific metadata directory (data catalogue); it was first publicly announced at the 2002 "EOGEO" Technical Workshop held at Ispra,
https://en.wikipedia.org/wiki/Osteophagy
Osteophagy is the practice in which animals, usually herbivores, consume bones. Most vegetation around the world lacks sufficient amounts of phosphate. Phosphorus is an essential mineral for all animals, as it plays a major role in the formation of the skeletal system, and is necessary for many biological processes including: energy metabolism, protein synthesis, cell signaling, and lactation. Phosphate deficiencies can cause physiological side effects, especially pertaining to the reproductive system, as well as side effects of delayed growth and failure to regenerate new bone. The importance of having sufficient amounts of phosphorus further resides in the physiological importance of maintaining a proper phosphorus to calcium ratio. Having a Ca:P ratio of 2:1 is important for the absorption of these minerals, as deviations from this optimal ratio can inhibit their absorption. Dietary calcium and phosphorus ratio, along with vitamin D, regulates bone mineralization and turnover by affecting calcium and phosphorus transport and absorption in the intestine. It has been suggested that osteophagy is an innate behavior that allows animals to supplement their phosphorus and calcium uptake in order to avoid the costly effects of deficiencies in these minerals. Osteophagic behavior has been observed in pastoral and wild animals, most notably ungulates and other herbivores, for over two hundred years. Osteophagy has been inferred from archaeological studies of dental wear in Pleistocene fossils dating back 780,000 years. It has been seen in domestic animals, as well as red deer, camels, giraffes, wildebeest, antelopes, tortoises, and grizzly bears. Due to differences in tooth structure, herbivores tend to chew old dry bones that are easier to break, while carnivores prefer to chew softer fresh bones. Variations of the behavior have also been observed in humans. While osteophagy has been regarded as a beneficial behavior to combat mineral deficiencies in animals, osteopha
https://en.wikipedia.org/wiki/Ambient%20isotopy
In the mathematical subject of topology, an ambient isotopy, also called an h-isotopy, is a kind of continuous distortion of an ambient space, for example a manifold, taking a submanifold to another submanifold. For example in knot theory, one considers two knots the same if one can distort one knot into the other without breaking it. Such a distortion is an example of an ambient isotopy. More precisely, let and be manifolds and and be embeddings of in . A continuous map is defined to be an ambient isotopy taking to if is the identity map, each map is a homeomorphism from to itself, and . This implies that the orientation must be preserved by ambient isotopies. For example, two knots that are mirror images of each other are, in general, not equivalent. See also Isotopy Regular homotopy Regular isotopy
https://en.wikipedia.org/wiki/Prime%20form
In algebraic geometry, the Schottky–Klein prime form E(x,y) of a compact Riemann surface X depends on two elements x and y of X, and vanishes if and only if x = y. The prime form E is not quite a holomorphic function on X × X, but is a section of a holomorphic line bundle over this space. Prime forms were introduced by Friedrich Schottky and Felix Klein. Prime forms can be used to construct meromorphic functions on X with given poles and zeros. If Σniai is a divisor linearly equivalent to 0, then ΠE(x,ai)ni is a meromorphic function with given poles and zeros. See also Fay's trisecant identity
https://en.wikipedia.org/wiki/Amanita%20xanthocephala
The vermilion grisette, also known as pretty grisette or vermilion amanita (Amanita xanthocephala) is a colourful mushroom of the genus Amanita. However, although it is often referred to by the common name "grisette", it is not closely related to other edible species that carry this common name, such as Amanita vaginata and Amanita fulva. It belongs to the same group of Amanita as A. muscaria and is reported to be toxic. It derives its specific epithet xanthocephala from the Greek xanthos/ξανθοѕ "yellow" and kephale/κεφαλη "head". It is a ringless mushroom with a yellowish- to reddish-orange cap up to in diameter, with deeper colour toward the centre, and paler similar-coloured warts. The gills and slim ringless stipe are pale yellow or white. The white volva has a neat outturned lip and is often bordered with orange or yellow. Amanita xanthocephala lives in an ectomycorrhizal relationship with Eucalyptus. It is found in eucalypt forests in the southwest of Western Australia, as well as southeastern Australia from near Adelaide around to Southeast Queensland. At one stage this fungus was known as Amanitopsis pulchella, a small genus that all grisettes (ringless Amanita species) were placed in. However, this genus has been later sunk back into Amanita. Unlike most ringless Amanita, which are part of Amanita section Vaginatae (A. vaginata and allies), A. xanthcephala belongs to Amanita section Amanita (A. muscaria and allies). There is one report of a person being quite ill after tasting a small piece of it in 1997. See also List of Amanita species
https://en.wikipedia.org/wiki/Grid%20%28spatial%20index%29
In the context of a spatial index, a grid or mesh is a regular tessellation of a manifold or 2-D surface that divides it into a series of contiguous cells, which can then be assigned unique identifiers and used for spatial indexing purposes. A wide variety of such grids have been proposed or are currently in use, including grids based on "square" or "rectangular" cells, triangular grids or meshes, hexagonal grids, and grids based on diamond-shaped cells. A "global grid" is a kind of grid that covers the entire surface of the globe. Types of grids Square or rectangular grids are frequently used for purposes such as translating spatial information expressed in Cartesian coordinates (latitude and longitude) into and out of the grid system. Such grids may or may not be aligned with the grid lines of latitude and longitude; for example, Marsden Squares, World Meteorological Organization squares, c-squares and others are aligned, while Universal Transverse Mercator coordinate system and various local grid based systems such as the British national grid reference system are not. In general, these grids fall into two classes, "equal angle" or "equal area". Grids that are "equal angle" have cell sizes that are constant in degrees of latitude and longitude but are unequal in area (particularly with varying latitude). Grids that are "equal area" (statistical grids), that have cell sizes that are constant in distance on the ground (e.g. 100 km, 10 km) but not in degrees of longitude, in particular. A commonly used triangular grid is the "Quaternary Triangular Mesh" (QTM), which was developed by Geoffrey Dutton in the early 1980s. It eventually resulted in a thesis entitled "A Hierarchical Coordinate System for Geoprocessing and Cartography" that was published in 1999. This grid was also employed as the basis of the rotatable globe that forms part of the Microsoft Encarta product. Hexagonal grids may also be used. In general, triangular and hexagonal grids are constructed so
https://en.wikipedia.org/wiki/Bee%20Card%20%28game%20cartridge%29
A is a ROM cartridge developed by Hudson Soft as a software distribution medium for MSX computers. Bee Cards are approximately the size of a credit card, but thicker. Compared to most game cartridges, the Bee Card is small and compact. Bee Cards were released in Japan and in Europe, but not in North America because the MSX was unsuccessful in North America. However, Atari Corporation adopted the Bee Card for the Atari Portfolio, a handheld PC released in 1989 in North America. Bee Cards were also used by some Korg Synthesizers and workstations as external storage of user content like sound programs or song data. Even though these systems all use Bee Cards, they are incompatible with each other. Only a small number of MSX software titles were published on Bee Card: six in Japan, and only two in Europe and Italy. In order to accept a Bee Card, the cartridge slot of the MSX had to be fitted with a removable adapter: the Hudson Soft BeePack. The first mass-produced Bee Cards, however, were EEPROM telephone cards manufactured by Mitsubishi Plastics; these were first sold in Japan in 1985. The trade names Bee Card and Bee Pack derive from Hudson Soft's corporate logo, which features a cartoon bee. MSX software published on Bee Card Hudson Soft and other software publishers distributed at least eleven MSX software titles on Bee Card: HuCard Hudson Soft later collaborated with NEC to develop the PC Engine video game console. The companies elected to use Hudson Soft's slim ROM cartridge technology to distribute PC Engine software. Hudson Soft adapted the design for their needs, and produced the HuCard. HuCards are slightly thicker than Bee Cards; also, whereas a Bee Card has 32 pins, a HuCard has 38.
https://en.wikipedia.org/wiki/General%20game%20playing
General game playing (GGP) is the design of artificial intelligence programs to be able to play more than one game successfully. For many games like chess, computers are programmed to play these games using a specially designed algorithm, which cannot be transferred to another context. For instance, a chess-playing computer program cannot play checkers. General game playing is considered as a necessary milestone on the way to artificial general intelligence. General video game playing (GVGP) is the concept of GGP adjusted to the purpose of playing video games. For video games, game rules have to be either learnt over multiple iterations by artificial players like TD-Gammon, or are predefined manually in a domain-specific language and sent in advance to artificial players like in traditional GGP. Starting in 2013, significant progress was made following the deep reinforcement learning approach, including the development of programs that can learn to play Atari 2600 games as well as a program that can learn to play Nintendo Entertainment System games. The first commercial usage of general game playing technology was Zillions of Games in 1998. General game playing was also proposed for trading agents in supply chain management thereunder price negotiation in online auctions from 2003 on. History In 1992, Barney Pell defined the concept of Meta-Game Playing, and developed the "MetaGame" system. This was the first program to automatically generate game rules of chess-like games, and one of the earliest programs to use automated game generation. Pell then developed the system Metagamer. This system was able to play a number of chess-like games, given game rules definition in a special language called Game Description Language (GDL), without any human interaction once the games were generated. In 1998, the commercial system Zillions of Games was developed by Jeff Mallett and Mark Lefler. The system used a LISP-like language to define the game rules. Zillions of Game
https://en.wikipedia.org/wiki/List%20of%20BPEL%20engines
This is a list of notable Business Process Execution Language (BPEL) and Business Process Model and Notation (BPMN) engines.
https://en.wikipedia.org/wiki/ObjectVision
ObjectVision was a forms-based programming language and environment for Windows 3.x developed by Borland. The latest version, 2.1, was released in 1992. An ObjectVision application is composed by forms designed in a graphic way that contains objects and events to provide interactivity. Forms are connected together with logic in the form of decision trees. ObjectVision applications also can interact with databases using multiple engines, like Paradox and dBase. A finished project is saved as an OVD file, that is executed by an interpreted runtime that can be freely distributed. ObjectVision was not used broadly except in some niche segments, but the visual programming ideas were the basis for Borland Delphi.
https://en.wikipedia.org/wiki/Cytohet
In genetics, a cytohet (or heteroplasmon) is a eukaryotic cell whose non-nuclear genome is heterozygous. The non-nucleic genome of eukaryotic cells exists in cytoplasmic organelles, namely the chloroplasts (only in plant cells) and the mitochondria (in all eukaryotic cells). Most of the genes in the mitochondria code for respiration-related proteins, and most of the genes in the chloroplasts code for photosynthesis-related proteins. The cytoplasmic genome, in contrast with the nucleic genome, exists in many copies in each cell: each cell contains numerous mitochondria and/or chloroplasts, and each such organelle contains multiple copies of its chromosome. Mutations in the cytoplasmic genome occur spontaneously and at a much higher rate than in the nucleus, since the mitochondria and chloroplasts are exposed to high concentrations of reactive oxygen species (ROS, by-products of respiration and photosynthesis). Mitochondria and chloroplasts with mutant genes have the ability to cause wildtype alleles in other mitochondria and chloroplasts to become mutant as well; the way in which this is done is still not clear. A certain cell in which a mutant gene exists only in some of the organelles, whereas the wildtype allele exists in the rest, is a cytohet (or heteroplasmon). External links http://www.oxfordreference.com/view/10.1093/acref/9780195307610.001.0001/acref-9780195307610-e-1540 Mitochondrial genetics
https://en.wikipedia.org/wiki/Mary%20Celine%20Fasenmyer
Mary Celine Fasenmyer, RSM (October 4, 1906, Crown, Pennsylvania – December 27, 1996, Erie, Pennsylvania) was an American mathematician and Catholic religious sister. She is most noted for her work on hypergeometric functions and linear algebra. Biography Fasenmyer grew up in Pennsylvania's oil country, and displayed mathematical talent in high school. For ten years after her graduation she taught and studied at Mercyhurst College in Erie, where she joined the Sisters of Mercy. She pursued her mathematical studies in Pittsburgh and the University of Michigan, obtaining her doctorate in 1946 under the direction of Earl Rainville, with a dissertation entitled Some Generalized Hypergeometric Polynomials. After earning her Ph.D., Fasenmyer published two papers which expanded on her doctorate work. These would be further elaborated by Doron Zeilberger and Herbert Wilf into "WZ theory", which allowed computerized proof of many combinatorial identities. After this, she returned to Mercyhurst to teach and did not engage in further research. Fasenmyer died in 1996. "Sister Celine's" method Fasenmyer is most remembered for the method that bears her name, first described in her Ph.D. thesis concerning recurrence relations in hypergeometric series. The thesis demonstrated a purely algorithmic method to find recurrence relations satisfied by sums of terms of a hypergeometric polynomial, and requires only the series expansions of the polynomial. The beauty of her method is that it lends itself readily to computer automation. The work of Wilf and Zeilberger generalized the algorithm and established its correctness. The hypergeometric polynomials she studied are called Sister Celine's polynomials.
https://en.wikipedia.org/wiki/Extramammary%20Paget%27s%20disease
Extramammary Paget's disease (EMPD) is a rare and slow-growing malignancy which occurs within the epithelium and accounts for 6.5% of all Paget's disease. The clinical presentation of this disease is similar to the characteristics of mammary Paget's disease (MPD). However, unlike MPD, which occurs in large lactiferous ducts and then extends into the epidermis, EMPD originates in glandular regions rich in apocrine secretions outside the mammary glands. EMPD incidence is increasing by 3.2% every year, affecting hormonally-targeted tissues such as the vulva and scrotum. In women, 81.3% of EMPD cases are related to the vulva, while for men, 43.2% of the manifestations present at the scrotum. The disease can be classified as being either primary or secondary depending on the presence or absence of associated malignancies. EMPD presents with typical symptoms such as scaly, erythematous, eczematous lesions accompanied by itchiness. In addition to this, 10% of patients are often asymptomatic. As a consequence, EMPD has high rates of misdiagnoses and delayed diagnoses. There are a variety of treatment options available, but most are unsuccessful. If caught early and treated, prognosis is generally good. Presentation Patients with EMPD present with typical symptoms, similar to MPD, such as severe itchiness (also called pruritus), rash, plaque formation, burning sensation, pain and tenderness. These symptoms are often confused for dermatitis or eczema. 10% of patients are asymptomatic resulting in delayed diagnosis. In rare cases bleeding can also be seen. Disease of the vulva Vulvar Paget's disease affect women and presents as erythematous (red), eczematous lesions. It is itchy and sometimes pain can be associated with the affected area. The lesion is clearly separated from normal skin in most cases, and sometimes scattered areas of white scale can be present, giving a "strawberries and cream" appearance. Involvement may be extensive including the perianal region, genit
https://en.wikipedia.org/wiki/Aitken%27s%20delta-squared%20process
In numerical analysis, Aitken's delta-squared process or Aitken extrapolation is a series acceleration method, used for accelerating the rate of convergence of a sequence. It is named after Alexander Aitken, who introduced this method in 1926. Its early form was known to Seki Kōwa (end of 17th century) and was found for rectification of the circle, i.e. the calculation of π. It is most useful for accelerating the convergence of a sequence that is converging linearly. Definition Given a sequence , one associates with this sequence the new sequence which can, with improved numerical stability, also be written as or equivalently as where and for Obviously, is ill-defined if contains a zero element, or equivalently, if the sequence of first differences has a repeating term. From a theoretical point of view, if that occurs only for a finite number of indices, one could easily agree to consider the sequence restricted to indices with a sufficiently large . From a practical point of view, one does in general rather consider only the first few terms of the sequence, which usually provide the needed precision. Moreover, when numerically computing the sequence, one has to take care to stop the computation when rounding errors in the denominator become too large, where the Δ2 operation may cancel too many significant digits. (It would be better for numerical calculation to use rather than Properties Aitken's delta-squared process is a method of acceleration of convergence, and a particular case of a nonlinear sequence transformation. Convergence of to limit is called "linear" if there is some number for which Which means that the distance between the sequence and its limit shrinks by nearly the same proportion on every step, and that rate of reduction becomes closer to being constant with every step. (This is also called "geometric convergence"; this form of convergence is common for power series.) Aitken's method will accelerate the sequence if is no
https://en.wikipedia.org/wiki/Ars%20Magna%20%28Cardano%20book%29
The Ars Magna (The Great Art, 1545) is an important Latin-language book on algebra written by Gerolamo Cardano. It was first published in 1545 under the title Artis Magnae, Sive de Regulis Algebraicis Liber Unus (Book number one about The Great Art, or The Rules of Algebra). There was a second edition in Cardano's lifetime, published in 1570. It is considered one of the three greatest scientific treatises of the early Renaissance, together with Copernicus' De revolutionibus orbium coelestium and Vesalius' De humani corporis fabrica. The first editions of these three books were published within a two-year span (1543–1545). History In 1535 Niccolò Fontana Tartaglia became famous for having solved cubics of the form x3 + ax = b (with a,b > 0). However, he chose to keep his method secret. In 1539, Cardano, then a lecturer in mathematics at the Piatti Foundation in Milan, published his first mathematical book, Pratica Arithmeticæ et mensurandi singularis (The Practice of Arithmetic and Simple Mensuration). That same year, he asked Tartaglia to explain to him his method for solving cubic equations. After some reluctance, Tartaglia did so, but he asked Cardano not to share the information until he published it. Cardano submerged himself in mathematics during the next several years working on how to extend Tartaglia's formula to other types of cubics. Furthermore, his student Lodovico Ferrari found a way of solving quartic equations, but Ferrari's method depended upon Tartaglia's, since it involved the use of an auxiliary cubic equation. Then Cardano became aware of the fact that Scipione del Ferro had discovered Tartaglia's formula before Tartaglia himself, a discovery that prompted him to publish these results. Contents The book, which is divided into forty chapters, contains the first published algebraic solution to cubic and quartic equations. Cardano acknowledges that Tartaglia gave him the formula for solving a type of cubic equations and that the same formula had b
https://en.wikipedia.org/wiki/Selective%20reabsorption
Selective reabsorption is the process whereby certain molecules (e.g. ions, glucose and amino acids), after being filtered out of the capillaries along with nitrogenous waste products (i.e. urea) and water in the glomerulus, are reabsorbed from the filtrate as they pass through the nephron. Selective reabsorbtion occurs in the PCT (proximal convoluted tubule). The PCT is highly permeable meaning it is easy for molecules to diffuse through it. A basic outline of the process The co-transport sodium-potassium pump actively transports sodium out of the PCT (proximal convoluted tubule) wall (using energy from converting ATP to ADP + Pi) to maintain a low Na+ concentration gradient in the wall. This low concentration gradient means that Na+ ions from the glomerulus filtrate can easily passively diffuse into the wall of the PCT. However, the Na+ ions cannot diffuse freely across the membrane, but can only enter through special transporter (carrier) proteins in the membrane of the wall. There are several different kinds of these transporter proteins, each of which transports another molecule, such as glucose or amino acids. The concentration gradient for the sodium provides the energy to pull in these other molecules into the wall of the PCT. As the substances listed above (Na+ ions, amino acids and glucose) enter the wall of the PCT, so does 65–70% of the water in the glomerulus filtrate via osmosis. Water can move freely through the wall of the PCT (it does not require a transporter protein). Nearly all the rest of the water is reabsorbed into the blood in the loop of Henle and the collecting duct system. However, as urea is a small molecule it can pass easily through the membrane of the PCT wall. As the concentration of urea in the filtrate is significantly higher than in the blood, around 50% of urea on the filtrate is reabsorbed.
https://en.wikipedia.org/wiki/Steve%20Jobs
Steven Paul Jobs (February 24, 1955 – October 5, 2011) was an American businessman who was the co-founder, chairman, and CEO of Apple; the chairman and majority shareholder of Pixar; a member of The Walt Disney Company's board of directors following its acquisition of Pixar; and the founder, chairman, and CEO of NeXT. He was a pioneer of the personal computer revolution of the 1970s and 1980s, along with his early business partner and fellow Apple co-founder Steve Wozniak. Jobs was born in San Francisco to a Syrian father and German-American mother. He was adopted shortly after his birth. Jobs attended Reed College in 1972 before withdrawing that same year. In 1974, he traveled through India seeking enlightenment before later studying Zen Buddhism. He and Wozniak co-founded Apple in 1976 to sell Wozniak's Apple I personal computer. Together the duo gained fame and wealth a year later with production and sale of the Apple II, one of the first highly successful mass-produced microcomputers. Jobs saw the commercial potential of the Xerox Alto in 1979, which was mouse-driven and had a graphical user interface (GUI). This led to the development of the unsuccessful Apple Lisa in 1983, followed by the breakthrough Macintosh in 1984, the first mass-produced computer with a GUI. The Macintosh introduced the desktop publishing industry in 1985 with the addition of the Apple LaserWriter, the first laser printer to feature vector graphics. In 1985, Jobs departed Apple after a long power struggle with the company's board and its then-CEO, John Sculley. That same year, Jobs took a few Apple employees with him to found NeXT, a computer platform development company that specialized in computers for higher-education and business markets. In addition, he helped to develop the visual effects industry when he funded the computer graphics division of George Lucas's Lucasfilm in 1986. The Graphics Group eventually spun off independently as Pixar, producing the first 3D computer-animate
https://en.wikipedia.org/wiki/Pangloss%20Collection
The Pangloss Collection is a digital library whose objective is to store and facilitate access to audio recordings in endangered languages of the world. Developed by the LACITO centre of CNRS in Paris, the collection provides free online access to documents of connected, spontaneous speech, in otherwise little-documented languages of all continents. Principles A sound archive with synchronized transcriptions For the science of linguistics, language is first and foremost spoken language. The medium of spoken language is sound. The Pangloss Collection gives access to original recordings simultaneously with transcriptions and translations, as a resource for further research. After being recorded in its cultural context, texts have been transcribed in collaboration with native speakers. A structured, open architecture The archived data is structured in accordance with the latest data-processing standards, as open architecture, in an open format, and may be downloaded under a Creative Commons license. The software used to prepare and disseminate it is open-source. The Pangloss Collection is a member of the OLAC network of archival repositories and of the Digital Endangered Languages and Music Archive Network (DELAMAN). History The collection was initially called the LACITO Archive. The project originated in 1996 from the collaboration of Boyd Michailovsky, linguist at LACITO, with John B. Lowe, engineer; they were later joined by Michel Jacobson, engineer, who developed some tools for the project, and brought it online. The purpose of the archive was “to conserve, and to make available for research, recorded and transcribed oral traditions and other linguistic materials in (mainly) unwritten languages, giving simultaneous access to sound recordings and text annotation.” The earliest archived corpora in the collection were languages from Nepal, from New Caledonia, from eastern Africa and French Guiana. The archive has grown steadily since the early 2000s, incorp
https://en.wikipedia.org/wiki/Microsatellite%20instability
Microsatellite instability (MSI) is the condition of genetic hypermutability (predisposition to mutation) that results from impaired DNA mismatch repair (MMR). The presence of MSI represents phenotypic evidence that MMR is not functioning normally. MMR corrects errors that spontaneously occur during DNA replication, such as single base mismatches or short insertions and deletions. The proteins involved in MMR correct polymerase errors by forming a complex that binds to the mismatched section of DNA, excises the error, and inserts the correct sequence in its place. Cells with abnormally functioning MMR are unable to correct errors that occur during DNA replication and consequently accumulate errors. This causes the creation of novel microsatellite fragments. Polymerase chain reaction-based assays can reveal these novel microsatellites and provide evidence for the presence of MSI. Microsatellites are repeated sequences of DNA. These sequences can be made of units of 1 to 6 base pairs in length that are repeated and reside adjacent to each other in the genome. Although the length of microsatellites can vary from person to person and contributes to the individual DNA "fingerprint", each individual has microsatellites of a set length. The most common microsatellite in humans is a dinucleotide repeat of the nucleotides C and A, which occurs tens of thousands of times across the genome. Microsatellites are also known as simple sequence repeats (SSRs). Structure Microsatellite instability structure consists of repeated nucleotides, most often seen as GT/CA repeats. Researchers have yet to confirm the precise definition of the MSI structure. While all researchers agree that microsatellites are repeat sequences, the lengths of the sequences remain in question. Some research suggests that MSIs are short tandem DNA repeat sequences of one to six base pairs throughout the genome, while other research suggests that the range may be two to five. Although researchers do not
https://en.wikipedia.org/wiki/Kopp%27s%20law
Kopp's law can refer to either of two relationships discovered by the German chemist Hermann Franz Moritz Kopp (1817–1892). Kopp found "that the molecular heat capacity of a solid compound is the sum of the atomic heat capacities of the elements composing it; the elements having atomic heat capacities lower than those required by the Dulong–Petit law retain these lower values in their compounds." In studying organic compounds, Kopp found a regular relationship between boiling points and the number of CH2 groups present. Kopp–Neumann law The Kopp–Neumann law, named for Kopp and Franz Ernst Neumann, is a common approach for determining the specific heat C (in J·kg−1·K−1) of compounds using the following equation: where N is the total number of compound constituents, and Ci and fi denote the specific heat and mass fraction of the i-th constituent. This law works surprisingly well at room-temperature conditions, but poorly at elevated temperatures. See also Rule of mixtures
https://en.wikipedia.org/wiki/HP%20OpenMail
HP OpenMail, also known simply as OpenMail, was an enterprise email messaging and collaboration product from Hewlett-Packard. It was known for its ability to interconnect several other APIs and protocols, including MAPI, cc:Mail, SMTP and MIME, and was originally based on the OSI standards such as X.400. In addition to email, it also supported directory, public folder, and contact-management functionality. It was notable for being supported not only on HP-UX, but also on IBM's AIX, Sun Microsystems' Solaris and Linux, which increased its attraction for enterprise customers. There were also lesser–used versions for SCO Unix, DG/UX, Ultrix and Windows NT. History From the initial designs in 1987, OpenMail was primarily designed and developed at HP's now-demolished Pinewood offices, near Wokingham, England (also the home of OpenMail's predecessor product family, HP DeskManager and the original developers of HP NewWave). HP stopped selling OpenMail to new customers in November 2001. HP subsequently twice licensed the source, allowing it to serve as the foundation for the Scalix and Samsung Contact products. External links
https://en.wikipedia.org/wiki/Sequential%20quadratic%20programming
Sequential quadratic programming (SQP) is an iterative method for constrained nonlinear optimization which may be considered a quasi-Newton method. SQP methods are used on mathematical problems for which the objective function and the constraints are twice continuously differentiable. SQP methods solve a sequence of optimization subproblems, each of which optimizes a quadratic model of the objective subject to a linearization of the constraints. If the problem is unconstrained, then the method reduces to Newton's method for finding a point where the gradient of the objective vanishes. If the problem has only equality constraints, then the method is equivalent to applying Newton's method to the first-order optimality conditions, or Karush–Kuhn–Tucker conditions, of the problem. Algorithm basics Consider a nonlinear programming problem of the form: The Lagrangian for this problem is where and are Lagrange multipliers. The standard Newton's Method searches for the solution by iterating the following equation, where denotes the Hessian matrix: . However, because the matrix is generally singular (and therefore non-invertible), the Newton step cannot be calculated directly. Instead the basic sequential quadratic programming algorithm defines an appropriate search direction at an iterate , as a solution to the quadratic programming subproblem Note that the term in the expression above may be left out for the minimization problem, since it is constant under the operator. Together, the SQP algorithm starts by first choosing the initial iterate , then calculating and . Then the QP subproblem is built and solved to find the Newton step direction which is used to update the parent problem iterate using . This process is repeated for until the parent problem satisfies a convergence test. Alternative approaches Sequential linear programming Sequential linear-quadratic programming Augmented Lagrangian method Implementations SQP methods have been impl
https://en.wikipedia.org/wiki/Zoho%20Office%20Suite
Zoho Office Suite is an Indian web-based online office suite containing word processing, spreadsheets, presentations, databases, note-taking, wikis, web conferencing, customer relationship management (CRM), project management, invoicing and other applications. It is developed by Zoho Corporation. History Zoho Office Suite was launched in 2005 with a web-based word processor. Additional products such as spreadsheets and presentations, were incorporated later into Zoho. Zoho applications are distributed as software as a service (SaaS). Zoho uses an open application programming interface for its Writer, Sheet, Show, Creator, Meeting, and Planner products. It also has plugins into Microsoft Word and Excel, an OpenOffice.org plugin, and a plugin for Firefox. Zoho Sites is an online, drag and drop website builder. It provides web hosting, unlimited storage, bandwidth and web pages. Features also include an array of website templates and mobile websites. Zoho CRM is a customer relationship management application with features like procurement, inventory, and some accounting functions from the realm of ERP. The free version is limited to 10 users. In October 2009, Zoho integrated some of their applications with the Google Apps online suite. This enabled users to sign into both suites under one login. Zoho and Google still remain separate, competing companies. In 2020, Zoho Workplace won Rank 1 in the Indian government's 'Atmanirbhar Bharat App Innovation challenge' in Office category while Zoho Invoice, Books & Expense won Rank 1 in business category. See also Comparison of office suites List of office suites
https://en.wikipedia.org/wiki/EusLisp%20Robot%20Programming%20Language
EusLisp is a Lisp-based programming system. Built on the basis of object orientation, it is designed specifically for developing robotics software. The first version of it ran in 1986 on Unix-System5/Ustation-E20.
https://en.wikipedia.org/wiki/Peskin%E2%80%93Takeuchi%20parameter
In particle physics, the Peskin–Takeuchi parameters are a set of three measurable quantities, called S, T, and U, that parameterize potential new physics contributions to electroweak radiative corrections. They are named after physicists Michael Peskin and Tatsu Takeuchi, who proposed the parameterization in 1990; proposals from two other groups (see References below) came almost simultaneously. The Peskin–Takeuchi parameters are defined so that they are all equal to zero at a reference point in the Standard Model, with a particular value chosen for the (then unmeasured) Higgs boson mass. The parameters are then extracted from a global fit to the high-precision electroweak data from particle collider experiments (mostly the Z pole data from the CERN LEP collider) and atomic parity violation. The measured values of the Peskin–Takeuchi parameters agree with the Standard Model. They can then be used to constrain models of new physics beyond the Standard Model. The Peskin–Takeuchi parameters are only sensitive to new physics that contributes to the oblique corrections, i.e., the vacuum polarization corrections to four-fermion scattering processes. Definitions The Peskin–Takeuchi parameterization is based on the following assumptions about the nature of the new physics: The electroweak gauge group is given by SU(2)L x U(1)Y, and thus there are no additional electroweak gauge bosons beyond the photon, Z boson, and W boson. In particular, this framework assumes there are no Z' or W' gauge bosons. If there are such particles, the S, T, U parameters do not in general provide a complete parameterization of the new physics effects. New physics couplings to light fermions are suppressed, and hence only oblique corrections need to be considered. In particular, the framework assumes that the nonoblique corrections (i.e., vertex corrections and box corrections) can be neglected. If this is not the case, then the process by which the S, T, U parameters are extracted fr
https://en.wikipedia.org/wiki/Constraint%20%28mathematics%29
In mathematics, a constraint is a condition of an optimization problem that the solution must satisfy. There are several types of constraints—primarily equality constraints, inequality constraints, and integer constraints. The set of candidate solutions that satisfy all constraints is called the feasible set. Example The following is a simple optimization problem: subject to and where denotes the vector (x1, x2). In this example, the first line defines the function to be minimized (called the objective function, loss function, or cost function). The second and third lines define two constraints, the first of which is an inequality constraint and the second of which is an equality constraint. These two constraints are hard constraints, meaning that it is required that they be satisfied; they define the feasible set of candidate solutions. Without the constraints, the solution would be (0,0), where has the lowest value. But this solution does not satisfy the constraints. The solution of the constrained optimization problem stated above is , which is the point with the smallest value of that satisfies the two constraints. Terminology If an inequality constraint holds with equality at the optimal point, the constraint is said to be , as the point cannot be varied in the direction of the constraint even though doing so would improve the value of the objective function. If an inequality constraint holds as a strict inequality at the optimal point (that is, does not hold with equality), the constraint is said to be , as the point could be varied in the direction of the constraint, although it would not be optimal to do so. Under certain conditions, as for example in convex optimization, if a constraint is non-binding, the optimization problem would have the same solution even in the absence of that constraint. If a constraint is not satisfied at a given point, the point is said to be infeasible. Hard and soft constraints If the problem mandates that the con
https://en.wikipedia.org/wiki/Bucladesine
Bucladesine is a cyclic nucleotide derivative which mimics the action of endogenous cAMP and is a phosphodiesterase inhibitor. Bucladesine is a cell permeable cAMP analog. The compound is used in a wide variety of research applications because it mimics cAMP and can induce normal physiological responses when added to cells in experimental conditions. cAMP is only able to elicit minimal responses in these situations. The neurite outgrowth instigated by bucladesine in cell cultures has been shown to be enhanced by nardosinone. Bucladesine and seizure The effect of bucladesine as a cAMP analog has been studied on the pentylenetetrazol-induced seizure in the wild-type mice. The data showed that bucladesine (300nM/mouse) reduced the seizure latency and threshold. In addition they found that combination of bucladesine and pentoxyfillin has additive effect on seizure latency and threshold. Bucladesine and morphine withdrawal syndrome Bucladesine (50-100nM/mouse) showed significant attenuation in the morphine withdrawal syndrome in the wild-type mice. In addition, its high dose (200nM/mouse) combination with H-89, as a protein kinase inhibitor, had additive attenuating effect on withdrawal syndromes.
https://en.wikipedia.org/wiki/The%20Centipede%27s%20Dilemma
"The Centipede's Dilemma" is a short poem that has lent its name to a psychological effect called the centipede effect or centipede syndrome. The centipede effect occurs when a normally automatic or unconscious activity is disrupted by consciousness of it or reflection on it. For example, a golfer thinking too closely about their swing or someone thinking too much about how they knot their tie may find their performance of the task impaired. The effect is also known as hyperreflection or Humphrey's law after English psychologist George Humphrey (1889–1966), who propounded it in 1923. As he wrote of the poem, "This is a most psychological rhyme. It contains a profound truth which is illustrated daily in the lives of all of us". The effect is the reverse of a solvitur ambulando. Poem The short poem is usually attributed to Katherine Craster (1841–1874) in Pinafore Poems, 1871. By 1881, it had begun appearing in journals such as The Spectator and Littell's Living Age. On May 23, 1889, the poem appeared in an article by British zoologist Ray Lankester, published in the scientific journal Nature, which discussed the work of photographer Eadweard Muybridge in capturing the motion of animals: "For my own part," wrote Lankester, "I should greatly like to apply Mr. Muybridge's cameras, or a similar set of batteries, to the investigation of a phenomenon more puzzling even than that of 'the galloping horse'. I allude to the problem of 'the running centipede. Lankester finished the article on a fanciful note by imagining the "disastrous results in the way of perplexity" that could result from such an investigation, quoting the poem and mentioning that the author was unknown to him or to the friend who sent it to him. It has since been variously attributed to specific authors, but without convincing evidence, and often appears under the title "The Centipede's Dilemma". The version in the article is: A centipede was happy – quite! Until a toad in fun Said, "Pray, which leg m
https://en.wikipedia.org/wiki/Minimum%20information%20required%20in%20the%20annotation%20of%20models
MIRIAM (Minimum Information Required In The Annotation of Models) is a community-level effort to standardize the annotation and curation processes of quantitative models of biological systems. It consists of a set of guidelines suitable for use with any structured format, allowing different groups to collaborate and share resulting models. Adherence to these guidelines also facilitates the sharing of software and service infrastructures built upon modeling activities. The idea of "a set of good practices" including "some obligatory metadata" was first proposed by Nicolas Le Novère in October 2004 as part of a discussion to develop a common database of models in systems biology (which led to the creation of BioModels Database). These initial ideas were further refined at a meeting in Heidelberg, during ICSB 2004, with representatives from many other interested groups. MIRIAM is a registered project of the MIBBI (minimum information for biological and biomedical investigations). MIRIAM Guidelines The MIRIAM Guidelines are composed of three parts, reference correspondence, attribution annotation, and external resource annotation, each of which deals with a different aspect of information that should be included within a model. Reference correspondence 'Reference correspondence' deals with the basic reference information needed to make use of the model, detailing on a gross level the format of the model file, and its instantiability for simulation purposes. The model file must be encoded in a public, standardized, machine-readable format (SBML, CellML, GENESIS, ...). The model file must be valid with respect to its encoding schema. The model must be associated with a reference description or publication detailing its origin, even if it is a composite. The encoded model structure must reflect the (biological) process(es) detailed in the reference description. The model must be instantiable; necessary quantitative parameters, such as initial conditions, shoul
https://en.wikipedia.org/wiki/Umesh%20Vazirani
Umesh Virkumar Vazirani is an Indian–American academic who is the Roger A. Strauch Professor of Electrical Engineering and Computer Science at the University of California, Berkeley, and the director of the Berkeley Quantum Computation Center. His research interests lie primarily in quantum computing. He is also a co-author of a textbook on algorithms. Biography Vazirani received a BS from MIT in 1981 and received his Ph.D. in 1986 from UC Berkeley under the supervision of Manuel Blum. He is the brother of University of California, Irvine professor Vijay Vazirani. Research Vazirani is one of the founders of the field of quantum computing. His 1993 paper with his student Ethan Bernstein on quantum complexity theory defined a model of quantum Turing machines which was amenable to complexity based analysis. This paper also gave an algorithm for the quantum Fourier transform, which was then used by Peter Shor within a year in his celebrated quantum algorithm for factoring integers. With Charles Bennett, Ethan Bernstein, and Gilles Brassard, he showed that quantum computers cannot solve black-box search problems faster than in the number of elements to be searched. This result shows that the Grover search algorithm is optimal. It also shows that quantum computers cannot solve NP-complete problems in polynomial time using only the certifier. Awards and honors In 2005, both Vazirani and his brother Vijay Vazirani were inducted as Fellows of the Association for Computing Machinery, Umesh for "contributions to theoretical computer science and quantum computation" and his brother Vijay for his work on approximation algorithms. Vazirani was awarded the Fulkerson Prize for 2012 for his work on improving the approximation ratio for graph separators and related problems (jointly with Satish Rao and Sanjeev Arora). In 2018, he was elected to the National Academy of Sciences. Selected publications . A preliminary version of this paper was also published in STOC '87. . . .
https://en.wikipedia.org/wiki/Trigonitis
Trigonitis is a condition of inflammation of the trigone region of the bladder. It is more common in women. The cause of trigonitis is not known, and there is no solid treatment. Electrocautery is sometimes used, but is generally unreliable as a treatment, and typically does not have quick results. Several drugs, such as muscle relaxants, antibiotics, antiseptics such as Urised, have varied and unreliable results. Other forms of treatment include urethrotomy, cryosurgery, and neurostimulation.
https://en.wikipedia.org/wiki/Lehmer%20mean
In mathematics, the Lehmer mean of a tuple of positive real numbers, named after Derrick Henry Lehmer, is defined as: The weighted Lehmer mean with respect to a tuple of positive weights is defined as: The Lehmer mean is an alternative to power means for interpolating between minimum and maximum via arithmetic mean and harmonic mean. Properties The derivative of is non-negative thus this function is monotonic and the inequality holds. The derivative of the weighted Lehmer mean is: Special cases is the minimum of the elements of . is the harmonic mean. is the geometric mean of the two values and . is the arithmetic mean. is the contraharmonic mean. is the maximum of the elements of . Sketch of a proof: Without loss of generality let be the values which equal the maximum. Then Applications Signal processing Like a power mean, a Lehmer mean serves a non-linear moving average which is shifted towards small signal values for small and emphasizes big signal values for big . Given an efficient implementation of a moving arithmetic mean called you can implement a moving Lehmer mean according to the following Haskell code. lehmerSmooth :: Floating a => ([a] -> [a]) -> a -> [a] -> [a] lehmerSmooth smooth p xs = zipWith (/) (smooth (map (**p) xs)) (smooth (map (**(p-1)) xs)) For big it can serve an envelope detector on a rectified signal. For small it can serve an baseline detector on a mass spectrum. Gonzalez and Woods call this a "contraharmonic mean filter" described for varying values of p (however, as above, the contraharmonic mean can refer to the specific case ). Their convention is to substitute p with the order of the filter Q: Q=0 is the arithmetic mean. Positive Q can reduce pepper noise and negative Q can reduce salt noise. See also Mean Power mean Notes External links Lehmer Mean at MathWorld Means Articles with example Haskell code
https://en.wikipedia.org/wiki/Helix%20Board%2012%20FireWire
The Helix Board 12 FireWire is a mixer developed by Phonic Corporation that features a FireWire Interface, able to connect the mixer to Windows and Mac computers. The product was released in 2005 and has since become a signature product for the company Outputs The Helix Board 12 FireWire has a stereo main output, control room output, auxiliary send, headphones outputs, RCA record outputs and alternate 3-4 outputs. Digital Effects A built-in 16 program, 32-bit effect processor is included with the Helix Board 12 FireWire. Effects include Hall, Room, Plate, Cathedral, and more. FireWire Interface The FireWire interface allows all 8 inputs (counting stereo channels as 2) of the Helix Board 12 FireWire to be sent to a computer for recording, through its ASIO drivers (no drivers are required for the Mac). The main stereo output of your computer can also be sent through the FireWire interface, back to the Helix Board. The returned signal can be routed by the touch of one of the buttons on the face of the mixer. The Helix Board 12 FireWire is bundled with Steinberg Cubase LE software for recording purposes, though it's suggested that users upgrade to SX or use other digital audio workstation software. System requirements Windows Microsoft Windows XP SP1 or SP2 Available FireWire port Intel Pentium 4 processor or equivalent AMD processor Motherboard with Intel or VIA chipset 5400RPM or faster hard disc drive (7200 RPM or faster with 8MB cache recommended) 256MB or more of RAM (512MB recommended) Macintosh Mac OS X 10.3.5 or later with native FireWire support G4 or newer processor 256MB or more of RAM MKII 2006 saw the release of the Helix Board 12 FireWire MKII, which included additional features. Among these was a new DFX processor, pre/post switch on each input channel and a channel 9/10 assign switch.
https://en.wikipedia.org/wiki/Energy%20budget
An energy budget is a balance sheet of energy income against expenditure. It is studied in the field of Energetics which deals with the study of energy transfer and transformation from one form to another. Calorie is the basic unit of measurement. An organism in a laboratory experiment is an open thermodynamic system, exchanging energy with its surroundings in three ways - heat, work and the potential energy of biochemical compounds. Organisms use ingested food resources (C=consumption) as building blocks in the synthesis of tissues (P=production) and as fuel in the metabolic process that power this synthesis and other physiological processes (R=respiratory loss). Some of the resources are lost as waste products (F=faecal loss, U=urinary loss). All these aspects of metabolism can be represented in energy units. The basic model of energy budget may be shown as: P = C - R - U - F or P = C - (R + U + F) or C = P + R + U + F All the aspects of metabolism can be represented in energy units (e.g. joules (J);1 calorie = 4.2 kJ). Energy used for metabolism will be R = C - (F + U + P) Energy used in the maintenance will be R + F + U = C - P Endothermy and ectothermy Energy budget allocation varies for endotherms and ectotherms. Ectotherms rely on the environment as a heat source while endotherms maintain their body temperature through the regulation of metabolic processes. The heat produced in association with metabolic processes facilitates the active lifestyles of endotherms and their ability to travel far distances over a range of temperatures in the search for food. Ectotherms are limited by the ambient temperature of the environment around them but the lack of substantial metabolic heat production accounts for an energetically inexpensive metabolic rate. The energy demands for ectotherms are generally one tenth of that required for endotherms.
https://en.wikipedia.org/wiki/Control%20reconfiguration
Control reconfiguration is an active approach in control theory to achieve fault-tolerant control for dynamic systems. It is used when severe faults, such as actuator or sensor outages, cause a break-up of the control loop, which must be restructured to prevent failure at the system level. In addition to loop restructuring, the controller parameters must be adjusted to accommodate changed plant dynamics. Control reconfiguration is a building block toward increasing the dependability of systems under feedback control. Reconfiguration problem Fault modelling The figure to the right shows a plant controlled by a controller in a standard control loop. The nominal linear model of the plant is The plant subject to a fault (indicated by a red arrow in the figure) is modelled in general by where the subscript indicates that the system is faulty. This approach models multiplicative faults by modified system matrices. Specifically, actuator faults are represented by the new input matrix , sensor faults are represented by the output map , and internal plant faults are represented by the system matrix . The upper part of the figure shows a supervisory loop consisting of fault detection and isolation (FDI) and reconfiguration which changes the loop by choosing new input and output signals from {} to reach the control goal, changing the controller internals (including dynamic structure and parameters), adjusting the reference input . To this end, the vectors of inputs and outputs contain all available signals, not just those used by the controller in fault-free operation. Alternative scenarios can model faults as an additive external signal influencing the state derivatives and outputs as follows: Reconfiguration goals The goal of reconfiguration is to keep the reconfigured control-loop performance sufficient for preventing plant shutdown. The following goals are distinguished: Stabilization Equilibrium recovery Output trajectory recovery State trajectory
https://en.wikipedia.org/wiki/List%20of%20Unix%20systems
Each version of the UNIX Time-Sharing System evolved from the version before, with version one evolving from the prototypal Unix. Not all variants and descendants are displayed. Research Unix {| style="background-color: transparent; width: 100%" | align="left" valign="top" | "Ken's new system" (Unics → Unix) (1969) UNIX Time-Sharing System v1 (1971) UNIX Time-Sharing System v2 (1972) UNIX Time-Sharing System v3 (1973) UNIX Time-Sharing System v4 (1973) UNIX Time-Sharing System v5 (1974) UNSW 01 (1978) UNIX Time-Sharing System v6 (1975) 1BSD (1978) AUSAM (1978) IS/1 (1977) LSI-UNIX (1977) Mini-UNIX (1977) PWB/UNIX 1.0 (1977) USG 1.0 CB UNIX 1 Wollongong Unix (1977) | align="left" valign="top" | UNIX Time-Sharing System v7 (1979) 2BSD (1979) UNIX/32V (1979) 3BSD (1979) UNIX System III (1981) UNIX/V7M (1979) UNIX Time-Sharing System v8 (1985) UNIX Time-Sharing System v9 (1986) UNIX Time-Sharing System v10 (1989) IX Multilevel-Secure UNIX System (1992) |} The versions leading to v7 are also sometimes called Ancient unix. After the release of Version 10, the Unix research team at Bell Labs turned its focus to Plan 9 from Bell Labs, a distinct operating system that was first released to the public in 1993. All versions of BSD from its inception up to 4.3BSD-Reno are based on Research Unix, with versions starting with 4.4 BSD and Net/2 instead becoming Unix-like. Commercial AT&T UNIX Systems and descendants Each of the systems in this list is evolved from the version before, with Unix System III evolving from both the UNIX Time-Sharing System v7 and the descendants of the UNIX Time-Sharing System v6. {| style="background-color: transparent; width: 100%" | align="left" valign="top" | UNIX System III (1981) UNIX System IV (1982) UNIX System V (1983) UNIX System V Release 2 (1984) UNIX System V Release 3.0 (1986) UNIX System V Release 3.2 (1987) UNIX System V Release 4 (1988) UNIX System V Release 4.2 (1992) UnixWare 1.1 (1993) UnixWare 1.1.1 (1994) UnixWare 2.0 (
https://en.wikipedia.org/wiki/Stolarsky%20mean
In mathematics, the Stolarsky mean is a generalization of the logarithmic mean. It was introduced by Kenneth B. Stolarsky in 1975. Definition For two positive real numbers x, y the Stolarsky Mean is defined as: Derivation It is derived from the mean value theorem, which states that a secant line, cutting the graph of a differentiable function at and , has the same slope as a line tangent to the graph at some point in the interval . The Stolarsky mean is obtained by when choosing . Special cases is the minimum. is the geometric mean. is the logarithmic mean. It can be obtained from the mean value theorem by choosing . is the power mean with exponent . is the identric mean. It can be obtained from the mean value theorem by choosing . is the arithmetic mean. is a connection to the quadratic mean and the geometric mean. is the maximum. Generalizations One can generalize the mean to n + 1 variables by considering the mean value theorem for divided differences for the nth derivative. One obtains for . See also Mean
https://en.wikipedia.org/wiki/Identric%20mean
The identric mean of two positive real numbers x, y is defined as: It can be derived from the mean value theorem by considering the secant of the graph of the function . It can be generalized to more variables according by the mean value theorem for divided differences. The identric mean is a special case of the Stolarsky mean. See also Mean Logarithmic mean
https://en.wikipedia.org/wiki/IBM%20App%20Connect%20Enterprise
IBM App Connect Enterprise (abbreviated as IBM ACE, formerly known as IBM Integration Bus or WebSphere Message Broker) is IBM's premier integration software offering, allowing business information to flow between disparate applications across multiple hardware and software platforms. Rules can be applied to the data flowing through user-authored integrations to route and transform the information. The product can be used as an Enterprise Service Bus supplying a communication channel between applications and services in a service-oriented architecture. IBM ACE provides capabilities to build integration flows needed to support diverse integration requirements through a set of connectors to a range of data sources, including packaged applications, files, mobile devices, messaging systems, and databases. A benefit of using IBM ACE is that the tool enables existing applications for Web Services without costly legacy application rewrites. IBM ACE avoids the point-to-point strain on development resources by connecting any application or service over multiple protocols, including SOAP, HTTP and JMS. Modern secure authentication mechanisms, including the ability to perform actions on behalf of masquerading or delegate users, through MQ, HTTP, and SOAP nodes are supported, such as LDAP, X-AUTH, O-AUTH, and two-way SSL. A major focus of IBM ACE in its recent releases has been the capability of the product's runtime to be fully hosted in a cloud. Hosting the runtime in the cloud provides certain advantages and potential cost savings compared to hosting the runtime on premises as it simplifies the maintenance and application of OS-level patches ,which can sometimes be disruptive to business continuity. Also, cloud hosting of IBM ACE runtime allows easy expansion of capacity by adding more horsepower to the CPU configuration of a cloud environment or by adding additional nodes in an Active-Active configuration. An additional advantage of maintaining IBM ACE runtime in the cloud
https://en.wikipedia.org/wiki/Message%20broker
A message broker (also known as an integration broker or interface engine) is an intermediary computer program module that translates a message from the formal messaging protocol of the sender to the formal messaging protocol of the receiver. Message brokers are elements in telecommunication or computer networks where software applications communicate by exchanging formally-defined messages. Message brokers are a building block of message-oriented middleware (MOM) but are typically not a replacement for traditional middleware like MOM and remote procedure call (RPC). Overview A message broker is an architectural pattern for message validation, transformation, and routing. It mediates communication among applications, minimizing the mutual awareness that applications should have of each other in order to be able to exchange messages, effectively implementing decoupling. Purpose The primary purpose of a broker is to take incoming messages from applications and perform some action on them. Message brokers can decouple end-points, meet specific non-functional requirements, and facilitate reuse of intermediary functions. For example, a message broker may be used to manage a workload queue or message queue for multiple receivers, providing reliable storage, guaranteed message delivery and perhaps transaction management. Life cycle The following represent other examples of actions that might be handled by the broker: Route messages to one or more destinations Transform messages to an alternative representation Perform message aggregation, decomposing messages into multiple messages and sending them to their destination, then recomposing the responses into one message to return to the user Interact with an external repository to augment a message or store it Invoke web services to retrieve data Respond to events or errors Provide content and topic-based message routing using the publish–subscribe pattern Message brokers are generally based on one of two fundamental a
https://en.wikipedia.org/wiki/Marantz%20PMD-660
Manufactured by Marantz, the Marantz PMD-660 is a portable, solid-state, compact flash audio field recorder. It has 2 XLR (balanced) inputs, 2 line-in inputs, 2 internal microphones and can record in raw WAV or MP3 formats. It is powered with four (non-rechargeable) AA-sized batteries which offers 3.5 to 4 hours of uninterrupted recording. Uses As a field recorder, the PMD-660 is designed to be used outside of a controlled studio environment. Uses are electronic news gathering (ENG), podcasting, live music recording. External links Oade PMD660 Mods Customer Reviews at amazon Transom.org PMD660 review Sound recording technology
https://en.wikipedia.org/wiki/Iometer
Iometer is an I/O subsystem measurement and characterization tool for single and clustered systems. It is used as a benchmark and troubleshooting tool and is easily configured to replicate the behaviour of many popular applications. One commonly quoted measurement provided by the tool is IOPS. History Created by Intel Corporation (Sean Hefty, David Levine and Fab Tillier are listed by the Iometer About dialog as the developers), the tool was officially announced at the Intel Developer Forum (IDF) on 17 February 1998. In 2001 Intel discontinued development and subsequently handed the sources to the Open Source Development Lab for release under the Intel Open Source License. On 15 November 2001 the Iometer project was registered at SourceForge.net and an initial version was made available. Experiencing no further development, the project was relaunched by Daniel Scheibli in February 2003. Since then it has been driven by an international group of individuals who have been improving and porting the product to additional platforms. Functionality Iometer is based on a client–server model, where one instance of the Iometer graphical user interface is managing one or more 'managers' (each one representing a separate Dynamo.exe process) which are doing the I/O with one or more worker threads. Iometer performs Asynchronous I/O - accessing files or block devices (later one allowing to bypass the file system buffers). Iometer allows the configuration of disk parameters such as the 'Maximum Disk Size', 'Starting Disk Sector' and '# of Outstanding I/Os'. This allows a user to configure a test file upon which the 'Access Specifications' configure the I/O types to the file. Configurable items within the Access Specifications are: Transfer Request Size Percent Random/Sequential distribution. Percent Read/Write Distribution Aligned I/O's. Reply Size TCP/IP status Burstiness. In conjunction with the Access Specifications, Iometer allows the specifications to be cycled with