source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Pelorus%20%28instrument%29
In marine navigation, a pelorus is a reference tool for maintaining bearing of a vessel at sea. It is a "simplified compass" without a directive element, suitably mounted and provided with vanes to permit observation of relative bearings. The instrument was named for one Pelorus, said to have been the pilot for Hannibal, circa 203 BC. Ancient use Harold Gatty described the use of a pelorus by Polynesians before the use of a compass. In equatorial waters the nightly course of stars overhead is nearly uniform during the year. This regularity simplified navigation for the Polynesians using a pelorus, or dummy compass: Reading from North to South, in their rising and setting positions, these stars are: {| class="wikitable" |+ Stars |- valign="top" ! Point ! Star |- valign="top" | N || Polaris |- valign="top" | NbE || "the Guards" (Ursa Minor) |- valign="top" | NNE || Alpha Ursa Major |- valign="top" | NEbN ||Alpha Cassiopeiae |- valign="top" | NE || Capella |- valign="top" | NEbE || Vega |- valign="top" | ENE || Arcturus |- valign="top" | EbN || the Pleiades |- valign="top" | E || Altair |- valign="top" | EbS || Orion's belt |- valign="top" | ESE || Sirius |- valign="top" | SEbE || Beta Scorpionis |- valign="top" | SE || Antares |- valign="top" | SEbS || Alpha Centauri |- valign="top" | SSE || Canopus |- valign="top" | SbE || Achernar |- valign="top" | S || Southern Cross |} The true position of these stars is only approximate to their theoretical equidistant rhumbs on the sidereal compass. Over time, the elaboration of the pelorus points led to the modern compass rose. Modern use In appearance and use, a pelorus resembles a compass or compass repeater, with sighting vanes or a sighting telescope attached, but it has no directive properties. That is, it remains at any relative direction to which it is set. It is generally used by setting 000° at the lubber's line. Relative bearings are then observed. They can be converted to bearings true, magnetic, gr
https://en.wikipedia.org/wiki/Azimuth
An azimuth (; from ) is the angular measurement in a spherical coordinate system which represents the horizontal angle from a cardinal direction, most commonly north. Mathematically, the relative position vector from an observer (origin) to a point of interest is projected perpendicularly onto a reference plane (the horizontal plane); the angle between the projected vector and a reference vector on the reference plane is called the azimuth. When used as a celestial coordinate, the azimuth is the horizontal direction of a star or other astronomical object in the sky. The star is the point of interest, the reference plane is the local area (e.g. a circular area with a 5 km radius at sea level) around an observer on Earth's surface, and the reference vector points to true north. The azimuth is the angle between the north vector and the star's vector on the horizontal plane. Azimuth is usually measured in degrees (°), in the positive range 0° to 360° or in the signed range -180° to +180°. The concept is used in navigation, astronomy, engineering, mapping, mining, and ballistics. Etymology The word azimuth is used in all European languages today. It originates from medieval Arabic السموت (al-sumūt, pronounced as-sumūt), meaning "the directions" (plural of Arabic السمت al-samt = "the direction"). The Arabic word entered late medieval Latin in an astronomy context and in particular in the use of the Arabic version of the astrolabe astronomy instrument. Its first recorded use in English is in the 1390s in Geoffrey Chaucer's Treatise on the Astrolabe. The first known record in any Western language is in Spanish in the 1270s in an astronomy book that was largely derived from Arabic sources, the Libros del saber de astronomía commissioned by King Alfonso X of Castile. In astronomy In the horizontal coordinate system, used in celestial navigation, azimuth is one of the two coordinates. The other is altitude, sometimes called elevation above the horizon. It is also used
https://en.wikipedia.org/wiki/Electroglottographic%20wavegram
An electroglottographic wavegram (short: EGG wavegram) is a tool for analyzing the voice source in speech and singing, based on electroglottographic (EGG) signals (and their first derivative, DEGG). Assessing the singing and speaking voice The wavegram, invented by Christian T. Herbst, provides an intuitive means for quickly assessing vocal fold contact phenomena and their variation over time. Vocal fold closings and openings appear here as a sequence of events rather than single incidents, taking place over a certain period of time, and changing with pitch, loudness and register. Wavegrams document systematic phenomena, indicating subtle changes of the vocal fold oscillatory regime. Electroglottographic wavegrams are created in 5 steps (see illustration): extraction of consecutive glottal cycles from the EGG signal; locally normalized data values are converted into monochrome color information, and are plotted as a strip representing one glottal cycle each; strips are rotated 90 degrees counter-clockwise; Glottal cycle duration is normalized by scaling the individual glottal cycle plots to the same height; the resulting graphs are combined to form the final display, the EGG wavegram. Wavegram data can be influenced by the anatomical baseline of the individual; physiological and habitual muscular patterns in phonation, e.g. degree of vocal fold adduction, register in singing and speech; organic voice disorders, i.e. pathological deviations from the norm. Wavegrams show a potential to be used in: screenings for voice disorders (e.g. in schools); (real-time) bio-feedback in voice pedagogy and education; voice therapy (functional voice disorders); conservative and post-operative treatment of organic voice disorders; voice recognition in forensics. To construct a wavegram, the time-varying fundamental frequency is calculated and consecutive individual glottal cycles are identified in the EGG or DEGG signal. Each cycle is locally normalized in durat
https://en.wikipedia.org/wiki/Epigenetics%20of%20schizophrenia
The epigenetics of schizophrenia is the study of how inherited epigenetic changes are regulated and modified by the environment and external factors and how these changes influence the onset and development of, and vulnerability to, schizophrenia. Epigenetics concerns the heritability of those changes, too. Schizophrenia is a debilitating and often misunderstood disorder that affects up to 1% of the world's population. Although schizophrenia is a heavily studied disorder, it has remained largely impervious to scientific understanding; epigenetics offers a new avenue for research, understanding, and treatment. Background History Historically, schizophrenia has been studied and examined through different paradigms, or schools of thought. In the late 1870s, Emil Kraepelin started the idea of studying it as an illness. Another paradigm, introduced by Zubin and Spring in 1977, was the stress-vulnerability model where the individual has unique characteristics that give him or her strengths or vulnerabilities to deal with stress, a predisposition for schizophrenia. More recently, with the decoding of the human genome, there had been a focus on identifying specific genes to study the disease. However, the genetics paradigm faced problems with inconsistent, inconclusive, and variable results. The most recent school of thought is studying schizophrenia through epigenetics. The idea of epigenetics has been described as far back as 1942, when Conrad Waddington described it as how the environment regulated genetics. As the field and available technology has progressed, the term has come to also refer to the molecular mechanisms of regulation. The concept that these epigenetic changes can be passed on to future generations has progressively become more accepted. While epigenetics is a relatively new field of study, specific applications and focus on mental disorders like schizophrenia is an even more recent area of research. Schizophrenia Symptoms The core symptoms of s
https://en.wikipedia.org/wiki/Born%E2%80%93Land%C3%A9%20equation
The Born–Landé equation is a means of calculating the lattice energy of a crystalline ionic compound. In 1918 Max Born and Alfred Landé proposed that the lattice energy could be derived from the electrostatic potential of the ionic lattice and a repulsive potential energy term. where: NA = Avogadro constant; M = Madelung constant, relating to the geometry of the crystal; z+ = numeric charge number of cation z− = numeric charge number of anion e = elementary charge, 1.6022 C ε0 = permittivity of free space 4πε0 = 1.112 C2/(J·m) r0 = distance between closest cation [ +ve ] & anion [ -ve ]. n = Born exponent, typically a number between 5 and 12, determined experimentally by measuring the compressibility of the solid, or derived theoretically. E = Lattice energy is expressed by 'E' . Derivation The ionic lattice is modeled as an assembly of hard elastic spheres which are compressed together by the mutual attraction of the electrostatic charges on the ions. They achieve the observed equilibrium distance apart due to a balancing short range repulsion. Electrostatic potential The electrostatic potential energy, Epair, between a pair of ions of equal and opposite charge is: where z = magnitude of charge on one ion e = elementary charge, 1.6022 C ε0 = permittivity of free space 4ε0 = 1.112 C2/(J·m) r = distance separating the ion centers For a simple lattice consisting ions with equal and opposite charge in a 1:1 ratio, interactions between one ion and all other lattice ions need to be summed to calculate EM, sometimes called the Madelung or lattice energy: where M = Madelung constant, which is related to the geometry of the crystal r = closest distance between two ions of opposite charge Repulsive term Born and Lande suggested that a repulsive interaction between the lattice ions would be proportional to so that the repulsive energy term, ER, would be expressed: where B = constant scaling the strength of the repulsive interaction r = closest distance between two i
https://en.wikipedia.org/wiki/Subtelomere
Subtelomeres are segments of DNA between telomeric caps and chromatin. Structure Telomeres are specialized protein–DNA constructs present at the ends of eukaryotic chromosomes, which prevent them from degradation and end-to-end chromosomal fusion. Most vertebrate telomeric DNA consists of long (TTAGGG)n repeats of variable length, often around 3-20kb. Subtelomeres are segments of DNA between telomeric caps and chromatin. In vertebrates, each chromosome has two subtelomeres immediately adjacent to the long (TTAGGG)n repeats. Subtelomeres are considered to be the most distal (farthest from the centromere) region of unique DNA on a chromosome, and they are unusually dynamic and variable mosaics of multichromosomal blocks of sequence. The subtelomeres of such diverse species as humans, Plasmodium falciparum, Drosophila melanogaster, and Saccharomyces cerevisiae are structurally similar in that they are composed of various repeated elements, but the extent of the subtelomeres and the sequence of the elements vary greatly among organisms. In yeast (S. cerevisiae), subtelomeres are composed of two domains: the proximal and distal (telomeric) domains. The two domains differ in sequence content and extent of homology to other chromosome ends, and they are often separated by a stretch of degenerate telomere repeats (TTAGGG) and an element called 'core X', which is found at all chromosome ends and contains an autonomously replicating sequence (ARS) and an ABF1 binding site. The proximal domain is composed of variable interchromosomal duplications (<1-30 kb); this region can contain genes such Pho, Mel, and Mal. The distal domain is composed of 0-4 tandem copies of the highly conserved Y' element; the number and chromosomal distribution of Y′ elements varies among yeast strains. Between the core X and the Y' element or the core X and TTAGGG sequence there is often a set of 4 subtelomeric repeats elements (STR): STR-A, STR-B, STR-C and STR-D which consists of multiple copies o
https://en.wikipedia.org/wiki/Isopropyl%20acetate
Isopropyl acetate is an ester, an organic compound which is the product of esterification of acetic acid and isopropanol. It is a clear, colorless liquid with a characteristic fruity odor. Isopropyl acetate is a solvent with a wide variety of manufacturing uses that is miscible with most other organic solvents, and slightly soluble in water (although less so than ethyl acetate). It is used as a solvent for cellulose, plastics, oil and fats. It is a component of some printing inks and perfumes. Isopropyl acetate decomposes slowly on contact with steel in the presence of air, producing acetic acid and isopropanol. It reacts violently with oxidizing materials and it attacks many plastics. Isopropyl acetate is quite flammable in both its liquid and vapor forms, and it may be harmful if swallowed or inhaled. The Occupational Safety and Health Administration has set a permissible exposure limit (PEL) of 250ppm (950mg/m3) over an eight-hour time-weighted average for workers handling isopropyl acetate.
https://en.wikipedia.org/wiki/Purple%20moor%20grass%20and%20rush%20pastures
Purple moor grass and rush pastures is a type of Biodiversity Action Plan habitat in the UK. It occurs on poorly drained neutral and acidic soils of the lowlands and upland fringe. It is found in the South West of England, especially in Devon. The vegetation consists of species-rich, semi-natural grassland containing abundant purple moor grass (Molinia caerulea) and one or more of several creeping rushes: sharp-flowered rush (Juncus acutiflorus), jointed rush (Juncus articulatus) and blunt-flowered rush (Juncus subnodulosus). Only 8% remains of the area thought to have existed in 1900. In the UK estimate the area is thought to be less than . Their importance is recognised and are included as a priority habitat in the United Kingdom Biodiversity Action Plan. Etymology In Devon and Cornwall it is known as culm grassland, after the Culm Measures on which it is predominantly found. In East Anglia it is known as litter meadow due to the practice of cutting it for bedding. In Wales it is known as rhôs pasture. Typical grasses Common bent (Agrostis capillaris), crested dog's-tail (Cynosurus cristatus), floating sweet grass (Glyceria fluitans), marsh foxtail (Alopecurus geniculatus), purple moor grass (Molinia caerulea), red fescue (Festuca rubra), sweet vernal grass (Anthoxanthum odoratum). Characteristic species The Natural England Higher Level Stewardship Farm Environmental Plan handbook defines the habitat as grassland with at least two of the following species are found frequently, with another two being found occasionally. Bog asphodel (Narthecium ossifragum) Bog mosses (Sphagnum spp) Bog pimpernel (Anagallis tenella) Bugle (Ajuga reptans) Common valerian (Valeriana officinalis) Common meadow-rue (Thalictrum flavum) Cross-leaved heath (Erica tetralix) Devil's-bit scabious (Succisa pratensis) Globeflower (Trollius europaeus) Greater burnet (Sanguisorba officinalis) Greater bird's foot trefoil (Lotus pedunculatus) Hemp-agrimony (Eupatorium cannabinum) Join
https://en.wikipedia.org/wiki/Bramble%E2%80%93Hilbert%20lemma
In mathematics, particularly numerical analysis, the Bramble–Hilbert lemma, named after James H. Bramble and Stephen Hilbert, bounds the error of an approximation of a function by a polynomial of order at most in terms of derivatives of of order . Both the error of the approximation and the derivatives of are measured by norms on a bounded domain in . This is similar to classical numerical analysis, where, for example, the error of linear interpolation can be bounded using the second derivative of . However, the Bramble–Hilbert lemma applies in any number of dimensions, not just one dimension, and the approximation error and the derivatives of are measured by more general norms involving averages, not just the maximum norm. Additional assumptions on the domain are needed for the Bramble–Hilbert lemma to hold. Essentially, the boundary of the domain must be "reasonable". For example, domains that have a spike or a slit with zero angle at the tip are excluded. Lipschitz domains are reasonable enough, which includes convex domains and domains with continuously differentiable boundary. The main use of the Bramble–Hilbert lemma is to prove bounds on the error of interpolation of function by an operator that preserves polynomials of order up to , in terms of the derivatives of of order . This is an essential step in error estimates for the finite element method. The Bramble–Hilbert lemma is applied there on the domain consisting of one element (or, in some superconvergence results, a small number of elements). The one-dimensional case Before stating the lemma in full generality, it is useful to look at some simple special cases. In one dimension and for a function that has derivatives on interval , the lemma reduces to where is the space of all polynomials of degree at most and indicates the th derivative of a function . In the case when , , , and is twice differentiable, this means that there exists a polynomial of degree one such that for all , T
https://en.wikipedia.org/wiki/IPC%20%28electronics%29
IPC is a trade association whose aim is to standardize the assembly and production requirements of electronic equipment and assemblies. It was founded in 1957 as the Institute of Printed Circuits. Its name was later changed to the Institute for Interconnecting and Packaging Electronic Circuits to highlight the expansion from bare boards to packaging and electronic assemblies. In 1999, the organization formally changed its name to IPC with the accompanying tagline, Association Connecting Electronics Industries. IPC is accredited by the American National Standards Institute (ANSI) as a standards developing organization and is known globally for its standards. It publishes the most widely used acceptability standards in the electronics industry. IPC is headquartered in Bannockburn, Illinois, United States with additional offices in Washington, D.C. and Atlanta, Ga. in the United States, and overseas offices in China, Thailand, Vietnam, India and Belgium. Standards IPC standards are used by the electronics manufacturing industry. IPC-A-610, Acceptability of Electronic Assemblies, is used worldwide by original equipment manufacturers and EMS companies. There are more than 3600 trainers worldwide who are certified to train and test on the standard. Standards are created by committees of industry volunteers. Task groups have been formed in China, the United States, and Denmark. Standards published by IPC include: General documents IPC-T-50 Terms and Definitions IPC-2615 Printed Board Dimensions and Tolerances IPC-D-325 Documentation Requirements for Printed Boards IPC-A-31 Flexible Raw Material Test Pattern IPC-ET-652 Guidelines and Requirements for Electrical Testing of Unpopulated Printed Boards Design specifications IPC-2612 Sectional Requirements for Electronic Diagramming Documentation (Schematic and Logic Descriptions) IPC-2141A Design Guide for High-Speed Controlled Impedance Circuit Boards IPC-2221 Generic Standard on Printed Board Design IPC-2223 S
https://en.wikipedia.org/wiki/Chomsky%E2%80%93Sch%C3%BCtzenberger%20representation%20theorem
In formal language theory, the Chomsky–Schützenberger representation theorem is a theorem derived by Noam Chomsky and Marcel-Paul Schützenberger about representing a given context-free language in terms of two simpler languages. These two simpler languages, namely a regular language and a Dyck language, are combined by means of an intersection and a homomorphism. A few notions from formal language theory are in order. A context-free language is regular, if it can be described by a regular expression, or, equivalently, if it is accepted by a finite automaton. A homomorphism is based on a function which maps symbols from an alphabet to words over another alphabet ; If the domain of this function is extended to words over in the natural way, by letting for all words and , this yields a homomorphism . A matched alphabet is an alphabet with two equal-sized sets; it is convenient to think of it as a set of parentheses types, where contains the opening parenthesis symbols, whereas the symbols in contains the closing parenthesis symbols. For a matched alphabet , the Dyck language is given by Chomsky–Schützenberger theorem. A language L over the alphabet is context-free if and only if there exists a matched alphabet a regular language over , and a homomorphism such that . Proofs of this theorem are found in several textbooks, e.g. or .
https://en.wikipedia.org/wiki/Hurdle%20technology
Hurdle technology is a method of ensuring that pathogens in food products can be eliminated or controlled. This means the food products will be safe for consumption, and their shelf life will be extended. Hurdle technology usually works by combining more than one approach. These approaches can be thought of as "hurdles" the pathogen has to overcome if it is to remain active in the food. The right combination of hurdles can ensure all pathogens are eliminated or rendered harmless in the final product. Hurdle technology has been defined by Leistner (2000) as an intelligent combination of hurdles which secures the microbial safety and stability as well as the organoleptic and nutritional quality and the economic viability of food products. The organoleptic quality of the food refers to its sensory properties, that is its look, taste, smell and texture. Examples of hurdles in a food system are high temperature during processing, low temperature during storage, increasing the acidity, lowering the water activity or redox potential, or the presence of preservatives. According to the type of pathogens and how risky they are, the intensity of the hurdles can be adjusted individually to meet consumer preferences in an economical way, without compromising the safety of the product. Hurdles Each hurdle aims to eliminate, inactivate or at least inhibit unwanted microorganisms. Common salt or organic acids can be used as hurdles to control microbials in food. Many natural antimicrobials such as nisin, natamycin and other bacteriocins, and essential oils derived from rosemary or thyme, also work well. "Traditionally, fermented seafood products common in Japan, provide a typical example of hurdle technology. Fermentation of sushi employs hurdles that favour growth of desirable bacteria but inhibit the growth of pathogens. The important hurdles in the early stages of fermentation are salt and vinegar. Raw fish is cured in salt (20–30%, w/w) for one month before being desalted a
https://en.wikipedia.org/wiki/Block%20matrix
In mathematics, a block matrix or a partitioned matrix is a matrix that is interpreted as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original matrix with a collection of horizontal and vertical lines, which break it up, or partition it, into a collection of smaller matrices. Any matrix may be interpreted as a block matrix in one or more ways, with each interpretation defined by how its rows and columns are partitioned. This notion can be made more precise for an by matrix by partitioning into a collection , and then partitioning into a collection . The original matrix is then considered as the "total" of these groups, in the sense that the entry of the original matrix corresponds in a 1-to-1 way with some offset entry of some , where and . Block matrix algebra arises in general from biproducts in categories of matrices. Example The matrix can be partitioned into four 2×2 blocks The partitioned matrix can then be written as Block matrix multiplication It is possible to use a block partitioned matrix product that involves only algebra on submatrices of the factors. The partitioning of the factors is not arbitrary, however, and requires "conformable partitions" between two matrices and such that all submatrix products that will be used are defined. Given an matrix with row partitions and column partitions and a matrix with row partitions and column partitions that are compatible with the partitions of , the matrix product can be performed blockwise, yielding as an matrix with row partitions and column partitions. The matrices in the resulting matrix are calculated by multiplying: Or, using the Einstein notation that implicitly sums over repeated indices: Block matrix inversion If a matrix is partitioned into four blocks, it can be inverted blockwise as follows: where A and D are square blocks of arbitrary size, and B and C are conformabl
https://en.wikipedia.org/wiki/Jean-Daniel%20Boissonnat
Jean-Daniel Boissonnat (born 18 May 1953) is a French computer scientist, who works as a director of research at the French Institute for Research in Computer Science and Automation (INRIA). He is an invited professor of computational geometry at the Collège de France, holding the Chair in Informatics and Computational Sciences for 2016–2017. Boissonat was one of the founders of the CGAL project for implementing geometric algorithms. With Mariette Yvinec, he is the author of the book Algorithmic Geometry (Cambridge University Press, 1998, translated from a 1995 edition in French). With Yvinec and Frédéric Chazal, he is the coauthor of Geometric and Topological Inference (Cambridge University Press, 2018). Awards and honours 1987: IBM award in Computer Science 2006: EADS award in Information Sciences 2006: Knight of National Order of Merit 2013: ANR Digital Technology Award
https://en.wikipedia.org/wiki/ReplayTV
ReplayTV was a former DVR company that from 1999 until 2005, produced a brand of digital video recorders (DVR), a term synonymous with personal video recorder (PVR). It is a consumer video device which allows users to capture television programming to internal hard disk storage for later viewing (and time shifting). ReplayTV was founded in September 1997 by future Roku founder Anthony Wood, who was president and CEO of ReplayTV until August 2001. The first ReplayTV model was introduced in January 1999 during the Consumer Electronics Show in Las Vegas, at the same time as a competing DVR model from rival company TiVo. After the sale of assets to DirecTV, ReplayTV's only ongoing activity was maintenance of the electronic program guide service by D&M Holdings, which was to be discontinued on July 31, 2011. However, on July 29, 2011, a notice was placed on the ReplayTV website stating that service would be continued without interruption for lifetime subscribers and monthly subscribers may have a short interruption in service. On September 2, 2011, programming contact through the ReplayTV dialup system was terminated without any update message being sent to subscribers or posted on replaytv.com. DNNA filed for bankruptcy on July 20, 2015. EPG data from their servers ran out on July 15, 2015. Even with the end of support from DNNA, third-party solutions are available to provide Electronic Program Guide data to ReplayTV units. History ReplayTV was founded in September 1997 by businessman Anthony Wood, who later founded Roku in October 2002. Initial sales to consumers were launched in April 1999, while volume production and sales did not begin until later in 2000. ReplayTV was purchased by SONICblue in 2001. On March 23, 2003, SONICblue filed for Chapter 11 bankruptcy, and on April 16 sold most of its assets, including ReplayTV, to the Japanese electronics giant D&M Holdings. SONICblue was fighting a copyright infringement suit over the ReplayTV's ability to skip commer
https://en.wikipedia.org/wiki/Diagnosis-related%20group
Diagnosis-related group (DRG) is a system to classify hospital cases into one of originally 467 groups, with the last group (coded as 470 through v24, 999 thereafter) being "Ungroupable". This system of classification was developed as a collaborative project by Robert B Fetter, PhD, of the Yale School of Management, and John D. Thompson, MPH, of the Yale School of Public Health. The system is also referred to as "the DRGs", and its intent was to identify the "products" that a hospital provides. One example of a "product" is an appendectomy. The system was developed in anticipation of convincing Congress to use it for reimbursement, to replace "cost based" reimbursement that had been used up to that point. DRGs are assigned by a "grouper" program based on ICD (International Classification of Diseases) diagnoses, procedures, age, sex, discharge status, and the presence of complications or comorbidities. DRGs have been used in the US since 1982 to determine how much Medicare pays the hospital for each "product", since patients within each category are clinically similar and are expected to use the same level of hospital resources. DRGs may be further grouped into Major Diagnostic Categories (MDCs). DRGs are also standard practice for establishing reimbursements for other Medicare related reimbursements such as to home healthcare providers. Purpose The original objective of diagnosis-related groups (DRG) was to develop a classification system that identified the "products" that the patient received. Since the introduction of DRGs in the early 1980s, the healthcare industry has evolved and developed an increased demand for a patient classification system that can serve its original objective at a higher level of sophistication and precision. To meet those evolving needs, the objective of the DRG system had to expand in scope. Several different DRG systems have been developed in the United States. They include: Medicare DRG (CMS-DRG & MS-DRG) Refined DRGs (R-DR
https://en.wikipedia.org/wiki/Constructive%20neutral%20evolution
Constructive neutral evolution (CNE) is a theory that seeks to explain how complex systems can evolve through neutral transitions and spread through a population by chance fixation (genetic drift). Constructive neutral evolution is a competitor for both adaptationist explanations for the emergence of complex traits and hypotheses positing that a complex trait emerged as a response to a deleterious development in an organism. Constructive neutral evolution often leads to irreversible or "irremediable" complexity and produces systems which, instead of being finely adapted for performing a task, represent an excess complexity that has been described with terms such as "runaway bureaucracy" or even a "Rube Goldberg machine". The groundworks for the concept of CNE were laid by two papers in the 1990s, although first explicitly proposed by Arlin Stoltzfus in 1999. The first proposals for the role CNE was in the evolutionary origins of complex macromolecular machines such as the spliceosome, RNA editing machinery, supernumerary ribosomal proteins, chaperones, and more. Since then and as an emerging trend of studies in molecular evolution, CNE has been applied to broader features of biology and evolutionary history including some models of eukaryogenesis, the emergence of complex interdependence in microbial communities, and de novo formation of functional elements from non-functional transcripts of junk DNA. Several approaches propose a combination of neutral and adaptive contributions in the evolutionary origins of various traits. Many evolutionary biologists posit that CNE must be the null hypothesis when explaining the emergence of complex systems to avoid assuming that a trait arose for an adaptive benefit. A trait may have arisen neutrally, even if later co-opted for another function. This approach stresses the need for rigorous demonstrations of adaptive explanations when describing the emergence of traits. This avoids the "adaptationist fallacy" which assumes that
https://en.wikipedia.org/wiki/Periodic%20breathing
Periodic breathing is clusters of breaths separated by intervals of apnea or near-apnea. As opposed to normal breathing which is usually regular, periodic breathing is defined as three or more episodes of central apnea lasting at least 4 seconds, separated by no more than 30 seconds of normal breathing. Periodic breathing was originally thought to arise from serious neurologic or cardiovascular disease and therefore to carry a poor outlook. It is now known that periodic breathing also tends to occur during sleep, it can occur in healthy persons, and the apnea in periodic breathing is usually central sleep apnea rather than obstructive sleep apnea. Periodic breathing during sleep occurs typically in adult patients with congestive heart failure. Periodic breathing is also a normal variation of breathing found in premature and full term infants. It occurs when the infant has pauses in breathing for no more than 10 seconds at a time followed by a series of rapid, shallow breaths. Then the breathing returns to normal without any stimulation or intervention. These pauses in breathing may be accompanied by minor oxygen desaturation and bradycardia. It usually occurs when the infant is sleeping deeply, but may occur with light sleep or even when awake. Studies have shown that 78% of healthy full-term infants experience episodes of periodic breathing in the first two weeks of life, which typically resolves in the first six months of life.
https://en.wikipedia.org/wiki/Non-dairy%20creamer
A non-dairy creamer, commonly also called tea whitener or coffee whitener or else just creamer, is a liquid or granular product intended to substitute for milk or cream as an additive to coffee, tea, hot chocolate or other beverages. They do not contain lactose and therefore are commonly described as being non-dairy products, although many contain casein, a milk-derived protein. Dry granular products do not need to be refrigerated and can be used and stored in locations which do not have a refrigerator. Liquid non-dairy creamers should be tightly capped and refrigerated after opening. Some non-dairy creamers contain sweeteners and flavors, such as vanilla, hazelnut or Irish cream. As with other processed food products, low calorie and low fat versions are available for non-dairy creamers. History Holton "Rex" Diamond, an employee of Rich Products, performed experiments from 1943 to 1945 with using a form of soybean protein" to make a "soy cream" that would not form curds when mixed with coffee. Diamond's experiments are the first English-language reference to a non-dairy creamer for coffee. Frank S. Mitchell, another Rich Products Corp. employee, and Diamond developed a non-dairy whipped topping for their employer in 1946. Mitchell also developed a non-dairy coffee creamer, Perx, which was successful in the market. In 1950, Melvin Morse and Dick Borne of Presto Foods developed "Mocha Mix Coffee Creamer", which was the first commercial non-dairy creamer and the first product with the term “coffee creamer” in the name. Another early commercial powdered creamer was "Pream", first marketed in 1952 and made from dehydrated cream and sugar. It did not dissolve easily because of the protein in the milk. Six years later, in 1958, the Carnation Company developed a product that easily dissolved in hot liquid because it replaced most of the milk fat with vegetable oil, and reduced the milk protein. The new product was marketed under the Carnation label with the brand nam
https://en.wikipedia.org/wiki/3Blue1Brown
3Blue1Brown is a math YouTube channel created and run by Grant Sanderson. The channel focuses on teaching higher mathematics from a visual perspective, and on the process of discovery and inquiry-based learning in mathematics, which Sanderson calls "inventing math". , the channel has 5.47 million subscribers. Grant Sanderson Early life and education Sanderson graduated from Stanford University in 2015 with a bachelor's degree in mathematics. He worked for Khan Academy from 2015 to 2016 as part of their content fellowship program, producing videos and articles about multivariable calculus, after which he started focusing his full attention on 3Blue1Brown. Career In 2020, Grant Sanderson became one of the creators and lecturers of the MIT course Introduction to Computational Thinking, together with Alan Edelman, David Sanders, James Schloss, and Benoit Forget. The course uses the Julia programming language and Grant Sanderson's animations to explain various topics: convolutions, image processing, COVID-19 data visualization, epidemic modelling, ray tracing, introduction to climate modelling, ocean modelling, and the algorithms that lie behind these topics. In February 2022, Sanderson determined that the best starting word on the game Wordle was CRANE using information theory. Later, he stated that the code he wrote to determine the best starting word had a bug in it, and the actual best word is SALET. In January 2020, Sanderson delivered a talk in An Evening with Grant Sanderson, hosted by the Stanford Speakers Bureau. Sanderson offered his perspective on engaging with math: instead of prioritizing usefulness, he emphasizes emotion, wonder and imagination. He aims to "bring life to math" with visuals, graphics, and animations. In August 2021, Sanderson was one of several featured speakers at SIGGRAPH 2021. In November 2022, Sanderson delivered a keynote speech, "What can algorithms teach us about education?", at the 17th Dutch National Informatics Congress Celer
https://en.wikipedia.org/wiki/Autosuggestion
Autosuggestion is a psychological technique related to the placebo effect, developed by pharmacist Émile Coué at the beginning of the 20th century. It is a form of self-induced suggestion in which individuals guide their own thoughts, feelings, or behavior. The technique is often used in self-hypnosis. Typological distinctions Émile Coué identified two very different types of self-suggestion: intentional, "reflective autosuggestion": made by deliberate and conscious effort, and unintentional, "spontaneous auto-suggestion": which is a "natural phenomenon of our mental life … which takes place without conscious effort [and has its effect] with an intensity proportional to the keenness of [our] attention". In relation to Coué's group of "spontaneous auto-suggestions", his student Charles Baudouin (1920, p. 41) made three further useful distinctions, based upon the sources from which they came: "Instances belonging to the representative domain   (sensations, mental images, dreams, visions, memories, opinions, and all intellectual phenomena)." "Instances belonging to the affective domain   (joy or sorrow, emotions, sentiments, tendencies, passions)." "Instances belonging to the active or motor domain   (actions, volitions, desires, gestures, movements at the periphery or in the interior of the body, functional or organic modifications)." Émile Coué Émile Coué, who had both B.A. and B.Sc. degrees before he was 21, graduated top of his class (with First Class Honours) with a degree in pharmacology from the prestigious Collège Sainte-Barbe in Paris in 1882. Having spent an additional six months as an intern at the Necker-Enfants Malades Hospital in Paris, he returned to Troyes, where he worked as an apothecary from 1882 to 1910. "Hypnosis" à la Ambroise-Auguste Liébeault and Hippolyte Bernheim In 1885, his investigations of hypnotism and the power of the imagination began with Ambroise-Auguste Liébeault and Hippolyte Bernheim, two leading exponents of "hypnosis", o
https://en.wikipedia.org/wiki/Fej%C3%A9r%27s%20theorem
In mathematics, Fejér's theorem, named after Hungarian mathematician Lipót Fejér, states the following: Explanation of Fejér's Theorem's Explicitly, we can write the Fourier series of f as where the nth partial sum of the Fourier series of f may be written as where the Fourier coefficients are Then, we can define with Fn being the nth order Fejér kernel. Then, Fejér's theorem asserts that with uniform convergence. With the convergence written out explicitly, the above statement becomes Proof of Fejér's Theorem We first prove the following lemma: Proof: Recall the definition of , the Dirichlet Kernel:We substitute the integral form of the Fourier coefficients into the formula for above Using a change of variables we get This completes the proof of Lemma 1. We next prove the following lemma: Proof: Recall the definition of the Fejér Kernel As in the case of Lemma 1, we substitute the integral form of the Fourier coefficients into the formula for This completes the proof of Lemma 2. We next prove the 3rd Lemma: This completes the proof of Lemma 3. We are now ready to prove Fejér's Theorem. First, let us recall the statement we are trying to prove We want to find an expression for . We begin by invoking Lemma 2: By Lemma 3a we know that Applying the triangle inequality yields and by Lemma 3b, we get We now split the integral into two parts, integrating over the two regions and . The motivation for doing so is that we want to prove that . We can do this by proving that each integral above, integral 1 and integral 2, goes to zero. This is precisely what we'll do in the next step. We first note that the function f is continuous on [-π,π]. We invoke the theorem that every periodic function on [-π,π] that is continuous is also bounded and uniformily continuous. This means that . Hence we can rewrite the integral 1 as follows Because and By Lemma 3a we then get for all n This gives the desired bound for integral 1 which we can exploit in fina
https://en.wikipedia.org/wiki/Institute%20of%20Physics%20Edward%20Appleton%20Medal%20and%20Prize
The Edward Appleton Medal and Prize is awarded by the Institute of Physics for distinguished research in environmental, earth or atmospheric physics. Originally named after Charles Chree, the British physicist and former President of the Physical Society of London, it was renamed in 2008 to commemorate Edward Victor Appleton, winner of the Nobel prize for proving the existence of the ionosphere. History The prize was established in 1941 by Chree's sister, Jessie, after his death, and it was originally awarded biennially. It was first awarded to Sydney Chapman. From 2001 it was awarded annually. After the 2008 renaming the prize was awarded in even-dated years until 2016, then as and when required. The cash prize part of the award has risen in value since its inception, reported at £150 in 1985 and £300 in 1987, to its present-day value of £1000. Winners Recipients of the Appleton medal and prize 2021 Philip Stier 2020 Adam Scaife 2019 Cathryn Mitchell 2016 Giles Harrison 2014 David Marshall 2012 Colin O'Dowd 2010 Myles Allen 2008 Ann Wintle Recipients of the Chree medal and prize See also Institute of Physics Awards List of physics awards List of awards named after people
https://en.wikipedia.org/wiki/Entomoplasmatales
Entomoplasmatales is a small order of mollicute bacteria. The genus Spiroplasma is part of this order. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI) See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/Runtime%20verification
Runtime verification is a computing system analysis and execution approach based on extracting information from a running system and using it to detect and possibly react to observed behaviors satisfying or violating certain properties. Some very particular properties, such as datarace and deadlock freedom, are typically desired to be satisfied by all systems and may be best implemented algorithmically. Other properties can be more conveniently captured as formal specifications. Runtime verification specifications are typically expressed in trace predicate formalisms, such as finite state machines, regular expressions, context-free patterns, linear temporal logics, etc., or extensions of these. This allows for a less ad-hoc approach than normal testing. However, any mechanism for monitoring an executing system is considered runtime verification, including verifying against test oracles and reference implementations . When formal requirements specifications are provided, monitors are synthesized from them and infused within the system by means of instrumentation. Runtime verification can be used for many purposes, such as security or safety policy monitoring, debugging, testing, verification, validation, profiling, fault protection, behavior modification (e.g., recovery), etc. Runtime verification avoids the complexity of traditional formal verification techniques, such as model checking and theorem proving, by analyzing only one or a few execution traces and by working directly with the actual system, thus scaling up relatively well and giving more confidence in the results of the analysis (because it avoids the tedious and error-prone step of formally modelling the system), at the expense of less coverage. Moreover, through its reflective capabilities runtime verification can be made an integral part of the target system, monitoring and guiding its execution during deployment. History and context Checking formally or informally specified properties again
https://en.wikipedia.org/wiki/Pterin-4%20alpha-carbinolamine%20dehydratase%20deficiency
Pterin-4 alpha-carbinolamine dehydratase deficiency (PCDD) is one of the known forms of tetrahydrobiopterin deficiency. This condition is associated with mutations of the PCBD1 gene. As of 2020, PCDD was the rarest form of BH4 deficiency in terms of cases described in medical literature. Symptoms Patients with PCDD are mostly asymptomatic, although they may have transient neurologic deficits in infancy, and sometimes hypomagnesemia and nonautoimmune diabetes mellitus in puberty. Diagnosis Pterin-4 alpha-carbinolamine dehydratase deficiency causes hyperphenylalaninemia and therefore can be suspected upon finding elevated levels of phenylalanine. To distinguish it from other forms of BH4 deficiency, further analyses are made. PCDD is associated with elevatel levels of primapterin, especially in urine, while biopterin levels range from low to normal, and neopterin levels from normal to high. Treatment Treatment involves a diet with a low phenylalanine content, and sapropterin to help normalize phenylalanine levels. Since phenylalanine levels in this disease have been reported to be only mildly elevated in the majority of patients, relaxation and discontinuation of phenylalanine-reduced diet and/or sapropterin supplementation can be attempted after the first year of life under careful monitoring of phenylalanine levels, according to a consensus guideline published in 2020. External links Hyperphenylalaninemia, BH4-deficient, D - description in the OMIM compendium.
https://en.wikipedia.org/wiki/Food%20self-provisioning
Food self-provisioning (FSP) is the growing of one's own food, especially fruits and vegetables. Also labelled as household food production, is a traditional activity persisting in the countries of the Global North. It is studied in Sustainability science and in ecofeminism on reason of its social, health and environmental outcomes.
https://en.wikipedia.org/wiki/Failure%20rate
Failure rate is the frequency with which an engineered system or component fails, expressed in failures per unit of time. It is usually denoted by the Greek letter λ (lambda) and is often used in reliability engineering. The failure rate of a system usually depends on time, with the rate varying over the life cycle of the system. For example, an automobile's failure rate in its fifth year of service may be many times greater than its failure rate during its first year of service. One does not expect to replace an exhaust pipe, overhaul the brakes, or have major transmission problems in a new vehicle. In practice, the mean time between failures (MTBF, 1/λ) is often reported instead of the failure rate. This is valid and useful if the failure rate may be assumed constant – often used for complex units / systems, electronics – and is a general agreement in some reliability standards (Military and Aerospace). It does in this case only relate to the flat region of the bathtub curve, which is also called the "useful life period". Because of this, it is incorrect to extrapolate MTBF to give an estimate of the service lifetime of a component, which will typically be much less than suggested by the MTBF due to the much higher failure rates in the "end-of-life wearout" part of the "bathtub curve". The reason for the preferred use for MTBF numbers is that the use of large positive numbers (such as 2000 hours) is more intuitive and easier to remember than very small numbers (such as 0.0005 per hour). The MTBF is an important system parameter in systems where failure rate needs to be managed, in particular for safety systems. The MTBF appears frequently in the engineering design requirements, and governs frequency of required system maintenance and inspections. In special processes called renewal processes, where the time to recover from failure can be neglected and the likelihood of failure remains constant with respect to time, the failure rate is simply the multiplica
https://en.wikipedia.org/wiki/NuttX
NuttX is a free and open-source Real-Time Operating System (RTOS) with an emphasis on technical standards compliance and on having a small footprint. Scalable from 8-bit to 64-bit microcontroller environments, the main governing standards in NuttX are from the Portable Operating System Interface (POSIX) and the American National Standards Institute (ANSI). Further standard application programming interfaces (APIs) from Unix and other common RTOSes (such as VxWorks) are adopted for functions unavailable under these standards, or inappropriate for deeply embedded environments, such as the fork() system call. NuttX was initially released in 2007 under the permissive BSD license. In December 2019, it began incubation at the Apache Software Foundation. changing its license form BSD to Apache License and graduated to a top-level project in November 2022. Usage NuttX RTOS is used in a variety of applications, including the Sony CXD5602/Spresense microcontroller as well as an audio recorder from Sony. The firmware for some of Motorola's MotoMod accessories for the Moto Z used NuttX RTOS, and NuttX is also used in the PX4 autopilot drones, which use NuttX to control a variety of autonomous platforms.
https://en.wikipedia.org/wiki/SPRY2
Sprouty homolog 2 (Drosophila), also known as SPRY2, is a protein which in humans is encoded by the SPRY2 gene. Function This gene encodes a protein belonging to the sprouty family. The encoded protein contains a carboxyl-terminal cysteine-rich domain essential for the inhibitory activity on receptor tyrosine kinase signaling proteins and is required for growth factor stimulated translocation of the protein to membrane ruffles. In primary dermal endothelial cells this gene is transiently upregulated in response to fibroblast growth factor two. This protein is indirectly involved in the non-cell autonomous inhibitory effect on fibroblast growth factor two signaling. The protein interacts with Cas-Br-M (murine) ectropic retroviral transforming sequence, and can function as a bimodal regulator of epidermal growth factor receptor/mitogen-activated protein kinase signaling. This protein may play a role in alveoli branching during lung development as shown by a similar mouse protein. SPRY2 is a negative feedback regulator of multiple receptor tyrosine kinases (RTKs) including receptors for fibroblast growth factor (FGF), epidermal growth factor (EGF), and hepatocyte growth factor (HGF). Antagonization of growth factor mediated pathways, cell migration, and cellular differentiation occurs through the ERK pathway. Spry2 can also enhance EGFR signaling by sequestering CBL. Spry gene expression has been reported silenced or repressed in cancer of the breast, liver, lung, prostate, and in lymphoma. Human spry2 expression is localized to the microtubules in unstimulated cells. All sprouty isoforms inhibit the ERK pathway by themselves, but can also form heterodimers and homodimers which have enhanced inhibition. Interactions SPRY2 has been shown to interact with Cbl gene. See also SPRED1 gene Neurofibromin 1 SPRY1
https://en.wikipedia.org/wiki/Transposase
A transposase is any of a class of enzymes capable of binding to the end of a transposon and catalysing its movement to another part of a genome, typically by a cut-and-paste mechanism or a replicative mechanism, in a process known as transposition. The word "transposase" was first coined by the individuals who cloned the enzyme required for transposition of the Tn3 transposon. The existence of transposons was postulated in the late 1940s by Barbara McClintock, who was studying the inheritance of maize, but the actual molecular basis for transposition was described by later groups. McClintock discovered that some segments of chromosomes changed their position, jumping between different loci or from one chromosome to another. The repositioning of these transposons (which coded for color) allowed other genes for pigment to be expressed. Transposition in maize causes changes in color; however, in other organisms, such as bacteria, it can cause antibiotic resistance. Transposition is also important in creating genetic diversity within species and generating adaptability to changing living conditions. Transposases are classified under EC number EC 2.7.7. Genes encoding transposases are widespread in the genomes of most organisms and are the most abundant genes known. During the course of human evolution, as much as 40% of the human genome has moved around via methods such as transposition of transposons. Transposase Tn5 Transposase (Tnp) Tn5 is a member of the RNase superfamily of proteins which includes retroviral integrases. Tn5 can be found in Shewanella and Escherichia bacteria. The transposon codes for antibiotic resistance to kanamycin and other aminoglycoside antibiotics. Tn5 and other transposases are notably inactive. Because DNA transposition events are inherently mutagenic, the low activity of transposases is necessary to reduce the risk of causing a fatal mutation in the host, and thus eliminating the transposable element. One of the reasons Tn5 is so unr
https://en.wikipedia.org/wiki/Intertrigo
Intertrigo refers to a type of inflammatory rash (dermatitis) of the superficial skin that occurs within a person's body folds. These areas are more susceptible to irritation and subsequent infection due to factors that promote skin breakdown such as moisture, friction, and exposure to bodily secretions and excreta such as sweat, urine, or feces. Areas of the body which are more likely to be affected by intertrigo include the inframammary fold, intergluteal cleft, armpits, and spaces between the fingers or toes. Skin affected by intertrigo is more prone to infection than intact skin. The term "intertrigo" commonly refers to a secondary infection with bacteria (such as Corynebacterium minutissimum), fungi (such as Candida albicans), or viruses. A frequent manifestation is candidal intertrigo. Intertrigo occurs more often in warm and humid conditions. Generally, intertrigo is more common in people with a weakened immune system including children, the elderly, and immunocompromised people. The condition is also more common in people who experience urinary incontinence and decreased ability to move. Cause An intertrigo usually develops from the chafing of warm, moist skin in the areas of the inner thighs and genitalia, the armpits, under the breasts, the underside of the belly, behind the ears, and the web spaces between the toes and fingers. An intertrigo usually appears red and raw-looking, and may also itch, ooze, and be sore. Intertrigos occur more often among overweight individuals, those with diabetes, those restricted to bed rest or diaper use, and those who use medical devices, like artificial limbs, that trap moisture against the skin. Also, there are several skin diseases that can cause an intertrigo to develop, such as dermatitis or inverse psoriasis. Bacterial Bacterial intertrigo can be caused by Streptococci and Corynebacterium minutissimum. Diagnosis Intertrigo can be diagnosed clinically by a medical professional after taking a thorough history and
https://en.wikipedia.org/wiki/Ibrahim%20Index%20of%20African%20Governance
The Ibrahim Index of African Governance (IIAG), established in 2007, provides an assessment of the quality of governance in African countries. The IIAG is compiled by 81 indicators and 265 variables from 54 data projects, coming from 47 independent African and international data sources. Published every two years, the IIAG is one of the world’s most comprehensive collections of data on African governance. The IIAG provides a framework for citizens, governments, institutions, academics and business to assess the delivery of public goods and services, and policy outcomes, across Africa. The Foundation defines governance as the provision of the political, social, economic and environmental goods that a citizen has the right to expect from their state, and that a state has the responsibility to deliver to its citizens. The IIAG assesses progress under four main conceptual categories: Security & Rule of Law, Participation, Rights & Inclusion, Foundations for Economic Opportunity, and Human Development. Scholars, development professionals, analysts, and policymakers have used the IIAG to benchmark governance performance across a number of dimensions at national, regional and continental levels. Scores and ranks are available for the latest 10-year period, enabling the analysis of trends over time. All underlying data used in the construction of the IIAG is freely available and transparently published alongside a comprehensive methodology. History The Ibrahim Index of African Governance (IIAG) is a key initiative of the Mo Ibrahim Foundation that was first published in 2007. The most recent iteration, the 2022 IIAG, was published in January 2023 and covers the period 2012-2021. The index was initially produced in association with Harvard University; academic and technical assistance has subsequently been provided by a range of African academics and research bodies. The IIAG is published every two years and receives extensive media attention from across the African
https://en.wikipedia.org/wiki/Archive%20for%20Mathematical%20Logic
Archive for Mathematical Logic is a peer-reviewed mathematics journal published by Springer Science+Business Media. It was established in 1950 and publishes articles on mathematical logic. Abstracting and indexing The journal is abstracted and indexed in: Mathematical Reviews Zentralblatt MATH Scopus SCImago According to the Journal Citation Reports, the journal has a 2020 impact factor of 0.287.
https://en.wikipedia.org/wiki/Cavalieri%27s%20quadrature%20formula
In calculus, Cavalieri's quadrature formula, named for 17th-century Italian mathematician Bonaventura Cavalieri, is the integral and generalizations thereof. This is the definite integral form; the indefinite integral form is: There are additional forms, listed below. Together with the linearity of the integral, this formula allows one to compute the integrals of all polynomials. The term "quadrature" is a traditional term for area; the integral is geometrically interpreted as the area under the curve y = xn. Traditionally important cases are y = x2, the quadrature of the parabola, known in antiquity, and y = 1/x, the quadrature of the hyperbola, whose value is a logarithm. Forms Negative n For negative values of n (negative powers of x), there is a singularity at x = 0, and thus the definite integral is based at 1, rather than 0, yielding: Further, for negative fractional (non-integer) values of n, the power xn is not well-defined, hence the indefinite integral is only defined for positive x. However, for n a negative integer the power xn is defined for all non-zero x, and the indefinite integrals and definite integrals are defined, and can be computed via a symmetry argument, replacing x by −x, and basing the negative definite integral at −1. Over the complex numbers the definite integral (for negative values of n and x) can be defined via contour integration, but then depends on choice of path, specifically winding number – the geometric issue is that the function defines a covering space with a singularity at 0. n = −1 There is also the exceptional case n = −1, yielding a logarithm instead of a power of x: (where "ln" means the natural logarithm, i.e. the logarithm to the base e = 2.71828...). The improper integral is often extended to negative values of x via the conventional choice: Note the use of the absolute value in the indefinite integral; this is to provide a unified form for the integral, and means that the integral of this odd function is
https://en.wikipedia.org/wiki/Syntax%20and%20semantics%20of%20logic%20programming
Logic programming is a programming paradigm that includes languages based on formal logic, including Datalog and Prolog. This article describes the syntax and semantics of the purely declarative subset of these languages. Confusingly, the name "logic programming" also refers to a programming language that roughly corresponds to the declarative subset of Prolog. Unfortunately, the term must be used in both senses in this article. Declarative logic programs consist entirely of rules of the form H :- B1, ..., BN. Each such rule can be read as an implication: meaning "If each is true, then is true". Logic programs compute the set of facts that are implied by their rules. Many implementations of Datalog, Prolog, and related languages add procedural features such as Prolog's cut operator or extra-logical features such as a foreign function interface. The formal semantics of such extensions are beyond the scope of this article. Datalog Datalog is the simplest widely-studied logic programming language. There are three major definitions of the semantics of Datalog, and they are all equivalent. The syntax and semantics of other logic programming languages are extensions and generalizations of those of Datalog. Syntax A Datalog program consists of a list of rules (Horn clauses). If constant and variable are two countable sets of constants and variables respectively and relation is a countable set of predicate symbols, then the following BNF grammar expresses the structure of a Datalog program: <program> ::= <rule> <program> | "" <rule> ::= <atom> ":-" <atom-list> "." <atom> ::= <relation> "(" <term-list> ")" <atom-list> ::= <atom> | <atom> "," <atom-list> | "" <term> ::= <constant> | <variable> <term-list> ::= <term> | <term> "," <term-list> | "" Atoms are also referred to as . The atom to the left of the :- symbol is called the of the rule; the atoms to the right are the . Every Datalog program must satisfy the condition that every variable that appears in th
https://en.wikipedia.org/wiki/List%20of%20pathology%20mnemonics
This is a list of pathology mnemonics, categorized and alphabetized. For mnemonics in other medical specialities, see this list of medical mnemonics. Acute intermittent porphyria: signs and symptoms 5 Ps: Pain in the abdomen Polyneuropathy Psychological abnormalities Pink urine Precipitated by drugs (including barbiturates, oral contraceptives, and sulfa drugs) Acute ischemia: signs [especially limbs] 6 P's: Pain Pallor Pulselessness Paralysis Paraesthesia Perishingly cold Anemia (normocytic): causes : Acute blood loss Bone marrow failure Chronic disease Destruction (hemolysis) Anemia causes (simplified) ANEMIA: Anemia of chronic disease No folate or B12 Ethanol Marrow failure & hemaglobinopathies Iron deficient Acute & chronic blood loss Atherosclerosis risk factors "You're a SAD BET with these risk factors": Sex: male Age: middle-aged, elderly Diabetes mellitus BP high: hypertension Elevated cholesterol Tobacco Carcinoid syndrome: components CARCinoid: Cutaneous flushing Asthmatic wheezing Right sided valvular heart lesions Cramping and diarrhea Cushing syndrome CUSHING: Central obesity/ Cervical fat pads/ Collagen fiber weakness/ Comedones (acne) Urinary free corisol and glucose increase Striae/ Suppressed immunity Hypercortisolism/ Hypertension/ Hyperglycemia/ Hirsutism Iatrogenic (Increased administration of corticosteroids) Noniatrogenic (Neoplasms) Glucose intolerance/Growth retardation Diabetic ketoacidosis: I vs. II ketONEbodies are seen in type ONEdiabetes. Gallstones: risk factors 5 F's: Fat Female Fair (gallstones more common in Caucasians) Fertile (premenopausal- increased estrogen is thought to increase cholesterol levels in bile and decrease gallbladder contractions) Forty or above (age) Hepatomegaly: 3 common causes, 3 rarer causes Common are 3 C's: Cirrhosis Carcinoma Cardiac failure Rarer are 3 C's: Cholestasis Cysts Cellular infiltration Hyperkalemia (signs and symptoms) MURDER Mus
https://en.wikipedia.org/wiki/TMEM275
TMEM275 (Transmembrane protein 275) is a protein that in humans is encoded by the TMEM275 gene. TMEM275 has two, highly-conserved, helical trans-membrane regions. It is predicted to reside within the plasma membrane or the endoplasmic reticulum's membrane. Gene Locus In humans, the gene is located on chromosomal band 1p33 on the minus strand. It specifically resides on chromosome 1 at chr1:46,532,166-46,543,969. TMEM275 is 11,804bp long. Gene neighborhood Upstream, on the minus strand, of TMEM275 is another gene; KNCN. Transcripts The TMEM275 gene encodes four exons, only three of which are included within the final mRNA transcript. Two of those three are within the 5’ UTR and the coding sequence codes for 177 amino acids. TMEM275 is found two have two potential isoforms outside of its reference transcript. The reference isoform does not include exon 1, while isoform X1 has a shortened 5' UTR and isoform X2 has a shortened 3' UTR. However, all versions of mRNA transcripts yield the same 177 amino acid sequence. Protein General properties The TMEM275 protein contains 177 amino acids with a predicted molecular mass of 17.2 kDa and a pI of 8.13. TMEM275 was found to have a higher presence of alanine and proline amino acid residues than most proteins. When looking at orthologous proteins sequences within other species, the alanine presence was conserved throughout, but the same could not be said of the proline presence. Primary sequence TMEM275's protein consists of 177 amino acids. The protein, or polypeptide chain, that is encoded by the coding sequence is made through the process of translation and is shown below among other regions of interest. There is also a poly-A, or polyadenylation signal, towards the end of the 3'UTR. Domains TMEM275 has two, highly-conserved, helical trans-membrane regions. The regions can be seem within the amino acid sequence above within the conceptual translation in purple. Evolutionary analysis showed that these trans-membr
https://en.wikipedia.org/wiki/Fuzzy%20number
A fuzzy number is a generalization of a regular real number in the sense that it does not refer to one single value but rather to a connected set of possible values, where each possible value has its own weight between 0 and 1. This weight is called the membership function. A fuzzy number is thus a special case of a convex, normalized fuzzy set of the real line. Just like fuzzy logic is an extension of Boolean logic (which uses absolute truth and falsehood only, and nothing in between), fuzzy numbers are an extension of real numbers. Calculations with fuzzy numbers allow the incorporation of uncertainty on parameters, properties, geometry, initial conditions, etc. The arithmetic calculations on fuzzy numbers are implemented using fuzzy arithmetic operations, which can be done by two different approaches: (1) interval arithmetic approach; and (2) the extension principle approach. A fuzzy number is equal to a fuzzy interval. The degree of fuzziness is determined by the a-cut which is also called the fuzzy spread. See also Fuzzy set Uncertainty Interval arithmetic Random variable
https://en.wikipedia.org/wiki/Lesser%20sciatic%20notch
Below the ischial spine is a small notch, the lesser sciatic notch; it is smooth, coated in the recent state with cartilage, the surface of which presents two or three ridges corresponding to the subdivisions of the tendon of the obturator internus, which winds over it. It is converted into a foramen, the lesser sciatic foramen, by the sacrotuberous and sacrospinous ligaments, and transmits the tendon of the obturator internus, the nerve which supplies that muscle, and the internal pudendal vessels and pudendal nerve. See also Lesser sciatic foramen Greater sciatic notch
https://en.wikipedia.org/wiki/Gustav%20Bergmann
Gustav Bergmann (May 4, 1906 – April 21, 1987) was an Austrian-born American philosopher. He studied at the University of Vienna and was a member of the Vienna Circle. Bergmann was influenced by the philosophers Moritz Schlick, Friedrich Waismann, and Rudolf Carnap, who were members of the Circle. In the United States, he was a professor of philosophy and psychology at the University of Iowa. Biography Bergmann was born in Vienna, Austria-Hungary. He earned his Ph.D. in mathematics at the University of Vienna in 1928. His dissertation, directed by Walther Mayer, was titled Zwei Beiträge zur mehrdimensionalen Differentialgeometrie. While studying for his doctorate, he was invited to join the Vienna Circle, a group of philosophers, mathematicians, scientists, and others committed to a scientific worldview under the name of logical positivism. In 1930–31, he worked with Albert Einstein in Berlin. Unable as a Jew to find academic employment, Bergmann obtained a J.D. degree from the University of Vienna in 1935, and practiced corporate law until he and his family fled to the United States in 1938. Settling at the University of Iowa in Iowa City in 1939, Bergmann eventually became professor of both philosophy and psychology. He died in Iowa City. Bibliography The Metaphysics of Logical Positivism. New York: Longmans, Green & Co. 1954. (Second edition: Madison, University of Wisconsin Press, 1967.) Philosophy of Science. Madison: University of Wisconsin Press 1957. Meaning and Existence. Madison: University of Wisconsin Press 1959. Logic and Reality. Madison: University of Wisconsin Press 1964. Realism: A Critique of Brentano and Meinong. Madison: University of Wisconsin Press 1967. New Foundations of Ontology. Madison: University of Wisconsin Press 1992. Edited by William Heald. Collected Works. Vol I. II. Frankfurt am Main: Ontos Verlag 2003. See also American philosophy List of American philosophers
https://en.wikipedia.org/wiki/Antiparallel%20%28biochemistry%29
In biochemistry, two biopolymers are antiparallel if they run parallel to each other but with opposite directionality (alignments). An example is the two complementary strands of a DNA double helix, which run in opposite directions alongside each other. Nucleic acids Nucleic acid molecules have a phosphoryl (5') end and a hydroxyl (3') end. This notation follows from organic chemistry nomenclature, and can be used to define the movement of enzymes such as DNA polymerases relative to the DNA strand in a non-arbitrary manner. G-quadruplexes G-quadruplexes, also known as G4 DNA are secondary structures found in nucleic acids that are rich in guanine. These structures are normally located at the telomeres (the ends of the chromosomes). The G-quadruplex can either be parallel or antiparallel depending on the loop configuration, which is a component of the structure. If all the DNA strands run in the same direction, it is termed to be a parallel quadruplex, and is known as a strand-reversal/propeller, connecting adjacent parallel strands. If one or more of the DNA strands run in opposite direction, it is termed as an anti-parallel quadruplex, and can either be in a form of a lateral/edgewise, connecting adjacent anti-parallel strands, or a diagonal, joining two diagonally opposite strands. The structure of these G-quadruplexes can be determined by a cation. DNA replication In DNA, the 5' carbon is located at the top of the leading strand, and the 3' carbon is located at the lower section of the lagging strand. The nucleic acid sequences are complementary and parallel, but they go in opposite directions, hence the antiparallel designation. The antiparallel structure of DNA is important in DNA replication because it replicates the leading strand one way and the lagging strand the other way. During DNA replication, the leading strand is replicated continuously whereas the lagging strand is replicated in segments known as Okazaki fragments. Anti-parallelism in biochemi
https://en.wikipedia.org/wiki/Lorenz%2096%20model
The Lorenz 96 model is a dynamical system formulated by Edward Lorenz in 1996. It is defined as follows. For : where it is assumed that and and . Here is the state of the system and is a forcing constant. is a common value known to cause chaotic behavior. It is commonly used as a model problem in data assimilation. Python simulation from scipy.integrate import odeint import matplotlib.pyplot as plt import numpy as np # These are our constants N = 5 # Number of variables F = 8 # Forcing def L96(x, t): """Lorenz 96 model with constant forcing""" # Setting up vector d = np.zeros(N) # Loops over indices (with operations and Python underflow indexing handling edge cases) for i in range(N): d[i] = (x[(i + 1) % N] - x[i - 2]) * x[i - 1] - x[i] + F return d x0 = F * np.ones(N) # Initial state (equilibrium) x0[0] += 0.01 # Add small perturbation to the first variable t = np.arange(0.0, 30.0, 0.01) x = odeint(L96, x0, t) # Plot the first three variables fig = plt.figure() ax = fig.add_subplot(projection="3d") ax.plot(x[:, 0], x[:, 1], x[:, 2]) ax.set_xlabel("$x_1$") ax.set_ylabel("$x_2$") ax.set_zlabel("$x_3$") plt.show() Julia simulation using DynamicalSystems, PyPlot PyPlot.using3D() # parameters and initial conditions N = 5 F = 8.0 u₀ = F * ones(N) u₀[1] += 0.01 # small perturbation # The Lorenz-96 model is predefined in DynamicalSystems.jl: ds = Systems.lorenz96(N; F = F) # Equivalently, to define a fast version explicitly, do: struct Lorenz96{N} end # Structure for size type function (obj::Lorenz96{N})(dx, x, p, t) where {N} F = p[1] # 3 edge cases explicitly (performance) @inbounds dx[1] = (x[2] - x[N - 1]) * x[N] - x[1] + F @inbounds dx[2] = (x[3] - x[N]) * x[1] - x[2] + F @inbounds dx[N] = (x[1] - x[N - 2]) * x[N - 1] - x[N] + F # then the general case for n in 3:(N - 1) @inbounds dx[n] = (x[n + 1] - x[n - 2]) * x[n - 1] - x[n] + F end return nothing end lor96 = Lorenz96
https://en.wikipedia.org/wiki/Acidianus%20infernus
Acidianus infernus is a species of archaeon. It is aerobic, extremely acidophilic, thermophilic (hence its name) and sulfur-metabolizing. Its type strain is strain DSM 3191.
https://en.wikipedia.org/wiki/Primordial%20phallus
In embryology, the primordial phallus refers to the clitoris of a female or the penis in the male, particularly during fetal development of the urinary and reproductive organs, before sexual differentiation is evident. This is also the case for the immature male analog, the immature glans penis. See also Aphallia Genital tubercle Embryology of urogenital system
https://en.wikipedia.org/wiki/Cru%20%28wine%29
Cru is a wine term used to indicate a high-quality vineyard or group of vineyards. It is a French word which was originally used to refer to anything grown in a region and that region but is now mostly used to refer to a vineyard and its wines. The term is often used within classifications of French wine. By implication, a wine that displays (or is allowed to display) the name of its cru on its wine label is supposed to exhibit the typical characteristics of this vineyard or group of vineyards. The terms premier cru and grand cru designate levels of presumed quality that are variously defined in different wine regions. Premier cru Premier cru is a French language wine term corresponding to "first growth" and which can be used to refer to classified vineyards, wineries and wines, with different meanings in different wine regions: For Bordeaux wine, the term is applied to classified wineries: In the Bordeaux Wine Official Classification of 1855, Premier cru or Premier cru classé is the highest level of five within the "Grand cru classé" designation for red wines from the Médoc and Graves, and the second-highest of three in Sauternes where the highest is Premier cru supérieur (superior first growth). These wines are often referred to as first growths in English. In the classification of Saint-Émilion wine, the highest level is Premier grand cru classé A and the second highest Premier grand cru classé B. The term Saint-Émilion grand cru refers to wineries or wines below the overall Grand cru classé level, and is integrated within the appellation rules. For Burgundy wine, the term is applied to classified vineyards, with Premier cru being the second highest classification level, below that of grand cru and above the basic village AOCs. For Burgundy wines, the terms premier cru (abbreviated 1er cru) are usually kept rather than being translated into English. Grand cru Grand cru (French for 'great growth') is a regional wine classification that designates a vineyar
https://en.wikipedia.org/wiki/Bioclimatology
Bioclimatology is the interdisciplinary field of science that studies the interactions between the biosphere and the Earth's atmosphere on time scales of the order of seasons or longer (in contrast to biometeorology). Examples of relevant processes Climate processes largely control the distribution, size, shape and properties of living organisms on Earth. For instance, the general circulation of the atmosphere on a planetary scale broadly determines the location of large deserts or the regions subject to frequent precipitation, which, in turn, greatly determine which organisms can naturally survive in these environments. Furthermore, changes in climates, whether due to natural processes or to human interferences, may progressively modify these habitats and cause overpopulation or extinction of indigenous species. The biosphere, for its part, and in particular continental vegetation, which constitutes over 99% of the total biomass, has played a critical role in establishing and maintaining the chemical composition of the Earth's atmosphere, especially during the early evolution of the planet (See History of Earth for more details on this topic). Currently, the terrestrial vegetation exchanges some 60 billion tons of carbon with the atmosphere on an annual basis (through processes of carbon fixation and carbon respiration), thereby playing a critical role in the carbon cycle. On a global and annual basis, small imbalances between these two major fluxes, as do occur through changes in land cover and land use, contribute to the current increase in atmospheric carbon dioxide.
https://en.wikipedia.org/wiki/Shebang%20%28Unix%29
In computing, a shebang is the character sequence consisting of the characters number sign and exclamation mark () at the beginning of a script. It is also called sharp-exclamation, sha-bang, hashbang, pound-bang, or hash-pling. When a text file with a shebang is used as if it is an executable in a Unix-like operating system, the program loader mechanism parses the rest of the file's initial line as an interpreter directive. The loader executes the specified interpreter program, passing to it as an argument the path that was initially used when attempting to run the script, so that the program may use the file as input data. For example, if a script is named with the path path/to/script, and it starts with the line #!/bin/sh, then the program loader is instructed to run the program /bin/sh, passing path/to/script as the first argument. The shebang line is usually ignored by the interpreter, because the "#" character is a comment marker in many scripting languages; some language interpreters that do not use the hash mark to begin comments still may ignore the shebang line in recognition of its purpose. Syntax The form of a shebang interpreter directive is as follows: #! interpreter [optional-arg] in which interpreter is a path to an executable program. The space between and interpreter is optional. There could be any number of spaces or tabs either before or after interpreter. The optional-arg will include any extra spaces up to the end-of-line. In Linux, the file specified by interpreter can be executed if it has the execute rights and is one of the following: a native executable, such as an ELF binary any kind of file for which an interpreter was registered via the binfmt_misc mechanism (such as for executing Microsoft .exe binaries using wine) another script starting with a shebang On Linux and Minix, an interpreter can also be a script. A chain of shebangs and wrappers yields a directly executable file that gets the encountered scripts as parameters in
https://en.wikipedia.org/wiki/Megaprime
A megaprime is a prime number with at least one million decimal digits. Other terms for large primes include titanic prime, coined by Samuel Yates in the 1980s for a prime with at least 1000 digits (of which the smallest is 10999+7), and gigantic prime for a prime with at least 10,000 digits (of which the smallest is 109999+33603). Bevaprime has been proposed for a prime with at least 1,000,000,000 digits. , there are 2171 known megaprimes and 104 further probable primes (PRPs) which have more than 1,000,000 digits. The first to be found was the Mersenne prime 26972593−1 with 2,098,960 digits, discovered in 1999 by Nayan Hajratwala, a participant in the distributed computing project GIMPS. Nayan was awarded a Cooperative Computing Award from the Electronic Frontier Foundation for this achievement. Almost all primes are megaprimes, as the number of primes with fewer than one million digits is finite. However, the vast majority of known primes are not megaprimes. All numbers from 10999999 through 10999999 + 593498 are known to be composite, and there is a very high probability that 10999999 + 593499, a strong probable prime, is the smallest megaprime. , the smallest number known to be a megaprime is 10999999 + 308267*10292000 + 1. The last prime that is not a megaprime is almost certainly 10999999 - 172473. See also List of largest known primes and probable primes, a list that includes the largest known megaprimes and probable megaprimes Largest known prime number
https://en.wikipedia.org/wiki/Thermophotovoltaic%20energy%20conversion
Thermophotovoltaic (TPV) energy conversion is a direct conversion process from heat to electricity via photons. A basic thermophotovoltaic system consists of a hot object emitting thermal radiation and a photovoltaic cell similar to a solar cell but tuned to the spectrum being admitted from the hot object. As TPV systems generally work at lower temperatures than solar cells, their efficiencies tend to be low. Offsetting this through the use of multi-junction cells based on non-silicon materials is common, but generally very expensive. This currently limits TPV to niche roles like spacecraft power and waste heat collection from larger systems like steam turbines. General concept PV Typical photovoltaics work by creating a p–n junction near the front surface of a thin semiconductor material. When photons above the bandgap energy of the material hit atoms within the bulk lower layer, below the junction, an electron is photoexcited and becomes free of its atom. The junction creates an electric field that accelerates the electron forward within the cell until it passes the junction and is free to move to the thin electrodes patterned on the surface. Connecting a wire from the front to the rear allows the electrons to flow back into the bulk and complete the circuit. Photons with less energy than the bandgap do not eject electrons. Photons with energy above the bandgap will eject higher-energy electrons which tend to thermalize within the material and lose their extra energy as heat. If the cell's bandgap is raised, the electrons that are emitted will have higher energy when they reach the junction and thus result in a higher voltage, but this will reduce the number of electrons emitted as more photons will be below the bandgap energy and thus generate a lower current. As electrical power is the product of voltage and current, there is a sweet spot where the total output is maximized. Terrestrial solar radiation is typically characterized by a standard known as Air M
https://en.wikipedia.org/wiki/Projective%20Set%20%28game%29
Projective Set (sometimes shortened to ProSet) is a real-time card game derived from the older game Set. The deck contains cards consisting of colored dots; some cards are laid out on the table and players attempt to find "Sets" among them. The word projective comes from the game's relation to Projective spaces over the finite field with two elements. Projective Set has been studied mathematically as well as played recreationally. It has been a popular game at Canada/USA Mathcamp. Rules A Projective Set card has six binary attributes, or bits, generally represented by colored dots. For each color of dot, each card either has that dot or does not. There is one card for each possible combination of dots except the combination of no dots at all, making cards total. Three cards are said to form a "set" if the total number of dots of each color is either 0 or 2. Similarly, four or more cards form a "set" if the number of dots of each color is an even number. A card and itself could be said to form a two-card set, but as the cards in the deck are all distinct, this does not arise in actual gameplay. Original Version In the original version, as in Set, 12 cards are laid out on the table. The first player to find three cards which form a set and call out "set" takes the three cards. Three new cards are then dealt and the play continues until the deck is depleted. If at any time the players agree there is no set among the cards, three new cards can be dealt, bringing the total number of cards on the table to 15. Other than this, new cards are not dealt out unless the number of cards on the table goes below 12. The game ends when the deck is depleted and no more sets can be found among the cards on the table. The player who captured the most sets is the winner. 7-card Version A variation of the game, more popular than the original, allows sets of any size, rather than just sets of size three. 7 cards are put out on the table at a time, and when a set is found (
https://en.wikipedia.org/wiki/Open%20range
In the Western United States and Canada, open range is rangeland where cattle roam freely regardless of land ownership. Where there are "open range" laws, those wanting to keep animals off their property must erect a fence to keep animals out; this applies to public roads as well. Land in open range that is designated as part of a "herd district" reverses liabilities, requiring an animal's owner to fence it in or otherwise keep it on the person's own property. Most eastern states and jurisdictions in Canada require owners to fence in or herd their livestock. History and practice The Western open-range tradition originated from the early practice of unregulated grazing of livestock in the newly acquired western territories of the United States and Canada. These practices were eventually codified in the laws of many Western US states as they developed written statutes. Over time, as the Western lands became more populated and more developed (through railroads, mining, farming, etc.), open-range laws began to be challenged and were significantly curtailed, though they still exist in certain areas of most Western US states and Canadian provinces. Open-range conditions existed in Western Canada prior to the 1889 amendments of the Dominion Lands Act, which prohibited cattle from grazing on unleased land, though the practice did not disappear immediately. Open-range management has also been practiced in other areas, including the Caribbean and some Eastern US states, such as South Carolina during the colonial period. The practice was also widespread in Mexico, and some argue that the Mexican tradition may have been the predecessor to open-range practices in the American West, much of which was part of Mexico prior to the 1840s. American ranchers borrowed many other cattle-raising techniques from Mexico. Unlike the Eastern United States, the Western prairies of the 19th century were vast, undeveloped, and uncultivated. The land was also generally much more arid, wit
https://en.wikipedia.org/wiki/Logical%20Disk%20Manager
The Logical Disk Manager (LDM) is an implementation of a logical volume manager for Microsoft Windows NT, developed by Microsoft and Veritas Software. It was introduced with the Windows 2000 operating system, and is supported in Windows XP, Windows Server 2003, Windows Vista, Windows 7, Windows 8, Windows 10 and Windows 11. The MMC-based Disk Management snap-in () hosts the Logical Disk Manager. On Windows 8 and Windows Server 2012, Microsoft deprecated LDM in favor of Storage Spaces. Logical Disk Manager enables disk volumes to be dynamic, in contrast to the standard basic volume format. Basic volumes and dynamic volumes differ in their ability to extend storage beyond one physical disk. Basic partitions are restricted to a fixed size on one physical disk. Dynamic volumes can be enlarged to include more free space - either from the same disk or another physical disk. (For more information on the difference, see Basic and dynamic disks and volumes, below.) Overview Basic storage involves dividing a disk into primary and extended partitions. This is the route that all versions of Windows that were reliant on DOS-handled storage took, and disks formatted in this manner are known as basic disks. Dynamic storage involves the use of a single partition that covers the entire disk, and the disk itself is divided into volumes or combined with other disks to form volumes that are greater in size than one disk itself. Volumes can use any supported file system. Basic disks can be upgraded to dynamic disks; however, when this is done the disk cannot easily be downgraded to a basic disk again. To perform a downgrade, data on the dynamic disk must first be backed up onto some other storage device. Second, the dynamic disk must be re-formatted as a basic disk (erasing all data). Finally, data from the backup must be copied back over to the newly re-formatted basic disk. Dynamic disks provide the capability for software implementations of RAID. The main disadvantage of dynami
https://en.wikipedia.org/wiki/ThYme%20%28database%29
ThYme (Thioester-active enzYme) is database of enzymes constituting the fatty acid synthesis and polyketide synthesis cycles. See also Thioester
https://en.wikipedia.org/wiki/United%20States%20Cyber%20Command
United States Cyber Command (USCYBERCOM) is one of the eleven unified combatant commands of the United States Department of Defense (DoD). It unifies the direction of cyberspace operations, strengthens DoD cyberspace capabilities, and integrates and bolsters DoD's cyber expertise. USCYBERCOM was established as a Sub-Unified command under U.S. Strategic Command at the direction of Secretary of Defense Robert Gates on June 23, 2009 at the National Security Agency (NSA) headquarters in Fort George G. Meade, Maryland. It cooperates with NSA networks and has been concurrently headed by the director of the National Security Agency since its inception. While originally created with a defensive mission in mind, it has increasingly been viewed as an offensive force. On 18 August 2017, it was announced that USCYBERCOM would be elevated to the status of a full and independent unified combatant command. On 23 May 2023, it was announced that President Biden nominated Lt. Gen. Timothy Haugh to the Senate to be the next USCYBERCOM Commander. Mission statement According to the US Department of Defense (DoD): The text "9ec4c12949a4f31474f299058ce2b22a", located in the command's emblem, is the MD5 hash of their mission statement. The command is charged with pulling together existing cyberspace resources, creating synergies and synchronizing war-fighting effects to defend the information security environment. USCYBERCOM is tasked with centralizing command of cyberspace operations, strengthening DoD cyberspace capabilities, and integrating and bolstering DoD's cyber expertise. Organizational structure Cyber teams Since 2015, the U.S. Cyber Command added 133 new cyber teams. The breakdown was: Thirteen National Mission Teams to defend against broad cyberattacks Sixty-eight Cyber Protection Teams to defend priority DoD networks and systems against priority threats Twenty-seven Combat Mission Teams to provide integrated cyberspace attacks in support of operational plans and con
https://en.wikipedia.org/wiki/Strip%20packing%20problem
The strip packing problem is a 2-dimensional geometric minimization problem. Given a set of axis-aligned rectangles and a strip of bounded width and infinite height, determine an overlapping-free packing of the rectangles into the strip minimizing its height. This problem is a cutting and packing problem and is classified as an Open Dimension Problem according to Wäscher et al. This problem arises in the area of scheduling, where it models jobs, that require a contiguous portion of the memory over a given time period. Another example is the area of industrial manufacturing, where rectangular pieces need to be cut out of a sheet of material (e.g., cloth or paper) that has a fixed width but infinite length, and one wants to minimize the wasted material. This problem was first studied in 1980. It is strongly-NP hard and there exists no polynomial time approximation algorithm with a ratio smaller than unless . However, the best approximation ratio achieved so far (by a polynomial time algorithm by Harren et al.) is , imposing an open question whether there is an algorithm with approximation ratio . Definition An instance of the strip packing problem consists of a strip with width and infinite height, as well as a set of rectangular items. Each item has a width and a height . A packing of the items is a mapping that maps each lower-left corner of an item to a position inside the strip. An inner point of a placed item is a point from the set . Two (placed) items overlap if they share an inner point. The height of the packing is defined as . The objective is to find an overlapping-free packing of the items inside the strip while minimizing the height of the packing. This definition is used for all polynomial time algorithms. For pseudo-polynomial time and FPT-algorithms, the definition is slightly changed for the simplification of notation. In this case, all appearing sizes are integral. Especially the width of the strip is given by an arbitrary integ
https://en.wikipedia.org/wiki/Dilly%20beans
Dilly beans, or pickled green beans, are a means of preserving this summer legume. Often flavored with dill, hence the name, they may also contain garlic, Tabasco sauce, and red pepper. Best kept in glass jars for safekeeping over the winter months, they can be served on their own as a snack or alongside a main dish or in salad. While they are made in kitchens all over the United States, they are particularly common in Vermont, where the overabundance of green beans produced during the short summer needs to be preserved for enjoyment during the long winter. Dilly beans were developed as a commercial product in 1958 by Sonya Hagna and Jacquelyn Park, who made them the subject of a well-known radio advertising campaign.
https://en.wikipedia.org/wiki/Swecha
Swecha is a non-profit organization formerly called as Free Software Foundation Andhra Pradesh (FSF-AP) later changed its name to Swecha. It is a Telugu Operating System released in the year 2005, and is a part of Free Software Movement of India (FSMI). The organization is a social movement working towards educating the masses with the essence of Free Software and to provide knowledge to the commoners. Swecha organizes workshops and seminars in the Indian state of Telangana and Andhra Pradesh. Presently Swecha is active GLUG (GNU/Linux User Group) in many engineering colleges like International Institute of Information Technology, Hyderabad , Jawaharlal Nehru Technological University, Hyderabad, Chaitanya Bharathi Institute of Technology, St. Martin's Engineering College, Sridevi Women's Engineering College, Mahatma Gandhi Institute of Technology, SCIENT Institute of Technology, CMR Institute of Technology, Hyderabad, Jyothishmathi College of Engineering and Technology, MVGR College of Engineering, K L University and Ace Engineering College. Objectives The main objectives of the organization are as follows: To take forward free software and its ideological implications to all corners of our country from the developed domains to the underprivileged. To create awareness among computer users in the use of free software. To work towards usage of free software in all streams of sciences and research. To take forward implementation and usage of free software in school education, academics and higher education. To work towards e-literacy and bridging digital divide based on free software and mobilizing the underprivileged. To work among developers on solutions catering to societal & national requirements. To work towards a policy change favoring free software in all walks of life. Activities Swecha hosted a National Convention for Academics and Research which was attended by researchers and academicians from different parts of the country. Former President of In
https://en.wikipedia.org/wiki/Globe%20effect
The globe effect, sometimes called the rolling ball effect or the spinning globe effect, is an optical phenomenon—perhaps partially an optical illusion—that occurs with visual optical instruments, in particular binoculars and telescopes, that are designed to be free of distortion. When these instruments are panned, the moving image appears to roll over a curved, convex surface. In 1949, Horst Koehler at Zeiss (Jena) suggested adding some pincushion distortion to the optical design to eliminate the globe effect. August Sonnefeld conducted experiments with volunteers, which supported the claim that a supplementary distortion could improve the imaging of visual optical instruments. Since that time, most binocular manufacturers have followed Zeiss's example and added pincushion distortion to their optical design. The origin of the globe effect initially remained unclear after its discovery in the first half of the past century. Koehler speculated about an "unnatural perspective generated by the binocular while panning over a three dimensional scenery", thereby ignoring the fact that the globe effect was observable at the night sky as well, where perspective distortions were absent. Recently, the globe effect has been linked to the peculiar properties of human visual perception, which adds a certain amount of barrel distortion to the visual field. The amount of barrel distortion is subject to individual differences, which explains the fact that the perceived intensity of the globe effect varies significantly between different observers. An alternative approach for explaining the globe effect comes from the technical journalist and optics specialist Walter E. Schön. He states that the observed effect is in fact not that of a rolling globe but that of a vertically rotating cylinder. The globe shape of the illusion seen by most observers is only because the field of view through the optical device is circular. This illusion of a rotating cylinder during panning is cause
https://en.wikipedia.org/wiki/Nicotinamide%20ribonucleoside%20uptake%20transporters
The Nicotinamide Ribonucleoside (NR) Uptake Permease (PnuC) Family (TC# 4.B.1) is a family of transmembrane transporters that is part of the TOG superfamily. Close PnuC homologues are found in a wide range of Gram-negative and Gram-positive bacteria, archaea and eukaryotes. Function PnuC of Salmonella typhimurium and Haemophilus influenzae are believed to function cooperatively with NadR homologues, multifunctional proteins that together with PnuC, participate in NR phosphorylation, transport and transcriptional regulation. NadR, a cytoplasmic protein that is partly membrane associated, contains one well conserved and one poorly conserved mononucleotide-binding consensus sequence (G-X4 GKS). It drives transport and may render transport responsive to internal pyridine nucleotide levels. While its N-terminal half functions as a repressor, its C-terminal half functions as an NR kinase in a putative group translocation process. PnuC of Haemophilus influenzae The H. influenzae homologue has been shown to transport NR from the periplasm into the cytoplasm. Phosphorylation of NR by NadR is required for NR uptake. The ribonucleoside kinase (RNK) domain has both Walker A and Walker B motifs, responsible for ATP binding and phosphoryl transfer. In addition, a proposed LID domain was identified in RNK. LID domains have been found in other kinases, and these domains are regions which are able to move after substrate binding. They are responsible for coordination of three distinct conformations, an open state in the absence of substrate, a partially closed state after substrate binding, and a fully closed state when both substrates are present. In H. influenzae, NR enters the NAD+ resynthesis pathway after phosphorylation to NMN, and subsequently, NAD+ is synthesized from NMN and ATP via an NMN adenylyl transferase activity. NadR represents a multifunctional regulator/enzyme complex able to integrate several functions, such as enzymatic catalysis, transport, and transcriptio
https://en.wikipedia.org/wiki/Sipser%E2%80%93Lautemann%20theorem
In computational complexity theory, the Sipser–Lautemann theorem or Sipser–Gács–Lautemann theorem states that bounded-error probabilistic polynomial (BPP) time is contained in the polynomial time hierarchy, and more specifically Σ2 ∩ Π2. In 1983, Michael Sipser showed that BPP is contained in the polynomial time hierarchy. Péter Gács showed that BPP is actually contained in Σ2 ∩ Π2. Clemens Lautemann contributed by giving a simple proof of BPP’s membership in Σ2 ∩ Π2, also in 1983. It is conjectured that in fact BPP=P, which is a much stronger statement than the Sipser–Lautemann theorem. Proof Here we present the Lautemann's proof. Without loss of generality, a machine M ∈ BPP with error ≤ 2−|x| can be chosen. (All BPP problems can be amplified to reduce the error probability exponentially.) The basic idea of the proof is to define a Σ2 sentence that is equivalent to stating that x is in the language, L, defined by M by using a set of transforms of the random variable inputs. Since the output of M depends on random input, as well as the input x, it is useful to define which random strings produce the correct output as A(x) = {r | M(x,r) accepts}. The key to the proof is to note that when x ∈ L, A(x) is very large and when x ∉ L, A(x) is very small. By using bitwise parity, ⊕, a set of transforms can be defined as A(x) ⊕ t={r ⊕ t | r ∈ A(x)}. The first main lemma of the proof shows that the union of a small finite number of these transforms will contain the entire space of random input strings. Using this fact, a Σ2 sentence and a Π2 sentence can be generated that is true if and only if x ∈ L (see conclusion). Lemma 1 The general idea of lemma one is to prove that if A(x) covers a large part of the random space then there exists a small set of translations that will cover the entire random space. In more mathematical language: If , then , where such that Proof. Randomly pick t1, t2, ..., t|r|. Let (the union of all transforms of A(x)). So, for all r
https://en.wikipedia.org/wiki/Marcescence
Marcescence is the withering and persistence of plant organs that normally are shed, and is a term most commonly applied to plant leaves. The underlying physiological mechanism is that trees transfer water and sap from the roots to the leaves through their vascular cells, but in some trees as autumn begins, the veins carrying the sap slowly close until a layer of cells called the abscission layer completely closes off the vein allowing the tree to rid itself of the leaf. Leaf marcescence is most often seen on juvenile plants and may disappear as the tree matures. It also may not affect the entire tree; sometimes leaves persist only on scattered branches. Marcescence is most obvious in deciduous trees that retain leaves through the winter. Trees that exhibit marcescence are known as "everciduous". Several trees normally have marcescent leaves such as oak (Quercus), beech (Fagus) and hornbeam (Carpinus), or marcescent stipules as in some but not all species of willows (Salix). All oak trees may display foliage marcescence, even species that are known to fully drop leaves when the tree is mature. Marcescent leaves of pin oak (Quercus palustris) complete development of their abscission layer in the spring. The base of the petiole remains alive over the winter. Many other trees may have marcescent leaves in seasons where an early freeze kills the leaves before the abscission layer develops or completes development. Diseases or pests can also kill leaves before they can develop an abscission layer. Marcescent leaves may be retained indefinitely and do not break off until mechanical forces (wind for instance) cause the dry and brittle petioles to snap. The evolutionary reasons for marcescence are not clear, theories include: protection of leaf buds from winter desiccation, and as a delayed source of nutrients or moisture-conserving mulch when the leaves finally fall and decompose in spring. Many palms form a skirt-like or shuttlecock-like crown of marcescent leaves unde
https://en.wikipedia.org/wiki/Committee%20machine
A committee machine is a type of artificial neural network using a divide and conquer strategy in which the responses of multiple neural networks (experts) are combined into a single response. The combined response of the committee machine is supposed to be superior to those of its constituent experts. Compare with ensembles of classifiers. Types Static structures In this class of committee machines, the responses of several predictors (experts) are combined by means of a mechanism that does not involve the input signal, hence the designation static. This category includes the following methods: Ensemble averaging In ensemble averaging, outputs of different predictors are linearly combined to produce an overall output. Boosting In boosting, a weak algorithm is converted into one that achieves arbitrarily high accuracy. Dynamic structures In this second class of committee machines, the input signal is directly involved in actuating the mechanism that integrates the outputs of the individual experts into an overall output, hence the designation dynamic. There are two kinds of dynamic structures: Mixture of experts In mixture of experts, the individual responses of the experts are non-linearly combined by means of a single gating network. Hierarchical mixture of experts In hierarchical mixture of experts, the individual responses of the individual experts are non-linearly combined by means of several gating networks arranged in a hierarchical fashion.
https://en.wikipedia.org/wiki/Radiant%20energy%20density
In radiometry, radiant energy density is the radiant energy per unit volume. The SI unit of radiant energy density is the joule per cubic metre (J/m3). Mathematical definition Radiant energy density, denoted we ("e" for "energetic", to avoid confusion with photometric quantities), is defined as where ∂ is the partial derivative symbol; Qe is the radiant energy; V is the volume. Relation to other radiometric quantities Because radiation always transmits the energy, it is useful to wonder what the speed of the transmission is. If all the radiation at given location propagates in the same direction, then the radiant flux through a unit area perpendicular to the propagation direction is given by the irradiance: where c is the radiation propagation speed. Contrarily if the radiation intensity is equal in all directions, like in a cavity in a thermodynamic equilibrium, then the energy transmission is best described by radiance: Radiant exitance through a small opening from such a cavity is: These relations can be used for example in the black-body radiation equation's derivation. SI radiometry units
https://en.wikipedia.org/wiki/Teva%20Active%20Pharmaceutical%20Ingredients
Teva API is an international pharmaceutical company headquartered in Israel. Teva API is a stand-alone business unit of Teva Pharmaceutical Industries limited, the largest generic drug manufacturer in the world and one of the 15 largest pharmaceutical companies worldwide. On top of supplying a major share of Teva's own needs, the Teva API division is an active competitor in world markets, investing both in the development of new products and manufacturing processes and in the upgrading of production facilities. In 2014, Teva API's sales to third parties totaled $724 million. In recent years growth occurred in all of Teva API's principal geographical markets: North America, Europe and International. History At the heart of the API division is the Israel-based Teva-Tech, formerly known as the Assia and Plantex plants which manufacture, develop and market raw materials for pharmaceuticals. Acquisitions Teva api has grown by acquiring manufacturing and development facilities around the world. Today, teva api operates 21 plants and sales offices worldwide. Major Teva api's acquisitions: 2011 - Teva Acquisition of Theramex in Monaco Teva api 2008 Acquisitions December 2008 – Teva Acquisition of Barr-Pliva in Croatia July 2008 – Teva Acquisition of Bentley in Spain April 2008 – Teva api Acquisition of Archimica in Puerto Rico Acquisitions & Foundations since the 1990s 2006 – Acquisition of Wanma in China 2006 – Acquisition of Ivax Corporation 2004 – Acquisition of Sicor API in Italy, Mexico and Switzerland 2003 – Acquisition of RDL in India 2002 – Acquisition of PFC in Italy 1996 – Acquisition of BIOCRAFT in Missouri (US) 1995 – Foundation of TEVA-TECH in Israel 1995 – Acquisition of ICI in Italy & BIOGAL in Hungary 1991 – Acquisition of PROSINTEX in Italy Facilities Research and development The R&D group at teva api consists of a team of over 760 top scientists located in 7 development centers worldwide: A large center in Israel
https://en.wikipedia.org/wiki/Guido%20Fubini
Guido Fubini (19 January 1879 – 6 June 1943) was an Italian mathematician, known for Fubini's theorem and the Fubini–Study metric. Life Born in Venice, he was steered towards mathematics at an early age by his teachers and his father, who was himself a teacher of mathematics. In 1896 he entered the Scuola Normale Superiore di Pisa, where he studied differential geometry under Ulisse Dini and Luigi Bianchi. His 1900 doctoral thesis was about Clifford's parallelism in elliptic spaces. After earning his doctorate, he took up a series of professorships. In 1901 he began teaching at the University of Catania in Sicily; shortly afterwards he moved to the University of Genoa; and in 1908 he moved to the Politecnico in Turin and then the University of Turin, where he stayed for a few decades. During this time his research focused primarily on topics in mathematical analysis, especially differential equations, functional analysis, and complex analysis; but he also studied the calculus of variations, group theory, non-Euclidean geometry, and projective geometry, among other topics. With the outbreak of World War I, he shifted his work towards more applied topics, studying the accuracy of artillery fire; after the war, he continued in an applied direction, applying results from this work to problems in electrical circuits and acoustics. In 1938, when Fubini at the age of 59 was nearing retirement, Benito Mussolini's Fascists adopted the anti-Jewish policies advocated for several years by Adolf Hitler's Nazis. As a Jew, Fubini feared for the safety of his family, and so accepted an invitation by Princeton University to teach there; he died in New York City four years later. Legacy A main-belt asteroid, 22495 Fubini, was named in his honour. Publications 1920: Lezioni di analisi matematica (Società Tipografico-Editrice Nazionale, Torino)
https://en.wikipedia.org/wiki/Journal%20of%20Biosocial%20Science
The Journal of Biosocial Science is a bimonthly peer-reviewed scientific journal covering the intersection of biology and sociology. It was the continuation of The Eugenics Review, published by the Galton Institute from 1909 till 1968. It obtained its current name in 1969, with volume numbering re-starting at 1, and switched publishers to Cambridge University Press. The editor-in-chief is Dr Alejandra Núñez-de la Mora (Universidad Veracruzana). Abstracting and indexing The journal is abstracted and indexed in: According to the Journal Citation Reports, the journal has a 2021 impact factor of 2.148. Notable studies In 2006, the journal published a controversial study arguing that Ashkenazi Jews are more intelligent than other ethnic groups as a result of human evolution. Past editors Alan Sterling Parkes (1969–1978) Derek F. Roberts (1979–1988) C. G. Nicholas Mascie-Taylor (1989–2021)
https://en.wikipedia.org/wiki/DNA%20sequencing%20theory
DNA sequencing theory is the broad body of work that attempts to lay analytical foundations for determining the order of specific nucleotides in a sequence of DNA, otherwise known as DNA sequencing. The practical aspects revolve around designing and optimizing sequencing projects (known as "strategic genomics"), predicting project performance, troubleshooting experimental results, characterizing factors such as sequence bias and the effects of software processing algorithms, and comparing various sequencing methods to one another. In this sense, it could be considered a branch of systems engineering or operations research. The permanent archive of work is primarily mathematical, although numerical calculations are often conducted for particular problems too. DNA sequencing theory addresses physical processes related to sequencing DNA and should not be confused with theories of analyzing resultant DNA sequences, e.g. sequence alignment. Publications sometimes do not make a careful distinction, but the latter are primarily concerned with algorithmic issues. Sequencing theory is based on elements of mathematics, biology, and systems engineering, so it is highly interdisciplinary. The subject may be studied within the context of computational biology. Theory and sequencing strategies Sequencing as a covering problem All mainstream methods of DNA sequencing rely on reading small fragments of DNA and subsequently reconstructing these data to infer the original DNA target, either via assembly or alignment to a reference. The abstraction common to these methods is that of a mathematical covering problem. For example, one can imagine a line segment representing the target and a subsequent process where smaller segments are "dropped" onto random locations of the target. The target is considered "sequenced" when adequate coverage accumulates (e.g., when no gaps remain). The abstract properties of covering have been studied by mathematicians for over a century. However, dir
https://en.wikipedia.org/wiki/AlkB
AlkB (Alkylation B) is a protein found in E. coli, induced during an adaptive response and involved in the direct reversal of alkylation damage. AlkB specifically removes alkylation damage to single stranded (SS) DNA caused by SN2 type of chemical agents. It efficiently removes methyl groups from 1-methyl adenines, 3-methyl cytosines in SS DNA. AlkB is an alpha-ketoglutarate-dependent hydroxylase, a superfamily non-haem iron-containing proteins. It oxidatively demethylates the DNA substrate. Demethylation by AlkB is accompanied with release of CO2, succinate, and formaldehyde. Human homologs There are nine human homologs of AlkB. They are: Alkb homolog 1, histone h2a dioxygenase, , , , AlkB homolog 5, RNA demethylase, , , , ABH3, like E. coli AlkB, is specific for SS DNA and RNA whereas ABH2 has higher affinity for damages in double-stranded DNA. ALKBH8 has a RNA recognition motif, a methyltransferase domain, and an AlkB-like domain. The methyltransferase domain generates the wobble nucleoside 5-methoxycarbonylmethyluridine (mcm5U) from its precursor 5-carboxymethyluridine (cm5U). The AlkB-like domain generates (S)-5-methoxycarbonylhydroxymethyluridine (mchm5U)in Gly-tRNA-UCC. FTO, which is associated with obesity in humans, is the first identified RNA demethylase. It demethylates N6-methyladenosine in mRNA. There is also another very different protein called AlkB or alkane hydroxylase. It is the catalytic subunit of a non-heme diiron protein, catalyzing the hydroxylation of alkanes, in aerobic bacteria that are able to utilize alkanes as a carbon source. Virus homologs AlkB domains are present within viral replication-associated proteins of plant RNA viruses of the families Closteroviridae, Alphaflexiviridae, Betaflexiviridae, and Secoviridae. Potyviridae is the largest family of plant RNA viruses; among these the AlkB domain is embedded in P1 proteases of endive necrotic mosaic virus (ENMV) of genus Potyvirus, French endive necrotic mosaic virus (FEN
https://en.wikipedia.org/wiki/Chronic%20deciduitis
Chronic deciduitis is a type of long-lasting inflammation that arises in pregnancy and affects the endometrial stromal tissue (decidua). It is associated with preterm labour. The diagnosis rests primarily on the presence of plasma cells. See also Chorioamnionitis Decidua
https://en.wikipedia.org/wiki/Cleversticks
Cleversticks is a plastic construction toy manufactured in the UK. The "Cleversticks Construction System" The original design manufactured in polypropylene used sticks with pairs of jaws on each end and rings, the jaws connect onto the body of the sticks or rings, The jaws also connected the sticks end to end. The resulting toy was best suited to young children; with these few components, a wide range of models could be assembled. Squares, triangles, rectangles and hexagon shapes were later added to allow the toy to be used for sorting into sets; the toy could now be used in primary education. Technical shapes were added to increase the complexity of the toy and appeal to older children. The most recent additions utilise the new patented design manufactured in Acrylonitrile butadiene styrene (ABS) including wheels with tyres manufactured in Santoprene, a synthetic rubber. Patent, registered design, and trademark The first patent in the UK GB 8313507 has a priority date of May 17, 1983, a US patent No 4548590 was granted in October 1985. The most recent UK patent GB2417434 was granted on 24 February 2009. The following registered designs have also been obtained: UK No 2066049, UK No 2106334, Ger No 40204339.1, Aus 1447/2002 and UK No 3019498. The name CLEVERSTICKS was registered under the No 1557320K, as a GB trademark in December 1993.
https://en.wikipedia.org/wiki/Slot-die%20coating
Slot-die coating is a coating technique for the application of solution, slurry, or extruded thin films onto typically flat substrates such as glass, metal, paper, fabric or plastic foils. The process was first developed for the industrial production of photographic papers in the 1950's. It has since become relevant in numerous commercial processes and nanomaterials related research fields. Slot-die coating produces thin films via solution processing. The desired coating material is typically dissolved or suspended into a precursor solution or slurry (sometimes referred to as "ink") and delivered onto the surface of the substrate through a precise coating head known as a slot-die. The slot-die has a high aspect ratio outlet controlling the final delivery of the coating liquid onto the substrate. This results in the continuous production of a wide layer of coated material on the substrate, with adjustable width depending on the dimensions of the slot-die outlet. By closely controlling the rate of solution deposition and the relative speed of the substrate, slot-die coating affords thin material coatings with easily controllable thicknesses in the range of 10 nanometers to hundreds of micrometers after evaporation of the precursor solvent. Commonly cited benefits of the slot-die coating process include its pre-metered thickness control, non-contact coating mechanism, high material efficiency, scalability of coating areas and throughput speeds, and roll-to-roll compatibility. The process also allows for a wide working range of layer thickness and precursor solution properties such as material choice, viscosity, and solids content. Commonly cited drawbacks of the slot-die coating process include its comparatively high complexity of apparatus and process optimization relative to similar coating techniques such as blade coating and spin coating. Furthermore, slot-die coating falls into the category of coating processes rather than printing processes. It is therefore bet
https://en.wikipedia.org/wiki/Curve
In mathematics, a curve (also called a curved line in older texts) is an object similar to a line, but that does not have to be straight. Intuitively, a curve may be thought of as the trace left by a moving point. This is the definition that appeared more than 2000 years ago in Euclid's Elements: "The [curved] line is […] the first species of quantity, which has only one dimension, namely length, without any width nor depth, and is nothing else than the flow or run of the point which […] will leave from its imaginary moving some vestige in length, exempt of any width." This definition of a curve has been formalized in modern mathematics as: A curve is the image of an interval to a topological space by a continuous function. In some contexts, the function that defines the curve is called a parametrization, and the curve is a parametric curve. In this article, these curves are sometimes called topological curves to distinguish them from more constrained curves such as differentiable curves. This definition encompasses most curves that are studied in mathematics; notable exceptions are level curves (which are unions of curves and isolated points), and algebraic curves (see below). Level curves and algebraic curves are sometimes called implicit curves, since they are generally defined by implicit equations. Nevertheless, the class of topological curves is very broad, and contains some curves that do not look as one may expect for a curve, or even cannot be drawn. This is the case of space-filling curves and fractal curves. For ensuring more regularity, the function that defines a curve is often supposed to be differentiable, and the curve is then said to be a differentiable curve. A plane algebraic curve is the zero set of a polynomial in two indeterminates. More generally, an algebraic curve is the zero set of a finite set of polynomials, which satisfies the further condition of being an algebraic variety of dimension one. If the coefficients of the polynomials bel
https://en.wikipedia.org/wiki/Indigenous%20data%20governance
Data governance in the context of Indigenous data involves supporting the data interests, gaps and priorities of Indigenous peoples, in order to enable Indigenous self-determination. Generally, data governance refers to who has ownership, control and access over the use of data. Indigenous data governance requires the data to surround Indigenous peoples and its purpose to reflect Indigenous needs and priorities, rather than omitting Indigenous peoples in the production of Indigenous data. Overview Indigenous data governance is key in enabling Indigenous self-determinism and rebuilding strong Indigenous nations. Oftentimes, Indigenous peoples do not have access to relevant Indigenous data. Currently in Canada, much information on Indigenous peoples are considered government data that fall under Crown copyright, limiting access to relevant data such as archeological sites that are of significance to Indigenous nations. Thus, Indigenous data that lacks strong data governance often misrepresent Indigenous peoples, help inform policies that have discriminatory impacts on Indigenous peoples, and uphold colonial practices. Definition of Indigenous data Indigenous data can include knowledge and information on census, health and other administrative data about Indigenous peoples, information on the environment, non-humans and resources, and information on cultural heritage such as oral histories, clan knowledge and cultural sites. Indigenous data be produced by Indigenous people, governments, other institutions, and corporations. In terms of rebuilding Indigenous nations, Indigenous data can be useful for tribal governments when making decisions about their resources and communities. Indigenous data sovereignty Companies and states often have the power in deciding what kind of data is produced and for what purposes. Data sovereignty in the context of Indigenous data is about ensuring that Indigenous people have a say in the data that is produced about them, how this d
https://en.wikipedia.org/wiki/Paving%20matroid
In the mathematical theory of matroids, a paving matroid is a matroid in which every circuit has size at least as large as the matroid's rank. In a matroid of rank every circuit has size at most , so it is equivalent to define paving matroids as the matroids in which the size of every circuit belongs to the set . It has been conjectured that almost all matroids are paving matroids. Examples Every simple matroid of rank three is a paving matroid; for instance this is true of the Fano matroid. The Vámos matroid provides another example, of rank four. Uniform matroids of rank have the property that every circuit is of length exactly and hence are all paving matroids; the converse does not hold, for example, the cycle matroid of the complete graph is paving but not uniform. A Steiner system is a pair where is a finite set of size and is a family of -element subsets of with the property that every distinct elements of are contained in exactly one set in . The elements of form a -partition of and hence are the hyperplanes of a paving matroid on . d-Partitions If a paving matroid has rank , then its hyperplanes form a set system known as a -partition. A family of two or more sets forms a -partition if every set in has size at least and every -element subset of is a subset of exactly one set in . Conversely, if is a -partition, then it can be used to define a paving matroid on for which is the set of hyperplanes. In this matroid, a subset of is independent whenever either or and is not a subset of any set in . Combinatorial enumeration Combinatorial enumeration of the simple matroids on up to nine elements has shown that a large fraction of them are also paving matroids. On this basis, it has been conjectured that almost all matroids are paving matroids. More precisely, according to this conjecture, the limit, as n goes to infinity, of the ratio between the number of paving matroids and the number of all matroids should equal one. If so, the
https://en.wikipedia.org/wiki/Karmarkar%27s%20algorithm
Karmarkar's algorithm is an algorithm introduced by Narendra Karmarkar in 1984 for solving linear programming problems. It was the first reasonably efficient algorithm that solves these problems in polynomial time. The ellipsoid method is also polynomial time but proved to be inefficient in practice. Denoting as the number of variables and as the number of bits of input to the algorithm, Karmarkar's algorithm requires operations on -digit numbers, as compared to such operations for the ellipsoid algorithm. The runtime of Karmarkar's algorithm is thus using FFT-based multiplication (see Big O notation). Karmarkar's algorithm falls within the class of interior-point methods: the current guess for the solution does not follow the boundary of the feasible set as in the simplex method, but moves through the interior of the feasible region, improving the approximation of the optimal solution by a definite fraction with every iteration and converging to an optimal solution with rational data. The algorithm Consider a linear programming problem in matrix form: Karmarkar's algorithm determines the next feasible direction toward optimality and scales back by a factor . It is described in a number of sources. Karmarkar also has extended the method to solve problems with integer constraints and non-convex problems. Since the actual algorithm is rather complicated, researchers looked for a more intuitive version of it, and in 1985 developed affine scaling, a version of Karmarkar's algorithm that uses affine transformations where Karmarkar used projective ones, only to realize four years later that they had rediscovered an algorithm published by Soviet mathematician I. I. Dikin in 1967. The affine-scaling method can be described succinctly as follows. While applicable to small scale problems, it is not a polynomial time algorithm. stopping criterion, . return unbounded end if end do Example Cons
https://en.wikipedia.org/wiki/Chlorin
In organic chemistry, chlorins are tetrapyrrole pigments that are partially hydrogenated porphyrins. The parent chlorin is an unstable compound which undergoes air oxidation to porphine. The name chlorin derives from chlorophyll. Chlorophylls are magnesium-containing chlorins and occur as photosynthetic pigments in chloroplasts. The term "chlorin" strictly speaking refers to only compounds with the same ring oxidation state as chlorophyll. Chlorins are excellent photosensitizing agents. Various synthetic chlorins analogues such as m-tetrahydroxyphenylchlorin (mTHPC) and mono-L-aspartyl chlorin e6 are effectively employed in experimental photodynamic therapy as photosensitizer. Chlorophylls The most abundant chlorin is the photosynthetic pigment chlorophyll. Chlorophylls have a fifth, ketone-containing ring unlike the chlorins. Diverse chlorophylls exists, such as chlorophyll a, chlorophyll b, chlorophyll d, chlorophyll e, chlorophyll f, and chlorophyll g. Chlorophylls usually feature magnesium as a central metal atom, replacing the two NH centers in the parent. Variation Microbes produce two reduced variants of chlorin, bacteriochlorins and isobacteriochlorins. Bacteriochlorins are found in some bacteriochlorophylls; the ring structure is produced by Chlorophyllide a reductase (COR) reducing a chlorin ring at the C7-8 double boud. Isobacteriochlorins are found in nature mostly as sirohydrochlorin, a biosynthetic intermediate of vitamin B12, produced without going through a chlorin. In living organisms, both are ultimately derived from uroporphyrinogen III, a near-universal intermediate in tetrapyrrole biosynthesis. Synthetic chlorins Numerous synthetic chlorins with different functional groups and/or ring modifications have been examined. Contracted chlorins can be synthesised by reduction of B(III)subporphyrin or by oxidation of corresponding B(III)subbacteriochlorin. The B(III)subchlorins were directly synthesized as meso-ester B(III)subchlorin from me
https://en.wikipedia.org/wiki/Motivating%20operation
Motivating operation (MO) is a behavioristic concept introduced by Jack Michael in 1982. It is used to explain variations in the effects in the consequences of behavior. Most importantly, an MO affects how strongly the person is reinforced or punished by the consequences of their behavior. For example, food deprivation is a motivating operation; if a person is hungry, food is strongly reinforcing, but if a person is satiated, food is less reinforcing. In 2003 Laraway suggested subdividing MOs into those that increase the reinforcing or punishing effects of a stimulus, which are termed establishing operations, and MOs that decrease the reinforcing or punishing effects of a stimulus, which are termed abolishing operations. Theory The concept of motivating operation deals with the observation that behavior depends not only on the stimuli present in the current situation and the organism's past experience with those stimuli, but also on the organism's recent past history of deprivation, satiation, pain, or other such influences. Such a past history can have two effects: it can change the value of a consequence by making it more or less reinforcing, and/or it can change the probability of behaviors that have produced that consequence. For example, food deprivation changes the value of food, making it more reinforcing, and it also evokes learned behaviors that have obtained food. Likewise, food satiation reduces both the reinforcing effect of food and the probability of food-getting behaviors. Note that a motivating operation differs from a discriminative stimulus (Sd). A discriminative stimulus signals the availability of reinforcement, while a motivating operation changes the effectiveness of a reinforcer. Nine main unconditioned (i.e. not learned) motivating operations, have been identified in humans. Deprivation of food, water, sleep, activity, sex, or oxygen; becoming too warm or too cold; and increase of a painful stimulus all function as establishing operation
https://en.wikipedia.org/wiki/Mr.%20Do%21
is a 1982 maze game developed by Universal. It is the first arcade video game to be released as a conversion kit for other arcade machines; Taito published the conversion kit in Japan. The game was inspired by Namco's Dig Dug released earlier in 1982. Mr. Do! was a commercial success in Japan and North America, selling 30,000 arcade units in the US, and it was followed by several arcade sequels. Gameplay The object of Mr. Do! is to score as many points as possible by digging tunnels through the ground and collecting cherries. The title character, Mr. Do (a circus clown—except for the original Japanese version of the game, in which he is a snowman), is constantly chased by red dinosaur-like monsters called creeps, and the player loses a life if Mr. Do is caught by one. The game ends when the last life is lost. Cherries are distributed throughout the level in groups of eight, and collecting all the cherries in one group without a pause awards bonus points. A level is complete either when all cherries are removed, all creeps are destroyed, "EXTRA" is spelled, or a diamond is found. Mr. Do can defeat creeps by hitting them with his bouncing "power ball" or by dropping large apples on them. While the power ball is bouncing toward a creep, Mr. Do is defenseless. If the ball bounces into an area where there are no creeps to hit (such as behind a fallen apple), Mr. Do cannot use it again until he has retrieved it. When the power ball hits a creep, it then reforms in Mr. Do's hands after a delay that increases with each use. Mr. Do or the creeps can push an apple off the edge of a vertical tunnel and crush one or more creeps. If an apple falls more than its own height, it breaks and disappears. Mr. Do can also be crushed by a falling apple, causing a loss of life. Occasionally, the creeps transform briefly into more powerful multicolored monsters that can tunnel through the ground. If one of these digs through a cherry, it leaves fewer cherries for Mr. Do to collect. W
https://en.wikipedia.org/wiki/Farallon%20virus
Farallon virus is a strain of Hughes orthonairovirus in the genus Orthonairovirus belonging to the Hughes serogroup. A known host of the virus is Ornithodoros. The virus is named after the Farallon Islands.
https://en.wikipedia.org/wiki/Fellgett%27s%20advantage
Fellgett's advantage or the multiplex advantage is an improvement in signal-to-noise ratio (SNR) that is gained when taking multiplexed measurements rather than direct measurements. The name is derived from P. B. Fellgett, who first made the observation as part of his PhD. When measuring a signal whose noise is dominated by detector noise, a multiplexed measurement such as the signal generated by a Fourier transform spectrometer can produce a relative improvement in SNR, compared to an equivalent scanning monochromator, of the order of the square root of m, where m is the number of sample points comprising the spectrum. Exit slit Sellar and Boreman have argued that this SNR improvement can be considered as a result of freedom from needing an exit slit inside the spectrometer, since an exit slit reduces the light collected by the detector by the same factor. Emission There is an additional multiplex advantage for emission lines of atomic and molecular spectra. At the peak of the emission line, a monochromator measurement will be noisy, since the noise is proportional to the square root of the signal. For the same reason, the measurement will be less noisy at the baseline of the spectrum. In a multiplexed measurement, however, the noise in a given measurement is spread more or less evenly across the spectrum, regardless of the local signal intensity. Thus, multiplexed measurements can achieve higher SNR at the emission line peaks. There is a corresponding multiplex disadvantage, however. When the signals of interest are absorption lines in the spectrum, then the same principle will produce increased noise at the valleys of the absorption lines relative to the noise of a scanning monochromator. Shot noise However, if the detector is shot noise dominated (which is typically the case for a photomultiplier tube), noise will be proportional to the square root of the power, so that for a broad flat spectrum the noise will be proportional to the square root of m, where m
https://en.wikipedia.org/wiki/Abel%27s%20summation%20formula
In mathematics, Abel's summation formula, introduced by Niels Henrik Abel, is intensively used in analytic number theory and the study of special functions to compute series. Formula Let be a sequence of real or complex numbers. Define the partial sum function by for any real number . Fix real numbers , and let be a continuously differentiable function on . Then: The formula is derived by applying integration by parts for a Riemann–Stieltjes integral to the functions and . Variations Taking the left endpoint to be gives the formula If the sequence is indexed starting at , then we may formally define . The previous formula becomes A common way to apply Abel's summation formula is to take the limit of one of these formulas as . The resulting formulas are These equations hold whenever both limits on the right-hand side exist and are finite. A particularly useful case is the sequence for all . In this case, . For this sequence, Abel's summation formula simplifies to Similarly, for the sequence and for all , the formula becomes Upon taking the limit as , we find assuming that both terms on the right-hand side exist and are finite. Abel's summation formula can be generalized to the case where is only assumed to be continuous if the integral is interpreted as a Riemann–Stieltjes integral: By taking to be the partial sum function associated to some sequence, this leads to the summation by parts formula. Examples Harmonic numbers If for and then and the formula yields The left-hand side is the harmonic number . Representation of Riemann's zeta function Fix a complex number . If for and then and the formula becomes If , then the limit as exists and yields the formula where is the Riemann zeta function. This may be used to derive Dirichlet's theorem that has a simple pole with residue 1 at . Reciprocal of Riemann zeta function The technique of the previous example may also be applied to other Dirichlet series. If is the Möbi
https://en.wikipedia.org/wiki/European%20Transactions%20on%20Telecommunications
European Transactions on Telecommunications is a peer-reviewed scientific journal covering all aspects of information technology and telecommunications. Abstracting and indexing The journal is abstracted and indexed in the following databases: According to the Journal Citation Reports, the journal has a 2010 impact factor of 0.448, ranking it 55th out of 78 journals in the category "Telecommunications".
https://en.wikipedia.org/wiki/Passive%20infrared%20sensor
A passive infrared sensor (PIR sensor) is an electronic sensor that measures infrared (IR) light radiating from objects in its field of view. They are most often used in PIR-based motion detectors. PIR sensors are commonly used in security alarms and automatic lighting applications. PIR sensors detect general movement, but do not give information on who or what moved. For that purpose, an imaging IR sensor is required. PIR sensors are commonly called simply "PIR", or sometimes "PID", for "passive infrared detector". The term passive refers to the fact that PIR devices do not radiate energy for detection purposes. They work entirely by detecting infrared radiation (radiant heat) emitted by or reflected from objects. Operating principles All objects with a temperature above absolute zero emit heat energy in the form of electromagnetic radiation. Usually this radiation isn't visible to the human eye because it radiates at infrared wavelengths, but it can be detected by electronic devices designed for such a purpose. PIR-based motion detector A PIR-based motion detector is used to sense movement of people, animals, or other objects. They are commonly used in burglar alarms and automatically activated lighting systems. Operation A PIR sensor can detect changes in the amount of infrared radiation impinging upon it, which varies depending on the temperature and surface characteristics of the objects in front of the sensor. When an object, such as a person, passes in front of the background, such as a wall, the temperature at that point in the sensor's field of view will rise from room temperature to body temperature, and then back again. The sensor converts the resulting change in the incoming infrared radiation into a change in the output voltage, and this triggers the detection. Objects of similar temperature but different surface characteristics may also have a different infrared emission pattern, and thus moving them with respect to the background may trig
https://en.wikipedia.org/wiki/Navigation%20Catalyst%20Systems
Navigation Catalyst Systems, previously known as Vendare, First Look, and qsrch.net, was a pay-per-click advertising company that specialized in monetizing parked domain names and registrars' wildcard DNS records. Navigation Catalyst Systems was a subsidiary of New.net and an affiliate of the ICANN-accredited registrar Basic Fusion. Lawsuit In 2007, Navigation Catalyst Systems, along with its affiliate registrar Basic Fusion, was sued by Verizon for the registration and domain tasting of 1,392 domain names incorporating Verizon's trademarks (Verizon California, Inc. v. Navigation Catalyst Systems, Inc., 2008 WL 2651163 (C.D. Cal. June 30, 2008). Navigation argued that it did not register the domain names in bad faith because it had used a proprietary software tool that, through both manual and automatic means, "scrubbed" its registered domains for registered trademarks. It claimed that it was merely "reserving" the domain names by taking advantage of Basic Fusion's Add Grace Period. Basic Fusion attempted to argue that it was not the same legal entity as Navigation Catalyst Systems, but the court stated that they were the same for the purposes of a preliminary injunction motion. Navigation further asserted an unclean hands defense, arguing that Verizon used a similar pay-per-click program to monetize its wildcard DNS records. The court rejected both arguments and ultimately granted a preliminary injunction against Navigation Catalyst Systems. In so doing, the court held that Navigation Catalyst Systems had intended to profit in bad faith from the typographical errors of Internet users that mistyped Verizon's trademarks in the URL bar of their web browsers: It is clear that their intent was to profit from the poor typing abilities of consumers trying to reach Plaintiffs' sites: what other value could there be in a name like ve3rizon.com? Further, the sites associated with these names often contained links to products directly competitive with Plaintiffs'
https://en.wikipedia.org/wiki/Signed-digit%20representation
In mathematical notation for numbers, a signed-digit representation is a positional numeral system with a set of signed digits used to encode the integers. Signed-digit representation can be used to accomplish fast addition of integers because it can eliminate chains of dependent carries. In the binary numeral system, a special case signed-digit representation is the non-adjacent form, which can offer speed benefits with minimal space overhead. History Challenges in calculation stimulated early authors Colson (1726) and Cauchy (1840) to use signed-digit representation. The further step of replacing negated digits with new ones was suggested by Selling (1887) and Cajori (1928). In 1928, Florian Cajori noted the recurring theme of signed digits, starting with Colson (1726) and Cauchy (1840). In his book History of Mathematical Notations, Cajori titled the section "Negative numerals". For completeness, Colson uses examples and describes addition (pp. 163–4), multiplication (pp. 165–6) and division (pp. 170–1) using a table of multiples of the divisor. He explains the convenience of approximation by truncation in multiplication. Colson also devised an instrument (Counting Table) that calculated using signed digits. Eduard Selling advocated inverting the digits 1, 2, 3, 4, and 5 to indicate the negative sign. He also suggested snie, jes, jerd, reff, and niff as names to use vocally. Most of the other early sources used a bar over a digit to indicate a negative sign for it. Another German usage of signed-digits was described in 1902 in Klein's encyclopedia. Definition and properties Digit set Let be a finite set of numerical digits with cardinality (If , then the positional number system is trivial and only represents the trivial ring), with each digit denoted as for is known as the radix or number base. can be used for a signed-digit representation if it's associated with a unique function such that for all This function, is what rigorously and formally
https://en.wikipedia.org/wiki/List%20of%20Eclipse-based%20software
The Eclipse IDE platform can be extended by adding different plug-ins. Notable examples include: Acceleo, an open source code generator that uses EMF-based models to generate any textual language (Java, PHP, Python, etc.). Actifsource, a modeling and code generation workbench. Adobe ColdFusion Builder, the official Adobe IDE for ColdFusion. Adobe Flash Builder (formerly Adobe Flex Builder), an Adobe IDE based on Eclipse for building Flex applications for the Flash Platform and mobile platforms. ADT Eclipse plugin developed by Google for the Android SDK. AnyLogic, a simulation modeling tool developed by The AnyLogic Company. Appcelerator, a cross platform mobile development tool by Axway Appcelerator Aptana, Web IDE based on Eclipse Avaya Dialog Designer, a commercial IDE to build scripts for voice self-service applications. Bioclipse, a visual platform for chemo- and bioinformatics. BIRT Project, open source software project that provides reporting and business intelligence capabilities for rich client and web applications. Bonita Open Solution relies on Eclipse for the modeling of processes, implementing a BPMN and a Web form editors. Cantata IDE is a computer program for software testing at run time of C and C++ programs. CityEngine procedural based city generator. Code Composer Studio Texas Instruments' IDE for microcontroller development. CodeWarrior Freescale's IDE for microcontrollers, since Version 10 (C/C++/Assembly compilers). Compuware OptimalJ, a model-driven development environment for Java Coverity Static Analysis, which finds crash-causing defects and security vulnerabilities in code DBeaver, universal database manager and SQL client ECLAIR, a tool for automatic program analysis, verification, testing and transformation EasyEclipse, bundled distributions of the Eclipse IDE Elysium, a frontend for the LilyPond music-engraving program g-Eclipse, an integrated workbench framework to access the power of existing Grid infrastructures
https://en.wikipedia.org/wiki/Peptide%20mass%20fingerprinting
Peptide mass fingerprinting (PMF) (also known as protein fingerprinting) is an analytical technique for protein identification in which the unknown protein of interest is first cleaved into smaller peptides, whose absolute masses can be accurately measured with a mass spectrometer such as MALDI-TOF or ESI-TOF. The method was developed in 1993 by several groups independently. The peptide masses are compared to either a database containing known protein sequences or even the genome. This is achieved by using computer programs that translate the known genome of the organism into proteins, then theoretically cut the proteins into peptides, and calculate the absolute masses of the peptides from each protein. They then compare the masses of the peptides of the unknown protein to the theoretical peptide masses of each protein encoded in the genome. The results are statistically analyzed to find the best match. The advantage of this method is that only the masses of the peptides have to be known. A disadvantage is that the protein sequence has to be present in the database of interest. Additionally most PMF algorithms assume that the peptides come from a single protein. The presence of a mixture can significantly complicate the analysis and potentially compromise the results. Typical for the PMF-based protein identification is the requirement for an isolated protein. Mixtures exceeding a number of 2–3 proteins typically require the additional use of MS/MS-based protein identification to achieve sufficient specificity of identification. Therefore, typical PMF samples are isolated proteins from two-dimensional gel electrophoresis (2D gels) or isolated SDS-PAGE bands. Additional analyses by MS/MS can either be direct, e.g., MALDI-TOF/TOF analysis or downstream nanoLC-ESI-MS/MS analysis of gel spot eluates. Origins Due to the long, tedious process of analyzing proteins, peptide mass fingerprinting was developed. Edman degradation was used in protein analysis, and it required
https://en.wikipedia.org/wiki/Free%20boundary%20condition
In image processing, the free boundary condition is the convention used when applying a convolution kernel to a digital image in which pixel locations that lie outside the image boundaries are interpreted as having a value of zero. The question of what value to assign out-of-bounds pixels may arise, for instance, when applying a 3×3 kernel to the corner pixel in an image. Image processing
https://en.wikipedia.org/wiki/Ulster%20Museum
The Ulster Museum, located in the Botanic Gardens in Belfast, has around 8,000 square metres (90,000 sq. ft.) of public display space, featuring material from the collections of fine art and applied art, archaeology, ethnography, treasures from the Spanish Armada, local history, numismatics, industrial archaeology, botany, zoology and geology. It is the largest museum in Northern Ireland, and one of the components of National Museums Northern Ireland. History The Ulster Museum was founded as the Belfast Natural History Society in 1821 and began exhibiting in 1833. It has included an art gallery since 1890. Originally called the Belfast Municipal Museum and Art Gallery, in 1929, it moved to its present location in Stranmillis. The new building was designed by James Cumming Wynne. In 1962, courtesy of the Museum Act (Northern Ireland) 1961, it was renamed as the Ulster Museum and was formally recognised as a national museum. A major extension constructed by McLaughlin & Harvey Ltd to designs by Francis Pym who won the 1964 competition was opened in 1972 and Pym's only completed work. It was published in several magazines and was until alteration the most important example of Brutalism in Northern Ireland. It was praised by David Evans for the "almost barbaric power of its great cubic projections and cantilevers brooding over the conifers of the botanic gardens like a mastodon". Since the 1940s the Ulster Museum has built up a good collection of art by modern Irish, and particularly Ulster-based artists. In 1998, the Ulster Museum merged with the Ulster Folk and Transport Museum and the Ulster-American Folk Park to form the National Museums and Galleries of Northern Ireland. In July 2005, a £17m refurbishment of the museum was announced, with grants from the Heritage Lottery Fund and the Department of Culture, Arts and Leisure (DCAL, usually pronounced as 'Dee-Kal'). In October 2006 the museum closed its doors until 2009, to allow for the work. Illustrations of
https://en.wikipedia.org/wiki/Certified%20broadcast%20technologist
Certified Broadcast Technologist (CBT) is a title granted to an individual that successfully meets the experience or examination requirements of the certification. The certification is regulated by the Society of Broadcast Engineers (SBE). The CBT title is protected by copyright laws. Individuals who use the title without consent from the Society of Broadcast Engineers could face legal action. The SBE certifications were created to recognize individuals who practice in career fields which are not regulated by state licensing or Professional Engineering programs. Marine Radio and radar systems still require a Federal Communications Commission (FCC) license apart from an SBE certification. Broadcast Engineering is regulated at the national level and not by individual states. See also List of post-nominal letters External links Certified Broadcast Technologist (CBT) Requirements & Application SBE official website Broadcast engineering Professional titles and certifications
https://en.wikipedia.org/wiki/Infinitesimal%20strain%20theory
In continuum mechanics, the infinitesimal strain theory is a mathematical approach to the description of the deformation of a solid body in which the displacements of the material particles are assumed to be much smaller (indeed, infinitesimally smaller) than any relevant dimension of the body; so that its geometry and the constitutive properties of the material (such as density and stiffness) at each point of space can be assumed to be unchanged by the deformation. With this assumption, the equations of continuum mechanics are considerably simplified. This approach may also be called small deformation theory, small displacement theory, or small displacement-gradient theory. It is contrasted with the finite strain theory where the opposite assumption is made. The infinitesimal strain theory is commonly adopted in civil and mechanical engineering for the stress analysis of structures built from relatively stiff elastic materials like concrete and steel, since a common goal in the design of such structures is to minimize their deformation under typical loads. However, this approximation demands caution in the case of thin flexible bodies, such as rods, plates, and shells which are susceptible to significant rotations, thus making the results unreliable. Infinitesimal strain tensor For infinitesimal deformations of a continuum body, in which the displacement gradient tensor (2nd order tensor) is small compared to unity, i.e. , it is possible to perform a geometric linearization of any one of the (infinitely many possible) finite strain tensors used in finite strain theory, e.g. the Lagrangian finite strain tensor , and the Eulerian finite strain tensor . In such a linearization, the non-linear or second-order terms of the finite strain tensor are neglected. Thus we have or and or This linearization implies that the Lagrangian description and the Eulerian description are approximately the same as there is little difference in the material and spatial coordi
https://en.wikipedia.org/wiki/X17%20particle
The X17 particle is a hypothetical subatomic particle proposed by Attila Krasznahorkay and his colleagues to explain certain anomalous measurement results; these anomalous measurements are known as ATOMKI anomaly or beryllium anomaly or X17 anomaly. The particle has been proposed to explain wide angles observed in the trajectory paths of particles produced during a nuclear transition of beryllium-8 atoms and in stable helium atoms. The X17 particle could be the force carrier for a postulated fifth force, possibly connected with dark matter, and has been described as a protophobic (i.e., ignoring protons) vector boson with a mass near . History In 2015, Krasznahorkay and his colleagues at ATOMKI, the Hungarian Institute for Nuclear Research, posited the existence of a new, light boson with a mass of about (i.e., 34 times heavier than the electron). In an effort to find a dark photon, the team fired protons at thin targets of lithium-7, which created unstable beryllium-8 nuclei that then decayed and produced pairs of electrons and positrons. Excess decays were observed at an opening angle of 140° between the and particles and a combined energy of approximately . This indicated that a small fraction of beryllium-8 might shed its excess energy in the form of a new particle. The result was successfully repeated by the team. Feng et al. (2016) proposed that a "protophobic" X boson, with a mass of , suppressed couplings to protons relative to neutrons and electrons at femtometer range, could explain the data. The force may explain the  − 2 muon anomaly and provide a dark matter candidate. , several research experiments are underway to attempt to validate or refute these results. Krasznahorkay (2019) posted a preprint announcing that he and his team at ATOMKI had successfully observed the same anomalies in the decay of stable helium atoms as had been observed in beryllium-8, strengthening the case for the existence of the X17 particle. This was covered in science jou
https://en.wikipedia.org/wiki/Robotic%20sensing
Robotic sensing is a subarea of robotics science intended to provide sensing capabilities to robots. Robotic sensing provides robots with the ability to sense their environments and is typically used as feedback to enable robots to adjust their behavior based on sensed input. Robot sensing includes the ability to see, touch, hear and move and associated algorithms to process and make use of environmental feedback and sensory data. Robot sensing is important in applications such as vehicular automation, robotic prosthetics, and for industrial, medical, entertainment and educational robots. Vision Method Visual sensing systems can be based on a variety of technologies and methods including the use of camera, sonar, laser and radio frequency identification (RFID) technology. All four methods aim for three procedures—sensation, estimation, and matching. Image processing Image quality is important in applications that require excellent robotic vision. Algorithms based on wavelet transform that are used for fusing images of different spectra and different foci result in improved image quality. Robots can gather more accurate information from the resulting improved image. Usage Visual sensors help robots to identify the surrounding environment and take appropriate action. Robots analyze the image of the immediate environment based on data input from the visual sensor. The result is compared to the ideal, intermediate or end image, so that appropriate movement or action can be determined to reach the intermediate or final goal. Touch Robot skin Types and examples Examples of the current state of progress in the field of robot skins as of mid-2022 are a robotic finger covered in a type of manufactured living human skin, an electronic skin giving biological skin-like haptic sensations and touch/pain-sensitivity to a robotic hand, a system of an electronic skin and a human-machine interface that can enable remote sensed tactile perception, and wearable or robotic sensin
https://en.wikipedia.org/wiki/HMA%20%28VPN%29
HMA (formerly HideMyAss!) is a VPN service founded in 2005 in the United Kingdom. It has been a subsidiary of the Czech cybersecurity company Avast since 2016. History HMA was created in 2005 in Norfolk, England by Jack Cator. At the time, Cator was sixteen years-old. He created HMA in order to circumvent restrictions his school had on accessing games or music from their network. According to Cator, the first HMA service was created in just a few hours using open-source code. The first product was a free proxy website where users typed in a URL and it delivered the website in the user's web browser. Cator promoted the tool in online forums and it was featured on the front page of digg. After attracting more than one thousand users, Cator incorporated ads. HMA did not take any venture capital funding. It generated about $1,000 - $2,000 per month while the founder went to college to pursue a degree in computer science. In 2009, Cator dropped out of college to focus on HMA and added a paid VPN service. Most early HMA employees were freelancers found on oDesk. In 2012, one of the freelancers set up a competing business. HMA responded by hiring its contractors as full-time employees and establishing physical offices in London. In 2012, the United Kingdom's government sent HMA a court order demanding it provide information about Cody Andrew Kretsinger's use of HMA's service to hack Sony as a member of the LulzSec hacking group. HMA provided the information to authorities. HMA said it was a violation of the company's terms of use to use its software for illegal activities. In 2013, HMA added software to anonymize internet traffic from mobile devices was first added in 2013. In 2014, the company introduced HideMyPhone! service, which allowed mobile phone users to make their calls appear to come from a different location. By 2014, the service had 10 million users and 215,000 paying subscribers of its VPN service. It made £11 million in revenue that year. HMA had 100 sta
https://en.wikipedia.org/wiki/ATX
ATX (Advanced Technology Extended) is a motherboard and power supply configuration specification, patented by David Dent in 1995 at Intel, to improve on previous de facto standards like the AT design. It was the first major change in desktop computer enclosure, motherboard and power supply design in many years, improving standardization and interchangeability of parts. The specification defines the dimensions; the mounting points; the I/O panel; and the power and connector interfaces among a computer case, a motherboard, and a power supply. Overview ATX is the most common motherboard design. Other standards for smaller boards (including microATX, FlexATX, nano-ITX, and mini-ITX) usually keep the basic rear layout but reduce the size of the board and the number of expansion slots. Dimensions of a full-size ATX board are , which allows many ATX chassis to accept microATX boards. The ATX specifications were released by Intel in 1995 and have been revised numerous times since. The most recent ATX motherboard specification is version 2.2. The most recent ATX12V power supply unit specification is ATX 3.0 released in February 2022. EATX (Extended ATX) is a bigger version of the ATX motherboard with dimensions. While some dual CPU socket motherboards have been implemented in ATX, the extra size of EATX makes it the typical form factor for dual socket systems, and with sockets that support four or eight memory channels, for single socket systems with a large number of memory slots. In 2004, Intel announced the BTX (Balanced Technology eXtended) standard, intended as a replacement for ATX. While some manufacturers adopted the new standard, Intel discontinued any future development of BTX in 2006. , the ATX design still remains the de facto standard for personal computers. Connectors On the back of the computer case, some major changes were made to the AT standard. Originally AT style cases had only a keyboard connector and expansion slots for add-on card backplates
https://en.wikipedia.org/wiki/Rake%20%28cellular%20automaton%29
A rake, in the lexicon of cellular automata, is a type of puffer train, which is an automaton that leaves behind a trail of debris. In the case of a rake, however, the debris left behind is a stream of spaceships, which are automata that "travel" by looping through a short series of iterations and end up in a new location after each cycle returns to the original configuration. In Conway's Game of Life, the discovery of rakes was one of the key components needed to form the breeder, the first known pattern in Life in which the number of live cells exhibits quadratic growth. A breeder is formed by arranging several rakes so that the gliders—the smallest possible spaceships—they generate interact to form a sequence of glider guns, patterns which emit gliders. The emitted gliders fill a growing triangle of the plane of the game. More generally, when a rake exists for a cellular automaton rule (a mathematical function defining the next iteration to be derived from a particular configuration of live and dead cells), one can often construct puffers which leave trails of many other kinds of objects, by colliding the streams of spaceships emitted by multiple rakes moving in parallel. As David Bell writes: The first rake to be discovered, in the early 1970s, was the "space rake", which moves with speed c/2 (or one unit every two steps), emitting a glider every twenty steps. For Life, rakes are now known that move orthogonally with speeds c/2, c/3, c/4, c/5, 2c/5, 2c/7, c/10 and 17c/45, and diagonally with speeds c/4 and c/12, with many different periods. Rakes are also known for some other cellular automata, including Highlife, Day & Night, and Seeds. Gotts (1980) shows that the space rake in Life can be formed by a "standard collision sequence" in which a single glider interacts with a widely separated set of 3-cell initial seeds (blinkers and blocks). As a consequence, he finds lower bounds on the probability that these patterns form in any sufficiently sparse and suf
https://en.wikipedia.org/wiki/Acoustic%20Doppler%20current%20profiler
An acoustic doppler current profiler (ADCP) is a hydroacoustic current meter similar to a sonar, used to measure water current velocities over a depth range using the Doppler effect of sound waves scattered back from particles within the water column. The term ADCP is a generic term for all acoustic current profilers, although the abbreviation originates from an instrument series introduced by RD Instruments in the 1980s. The working frequencies range of ADCPs range from 38 kHz to several megahertz. A similar device is a SODAR, which works in the air and uses the same principles for wind speed profiling. Working principle ADCPs contain piezoelectric transducers to transmit and receive sound signals. The traveling time of sound waves gives an estimate of the distance. The frequency shift of the echo is proportional to the water velocity along the acoustic path. To measure 3D velocities, at least three beams are required. In rivers, only the 2D velocity is relevant and ADCPs typically have two beams. In recent years, more functionality has been added to ADCPs (notably wave and turbulence measurements) and systems can be found with 2,3,4,5 or even 9 beams. Further components of an ADCP are an electronic amplifier, a receiver, a clock to measure the traveling time, a temperature sensor, a compass to know the heading, and a pitch/roll sensor to know the orientation. An analog-to-digital converter and a digital signal processor are required to sample the returning signal in order to determine the Doppler shift. A temperature sensor is used to estimate the sound velocity at the instrument position using the seawater equation of state, and uses this to estimate scale the frequency shift to water velocities. This procedure assumes that the salinity has a preconfigured constant value. Finally, the results are saved to internal memory or output online to an external display software. Processing methods Three common methods are used to calculate the Doppler shift and
https://en.wikipedia.org/wiki/Vkernel
A virtual kernel architecture (vkernel) is an operating system virtualisation paradigm where kernel code can be compiled to run in the user space, for example, to ease debugging of various kernel-level components, in addition to general-purpose virtualisation and compartmentalisation of system resources. It is used by DragonFly BSD in its vkernel implementation since DragonFly 1.7, having been first revealed in , and first released in the stable branch with DragonFly 1.8 in . The long-term goal, in addition to easing kernel development, is to make it easier to support internet-connected computer clusters without compromising local security. Similar concepts exist in other operating systems as well; in Linux, a similar virtualisation concept is known as user-mode Linux; whereas in NetBSD since the summer of 2007, it has been the initial focus of the rump kernel infrastructure. The virtual kernel concept is nearly the exact opposite of the unikernel concept — with vkernel, kernel components get to run in userspace to ease kernel development and debugging, supported by a regular operating system kernel; whereas with a unikernel, userspace-level components get to run directly in kernel space for extra performance, supported by baremetal hardware or a hardware virtualisation stack. However, both vkernels and unikernels can be used for similar tasks as well, for example, to self-contain software to a virtualised environment with low overhead. In fact, NetBSD's rump kernel, originally having a focus of running kernel components in userspace, has since shifted into the unikernel space as well (going after the anykernel moniker for supporting both paradigms). The vkernel concept is different from a FreeBSD jail in that a jail is only meant for resource isolation, and cannot be used to develop and test new kernel functionality in the userland, because each jail is sharing the same kernel. (DragonFly, however, still has FreeBSD jail support as well.) In DragonFly, the v
https://en.wikipedia.org/wiki/Mass%20action%20law%20%28electronics%29
In electronics and semiconductor physics, the law of mass action relates the concentrations of free electrons and electron holes under thermal equilibrium. It states that, under thermal equilibrium, the product of the free electron concentration and the free hole concentration is equal to a constant square of intrinsic carrier concentration . The intrinsic carrier concentration is a function of temperature. The equation for the mass action law for semiconductors is: Carrier concentrations In semiconductors, free electrons and holes are the carriers that provide conduction. For cases where the number of carriers are much less than the number of band states, the carrier concentrations can be approximated by using Boltzmann statistics, giving the results below. Electron concentration The free-electron concentration n can be approximated by where Ec is the energy of the conduction band, EF is the energy of the Fermi level, kB is the Boltzmann constant, T is the absolute temperature in kelvins, Nc is the effective density of states at the conduction band edge given by , with m*e being the electron effective mass and h being Planck's constant. Hole concentration The free-hole concentration p is given by a similar formula where EF is the energy of the Fermi level, Ev is the energy of the valence band, kB is the Boltzmann constant, T is the absolute temperature in kelvins, Nv is the effective density of states at the valence band edge given by , with m*h being the hole effective mass and h Planck's constant. Mass action law Using the carrier concentration equations given above, the mass action law can be stated as where Eg is the band gap energy given by Eg = Ec − Ev. The above equation holds true even for lightly doped extrinsic semiconductors as the product is independent of doping concentration. See also Law of mass action