source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/SS%20blood%20group%20tattoo
SS blood group tattoos () were worn by members of the Waffen-SS in Nazi Germany during World War II to identify the individual's blood type. After the war, the tattoo was taken to be prima facie evidence of being part of the Waffen-SS, leading to potential arrest and prosecution. Description and purpose The SS blood group tattoo was applied, in theory, to all Waffen-SS members, except members of the British Free Corps. It was a small black ink tattoo located on the underside of the left arm. It generally measured around long and was placed roughly above the elbow. The tattoo consisted of the soldier's blood type letter, either A, B, AB or O. The discovery of the Rh D antigen had been made in 1937, but was not fully understood during World War II, so was not implemented. In the early part of the war, tattoos were printed in Fraktur, while later on they were printed in Latin-style. The purpose of the tattoo was to identify a soldier's blood type in case a blood transfusion was needed while unconscious, or his Erkennungsmarke (dog tag) or Soldbuch (pay book) were missing. The tattoo was generally applied by the unit's Sanitäter (medic) in basic training but could have been applied by anyone assigned to do it at any time during his term of service. Usage Not all Waffen-SS men had the tattoo, particularly those who had transferred from other branches of the military to the Waffen-SS, or those who transferred from the Allgemeine SS, the "General" or non-military SS. Some non-SS men also had the tattoo: if a member of a branch of the Wehrmacht was treated in an SS hospital, he would often have the tattoo applied. Although the tattoo was widely used in the early war years, over the course of the war it was gradually applied to fewer and fewer soldiers, and towards the end of the war, having the tattoo was more the exception than the rule. The application of the tattoo to foreign volunteers was apparently an issue of contention with some, such as the British Free Corp
https://en.wikipedia.org/wiki/Beta%20attenuation%20monitoring
Beta attenuation monitoring (BAM) is a widely used air monitoring technique employing the absorption of beta radiation by solid particles extracted from air flow. This technique allows for the detection of PM10 and PM2.5, which are monitored by most air pollution regulatory agencies. The main principle is based on a kind of Bouguer (Lambert–Beer) law: the amount by which the flow of beta radiation (electrons) is attenuated by a solid matter is exponentially dependent on its mass and not on any other feature (such as density, chemical composition or some optical or electrical properties) of this matter. So, the air is drawn from outside of the detector through an "infinite" (cycling) ribbon made from some filtering material so that the particles are collected on it. There are two sources of beta radiation placed one before and one after the region where air flow passes through the ribbon leaving particles on it; and there are also two detectors on the opposite side of the ribbon, facing the detectors. The sources' intensity and detectors' sensitivity being the same (or corrected with appropriate calibration lookup table), the intensity of beta rays detected by one of detectors is compared to that of the other. Thus one can deduce how much mass has the ribbon acquired upon being exposed to air flow; knowing the drain velocity, actual particle mass concentration in air could be assessed. The radiation source can be a gas chamber, filled with 86Kr gas, or a pieces of 14C-rich polymer plastic, such as PMMA. Detector is simply a Geiger–Mueller counter. The particulate matter content measured is affected by the moisture content in the air, unfortunately. To discriminate between particle of different sizes (e. g., between PM10 and PM2.5), some preliminary separation could be accomplished, for example, by cyclone battery. A similar method exists, where instead of beta particle flow an X-ray Fluorescence Spectroscopic monitoring is applied on the either side of air flow co
https://en.wikipedia.org/wiki/Stationary%20phase%20approximation
In mathematics, the stationary phase approximation is a basic principle of asymptotic analysis, applying to functions given by integration against a rapidly-varying complex exponential. This method originates from the 19th century, and is due to George Gabriel Stokes and Lord Kelvin. It is closely related to Laplace's method and the method of steepest descent, but Laplace's contribution precedes the others. Basics The main idea of stationary phase methods relies on the cancellation of sinusoids with rapidly varying phase. If many sinusoids have the same phase and they are added together, they will add constructively. If, however, these same sinusoids have phases which change rapidly as the frequency changes, they will add incoherently, varying between constructive and destructive addition at different times. Formula Letting denote the set of critical points of the function (i.e. points where ), under the assumption that is either compactly supported or has exponential decay, and that all critical points are nondegenerate (i.e. for ) we have the following asymptotic formula, as : Here denotes the Hessian of , and denotes the signature of the Hessian, i.e. the number of positive eigenvalues minus the number of negative eigenvalues. For , this reduces to: In this case the assumptions on reduce to all the critical points being non-degenerate. This is just the Wick-rotated version of the formula for the method of steepest descent. An example Consider a function . The phase term in this function, , is stationary when or equivalently, . Solutions to this equation yield dominant frequencies for some and . If we expand as a Taylor series about and neglect terms of order higher than , we have where denotes the second derivative of . When is relatively large, even a small difference will generate rapid oscillations within the integral, leading to cancellation. Therefore we can extend the limits of integration beyond the limit for a Taylor expan
https://en.wikipedia.org/wiki/Hollow%20fiber%20bioreactor
A Hollow fiber bioreactor is a 3 dimensional cell culturing system based on hollow fibers, which are small, semi-permeable capillary membranes arranged in parallel array with a typical molecular weight cut-off (MWCO) range of 10-30 kDa. These hollow fiber membranes are often bundled and housed within tubular polycarbonate shells to create hollow fiber bioreactor cartridges. Within the cartridges, which are also fitted with inlet and outlet ports, are two compartments: the intracapillary (IC) space within the hollow fibers, and the extracapillary (EC) space surrounding the hollow fibers. Cells are seeded into the EC space of the hollow fiber bioreactor and expand there. Cell culture medium is pumped through the IC space and delivers oxygen and nutrients to the cells via hollow fiber membrane perfusion. As the cells expand, their waste products and CO2 also perfuse the hollow fiber membranes and are carried away by the pumping of medium through the IC space. As waste products build up due to increased cell mass, the rate of medium flow can also be increased so that cell growth is not inhibited by waste product toxicity. Because thousands of hollow fibers may be packed into a single hollow fiber bioreactor, they increase the surface area of the cartridge considerably. As a result, cells can fill up the EC space to densities >108 cells/ml. However, the cartridge itself takes up a very small volume (oftentimes the volume of a 12-oz soda can). The fact that hollow fiber bioreactors are very small and yet enable incredibly high cell densities has led to their development for both research and commercial applications, including monoclonal antibody and influenza vaccine production. Likewise, because hollow fiber bioreactors use up significantly less medium and growth factors than traditional cell culture methods such as stirred-tank bioreactors, they offer a significant cost savings. Finally, hollow fiber bioreactors are sold as single-use disposables, resulting in signifi
https://en.wikipedia.org/wiki/OVPsim
OVPsim is a multiprocessor platform emulator (often called a full-system simulator) used to run unchanged production binaries of the target hardware. It has public APIs allowing users to create their own processor, peripheral and platform models. Various models are available as open source. OVPsim is a key component of the Open Virtual Platforms initiative (OVP), an organization created to promote the use of open virtual platforms for embedded software development. OVPsim requires OVP registration to download. Licensing OVPsim is developed and maintained by Imperas. The core simulation platform is proprietary software; it is available free of charge for non-commercial usage. Commercial usage requires a low-cost license from Imperas to cover maintenance. Various processor, peripheral and platform models are available as free software under the Apache License version 2.0. Details There are three main components of OVP: open-source models, fast OVPsim simulator, and modeling APIs. These components are designed to make it easy to assemble multi-core heterogeneous or homogeneous platforms with complex memory hierarchies, cache systems and layers of embedded software that can run at hundreds of MIPS on standard desktop PCs. OVPSim is considered instruction accurate, but not cycle-accurate. There are many examples of components, and complete virtual platforms that can boot a Linux Kernel in under 5 seconds at OVP Homepage. Open source models Within OVP there are several different model categories. These models are provided as both pre-compiled object code, and as in some cases, source files. OVPsim no longer supplies source code for the ARM and MIPS processor models. Currently there are processor models of ARM (processors using the ARMv4, ARMv5, ARMv6, ARMv7, ARMv8 instruction sets) up to the ARM Cortex-A72MPx4 (and including multi-cluster ARMv8 models with GICv3), Imagination MIPS (processors using MIPS32, MIPS64, microMIPS, nanoMIPS and MIPS R6 instruction sets)
https://en.wikipedia.org/wiki/Discovery%20of%20the%20nonmetals
Most nonmetallic elements were discovered after the freezing of mercury in 1759 by the German-Russian physicist and the Russian polymath Mikhail Lomonosov. Before then, carbon, sulfur and antimony were known in antiquity. Arsenic and phosphorus were discovered in the middle ages and in the Renaissance, respectively. In the ensuing century and a half, from 1766 to 1895, all the remaining nonmetallic elements, bar radon had been isolated. The latter three were discovered in 1898. Isolation by periods Antiquity Carbon (C) and sulfur (S) were known in antiquity. The earliest known use of charcoal dates to around 3750 BCE. The Egyptians and Sumerians employed it for the reduction of copper, zinc, and tin ores in the manufacture of bronze. Diamonds were probably known from as early as 2500 BCE. The first true chemical analyses were made in the 18th century; Antoine Lavoisier recognized carbon as an element in 1789. Sulfur usage dates from before 2500 BCE; it was also recognized as an element by Lavoisier, in 1777. 17th century Phosphorus (P) was prepared from urine, by Hennig Brand, in 1669. 18th century Henry Cavendish, in 1766, was the first to distinguish hydrogen (H) from other gases, although Paracelsus around 1500, Robert Boyle (1670), and Joseph Priestley (?) had observed its production by reacting strong acids with metals. Lavoisier named it in 1793. Carl Wilhelm Scheele obtained oxygen (O) by heating mercuric oxide (HgO) and nitrates in 1771, but did not publish his findings until 1777. Priestley also prepared this new "air" by 1774, but only Lavoisier recognized it as a true element; he named it in 1777. Ernest Rutherford discovered nitrogen (N) while he was studying at the University of Edinburgh. He showed that the air in which animals breathed, after removal of exhaled carbon dioxide (CO2), was no longer able to burn a candle. Scheele, Cavendish, and Priestley also studied this element at about the same time; Lavoisier named it in 1775 or 1776.
https://en.wikipedia.org/wiki/Wishart%20distribution
In statistics, the Wishart distribution is a generalization to multiple dimensions of the gamma distribution. It is named in honor of John Wishart, who first formulated the distribution in 1928. Other names include Wishart ensemble (in random matrix theory, probability distributions over matrices are usually called "ensembles"), or Wishart–Laguerre ensemble (since its eigenvalue distribution involve Laguerre polynomials), or LOE, LUE, LSE (in analogy with GOE, GUE, GSE). It is a family of probability distributions defined over symmetric, positive-definite random matrices (i.e. matrix-valued random variables). These distributions are of great importance in the estimation of covariance matrices in multivariate statistics. In Bayesian statistics, the Wishart distribution is the conjugate prior of the inverse covariance-matrix of a multivariate-normal random-vector. Definition Suppose is a matrix, each column of which is independently drawn from a -variate normal distribution with zero mean: Then the Wishart distribution is the probability distribution of the random matrix known as the scatter matrix. One indicates that has that probability distribution by writing The positive integer is the number of degrees of freedom. Sometimes this is written . For the matrix is invertible with probability if is invertible. If then this distribution is a chi-squared distribution with degrees of freedom. Occurrence The Wishart distribution arises as the distribution of the sample covariance matrix for a sample from a multivariate normal distribution. It occurs frequently in likelihood-ratio tests in multivariate statistical analysis. It also arises in the spectral theory of random matrices and in multidimensional Bayesian analysis. It is also encountered in wireless communications, while analyzing the performance of Rayleigh fading MIMO wireless channels . Probability density function The Wishart distribution can be characterized by its probability density fu
https://en.wikipedia.org/wiki/CPLEX
IBM ILOG CPLEX Optimization Studio (often informally referred to simply as CPLEX) is an optimization software package. History The CPLEX Optimizer was named for the simplex method implemented in the C programming language, although today it also supports other types of mathematical optimization and offers interfaces other than C. It was originally developed by Robert E. Bixby and sold commercially from 1988 by CPLEX Optimization Inc. This was acquired by ILOG in 1997 and ILOG was subsequently acquired by IBM in January 2009. CPLEX continues to be actively developed by IBM. Features The IBM ILOG CPLEX Optimizer solves integer programming problems, very large linear programming problems using either primal or dual variants of the simplex method or the barrier interior point method, convex and non-convex quadratic programming problems, and convex quadratically constrained problems (solved via second-order cone programming, or SOCP). The CPLEX Optimizer has a modeling layer called Concert that provides interfaces to the C++, C#, and Java languages. There is a Python language interface based on the C interface. Finally, a stand-alone Interactive Optimizer executable is provided for debugging and other purposes. The CPLEX Optimizer is accessible through independent modeling systems such as AIMMS, AMPL, GAMS, OptimJ and TOMLAB. In addition to that AMPL provides an interface to the CPLEX CP Optimizer. The full IBM ILOG CPLEX Optimization Studio consists of the CPLEX Optimizer for mathematical programming, the CP Optimizer for constraint programming, the Optimization Programming Language (OPL), and a tightly integrated IDE. Release history Prior to IBM acquiring ILOG, the CPLEX team published a release history of CPLEX. See also FICO Xpress GLPK Gurobi Optimizer SCIP
https://en.wikipedia.org/wiki/Occipital%20lymph%20nodes
The occipital lymph nodes, one to three in number, are located on the back of the head close to the margin of the trapezius and resting on the insertion of the . Their afferent vessels drain the occipital region of the scalp, while their efferents pass to the superior deep cervical glands. Additional images Etymology The word occipital comes from the ("the back of the head").
https://en.wikipedia.org/wiki/Myclobutanil
Myclobutanil is a triazole chemical used as a fungicide. It is a steroid demethylation inhibitor, specifically inhibiting ergosterol biosynthesis. Ergosterol is a critical component of fungal cell membranes. Stereoisomerism Safety The Safety Data Sheet indicates the following hazards: Suspected of damaging fertility or the unborn child. Toxic to aquatic life with long lasting effects. The first hazard has caused this chemical to be placed on the 1986 California Proposition 65 toxics list. When heated, myclobutanil decomposes to produce corrosive and/or toxic fumes, including carbon monoxide, carbon dioxide, hydrogen chloride, hydrogen cyanide, and nitrogen oxides. Banned for cannabis cultivation Myclobutanil is banned in Canada, Colorado, Washington, Oregon, and Oklahoma for the production of medical and recreational cannabis. In 2014, a Canadian news investigation by The Globe and Mail reported the discovery of myclobutanil in medical cannabis produced by at least one government licensed grower. In September 2019, NBC News commissioned CannaSafe to test THC cartridges for heavy metals, pesticides, and residual solvents like Vitamin E; pesticides, including myclobutanil, was found in products from unlicensed dealers. In Michigan, the current state action limit for myclobutanil is 200 ppb in cannabis products.
https://en.wikipedia.org/wiki/QNS
QNS is a clinical laboratory abbreviation for quantity not sufficient. This indicates that either: There is not enough specimen for the lab tests ordered to be performed. In the case of Vacutainers or other tubes with pre-added anticoagulant, the amount of blood invacuated into the tube at the time of phlebotomy was insufficient to attain the correct blood:anticoagulant ratio. This can cause false results in assays such as coagulation assays (causing falsely increased clotting times) or blood cell differentials (causing a false increase in poikilocytes, particularly burr cells.) In either case, the most common and feasible way to correct the problem is to simply recollect the specimen. Quantity not sufficient implies that the final volume of diluent is not sufficient for molecular testing. Medical diagnosis
https://en.wikipedia.org/wiki/Apioperdon
Apioperdon pyriforme commonly known as the pear-shaped puffball or stump puffball, is a saprobic fungus present throughout much of the world. Emerging in autumn, this puffball is common and abundant on decaying logs of both deciduous and coniferous wood. It is considered a choice edible when still immature and the inner flesh is white. It is often called Lycoperdon pyriforme, but was transferred to Apioperdon in 2017 based on phylogenetic and morphological differences. It is the only species in the genus. Taxonomy The fungus was first described in the scientific literature by Jacob Christian Schaeffer in 1774. In 2001, DNA evidence compiled by Dirk Krüger and a number of other mycologists suggested that the genus Lycoperdon was polyphyletic, with the pear-shaped puffball differing most significantly from the group. This finding was supported by several morphological differences including the presence of rhizomorphs and its preference for wood. A 2003 publication of further research moved the puffball to the genus Morganella with the specific epithet's ending changed for agreement. In 2008, however, Larsson and Jeppson revisited the phylogeny of the Lycoperdaceae, with a broader sampling of species, and retained the taxon in Lycoperdon. It was renamed Apioperdon pyriforme in 2017. The specific epithet pyriforme is Latin for "pear-shaped". Description The fruit body of the pear-shaped puffball measures 1.5 to 4.5 cm in width by 2 to 4.5 cm in height. They are often pear-shaped as the name suggests, but they may also be nearly spherical. When very young they are covered in small white spines that typically fall off before maturity. A small developing pore may be visible at the top, while the sterile base of the mushroom is small and appears to be pinched in. Colour ranges from nearly white to yellowish brown with the darker shades developing with age, though the base remains white. The central pore ruptures at late maturity to allow the wind and rain to disperse the
https://en.wikipedia.org/wiki/Dewar%E2%80%93Chatt%E2%80%93Duncanson%20model
The Dewar–Chatt–Duncanson model is a model in organometallic chemistry that explains the chemical bonding in transition metal alkene complexes. The model is named after Michael J. S. Dewar, Joseph Chatt and L. A. Duncanson. The alkene donates electron density into a π-acid metal d-orbital from a π-symmetry bonding orbital between the carbon atoms. The metal donates electrons back from a (different) filled d-orbital into the empty π* antibonding orbital. Both of these effects tend to reduce the carbon-carbon bond order, leading to an elongated C−C distance and a lowering of its vibrational frequency. In Zeise's salt K[PtCl3(C2H4)].H2O the C−C bond length has increased to 134 picometres from 133 pm for ethylene. In the nickel compound Ni(C2H4)(PPh3)2 the value is 143 pm. The interaction also causes carbon atoms to "rehybridise" from sp2 towards sp3, which is indicated by the bending of the hydrogen atoms on the ethylene back away from the metal. In silico calculations show that 75% of the binding energy is derived from the forward donation and 25% from backdonation. This model is a specific manifestation of the more general π backbonding model.
https://en.wikipedia.org/wiki/Celery%20%28software%29
Celery is an open source asynchronous task queue or job queue which is based on distributed message passing. While it supports scheduling, its focus is on operations in real time. Overview The execution units, called tasks, are executed concurrently on one or more worker nodes using multiprocessing, or . Tasks can execute asynchronously (in the background) or synchronously (wait until ready). Celery is used in production systems, for services such as Instagram, to process millions of tasks every day. Technology Celery is written in Python, but the protocol can be implemented in any language. It can also operate with other languages using webhooks. There is also a Ruby-Client called RCelery, a PHP client, a Go client, a Rust client, and a Node.js client. The recommended message brokers are RabbitMQ or Redis. In comparison with RabbitMQ, Redis represents a good start. However, if there is a valid reason and Redis is no longer meeting the requirements for the project, it is simple to switch to RabbitMQ. Additionally, MongoDB, Amazon SQS, CouchDB, IronMQ, and databases (using SQLAlchemy or the Django ORM) are supported in status experimental. See also Advanced Message Queuing Protocol
https://en.wikipedia.org/wiki/Detlef%20Laugwitz
Detlef Laugwitz (1932–2000) was a German mathematician and historian, who worked in differential geometry, history of mathematics, functional analysis, and non-standard analysis. Biography He was born on 11 May 1932 in Breslau, Germany. Starting in 1949, he studied mathematics, physics, and philosophy at the Georg-August-University at Göttingen, where he received his doctorate in 1954. Until 1956 he worked in the Mathematical Research Institute of Oberwolfach. In 1958 he became a lecturer at the Technical University of Munich, where he obtained his Habilitation. In 1958 he moved to the Technical University of Darmstadt, where in 1962 he became a professor, and remained until his retirement. From 1976 to 1984 he was a visiting professor at Caltech. Work Laugwitz worked in differential geometry of infinite dimensional vector spaces (his dissertation) and in Finsler geometry. In 1958 he and Curt Schmieden developed their own approach to infinitesimals through field extensions, independently of Abraham Robinson. They described this as "infinitesimal mathematics" and leading back to the historical roots in Leibniz. In 1996 he published the standard biography of Bernhard Riemann. Notes Publications . 20th-century German mathematicians German historians of mathematics Differential geometers Technical University of Munich alumni Academic staff of the Technical University of Munich 1932 births 2000 deaths 20th-century German historians Academic staff of Technische Universität Darmstadt
https://en.wikipedia.org/wiki/List%20of%20the%20bryophytes%20of%20Canada
Parent page: Flora of Canada This is a listing of the non-vascular plants of Canada, and includes the mosses, liverworts and hornworts. IDD - incomplete distribution data Anthocerotophyta (hornworts) Anthocerotaceae Notothyladaceae Bryophyta (mosses) Andreaeaceae Andreaeobryaceae Archidiaceae Aulacomniaceae Bartramiaceae Bryaceae Catoscopiaceae Hypopterygiaceae Meesiaceae Mniaceae Lists of plants Non-vascular plants Bryophytes
https://en.wikipedia.org/wiki/Southeast%20Asian%20Federation%20of%20Organizations%20of%20Medical%20Physics
South East Asian Federation of Organizations for Medical Physics (SEAFOMP) was officially formed when it was accepted as a regional chapter of the International Organization for Medical Physics (IOMP) at the World Congress on Medical Physics and Biomedical Engineering, held in Chicago, Illinois, in 2000. The founding member countries of SEAFOMP were Indonesia, Malaysia, Philippines, Singapore and Thailand. Two other ASEAN countries joined SEAFOMP later in 2002 (Brunei) and 2005 (Vietnam). SEAFOMP was established with the primary aim of helping to raise the profile of medical physics, via education and training, in ASEAN. The objectives of SEAFOMP are to promote (i) cooperation and communication between medical physics organizations in the region; (ii) medical physics and related activities in the region; (iii) the advancement in status and standard of practice of the medical physics profession; (iv) to organize and/or sponsor international and regional conferences, meetings or courses; (v) to collaborate or affiliate with other scientific organizations. Since then the initiative has been supported by many international organisations, including the American Association of Physicists in Medicine (AAPM), International Organization for Medical Physics (IOMP), International Atomic Energy Agency (IAEA), World Health Organization (WHO), and International Center for Theoretical Physics (ICTP). The support is primarily financial, via provisions of co-sponsorship of workshops and conferences, travel grants, medical physics libraries programs, and research experts and educators for educational activities. SEAFOMP has organised a series of congresses to promote scientific exchange and mutual support among its members. The South East Asian Congress for Medical Physics (SEACOMP) series were held respectively in Malaysia (Kuala Lumpur, 2001 and 2004), Thailand (Bangkok, 2003; Chiang Mai, 2009), Indonesia (Jakarta, 2006; Bandung, 2010), The Philippines (Manila, 2007; Manila & B
https://en.wikipedia.org/wiki/Pyramidal%20inversion
In chemistry, pyramidal inversion (also umbrella inversion) is a fluxional process in compounds with a pyramidal molecule, such as ammonia (NH3) "turns inside out". It is a rapid oscillation of the atom and substituents, the molecule or ion passing through a planar transition state. For a compound that would otherwise be chiral due to a stereocenter, pyramidal inversion allows its enantiomers to racemize. The general phenomenon of pyramidal inversion applies to many types of molecules, including carbanions, amines, phosphines, arsines, stibines, and sulfoxides. Energy barrier The identity of the inverting atom has a dominating influence on the barrier. Inversion of ammonia is rapid at room temperature, inverting 30 billion times per second. Three factors contribute to the rapidity of the inversion: a low energy barrier (24.2 kJ/mol; 5.8 kcal/mol), a narrow barrier width (distance between geometries), and the low mass of hydrogen atoms, which combine to give a further 80-fold rate enhancement due to quantum tunnelling. In contrast, phosphine (PH3) inverts very slowly at room temperature (energy barrier: 132 kJ/mol). Consequently, amines of the type RR′R"N usually are not optically stable (enantiomers racemize rapidly at room temperature), but P-chiral phosphines are. Appropriately substituted sulfonium salts, sulfoxides, arsines, etc. are also optically stable near room temperature. Steric effects can also influence the barrier. Nitrogen inversion Pyramidal inversion in nitrogen and amines is known as nitrogen inversion. It is a rapid oscillation of the nitrogen atom and substituents, the nitrogen "moving" through the plane formed by the substituents (although the substituents also move - in the other direction); the molecule passing through a planar transition state. For a compound that would otherwise be chiral due to a nitrogen stereocenter, nitrogen inversion provides a low energy pathway for racemization, usually making chiral resolution impossible. Quantum
https://en.wikipedia.org/wiki/Structural%20fracture%20mechanics
Structural fracture mechanics is the field of structural engineering concerned with the study of load-carrying structures that includes one or several failed or damaged components. It uses methods of analytical solid mechanics, structural engineering, safety engineering, probability theory, and catastrophe theory to calculate the load and stress in the structural components and analyze the safety of a damaged structure. There is a direct analogy between fracture mechanics of solid and structural fracture mechanics: There are different causes of the first component failure: mechanical overload, fatigue (material), unpredicted scenario, etc. “human intervention” like unprofessional behavior or a terrorist attack. There are two typical scenarios: A localized failure does NOT cause immediate collapse of the entire structure. The entire structure fails immediately after one of its components fails. If the structure does not collapse immediately there is a limited period of time until the catastrophic structural failure of the entire structure. There is a critical number of structural elements that defines whether the system has reserve ability or not. Safety engineers use the failure of the first component as an indicator and try to intervene during the given period of time to avoid the catastrophe of the entire structure. For example, “Leak-Before-Break” methodology means that a leak will be discovered prior to a catastrophic failure of the entire piping system occurring in service. It has been applied to pressure vessels, nuclear piping, gas and oil pipelines, etc. The methods of structural fracture mechanics are used as checking calculations to estimate sensitivity of a structure to its component failure. The failure of a complex system with parallel redundancy can be estimated based on probabilistic properties of the system elements. See also
https://en.wikipedia.org/wiki/Modulo%20%28mathematics%29
In mathematics, the term modulo ("with respect to a modulus of", the Latin ablative of modulus which itself means "a small measure") is often used to assert that two distinct mathematical objects can be regarded as equivalent—if their difference is accounted for by an additional factor. It was initially introduced into mathematics in the context of modular arithmetic by Carl Friedrich Gauss in 1801. Since then, the term has gained many meanings—some exact and some imprecise (such as equating "modulo" with "except for"). For the most part, the term often occurs in statements of the form: A is the same as B modulo C which is often equivalent to "A is the same as B up to C", and means A and B are the same—except for differences accounted for or explained by C. History Modulo is a mathematical jargon that was introduced into mathematics in the book Disquisitiones Arithmeticae by Carl Friedrich Gauss in 1801. Given the integers a, b and n, the expression "a ≡ b (mod n)", pronounced "a is congruent to b modulo n", means that a − b is an integer multiple of n, or equivalently, a and b both share the same remainder when divided by n. It is the Latin ablative of modulus, which itself means "a small measure." The term has gained many meanings over the years—some exact and some imprecise. The most general precise definition is simply in terms of an equivalence relation R, where a is equivalent (or congruent) to b modulo R if aRb. More informally, the term is found in statements of the form: A is the same as B modulo C which means A and B are the same—except for differences accounted for or explained by C. Usage Original use Gauss originally intended to use "modulo" as follows: given the integers a, b and n, the expression a ≡ b (mod n) (pronounced "a is congruent to b modulo n") means that a − b is an integer multiple of n, or equivalently, a and b both leave the same remainder when divided by n. For example: 13 is congruent to 63 modulo 10 means that 13 − 63 is a
https://en.wikipedia.org/wiki/Subcostal%20nerve
The subcostal nerve (anterior division of the twelfth thoracic nerve) is a mixed motor and sensory nerve contributing to the lumbar plexus. It runs along the lower border of the twelfth rib, often gives a communicating branch to the first lumbar nerve, and passes under the lateral lumbocostal arch. It then runs in front of the quadratus lumborum, innervates the transversus, and passes forward between it and the abdominal internal oblique to be distributed in the same manner as the lower intercostal nerves. It communicates with the iliohypogastric nerve and the ilioinguinal nerve of the lumbar plexus, and gives a branch to the pyramidalis muscle and the quadratus lumborum muscle. It also gives off a lateral cutaneous branch that supplies sensory innervation to the skin over the hip. Additional images See also Subcostal artery Subcostal vein
https://en.wikipedia.org/wiki/172nd%20meridian%20east
The meridian 172° east of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Asia, the Pacific Ocean, New Zealand, the Southern Ocean, and Antarctica to the South Pole. The 172nd meridian east forms a great circle with the 8th meridian west. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 172nd meridian east passes through: {| class="wikitable plainrowheaders" ! scope="col" width="130" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | East Siberian Sea | style="background:#b0e0e6;" | |-valign="top" | ! scope="row" | | Chukotka Autonomous Okrug Kamchatka Krai — from |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Bering Sea | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Pacific Ocean | style="background:#b0e0e6;" | Passing just east of Arno Atoll, (at ) |- | ! scope="row" | | Mili Atoll |-valign="top" | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Pacific Ocean | style="background:#b0e0e6;" | Passing just west of Hunter Island, (claimed by ) (at ) Passing just west of the Three Kings Islands, (at ) |- | ! scope="row" | | South Island – West Coast and Canterbury regions |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Pacific Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Southern Ocean | style="background:#b0e0e6;" | |- | ! scope="row" | Antarctica | Ross Dependency, claimed by |- |} See also 171st meridian east 173rd meridian east e172 meridian east
https://en.wikipedia.org/wiki/Powder%20mill
A powder mill was a mill where gunpowder is made from sulfur, saltpeter and charcoal. Milling steps Crude grinding and mixing operations such as the Frankford Powder-Mill of Philadelphia were a cottage industry until the industrial revolution brought improved product quality through the following procedures: Charcoal was often manufactured nearby from locally available trees, but the heating retorts were typically separated from the other buildings to minimize fire danger. Trees with low value as sources of lumber were debarked, dried, and cut to uniform length to fit into iron retorts with cast iron doors. The retorts were carefully packed to leave as little air space as possible, and the retort doors were closed and sealed with clay to prevent entry of air as the retorts were heated by external fires. Volatile gas generated by the heating process was vented through a small flue at the top of each retort. After the external fires were extinguished, the retorts were allowed to cool before the doors were opened to remove the charcoal. High purity sulfur usually required little preparation other than grinding to a powder. Separate grinding mills reduced cool charcoal and sulfur to fine powders. Crude saltpeter (KNO3) was dissolved in a pot of boiling water. After boiling, a small amount of floating glue might be added to the hot water to aid skimming off floating impurities. The hot solution was then carefully decanted to avoid suspending insoluble solid impurities settled to the bottom of the pot. The hot decanted liquid was distributed into shallow wooden vats where evaporative cooling would remove excess water. Purified potassium nitrate recrystallized while the cooling solution was gently agitated with broad hoes to prevent formation of large crystals. Powdered sulfur and charcoal were combined in appropriate ratios with moist saltpeter crystals, and the moist paste was mixed and compacted by kneading beneath heavy rotating wheels in a press mill. The compr
https://en.wikipedia.org/wiki/Serial%20homology
Serial homology is a special type of homology, defined by Owen as "representative or repetitive relation in the segments of the same organism." Ernst Haeckel preferred the term "homotypy" for the same phenomenon. Classical examples of serial homologies are the development of forelimbs and hind limbs of tetrapods and the iterative structure of the vertebrae. See also Deep homology Evolutionary developmental biology
https://en.wikipedia.org/wiki/Comodule%20over%20a%20Hopf%20algebroid
In mathematics, at the intersection of algebraic topology and algebraic geometry, there is the notion of a Hopf algebroid which encodes the information of a presheaf of groupoids whose object sheaf and arrow sheaf are represented by algebras. Because any such presheaf will have an associated site, we can consider quasi-coherent sheaves on the site, giving a topos-theoretic notion of modules. Duallypg 2, comodules over a Hopf algebroid are the purely algebraic analogue of this construction, giving a purely algebraic description of quasi-coherent sheaves on a stack: this is one of the first motivations behind the theory. Definition Given a commutative Hopf-algebroid a left comodule pg 302 is a left -module together with an -linear mapwhich satisfies the following two properties (counitary) (coassociative) A right comodule is defined similarly, but instead there is a mapsatisfying analogous axioms. Structure theorems Flatness of Γ gives an abelian category One of the main structure theorems for comodulespg 303 is if is a flat -module, then the category of comodules of the Hopf-algebroid is an Abelian category. Relation to stacks There is a structure theorempg 7 relating comodules of Hopf-algebroids and modules of presheaves of groupoids. If is a Hopf-algebroid, there is an equivalence between the category of comodules and the category of quasi-coherent sheaves for the associated presheaf of groupoidsto this Hopf-algebroid. Examples From BP-homology Associated to the Brown-Peterson spectrum is the Hopf-algebroid classifying p-typical formal group laws. Notewhere is the localization of by the prime ideal . If we let denote the idealSince is a primitive in , there is an associated Hopf-algebroid There is a structure theorem on the Adams-Novikov spectral sequence relating the Ext-groups of comodules on to Johnson-Wilson homology, giving a more tractable spectral sequence. This happens through an equivalence of categories of comodules of to the
https://en.wikipedia.org/wiki/Fractional-order%20system
In the fields of dynamical systems and control theory, a fractional-order system is a dynamical system that can be modeled by a fractional differential equation containing derivatives of non-integer order. Such systems are said to have fractional dynamics. Derivatives and integrals of fractional orders are used to describe objects that can be characterized by power-law nonlocality, power-law long-range dependence or fractal properties. Fractional-order systems are useful in studying the anomalous behavior of dynamical systems in physics, electrochemistry, biology, viscoelasticity and chaotic systems. Definition A general dynamical system of fractional order can be written in the form where and are functions of the fractional derivative operator of orders and and and are functions of time. A common special case of this is the linear time-invariant (LTI) system in one variable: The orders and are in general complex quantities, but two interesting cases are when the orders are commensurate and when they are also rational: When , the derivatives are of integer order and the system becomes an ordinary differential equation. Thus by increasing specialization, LTI systems can be of general order, commensurate order, rational order or integer order. Transfer function By applying a Laplace transform to the LTI system above, the transfer function becomes For general orders and this is a non-rational transfer function. Non-rational transfer functions cannot be written as an expansion in a finite number of terms (e.g., a binomial expansion would have an infinite number of terms) and in this sense fractional orders systems can be said to have the potential for unlimited memory. Motivation to study fractional-order systems Exponential laws are a classical approach to study dynamics of population densities, but there are many systems where dynamics undergo faster or slower-than-exponential laws. In such case the anomalous changes in dynamics may b
https://en.wikipedia.org/wiki/Swiss%20Ornithological%20Institute
The Swiss Ornithological Institute () is a non-profit foundation headquartered in Sempach in the district of Sursee in the canton of Lucerne in Switzerland dedicated to the study and conservation of birds. The Swiss Ornithological Institute was founded in 1924 as a ringing station for researching bird migration in the Alpine region. Since then, it has grown from a one-man volunteer operation into an internationally recognised institute with more than 130 employees. In 1954, the Swiss Ornithological Institute was registered as an independent foundation under Swiss law. In 1955, it moved from the town hall in Sempach to a new building on Lake Sempach. Since 1958, the institute has also operated a bird ringing station on Col de Bretolet at 1923 m asl. The Swiss Ornithological Institute monitors Switzerland's native birdlife, studies the behaviour of wild birds and seeks to understand the threats they face. It develops protection and conservation measures for threatened species and, together with its partners, works to ensure that the results of its research are used for the benefit of nature and birds. The Swiss Ornithological Institute operates a care centre for sick, injured and orphaned birds. As a national centre of expertise in ornithology and conservation, it also provides information and consulting services to the public, the media and the authorities. The institute offers programmes for schools and environmental education services to raise awareness for the protection of birds, a purpose primarily served by the state-of-the-art visitor centre that opened in 2015. The three-storey loam construction is the first of its kind in Switzerland. The private status of the Swiss Ornithological Institute is unique and distinguishes it from other ornithological institutes, many of which are governmental organisations. The Swiss Ornithological Institute is funded by donations and can count on more than 2,000 volunteers to support its nationwide bird-monitoring programmes
https://en.wikipedia.org/wiki/International%20Society%20for%20Structural%20and%20Multidisciplinary%20Optimization
The International Society for Structural and Multidisciplinary Optimization is a learned society in the field of multidisciplinary design optimization that was founded in October 1991. It has more than 1000 members in 45 countries. The current president is Wei Chen (Northwestern University). The society is an affiliated organization of the International Union of Theoretical and Applied Mechanics (IUTAM). Objectives The objectives are: to stimulate and promote research into all aspects of the optimal design of structures and related topics, including engineering systems consisting partially of structures and/or fluids; to encourage practical applications of optimization methods and the corresponding software development in all branches of technology; to foster the interchange of ideas amongst various fields contributing to structural and multidisciplinary optimization; to support the role of optimization in multidisciplinary design; to provide a framework for the organization of meetings and other means for the dissemination of knowledge on structural optimization; and to promote teaching of structural optimization in tertiary institutions. The society works towards these objectives by organizing a biennial "World Congress of Structural and Multidisciplinary Optimization" and publishing an official journal, Structural and Multidisciplinary Optimization, in collaboration with Springer Science+Business Media. Past Presidents 1991-1995: George Rozvany (Founder President) 1995-1999: Raphael Haftka 1999-2003: Niels Olhoff 2003-2007: Martin P. Bendsøe 2007-2011: KK Choi 2011-2015: Ole Sigmund 2015-2019: Cheng Gengdong
https://en.wikipedia.org/wiki/Cloned%20enzyme%20donor%20immunoassay
A cloned enzyme donor immunoassay (CEDIA) is a competitive homogenous enzyme immunoassay. This assay makes use of two component fragments of an enzyme which are each individually inactive. Under the right conditions in solution these fragments can spontaneously reassemble to form the active enzyme. For use in biochemical assays one of the enzyme fragments is attached to an analyte of interest. The analyte-enzyme-fragment-conjugate is still able to reassemble with the other enzyme fragment to form an active enzyme. However it is unable to do this if the analyte is bound to an antibody. To determine the quantity of analyte in a sample, an aliquot of sample must be added to a solution containing enzyme-fragment-analyte-conjugate, the other enzyme fragment, antibody directed against the analyte and substrate for the enzyme reaction. Competition for the antibody occurs between the analyte in the sample and the enzyme-fragment-analyte-conjugate. High concentrations of analyte in the sample lead to a relatively small amount of the enzyme-fragment-analyte-conjugate being prevented from forming active enzyme and therefore high enzyme activity. Conversely, low concentrations of analyte in the sample lead to a relatively large amount of the enzyme-fragment-analyte-conjugate being prevented from forming active enzymes and therefore low enzyme activity.
https://en.wikipedia.org/wiki/Active%20transport
In cellular biology, active transport is the movement of molecules or ions across a cell membrane from a region of lower concentration to a region of higher concentration—against the concentration gradient. Active transport requires cellular energy to achieve this movement. There are two types of active transport: primary active transport that uses adenosine triphosphate (ATP), and secondary active transport that uses an electrochemical gradient. This process is in contrast to passive transport, which allows molecules or ions to move down their concentration gradient, from an area of high concentration to an area of low concentration, without energy. Active transport is essential for various physiological processes, such as nutrient uptake, hormone secretion, and nerve impulse transmission. For example, the sodium-potassium pump uses ATP to pump sodium ions out of the cell and potassium ions into the cell, maintaining a concentration gradient essential for cellular function. Active transport is highly selective and regulated, with different transporters specific to different molecules or ions. Dysregulation of active transport can lead to various disorders, including cystic fibrosis, caused by a malfunctioning chloride channel, and diabetes, resulting from defects in glucose transport into cells. Active cellular transportation (ACT) Unlike passive transport, which uses the kinetic energy and natural entropy of molecules moving down a gradient, active transport uses cellular energy to move them against a gradient, polar repulsion, or other resistance. Active transport is usually associated with accumulating high concentrations of molecules that the cell needs, such as ions, glucose and amino acids. Examples of active transport include the uptake of glucose in the intestines in humans and the uptake of mineral ions into root hair cells of plants. History In 1848, the German physiologist Emil du Bois-Reymond suggested the possibility of active transport of subst
https://en.wikipedia.org/wiki/History%20of%20smallpox%20in%20Mexico
The history of smallpox in Mexico spans approximately 430 years from the arrival of the Spanish to the official eradication in 1951. It was brought to what is now Mexico by the Spanish, then spread to the center of Mexico, where it became a significant factor in the fall of Tenochtitlan. During the colonial period, there were major epidemic outbreaks which led to the implementation of sanitary and preventive policy. The introduction of smallpox vaccination in New Spain by Francisco Javier de Balmis and the work of Ignacio Bartolache reduced the mortality and morbidity of the disease. Introduction of smallpox Smallpox was an unknown disease not only in 16th century Mexico, but in all the Americas, before the arrival of Europeans. People in the Americas had not yet been exposed to the type of diseases that plagued the East, which meant that they had no resistance or immunity against them. It was introduced to Mexican lands by the Spanish and played a significant role in the downfall of the Aztec Empire. Hernán Cortés departed from Cuba and arrived in Mexico in 1519, sent to start trade relations only on the Veracruz Coast. However, he disobeyed the Cuban governor and began to invade the mainland. The governor sent Pánfilo de Narváez after Cortés. Narvaez's forces had at least one active case of smallpox, and when the Narvaez expedition stopped at Cozumel and Veracruz in 1520, the disease gained a foothold in the region. The introduction of smallpox among the Aztecs has been attributed to an African slave (by the name of Francisco Eguía, according to one account) but this has been disputed. From May to September, smallpox spread slowly to Tepeaca and Tlaxcala, and to Tenochtitlán by the fall of 1520. At this time, Cortes was returning to conquer the city after being thrown out on the Noche Triste. Cortes names only one indigenous leader who died of smallpox, Maxixcatzin. However, Cuitláhuac and other native rulers also died of smallpox. Chimalpahin reports the death
https://en.wikipedia.org/wiki/PhonePe
PhonePe is an Indian digital payments and financial services company headquartered in Bengaluru, Karnataka, India. PhonePe was founded in December 2015, by Sameer Nigam, Rahul Chari and Burzin Engineer. The PhonePe app, based on the Unified Payments Interface (UPI), went live in August 2016. The PhonePe app is available in 11 Indian languages. Using PhonePe, users can send and receive money, recharge mobile, DTH, data cards, make utility payments, pay at shops, invest in tax saving funds, buy insurance, mutual funds, and digital gold. PhonePe is accepted as a payment option by over 3.5 crore offline and online merchant outlets, constituting 99% of pin codes in the country. The app served more than 10 crore users as of June 2018, processed 500 crore transactions by December 2019, and crossed 10 crore transactions a day in April 2022. It currently has over 44 crore registered users with over 20 crore monthly active users. PhonePe is licensed by the Reserve Bank of India for the issuance and operation of a Semi Closed Prepaid Payment system. History PhonePe was incorporated in December 2015. In April 2016, the company was acquired by Flipkart and as part of the acquisition, the FxMart license was transferred to PhonePe and rebranded as the PhonePe wallet. PhonePe's founder Sameer Nigam was appointed as the CEO of the company. In August 2016, the company partnered with Yes Bank to launch a UPI-based mobile payment app, based on the government-backed UPI platform. Within three months of launch, the app was downloaded by over one crore users. In 2018, PhonePe became the fastest Indian payment app to get a five crore badge on the Google Play Store. The PhonePe app overtook BHIM to become the market leader in UPI transactions in August 2017. In 2022, PhonePe became the first UPI TPAP (Third Party Application Providers) App to allow UPI activation through Aadhaar. A year later, it further expanded its services by launching international UPI payments, allowing Indian
https://en.wikipedia.org/wiki/Eijir%C5%8D
is a very large database of English–Japanese translations. Developed by the editors of the Electronic Dictionary Project and aimed at translators, Eijirō is currently one of the most popular dictionaries on the Internet. Although the contents are technically the same, EDP refers to the accompanying Japanese–English database as . History The Eijirō project was started by an anonymous Japanese translator. Noting the favorable reception it received when he shared it with his friends, he started the Electronic Dictionary Project, a wiki-like structure that allowed for and even encouraged contributions to the dictionary. This resulted in a comprehensive database that grew to include over 1.66 million entries in the fourth edition. Characteristics Although commonly termed a dictionary, Eijirō differs from other Japanese dictionaries such as the Kōjien by not distinguishing examples from definitions. Access Eijirō can be purchased online as either a CD-R or downloadable dictionary file for a comparatively low price. Eijirō was also released from SpaceALC in 2002, and the SpaceALC version has since gone through eight revisions as of 2016. In addition, an online version of Eijirō is provided free of charge through the SpaceALC Japanese portal. Notes External Links and References Eijirō Homepage What is Eijirō? SpaceALC – a portal site which includes an online dictionary based on Eijirō. Pocket Eijirō The Story Behind Eijiro – A first-hand blog entry outlining the history of Eijirō Japanese dictionaries Online dictionaries
https://en.wikipedia.org/wiki/New%20England%20Enzyme%20Center
The New England Enzyme Center (NEEC) was created at the Tufts University School of Medicine in Boston, Massachusetts in 1964 as a federally supported biochemical resource center. History According to Doogab Yi, by the late 1970s NEEC had been transformed into "several commercial biotech companies." Roscoe O. Brady and his colleagues at National Institutes of Health (NIH) were almost ready for a clinical trial for an enzyme replacement therapy for Gaucher's disease that they had been working on for over a decade. They could not purify the enzyme in large enough quantities. Blair had started his career in the biotechnology industry working as a technician at Tufts medical school. In 1978 Henry E. Blair, from the NEEC and a team of researchers including Peter G. Pentchev, Roscoe O. Brady, Daniel E. Britton and Susan H. Sorrell from the National Institutes of Health co-authored a paper in the PNAS isolating and comparing enzymes in search of a treatment for Gaucher disease. In 1981 venture capitalist Sheridan Snyder, Henry Blair and George M. Whitesides created the start-up Genzyme and continued to produce the enzymes for the NIH. Genzyme's first office was an old clothing warehouse adjacent to Tufts Medical School.
https://en.wikipedia.org/wiki/Unicentric%20Castleman%20disease
Unicentric Castleman disease is a subtype of Castleman disease (also known as giant lymph node hyperplasia, lymphoid hamartoma, or angiofollicular lymph node hyperplasia), a group of lymphoproliferative disorders characterized by lymph node enlargement, characteristic features on microscopic analysis of enlarged lymph node tissue, and a range of symptoms and clinical findings. People with unicentric Castleman disease (UCD) have an enlarged lymph node or multiple enlarged lymph nodes in a single lymph node region. It is the most common subtype of Castleman disease, symptoms are typically mild, abnormalities on blood tests are uncommon, organ dysfunction is uncommon, and surgical treatment is curative in the majority of patients. The cause of UCD is not known. Castleman disease is named after Benjamin Castleman, who first described the disease in 1956. The Castleman Disease Collaborative Network is the largest organization focused on the disease and is involved in research, awareness, and patient support. Signs and symptoms In approximately half of patients, UCD presents without symptoms; however, people with the disease may experience enlarged lymph nodes in a single lymph node region or report symptoms related to compression of neighboring structures by enlarged lymph nodes, such as difficulty breathing and pain or pressure in the abdomen or chest. Systemic symptoms (fever, night sweats, weight loss, fatigue) are seen in up to half of UCD patients. Some symptoms that are commonly seen in HHV-8-associated MCD and iMCD like extravascular fluid accumulation (peripheral edema, ascites, pleural effusions), and enlargement of the liver and/or spleen are uncommon in UCD. Associated diseases UCD is associated with increased risk of paraneoplastic pemphigus as well as bronchiolitis obliterans (BO), AA amyloidosis, vascular neoplasms (e.g., FDC sarcoma), and possibly lymphomas. Causes UCD has no known causes or risk factors. Cases of Castleman disease running in fa
https://en.wikipedia.org/wiki/Expansive%20clay
Expansive clay is a clay soil that is prone to large volume changes (swelling and shrinking) that are directly related to changes in water content. Soils with a high content of expansive minerals can form deep cracks in drier seasons or years; such soils are called vertisols. Soils with smectite clay minerals, including montmorillonite and bentonite, have the most dramatic shrink-swell capacity. The mineral make-up of this type of soil is responsible for the moisture retaining capabilities. All clays consist of mineral sheets packaged into layers, and can be classified as either 1:1 or 2:1. These ratios refer to the proportion of tetrahedral sheets to octahedral sheets. Octahedral sheets are sandwiched between two tetrahedral sheets in 2:1 clays, while 1:1 clays have sheets in matched pairs. Expansive clays have an expanding crystal lattice in a 2:1 ratio; however, there are 2:1 non-expansive clays. Mitigation of the effects of expansive clay on structures built in areas with expansive clays is a major challenge in geotechnical engineering. Some areas mitigate foundation cracking by watering around the foundation with a soaker hose during dry conditions. This process can be automated by a timer, or using a soil moisture sensor controller. Even though irrigation is expensive, the cost is small compared to repairing a cracked foundation. Admixtures can be added to expansive clays to reduce the shrink-swell properties, as well. One laboratory test to measure the expansion potential of soil is ASTM D 4829. See also Argillipedoturbation Dispersion (soil)
https://en.wikipedia.org/wiki/Vanishing%20twin
A vanishing twin, also known as twin resorption, is a fetus in a multigestation pregnancy that dies in utero and is then partially or completely reabsorbed. In some instances, the dead twin is compressed into a flattened, parchment-like state known as fetus papyraceus. Vanishing twins occur in up to one of every eight multifetus pregnancies and may not even be known in most cases. "High resorption rates, which cannot be explained on the basis of the expected abortion rate, suggest intense fetal competition for space, nutrition, or other factors during early gestation, with frequent loss or resorption of the other twin(s)." According to Boklage, most twins are born as singles and vanished twins are a possible source of abnormal cells. In pregnancies achieved by in vitro fertilization, "it frequently happens that more than one amniotic sac can be seen in early pregnancy, whereas a few weeks later there is only one to be seen and the other has 'vanished. See also Chimera (genetics) Mosaicism Parasitic twin
https://en.wikipedia.org/wiki/Subcarrier
A subcarrier is a sideband of a radio frequency carrier wave, which is modulated to send additional information. Examples include the provision of colour in a black and white television system or the provision of stereo in a monophonic radio broadcast. There is no physical difference between a carrier and a subcarrier; the "sub" implies that it has been derived from a carrier, which has been amplitude modulated by a steady signal and has a constant frequency relation to it. FM stereo Stereo broadcasting is made possible by using a subcarrier on FM radio stations, which takes the left channel and "subtracts" the right channel from it — essentially by hooking up the right-channel wires backward (reversing polarity) and then joining left and reversed-right. The result is modulated with suppressed carrier AM, more correctly called sum and difference modulation or SDM, at 38 kHz in the FM signal, which is joined at 2% modulation with the mono left+right audio (which ranges 50 Hz ~ 15 kHz). A 19 kHz pilot tone is also added at a 9% modulation to trigger radios to decode the stereo subcarrier, making FM stereo fully compatible with mono. Once the receiver demodulates the L+R and L−R signals, it adds the two signals ([L+R] + [L−R] = 2L) to get the left channel and subtracts ([L+R] − [L−R] = 2R) to get the right channel. Rather than having a local oscillator, the 19 kHz pilot tone provides an in-phase reference signal used to reconstruct the missing carrier wave from the 38 kHz signal. For AM broadcasting, different analog (AM stereo) and digital (HD Radio) methods are used to produce stereophonic audio. Modulated subcarriers of the type used in FM broadcasting are impractical for AM broadcast due to the relatively narrow signal bandwidth allocated for a given AM signal. On standard AM broadcast radios, the entire 9 kHz to 10 kHz allocated bandwidth of the AM signal may be used for audio. Television Likewise, analog TV signals are transmitted with the black and whit
https://en.wikipedia.org/wiki/Genetic%20variation
Genetic variation is the difference in DNA among individuals or the differences between populations among the same species. The multiple sources of genetic variation include mutation and genetic recombination. Mutations are the ultimate sources of genetic variation, but other mechanisms, such as genetic drift, contribute to it, as well. Among individuals within a population Genetic variation can be identified at many levels. Identifying genetic variation is possible from observations of phenotypic variation in either quantitative traits (traits that vary continuously and are coded for by many genes (e.g., leg length in dogs)) or discrete traits (traits that fall into discrete categories and are coded for by one or a few genes (e.g., white, pink, or red petal color in certain flowers)). Genetic variation can also be identified by examining variation at the level of enzymes using the process of protein electrophoresis. Polymorphic genes have more than one allele at each locus. Half of the genes that code for enzymes in insects and plants may be polymorphic, whereas polymorphisms are less common among vertebrates. Ultimately, genetic variation is caused by variation in the order of bases in the nucleotides in genes. New technology now allows scientists to directly sequence DNA, which has identified even more genetic variation than was previously detected by protein electrophoresis. Examination of DNA has shown genetic variation in both coding regions and in the noncoding intron region of genes. Genetic variation will result in phenotypic variation if variation in the order of nucleotides in the DNA sequence results in a difference in the order of amino acids in proteins coded by that DNA sequence, and if the resultant differences in amino-acid sequence influence the shape, and thus the function of the enzyme. Between populations Differences between populations resulting from geographic separation is known as geographic variation. Natural selection, genetic
https://en.wikipedia.org/wiki/Xerophile
A xerophile () is an extremophilic organism that can grow and reproduce in conditions with a low availability of water, also known as water activity. Water activity (aw) is measured as the humidity above a substance relative to the humidity above pure water (Aw = 1.0). Xerophiles are "xerotolerant", meaning tolerant of dry conditions. They can often survive in environments with water activity below 0.8; above which is typical for most life on Earth. Typically xerotolerance is used with respect to matric drying, where a substance has a low water concentration. These environments include arid desert soils. The term osmotolerance is typically applied to organisms that can grow in solutions with high solute concentrations (salts, sugars), such as halophiles. The common food preservation method of reducing water (food drying) activities may not prevent the growth of xerophilic organisms, often resulting in food spoilage. Some mold and yeast species are xerophilic. Mold growth on bread is an example of food spoilage by xerophilic organisms. Examples of xerophiles include Trichosporonoides nigrescens, Zygosaccharomyces, and cacti. See also Xerocole Xerophyte
https://en.wikipedia.org/wiki/Malignant%20transformation
Malignant transformation is the process by which cells acquire the properties of cancer. This may occur as a primary process in normal tissue, or secondarily as malignant degeneration of a previously existing benign tumor. Causes There are many causes of primary malignant transformation, or tumorigenesis. Most human cancers in the United States are caused by external factors, and these factors are largely avoidable. These factors were summarized by Doll and Peto in 1981, and were still considered to be valid in 2015. These factors are listed in the table. a Reproductive and sexual behaviors include: number of partners; age at first menstruation; zero versus one or more live births Examples of diet-related malignant transformation Diet and colon cancer Colon cancer provides one example of the mechanisms by which diet, the top factor listed in the table, is an external factor in cancer. The Western diet of African Americans in the United States is associated with a yearly colon cancer rate of 65 per 100,000 individuals, while the high fiber/low fat diet of rural Native Africans in South Africa is associated with a yearly colon cancer rate of <5 per 100,000. Feeding the Western diet for two weeks to Native Africans increased their secondary bile acids, including carcinogenic deoxycholic acid, by 400%, and also changed the colonic microbiota. Evidence reviewed by Sun and Kato indicates that differences in human colonic microbiota play an important role in the progression of colon cancer. Diet and lung cancer A second example, relating a dietary component to a cancer, is illustrated by lung cancer. Two large population-based studies were performed, one in Italy and one in the United States. In Italy, the study population consisted of two cohorts: the first, 1721 individuals diagnosed with lung cancer and no severe disease, and the second, 1918 control individuals with absence of lung cancer history or any advanced diseases. All individuals filled out a
https://en.wikipedia.org/wiki/%28Histone-H3%29-lysine-36%20demethylase
(Histone-H3)-lysine-36 demethylase (, JHDM1A, JmjC domain-containing histone demethylase 1A, H3-K36-specific demethylase, histone-lysine (H3-K36) demethylase, histone demethylase, protein-6-N,6-N-dimethyl-L-lysine,2-oxoglutarate:oxygen oxidoreductase) is an enzyme with systematic name protein-N6,N6-dimethyl-L-lysine,2-oxoglutarate:oxygen oxidoreductase. This enzyme catalyses the following chemical reaction protein N6,N6-dimethyl-L-lysine + 2 2-oxoglutarate + 2 O2 protein L-lysine + 2 succinate + 2 formaldehyde + 2 CO2 (overall reaction) (1a) protein N6,N6-dimethyl-L-lysine + 2-oxoglutarate + O2 protein N6-methyl-L-lysine + succinate + formaldehyde + CO2 (1b) protein N6-methyl-L-lysine + 2-oxoglutarate + O2 protein L-lysine + succinate + formaldehyde + CO2 (Histone-H3)-lysine-36 demethylase contains iron(II).
https://en.wikipedia.org/wiki/Number%20Theory%20Foundation
The Number Theory Foundation (NTF) is a non-profit organization based in the United States which supports research and conferences in the field of number theory, with a particular focus on computational aspects and explicit methods. The NTF funds the Selfridge prize awarded at each Algorithmic Number Theory Symposium (ANTS) and is a regular supporter of several conferences and organizations in number theory, including the Canadian Number Theory Association (CNTA), Women in Numbers (WIN), and the West Coast Number Theory (WCNT) conference. History The NTF was created in 1999 via a grant from John Selfridge with an initial board of directors including Paul Bateman, John Brillhart, Richard Blecksmith, Brian Conrey, Ronald Graham, Richard Guy, Carl Pomerance, John Selfridge, Sam Wagstaff, and Hugh Williams. Carl Pomerance served as President of the foundation for its first two decades and was succeeded by Andrew Sutherland in 2019.
https://en.wikipedia.org/wiki/Brown%20rot
Brown rot may refer to the following diseases: Wood-decay fungus, fungi that digest moist wood, causing rot, includes various species that infect living trees and cured wood Ralstonia solanacearum, an aerobic, non-sporing, plant pathogenic bacterium that causes brown rot in a wide range of crops Monilinia fructicola, a plant pathogenic fungus, the cause of brown rot in stone fruits, such as plums, peaches, nectarines and almonds Two of the Gnomoniopsis - G. smithogilvyii or G. castaneae - in chestnuts Almond brown rot caused by Monilinia fructicola
https://en.wikipedia.org/wiki/Time%20to%20first%20byte
Time to first byte (TTFB) is a measurement used as an indication of the responsiveness of a webserver or other network resource. TTFB measures the duration from the user or client making an HTTP request to the first byte of the page being received by the client's browser. This time is made up of the socket connection time, the time taken to send the HTTP request, and the time taken to get the first byte of the page. Although sometimes misunderstood as a post-DNS calculation, the original calculation of TTFB in networking always includes network latency in measuring the time it takes for a resource to begin loading. Often, a smaller (faster) TTFB size is seen as a benchmark of a well-configured server application. For example, a lower time to first byte could point to fewer dynamic calculations being performed by the webserver, although this is often due to caching at either the DNS, server, or application level. More commonly, a very low TTFB is observed with statically served web pages, while larger TTFB is often seen with larger, dynamic data requests being pulled from a database. Uses in web development Time to first byte is important to a webpage since it indicates pages that load slowly due to server-side calculations that might be better served as client-side scripting. Often this includes simple scripts and calculations like transitioning images that are not gifs and are transitioned using JavaScript to modify their transparency levels. This can often speed up a website by downloading multiple smaller images through sockets instead of one large image. However this technique is more intensive on the client's computer and on older PCs can slow the webpage down when actually rendering. Importance TTFB is often used by web search engines like Google and Yahoo to improve search rankings since a website will respond to the request faster and be usable before other websites would be able to. There are downsides to this metric since a web-server can send only the
https://en.wikipedia.org/wiki/Relocation%20%28computing%29
Relocation is the process of assigning load addresses for position-dependent code and data of a program and adjusting the code and data to reflect the assigned addresses. Prior to the advent of multiprocess systems, and still in many embedded systems, the addresses for objects were absolute starting at a known location, often zero. Since multiprocessing systems dynamically link and switch between programs it became necessary to be able to relocate objects using position-independent code. A linker usually performs relocation in conjunction with symbol resolution, the process of searching files and libraries to replace symbolic references or names of libraries with actual usable addresses in memory before running a program. Relocation is typically done by the linker at link time, but it can also be done at load time by a relocating loader, or at run time by the running program itself. Some architectures avoid relocation entirely by deferring address assignment to run time; as, for example, in stack machines with zero address arithmetic or in some segmented architectures where every compilation unit is loaded into a separate segment. Segmentation Object files are segmented into various memory segment types. Example segments include code segment (.text), initialized data segment (.data), uninitialized data segment (.bss), or others. Relocation table The relocation table is a list of pointers created by the translator (a compiler or assembler) and stored in the object or executable file. Each entry in the table, or "fixup", is a pointer to an absolute address in the object code that must be changed when the loader relocates the program so that it will refer to the correct location. Fixups are designed to support relocation of the program as a complete unit. In some cases, each fixup in the table is itself relative to a base address of zero, so the fixups themselves must be changed as the loader moves through the table. In some architectures a fixup that crosses ce
https://en.wikipedia.org/wiki/Flag%20of%20the%20Western%20European%20Union
The flag of the Western European Union (WEU) was dark blue with a semicircle of ten yellow five pointed stars, broken at the top, with the organisation's initials in the centre. Although it is the flag of a military organisation, it has rarely been flown in military situations. Design The flag is dark blue with a semicircle of ten yellow five pointed stars, broken at the top, with the white letters WEU horizontally across the centre and UEO vertically across the centre sharing the letter E with the former set of initials. UEO is the French abbreviation for Western European Union (. The flag's blue colour with yellow stars is taken from the flag of the Council of Europe and European Union, however the number of stars is ten due to the WEU membership being of that number. Use The flag was rarely used, with the WEU being largely dormant before it was succeeded by the European Union's (EU) military activities. It was once flown on board an operational United States Navy warship, , when it was used as the flagship of an Italian general (with a WEU crew) commanding WEU relief operations in Bosnia and Herzegovina. There are variants for its bodies, such as its assembly, used on formal occasions. The WEU treaty is now terminated, with WEU activities wound up by June 2011, so no further use of the flag is foreseeable. Previous design The current flag was in use only since 1993. Prior to that there was a similar design but with only nine stars (before Greece became a member) and the stars towards the base of the flag were progressively larger than those at the fringe. This design replaced the flag of the Western Union, the organisation that was transformed into the WEU upon the entry into force of the Modified Brussels Treaty. See also Flag of the Western Union Flag of Europe Flag of NATO Flag of the European Coal and Steel Community Federalist flag
https://en.wikipedia.org/wiki/Polymorphic%20association
Polymorphic association is a term used in discussions of Object-Relational Mapping with respect to the problem of representing in the relational database domain, a relationship from one class to multiple classes. In statically typed languages such as Java these multiple classes are subclasses of the same superclass. In languages with duck typing, such as Ruby, this is not necessarily the case. See also Polymorphism in object-oriented programming Hibernate (Java)
https://en.wikipedia.org/wiki/Advanced%20Comprehensive%20Operating%20System
Advanced Comprehensive Operating System (ACOS) is a family of mainframe computer operating systems developed by NEC for the Japanese market. It consists of three systems, based on the General Comprehensive Operating System family developed by General Electric, Honeywell, and Bull. Two of these systems, ACOS-2 (based on GCOS 4) and ACOS-4 (based on GCOS 7) are still sold, although only ACOS-4 is under active development. ACOS-6 (based on GCOS 8) is an obsolete high-end mainframe platform, which ceased active development in the early 2000s. The first two models in NEC's SX series of supercomputers, the SX-1 and the SX-2 (released in 1985), ran an operating system derived from ACOS-4, which was variously called either SX-OS or SXCP (SX System Control Program). However, subsequent SX supercomputers, starting with the SX-3 (released in 1990), instead ran a derivative of Unix. In late September 2012, NEC announced a return from IA-64 to the previous NOAH line of proprietary mainframe processors for ACOS-4, now produced in a quad-core variant on 40 nm, called NOAH-6. ACOS-2 runs on Intel Xeon servers. In June 2022, i-PX AKATSUKI server equipped with NEC's original processor (NOAH-7) was released. See also Timeline of operating systems
https://en.wikipedia.org/wiki/Amanita%20flavella
Amanita flavella is a species of mycorrhizal fungus from family Amanitaceae that can be found in New South Wales and Queensland Australia. The species have a convex lemon-yellow coloured cap that is up to in diameter. They can also be yellowish-orange coloured and have crowded gills that are pale yellow in colour. The stipe is central and just like the cap is high and yellowish white in colour. It is slightly bulbous, and is enclosed into a volva. The ring is flared and white coloured. The ring is ample, membranous, and yellowish in colour. The spores are 8.5–10 μm long and 6–6.5 μm wide, and are white coloured, amyloid and ellipsoid. The species is similar in appearance to Amanita flavoconia and Amanita flavipes. See also List of Amanita species
https://en.wikipedia.org/wiki/Cyber%20Insider%20Threat
Cyber Insider Threat, or CINDER, is a digital threat method. In 2010, DARPA initiated a program under the same name (Cyber Insider Threat (CINDER) Program) to develop novel approaches to the detection of activities within military-interest networks that are consistent with the activities of cyber espionage. The CINDER threat is unlike other vulnerability based attacks in that the action taken by the initiator is not based on unauthorized access by unauthorized objects or authorized objects, it is based on the concept that authorized access by authorized objects will normally occur (along with their subsequent actions) within the security boundary. This object action will not be viewed as an attack, but normal use when analyzed by standard IDS-IPS, logging and expert systems. The CINDER Mission will be seen as an unauthorized disclosure once data exfiltration has been realized. At that time, the resultant CINDER Case would change all object actions related to the disclosure from "Authorized Use by an Authorized Object" to "Unauthorized Use by an Authorized Object". Note: For the initial CINDER case, the controlling agent will still be seen as an Authorized Object based on the fact that the security system has passed an evaluation for Assurance and Functionality. The Cyber Insider Threat has continued to be a known issue since the mid-1980s. The following NIST material dated March 1994, "Internal Threats", shows how it was defined in its infancy. "System controls are not well matched to the average organization's security policy. As a direct result, the typical user is permitted to circumvent that policy on a frequent basis. The administrator is unable to enforce the policy because of the weak access controls, and cannot detect the violation of policy because of weak audit mechanisms. Even if the audit mechanisms are in place, the daunting volume of data produced makes it unlikely that the administrator will detect policy violations. Ongoing research in integrity a
https://en.wikipedia.org/wiki/Publish%E2%80%93subscribe%20pattern
In software architecture, publish–subscribe is a messaging pattern where publishers categorize messages into classes that are received by subscribers. This is contrasted to the typical messaging pattern model where publishers sends messages directly to subscribers. Similarly, subscribers express interest in one or more classes and only receive messages that are of interest, without knowledge of which publishers, if any, there are. Publish–subscribe is a sibling of the message queue paradigm, and is typically one part of a larger message-oriented middleware system. Most messaging systems support both the pub/sub and message queue models in their API; e.g., Java Message Service (JMS). This pattern provides greater network scalability and a more dynamic network topology, with a resulting decreased flexibility to modify the publisher and the structure of the published data. Message filtering In the publish-subscribe model, subscribers typically receive only a subset of the total messages published. The process of selecting messages for reception and processing is called filtering. There are two common forms of filtering: topic-based and content-based. In a topic-based system, messages are published to "topics" or named logical channels. Subscribers in a topic-based system will receive all messages published to the topics to which they subscribe. The publisher is responsible for defining the topics to which subscribers can subscribe. In a content-based system, messages are only delivered to a subscriber if the attributes or content of those messages matches constraints defined by the subscriber. The subscriber is responsible for classifying the messages. Some systems support a hybrid of the two; publishers post messages to a topic while subscribers register content-based subscriptions to one or more topics. Topologies In many publish-subscribe systems, publishers post messages to an intermediary message broker or event bus, and subscribers register subscription
https://en.wikipedia.org/wiki/Jurkat%E2%80%93Richert%20theorem
The Jurkat–Richert theorem is a mathematical theorem in sieve theory. It is a key ingredient in proofs of Chen's theorem on Goldbach's conjecture. It was proved in 1965 by Wolfgang B. Jurkat and Hans-Egon Richert. Statement of the theorem This formulation is from Diamond & Halberstam. Other formulations are in Jurkat & Richert, Halberstam & Richert, and Nathanson. Suppose A is a finite sequence of integers and P is a set of primes. Write Ad for the number of items in A that are divisible by d, and write P(z) for the product of the elements in P that are less than z. Write ω(d) for a multiplicative function such that ω(p)/p is approximately the proportion of elements of A divisible by p, write X for any convenient approximation to |A|, and write the remainder as Write S(A,P,z) for the number of items in A that are relatively prime to P(z). Write Write ν(m) for the number of distinct prime divisors of m. Write F1 and f1 for functions satisfying certain difference differential equations (see Diamond & Halberstam for the definition and properties). We assume the dimension (sifting density) is 1: that is, there is a constant C such that for 2 ≤ z < w we have (The book of Diamond & Halberstam extends the theorem to dimensions higher than 1.) Then the Jurkat–Richert theorem states that for any numbers y and z with 2 ≤ z ≤ y ≤ X we have and Notes Sieve theory Theorems in analytic number theory
https://en.wikipedia.org/wiki/Garden%20of%20Cosmic%20Speculation
The Garden of Cosmic Speculation is a 30 acre (12 hectare) sculpture garden created by landscape architect and theorist Charles Jencks at his home, Portrack House, in Dumfriesshire, Scotland. Like much of Jencks' work, the garden is inspired by modern cosmology. History Features The garden is inspired by science and mathematics, with sculptures and landscaping on these themes, such as black holes and fractals. The garden is not abundant with plants, but sets mathematical formulae and scientific phenomena in a setting which elegantly combines natural features and artificial symmetry and curves. It is probably unique among gardens, drawing comparisons with a similarly abstract garden in Scotland, Little Sparta. Access The garden is private but usually opens for only five hours on one day each year for 1500 ticket holders through the Scotland's Gardens programme and raises money for Maggie's Centres, a cancer care charity named for Maggie Keswick Jencks, the late wife of Charles Jencks. Depiction in music The garden is the subject of an orchestral composition by American composer, Michael Gandolfi, which he composed for a joint commission from the Boston Symphony Orchestra and the Tanglewood Music Center. The piece was subsequently recorded by the Atlanta Symphony Orchestra conducted by Robert Spano, and nominated for "Best Contemporary Classical Composition" at the 2009 Grammy Awards. In literature Louise Penny uses The Garden of Cosmic Speculation as an important plot device in her tenth Inspector Gamache mystery, The Long Way Home (St. Martin Press, 2014). Cameron Jace makes creative use of The Garden of Cosmic Speculation in his fictional novel titled Circus, which is the third installment of his Insanity series. In the book, Jace uses many facts when referring to 'public' knowledge of the garden (per character conversation), but changed the name of the designer to better fit into the story's plot line. See also Crawick Multiverse Jupiter Artland Land a
https://en.wikipedia.org/wiki/Brown%20dwarf
Brown dwarfs (also called failed stars) are substellar objects that, while more massive than the most massive gas giant planets, are (unlike a main-sequence star), not massive enough to sustain nuclear fusion of ordinary hydrogen (1H) into helium in their cores. Smaller than the least massive stars, approximately 13 to 80 times that of Jupiter (), they are massive enough to fuse deuterium (2H) and the most massive ones (> ) can fuse lithium (7Li). Astronomers classify self-luminous objects by spectral type, a distinction intimately tied to the surface temperature, and brown dwarfs occupy types M, L, T, and Y. As brown dwarfs do not undergo stable hydrogen fusion, they cool down over time, progressively passing through later spectral types as they age. Despite their name, to the naked eye, brown dwarfs would appear in different colors depending on their temperature. The warmest ones are possibly orange or red, while cooler brown dwarfs would likely appear magenta or black to the human eye. Brown dwarfs may be fully convective, with no layers or chemical differentiation by depth. Though their existence was initially theorized in the 1960s, it was not until the mid-1990s that the first unambiguous brown dwarfs were discovered. As brown dwarfs have relatively low surface temperatures, they are not very bright at visible wavelengths, emitting most of their light in the infrared. However, with the advent of more capable infrared detecting devices, thousands of brown dwarfs have been identified. The nearest known brown dwarfs are located in the Luhman 16 system, a binary of L- and T-type brown dwarfs about from the Sun. Luhman 16 is the third closest system to the Sun after Alpha Centauri and Barnard's Star. History Early theorizing The objects now called "brown dwarfs" were theorized by Shiv S. Kumar in the 1960s to exist and were originally called black dwarfs, a classification for dark substellar objects floating freely in space that were not massive enough
https://en.wikipedia.org/wiki/Donald%20A.%20S.%20Fraser
Donald Alexander Stuart Fraser (April 29, 1925–December 21, 2020) was a Canadian statistician, and Professor Emeritus at the University of Toronto. In 2012 he was appointed an Officer of the Order of Canada for his influence in the advancement of the statistical sciences in Canada. In 1961 he was elected as a Fellow of the American Statistical Association. In 1985, he was awarded the first Gold Medal of the Statistical Society of Canada. In 2014 he was chosen as a fellow of the American Mathematical Society "for contributions to the theory and foundations of statistics, as well as for leadership and influence on the advancement of the statistical sciences." Early life Donald Fraser was born in Toronto, and raised in Stratford, Ontario. He attended St. Andrew's College in Aurora from 1939 to 1942, and the University of Toronto from 1942 to 1947, completing a bachelor's degree in mathematics, physics and chemistry. While at the University of Toronto he distinguished himself in the sciences, and especially in mathematics. He was a member of the first place team and a Putnam fellow in the William Lowell Putnam Mathematical Competition in 1946.
https://en.wikipedia.org/wiki/Finitely%20generated%20object
In category theory, a finitely generated object is the quotient of a free object over a finite set, in the sense that it is the target of a regular epimorphism from a free object that is free on a finite set. For instance, one way of defining a finitely generated group is that it is the image of a group homomorphism from a finitely generated free group. See also Finitely generated (disambiguation)
https://en.wikipedia.org/wiki/U-Report
U-Report is a social messaging tool and data collection system developed by UNICEF to improve citizen engagement, inform leaders, and foster positive change. The program sends SMS polls and alerts to its participants, collecting real-time responses, and subsequently publishes gathered data. Issues polled include health, education, water, sanitation and hygiene, youth unemployment, HIV/AIDS, and disease outbreaks. The program currently has 28 million u-reporters in 95 countries. History In 2007, UNICEF Innovation used RapidSMS to develop U-Report, a platform that would allow anyone to publish real-time information and data analytics in SMS format without the need of a programmer. In May 2011, Uganda became the first country in which UNICEF launched the U-Report mobile initiative, due to its population being, on average, one of the youngest in the world. Another reason UNICEF cited for introducing the program in Uganda was the nation's high cellphone use compared to other developing nations, with 48% of the nation's citizens owning a cellphone. Due to U-Report's success in Uganda, UNICEF expanded the program to Zambia in December 2012 and to Nigeria in June 2014. In Zambia, U-report was used to prevent HIV among adolescents and young people, with voluntary HIV testing in the country rising from 24% of the population to 40%. In Nigeria, U-Report primarily conducts surveys on social and medical issues. In July 2015, U-Report reached a total of one million reporters in fifteen countries. In October 2015, Ukraine became the first country in Europe to join the U-Report program, growing to 68,273 participants by September 30, 2018. See also World Health Report Human Development Report The State of the World's Children World Development Report
https://en.wikipedia.org/wiki/Cancer%20selection
Cancer selection can be viewed through the lens of natural selection. The animal host's body is the environment which applies the selective pressures upon cancer cells. The most fit cancer cells will have traits that will allow them to out compete other cancer cells which they are related to, but are genetically different from. This genetic diversity of cells within a tumor gives cancer an evolutionary advantage over the host's ability to inhibit and destroy tumors. Therefore, other selective pressures such as clinical treatments and pharmaceutical treatments are needed to help destroy the large amount of genetically diverse cancerous cells within a tumor. It is because of the high genetic diversity between cancer cells within a tumor that makes cancer a formidable foe for the survival of animal hosts. It has also been proposed that cancer selection is a selective force that has driven the evolution of animals. Therefore, cancer and animals have been paired as competitors in co-evolution throughout time. Natural selection Evolution, which is driven by natural selection, is the cornerstone for nearly all branches of biology including cancer biology. In 1859, Charles Darwin's book On the Origin of Species was published, in which Darwin proposed his theory of evolution by means of natural selection. Natural selection is the force that drives changes in the phenotypes observed in populations over time, and is therefore responsible for the diversity amongst all living things. It is through the pressures applied by natural selection upon individuals that leads to evolutionary change over time. Natural selection is simply the selective pressures acting upon individuals within a population due to changes in their environment which picks the traits that are best fit for the selective change. Selection and cancer These same observations that Darwin proposed for the diversity in phenotypes amongst all living things can also be applied to cancer biology to explai
https://en.wikipedia.org/wiki/Internet%20interventions%20for%20post-traumatic%20stress
Internet interventions for post-traumatic stress have grown in popularity due to the limits that many patients face in their ability to seek therapy to treat their symptoms. These limits include lack of resources and residing in small towns or in the countryside. These patients may find it difficult to seek treatment because they do not have geographical access to treatment, and this can also limit the time they have to seek help. Additionally, those who live in rural areas may experience more stigma related to mental health issues. Internet interventions can increase the possibility that those who suffer from PTSD can seek help by eliminating these barriers to treatment. Most of the internet interventions for PTSD currently being studied use Cognitive Behavioral Therapy (CBT) tenants to provide treatment. Often these internet interventions also pull from Cognitive Processing Therapy (CPT) and exposure therapy as well. There are two types of internet interventions. Those that are “therapist-assisted,” which means there is an actual therapist guiding the patient through some, but not all of the intervention, and those that are self-guided, which means they do not provide this service. In therapist assisted interventions, patients have access to a live therapist either via video conferencing, instant messenger, or telephone. Therapists can provide feedback to the patient's assignments, and help them process their trauma. In self-guided interventions, patients do not have contact with therapists, unless there is an emergency in which they are a risk to themselves or others. Throughout these interventions, patients are given coping skills and resources. The resources available to patients participating in a self-guided intervention are typically crisis lines, emergency services, and outside sources in which the person can seek help or treatment. The resources provided in a self-guided treatment protocol are not a part of the intervention itself. While there is a recen
https://en.wikipedia.org/wiki/Beer%20can%20pyramid
A beer can pyramid, often called a beeramid as a portmanteau, is a pyramid made from discarded beer cans. Beer can pyramids are built as empty beer cans became available, slowly growing as the night (or week or month) wears on. In most cases, though, they are temporary structures, eventually being cleaned up or accidentally knocked over. On 23 September 2000, the Malaysian Can team, consisting of 12 college students from the INTI College Subang Jaya, Malaysia built a free standing can pyramid created from 9,455 empty aluminium drink cans in 24 minutes at the Mid Valley Megamall in Kuala Lumpur, Malaysia. It had a square base of cans, measuring . This feat made a successful entry into the Guinness World Record and to-date this record has yet to be broken. Another attempt to break the world record Beer Can Pyramid was made with beer cans over 5 metres high and contained 10,660 cans. It was built by the Melbourne University Student Union in 2005, and was featured on Blokesworld and in mX. A beer can pyramid was shown on the outside of Duff Gardens in The Simpsons episode 9F11 "Selma's Choice", in 1993, as a parody of Cinderella's Castle at Walt Disney World.
https://en.wikipedia.org/wiki/Electrostatic%20particle%20accelerator
An electrostatic particle accelerator is a particle accelerator in which charged particles are accelerated to a high energy by a static high voltage potential. This contrasts with the other major category of particle accelerator, oscillating field particle accelerators, in which the particles are accelerated by oscillating electric fields. Owing to their simpler design, electrostatic types were the first particle accelerators. The two most common types are the Van de Graaf generator invented by Robert Van de Graaff in 1929, and the Cockcroft-Walton accelerator invented by John Cockcroft and Ernest Walton in 1932. The maximum particle energy produced by electrostatic accelerators is limited by the maximum voltage which can be achieved the machine. This is in turn limited by insulation breakdown to a few megavolts. Oscillating accelerators do not have this limitation, so they can achieve higher particle energies than electrostatic machines. The advantages of electrostatic accelerators over oscillating field machines include lower cost, the ability to produce continuous beams, and higher beam currents that make them useful to industry. As such, they are by far the most widely used particle accelerators, with industrial applications such as plastic shrink wrap production, high power X-ray machines, radiation therapy in medicine, radioisotope production, ion implanters in semiconductor production, and sterilization. Many universities worldwide have electrostatic accelerators for research purposes. High energy oscillating field accelerators usually incorporate an electrostatic machine as their first stage, to accelerate particles to a high enough velocity to inject into the main accelerator. Electrostatic accelerators are a subset of linear accelerators (linacs). While all linacs accelerate particles in a straight line, electrostatic accelerators use a fixed accelerating field from a single high voltage source, while radiofrequency linacs use oscillating electri
https://en.wikipedia.org/wiki/Ioana%20Dumitriu
Ioana Dumitriu (born July 6, 1976) is a Romanian-American mathematician who works as a professor of mathematics at the University of California, San Diego. Her research interests include the theory of random matrices, numerical analysis, scientific computing, and game theory. Life Dumitriu is the daughter of two Romanian electrical engineering professors from Bucharest. Early in her life she was identified as having mathematical talent, and at age 11 won a national mathematics contest. She entered mathematics training camps in preparation for participation on the Romanian team at the International Mathematical Olympiad, although her highest level of participation in the olympiad was the national semifinal. As a 19-year-old freshman at NYU, Dumitriu already was taking graduate-level classes in mathematics. She graduated summa cum laude from NYU in 1999 with a B.A. in mathematics and a minor in computer science. She earned her Ph.D. in 2003 from the Massachusetts Institute of Technology under the supervision of Alan Edelman, with a thesis on Eigenvalue statistics for beta-ensembles. After postdoctoral research as a Miller Research Fellow at the University of California, Berkeley, she joined the faculty of the University of Washington in 2006, moving to UC San Diego in 2019. Awards and honors Dumitriu won the Alice T. Schafer prize for excellence in mathematics by an undergraduate woman in 1996. Also in 1996, as a sophomore at New York University, Dumitriu became the first woman to become a Putnam Fellow, meaning that she earned one of the top five scores at the William Lowell Putnam Mathematical Competition. In 1995, 1996, and 1997 she won the Elizabeth Lowell Putnam Award that is given to the top woman in the contest, a record that was not matched until ten years later when Alison Miller also won the same award in three consecutive years. She won the Leslie Fox Prize for Numerical Analysis (given to a young numerical analysis researcher who excels both mathemati
https://en.wikipedia.org/wiki/Euler%20operator%20%28digital%20geometry%29
In solid modeling and computer-aided design, the Euler operators modify the graph of connections to add or remove details of a mesh while preserving its topology. They are named by Baumgart after the Euler–Poincaré characteristic. He chose a set of operators sufficient to create useful meshes, some lose information and so are not invertible. The boundary representation for a solid object, its surface, is a polygon mesh of vertices, edges and faces. Its topology is captured by the graph of the connections between faces. A given mesh may actually contain multiple unconnected shells (or bodies); each body may be partitioned into multiple connected components each defined by their edge loop boundary. To represent a hollow object, the inside and outside surfaces are separate shells. Let the number of vertices be V, edges be E, faces be F, components H, shells S, and let the genus be G (S and G correspond to the b0 and b2 Betti numbers respectively). Then, to denote a meaningful geometric object, the mesh must satisfy the generalized Euler–Poincaré formula V – E + F = H + 2 * (S – G) The Euler operators preserve this characteristic. The Eastman paper lists the following basic operators, and their effects on the various terms: Geometry Euler operators modify the mesh's graph creating or removing faces, edges and vertices according to simple rules while preserving the overall topology thus maintaining a valid boundary (i.e. not introducing holes). The operators themselves don't define how geometric or graphical attributes map to the new graph: e.g. position, gradient, uv texture coordinate, these will depend on the particular implementation. See also Boundary representation Lecture 31 of AML710 Computer Aided Design – Dr S. Hegde of Indian Institute of Technology Delhi
https://en.wikipedia.org/wiki/Sparse%20network
In network science, a sparse network has much fewer links than the possible maximum number of links within that network (the opposite is a dense network). The study of sparse networks is a relatively new area primarily stimulated by the study of real networks, such as social and computer networks. The notion of much fewer links is, of course, colloquial and informal. While a threshold for a particular network may be invented, there is no universal threshold that defines what much fewer actually means. As a result, there is no formal sense of sparsity for any finite network, despite widespread agreement that most empirical networks are indeed sparse. There is, however, a formal sense of sparsity in the case of infinite network models, determined by the behavior of the number of edges (M) and/or the average degree () as the number of nodes (N) goes to infinity. Definitions A simple unweighted network of size is called sparse if the number of links in it is much smaller than the maximum possible number of links : . In any given (real) network, the number of nodes N and links M are just two numbers, therefore the meaning of the much smaller sign ( above) is purely colloquial and informal, and so are statements like "many real networks are sparse." However, if we deal with a synthetic graph sequence , or a network model that is well defined for networks of any size N = 1,2,...,, then the attains its usual formal meaning: . In other words, a network sequence or model is called dense or sparse depending on whether the (expected) average degree in scales linearly or sublinearly with N: is dense if ; is sparse if . An important subclass of sparse networks are networks whose average degree is either constant or converges to a constant. Some authors call only such networks sparse, while others reserve special names for them: is truly sparse or extremely sparse or ultrasparse if . There also exist alternative, stricter definitions of network spa
https://en.wikipedia.org/wiki/Midparent
The midparent value is defined as the average of the trait value of father and a scaled version of the mother. This value can be used in the study to analyze the data set without heeding sex effects. Studying quantitative traits in heritability studies may be complicated by sex differences observed for the trait. Well-known examples include studies of stature height, whose midparent value hmp is given by: where hf and hm are, respectively, the father's and mother's heights. The coefficient 1.08 serves as a scaling factor. After the 1.08 scaling, the mean of the mother's height is the same as that of the father's, and the variance is closer to the father's; in this way, sex difference can be ignored.
https://en.wikipedia.org/wiki/Torpor
Torpor is a state of decreased physiological activity in an animal, usually marked by a reduced body temperature and metabolic rate. Torpor enables animals to survive periods of reduced food availability. The term "torpor" can refer to the time a hibernator spends at low body temperature, lasting days to weeks, or it can refer to a period of low body temperature and metabolism lasting less than 24 hours, as in "daily torpor". Animals that undergo daily torpor include birds (even tiny hummingbirds, notably Cypselomorphae) and some mammals, including many marsupial species, rodent species (such as mice), and bats. During the active part of their day, such animals maintain normal body temperature and activity levels, but their metabolic rate and body temperature drop during a portion of the day (usually night) to conserve energy. Some animals seasonally go into long periods of inactivity, with reduced body temperature and metabolism, made up of multiple bouts of torpor. This is known as hibernation if it occurs during winter or aestivation if it occurs during the summer. Daily torpor, on the other hand, is not seasonally dependent and can be an important part of energy conservation at any time of year. Torpor is a well-controlled thermoregulatory process and not, as previously thought, the result of switching off thermoregulation. Marsupial torpor differs from non-marsupial mammalian (eutherian) torpor in the characteristics of arousal. Eutherian arousal relies on a heat-producing brown adipose tissue as a mechanism to accelerate rewarming. The mechanism of marsupial arousal is unknown, but appears not to rely on brown adipose tissue. Evolution The evolution of torpor likely accompanied the development of homeothermy. Animals capable of maintaining a body temperature above ambient temperature when other members of its species would not have a fitness advantage. Benefits of maintaining internal temperatures include increased foraging time and less susceptibility to
https://en.wikipedia.org/wiki/Flora%20of%20Borneo
The flora of Borneo include 15 species of dicot tree, 37 species of non-tree dicot and 49 species of monocot endemic to the rich forest of Brunei Darussalam. Borneo is also home to the world's largest flower, the "corpse flower" (Rafflesia arnoldii), which can reach nearly in diameter and up to in weight. Borneo is the third largest island in the world and is divided between three countries: Brunei in the north, the Malaysian constituent states of Sarawak and Sabah, and the 5 Kalimantan provinces of Indonesia (note that in Indonesian, "Kalimantan" refers to the entire island of Borneo). The tallest tropical trees of the world are in Borneo. They are in the family Dipterocarpaceae. See also Biodiversity of Borneo Fauna of Borneo
https://en.wikipedia.org/wiki/NUTS%20statistical%20regions%20of%20Ireland
Ireland uses the Nomenclature of Territorial Units for Statistics (NUTS) geocode standard for referencing country subdivisions for statistical purposes. The standard is developed and regulated by the European Union. The NUTS standard is instrumental in delivering European Structural and Investment Funds. The NUTS code for Ireland is IE and a hierarchy of three levels is established by Eurostat. A further level of geographic organisation, the local administrative unit (LAU), in Ireland is the local electoral area. Overview NUTS levels 1, 2 and 3 The most recent revision of NUTS regions was made in 2016 and took effect in 2018. The eligibility of regions for funding under the European Regional Development Fund and the European Social Fund Plus was revised in 2021. NUTS 2 Regions may be classified as less developed regions, transition regions, or more developed regions. Demographic statistics by NUTS 3 region Local administrative units The local administrative units in Ireland are the local electoral areas. These are subdivisions of local government areas used for local elections. In counties outside Dublin and in the cities and counties, they also form the basis of municipal districts within local authorities. Regional Assemblies Each of the three NUTS 2 regions has a Regional Assembly. These are divided into strategic planning areas, which correspond to the NUTS 3 regions. Prior to 2014, the eight NUTS 3 regions had Regional Authorities. The 2014 act abolished these and transferred their functions to the Regional Assemblies. Assembly members are nominated by constituent local authorities from among their elected councillors. See also ISO 3166-2 codes of Ireland FIPS region codes of Ireland List of Irish regions by Human Development Index Local government in the Republic of Ireland Sources LAU codes Hierarchical list of the Nomenclature of territorial units for statistics - NUTS and the Statistical regions of Europe Overview map of EU Countries - NUTS le
https://en.wikipedia.org/wiki/SpaceChem
SpaceChem is an indie puzzle game developed by Zachtronics Industries, based on principles of automation and chemical bonding. In the game, the player is tasked to produce one or more specific chemical molecules via an assembly line by programming two remote manipulators (called "waldos" in the game) that interact with atoms and molecules through a visual programming language. SpaceChem was the developer's first foray into a commercial title after a number of free Flash-based browser games that feature similar puzzle-based assembly problems. The game was initially released for Microsoft Windows at the start of 2011 via Zachtronics' own website. Though it was initially rejected for sale on the Steam platform, Valve later offered to sell the game after it received high praise from game journalists; further attention came from the game's release alongside one of the Humble Indie Bundles. The game has since been ported to other computing platforms and mobile devices. Reviewers found the game's open-ended problem-solving nature as a highlight of the title. SpaceChem was incorporated into some academic institutions for teaching concepts related to both chemistry and programming. Gameplay In SpaceChem, the player takes the role of a SpaceChem Reactor Engineer whose task is to create circuits through which atoms and molecules flow with the aid of waldos to produce particular batches of chemical shipments for each level. The primary game mode of SpaceChem depicts the internal workings of a Reactor, mapped out to a 10 × 8 regular grid. Each reactor has up to two input and up to two output quadrants, and supports two waldos, red and blue, manipulated through command icons placed on the grid. The player adds commands from an array to direct each waldo independently through the grid. The commands direct the movement of the waldo, to pick up, rotate, and drop atoms and molecules, and to trigger reactor events such as chemical bond formation. The two waldos can also be sy
https://en.wikipedia.org/wiki/Pleated%20surface
In geometry, a pleated surface is roughly a surface that may have simple folds but is not crumpled in more complicated ways. More precisely, a pleated surface is an isometry from a complete hyperbolic surface S to a hyperbolic 3-fold such that every point of S is in the interior of a geodesic that is mapped to a geodesic. They were introduced by , where they were called uncrumpled surfaces. The Universal Book of Mathematics provides the following information about pleated surfaces: It is a surface in Euclidean space or hyperbolic space that resembles a polyhedron in the sense that it has flat faces that meet along edges. Unlike a polyhedron, a pleated surface has no corners, but it may have infinitely many edges that form a lamination.
https://en.wikipedia.org/wiki/Channel%20I/O
In computing, channel I/O is a high-performance input/output (I/O) architecture that is implemented in various forms on a number of computer architectures, especially on mainframe computers. In the past, channels were generally implemented with custom devices, variously named channel, I/O processor, I/O controller, I/O synchronizer, or DMA controller. Overview Many I/O tasks can be complex and require logic to be applied to the data to convert formats and other similar duties. In these situations, the simplest solution is to ask the CPU to handle the logic, but because I/O devices are relatively slow, a CPU could waste time waiting for the data from the device. This situation is called 'I/O bound'. Channel architecture avoids this problem by processing some or all of the I/O task without the aid of the CPU by offloading the work to dedicated logic. Channels are logically self-contained, with sufficient logic and working storage to handle I/O tasks. Some are powerful or flexible enough to be used as a computer on their own and can be construed as a form of coprocessor, for example, the 7909 Data Channel on an IBM 7090 or IBM 7094; however, most are not. On some systems the channels use memory or registers addressable by the central processor as their working storage, while on other systems it is present in the channel hardware. Typically, there are standard interfaces between channels and external peripheral devices, and multiple channels can operate concurrently. A CPU typically designates a block of storage as, or sends, a relatively small channel program to the channel in order to handle I/O tasks, which the channel and controller can, in many cases, complete without further intervention from the CPU (exception: those channel programs which utilize 'program controlled interrupts', PCIs, to facilitate program loading, demand paging and other essential system tasks). When I/O transfer is complete or an error is detected, the controller typically communicates wi
https://en.wikipedia.org/wiki/Hellenic%20Geodetic%20Reference%20System%201987
The Hellenic Geodetic Reference System 1987 or HGRS87 () is a geodetic system commonly used in Greece (SRID=2100). The system specifies a local geodetic datum and a projection system. In some documents it is called Greek Geodetic Reference System 1987 or GGRS87. HGRS87 datum HGRS87 specifies a non-geocentric datum that is tied to the coordinates of the key geodetic station at the Dionysos Satellite Observatory (DSO) northeast of Athens (). The central pedestal (CP) at this location has by definition HGRS87 coordinates 38° 4'  33.8000" N - 23° 55'  51.0000"E, N = +7 m. Although HGRS87 uses the GRS80 ellipsoid, the origin is shifted relative to the GRS80 geocenter, so that the ellipsoidal surface is best for Greece. The specified offsets relative to WGS84 (WGS84-HGRS87) are: Δx = -199.87 m, Δy = 74.79 m, Δz = 246.62 m. The HGRS87 datum is implemented by a first order geodetic network, which consists of approximately 30 triangulation stations throughout Greece and is maintained by the Hellenic Military Geographical Service. The initial uncertainty was estimated as 0.1 ppm (1x10−7). However, there are considerable tectonic movements that move parts of Greece towards different directions, causing incompatibilities between surveys taking place at different times. HGRS87 replaced an earlier de facto geodetic system. The datum of that system was based on the Bessel ellipsoid, with an accurate determination of the geodetic coordinates at the central premises of the National Observatory of Athens 37° 58'  20.1" N - 23° 42'  58.5"E with current Google Earth TM coordinates:37° 58'  20.20" N - 23° 43'  05.36"E and supplemented by an accurately measured azimuth from the observatory to Mount Parnes. Cartographic projections for civilian use were based on the Hatt projection system, with different projection parameters for each 1:100000 map. HGRS87 projection HGRS87 also specifies a transverse Mercator cartographic projection (TM) with m0=0.9996, covering six degrees of long
https://en.wikipedia.org/wiki/Rasterisation
In computer graphics, rasterisation (British English) or rasterization (American English) is the task of taking an image described in a vector graphics format (shapes) and converting it into a raster image (a series of pixels, dots or lines, which, when displayed together, create the image which was represented via shapes). The rasterized image may then be displayed on a computer display, video display or printer, or stored in a bitmap file format. Rasterization may refer to the technique of drawing 3D models, or the conversion of 2D rendering primitives such as polygons, line segments into a rasterized format. Etymology The term "rasterisation" comes . 2D images Line primitives Bresenham's line algorithm is an example of an algorithm used to rasterize lines. Circle primitives Algorithms such as Midpoint circle algorithm are used to render circle onto a pixelated canvas. 3D images Rasterization is one of the typical techniques of rendering 3D models. Compared with other rendering techniques such as ray tracing, rasterization is extremely fast and therefore used in most realtime 3D engines. However, rasterization is simply the process of computing the mapping from scene geometry to pixels and does not prescribe a particular way to compute the color of those pixels. The specific color of each pixel is assigned by a pixel shader (which in modern GPUs is completely programmable). Shading may take into account physical effects such as light position, their approximations or purely artistic intent. The process of rasterizing 3D models onto a 2D plane for display on a computer screen ("screen space") is often carried out by fixed function (non-programmable) hardware within the graphics pipeline. This is because there is no motivation for modifying the techniques for rasterization used at render time and a special-purpose system allows for high efficiency. Triangle rasterization Polygons are a common representation of digital 3D models. Before rasterization, i
https://en.wikipedia.org/wiki/Keyhole%20Markup%20Language
Keyhole Markup Language (KML) is an XML notation for expressing geographic annotation and visualization within two-dimensional maps and three-dimensional Earth browsers. KML was developed for use with Google Earth, which was originally named Keyhole Earth Viewer. It was created by Keyhole, Inc, which was acquired by Google in 2004. KML became an international standard of the Open Geospatial Consortium in 2008. Google Earth was the first program able to view and graphically edit KML files, but other projects such as Marble have added KML support. Structure The KML file specifies a set of features (place marks, images, polygons, 3D models, textual descriptions, etc.) that can be displayed on maps in geospatial software implementing the KML encoding. Every place has a longitude and a latitude. Other data can make a view more specific, such as tilt, heading, or altitude, which together define a "camera view" along with a timestamp or timespan. KML shares some of the same structural grammar as Geography Markup Language (GML). Some KML information cannot be viewed in Google Maps or Mobile. KML files are very often distributed as KMZ files, which are zipped KML files with a .kmz extension. The contents of a KMZ file are a single root KML document (notionally "doc.kml") and optionally any overlays, images, icons, and COLLADA 3D models referenced in the KML including network-linked KML files. The root KML document by convention is a file named "doc.kml" at the root directory level, which is the file loaded upon opening. By convention the root KML document is at root level and referenced files are in subdirectories (e.g. images for overlay). An example KML document is: <?xml version="1.0" encoding="UTF-8"?> <kml xmlns="http://www.opengis.net/kml/2.2"> <Document> <Placemark> <name>New York City</name> <description>New York City</description> <Point> <coordinates>-74.006393,40.714172,0</coordinates> </Point> </Placemark> </Document> </kml> The MIME type associ
https://en.wikipedia.org/wiki/Carboxyatractyloside
Carboxyatractyloside (CATR) is a highly toxic diterpene glycoside that inhibits the ADP/ATP translocase. It is about 10 times more potent than its analog atractyloside. While atractyloside is effective in the inhibition of oxidative phosphorylation, carboxyatractyloside is considered to be more effective. The effects of carboxyatractyloside on the ADP/ATP translocase are not reversed by increasing the concentration of adenine nucleotides, unlike its counterpart atractyloside. Carboxyatractyloside behavior resembles bongkrekic acid while in the mitochondria. Carboxyatractyloside is poisonous to humans as well as livestock, including cows and horses. Symptoms of carboxyatractyloside poisoning may include abdominal pain, nausea and vomiting, drowsiness, palpitations, sweating and trouble breathing. In severe cases, convulsions, liver failure and loss of consciousness may develop, which can lead to death. Carboxyatractyloside can be found in Xanthium species plants, including Xanthium strumarium. Consumption of Xanthium containing the toxin led to the deaths of at least 19 people in Sylhet, Bangladesh during a period of food scarcity. Along with atractyloside, it is also one of the main poisonous substances in the Atractylis gummifera thistle.
https://en.wikipedia.org/wiki/Lyapunov%20redesign
In nonlinear control, the technique of Lyapunov redesign refers to the design where a stabilizing state feedback controller can be constructed with knowledge of the Lyapunov function . Consider the system where is the state vector and is the vector of inputs. The functions , , and are defined for , where is a domain that contains the origin. A nominal model for this system can be written as and the control law stabilizes the system. The design of is called Lyapunov redesign. Further reading Nonlinear control
https://en.wikipedia.org/wiki/Mellified%20man
A mellified man, also known as a human mummy confection, was a legendary medicinal substance created by steeping a human cadaver in honey. The concoction is detailed in Chinese medical sources, including the Bencao Gangmu of the 16th century. Relying on a second-hand account, the text reports a story that some elderly men in Arabia, nearing the end of their lives, would submit themselves to a process of mummification in honey to create a healing confection. This process differed from a simple body donation because of the aspect of self-sacrifice; the mellification process would ideally start before death. The donor would stop eating any food other than honey, going as far as to bathe in the substance. Shortly, the donor's feces and even sweat would consist of honey. When this diet finally proved fatal, the donor's body would be placed in a stone coffin filled with honey. After a century or so, the contents would have turned into a sort of confection reputedly capable of healing broken limbs and other ailments. This confection would then be sold in street markets as a hard to find item with a hefty price. Origins Some of the earliest known records of mellified corpses come from Greek historian Herodotus (4th century BCE) who recorded that the Assyrians used to embalm their dead with honey. A century later, Alexander the Great's body was reportedly preserved in a honey-filled sarcophagus, and there are also indications that this practice was known to the Egyptians. Another record of mellification is found in the Bencao Gangmu (section 52, "Man as medicine") under the entry for munaiyi (木乃伊 "mummy"). It quotes the Chuogeng lu (輟耕錄 "Talks while the Plough is Resting", c. 1366) by the Yuan dynasty scholar Tao Zongyi (陶宗儀) and Tao Jiucheng (陶九成). According to [Tao Jiucheng] in his [Chuogenglu], in the lands of the Arabs there are men 70 or 80 years old who are willing to give their bodies to save others. Such a one takes no more food or drink, only bathing and eatin
https://en.wikipedia.org/wiki/Multiplicity%20%28statistical%20mechanics%29
In statistical mechanics, multiplicity (also called statistical weight) refers to the number of microstates corresponding to a particular macrostate of a thermodynamic system. Commonly denoted , it is related to the configuration entropy of an isolated system via Boltzmann's entropy formula where is the entropy and is Boltzmann's constant. Example: the two-state paramagnet A simplified model of the two-state paramagnet provides an example of the process of calculating the multiplicity of particular macrostate. This model consists of a system of microscopic dipoles which may either be aligned or anti-aligned with an externally applied magnetic field . Let represent the number of dipoles that are aligned with the external field and represent the number of anti-aligned dipoles. The energy of a single aligned dipole is while the energy of an anti-aligned dipole is thus the overall energy of the system is The goal is to determine the multiplicity as a function of ; from there, the entropy and other thermodynamic properties of the system can be determined. However, it is useful as an intermediate step to calculate multiplicity as a function of and This approach shows that the number of available macrostates is . For example, in a very small system with dipoles, there are three macrostates, corresponding to Since the and macrostates require both dipoles to be either anti-aligned or aligned, respectively, the multiplicity of either of these states is 1. However, in the either dipole can be chosen for the aligned dipole, so the multiplicity is 2. In the general case, the multiplicity of a state, or the number of microstates, with aligned dipoles follows from combinatorics, resulting in where the second step follows from the fact that Since the energy can be related to and as follows: Thus the final expression for multiplicity as a function of internal energy is This can be used to calculate entropy in accordance with Boltzmann's entropy f
https://en.wikipedia.org/wiki/Lambert%27s%20cosine%20law
In optics, Lambert's cosine law says that the radiant intensity or luminous intensity observed from an ideal diffusely reflecting surface or ideal diffuse radiator is directly proportional to the cosine of the angle θ between the observer's line of sight and the surface normal; . The law is also known as the cosine emission law or Lambert's emission law. It is named after Johann Heinrich Lambert, from his Photometria, published in 1760. A surface which obeys Lambert's law is said to be Lambertian, and exhibits Lambertian reflectance. Such a surface has the same radiance/luminance when viewed from any angle. This means, for example, that to the human eye it has the same apparent brightness. It has the same radiance because, although the emitted power from a given area element is reduced by the cosine of the emission angle, the solid angle, subtended by surface visible to the viewer, is reduced by the very same amount. Because the ratio between power and solid angle is constant, radiance (power per unit solid angle per unit projected source area) stays the same. Lambertian scatterers and radiators When an area element is radiating as a result of being illuminated by an external source, the irradiance (energy or photons /time/area) landing on that area element will be proportional to the cosine of the angle between the illuminating source and the normal. A Lambertian scatterer will then scatter this light according to the same cosine law as a Lambertian emitter. This means that although the radiance of the surface depends on the angle from the normal to the illuminating source, it will not depend on the angle from the normal to the observer. For example, if the moon were a Lambertian scatterer, one would expect to see its scattered brightness appreciably diminish towards the terminator due to the increased angle at which sunlight hit the surface. The fact that it does not diminish illustrates that the moon is not a Lambertian scatterer, and in fact tends to scatter
https://en.wikipedia.org/wiki/Zone%20theorem
In geometry, the zone theorem is a result that establishes the complexity of the zone of a line in an arrangement of lines. Definition A line arrangement, denoted as , is a subdivision of the plane, induced by a set of lines , into cells (-dimensional faces), edges (-dimensional faces) and vertices (-dimensional faces). Given a set of lines , the line arrangement , and a line (not belonging to ), the zone of is the set of faces intersected by . The complexity of a zone is the total number of edges in its boundary, expressed as a function of . The zone theorem states that said complexity is . History This result was published for the first time in 1985; Chazelle et al. gave the upper bound of for the complexity of the zone of a line in an arrangement. In 1991, this bound was improved to , and it was also shown that this is the best possible upper bound up to a small additive factor. Then, in 2011, Rom Pinchasi proved that the complexity of the zone of a line in an arrangement is at most , and this is a tight bound. Some paradigms used in the different proofs of the theorem are induction, sweep technique, tree construction, and Davenport-Schinzel sequences. Generalizations Although the most popular version is for arrangements of lines in the plane, there exist some generalizations of the zone theorem. For instance, in dimension , considering arrangements of hyperplanes, the complexity of the zone of a hyperplane is the number of facets ( - dimensional faces) bounding the set of cells (-dimensional faces) intersected by . Analogously, the -dimensional zone theorem states that the complexity of the zone of a hyperplane is . There are considerably fewer proofs for the theorem for dimension . For the -dimensional case, there are proofs based on sweep techniques and for higher dimensions is used Euler’s relation: Another generalization is considering arrangements of pseudolines (and pseudohyperplanes in dimension ) instead of lines (and hyperplanes). Some pr
https://en.wikipedia.org/wiki/RNA-targeting%20small%20molecule%20drugs
RNA-targeting small molecules represent a class of small molecules, organic compounds with traditional drug properties (e.g., Lipinski's rule of five) that can bind to RNA secondary or tertiary structures and alter translation patterns, localization, and degradation. Origins Recent discoveries implicating RNA in the pathogenesis of several forms of cancer and neuromuscular diseases have created a paradigm shift in drug discovery. This work combined with advances in structural characterization techniques such as NMR spectroscopy and X-ray crystallography together with computational modeling, has pushed forward the realization that RNA is a dynamic yet viable drug target. Traditionally, RNA was thought to be a mediator between DNA sequence-encoded instructions and functional protein. However, recent reports have shown that there are a large number of non-coding RNAs (ncRNAs) that are not translated into protein. Whereas 85% of the human genome is transcribed into RNA only 3% of the transcripts code for functional protein. Although, ncRNAs do affect gene expression levels by a variety of mechanisms. Further, RNA can adopt discrete secondary or tertiary structures which play a pivotal role in many biological processes and disease pathology. For these reasons, RNA is being recognized as an attractive drug target for small molecules. The earliest attempts to target RNA led to the discovery that aminoglycosides could bind to human RNA. In an early report, Noller discovered that several classes of antibiotics (streptomycin, tetracycline, spectinomycin, edeine, hygromycin, and the neomycins) could "protect" nucleotides in 16S ribosomal RNA by binding to this RNA. Subsequent studies by Schroeder and Green began to plant the seed that RNA could be targeted. Schroeder uncovered that aminoglycosides could inhibit protein synthesis by interacting with the ribosome through interactions with the 3’ end of the 16S RNA of E. coli taking advantage of RNA conformational changes. Gre
https://en.wikipedia.org/wiki/Microsoft%20Assistance%20Markup%20Language
Microsoft Assistance Markup Language (Microsoft AML, generally referred to as MAML) is an XML-based markup language developed by the Microsoft User Assistance Platform team to provide user assistance ("online help") for the Microsoft Windows Vista operating system. It makes up the Assistance Platform on Windows Vista. MAML is also used to provide help information for PowerShell V2 Cmdlets, modules and advanced functions. Concept MAML is a departure from all previous types of user assistance for Windows operating systems. Some of its features have been available in .NET Framework 2, but more options shipped with the release of .NET Framework 3. Previously, user assistance for Windows operating systems used Microsoft Compiled HTML Help files, which contains little machine-readable semantic information. The most significant aspect of MAML is that it shifts the production of user assistance to the concept of structured authoring (somewhat similar to DITA or DocBook). Documents and their constituent elements are defined by their context. With MAML, the emphasis is on content and the tasks a user performs with a computer, not the features of the software. Presentation is managed as part of the rendering engine when a user requests a topic. The structured feature of MAML means that it can express a wide range of active concepts as well. One notable feature is guided help (active content wizard), which allows the help file to either run a task automatically or highlight the parts of the screen a user should interact with step-by-step. This feature was considered a highlight of Project Longhorn, but Microsoft decided against including it in the final Vista release. The MAML authoring structure is divided into segments related to a type of content: conceptual, FAQ, glossary, procedure, reference, reusable content, task, troubleshooting, and tutorial. Presentation Three levels of transformation occur when a topic is displayed: structure, presentation, and rendering:
https://en.wikipedia.org/wiki/Pentagonal%20gyrocupolarotunda
In geometry, the pentagonal gyrocupolarotunda is one of the Johnson solids (). Like the pentagonal orthocupolarotunda (), it can be constructed by joining a pentagonal cupola () and a pentagonal rotunda () along their decagonal bases. The difference is that in this solid, the two halves are rotated 36 degrees with respect to one another. Formulae The following formulae for volume and surface area can be used if all faces are regular, with edge length a:
https://en.wikipedia.org/wiki/Figure%20of%20merit
A figure of merit (FOM) is a performance metric that characterizes the performance of a device, system, or method, relative to its alternatives. Examples Accuracy of a rifle Audio amplifier figures of merit such as gain or efficiency Battery life of a laptop computer Calories per serving Clock rate of a CPU is often given as a figure of merit, but is of limited use in comparing between different architectures. FLOPS may be a better figure, though these too aren't completely representative of the performance of a CPU. Contrast ratio of an LCD Frequency response of a speaker Fill factor of a solar cell Resolution of the image sensor in a digital camera Measure of the detection performance of a sonar system, defined as the propagation loss for which a 50% detection probability is achieved Noise figure of a radio receiver The thermoelectric figure of merit, zT, a material constant proportional to the efficiency of a thermoelectric couple made with the material The figure of merit of digital-to-analog converter, calculated as (power dissipation)/(2ENOB × effective bandwidth) [J/Hz] Luminous efficacy of lighting Profit of a company Residual noise remaining after compensation in an aeromagnetic survey Heat absorption and transfer quality for a solar cooker Computational benchmarks are synthetic figures of merit that summarize the speed of algorithms or computers in performing various typical tasks.
https://en.wikipedia.org/wiki/Eurasian%20Conformity%20mark
The Eurasian Conformity mark (EAC, ) is a certification mark to indicate products that conform to all technical regulations of the Eurasian Customs Union. It means that the EAC-marked products meet all requirements of the corresponding technical regulations and have passed all conformity assessment procedures. Before this marking came into use, the prevalent product marking was GOST R 50460-92: Mark of conformity for mandatory certification. History Originally, the conformity assessment of the products, which were brought on the market of the USSR or Russia for the first time, was defined by GOST and TR standards. From July 1, 2010, the Customs Code of the Customs Union between Russia, Belarus and Kazakhstan entered into force. The new Customs Code should replace the existing national certification rules such as Russian GOST-R, TR or Kazakh GOST-K standards by the uniform technical regulations and technical guidelines of the Customs Union. Products, which thus meet the requirements of the technical regulations of the Customs Union or the Eurasian Economic Union (EAEU) (consisting of Russia, Belarus, Armenia, Kazakhstan, Kyrgyzstan), are marked with the EAC mark. The uniform EAC marking of the customs union was defined by the board decision of the EAEU Commission No. 711 of July 15, 2011. The term TR marking in conjunction with the conformity assurance with the technical regulations of the CU or EAEU is therefore incorrect. Since February 2010, the new list of products subject to certification and declaration has been valid for the GOST-regulated area. In accordance with Federal Law No. 385 of December 30, 2009 On amendments to the Federal Law on Technical Regulation, the original deadline for 2010 was lifted on the personal request of the former President of Russia, Dmitry Medvedev, and the following items were ordered: To simplify the procedure for the introduction or modification of the TR in Russia (Ministry of Industry and Trade) To permit the application
https://en.wikipedia.org/wiki/WHOIS
WHOIS (pronounced as the phrase "who is") is a query and response protocol that is used for querying databases that store an Internet resource's registered users or assignees. These resources include domain names, IP address blocks and autonomous systems, but it is also used for a wider range of other information. The protocol stores and delivers database content in a human-readable format. The current iteration of the WHOIS protocol was drafted by the Internet Society, and is documented in . Whois is also the name of the command-line utility on most UNIX systems used to make WHOIS protocol queries. In addition, WHOIS has a sister protocol called Referral Whois (RWhois). History Elizabeth Feinler and her team (who had created the Resource Directory for ARPANET) were responsible for creating the first WHOIS directory in the early 1970s. Feinler set up a server in Stanford's Network Information Center (NIC) which acted as a directory that could retrieve relevant information about people or entities. She and the team created domains, with Feinler's suggestion that domains be divided into categories based on the physical address of the computer. The process of registration was established in . WHOIS was standardized in the early 1980s to look up domains, people, and other resources related to domain and number registrations. As all registration was done by one organization at that time, one centralized server was used for WHOIS queries. This made looking up such information very easy. At the time of the emergence of the internet from the ARPANET, the only organization that handled all domain registrations was the Defense Advanced Research Projects Agency (DARPA) of the United States government (created during 1958.). The responsibility of domain registration remained with DARPA as the ARPANET became the Internet during the 1980s. UUNET began offering domain registration service; however, they simply handled the paperwork which they forwarded to the DARPA Network In
https://en.wikipedia.org/wiki/Aitken%27s%20delta-squared%20process
In numerical analysis, Aitken's delta-squared process or Aitken extrapolation is a series acceleration method, used for accelerating the rate of convergence of a sequence. It is named after Alexander Aitken, who introduced this method in 1926. Its early form was known to Seki Kōwa (end of 17th century) and was found for rectification of the circle, i.e. the calculation of π. It is most useful for accelerating the convergence of a sequence that is converging linearly. Definition Given a sequence , one associates with this sequence the new sequence which can, with improved numerical stability, also be written as or equivalently as where and for Obviously, is ill-defined if contains a zero element, or equivalently, if the sequence of first differences has a repeating term. From a theoretical point of view, if that occurs only for a finite number of indices, one could easily agree to consider the sequence restricted to indices with a sufficiently large . From a practical point of view, one does in general rather consider only the first few terms of the sequence, which usually provide the needed precision. Moreover, when numerically computing the sequence, one has to take care to stop the computation when rounding errors in the denominator become too large, where the Δ2 operation may cancel too many significant digits. (It would be better for numerical calculation to use rather than Properties Aitken's delta-squared process is a method of acceleration of convergence, and a particular case of a nonlinear sequence transformation. Convergence of to limit is called "linear" if there is some number for which Which means that the distance between the sequence and its limit shrinks by nearly the same proportion on every step, and that rate of reduction becomes closer to being constant with every step. (This is also called "geometric convergence"; this form of convergence is common for power series.) Aitken's method will accelerate the sequence if is no
https://en.wikipedia.org/wiki/Strobogrammatic%20number
A strobogrammatic number is a number whose numeral is rotationally symmetric, so that it appears the same when rotated 180 degrees. In other words, the numeral looks the same right-side up and upside down (e.g., 69, 96, 1001). A strobogrammatic prime is a strobogrammatic number that is also a prime number, i.e., a number that is only divisible by one and itself (e.g., 11). It is a type of ambigram, words and numbers that retain their meaning when viewed from a different perspective, such as palindromes. Description When written using standard characters (ASCII), the numbers, 0, 1, 8 are symmetrical around the horizontal axis, and 6 and 9 are the same as each other when rotated 180 degrees. In such a system, the first few strobogrammatic numbers are: 0, 1, 8, 11, 69, 88, 96, 101, 111, 181, 609, 619, 689, 808, 818, 888, 906, 916, 986, 1001, 1111, 1691, 1881, 1961, 6009, 6119, 6699, 6889, 6969, 8008, 8118, 8698, 8888, 8968, 9006, 9116, 9696, 9886, 9966, ... The first few strobogrammatic primes are: 11, 101, 181, 619, 16091, 18181, 19861, 61819, 116911, 119611, 160091, 169691, 191161, 196961, 686989, 688889, ... The years 1881 and 1961 were the most recent strobogrammatic years; the next strobogrammatic year will be 6009. Although amateur aficionados of mathematics are quite interested in this concept, professional mathematicians generally are not. Like the concept of repunits and palindromic numbers, the concept of strobogrammatic numbers is base-dependent (expanding to base-sixteen, for example, produces the additional symmetries of 3/E; some variants of duodecimal systems also have this and a symmetrical x). Unlike palindromes, it is also font dependent. The concept of strobogrammatic numbers is not neatly expressible algebraically, the way that the concept of repunits is, or even the concept of palindromic numbers. Nonstandard systems The strobogrammatic properties of a given number vary by typeface. For instance, in an ornate serif type, the numbers 2 and 7
https://en.wikipedia.org/wiki/International%20Workshop%20on%20First-Order%20Theorem%20Proving
The International Workshop on First-Order Theorem Proving (FTP) is a scientific meeting of researchers interested in automated theorem proving for first-order logic and related fields. FTP workshops are less formal than many conferences, but more formal than most workshops. While FTP proceedings are published informally, most FTP workshops have resulted in a special issue of a recognized peer-reviewed academic journal. FTP is one of the constituent meetings of the International Joint Conference on Automated Reasoning, and is merged with this conference in years where it takes place. FTP meetings FTP '97 took place at the Research Institute for Symbolic Computation in Linz, Austria in Austria, October 27–28, 1997 FTP '98 took place in Vienna, Austria, November 23–25, 1998 FTP '00 took place in St. Andrews, Scotland, July 3–5, 2000 In 2001 FTP was merged into the first IJCAR FTP '03 took place in Valencia, Spain, June 12–14, 2003, as part of the Federated Conference on Rewriting, Deduction and Programming. In 2004, FTP was merged into the second IJCAR FTP '05 took place in Koblenz, Germany, September 14–17, 2005 In 2006, FTP was merged into the third IJCAR FTP '07 took place in Liverpool, United Kingdom, September 12–13, 2007 In 2008, FTP was merged into the fourth IJCAR FTP '09 took place in Oslo, Norway, July 6–7, 2009 In 2010, FTP was merged into the fifth IJCAR FTP '11 took place in Bern, Switzerland, June 13, 2011 Theoretical computer science conferences Logic conferences
https://en.wikipedia.org/wiki/Metre%20sea%20water
The metre (or meter) sea water (msw) is a metric unit of pressure used in underwater diving. It is defined as one tenth of a bar. The unit used in the US is the foot sea water (fsw), based on standard gravity and a sea-water density of 64 lb/ft3. According to the US Navy Diving Manual, one fsw equals 0.30643 msw, , or , though elsewhere it states that 33 fsw is (one atmosphere), which gives one fsw equal to about 0.445 psi. The msw and fsw are the conventional units for measurement of diver pressure exposure used in decompression tables and the unit of calibration for pneumofathometers and hyperbaric chamber pressure gauges. Feet of sea water One atmosphere is approximately equal to 33 feet of sea water or 14.7 psi, which gives 4.9/11 or about 0.445 psi per foot. Atmospheric pressure may be considered constant at sea level, and minor fluctuations caused by the weather are usually ignored. Pressures measured in fsw and msw are gauge pressure, relative to the surface pressure of 1 atm absolute, except when a pressure difference is measured between the locks of a hyperbaric chamber, which is also generally measured in fsw and msw. The pressure of seawater at a depth of 33 feet equals one atmosphere. The absolute pressure at 33 feet depth in sea water is the sum of atmospheric and hydrostatic pressure for that depth, and is 66 fsw, or two atmospheres absolute. For every additional 33 feet of depth, another atmosphere of pressure accumulates. Therefore at the surface the gauge pressure of 0 fsw is equivalent to an absolute pressure of , and the gauge pressure in fsw at any depth is incremented by 1 ata to provide absolute pressure. (Pressure in ata = Depth in feet/33 + 1) Usage In diving the absolute pressure is used in most computations, particularly for decompression and breathing gas consumption but depth is measured by way of hydrostatic pressure. In metric units the ambient pressure is usually measured in metres sea water (msw), and converted to bar for calc
https://en.wikipedia.org/wiki/Rubik%27s%20Magic
Rubik's Magic, like the Rubik's Cube, is a mechanical puzzle invented by Ernő Rubik and first manufactured by Matchbox in the mid-1980s. The puzzle consists of eight black square tiles (changed to red squares with goldish rings in 1997) arranged in a 2 × 4 rectangle; diagonal grooves on the tiles hold wires that connect them, allowing them to be folded onto each other and unfolded again in two perpendicular directions (assuming that no other connections restrict the movement) in a manner similar to a Jacob's ladder toy. The front side of the puzzle shows, in the initial state, three separate, rainbow-coloured rings; the back side consists of a scrambled picture of three interconnected rings. The goal of the game is to fold the puzzle into a heart-like shape and unscramble the picture on the back side, thus interconnecting the rings. Numerous ways to accomplish this exist, and experienced players can transform the puzzle from its initial into the solved state in less than 2 seconds. Other challenges for Rubik's Magic include reproducing given shapes (which are often three-dimensional), sometimes with certain tiles required to be in certain positions and/or orientations. History Rubik's Magic was first manufactured by Matchbox in 1986. Professor Rubik holds both a Hungarian patent (HU 1211/85, issued 19 March 1985) and a US patent (US 4,685,680, issued 11 August 1987) on the mechanism of Rubik's Magic. In 1987, Rubik's Magic: Master Edition was published by Matchbox; it consisted of 12 silver tiles arranged in a 2 × 6 rectangle, showing 5 interlinked rings that had to be unlinked by transforming the puzzle into a shape reminiscent of a W. Around the same time, Matchbox also produced Rubik's Magic Create the Cube, a "Level Two" version of Rubik's Magic, in which the puzzle is solved when folded into a cube with a base of two tiles, and the tile colors match at the corners of the cube. It did not have as wide a release, and is rare to find. In 1996, the original v
https://en.wikipedia.org/wiki/Feza%20G%C3%BCrsey%20Institute
Feza Gürsey Institute () is a joint institute of Boğaziçi University and TÜBİTAK (Scientific and Technological Research Council of Turkey) on physics research, founded in 1983 by Erdal İnönü with the name Research Institute for Basic Sciences. It now continues as the Feza Gürsey Institute, having been renamed in honor of Feza Gürsey, a distinguished Turkish physicist. The institute is located within the Kandilli Campus of the Boğaziçi University in Istanbul, Turkey. Currently it hosts researchers in mathematics and theoretical physics. External links Feza Gürsey Institute, official website of the institute Physics research institutes Research institutes in Turkey Scientific and Technological Research Council of Turkey Boğaziçi University
https://en.wikipedia.org/wiki/NASU%20Institute%20of%20Cryobiology%20and%20Cryomedicine%20Issues
The Institute for Problems of Cryobiology and Cryomedicine in Kharkiv is one of the institutes of the National Academy of Science of Ukraine, and is the largest institute devoted to cryobiology research in the world. Background Established in 1972, the focus of the research is on cryoinjury, cryosurgery, cryopreservation, lyophilization and hypothermia. Since 1985 the Institute has published the open access peer-reviewed scientific journal Problems of Cryobiology and Cryomedicine. See also Cryobiology National Academy of Science of Ukraine
https://en.wikipedia.org/wiki/Manure-derived%20synthetic%20crude%20oil
Manure-derived synthetic crude oil is a synthetic bio-oil chemically engineered (converted) from animal or human manure. Research into the production of manure-derived synthetic fuel began with pig manure in 1996 at the University of Illinois at Urbana–Champaign by the research team led by professors Yuanhui Zhang and Lance Schideman. They developed a method for converting raw pig manure into bio-oil through thermal depolymerization (thermochemical conversion). This process uses a thermochemical conversion reactor to apply heat and pressure for breaking down carbohydrate materials. As a result, bio-oil, methane and carbon dioxide are produced. With further research, large-scale chemical processing in a refinery-style environment could help process millions of gallons of "pig biocrude" per day. However, this technology is still in its infancy and could produce only of oil per of manure. In 2006, preparations for a construction of a pilot plant started. It is developed by Snapshot Energy, a start-up firm. According to the tests conducted by the National Institute of Standards and Technology pig manure biocrude produced by current technology contains 15% water, sulfur and char waste containing heavy metals, which should be removed to improve the quality of oil. See also Alternative fuels Energy and the environment Poultry litter
https://en.wikipedia.org/wiki/Yerevan%20Physics%20Institute
The A.I. Alikhanyan National Science Laboratory () is a research institute located in Yerevan, Armenia. It was founded in 1943 as a branch of the Yerevan State University by brothers Abram Alikhanov and Artem Alikhanian. It was often referred to by the acronym YerPhI (Yerevan Physics Institute). In 2011 it was renamed to its current name A.I. Alikhanyan National Science Laboratory. History and strategy The Yerevan Physics Institute was founded in 1943 as a branch of the Yerevan State University by brothers Abraham Alikhanov and Artem Alikhanian. Later two high-altitude cosmic ray stations were founded on Mount Aragats (3,200 m) and Nor Amberd (2,000 m). In 1963 the institute was transferred to the Soviet Union Atomic Energy State Committee. The construction of a 6 GeV electron synchrotron accomplished in 1967 became an important landmark in the history of institute, it is the first particle accelerator in Armenia (Arus "ԱՐՈՒՍ"). After collapse of Soviet Union YerPhI continued research in the fields of high-energy physics and astrophysics in Armenia and worldwide using world biggest accelerators and cosmic ray detectors. Now YerPhI get status of A. Alikhanyan National Laboratory. Brief summary of scientific activities Among the key results of YerPhI in the early years were the discovery of protons and neutrons in cosmic rays, and the establishment of the first evidence of existence of the particles with masses between that of muons and protons. The high altitude research stations have remained the main research base of the Cosmic Ray Division (CRD) of YerPhI until now. Among the CRD achievements there were: discovery of sharp knee in light components of primary cosmic rays, detection of the highest energy protons accelerated on the Sun, and the creation of the Aragats Space environ¬mental Center in 2000 for studies of the solar-terrestrial connection, where CRD becomes one of the world's leaders. The 6 GeV electron synchrotron was accomplished in 1967. During 197
https://en.wikipedia.org/wiki/Reconfigurable%20computing
Reconfigurable computing is a computer architecture combining some of the flexibility of software with the high performance of hardware by processing with very flexible high speed computing fabrics like field-programmable gate arrays (FPGAs). The principal difference when compared to using ordinary microprocessors is the ability to make substantial changes to the datapath itself in addition to the control flow. On the other hand, the main difference from custom hardware, i.e. application-specific integrated circuits (ASICs) is the possibility to adapt the hardware during runtime by "loading" a new circuit on the reconfigurable fabric. History The concept of reconfigurable computing has existed since the 1960s, when Gerald Estrin's paper proposed the concept of a computer made of a standard processor and an array of "reconfigurable" hardware. The main processor would control the behavior of the reconfigurable hardware. The latter would then be tailored to perform a specific task, such as image processing or pattern matching, as quickly as a dedicated piece of hardware. Once the task was done, the hardware could be adjusted to do some other task. This resulted in a hybrid computer structure combining the flexibility of software with the speed of hardware. In the 1980s and 1990s there was a renaissance in this area of research with many proposed reconfigurable architectures developed in industry and academia, such as: Copacobana, Matrix, GARP, Elixent, NGEN, Polyp, MereGen, PACT XPP, Silicon Hive, Montium, Pleiades, Morphosys, and PiCoGA. Such designs were feasible due to the constant progress of silicon technology that let complex designs be implemented on one chip. Some of these massively parallel reconfigurable computers were built primarily for special subdomains such as molecular evolution, neural or image processing. The world's first commercial reconfigurable computer, the Algotronix CHS2X4, was completed in 1991. It was not a commercial success, but was pro
https://en.wikipedia.org/wiki/Einstein%20ring
An Einstein ring, also known as an Einstein–Chwolson ring or Chwolson ring (named for Orest Chwolson), is created when light from a galaxy or star passes by a massive object en route to the Earth. Due to gravitational lensing, the light is diverted, making it seem to come from different places. If source, lens, and observer are all in perfect alignment (syzygy), the light appears as a ring. Introduction Gravitational lensing is predicted by Albert Einstein's theory of general relativity. Instead of light from a source traveling in a straight line (in three dimensions), it is bent by the presence of a massive body, which distorts spacetime. An Einstein Ring is a special case of gravitational lensing, caused by the exact alignment of the source, lens, and observer. This results in symmetry around the lens, causing a ring-like structure. The size of an Einstein ring is given by the Einstein radius. In radians, it is where is the gravitational constant, is the mass of the lens, is the speed of light, is the angular diameter distance to the lens, is the angular diameter distance to the source, and is the angular diameter distance between the lens and the source. Over cosmological distances in general. History The bending of light by a gravitational body was predicted by Albert Einstein in 1912, a few years before the publication of general relativity in 1916 (Renn et al. 1997). The ring effect was first mentioned in the academic literature by Orest Khvolson in a short article in 1924, in which he mentioned the “halo effect” of gravitation when the source, lens, and observer are in near-perfect alignment. Einstein remarked upon this effect in 1936 in a paper prompted by a letter by a Czech engineer, R W Mandl, but stated (In this statement, β is the Einstein Radius currently denoted by as in the expression above.) However, Einstein was only considering the chance of observing Einstein rings produced by stars, which is low – the chance of observing
https://en.wikipedia.org/wiki/Sparrowiq
SparrowIQ is a packet-based traffic analysis and network performance monitoring solution that provides network managers with near real-time traffic visibility into network usage based on conversations, applications, users and class of service. The product was developed by Solana Networks (Ottawa, Ontario, Canada) to allow smaller businesses to gain access to flow-based network traffic monitoring solutions - normally too complex or unaffordable. SparrowIQ was awarded the "Best New Product" by the ASCII Group in June 2015 and "Strong Value" award by Enterprise Management Associates in 2013. Features SparrowIQ key features Traffic Forensics for identifying business-relevant versus recreational and unauthorized traffic Real-time Traffic Alerting for instant notification when bandwidth crosses preset thresholds Automatic Report Generation and Distribution
https://en.wikipedia.org/wiki/De%20Franchis%20theorem
In mathematics, the de Franchis theorem is one of a number of closely related statements applying to compact Riemann surfaces, or, more generally, algebraic curves, X and Y, in the case of genus g > 1. The simplest is that the automorphism group of X is finite (see though Hurwitz's automorphisms theorem). More generally, the set of non-constant morphisms from X to Y is finite; fixing X, for all but a finite number of such Y, there is no non-constant morphism from X to Y. These results are named for (1875–1946). It is sometimes referenced as the De Franchis-Severi theorem. It was used in an important way by Gerd Faltings to prove the Mordell conjecture. See also Castelnuovo–de Franchis theorem