source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Bipartite%20realization%20problem
The bipartite realization problem is a classical decision problem in graph theory, a branch of combinatorics. Given two finite sequences and of natural numbers, the problem asks whether there is a labeled simple bipartite graph such that is the degree sequence of this bipartite graph. Solutions The problem belongs to the complexity class P. This can be proven using the Gale–Ryser theorem, i.e., one has to validate the correctness of inequalities. Other notations The problem can also be stated in terms of zero-one matrices. The connection can be seen if one realizes that each bipartite graph has a biadjacency matrix where the column sums and row sums correspond to and . The problem is then often denoted by 0-1-matrices for given row and column sums. In the classical literature the problem was sometimes stated in the context of contingency tables by contingency tables with given marginals. A third formulation is in terms of degree sequences of simple directed graphs with at most one loop per vertex. In this case the matrix is interpreted as the adjacency matrix of such a directed graph. When are pairs of non-negative integers ((a1,b1), ..., (an,bn)) the indegree-outdegree pairs of a labeled directed graph with at most one loop per vertex? Related problems Similar problems describe the degree sequences of simple graphs and simple directed graphs. The first problem is the so-called graph realization problem. The second is known as the digraph realization problem. The bipartite realization problem is equivalent to the question, if there exists a labeled bipartite subgraph of a complete bipartite graph to a given degree sequence. The hitchcock problem asks for such a subgraph minimizing the sum of the costs on each edge which are given for the complete bipartite graph. A further generalization is the f-factor problem for bipartite graphs, i.e. for a given bipartite graph one searches for a subgraph possessing a certain degree sequence. The problem uniform sampling
https://en.wikipedia.org/wiki/Tagged%20Deterministic%20Finite%20Automaton
In the automata theory, a tagged deterministic finite automaton (TDFA) is an extension of deterministic finite automaton (DFA). In addition to solving the recognition problem for regular languages, TDFA is also capable of submatch extraction and parsing. While canonical DFA can find out if a string belongs to the language defined by a regular expression, TDFA can also extract substrings that match specific subexpressions. More generally, TDFA can identify positions in the input string that match tagged positions in a regular expression (tags are meta-symbols similar to capturing parentheses, but without the pairing requirement). History TDFA were first described by Ville Laurikari in 2000. Prior to that it was unknown whether it is possible to perform submatch extraction in one pass on a deterministic finite-state automaton, so this paper was an important advancement. Laurikari described TDFA construction and gave a proof that the determinization process terminates, however the algorithm did not handle disambiguation correctly. In 2007 Chris Kuklewicz implemented TDFA in a Haskell library Regex-TDFA with POSIX longest-match semantics. Kuklewicz gave an informal description of the algorithm and answered the principal question whether TDFA are capable of POSIX longest-match disambiguation, which was doubted by other researchers. In 2017 Ulya Trafimovich described TDFA with one-symbol lookahead. The use of a lookahead symbol reduces the number of registers and register operations in a TDFA, which makes it faster and often smaller than Laurikari TDFA. Trafimovich called TDFA variants with and without lookahead TDFA(1) and TDFA(0) by analogy with LR parsers LR(1) and LR(0). The algorithm was implemented in the open-source lexer generator RE2C. Trafimovich formalized Kuklewicz disambiguation algorithm. In 2018 Angelo Borsotti worked on an experimental Java implementation of TDFA; it was published later in 2021. In 2019 Borsotti and Trafimovich adapted POSIX di
https://en.wikipedia.org/wiki/Cleavage%20stimulation%20factor
Cleavage stimulatory factor or cleavage stimulation factor (CstF or CStF) is a heterotrimeric protein, made up of the proteins CSTF1 (55kDa), CSTF2 (64kDa) and CSTF3 (77kDa), totalling about 200 kDa. It is involved in the cleavage of the 3' signaling region from a newly synthesized pre-messenger RNA (mRNA) molecule. CstF is recruited by cleavage and polyadenylation specificity factor (CPSF) and assembles into a protein complex on the 3' end to promote the synthesis of a functional polyadenine tail, which results in a mature mRNA molecule ready to be exported from the cell nucleus to the cytosol for translation. The amount of CstF in a cell is dependent on the phase of the cell cycle, increasing significantly during the transition from G0 phase to S phase in mouse fibroblast and human splenic B cells. Genes CSTF1, CSTF2 or CSTF2T, CSTF3
https://en.wikipedia.org/wiki/Reptilase%20time
Reptilase time (RT) is a blood test used to detect deficiency or abnormalities in fibrinogen, especially in cases of heparin contamination. Reptilase, an enzyme found in the venom of Bothrops snakes, has activity similar to thrombin. Unlike thrombin, reptilase is resistant to inhibition by antithrombin III. Thus, the reptilase time is not prolonged in blood samples containing heparin, hirudin, or direct thrombin inhibitors, whereas the thrombin time will be prolonged in these samples. Reptilase also differs from thrombin by releasing fibrinopeptide A, but not fibrinopeptide B, in its cleavage of fibrinogen.
https://en.wikipedia.org/wiki/Workspace%20virtualization
Workspace virtualization is a way of distributing applications to client computers using application virtualization; however, it also bundles several applications together into one complete workspace. Overview Workspace virtualization is an approach that encapsulates and isolates an entire computing workspace. At a minimum, the workspace consists of everything above the operating system kernel – applications, data, settings, and any non-privileged operating system subsystems required to provide a functional desktop computing environment. By doing this, applications within the workspace can interact with each other, enabling them to do some of the things users are accustomed to and which are missing in application virtualization such as embedding a Microsoft Excel worksheet into a Microsoft Word document. Further, the workspace contains applications settings and user data enabling the user to move to a different operating system or to a different computer and still preserve applications, settings and data in one complete working unit. For deeper workspace virtualization, the virtualization engine implementation virtualizes privileged code modules and full operating system subsystems through a kernel-mode Workspace Virtualization Engine (WVE). Advantages and disadvantages Workspace Virtualization vs Application Virtualization Workspace virtualization enables individual applications to interact with each other and also enables user settings/configuration and user data to stay within the workspace. Application virtualization does not. Application Virtualization shields independent applications from each other better should one of them prove to be hostile (i.e. contains a virus of some sort). Workspace Virtualization vs Desktop Virtualization Workspace Virtualization runs directly on the client computer hardware whereas Desktop Virtualization in many cases runs on a remote computer somewhere over a corporate LAN/WAN or over the Internet (called Hosted Desktop Virt
https://en.wikipedia.org/wiki/Boole%20polynomials
In mathematics, the Boole polynomials sn(x) are polynomials given by the generating function , . See also Umbral calculus Peters polynomials, a generalization of Boole polynomials.
https://en.wikipedia.org/wiki/Platial
Platial was a collaborative, user generated, cartographic website. It was founded in 2004 by Di-Ann Eisnor, Jake Olsen, and Jason Wilson. Platial, Inc. was headquartered in Portland, Oregon. Platial was used by hundreds of thousands of people globally to share and associate relevant content with geographic places of interest. Platial claims 25% of the web's map widgets, used on over 10,000 sites. On October 18, 2007, Platial acquired Frappr!, another social mapping website, increasing its total user-generated locations to more than 100 million. On February 27, 2010 Platial announced that the service would be closing down within a few days. In the Geographic Information Science literature, the concept of 'platial' refers to place-based studies in parallel with the use of 'spatial' for space-based (coordinate-system oriented) studies.
https://en.wikipedia.org/wiki/Greater%20palatine%20nerve
The greater palatine nerve is a branch of the pterygopalatine ganglion. This nerve is also referred to as the anterior palatine nerve, due to its location anterior to the lesser palatine nerve. It carries both general sensory fibres from the maxillary nerve, and parasympathetic fibers from the nerve of the pterygoid canal. It may be anaesthetised for procedures of the mouth and maxillary (upper) teeth. Structure The greater palatine nerve is a branch of the pterygopalatine ganglion. It descends through the greater palatine canal, moving anteriorly and inferiorly. Here, it is accompanied by the descending palatine artery. It emerges upon the hard palate through the greater palatine foramen. It then passes forward in a groove in the hard palate, nearly as far as the incisor teeth. While in the pterygopalatine canal, it gives off lateral posterior inferior nasal branches, which enter the nasal cavity through openings in the palatine bone, and ramify over the inferior nasal concha and middle and inferior meatuses. At its exit from the canal, a palatine branch is distributed to both surfaces of the soft palate. Function The greater palatine nerve carries both general sensory fibres from the maxillary nerve, and parasympathetic fibers from the nerve of the pterygoid canal. It supplies the gums, the mucous membrane and glands of the hard palate, and communicates in front with the terminal filaments of the nasopalatine nerve. Clinical significance The greater palatine nerve may be anaesthetised to perform dental procedures on the maxillary (upper) teeth, and sometimes for cleft lip and cleft palate surgery.
https://en.wikipedia.org/wiki/Non-convexity%20%28economics%29
In economics, non-convexity refers to violations of the convexity assumptions of elementary economics. Basic economics textbooks concentrate on consumers with convex preferences (that do not prefer extremes to in-between values) and convex budget sets and on producers with convex production sets; for convex models, the predicted economic behavior is well understood. When convexity assumptions are violated, then many of the good properties of competitive markets need not hold: Thus, non-convexity is associated with market failures, where supply and demand differ or where market equilibria can be inefficient. Non-convex economies are studied with nonsmooth analysis, which is a generalization of convex analysis. Demand with many consumers If a preference set is non-convex, then some prices determine a budget-line that supports two separate optimal-baskets. For example, we can imagine that, for zoos, a lion costs as much as an eagle, and further that a zoo's budget suffices for one eagle or one lion. We can suppose also that a zoo-keeper views either animal as equally valuable. In this case, the zoo would purchase either one lion or one eagle. Of course, a contemporary zoo-keeper does not want to purchase half of an eagle and half of a lion. Thus, the zoo-keeper's preferences are non-convex: The zoo-keeper prefers having either animal to having any strictly convex combination of both. When the consumer's preference set is non-convex, then (for some prices) the consumer's demand is not connected; A disconnected demand implies some discontinuous behavior by the consumer, as discussed by Harold Hotelling: If indifference curves for purchases be thought of as possessing a wavy character, convex to the origin in some regions and concave in others, we are forced to the conclusion that it is only the portions convex to the origin that can be regarded as possessing any importance, since the others are essentially unobservable. They can be detected only by the discontinuities
https://en.wikipedia.org/wiki/Detailed%20balance
The principle of detailed balance can be used in kinetic systems which are decomposed into elementary processes (collisions, or steps, or elementary reactions). It states that at equilibrium, each elementary process is in equilibrium with its reverse process. History The principle of detailed balance was explicitly introduced for collisions by Ludwig Boltzmann. In 1872, he proved his H-theorem using this principle. The arguments in favor of this property are founded upon microscopic reversibility. Five years before Boltzmann, James Clerk Maxwell used the principle of detailed balance for gas kinetics with the reference to the principle of sufficient reason. He compared the idea of detailed balance with other types of balancing (like cyclic balance) and found that "Now it is impossible to assign a reason" why detailed balance should be rejected (pg. 64). Albert Einstein in 1916 used the principle of detailed balance in a background for his quantum theory of emission and absorption of radiation. In 1901, Rudolf Wegscheider introduced the principle of detailed balance for chemical kinetics. In particular, he demonstrated that the irreversible cycles A1 -> A2 -> \cdots -> A_\mathit{n} -> A1 are impossible and found explicitly the relations between kinetic constants that follow from the principle of detailed balance. In 1931, Lars Onsager used these relations in his works, for which he was awarded the 1968 Nobel Prize in Chemistry. The principle of detailed balance has been used in Markov chain Monte Carlo methods since their invention in 1953. In particular, in the Metropolis–Hastings algorithm and in its important particular case, Gibbs sampling, it is used as a simple and reliable condition to provide the desirable equilibrium state. Now, the principle of detailed balance is a standard part of the university courses in statistical mechanics, physical chemistry, chemical and physical kinetics. Microscopic background The microscopic "reversing of time" turns at
https://en.wikipedia.org/wiki/Main%20conjecture%20of%20Iwasawa%20theory
In mathematics, the main conjecture of Iwasawa theory is a deep relationship between p-adic L-functions and ideal class groups of cyclotomic fields, proved by Kenkichi Iwasawa for primes satisfying the Kummer–Vandiver conjecture and proved for all primes by . The Herbrand–Ribet theorem and the Gras conjecture are both easy consequences of the main conjecture. There are several generalizations of the main conjecture, to totally real fields, CM fields, elliptic curves, and so on. Motivation was partly motivated by an analogy with Weil's description of the zeta function of an algebraic curve over a finite field in terms of eigenvalues of the Frobenius endomorphism on its Jacobian variety. In this analogy, The action of the Frobenius corresponds to the action of the group Γ. The Jacobian of a curve corresponds to a module X over Γ defined in terms of ideal class groups. The zeta function of a curve over a finite field corresponds to a p-adic L-function. Weil's theorem relating the eigenvalues of Frobenius to the zeros of the zeta function of the curve corresponds to Iwasawa's main conjecture relating the action of the Iwasawa algebra on X to zeros of the p-adic zeta function. History The main conjecture of Iwasawa theory was formulated as an assertion that two methods of defining p-adic L-functions (by module theory, by interpolation) should coincide, as far as that was well-defined. This was proved by for Q, and for all totally real number fields by . These proofs were modeled upon Ken Ribet's proof of the converse to Herbrand's theorem (the Herbrand–Ribet theorem). Karl Rubin found a more elementary proof of the Mazur–Wiles theorem by using Thaine's method and Kolyvagin's Euler systems, described in and , and later proved other generalizations of the main conjecture for imaginary quadratic fields. In 2014, Christopher Skinner and Eric Urban proved several cases of the main conjectures for a large class of modular forms. As a consequence, for a modular ellip
https://en.wikipedia.org/wiki/Heneage%20knot
The Heneage knot is a decorative heraldic knot, the badge of the Heneage family of Lincolnshire, England. It was awarded to Sir Thomas Heneage by Queen Elizabeth I in 1594.
https://en.wikipedia.org/wiki/Algebraic%20element
In mathematics, if is a field extension of , then an element of is called an algebraic element over , or just algebraic over , if there exists some non-zero polynomial with coefficients in such that . Elements of which are not algebraic over are called transcendental over . These notions generalize the algebraic numbers and the transcendental numbers (where the field extension is , being the field of complex numbers and being the field of rational numbers). Examples The square root of 2 is algebraic over , since it is the root of the polynomial whose coefficients are rational. Pi is transcendental over but algebraic over the field of real numbers : it is the root of , whose coefficients (1 and −) are both real, but not of any polynomial with only rational coefficients. (The definition of the term transcendental number uses , not .) Properties The following conditions are equivalent for an element of : is algebraic over , the field extension is algebraic, i.e. every element of is algebraic over (here denotes the smallest subfield of containing and ), the field extension has finite degree, i.e. the dimension of as a -vector space is finite, , where is the set of all elements of that can be written in the form with a polynomial whose coefficients lie in . To make this more explicit, consider the polynomial evaluation . This is a homomorphism and its kernel is . If is algebraic, this ideal contains non-zero polynomials, but as is a euclidean domain, it contains a unique polynomial with minimal degree and leading coefficient , which then also generates the ideal and must be irreducible. The polynomial is called the minimal polynomial of and it encodes many important properties of . Hence the ring isomorphism obtained by the homomorphism theorem is an isomorphism of fields, where we can then observe that . Otherwise, is injective and hence we obtain a field isomorphism , where is the field of fractions of , i.e. the field of
https://en.wikipedia.org/wiki/Ehresmann%20connection
In differential geometry, an Ehresmann connection (after the French mathematician Charles Ehresmann who first formalized this concept) is a version of the notion of a connection, which makes sense on any smooth fiber bundle. In particular, it does not rely on the possible vector bundle structure of the underlying fiber bundle, but nevertheless, linear connections may be viewed as a special case. Another important special case of Ehresmann connections are principal connections on principal bundles, which are required to be equivariant in the principal Lie group action. Introduction A covariant derivative in differential geometry is a linear differential operator which takes the directional derivative of a section of a vector bundle in a covariant manner. It also allows one to formulate a notion of a parallel section of a bundle in the direction of a vector: a section s is parallel along a vector X if . So a covariant derivative provides at least two things: a differential operator, and a notion of what it means to be parallel in each direction. An Ehresmann connection drops the differential operator completely and defines a connection axiomatically in terms of the sections parallel in each direction . Specifically, an Ehresmann connection singles out a vector subspace of each tangent space to the total space of the fiber bundle, called the horizontal space. A section s is then horizontal (i.e., parallel) in the direction X if lies in a horizontal space. Here we are regarding s as a function from the base M to the fiber bundle E, so that is then the pushforward of tangent vectors. The horizontal spaces together form a vector subbundle of . This has the immediate benefit of being definable on a much broader class of structures than mere vector bundles. In particular, it is well-defined on a general fiber bundle. Furthermore, many of the features of the covariant derivative still remain: parallel transport, curvature, and holonomy. The missing ingredient
https://en.wikipedia.org/wiki/Jug%20wine
"Jug wine" is a term in the United States for inexpensive table wine (or "bulk wine") typically bottled in a glass bottle or jug. Historically, jug wines were labeled semi-generically, often sold to third parties to be relabeled, or sold directly from the winery's tasting room to customers who would often bring their own bottles. For a period following Prohibition, jug wines were the only domestic wine available for most Americans. Beginning in the 1960s, when Americans began to consume more premium wine, jug wine took on a reputation for being "extreme value" (bargain-priced premium wine). Beginning in the late 1980s jug wines have increasingly been labeled varietally to meet consumer demand. Common brands Common brands include Gallo, Carlo Rossi, Almaden Vineyards, and Inglenook Winery. Typical formats include 750 ml and 1 liter glass bottles, as well as 3 and 5-liter jugs. More recent packaging methods include lined boxes, and plastic bags inside corrugated fiberboard boxes ("bag in a box"). See also Box wine Fighting varietal Flavored fortified wine Plonk (wine)
https://en.wikipedia.org/wiki/Ivan%20Smith%20%28mathematician%29
Ivan Smith is a British mathematician who deals with symplectic manifolds and their interaction with algebraic geometry, low-dimensional topology, and dynamics. He is a professor at the University of Cambridge. Education and career Smith was born in 1973 to Neil Smith, a professor of linguistics at University College London. Smith studied at the University of Oxford, where he received his doctorate in 1999 under the supervision of Simon Donaldson with thesis Symplectic Geometry of Lefschetz Fibrations. Smith is now a professor in Cambridge at Gonville & Caius College. Among other things, Smith derived nodal invariants from symplectic geometry. He received in 2007 the Whitehead Prize for his work in symplectic topology (highlighting the breadth of applied techniques from algebraic geometry and topology) and in 2013 the Adams Prize. In 2018 he was an invited speaker at the International Congress of Mathematicians in Rio de Janeiro. Smith was elected a Fellow of the Royal Society in 2023. Selected publications with Denis Auroux: Lefschetz pencils, branched covers and symplectic invariants . In: Symplectic 4-manifolds and algebraic surfaces (Cetraro, 2003), Lect. Notes in Math. 1938, Springer, 2008, 1–53, Arxiv with Mohammed Abouzaid: "Homological mirror symmetry for the 4-torus", Duke Math. J., Vol. 152, 2010, pp. 373–440, Arxiv Floer cohomology and pencils of quadrics , Inventiones Mathematicae, Vol. 189, 2012, pp. 149–250, Arxiv "A symplectic prolegomenon", Bulletin AMS, Vol. 52, 2015, pp. 415–464, Quiver algebras as Fukaya categories , Geom. Topol., Vol. 19, 2015, 2557–2617, Arxiv with Mohammed Abouzaid: Khovanov homology from Floer cohomology , Arxiv 2015 with Mohammed Abouzaid: The symplectic arc algebra is formal , Duke Math. J., Vol. 165, 2016, pp. 985–1060, Arxiv
https://en.wikipedia.org/wiki/National%20Physical%20Laboratory%20%28United%20Kingdom%29
The National Physical Laboratory (NPL) is the national measurement standards laboratory of the United Kingdom. It sets and maintains physical standards for British industry. Founded in 1900, the NPL is one of the oldest metrology institutes in the world. Research and development work at the laboratory has contributed to the advancement of many disciplines of science, including the development of early computers in the late 1940s and 1950s, construction of the first accurate atomic clock in 1955, and the invention and pioneering implementation of packet switching in the 1960s, which is today one of the fundamental technologies of the Internet. The former heads of NPL include many individuals who were pillars of the British scientific establishment. NPL is based at Bushy Park in Teddington, west London. It is managed under the Department for Science, Innovation and Technology and is one of the most extensive government laboratories in the United Kingdom. History Precursors In the 19th century, the Kew Observatory was run by self-funded devotees of science. In the early 1850s, the observatory began charging fees for testing meteorological instruments and other scientific equipment. As universities in the United Kingdom created and expanded physics departments, the governing committee of the Observatory became increasingly dominated by paid university physicists in the last two decades of the nineteenth century. By this time, instrument-testing was the observatory's main role. Physicists sought the establishment of a state-funded scientific institution for testing electrical standards. Founding The National Physical Laboratory was established in 1900 at Bushy House in Teddington on the site of the Kew Observatory. Its purpose was "for standardising and verifying instruments, for testing materials, and for the determination of physical constants". The laboratory was run by the UK government, with members of staff being part of the civil service. It grew to fill a
https://en.wikipedia.org/wiki/Attila%20Szabo%20%28scientist%29
Attila Szabo is a biophysicist who is a Distinguished Investigator and Section Chief of the Theoretical Biophysical Chemistry Section in the Laboratory of Chemical Physics at the National Institute of Diabetes and Digestive and Kidney Diseases, part of the United States National Institutes of Health. Early life and education Szabo was born in Budapest, Hungary in 1947. His family left Hungary in response to the Hungarian Revolution of 1956 - he later recalled a dramatic escape over the border to Austria - and immigrated to Canada, settling in Montreal. Interested in chemistry since he was a child, Szabo studied the subject at McGill University as an undergraduate and received his bachelor's degree in 1968. He then began graduate school at Harvard University, where he joined Martin Karplus' research group. There he was introduced to theoretical work on NMR spectroscopy and developed a statistical mechanical model of the hemoglobin protein. Szabo accompanied Karplus on a sabbatical in Paris in 1972, and then spent a year in Cambridge working jointly with David Buckingham and Max Perutz. Szabo received his Ph.D. from Harvard in 1973. Academic career Szabo joined the faculty at the University of Indiana, Bloomington in 1974. NMR was an active area of research there at the time and Szabo became involved in developing theoretical models to interpret NMR studies of biomolecules. Among the resulting work on NMR relaxation, the Lipari-Szabo model-free formalism for modeling the dynamics of molecules was most influential and remains in active use in the field. While at Indiana Szabo also co-authored a textbook on quantum chemistry with fellow Karplus lab member Neil Ostlund. In 1979, Szabo was informed that he had not received tenure at Indiana. As he later recalled, his Ph.D. advisor Karplus connected him with Bill Eaton at the National Institute of Diabetes and Digestive and Kidney Diseases, who recruited Szabo to NIDDK. Szabo has remained at NIDDK since and is currently
https://en.wikipedia.org/wiki/Internet%20Communications%20Engine
The Internet Communications Engine, or Ice, is an open-source RPC framework developed by ZeroC. It provides SDKs for C++, C#, Java, JavaScript, MATLAB, Objective-C, PHP, Python, Ruby and Swift, and can run on various operating systems, including Linux, Windows, macOS, iOS and Android. Ice implements a proprietary application layer communications protocol, called the Ice protocol, that can run over TCP, TLS, UDP, WebSocket and Bluetooth. As its name indicates, Ice can be suitable for applications that communicate over the Internet, and includes functionality for traversing firewalls. History Initially released in February 2003, Ice was influenced by the Common Object Request Broker Architecture (CORBA) in its design, and indeed was created by several influential CORBA developers, including Michi Henning. However, according to ZeroC, it was smaller and less complex than CORBA because it was designed by a small group of experienced developers, instead of suffering from design by committee. In 2004, it was reported that a game called "Wish" by a company named Mutable Realms used Ice. In 2008, it was reported that Big Bear Solar Observatory had used the software since 2005. The source code repository for Ice is on GitHub since May 2015. Components Ice components include object-oriented remote-object-invocation, replication, grid-computing, failover, load-balancing, firewall-traversals and publish-subscribe services. To gain access to those services, applications are linked to a stub library or assembly, which is generated from a language-independent IDL-like syntax called slice. IceStorm is an object-oriented publish-and-subscribe framework that also supports federation and quality-of-service. Unlike other publish-subscribe frameworks such as Tibco Software's Rendezvous or SmartSockets, message content consist of objects of well defined classes rather than of structured text. IceGrid is a suite of frameworks that provide object-oriented load balancing, failover
https://en.wikipedia.org/wiki/Morganucodon
Morganucodon ("Glamorgan tooth") is an early mammaliaform genus that lived from the Late Triassic to the Middle Jurassic. It first appeared about 205 million years ago. Unlike many other early mammaliaforms, Morganucodon is well represented by abundant and well preserved (though in the vast majority of cases disarticulated) material. Most of this comes from Glamorgan in Wales (Morganucodon watsoni), but fossils have also been found in Yunnan Province in China (Morganucodon oehleri) and various parts of Europe and North America. Some closely related animals (Megazostrodon) are known from exquisite fossils from South Africa. The name comes from a Latinization of Morganuc, the name for South Glamorgan in the Domesday Book, the county of Wales where it was discovered by Walter Georg Kühne, giving the meaning "Glamorgan tooth". History of discovery In the summer of 1947, fieldwork was done at Duchy Quarry in Glamorgan in southern Wales. Grey conglomerate that formed fissure fill deposits within karstic voids in Carboniferous limestone was extracted. In 1949, Walter Georg Kühne noted the lower cheek tooth of a primitive mammal while examining samples of the rock. He named it Morganucodon watsoni, with the genus name being derived from Morganuc, which Kühne stated was the name of South Glamorgan in the Domesday Book, with the species name being in honour of D. M. S. Watson. Additional remains of M. watsoni were described by Kühne in 1958. Also in 1958, Kenneth Kermack and Frances Mussett described additional remains from Pant Quarry, about a mile from Duchy Quarry, that had been collected in 1956. In August 1948, an expedition to Lufeng in Yunnan, China yielded a 1 in (2.5 cm) long skull. It was shortly sent to Beijing (then Peking) and then eventually sent out of China, and deposited with Kenneth Kermack at University College London in 1960. The specimen was preliminarily described in 1963 by Harold W. Rigney, who noted the similarity to Morganucodon from Britain, and
https://en.wikipedia.org/wiki/Resource%20Public%20Key%20Infrastructure
Resource Public Key Infrastructure (RPKI), also known as Resource Certification, is a specialized public key infrastructure (PKI) framework to support improved security for the Internet's BGP routing infrastructure. RPKI provides a way to connect Internet number resource information (such as Autonomous System numbers and IP addresses) to a trust anchor. The certificate structure mirrors the way in which Internet number resources are distributed. That is, resources are initially distributed by the IANA to the regional Internet registries (RIRs), who in turn distribute them to local Internet registries (LIRs), who then distribute the resources to their customers. RPKI can be used by the legitimate holders of the resources to control the operation of Internet routing protocols to prevent route hijacking and other attacks. In particular, RPKI is used to secure the Border Gateway Protocol (BGP) through BGP Route Origin Validation (ROV), as well as Neighbor Discovery Protocol (ND) for IPv6 through the Secure Neighbor Discovery protocol (SEND). The RPKI architecture is documented in RFC 6480. The RPKI specification is documented in a spread out series of RFCs: RFC 6481, RFC 6482, RFC 6483, RFC 6484, RFC 6485, RFC 6486, RFC 6487, RFC 6488, RFC 6489, RFC 6490, RFC 6491, RFC 6492, and RFC 6493. SEND is documented in RFC 6494 and RFC 6495. These RFCs are a product of the IETF's SIDR ("Secure Inter-Domain Routing") working group, and are based on a threat analysis which was documented in RFC 4593. These standards cover BGP origin validation, while path validation is provided by BGPsec, which has been standardized separately in RFC 8205. Several implementations for prefix origin validation already exist. Resource Certificates and child objects RPKI uses X.509 PKI certificates (RFC 5280) with extensions for IP addresses and AS identifiers (RFC 3779). It allows the members of regional Internet registries, known as local Internet registries (LIRs), to obtain a resource certif
https://en.wikipedia.org/wiki/Special%20function%20register
A Special Function Register (or Special Purpose Register, or simply Special Register) is a register within a microprocessor, which controls or monitors various aspects of the microprocessor's function. Depending on the processor architecture, this can include, but is not limited to: I/O and peripheral control (such as serial ports or general-purpose IOs) timers stack pointer stack limit (to prevent overflows) program counter subroutine return address processor status (servicing an interrupt, running in protected mode, etc.) condition codes (result of previous comparisons) Because special registers are closely tied to some special function or status of the processor, they might not be directly writeable by normal instructions (such as adds, moves, etc.). Instead, some special registers in some processor architectures require special instructions to modify them. For example, the program counter is not directly writeable in many processor architectures. Instead, the programmer uses instructions such as return from subroutine, jump, or branch to modify the program counter. For another example, the condition code register might not directly writable, instead being updated only by compare instructions. Intel Processors Some of SFR (Special Function Register) bits may be set directly using SETB/LDB instructions on proper address, whereas others may require usage of specific instructions. The Intel 80196 class microcontroller has 24 SFRs, each 1 Byte in size; standard Intel 8051 chips have 21 SFRs. The Special Function Register (SFR) is the upper area of addressable memory, from address 0x80 to 0xFF. This area of memory can't be used for data or program storage, but is instead a series of memory-mapped ports and registers. All port input and output can therefore be performed by memory move operations on specified addresses in the SFR. Also, different status registers are mapped into the SFR, for use in checking the status of the 8051, and changing some operation
https://en.wikipedia.org/wiki/Pramlintide
Pramlintide (trade name Symlin) is an injectable amylin analogue drug for diabetes (both type 1 and 2), developed by Amylin Pharmaceuticals (now a wholly owned subsidiary of AstraZeneca). Pramlintide is sold as an acetate salt. Pharmacology Pramlintide is an analogue of amylin, a small peptide hormone that is released into the bloodstream by the β cells of the pancreas along with insulin after a meal. Like insulin, amylin is completely absent in individuals with Type I diabetes. In synergy with endogenous amylin, pramlintide aids in the regulation of blood glucose by slowing gastric emptying, promoting satiety via hypothalamic receptors (different receptors than for GLP-1), and inhibiting inappropriate secretion of glucagon, a catabolic hormone that opposes the effects of insulin and amylin. Pramlintide also has effects in raising the acute first-phase insulin response threshold following a meal. Both a reduction in glycated hemoglobin and weight loss have been shown in insulin-treated patients with type 2 diabetes taking pramlintide as an adjunctive therapy. Research Applications In the research field, pramlintide has been experimented with and used as a potential treatment drug. Pramlintide has demonstrated its ability to decrease amyloid beta plaques in Alzheimer's disease mouse models. Approval Pramlintide has been approved on 3/16/2005 by the FDA, for use by type 1 and type 2 diabetic patients who use insulin. Pramlintide allows patients to use less insulin, lowers average blood sugar levels, and substantially reduces what otherwise would be a large unhealthy rise in blood sugar that occurs in diabetics right after eating. Apart from insulin analogs, pramlintide is the only drug approved by the FDA to lower blood sugar in type 1 diabetics since insulin in the early 1920s. Design and structure Since native human amylin is highly amyloidogenic and potentially toxic, the strategy for designing pramlintide was to substitute residues from rat amylin, which
https://en.wikipedia.org/wiki/Clef
A clef (from French: 'key') is a musical symbol used to indicate which notes are represented by the lines and spaces on a musical staff. Placing a clef on a staff assigns a particular pitch to one of the five lines or four spaces, which defines the pitches on the remaining lines and spaces. The three clef symbols used in modern music notation are the G-clef, F-clef, and C-clef. Placing these clefs on a line fixes a reference note to that line—an F-clef fixes the F below middle C, a C-clef fixes middle C, and a G-clef fixes the G above middle C. In modern music notation, the G-clef is most frequently seen as treble clef (placing G4 on the second line of the staff), and the F-clef as bass clef (placing F3 on the fourth line). The C-clef is mostly encountered as alto clef (placing middle C on the third line) or tenor clef (middle C on the fourth line). A clef may be placed on a space instead of a line, but this is rare. The use of different clefs makes it possible to write music for all instruments and voices, regardless of differences in range. Using different clefs for different instruments and voices allows each part to be written comfortably on a staff with a minimum of ledger lines. To this end, the G-clef is used for high parts, the C-clef for middle parts, and the F-clef for low parts. Transposing instruments can be an exception to this—the same clef is generally used for all instruments in a family, regardless of their sounding pitch. For example, even the low saxophones read in treble clef. A symmetry exists surrounding middle C regarding the F-, C- and G-clefs. C-clef defines middle C whereas G-clef and F-clef define the note at the interval of a fifth above middle C and below middle C, respectively. Two common mnemonics for learning the clef lines are: Good Boys Do Fine Always (bass clef) Every Good Boy Does Fine (treble clef){{efn|It is perhaps also useful to note that, for the treble clef, the four spaces between the five clef lines spell out Plac
https://en.wikipedia.org/wiki/Vino%20Versum%20Poysdorf
The Vino Versum Poysdorf is a municipal museum dedicated to the wine and culture history of Poysdorf and the Weinviertel ("wine quarter"). History Poysdorf is a center of wine growing in Austria with 500 Hectare vineyards. The wine growing tradition goes back to the Middle Ages and was first mentioned in 1338. 1582 the then village received a market right and developed huge wealth through growing and trading wine. 1910 a first municipal museum was established, 1978 it was reopened in the city's old almshouse. 2013 the Lower Austria Provincial Exhibition took place in Poysdorf, 4 million Euros were invested in the enlargement and renovation of the museums area. In April 2014 a new designed permanent exhibition was reopened with the name Vino Versum Poysdorf. Buildings The main building is a former almshouse, built in 1653. 1664 a baroque chapel, dedicated to St. Barbara, was added. The building is surrounded by an open-air area with vineyards, press houses with cellars and a barn with an exhibition about the development of wine presses. A historic wine cellar, passing under a highway, connects the old almshouse and open air area with a plaza, surrounded by a contemporary exhibition building, the Hall of Grapes, and an adapted vintner´s house with a vinothek, souvenir shop and visitor center. Gallery See also List of food and beverage museums
https://en.wikipedia.org/wiki/Alfred%20Tauber
Alfred Tauber (5 November 1866 – 26 July 1942) was an Austrian Empire-born Austrian mathematician, known for his contribution to mathematical analysis and to the theory of functions of a complex variable: he is the eponym of an important class of theorems with applications ranging from mathematical and harmonic analysis to number theory. He was murdered in the Theresienstadt concentration camp. Life and academic career Born in Pressburg, Kingdom of Hungary, Austrian Empire (now Bratislava, Slovakia), he began studying mathematics at Vienna University in 1884, obtained his Ph.D. in 1889, and his habilitation in 1891. Starting from 1892, he worked as chief mathematician at the Phönix insurance company until 1908, when he became an a.o. professor at the University of Vienna, though, already from 1901, he had been honorary professor at TU Vienna and director of its insurance mathematics chair. In 1933, he was awarded the Grand Decoration of Honour in Silver for Services to the Republic of Austria, and retired as emeritus extraordinary professor. However, he continued lecturing as a privatdozent until 1938, when he was forced to resign as a consequence of the "Anschluss". On 28–29 June 1942, he was deported with transport IV/2, č. 621 to Theresienstadt, where he was murdered on 26 July 1942. Work list 35 publications in the bibliography appended to his obituary, and also a search performed on the "Jahrbuch über die Fortschritte der Mathematik" database results in a list 35 mathematical works authored by him, spanning a period of time from 1891 to 1940. However, cites two papers on actuarial mathematics which do not appear in these two bibliographical lists and Binder's bibliography of Tauber's works (1984, pp. 163–166), while listing 71 entries including the ones in the bibliography of and the two cited by Hlawka, does not includes the short note so the exact number of his works is not known. According to , his scientific research can be divided into three areas: t
https://en.wikipedia.org/wiki/Geriatric%20medicine%20in%20Egypt
Geriatric medicine, as a speciality, was introduced in Egypt in 1982, and in 1984 a geriatrics and gerontology unit in Ain Shams University Faculty of Medicine was established. Egyptian history Many ancient physicians were interested in the field of geriatrics and some of them wrote separate books on this subject. Abu Ali Ibn-Sina Avicenna's The Canon of Medicine (1025) was the first book to offer instruction for the care of the aged. In a chapter entitled "Regimen of Old Age", Avicenna was concerned with how "old folk need plenty of sleep", how their bodies should be anointed with oil, and recommended exercises such as walking or horse-riding. Thesis III of the Canon discussed the diet suitable for old people, and dedicated several sections to elderly patients who become constipated. Also the famous Physician Ibn Al-Jazzar (Algizar) Al-Qayrawani (circa 898–980) wrote a special book on the medicine and health of elderly (Kitab Tibb Al Machayikh) or (Teb Al-Mashaikh wa hefz sehatahom). Also a book on sleep disorders and another one on forgetfulness and how to strengthen memory (Kitab Al Nissian wa Toroq Taqwiati Adhakira) and a Treatise on causes of mortality (Rissala Fi Asbab Al Wafah). Another Arabic physician in the 9th century, Ishaq ibn Hunayn (died in 910) wrote a Treatise on Drugs for Forgetfulness (Risalah al-Shafiyah fi adwiyat al-nisyan) Current situation Covering the cost of healthcare is a major challenge for seniors (defined as 60 years or older in Egypt). Currently, 62 percent of elderly men have health insurance compared to 35 percent of elderly women. However, medical care for the elderly, in the absence of a comprehensive health insurance plan, is extremely expensive. Few hospitals have wings just for the elderly, two percent of whom have a permanent disability (i.e., 81,000 people). There are only 52 physical therapy centers in Egypt serving 1,258 of those over 60. Geriatric teaching programs were available in Egypt. Geriatrics governmental
https://en.wikipedia.org/wiki/Interbilayer%20forces%20in%20membrane%20fusion
Membrane fusion is a key biophysical process that is essential for the functioning of life itself. It is defined as the event where two lipid bilayers approach each other and then merge to form a single continuous structure. In living beings, cells are made of an outer coat made of lipid bilayers; which then cause fusion to take place in events such as fertilization, embryogenesis and even infections by various types of bacteria and viruses. It is therefore an extremely important event to study. From an evolutionary angle, fusion is an extremely controlled phenomenon. Random fusion can result in severe problems to the normal functioning of the human body. Fusion of biological membranes is mediated by proteins. Regardless of the complexity of the system, fusion essentially occurs due to the interplay of various interfacial forces, namely hydration repulsion, hydrophobic attraction and van der Waals forces. Inter-bilayer forces Lipid bilayers are structures of lipid molecules consisting of a hydrophobic tail and a hydrophilic head group. Therefore, these structures experience all the characteristic Interbilayer forces involved in that regime. Hydration repulsion Two hydrated bilayers experience strong repulsion as they approach each other. These forces have been measured using the Surface forces apparatus (S.F.A), an instrument used for measuring forces between surfaces. This repulsion was first proposed by Langmuir and was thought to arise due to water molecules that hydrate the bilayers. Hydration repulsion can thus be defined as the work required in removing the water molecules around hydrophilic molecules (like lipid head groups) in the bilayer system. As water molecules have an affinity towards hydrophilic head groups, they try to arrange themselves around the head groups of the lipid molecules and it becomes very hard to separate this favorable combination. Experiments performed through SFA have confirmed that the nature of this force is an exponential declin
https://en.wikipedia.org/wiki/Ionizing%20radiation
Ionizing radiation (or ionising radiation), including nuclear radiation, consists of subatomic particles or electromagnetic waves that have sufficient energy to ionize atoms or molecules by detaching electrons from them. Some particles can travel up to 99% of the speed of light, and the electromagnetic waves are on the high-energy portion of the electromagnetic spectrum. Gamma rays, X-rays, and the higher energy ultraviolet part of the electromagnetic spectrum are ionizing radiation, whereas the lower energy ultraviolet, visible light, nearly all types of laser light, infrared, microwaves, and radio waves are non-ionizing radiation. The boundary between ionizing and non-ionizing radiation in the ultraviolet area cannot be sharply defined, as different molecules and atoms ionize at different energies. The energy of ionizing radiation starts between 10 electronvolts (eV) and 33 eV. Typical ionizing subatomic particles include alpha particles, beta particles, and neutrons. These are typically created by radioactive decay, and almost all are energetic enough to ionize. There are also secondary cosmic particles produced after cosmic rays interact with Earth's atmosphere, including muons, mesons, and positrons. Cosmic rays may also produce radioisotopes on Earth (for example, carbon-14), which in turn decay and emit ionizing radiation. Cosmic rays and the decay of radioactive isotopes are the primary sources of natural ionizing radiation on Earth, contributing to background radiation. Ionizing radiation is also generated artificially by X-ray tubes, particle accelerators, and nuclear fission. Ionizing radiation is not immediately detectable by human senses, so instruments such as Geiger counters are used to detect and measure it. However, very high energy particles can produce visible effects on both organic and inorganic matter (e.g. water lighting in Cherenkov radiation) or humans (e.g. acute radiation syndrome). Ionizing radiation is used in a wide variety of field
https://en.wikipedia.org/wiki/Haplogroup%20S%20%28mtDNA%29
In human genetics, Haplogroup S is a human mitochondrial DNA (mtDNA) haplogroup found only among Indigenous Australians. It is a descendant of macrohaplogroup N. Origin Haplogroup S mtDNA evolved within Australia between 64,000 and 40,000 years ago (51 kya). Distribution It is found in the Indigenous Australian population. Haplogroup S2 found in Willandra Lakes human remain WLH4 dated back Late Holocene (3,000-500 years ago). The following table lists relevant GenBank samples: Subclades Tree This phylogenetic tree of haplogroup S subclades is based on the paper by Mannis van Oven and Manfred Kayser Updated comprehensive phylogenetic tree of global human mitochondrial DNA variation and subsequent published research. The TMRCA for haplogroup S is between 49 and 51 KYA according to Nano Nagle's Aboriginal Australian mitochondrial genome variation – an increased understanding of population antiquity and diversity publication that published in 2017. S (64-40 kya) in Australia S1 (53-32 kya) in Australia S1a (44-29 kya) found in WA, NT, QLD and NSW S1b (37-22 kya) found in NT, QLD and NSW S1b1 (30-10 kya) found in NT and QLD S1b1a (24-6 kya) found in QLD S1b2 (17-3 kya) found in QLD S1b3 (20-4 kya) found in QLD and NSW S2 (44-22 kya) in Australia S2a (38-18 kya) found in NT, QLD, NSW and TAS S2a1 (31-12 kya) found in NSW, QLD and TAS S2a1a (19-6 kya) found in NSW and QLD S2a2 (38-11 kya) found in NT, QLD and NSW S2b (42-18 kya) found in WA, NT, QLD and VIC S2b1(27-9 kya) found in NT, QLD and VIC S2b2 (37-12 kya) found in WA, NT and QLD S3 (17-1 kya) found in NT S4 found in NT S5 found in WA S6 found in NSW See also Genealogical DNA test Genetic genealogy Human mitochondrial genetics Population genetics
https://en.wikipedia.org/wiki/Global%20Goal%3A%20Unite%20for%20Our%20Future
Global Goal: Unite for Our Future was a virtual event held on June 27, 2020. Created by Global Citizen and the European Commission, it consisted of a summit and a concert featuring different personalities aimed to highlight the disproportionate impact of the COVID-19 pandemic on marginalized communities. The event was produced in collaboration with companies such as Live Nation Entertainment, The Lede Company, and Roc Nation, and also by individuals such as Michele Anthony of the Universal Music Group, Declan Kelly of Teneo, Scooter Braun (with his company SB Projects), Adam Leber (on behalf of Maverick), and Derrick Johnson (on behalf of the NAACP). Summit The summit, titled Global Goal: Unite for Our Future—The Summit, featured panel discussions and interviews with world leaders, corporations and philanthropists as they announce new commitments to help develop equitable distribution of COVID-19 tests, treatments and vaccines, as well as rebuild communities devastated by the pandemic. It will be produced by Michael Dempsey and hosted by journalists Katie Couric, Mallika Kapur, Morgan Radford, Isha Sesay, and Keir Simmons. Participants Concert The concert, titled Global Goal: Unite for Our Future—The Concert, was hosted by actor Dwayne Johnson. It was a worldwide music and entertainment special claiming to celebrate the commitments made due to the actions Global Citizens have taken. Performers Additional performers Chloe x Halle Christine and the Queens J’Nai Bridges with Gustavo Dudamel, the Los Angeles Philharmonic and Youth Orchestra Los Angeles For Love Choir Appearances Chris Rock Hugh Jackman Kerry Washington Charlize Theron Forest Whitaker David Beckham Salma Hayek Billy Porter Diane Kruger Antoni Porowski Ken Jeong Naomi Campbell Nikolaj Coster-Waldau Olivia Colman Broadcast The event was broadcast in the United States on NBC, MSNBC, Bloomberg Television, iHeartRadio, Sirius XM, and InsightTV. International broadcasters Africa
https://en.wikipedia.org/wiki/Mojave%20phone%20booth
The Mojave phone booth ( ) was a lone telephone booth in what is now the Mojave National Preserve in California. It attracted online attention in 1997 for its unusual location – it was located at the intersection of two dirt roads in a remote part of the Mojave Desert, from the nearest paved road (Interstate 15 to the northeast, Kelbaker Road to the southwest) and miles from any buildings. History The phone booth was originally set up in 1948 to provide telephone service to local volcanic cinder miners and others living in the area, at the request of Emerson Ray, who owned the Cima Cinder Mine nearby. It was assigned the name Cinder Peak Policy Station, part of a network of "policy stations" placed by mandate of the California government to serve residents of isolated parts of the state. The Mojave phone booth probably replaced an earlier booth to the south. The original hand-cranked magneto phone was replaced with a payphone in the 1960s. The rotary phone was then replaced with a touch-tone model in the 1970s. The booth phone's original number was BAker-3-9969. After area codes were established in 1947, it shared the same area code as all of Southern California, 213. In 1951, the area code 714 was split off from 213, and 714 was the booth's area code until 1982 when parts of Riverside and San Bernardino counties were given the area code 619. The booth's area code for its final few years of existence was 760. In 1997, a Los Angeles man spotted a telephone icon on a map of the Mojave Desert and decided to visit it. He wrote a letter about his adventure to an underground magazine and included the booth's telephone number. An Arizona man, Godfrey Daniels, read the letter and started a website devoted to the Mojave telephone booth. The booth became an Internet sensation after it was mentioned in a New York Times article. Soon other people began calling the booth, others made websites about it, and a few even took trips to the booth to answer, often camping out at t
https://en.wikipedia.org/wiki/Lambert%20quadrilateral
In geometry, a Lambert quadrilateral (also known as Ibn al-Haytham–Lambert quadrilateral), is a quadrilateral in which three of its angles are right angles. Historically, the fourth angle of a Lambert quadrilateral was of considerable interest since if it could be shown to be a right angle, then the Euclidean parallel postulate could be proved as a theorem. It is now known that the type of the fourth angle depends upon the geometry in which the quadrilateral exists. In hyperbolic geometry the fourth angle is acute, in Euclidean geometry it is a right angle and in elliptic geometry it is an obtuse angle. A Lambert quadrilateral can be constructed from a Saccheri quadrilateral by joining the midpoints of the base and summit of the Saccheri quadrilateral. This line segment is perpendicular to both the base and summit and so either half of the Saccheri quadrilateral is a Lambert quadrilateral. Lambert quadrilateral in hyperbolic geometry In hyperbolic geometry a Lambert quadrilateral AOBF where the angles are right, and F is opposite O , is an acute angle , and the curvature = -1 the following relations hold: Where are hyperbolic functions Examples See also Non-Euclidean geometry Notes
https://en.wikipedia.org/wiki/Institute%20of%20Acoustics%20%28United%20Kingdom%29
The Institute of Acoustics (IOA) is a British professional engineering institution founded in 1974. It is licensed by the Engineering Council UK to assess candidates for inclusion on ECUK's Register of professional Engineers. The institute's address is Silbury Court, 406 Silbury Boulevard, Milton Keynes MK9 2AF, United Kingdom. The current president of the IOA is Alistair Somerville. Past presidents include Barry Gibbs, John Hinton OBE, Colin English, David Weston, Tony Jones, Professor Trevor Cox, William Egan, Professor Bridget Shield, and Jo Webb. History In 1963 a Society of Acoustic Technology was formed in the UK for those interested in this subject: the President was Elfyn Richards. Because of the interest in establishing a professional body, meetings were held with various societies and institutions, and in 1965 a British Acoustical Society was set up, absorbing the earlier society. In 1974 the British Acoustical Society amalgamated with the Acoustics Group of the Institute of Physics to form the Institute of Acoustics. Specialist groups Building acoustics Electroacoustics Environmental noise Measurement and instrumentation Musical acoustics Noise and vibration engineering Physical acoustics Speech and hearing Underwater acoustics Medals and awards The following prizes are awarded by the Institute Rayleigh Medal Tyndall Medal A B Wood Medal R W B Stephens Medal IOA Engineering Medal Honorary fellowship Peter Barnett Memorial Award The Award for Promoting Acoustics to the Public Award for Services to the Institute IOA Young Persons' Award for Innovation in Acoustical Engineering IOA Prize for best diploma student ANC prize for the best diploma project ANC prize for the best paper at an IOA conference See also Chartered engineer Incorporated engineer The Association of Noise Consultants
https://en.wikipedia.org/wiki/Immunopathology
Immunopathology is a branch of medicine that deals with immune responses associated with disease. It includes the study of the pathology of an organism, organ system, or disease with respect to the immune system, immunity, and immune responses. In biology, it refers to damage caused to an organism by its own immune response, as a result of an infection. It could be due to mismatch between pathogen and host species, and often occurs when an animal pathogen infects a human (e.g. avian flu leads to a cytokine storm which contributes to the increased mortality rate). Types of Immunity In all vertebrates, there are two different kinds of immunities: Innate and Adaptive immunity. Innate immunity is used to fight off non-changing antigens and is therefore considered nonspecific. It is usually a more immediate response than the adaptive immune system, usually responding within minutes to hours. It is composed of physical blockades such as the skin, but also contains nonspecific immune cells such as dendritic cells, macrophages, and basophils. The second form of immunity is Adaptive immunity. This form of immunity requires recognition of the foreign antigen before a response is produced. Once the antigen is recognized, a specific response is produced in order to destroy the specific antigen. Because of its tailored response characteristic, adaptive immunity is considered to be specific immunity. A key part of adaptive immunity that separates it from innate is the use of memory to combat the antigen in the future. When the antigen is originally introduced, the organism does not have any receptors for the antigen so it must generate them from the first time the antigen is present. The immune system then builds a memory of that antigen, which enables it to recognize the antigen quicker in the future and be able to combat it quicker and more efficiently. The more the system is exposed to the antigen, the quicker it will build up its responsiveness. Nested within Adaptive immu
https://en.wikipedia.org/wiki/Mihai%20P%C4%83tra%C8%99cu%20%28computer%20scientist%29
Mihai Pătrașcu (17 July 1982 – 5 June 2012) was a Romanian-American computer scientist at AT&T Labs in Florham Park, New Jersey, USA. Pătrașcu attended Carol I National College in Craiova. As a high school student, he won 2 gold medals and 1 silver medal at the International Olympiad in Informatics. He completed his undergraduate and graduate studies in Computer Science at Massachusetts Institute of Technology, completing his thesis under the supervision of Erik Demaine in 2008. Pătrașcu’s work was concerned with fundamental questions about basic data structures. Pătrașcu received the Machtey Award for the best student paper at the Symposium on Foundations of Computer Science in 2008, and the Presburger Award from the European Association for Theoretical Computer Science in 2012, for breaking "many old barriers on fundamental data structure problems, not only revitalizing but also revolutionizing a field that was almost silent for over a decade." Pătrașcu died on June 5, 2012 after being diagnosed with glioblastoma in January, 2011, and was buried in his native city. Selected publications Preliminary version published in FOCS 2008, . Preliminary version published in FOCS 2006, . Preliminary version published in FOCS 2004, . See Tango tree.
https://en.wikipedia.org/wiki/Rubratoxin
Rubratoxins are hepatotoxic mycotoxin produced by Penicillium rubrum and Penicillium purpurogenum. Rubratoxin A and rubratoxin B have been known since 1950s, while rubratoxin C was recently reported. Rubratoxins are also known as protein phosphatase 2A (PP2A) specific inhibitor. The PP2A inhibitory activity of rubratoxin A is about 100-fold higher than rubratoxin B and rubratoxin A is now used as a chemical probe for PP2A research.
https://en.wikipedia.org/wiki/DNA%20polymerase%20III%2C%20delta%20subunit
In molecular biology, the δ (delta) subunit of DNA polymerase III is encoded by the holA gene in E. coli and other bacteria. Along with the γ, δ', χ, and ψ subunits that make up the core polymerase, and the β accessory proteins, the δ subunit is responsible for the high speed and processivity of polIII.
https://en.wikipedia.org/wiki/Unbeatable%20strategy
In biology, the idea of an unbeatable strategy was proposed by W.D. Hamilton in his 1967 paper on sex ratios in Science. In this paper Hamilton discusses sex ratios as strategies in a game, and cites Verner as using this language in his 1965 paper which "claims to show that, given factors causing fluctuations of the population's primary sex ratio, a 1:1 sex-ratio production proves the best overall genotypic strategy". "In the way in which the success of a chosen sex ratio depends on choices made by the co-parasitizing females, this problem resembles certain problems discussed in the "theory of games." In the foregoing analysis a game-like element, of a kind, was present and made necessary the use of the word unbeatable to describe the ratio finally established. This word was applied in just the same sense in which it could be applied to the "minimax" strategy of a zero-sum two-person game. Such a strategy should not, without qualification, be called optimum because it is not optimum against -although unbeaten by- any strategy differing from itself. This is exactly the case with the "unbeatable" sex ratios referred to." Hamilton (1967). "[...] But if, on the contrary, players of such a game were motivated to outscore, they would find that is beaten by a higher ratio, ; the value of which gives its player the greatest possible advantage over the player playing , is found to be given by the relationship and shows to be the unbeatable play." Hamilton (1967). The concept can be traced through R.A. Fisher (1930) to Darwin (1859); see Edwards (1998). Hamilton did not explicitly define the term "unbeatable strategy" or apply the concept beyond the evolution of sex-ratios, but the idea was very influential. George R. Price generalised the verbal argument, which was then formalised mathematically by John Maynard Smith, into the evolutionarily stable strategy (ESS).
https://en.wikipedia.org/wiki/Magneto-optic%20effect
A magneto-optic effect is any one of a number of phenomena in which an electromagnetic wave propagates through a medium that has been altered by the presence of a quasistatic magnetic field. In such a medium, which is also called gyrotropic or gyromagnetic, left- and right-rotating elliptical polarizations can propagate at different speeds, leading to a number of important phenomena. When light is transmitted through a layer of magneto-optic material, the result is called the Faraday effect: the plane of polarization can be rotated, forming a Faraday rotator. The results of reflection from a magneto-optic material are known as the magneto-optic Kerr effect (not to be confused with the nonlinear Kerr effect). In general, magneto-optic effects break time reversal symmetry locally (i.e. when only the propagation of light, and not the source of the magnetic field, is considered) as well as Lorentz reciprocity, which is a necessary condition to construct devices such as optical isolators (through which light passes in one direction but not the other). Two gyrotropic materials with reversed rotation directions of the two principal polarizations, corresponding to complex-conjugate ε tensors for lossless media, are called optical isomers. Gyrotropic permittivity In particular, in a magneto-optic material the presence of a magnetic field (either externally applied or because the material itself is ferromagnetic) can cause a change in the permittivity tensor ε of the material. The ε becomes anisotropic, a 3×3 matrix, with complex off-diagonal components, depending on the frequency ω of incident light. If the absorption losses can be neglected, ε is a Hermitian matrix. The resulting principal axes become complex as well, corresponding to elliptically-polarized light where left- and right-rotating polarizations can travel at different speeds (analogous to birefringence). More specifically, for the case where absorption losses can be neglected, the most general form
https://en.wikipedia.org/wiki/Dark%20Horse%20Miniatures
Dark Horse Miniatures was an American company that produced 25 mm metal miniatures for fantasy role-playing games in the 1980s. History Dark Horse Miniatures was formed in 1979 to produce metal miniatures for the increasingly popular fantasy RPG market, including orcs, adventurers, elves, warriors, and monsters. The company started its own newsletter, Dark Times, for its mail order customers. Critic Bob Kindel noted that the newsletter "parodies, satirizes, and otherwise mocks more conventional house organs. It's a punster’s delight." In 1981, the company moved to Idaho. Unlike other miniatures companies of the time, Dark Horse provided alternate weapons, accessories and even pets for their miniatures, cast on the same "tree" in order to save costs. In 1984, Dark Horse obtained the first-ever license granted by the Teenage Mutant Ninja Turtles creators to manufacture 25 mm miniatures, and produced several lines of figures to be used with the Teenage Mutant Ninja Turtles & Other Strangeness role-playing game published by Palladium Books. Dark Horse subsequently obtained licenses for other comic book-related properties, including Groo The Wanderer and the Robotech franchise owned by Harmony Gold. In 1988, Robert Bigelow reported that Dark Horse was out of business, and that Goldring Miniatures of Arcadia, California was using the Dark Horse logo and calling themselves Dark Horse's successor. Reception In Issue 73 of Space Gamer, Bob Kindel thought the company had "an excellent line of well-crafted miniatures." Kindel noted that "Its human figures tend to be ordinary ... although they can rise to the sublime, as shown by the turbaned thief (reminiscent of Errol Flynn as The Thief of Bagdad), or the robed cleric 'armed' with turkey leg and ale tankard." Kindel preferred the non-humanoid lines, commenting, "It is with the nonhuman, non-orc figures that Dark Horse begins to shine. A line of fur-clad, mohawk-tressed barbarian elves, mounted and afoot, make a fine add
https://en.wikipedia.org/wiki/National%20Fenestration%20Rating%20Council
The National Fenestration Rating Council (NFRC) is a United States 501(c)3 non-profit organization which sponsors an energy efficiency certification and labeling program for windows, doors, and skylights. NFRC labels provide performance ratings for such products in five categories: U-value, Solar Heat Gain Coefficient, Visible Transmittance, Air Leakage, and Condensation Resistance. This allows architects, builders, code officials, contractors, home owners, and specifiers to compare the energy efficiency among products, and determine whether a product meets code. The certification methods determined by NFRC are the only pathway to Energy Star and Energy Code compliance for windows in the US. History The organization was founded in 1989 by a group of window and door industry professionals in response to mounting pressure from local and national governments to standardize energy efficiency ratings for windows and doors. Before NFRC standardized energy efficiency ratings, each manufacturer developed their own methods. They often published misleading claims on their products' energy efficiency. For example, many manufacturers published center-of-glass u-factors rather than entire-unit u-factors, even though most of the heat loss occurs around the edges of a window. When NFRC was formed, the members agreed to rate entire-unit u-factor information. They agreed upon standard modeling and test methods to do so. They also agreed upon the roles of simulation and modeling laboratories and independent Inspection Agencies. NFRC published the first Certified Products Directory in 1991. The organization gained credibility and recognition with the passage of the 1992 National Energy Policy Act and the 1997 launch of the Energy Star program for windows. In 2003, NFRC implemented one-size testing, moving away from the dual u-factor ratings that it had previously used (which were a compromise between founding manufacturers). Product Ratings NFRC product ratings are determine
https://en.wikipedia.org/wiki/Amazon%20Lumberyard
Amazon Lumberyard is a now-superseded freeware cross-platform game engine developed by Amazon and based on CryEngine (initially released in 2002), which was licensed from Crytek in 2015. In July 2021, Amazon and the Linux Foundation announced that parts of the engine would be used to create a new open source game engine called Open 3D Engine, which would replace it. A new Open 3D Foundation, run by the Linux Foundation, will manage the new engine, which will be licensed under the open source Apache 2.0 license. The new engine is reportedly partially based on Lumberyard but with many parts rewritten, and is considered a new engine. Overview The Lumberyard engine features integration with Amazon Web Services to allow developers to build or host their games on Amazon's servers, as well as support for livestreaming via Twitch. Additionally, the engine includes Twitch ChatPlay, allowing viewers of the Twitch stream to influence the game through the associated chat, a method of play inspired by the Twitch Plays Pokémon phenomenon. The source code is available to end users with limitations: Users may not publicly release the Lumberyard engine source code or use it to release their own game engine. Lumberyard launched on February 9, 2016 alongside GameLift, a fee-based managed service for deploying and hosting multiplayer games, intended to allow developers the easy development of games that attract "large and vibrant communities of fans." , the software is currently in beta status and can be used to build games for Microsoft Windows, PlayStation 4, and Xbox One, with limited support for iOS and Android and support for macOS being planned for future releases. Virtual reality integration was added in Beta 1.3, allowing developers to build games supporting devices like Oculus Rift and HTC Vive. Despite being based on the architecture of Crytek's CryEngine, the engine has been developed to use many of its own custom-developed systems, some of which are in a preview mode. A
https://en.wikipedia.org/wiki/Vesical%20venous%20plexus
The vesical venous plexus is a venous plexus situated at the fundus of the urinary bladder. It collects venous blood from the urinary bladder in both sexes, from the accessory sex glands in males, and from the corpora cavernosa of clitoris in females (via the v. dorsalis profunda clitoridis). It drains into the internal iliac veins via several vesical veins. Anatomy The vesical venous plexus envelops the inferior part of the bladder and the base of the prostate. Anastomoses It communicates with the pudendal and prostatic plexuses.
https://en.wikipedia.org/wiki/Quantum%20%28journal%29
Quantum is an online-only, open-access, peer-reviewed scientific journal for quantum science and related fields. The journal was established in 2017. Quantum is an arXiv overlay journal, meaning the journal's content is hosted on the arXiv. Quantum is listed in the Directory of Open Access Journals and the Emerging Sources Citation Index, and it is recognized by the European Physical Society as maintaining a high standard of peer review. See also Discrete Analysis
https://en.wikipedia.org/wiki/Internet%20Digital%20DIOS
The LG Internet Digital DIOS (also known as R-S73CT) is an internet refrigerator released by LG Electronics in June 2000. The technology is the result of a project that started in 1997 and staffed by a team of 55 researchers with a budget cost of 15 billion won (US$49.2 million). Features The refrigerator has a TFT-LCD (thin-film transistor-liquid crystal display) screen with TV functionality and Local Area Network (LAN) port. It includes a LCD information window that features electronic pen, data memo, video messaging and schedule management functions and provides information, such as inside temperature, the freshness of stored foods, nutrition information and recipes. Other features are a webcam that is used as a scanner and tracks what is inside the refrigerator, a MP3 player and a three-level automatic icemaker. In addition, the electricity consumption is half the level of conventional refrigerators and the noise level is only 23 decibels.
https://en.wikipedia.org/wiki/Biological%20applications%20of%20bifurcation%20theory
Biological applications of bifurcation theory provide a framework for understanding the behavior of biological networks modeled as dynamical systems. In the context of a biological system, bifurcation theory describes how small changes in an input parameter can cause a bifurcation or qualitative change in the behavior of the system. The ability to make dramatic change in system output is often essential to organism function, and bifurcations are therefore ubiquitous in biological networks such as the switches of the cell cycle. Biological networks and dynamical systems Biological networks originate from evolution and therefore have less standardized components and potentially more complex interactions than networks designed by humans, such as electrical networks. At the cellular level, components of a network can include a large variety of proteins, many of which differ between organisms. Network interactions occur when one or more proteins affect the function of another through transcription, translation, translocation, phosphorylation, or other mechanisms. These interactions either activate or inhibit the action of the target protein in some way. While humans build networks with a concern for simplicity and order, biological networks acquire redundancy and complexity over the course of evolution. Therefore, it can be impossible to predict the quantitative behavior of a biological network from knowledge of its organization. Similarly, it is impossible to describe its organization purely from its behavior, though behavior can indicate the presence of certain network motifs. However, with knowledge of network interactions and a set of parameters for the proteins and protein interactions (usually obtained through empirical research), it is often possible to construct a model of the network as a dynamical system. In general, for n proteins, the dynamical system takes the following form where x is typically protein concentration: These systems are often very diffi
https://en.wikipedia.org/wiki/Internal%20sort
An internal sort is any data sorting process that takes place entirely within the main memory of a computer. This is possible whenever the data to be sorted is small enough to all be held in the main memory. like a hard-disk. Any reading or writing of data to and from this slower media can slow the sortation process considerably. This issue has implications for different sort algorithms. Some common internal sorting algorithms include: Bubble Sort Insertion Sort Quick Sort Heap Sort Radix Sort Selection sort Consider a Bubblesort, where adjacent records are swapped in order to get them into the right order, so that records appear to “bubble” up and down through the dataspace. If this has to be done in chunks, then when we have sorted all the records in chunk 1, we move on to chunk 2, but we find that some of the records in chunk 1 need to “bubble through” chunk 2, and vice versa (i.e., there are records in chunk 2 that belong in chunk 1, and records in chunk 1 that belong in chunk 2 or later chunks). This will cause the chunks to be read and written back to disk many times as records cross over the boundaries between them, resulting in a considerable degradation of performance. If the data can all be held in memory as one large chunk, then this performance hit is avoided. On the other hand, some algorithms handle external sorting rather better. A Merge sort breaks the data up into chunks, sorts the chunks by some other algorithm (maybe bubblesort or Quick sort) and then recombines the chunks two by two so that each recombined chunk is in order. This approach minimises the number or reads and writes of data-chunks from disk, and is a popular external sort method. Sorting algorithms
https://en.wikipedia.org/wiki/Digital%20control
Digital control is a branch of control theory that uses digital computers to act as system controllers. Depending on the requirements, a digital control system can take the form of a microcontroller to an ASIC to a standard desktop computer. Since a digital computer is a discrete system, the Laplace transform is replaced with the Z-transform. Since a digital computer has finite precision (See quantization), extra care is needed to ensure the error in coefficients, analog-to-digital conversion, digital-to-analog conversion, etc. are not producing undesired or unplanned effects. Since the creation of the first digital computer in the early 1940s the price of digital computers has dropped considerably, which has made them key pieces to control systems because they are easy to configure and reconfigure through software, can scale to the limits of the memory or storage space without extra cost, parameters of the program can change with time (See adaptive control) and digital computers are much less prone to environmental conditions than capacitors, inductors, etc. Digital controller implementation A digital controller is usually cascaded with the plant in a feedback system. The rest of the system can either be digital or analog. Typically, a digital controller requires: Analog-to-digital conversion to convert analog inputs to machine-readable (digital) format Digital-to-analog conversion to convert digital outputs to a form that can be input to a plant (analog) A program that relates the outputs to the inputs Output program Outputs from the digital controller are functions of current and past input samples, as well as past output samples - this can be implemented by storing relevant values of input and output in registers. The output can then be formed by a weighted sum of these stored values. The programs can take numerous forms and perform many functions A digital filter for low-pass filtering A state space model of a system to act as a state observer A telemetr
https://en.wikipedia.org/wiki/Subpreputial%20wetness
Subpreputial wetness is a lanolin-like wetness commonly found under the prepuce in most female and male individuals who have not undergone circumcision, as well as in the female vulva area. If allowed to accumulate along with dead epithelial cells, the combination is commonly called smegma. Smegma is a whitish substance which can be removed by washing, whereas subpreputial wetness is a transparent lubricating secretion. The two substances are commonly confused.
https://en.wikipedia.org/wiki/Softonic
Softonic.com is a web portal based in Barcelona, Catalonia, Spain. It was founded in June 1997 and is owned by Softonic International. History Softonic started in 1996 as a file-oriented download service called Shareware Intercom, at Intercom Online (Grupo Intercom), a provider of Internet services in Cerdanyola del Vallès, near Barcelona (Catalonia, Spain). Files for the website originally came from the "Files Library Intercom BBS" at Intercom Online, using a "5-step semi-automated process", which was later greatly simplified. The company also offered monthly CDs of its software library for users tired of "having to wait for downloads to complete or pay exorbitant phone bills." In 1999, the service was described as having a "generosity of games, antivirus, education and the long, seductive etcetera." In 2000, the company became independent under the name Softonic. In 2004, Softonic was made available in German, and in 2005, English. Initially the service was oriented only to downloads of DOS and Windows software, but it evolved to offer software downloads for Mac and mobile platforms. Softonic was listed as the "Best place to work in Spain (100–250 employees)" in an annual study by the Great Place to Work Institute from 2009–2011, and as second-best in 2008. In 2009 and 2011, the site was listed as having the most unique visitor traffic in Spain by audit bureau OJD Interactiva. At the end of 2014, Softonic announced that a headcount reduction procedure would be carried out as a cost-cutting measure, allegedly due to financial and organizational reasons. In December, the company formalized the dismissal of 156 employees, against the initial estimation of 207. In February 2015, Softonic announced that industry veteran and Download.com co-founder Scott Arpajian had been appointed as CEO of Softonic, replacing Founder Tomás Diago. In July 2016, Softonic announced a partnership with business software platform Crozdesk. In May 2017, the company announced that it h
https://en.wikipedia.org/wiki/Borg%20%28cluster%20manager%29
Borg is a cluster manager used by Google. It led to widespread use of similar approaches, such as Docker and Kubernetes. See also Apache Mesos List of cluster management software Kubernetes DC/OS Operating-system-level virtualization (containerization)
https://en.wikipedia.org/wiki/Pyruvate%20cycling
Pyruvate cycling commonly refers to an intracellular loop of spatial movements and chemical transformations involving pyruvate. Spatial movements occur between mitochondria and cytosol and chemical transformations create various Krebs cycle intermediates. In all variants, pyruvate is imported into the mitochondrion for processing through part of the Krebs cycle. In addition to pyruvate, alpha-ketoglutarate may also be imported. At various points, the intermediate product is exported to the cytosol for additional transformations and then re-imported. Three specific pyruvate cycles are generally considered, each named for the principal molecule exported from the mitochondrion: malate, citrate, and isocitrate. Other variants may exist, such as dissipative or "futile" pyruvate cycles. This cycle is usually studied in relation to Glucose Stimulated Insulin Secretion ( or GSIS ) and there is thought to be a relationship between the insulin response and NADPH produced from this cycle but the specifics are not clear and particular confusion exists about the role of malic enzymes. It has been observed in various cell types including islet cells. The pyruvate-malate cycle was described in liver and kidney preparations as early as 1971.
https://en.wikipedia.org/wiki/Glitter%20cell
Glitter cells (also called Sternheimer-Malbin positive cells) are polymorphonuclear leukocyte neutrophils with granules that show a Brownian movement and that are found in the urine, most commonly associated with urinary tract infections or pyelonephritis and especially prevalent under conditions of hypotonic urine (samples with specific gravity less than 1.01). First described in 1908, they derive their name from their appearance when viewed on a wet mount preparation under a microscope; the granules within their cytoplasm can be seen moving, giving them a "glittering appearance." due to swelling of the neutrophil as result of hypotonicity. In addition to a glittering morphology, glitter cells also exhibit a colorless or pale blue nuclei and pale blue or gray cytoplasmic region when stained with Sternheimer-Malbin Stain. The presence of glitter cells may be indicative of inflammatory changes in the bladder and kidney.
https://en.wikipedia.org/wiki/Hexagon
In geometry, a hexagon (from Greek , , meaning "six", and , , meaning "corner, angle") is a six-sided polygon. The total of the internal angles of any simple (non-self-intersecting) hexagon is 720°. Regular hexagon A regular hexagon has Schläfli symbol {6} and can also be constructed as a truncated equilateral triangle, t{3}, which alternates two types of edges. A regular hexagon is defined as a hexagon that is both equilateral and equiangular. It is bicentric, meaning that it is both cyclic (has a circumscribed circle) and tangential (has an inscribed circle). The common length of the sides equals the radius of the circumscribed circle or circumcircle, which equals times the apothem (radius of the inscribed circle). All internal angles are 120 degrees. A regular hexagon has six rotational symmetries (rotational symmetry of order six) and six reflection symmetries (six lines of symmetry), making up the dihedral group D6. The longest diagonals of a regular hexagon, connecting diametrically opposite vertices, are twice the length of one side. From this it can be seen that a triangle with a vertex at the center of the regular hexagon and sharing one side with the hexagon is equilateral, and that the regular hexagon can be partitioned into six equilateral triangles. Like squares and equilateral triangles, regular hexagons fit together without any gaps to tile the plane (three hexagons meeting at every vertex), and so are useful for constructing tessellations. The cells of a beehive honeycomb are hexagonal for this reason and because the shape makes efficient use of space and building materials. The Voronoi diagram of a regular triangular lattice is the honeycomb tessellation of hexagons. It is not usually considered a triambus, although it is equilateral. Parameters The maximal diameter (which corresponds to the long diagonal of the hexagon), D, is twice the maximal radius or circumradius, R, which equals the side length, t. The minimal diameter or the diameter
https://en.wikipedia.org/wiki/Hotelling%27s%20lemma
Hotelling's lemma is a result in microeconomics that relates the supply of a good to the maximum profit of the producer. It was first shown by Harold Hotelling, and is widely used in the theory of the firm. Specifically, it states: The rate of an increase in maximized profits with respect to a price increase is equal to the net supply of the good. In other words, if the firm makes its choices to maximize profits, then the choices can be recovered from the knowledge of the maximum profit function. Formal Statement Let denote a variable price, and be a constant cost of each input. Let be a mapping from the price to a set of feasible input choices . Let be the production function, and be the net supply. The maximum profit can be written by Then the lemma states that if the profit is differentiable at , the maximizing net supply is given by Proof for Hotelling's lemma The lemma is a corollary of the envelope theorem. Specifically, the maximum profit can be rewritten as where is the maximizing input corresponding to . Due to the optimality, the first order condition gives By taking the derivative by at , where the second equality is due to (). QED Application of Hotelling's lemma Consider the following example. Let output have price and inputs and have prices and . Suppose the production function is . The unmaximized profit function is . From this can be derived the profit-maximizing choices of inputs and the maximized profit function, a function just of the input and output prices, which is Hotelling's Lemma says that from the maximized profit function we can find the profit-maximizing choices of output and input by taking partial derivatives: Note that Hotelling's lemma gives the net supplies, which are positive for outputs and negative for inputs, since profit rises with output prices and falls with input prices. Criticisms and empirical evidence A number of criticisms have been made with regards to the use and application of Hotellin
https://en.wikipedia.org/wiki/SequenceBase
SequenceBase is a privately held company, is an international patent sequence information provider with headquarters located in Edison, NJ, USA. SequenceBase develops and markets the SequenceBase Research Portal to the biotechnology, legal, pharmaceutical, scientific, technical and academic bioinformatics communities. Clarivate Analytics has acquired SequenceBase on 9th September 2019. USGENE provides searchable access to all available peptide and nucleotide sequences from the published applications and issued patents of the United States Patent and Trademark Office (USPTO). USGENE can be searched directly via the SequenceBase Research Portal or via STN International by FIZ Karlsruhe. The SequenceBase Research Portal offers BLAST+ as a sequence searching method.
https://en.wikipedia.org/wiki/Tympanic%20part%20of%20the%20temporal%20bone
The tympanic part of the temporal bone is a curved plate of bone lying below the squamous part of the temporal bone, in front of the mastoid process, and surrounding the external part of the ear canal. It originates as a separate bone (tympanic bone), which in some mammals stays separate through life. Evolutionarily, a portion of it is derived from the angular bone of the reptilian lower jaw. Surfaces Its postero-superior surface is concave, and forms the anterior wall, the floor, and part of the posterior wall of the bony ear canal. Medially, it presents a narrow furrow, the tympanic sulcus, for the attachment of the tympanic membrane. Its antero-inferior surface is quadrilateral and slightly concave; it constitutes the posterior boundary of the mandibular fossa, and is in contact with the retromandibular part of the parotid gland. Borders Its lateral border is free and rough, and gives attachment to the cartilaginous part of the ear canal. Internally, the tympanic part is fused with the petrous portion, and appears in the retreating angle between it and the squama, where it lies below and lateral to the orifice of the auditory tube. Posteriorly, it blends with the squama and mastoid part, and forms the anterior boundary of the tympanomastoid fissure. Its upper border fuses laterally with the back of the postglenoid process, while medially it bounds the petrotympanic fissure. The medial part of the lower border is thin and sharp; its lateral part splits to enclose the root of the styloid process, and is therefore named the vaginal process. Ear canal The central portion of the tympanic part is thin, as it gives rise to the bony inner two-thirds of the ear canal, and in 5 - 20% of skulls the lower surface is perforated by a hole, the foramen of Huschke that opens onto the temporomandibular joint due to incomplete fusion of the anterior and posterior prominences during development. The bony portion of the ear canal is nearly 2 cm long and is directed inwa
https://en.wikipedia.org/wiki/SequenceVariantAnalyzer
SequenceVariantAnalyzer (SVA) is a computer program for annotating and analyzing genetic variants called (identified) from a whole genome or exome sequencing study (Shotgun sequencing). Introduction Background DNA sequence information underpins genetic research, enabling discoveries of important biological or medical benefit. Compared with previous discovery strategies, a whole-genome sequencing study is no longer constrained by differing patterns of linkage disequilibrium, thus, in theory, is more possible to directly identify the genetic variants contributing to biological traits or medical outcomes. The rapidly evolving high-throughput DNA sequencing technologies have now allowed the rapid generation of large amounts of sequence data for the purpose of performing such whole-genome sequencing studies, at a reasonable cost. SequenceVariantAnalyzer, or SVA, is software that analyzes genetic variants identified in such studies. Functions SVA is designed for two specific aims: (1) To annotate the biological functions of the identified genetic variants and group them, conveniently; (2) To find the genetic variants that are associated with or responsible for the biological traits or medical outcomes of interest. Language SVA is developed on the Java platform. Authors SVA is developed and maintained by Dr. Dongliang Ge and Dr. David B. Goldstein at Duke University, Center for Human Genome Variation.
https://en.wikipedia.org/wiki/Objective-collapse%20theory
Objective-collapse theories, also known as models of spontaneous wave function collapse or dynamical reduction models, are proposed solutions to the measurement problem in quantum mechanics. As with other theories called interpretations of quantum mechanics, they are possible explanations of why and how quantum measurements always give definite outcomes, not a superposition of them as predicted by the Schrödinger equation, and more generally how the classical world emerges from quantum theory. The fundamental idea is that the unitary evolution of the wave function describing the state of a quantum system is approximate. It works well for microscopic systems, but progressively loses its validity when the mass / complexity of the system increases. In collapse theories, the Schrödinger equation is supplemented with additional nonlinear and stochastic terms (spontaneous collapses) which localize the wave function in space. The resulting dynamics is such that for microscopic isolated systems, the new terms have a negligible effect; therefore, the usual quantum properties are recovered, apart from very tiny deviations. Such deviations can potentially be detected in dedicated experiments, and efforts are increasing worldwide towards testing them. An inbuilt amplification mechanism makes sure that for macroscopic systems consisting of many particles, the collapse becomes stronger than the quantum dynamics. Then their wave function is always well-localized in space, so well-localized that it behaves, for all practical purposes, like a point moving in space according to Newton's laws. In this sense, collapse models provide a unified description of microscopic and macroscopic systems, avoiding the conceptual problems associated to measurements in quantum theory. The most well-known examples of such theories are: Ghirardi–Rimini–Weber (GRW) model Continuous spontaneous localization (CSL) model Diósi–Penrose (DP) model Collapse theories stand in opposition to many-world
https://en.wikipedia.org/wiki/List%20of%20allergens
This is a list of allergies, which includes the allergen, potential reactions, and a brief description of the cause where applicable. Allergens Food Medical case of dying Environmental Contact Many substances can cause an allergic reaction when in contact with the human integumentary system. See also Allergic inflammation Elimination diet Food intolerance Oral allergy syndrome Sweat allergy List of inclusion bodies that aid in diagnosis of cutaneous conditions List of cutaneous conditions List of genes mutated in cutaneous conditions List of target antigens in pemphigus List of specialized glands within the human integumentary system
https://en.wikipedia.org/wiki/Zero-knowledge%20password%20proof
In cryptography, a zero-knowledge password proof (ZKPP) is a type of zero-knowledge proof that allows one party (the prover) to prove to another party (the verifier) that it knows a value of a password, without revealing anything other than the fact that it knows the password to the verifier. The term is defined in IEEE P1363.2, in reference to one of the benefits of using a password-authenticated key exchange (PAKE) protocol that is secure against off-line dictionary attacks. A ZKPP prevents any party from verifying guesses for the password without interacting with a party that knows it and, in the optimal case, provides exactly one guess in each interaction. A common use of a zero-knowledge password proof is in authentication systems where one party wants to prove its identity to a second party using a password but doesn't want the second party or anybody else to learn anything about the password. For example, apps can validate a password without processing it and a payment app can check the balance of an account without touching or learning anything about the amount. History The first methods to demonstrate a ZKPP were the encrypted key exchange methods (EKE) described by Steven M. Bellovin and Michael Merritt in 1992. A considerable number of refinements, alternatives, and variations in the growing class of password-authenticated key agreement methods were developed in subsequent years. Standards for these methods include IETF , IEEE P1363.2, and ISO-IEC 11770-4. See also Cryptographic protocol Outline of cryptography Key-agreement protocol Secure Remote Password protocol
https://en.wikipedia.org/wiki/Kunitz%20STI%20protease%20inhibitor
Kunitz soybean trypsin inhibitor is a type of protein contained in legume seeds which functions as a protease inhibitor. Kunitz-type Soybean Trypsin Inhibitors are usually specific for either trypsin or chymotrypsin. They are thought to protect seeds against consumption by animal predators. Background Two types of trypsin inhibitors are found in soy: the Kunitz-type soybean trypsin inhibitor (STI, discovered by Moses Kunitz and sometimes abbreviated as KTI) and the Bowman-Birk inhibitor (BBI). STI is a large (20,100 daltons), strong inhibitor of trypsin, while BBI is much smaller (8,000 daltons) and inhibits both trypsin and chymotrypsin. Both inhibitors have significant anti-nutritive effects in the body, affecting digestion by hindering protein hydrolysis and activation of other enzymes in the gut. STI is found in much larger concentrations than BBI in soy, however, to achieve the highest nutritional value from soy, both of these inhibitors must be denatured in some way. Whole soybeans have been reported to contain 17–27 mg of trypsin inhibitor per gram. Protease inhibitory activity is decreased by cooking soybeans, leading to low levels in soy products such as tofu and soy milk. Structure Proteins from the Kunitz family contain from 170 to 200 amino acid residues and one or two intra-chain disulfide bonds. The best conserved region is found in their N-terminal section. The crystal structures of soybean trypsin inhibitor (STI), trypsin inhibitor DE-3 from the Kaffir tree Erythrina caffra (ETI) and the bifunctional proteinase K/alpha-amylase inhibitor from wheat (PK13) have been solved, showing them to share the same beta trefoil fold structure as those of interleukin 1 and heparin-binding growth factors. Despite the structural similarity, STI shows no interleukin-1 bioactivity, presumably as a result of their primary sequence disparities. The active inhibitory site containing the scissile bond is located in the loop between beta-strands 4 and 5 in STI and ETI.
https://en.wikipedia.org/wiki/Electronic%20symbol
An electronic symbol is a pictogram used to represent various electrical and electronic devices or functions, such as wires, batteries, resistors, and transistors, in a schematic diagram of an electrical or electronic circuit. These symbols are largely standardized internationally today, but may vary from country to country, or engineering discipline, based on traditional conventions. Standards for symbols The graphic symbols used for electrical components in circuit diagrams are covered by national and international standards, in particular: IEC 60617 (also known as BS 3939). There is also IEC 61131-3 – for ladder-logic symbols. JIC JIC (Joint Industrial Council) symbols as approved and adopted by the NMTBA (National Machine Tool Builders Association). They have been extracted from the Appendix of the NMTBA Specification EGPl-1967. ANSI Y32.2-1975 (also known as IEEE Std 315-1975 or CSA Z99-1975). IEEE Std 91/91a: graphic symbols for logic functions (used in digital electronics). It is referenced in ANSI Y32.2/IEEE Std 315. Australian Standard AS 1102 (based on a slightly modified version of IEC 60617; withdrawn without replacement with a recommendation to use IEC 60617). The number of standards leads to confusion and errors. Symbols usage is sometimes unique to engineering disciplines, and national or local variations to international standards exist. For example, lighting and power symbols used as part of architectural drawings may be different from symbols for devices used in electronics. Common electronic symbols Symbols shown are typical examples, not a complete list. Traces Grounds The shorthand for ground is GND. Optionally, the triangle in the middle symbol may be filled in. Sources Resistors It is very common for potentiometer and rheostat symbols to be used for many types of variable resistors, including trimmers. Capacitors Diodes Optionally, the triangle in these symbols may be filled in. Note: The words anode and cathode typically
https://en.wikipedia.org/wiki/Swedish%20grid
The Swedish grid (in Swedish Rikets Nät, RT 90) is a coordinate system that was previously used for government maps in Sweden. RT 90 is a slightly modified version of the RT 38 from 1938. RT 90 has been replaced with SWEREF 99 as the official Swedish spatial reference system. While the system could be used with negative numbers to represent all four "quarters" of the earth (NE, NW, SE, and SW hemispheres), the standard application of RT 90 is only useful for the northern half of the eastern hemisphere where numbers are positive. The coordinate system is based on metric measures rooting from the crossing of the Prime Meridian and the Equator at 0,0. The Central Meridian used to be based on a meridian located at the old observatory in Stockholm, but today it is based on the Prime Meridian at Greenwich. The numbering system's first digit represents the largest distance, followed by what can be seen as fractional decimal digits (though without an explicit decimal point). Therefore, X 65 is located halfway between X 6 and X 7. The coordinate grid is specified using two numbers, named X and Y, X being the south–north axis and Y the west–east axis. Two seven-digit numbers are sufficient to specify a location with a one m resolution. Example: X=6620000 Y=1317000 (X is the northing and Y is the easting) denotes a position 6620 km north of the Equator and -183 km (1317 km-1500 km) west of the Central Meridian, which happens to be somewhere near the town center of Arvika. RT90 Map Projection Parameters
https://en.wikipedia.org/wiki/Solaris%20Multiplexed%20I/O
Solaris Multiplexed I/O (MPxIO), known also as Sun StorageTek Traffic Manager (SSTM, earlier Sun StorEdge Traffic Manager), is multipath I/O software for Solaris/illumos. It enables a storage device to be accessed through multiple host controller interfaces from a single operating system instance. The MPxIO architecture helps protect against I/O outages due to I/O controller failures. Should one I/O controller fail, MPxIO automatically switches to an alternate controller. This architecture also increases I/O performance by load balancing across multiple I/O channels. It was integrated within the Solaris operating system beginning in February 2000 with Solaris 8 release. The file to enable or disable mpxio has been moved in Solaris 10 from /kernel/drv/scsi_vhci.conf to the bottom of the file /kernel/drv/fp.conf and /kernel/drv/mpt.conf. See also Multipath I/O
https://en.wikipedia.org/wiki/Easter%20egg%20%28media%29
An Easter egg is a message, image, or feature hidden in software, a video game, a film, or another — usually electronic — medium. The term used in this manner was coined around 1979 by Steve Wright, the then-Director of Software Development in the Atari Consumer Division, to describe a hidden message in the Atari video game Adventure, in reference to an Easter egg hunt. The earliest known video game Easter egg is in the 1973 video game Moonlander, in which the player tries to land a Lunar module on the moon; if the player opts to fly the module horizontally through several of the game's screens, they encounter a McDonald's restaurant, and if they land next to it the astronaut will visit it instead of standing next to the ship. The earliest known Easter egg in software in general is one placed in the "make" command for PDP-6/PDP-10 computers sometime in October 1967–October 1968, wherein if the user attempts to create a file named "love" by typing "make love", the program responds "not war?" before proceeding. Origin The use of the term "Easter egg" to describe secret features in video games originates from the 1980 video game Adventure for the Atari 2600 game console, programmed by employee Warren Robinett. At the time, Atari did not include programmers' names in the game credits, both to prevent competitors from poaching its developers as well as to deny developers a means to bargain with the management of the new owners, Warner Communications. Robinett, who disagreed with his supervisor over this lack of acknowledgment, secretly programmed the message "Created by Warren Robinett" to appear only if a player moves their avatar over a specific pixel (dubbed the "Gray Dot") during a certain part of the game and enters a previously "forbidden" part of the map where the message can be found. When Robinett left Atari, he did not inform the company of the acknowledgment that he included in the game. Shortly after his departure, the "Gray Dot" and his message were discov
https://en.wikipedia.org/wiki/Bromacil
Bromacil is an organic compound with the chemical formula C9H13BrN2O2, commercially available as a herbicide. Bromacil was first registered as a pesticide in the U.S. in 1961. It is used for brush control on non-cropland areas. It works by interfering with photosynthesis by entering the plant through the root zone and moving throughout the plant. Bromacil is one of a group of compounds called substituted uracils. These materials are broad spectrum herbicides used for nonselective weed and brush control on non-croplands, as well as for selective weed control on a limited number of crops, such as citrus fruit and pineapple. Bromacil is also found to be excellent at controlling perennial grasses. Safety There are quite a few safety precautions that should be taken when dealing with Bromacil. Dry formulations containing bromacil must bear the word "Caution" and liquid formulas must signal "Warning." Care should be exercised when spraying Bromacil on plants because it will also stop the photosynthesis of the adjacent non-target plants, therefore killing them. Bromacil should never be used in residential or recreation areas for risk of exposure. Bromacil is slightly toxic if individuals accidentally eat or touch residues and practically nontoxic if inhaled. Bromacil is a mild eye irritant and a very slight skin irritant. It is not a skin sensitizer. In studies using laboratory animals, bromacil is slightly toxic by the oral, dermal, and inhalation routes and has been placed in Toxicity Category IV (the lowest of four categories) for these effects. This herbicide should be stored in a cool, dry place and after any handling a thorough hand-washing is advised. In regards to occupational exposure, the National Institute for Occupational Safety and Health has recommended workers handing bromacil not exceed an exposure of 1 ppm (10 mg/m3) over an eight-hour time-weighted average. Facts Bromacil (40%) is combined with the active ingredient diuron in the herbicide Krovar, whic
https://en.wikipedia.org/wiki/Hawthorne%20effect
The Hawthorne effect is a type of human behavior reactivity in which individuals modify an aspect of their behavior in response to their awareness of being observed. The effect was discovered in the context of research conducted at the Hawthorne Western Electric plant; however, some scholars think the descriptions are fictitious. The original research involved workers who made electrical relays at the Hawthorne Works, a Western Electric plant in Cicero, Illinois. Between 1924 and 1927, the lighting study was conducted. Workers experienced a series of lighting changes in which productivity was said to increase with almost any change in the lighting. This turned out not to be true. In the study that was associated with Elton Mayo, which ran from 1928 to 1932, a series of changes in work structure were implemented (e.g., changes in rest periods) in a group of five women. However, this was a methodologically poor, uncontrolled study that did not permit any firm conclusions to be drawn. One of the later interpretations by Landsberger suggested that the novelty of being research subjects and the increased attention from such could lead to temporary increases in workers' productivity. This interpretation was dubbed "the Hawthorne effect". History The term "Hawthorne effect" was coined in 1958 by Henry A. Landsberger when he was analyzing the Hawthorne studies conducted between 1924 and 1932 at the Hawthorne Works (a Western Electric factory in Cicero, outside Chicago). The Hawthorne Works had commissioned a study to determine if its workers would become more productive in higher or lower levels of light. The workers' productivity seemed to improve when changes were made, and slumped when the study ended. It was suggested that the productivity gain occurred as a result of the motivational effect on the workers of the interest being shown in them. This effect was observed for minute increases in illumination. In these lighting studies, light intensity was altered t
https://en.wikipedia.org/wiki/Mercer%20%28consulting%20firm%29
Mercer is an American consulting firm founded in 1945. It is one of the four operating subsidiaries of global professional services firm Marsh McLennan (NYSE: MMC). Mercer is headquartered in New York City with offices in 43 countries and operations in 130 countries. The company primarily provides human resources and financial services consulting services to its clients. Mercer has several distinct lines of business, namely: health and benefits, investments and retirement, workforce and careers, and M&A advisory services. It is the world's largest investment advisory with over US$300 billion outsourced assets under management and US$15 trillion under advisement in total. History Foundation and early years (1937–1959) William Manson Mercer founded William M. Mercer, Limited in Vancouver, Canada in 1945. It was acquired by Marsh McLennan and merged into their employee benefits department in 1959. Post acquisition growth (1959–2002) Mercer Consulting Group In 1975, Marsh McLennan converted their benefits operations into a wholly owned subsidiary, William M. Mercer, Inc. In 1992, a holding company was created for Marsh McLennan's three global consulting businesses, known as Mercer Consulting Group. William M. Mercer, Inc. continued to provide actuarial and employee benefits consulting within the group alongside two sister companies: Mercer Management Consulting and National Economic Research Associates, Inc, which provided corporate strategy consulting and economic consulting, respectively. Mercer Delta Consulting In 2000, Mercer Consulting Group acquired Delta Consulting Group for its organizational development and change management expertise. Founded by organizational theorist David A. Nadler in 1980, Delta Consulting Group worked to structure effective executive teams. The firm had an influential client list, including corporations such as 3M, Citicorp, Procter & Gamble, The New York Times, and Xerox. The new entity was renamed Mercer Delta Consulting, and ma
https://en.wikipedia.org/wiki/Jacobi%20bound%20problem
The Jacobi Bound Problem concerns the veracity of Jacobi's inequality which is an inequality on the absolute dimension of a differential algebraic variety in terms of its defining equations. The inequality is the differential algebraic analog of Bezout's theorem in affine space. Although first formulated by Jacobi, In 1936 Joseph Ritt recognized the problem as non-rigorous in that Jacobi didn't even have a rigorous notion of absolute dimension (Jacobi and Ritt used the term "order" - which Ritt first gave a rigorous definition for using the notion of transcendence degree). Intuitively, the absolute dimension is the number of constants of integration required to specify a solution of a system of ordinary differential equations. A mathematical proof of the inequality has been open since 1936. Statement Let be a differential field of characteristic zero and consider a differential algebraic variety determined by the vanishing of differential polynomials . If is an irreducible component of of finite absolute dimension then In the above display is the *jacobi number*. It is defined to be .
https://en.wikipedia.org/wiki/Lightweight%20software%20test%20automation
Lightweight software test automation is the process of creating and using relatively short and simple computer programs, called lightweight test harnesses, designed to test a software system. Lightweight test automation harnesses are not tied to a particular programming language but are most often implemented with the Java, Perl, Visual Basic .NET, and C# programming languages. Lightweight test automation harnesses are generally four pages of source code or less, and are generally written in four hours or less. Lightweight test automation is often associated with Agile software development methodology. The three major alternatives to the use of lightweight software test automation are commercial test automation frameworks, Open Source test automation frameworks, and heavyweight test automation. The primary disadvantage of lightweight test automation is manageability. Because lightweight automation is relatively quick and easy to implement, a test effort can be overwhelmed with harness programs, test case data files, test result files, and so on. However, lightweight test automation has significant advantages. Compared with commercial frameworks, lightweight automation is less expensive in initial cost and is more flexible. Compared with Open Source frameworks, lightweight automation is more stable because there are fewer updates and external dependencies. Compared with heavyweight test automation, lightweight automation is quicker to implement and modify. Lightweight test automation is generally used to complement, not replace these alternative approaches. Lightweight test automation is most useful for regression testing, where the intention is to verify that new source code added to the system under test has not created any new software failures. Lightweight test automation may be used for other areas of software testing such as performance testing, stress testing, load testing, security testing, code coverage analysis, mutation testing, and so on. The most widel
https://en.wikipedia.org/wiki/Pointed%20space
In mathematics, a pointed space or based space is a topological space with a distinguished point, the basepoint. The distinguished point is just simply one particular point, picked out from the space, and given a name, such as that remains unchanged during subsequent discussion, and is kept track of during all operations. Maps of pointed spaces (based maps) are continuous maps preserving basepoints, i.e., a map between a pointed space with basepoint and a pointed space with basepoint is a based map if it is continuous with respect to the topologies of and and if This is usually denoted Pointed spaces are important in algebraic topology, particularly in homotopy theory, where many constructions, such as the fundamental group, depend on a choice of basepoint. The pointed set concept is less important; it is anyway the case of a pointed discrete space. Pointed spaces are often taken as a special case of the relative topology, where the subset is a single point. Thus, much of homotopy theory is usually developed on pointed spaces, and then moved to relative topologies in algebraic topology. Category of pointed spaces The class of all pointed spaces forms a category Top with basepoint preserving continuous maps as morphisms. Another way to think about this category is as the comma category, ( Top) where is any one point space and Top is the category of topological spaces. (This is also called a coslice category denoted Top.) Objects in this category are continuous maps Such maps can be thought of as picking out a basepoint in Morphisms in ( Top) are morphisms in Top for which the following diagram commutes: It is easy to see that commutativity of the diagram is equivalent to the condition that preserves basepoints. As a pointed space, is a zero object in Top, while it is only a terminal object in Top. There is a forgetful functor Top Top which "forgets" which point is the basepoint. This functor has a left adjoint which assigns to each topologic
https://en.wikipedia.org/wiki/Nimble%20Storage
Nimble Storage, founded in 2008, is a subsidiary of Hewlett Packard Enterprise. It specializes in producing hardware and software products for data storage, particularly data storage arrays that utilize the iSCSI and Fibre Channel protocols, and includes data backup and data protection features. History Nimble Storage was established in January 2008 by Varun Mehta and Umesh Maheshwari. In July 2010, the company announced its first product, the CS200 series hybrid arrays, at Tech Field Day. In September 2012, Nimble Storage secured $40.7 million in funding from both original and new investors, including Artis Capital Management and GGV Capital. Varun Mehta served as the chief executive until March 2011 when he became the vice president of engineering. Suresh Vasudevan assumed the role of CEO, and Umesh Maheshwari became the chief technology officer. In October 2013, the company filed for its initial public offering on the New York Stock Exchange and subsequently went public on December 13, 2013, under the ticker symbol NMBL. Throughout its history, Nimble Storage introduced various product updates and expansions. In June 2014, the company announced the CS700 Series Arrays and an All-Flash Shelf, along with its Adaptive Flash technology. In November 2014, Nimble Storage released arrays supporting the Fibre Channel protocol. In July 2015, the company announced updates to the Adaptive Flash platform, including Nimble SmartSecure (software-based encryption), all-flash service levels, REST APIs, InfoSight-VMVision per-VM monitoring, and integrated data protection. Nimble Storage also achieved Federal Information Processing Standard 140-2 Certification for the Adaptive Flash platform in August 2015. The company experienced significant growth and recognition, ranking sixth on Deloitte's list of the 500 fastest-growing technology, media, telecommunications, life sciences, and energy tech companies in North America in November 2015. On February 23, 2016, Nimble Storag
https://en.wikipedia.org/wiki/Anisophylly
Anisophylly is when leaves of a pair differ from one another, either in size or in shape. When a horizontal stem (plagiotropic shoot) also exhibits anisophylly, the photosynthetic leaf surfaces interfere less with light from above, and rotation of the leaf or the petiole can enhance that effect. The phenomenon is relatively common in some tropical plant families with decussate leaf arrangement, such as Melastomataceae, Gesneriaceae and Urticaceae as well as in certain species of other families.
https://en.wikipedia.org/wiki/Thermal%20barrier%20coating
Thermal barrier coatings (TBCs) are advanced materials systems usually applied to metallic surfaces on parts operating at elevated temperatures, such as gas turbine combustors and turbines, and in automotive exhaust heat management. These 100 μm to 2 mm thick coatings of thermally insulating materials serve to insulate components from large and prolonged heat loads and can sustain an appreciable temperature difference between the load-bearing alloys and the coating surface. In doing so, these coatings can allow for higher operating temperatures while limiting the thermal exposure of structural components, extending part life by reducing oxidation and thermal fatigue. In conjunction with active film cooling, TBCs permit working fluid temperatures higher than the melting point of the metal airfoil in some turbine applications. Due to increasing demand for more efficient engines running at higher temperatures with better durability/lifetime and thinner coatings to reduce parasitic mass for rotating/moving components, there is significant motivation to develop new and advanced TBCs. The material requirements of TBCs are similar to those of heat shields, although in the latter application emissivity tends to be of greater importance. Structure An effective TBC needs to meet certain requirements to perform well in aggressive thermo-mechanical environments. To deal with thermal expansion stresses during heating and cooling, adequate porosity is needed, as well as appropriate matching of thermal expansion coefficients with the metal surface that the TBC is coating. Phase stability is required to prevent significant volume changes (which occur during phase changes), which would cause the coating to crack or spall. In air-breathing engines, oxidation resistance is necessary, as well as decent mechanical properties for rotating/moving parts or parts in contact. Therefore, general requirements for an effective TBC can be summarized as needing: 1) a high melting point. 2) no p
https://en.wikipedia.org/wiki/Episodic%20tremor%20and%20slip
Episodic tremor and slip (ETS) is a seismological phenomenon observed in some subduction zones that is characterized by non-earthquake seismic rumbling, or tremor, and slow slip along the plate interface. Slow slip events are distinguished from earthquakes by their propagation speed and focus. In slow slip events, there is an apparent reversal of crustal motion, although the fault motion remains consistent with the direction of subduction. ETS events themselves are imperceptible to human beings and do not cause damage. Discovery Nonvolcanic, episodic tremor was first identified in southwest Japan in 2002. Shortly afterwards, the Geological Survey of Canada coined the term "episodic tremor and slip" to characterize observations of GPS measurements in the Vancouver Island area. Vancouver Island lies in the eastern, North American region of the Cascadia subduction zone. ETS events in Cascadia were observed to reoccur cyclically with a period of approximately 14 months. Analysis of measurements led to the successful prediction of ETS events in following years (e.g., 2003, 2004, 2005, and 2007). In Cascadia, these events are marked by about two weeks of 1 to 10 Hz seismic trembling and non-earthquake ("aseismic") slip on the plate boundary equivalent to a magnitude 7 earthquake. (Tremor is a weak seismological signal only detectable by very sensitive seismometers.) Recent episodes of tremor and slip in the Cascadia region have occurred down-dip of the region ruptured in the 1700 Cascadia earthquake. Since the initial discovery of this seismic mode in the Cascadia region, slow slip and tremor have been detected in other subduction zones around the world, including Japan and Mexico. Slow slip is not accompanied by tremor in the Hikurangi Subduction Zone. Every five years a year-long quake of this type occurs beneath the New Zealand capital, Wellington. It was first measured in 2003, and has reappeared in 2008 and 2013. Characteristics Slip behaviour In the Casca
https://en.wikipedia.org/wiki/Herbarium
A herbarium (plural: herbaria) is a collection of preserved plant specimens and associated data used for scientific study. The specimens may be whole plants or plant parts; these will usually be in dried form mounted on a sheet of paper (called exsiccatum, plur. exsiccata) but, depending upon the material, may also be stored in boxes or kept in alcohol or other preservative. The specimens in a herbarium are often used as reference material in describing plant taxa; some specimens may be types. The same term is often used in mycology to describe an equivalent collection of preserved fungi, otherwise known as a fungarium. A xylarium is a herbarium specialising in specimens of wood. The term hortorium (as in the Liberty Hyde Bailey Hortorium) has occasionally been applied to a herbarium specialising in preserving material of horticultural origin. History The making of herbaria is an ancient phenomenon, at least six centuries old, although the techniques have changed little, and has been an important step in the transformation of the study of plants from a branch of medicine to an independent discipline, and to make available plant material from far away places and over a long period of time. The oldest traditions of making herbarium collections have been traced to Italy. The Bologna physician and botanist, Luca Ghini (1490–1556) reintroduced the study of actual plants as opposed to relying on classical texts, such as Dioscorides, which lacked sufficient accuracy for identification. At first, he needed to make available plant material, even in winter, hence his Hortus hiemalis (winter garden) or Hortus siccus (dry garden). He and his students placed freshly gathered plants between two sheets of paper and applied pressure to flatten them and absorb moisture. The dried specimen was then glued onto a page in a book and annotated. This practice was supplemented by the parallel development of the Hortus simplicium or Orto botanico (botanical garden) to supply material,
https://en.wikipedia.org/wiki/Bacterioplankton
Bacterioplankton refers to the bacterial component of the plankton that drifts in the water column. The name comes from the Ancient Greek word (), meaning "wanderer" or "drifter", and , a Latin term coined in the 19th century by Christian Gottfried Ehrenberg. They are found in both seawater and freshwater. Bacterioplankton occupy a range of ecological niches in marine and aquatic ecosystems. They are both primary producers and primary consumers in these ecosystems and drive global biogeochemical cycling of elements essential for life (e.g., carbon and nitrogen). Many bacterioplankton species are autotrophic, and derive energy from either photosynthesis or chemosynthesis. Photosynthetic bacterioplankton are often categorized as picophytoplankton, and include major cyanobacterial groups such as Prochlorococcus and Synechococcus. Other heterotrophic bacterioplankton are saprotrophic, and obtain energy by consuming organic material produced by other organisms. This material may be dissolved in the medium and taken directly from there, or bacteria may live and grow in association with particulate material such as marine snow. Bacterioplankton play critical roles in global nitrogen fixation, nitrification, denitrification, remineralisation and methanogenesis. Bacterioplankton abundance depends on environmental variables like temperature, nutrient availability and predation. Like other small plankton, the bacterioplankton are preyed upon by zooplankton (usually protozoans), and their numbers are also controlled through infection by bacteriophages. Major groups Photosynthetic bacterioplankton Photosynthetic bacterioplankton are responsible for a large proportion of the total primary production of aquatic food webs, supplying organic compounds to higher trophic levels. These bacteria undergo oxygenic and anoxygenic photosynthesis. Differences between these processes can be seen in the byproducts produced, the primary electron donor, and the light harvesting pigments
https://en.wikipedia.org/wiki/Velvet%20complex
Velvet complex is a group of proteins found in fungi and especially molds that are important in reproduction and production of secondary metabolites including penicillin. The core members of the complex include VeA, LaeA (loss of aflR-expression A), and VelB. Other proteins including VelC and VosA sometimes function in the complex. The proteins were first characterized in Aspergillus nidulans. Some proteins in the complex are light-sensitive, including the founding member, VeA (Velvet A), which was first described in 1965. Four of these proteins, VeA, VelB, VelC, and VosA, have an approximately 200 amino acid domain called the velvet domain. Some fungal infections that are present in humans and sometimes plants have been traced down to certain velvet complex elements. The Velvet complex seems to affect a number of functions that are of pathogenic nature. This process is facilitated, managed and administered by the proteins of velvet complex. Velvet complex proteins have also been encountered in the genome sequence of fungal organisms in the form of transcription factors.
https://en.wikipedia.org/wiki/Disease%20registry
Disease or patient registries are collections of secondary data related to patients with a specific diagnosis, condition, or procedure, and they play an important role in post marketing surveillance of pharmaceuticals. Registries are different from indexes in that they contain more extensive data. In its simplest form, a disease registry could consist of a collection of paper cards kept inside "a shoe box" by an individual physician. Most frequently registries vary in sophistication from simple spreadsheets that only can be accessed by a small group of physicians to very complex databases that are accessed online across multiple institutions. They can provide health providers (or even patients) with reminders to check certain tests in order to reach certain quality goals. Versus electronic medical records Registries are less complex and simpler to set up than electronic medical records that according to a recent survey are only used by 9% of small offices where almost half of the US doctors work. An electronic medical record keeps track of all the patients a doctor follows but a registry only keeps track of a small sub population of patients with a specific condition. Types of medical conditions tracked More than 130 million Americans live with chronic diseases and chronic diseases account for 70% of all deaths in the US. "The medical care costs of people with chronic diseases account for more than 75% of the nation's $2 trillion medical care costs." Registries target certain conditions because medical expenses are unevenly distributed: most health care expenses are spent treating patients with a few chronic conditions. For example, the 2002 expenses with diabetes in the US was $132 billion, and this was around 12% of the US medical budget. Diabetes accounts for 25% of the Medicare budget. Given this, diabetes is one of the conditions targeted by registries. Diabetes is also amenable to this because there is a target population that can be defined according
https://en.wikipedia.org/wiki/Conversations%20%28software%29
Conversations is a free software, instant messaging client application software for Android. It is largely based on recognized open standards such as the Extensible Messaging and Presence Protocol (XMPP) and Transport Layer Security (TLS). The development focus is on secure communication and implementation of XMPP extensions that are important for mobile use. The trade press praised the decentralized and open nature of the transmission network and simple, intuitive usability with user guidance familiar from other applications. It enjoys recognition as a serious attempt to raise the mass suitability of XMPP-based messaging to a competitive level. The source code of the software is maintained at GitHub, and is subject to the terms of the GPL-3.0-only license. The application can be installed for free (or with donations) using F-Droid, or for a fee in the Google Play store. Google recorded over 100,000 installations as of November 2020. Features Conversations includes optional end-to-end encryption (OpenPGP or OMEMO), and in-transit encryption (Transport Layer Security). The latter must be configured by the server computers involved in the messaging. Conversations allows users to have multiple client devices (endpoints) logged into an account simultaneously (through XMPP), and also delivers messages to multiple client devices (synchronization) using the protocol extension "Message Carbons" ("carbon copies", XEP-0280). Files, also encrypted, can be sent between users. Transmitted images are displayed in the conversation view as inline image messages. As an XMPP client, Conversations can exchange messages with other, different XMPP client software, in principle, and is also not tied to a particular vendor's server infrastructure. The following features are also included: Multi-User (Group) chat (MUC) Optional address book integration Support for multiple user accounts or addresses Implemented XMPP extension protocols Conversations handles many XMPP extension p
https://en.wikipedia.org/wiki/Additive%20map
In algebra, an additive map, -linear map or additive function is a function that preserves the addition operation: for every pair of elements and in the domain of For example, any linear map is additive. When the domain is the real numbers, this is Cauchy's functional equation. For a specific case of this definition, see additive polynomial. More formally, an additive map is a -module homomorphism. Since an abelian group is a -module, it may be defined as a group homomorphism between abelian groups. A map that is additive in each of two arguments separately is called a bi-additive map or a -bilinear map. Examples Typical examples include maps between rings, vector spaces, or modules that preserve the additive group. An additive map does not necessarily preserve any other structure of the object; for example, the product operation of a ring. If and are additive maps, then the map (defined pointwise) is additive. Properties Definition of scalar multiplication by an integer Suppose that is an additive group with identity element and that the inverse of is denoted by For any and integer let: Thus and it can be shown that for all integers and all and This definition of scalar multiplication makes the cyclic subgroup of into a left -module; if is commutative, then it also makes into a left -module. Homogeneity over the integers If is an additive map between additive groups then and for all (where negation denotes the additive inverse) and Consequently, for all (where by definition, ). In other words, every additive map is homogeneous over the integers. Consequently, every additive map between abelian groups is a homomorphism of -modules. Homomorphism of -modules If the additive abelian groups and are also a unital modules over the rationals (such as real or complex vector spaces) then an additive map satisfies: In other words, every additive map is homogeneous over the rational numbers. Consequently, every additive
https://en.wikipedia.org/wiki/Cremona%E2%80%93Richmond%20configuration
In mathematics, the Cremona–Richmond configuration is a configuration of 15 lines and 15 points, having 3 points on each line and 3 lines through each point, and containing no triangles. It was studied by and . It is a generalized quadrangle with parameters (2,2). Its Levi graph is the Tutte–Coxeter graph. Symmetry The points of the Cremona–Richmond configuration may be identified with the unordered pairs of elements of a six-element set; these pairs are called duads. Similarly, the lines of the configuration may be identified with the 15 ways of partitioning the same six elements into three pairs; these partitions are called synthemes. Identified in this way, a point of the configuration is incident to a line of the configuration if and only if the duad corresponding to the point is one of the three pairs in the syntheme corresponding to the line. The symmetric group of all permutations of the six elements underlying this system of duads and synthemes acts as a symmetry group of the Cremona–Richmond configuration, and gives the automorphism group of the configuration. Every flag of the configuration (an incident point-line pair) can be taken to every other flag by a symmetry in this group. The Cremona–Richmond configuration is self-dual: it is possible to exchange points for lines while preserving all the incidences of the configuration. This duality gives the Tutte–Coxeter graph additional symmetries beyond those of the Cremona–Richmond configuration, which swap the two sides of its bipartition. These symmetries correspond to the outer automorphisms of the symmetric group on six elements. Realization Any six points in general position in four-dimensional space determine 15 points where a line through two of the points intersects the hyperplane through the other four points; thus, the duads of the six points correspond one-for-one with these 15 derived points. Any three duads that together form a syntheme determine a line, the intersection line of the three h
https://en.wikipedia.org/wiki/Pythagorean%20tuning
Pythagorean tuning is a system of musical tuning in which the frequency ratios of all intervals are based on the ratio 3:2. This ratio, also known as the "pure" perfect fifth, is chosen because it is one of the most consonant and easiest to tune by ear and because of importance attributed to the integer 3. As Novalis put it, "The musical proportions seem to me to be particularly correct natural proportions." Alternatively, it can be described as the tuning of the syntonic temperament in which the generator is the ratio 3:2 (i.e., the untempered perfect fifth), which is ≈ 702 cents wide. The system dates to Ancient Mesopotamia; see . The system is named, and has been widely misattributed, to Ancient Greeks, notably Pythagoras (sixth century BC) by modern authors of music theory, while Ptolemy, and later Boethius, ascribed the division of the tetrachord by only two intervals, called "semitonium", "tonus", "tonus" in Latin (256:243 × 9:8 × 9:8), to Eratosthenes. The so-called "Pythagorean tuning" was used by musicians up to the beginning of the 16th century. "The Pythagorean system would appear to be ideal because of the purity of the fifths, but some consider other intervals, particularly the major third, to be so badly out of tune that major chords [may be considered] a dissonance." The Pythagorean scale is any scale which can be constructed from only pure perfect fifths (3:2) and octaves (2:1). In Greek music it was used to tune tetrachords, which were composed into scales spanning an octave. A distinction can be made between extended Pythagorean tuning and a 12-tone Pythagorean temperament. Extended Pythagorean tuning corresponds 1-on-1 with western music notation and there is no limit to the number of fifths. In 12-tone Pythagorean temperament however one is limited by 12-tones per octave and one cannot play most music according to the Pythagorean system corresponding to the enharmonic notation, instead one finds that for instance the diminished sixth becomes a
https://en.wikipedia.org/wiki/Air2Web
The Air2Web company is an Atlanta-based Wireless Application Service Provider that was launched in 1999. They provided integrated wireless services, functioning as an intermediary between a company and its wireless carrier. Since 2011, Air2Web is owned by mobile marketing solutions provider Velti. Their product Always Interactive accepted incoming data requests, then packaged up the request in XML and sent it to the corporate database. The result was returned in XML, which the Air2Web software provided as a graphical display or a voice message to the wireless device. Among the custom services that Air2Web provided is wireless package tracking for United Parcel Service. In 2005, they acquired UK company Consult Mobile. Air2Web was acquired by Velti for $19M in 2011. Thomas Cotney, then CEO of Air2Web, executed the sale to Velti. Between 1999 and 2011 the company won multiple awards, including the 2007 Mobility Award for Best Enterprise Messaging Solution from consulting company Frost & Sullivan. Products As of 2011 major products were (some or all services now integrated into Velti's offer): AirCARE – application suite for mobile communications. CampaignPRO – create and administer interactive sales campaigns for mobile marketing. Mobile Marketing – create personalized sales offers through mobile communications. Mobile Assist – interactive chat through mobile devices. DirectTEXT – message routing and delivery. Enterprise Agent – create customized programs to access data on mobile device. Mobile Dev Suite – develop applications for mobile devices.
https://en.wikipedia.org/wiki/The%20Gloomy%20Prospect
In behavioral genetics and epidemiology, the "Gloomy Prospect" refers to the notion that non-shared environmental influences are unsystematic, idiosyncratic, serendipitous events. It is generally used to describe the messy and individualized tiny and innumerable, but causal environmental effects. It can also be used as a label. The phrase, "gloomy prospect", was coined in a paper by Plomin and Daniels (1987), a review of biometric family studies, which intended to study why there existed such a large variation of traits of siblings despite being within the same family. They explain that, "One gloomy prospect is that the salient environment might be unsystematic, idiosyncratic, or serendipitous events such as accidents, illnesses, and other traumas, as biographies often attest... It is possible that nonshared environmental influences could be unsystematic in the sense of stochastic events that, when compounded over time, make children in the same family different in unpredictable ways" Turkheimer and Waldron (2000) conducted a meta-analysis on the influx of new literature inspired by Plomin and Daniels (1987) regarding the systematic effects of environmental differences within families. The authors conclude in support of the gloomy prospect that their "quantitative review shows that measured nonshared environmental variables do not account for a substantial portion of the nonshared variability posited by biometric studies of behavior". In The Three Laws of Behavioral Genetics, from Turkheimer (2000), it states that "The gloomy prospect is true. Non-shared environmental variability predominates not because of the systematic effects of environmental events that are not shared among siblings, but rather because of the unsystematic effects of all environmental events, compounded by the equally unsystematic processes that expose us to environmental events in the first place." Smith (2011) brings the gloomy prospect with epidemiology as an underlying mechanism of chance
https://en.wikipedia.org/wiki/Low%20Frequency%20Analyzer%20and%20Recorder
Two closely related terms, Low Frequency Analyzer and Recorder and Low Frequency Analysis and Recording bearing the acronym LOFAR, deal with the equipment and process respectively for presenting a visual spectrum representation of low frequency sounds in a time–frequency analysis. The process was originally applied to fixed surveillance passive antisubmarine sonar systems and later to sonobuoy and other systems. Originally the analysis was electromechanical and the display was produced on electrostatic recording paper, a Lofargram, with stronger frequencies presented as lines against background noise. The analysis migrated to digital and both analysis and display were digital after a major system consolidation into centralized processing centers during the 1990s. Both the equipment and process had specific and classified application to fixed surveillance sonar systems and was the basis for the United States Navy's ocean wide Sound Surveillance System (SOSUS) established in the early 1950s. The research and development of systems utilizing LOFAR was given the code name Project Jezebel. The installation and maintenance of SOSUS was under the unclassified code name Project Caesar. The principle was later applied to air, surface and submarine tactical sonar systems with some incorporating the name "Jezebel". Origin In 1949 when the US Navy approached the Committee for Undersea Warfare, an academic advisory group formed in 1946 under the National Academy of Sciences, to research antisubmarine warfare. As a result, the Navy formed a study group designated Project Hartwell under Massachusetts Institute of Technology (MIT) leadership. The Hartwell panel recommended that spending of annually to develop systems to counter the Soviet submarine threat consisting primarily of a large fleet of diesel submarines. One recommendation was a system to monitor low-frequency sound in the SOFAR channel using multiple listening sites equipped with hydrophones and a processing facility
https://en.wikipedia.org/wiki/Flux-corrected%20transport
Flux-corrected transport (FCT) is a conservative shock-capturing scheme for solving Euler equations and other hyperbolic equations which occur in gas dynamics, aerodynamics, and magnetohydrodynamics. It is especially useful for solving problems involving shock or contact discontinuities. An FCT algorithm consists of two stages, a transport stage and a flux-corrected anti-diffusion stage. The numerical errors introduced in the first stage (i.e., the transport stage) are corrected in the anti-diffusion stage.
https://en.wikipedia.org/wiki/Bipolar%20magnetic%20semiconductor
Bipolar magnetic semiconductors (BMSs) are a special class of magnetic semiconductors characterized by a unique electronic structure, where valence band maximum (VBM) and conduction band minimum (CBM) are fully spin polarized in the opposite spin direction. BMSs can be described by three energy gaps, the spin-flip gap Δ2 in valence band (VB), band gap Δ1 and spin-flip gap Δ3 in conduction band (CB). Up to now, bipolar magnetic semiconductors, together with half-metal and spin gapless semiconductor, have been viewed as three important classes of spintronic materials. Properties and potential applications The proposal of bipolar magnetic semiconductor (BMS) is aimed to realize electrical control of carriers' spin orientation, which is a key scientific problem in developing high performance spintronics devices, since electric field can be easily applied locally, in contrast to magnetic field. In BMS, the carriers' spin orientation can be controlled simply by altering the sign of the applied gate voltage. Under zero gate voltage (VG = 0), BMS is semiconducting. Under negative gate voltages (VG < 0) which shift down the material's Fermi level (EF) into spin-flip gap Δ2 in valence band, BMS conducts with carriers fully spin up polarized, while the conducting carriers change to be fully spin down polarized when positive gate voltages (VG > 0) push the Fermi level (EF) up into spin-flip gap Δ3 in conduction band. BMS is expected to be applied as bipolar field effect spin filter and field effect spin valve, or entangled electron detectors and separators. Materials developments A number of BMS materials have been theoretically predicted, such as MnPSe3 nanosheets, Heusler alloys FeVXSi (X = Ti, Zr), double perovskites A2CrOsO6 (A=Ca, Sr, Ba) and DPP-based metal–organic framework. However, the experimental realization of electrical control of spin orientation in these materials still keeps a challenge and needs further experimental efforts.
https://en.wikipedia.org/wiki/Ortho-Vanillin
ortho-Vanillin (2-hydroxy-3-methoxybenzaldehyde) is an organic solid present in the extracts and essential oils of many plants. Its functional groups include aldehyde, ether and phenol. ortho-Vanillin, a compound of the formula C8H8O3, is distinctly different from its more prevalent isomer, meta-vanillin. The "ortho-" prefix refers to the position of the compound’s hydroxyl moiety, which is found in the para-position in vanillin. ortho-Vanillin is a fibrous, light-yellow, crystalline solid. Present in a variety of food products, it is not specifically sought after, and is therefore a less-commonly produced and encountered food additive. History ortho-Vanillin was first isolated, in 1876, by renowned German chemist Ferdinand Tiemann. By 1910, methods for its purification had been developed by Francis Noelting, who similarly demonstrated its versatility as a general synthetic precursor for a diverse array of compounds, such as the coumarins. By 1920, the compound began to show use as a dye for hides. Biological properties ortho-Vanillin is harmful if ingested, irritating to eyes, skin and respiratory system, but has an unmistakable high of 1330 mg/kg in mice. It is a weak inhibitor of tyrosinase, and displays both antimutagenic and comutagenic properties in Escherichia coli. However, its net effect makes it a “potent comutagen”. ortho-Vanillin possesses moderate antifungal and antibacterial properties. Uses Today, most ortho-vanillin is used in the study of mutagenesis and as a synthetic precursor for pharmaceuticals, for example, benafentrine and an antiandrogen compound called Pentomone. See also Vanillin 2-Hydroxy-5-methoxybenzaldehyde Isovanillin 2-Hydroxy-4-methoxybenzaldehyde Notes Hydroxybenzaldehydes Flavors Perfume ingredients
https://en.wikipedia.org/wiki/Universal%20Verification%20Methodology
The Universal Verification Methodology (UVM) is a standardized methodology for verifying integrated circuit designs. UVM is derived mainly from the OVM (Open Verification Methodology) which was, to a large part, based on the eRM (e Reuse Methodology) for the e verification language developed by Verisity Design in 2001. The UVM class library brings much automation to the SystemVerilog language such as sequences and data automation features (packing, copy, compare) etc., and unlike the previous methodologies developed independently by the simulator vendors, is an Accellera standard with support from multiple vendors: Aldec, Cadence, Mentor Graphics, Synopsys, Xilinx Simulator(XSIM). History In December 2009, a technical subcommittee of Accellera — a standards organization in the electronic design automation (EDA) industry — voted to establish the UVM and decided to base this new standard on the Open Verification Methodology (OVM-2.1.1), a verification methodology developed jointly in 2007 by Cadence Design Systems and Mentor Graphics. On February 21, 2011, Accellera approved the 1.0 version of UVM. UVM 1.0 includes a Reference Guide, a Reference Implementation in the form of a SystemVerilog base class library, and a User Guide. Factory A factory is a commonly-used concept in object-oriented programming. It is an object that is used for instantiating other objects. There are two ways to register an object with the UVM factory. In the declaration of class A, one can invoke the `uvm_object_utils(A) or `uvm_component_utils(A) registration macros. Otherwise, the `uvm_object_registry(A,B) or `uvm_component_registry(A,B) macros can be used to map a string B to a class type A. The UVM factory provides a variety of create methods that allow the user to instantiate an object with a particular instance name and of a registered type. Sequencer The sequencer is responsible for three main functions: Put the DUV (Design Under Verification) and the verification environment in
https://en.wikipedia.org/wiki/Borel%20isomorphism
In mathematics, a Borel isomorphism is a measurable bijective function between two standard Borel spaces. By Souslin's theorem in standard Borel spaces (which says that a set that is both analytic and coanalytic is necessarily Borel), the inverse of any such measurable bijective function is also measurable. Borel isomorphisms are closed under composition and under taking of inverses. The set of Borel isomorphisms from a space to itself clearly forms a group under composition. Borel isomorphisms on standard Borel spaces are analogous to homeomorphisms on topological spaces: both are bijective and closed under composition, and a homeomorphism and its inverse are both continuous, instead of both being only Borel measurable. Borel space A measurable space that is Borel isomorphic to a measurable subset of the real numbers is called a Borel space. See also Federer–Morse theorem
https://en.wikipedia.org/wiki/Puck%20puzzle
The Puck puzzle (Hungarian: Kép Korong), also known as Picture Disc, is a combination puzzle invented in 1980 by Hungarian physicist András Végh. The disc shaped puzzle consists of 2 central pieces: the inner disc, and segments forming a ring around the inner disc. The number of segments is an even number, most commonly six, eight, or twelve. The outer ring can be rotated around the central disc, similar to a wheel being rotated around its axis. The two parts of the central disc can also be rotated, together with the attached ring. With these operations, the combination of the parts can be changed. The goal of the game is to obtain a specific combination, such as forming a picture or ordering a sequence of numbers on the segments.
https://en.wikipedia.org/wiki/Intrinsic%20value%20in%20animal%20ethics
The intrinsic value of a human or any other sentient animal comes from within itself. It is the value it places on its own lived experience, valuing it as an end in itself. Intrinsic value exists wherever there are beings that value themselves. Intrinsic value is considered self-ascribed, all animals have it, unlike instrumental or extrinsic values. Instrumental value is the value that others confer on an animal (or on any other entity) because of its value as a resource (e.g. as property, labour, food, fibre, "ecosystem services") or as a source of emotional, recreational, aesthetic or spiritual gratification. Intrinsic values are conferred from within an animal, and are therefore not directly measurable by economists, while extrinsic values are conferred from outside and can, in principle, be measured econometrically. The phrase "intrinsic value" (often used synonymously with inherent value) has been adopted by animal rights advocates. The Dutch Animal Health and Welfare Act referred to it in 1981: "Acknowledgment of the intrinsic value of animals means that animals have value in their own right and as a consequence, their interests are no longer automatically subordinate to man's interests." This acknowledgement has stirred a debate on what it entails in the context of animal husbandry, animal breeding, vivisection, animal testing, and biotechnology. It is also used by environmental advocates and in law to holistically encompass the totality of intrinsic values in an ecosystem. Article 7(d) of New Zealand's Resource Management Act (RMA), for example, requires particular regard to being given to the "intrinsic values of ecosystems". History of the moral status of animals (1880–1980) Moral attitudes towards animals in the West (as expressed in public debate and legislation) have changed considerably over time. Britain's first anti-cruelty laws were introduced in the Cruelty to Animals Act 1835. This was followed by similar laws in many other countries, especia
https://en.wikipedia.org/wiki/Robert%20C.%20Gunning
Robert Clifford Gunning (born 1931) is a professor of mathematics at Princeton University specializing in complex analysis, who introduced indigenous bundles. Gunning was born in Longmont, Colorado, and attended to high school in his hometown. In 1947 he was admitted into the University of Colorado, graduating with a bachelor's degree in 1952. For his graduate studies he went to Princeton University, where he earned his Ph.D in 1955 under Salomon Bochner with thesis A classification of factors of automorphy. He then taught at the University of Chicago and in 1956 as Higgins-Lecturer at Princeton University. At Princeton, Gunning became in 1957 assistant professor, in 1962 associate professor, and in 1966 professor. He was a visiting professor in São Paulo in 1958, Cambridge in 1959/60, Munich in 1967, Oxford in 1968, Boulder in 1970, and Los Angeles in 1972. Gunning is known as the author of important books on functions of several complex variables. From 1958 to 1961 he was a Sloan Research Fellow. He served as Princeton University's dean of the faculty from 1989 to 1995. In 2003 he received Princeton University's prize for outstanding teaching. For a number of years he was an editor for Princeton University Press and for the Annals of Mathematical Studies. He was also the editor of the collected works of Salomon Bochner. In 1970 he was an invited speaker at the International Mathematical Congress in Nice (Some multivariable problems arising from Riemann surfaces). Among his doctoral students are Sheldon Katz, Henry Laufer, Richard S. Hamilton, Yum-Tong Siu, and Michael Eastwood. In 2012 he became a fellow of the American Mathematical Society. Selected works Analytic functions of several complex variables. Prentice-Hall 1965. Lectures on Riemann Surfaces. Princeton University Press 1966. Lectures on Vector Bundles over Riemann Surfaces. Princeton University Press 1967. Riemann Surfaces and generalized Theta Functions. Springer, Ergebnisse der Mathematik un
https://en.wikipedia.org/wiki/Whonamedit%3F
Whonamedit? is an online English-language dictionary of medical eponyms and the people associated with their identification. Though it is a dictionary, many eponyms and persons are presented in extensive articles with comprehensive bibliographies. The dictionary is hosted in Norway and maintained by medical historian Ole Daniel Enersen.
https://en.wikipedia.org/wiki/Viehland%E2%80%93Mason%20theory
The Viehland–Mason theory is a two-temperature theory for charged and neutral atoms, which explains how trace ions can have a substantially different temperature than dilute gas atoms. It is one of any of a number of kinetic theories of the transport of trace amounts of molecular ions through neutral gases under the influence of a uniform electrostatic field. Drs Larry Viehland and Ed Mason developed it in the late 1970s. They later extended this theory into a three-temperature theory that allowed for different ion temperatures parallel and perpendicular to the electric field. Current work for atomic ion-neutral systems uses a Gram–Charlier probability function as a zero-order approximation to the ion velocity distribution function. The Gram–Charlier theory has been remarkably successful in producing calculated mobilities and diffusion coefficients that are in excellent agreement with experimental results if the microscopic force between the ion and atom is accurately known over a wide range of separation. The Viehland–Mason theories for molecular ions in molecular gases are more elaborate than those for atoms, since the forces are angle-dependent and since internal degrees of freedom must be included. Theories have been developed using quantum-mechanical and semi-classical approaches, but there have been no numerical applications because it is extremely difficult to calculate the necessary cross-sections. To circumvent this difficulty, completely classical kinetic theories for atomic ions in non-vibrating (rigid rotor) diatomic gases and for non-vibrating diatomic ions in atomic or non-vibrating diatomic gases have been developed.
https://en.wikipedia.org/wiki/Ralph%20A.%20Bradley
Ralph Allan Bradley (November 28, 1923 – October 30, 2001) was a Canadian-American statistician and statistics educator, whose research lie in the fields of design of experiments, nonparametric statistics, sequential analysis, and multivariate analysis. He is known for the Bradley–Terry model in pairwise comparison and foundation of the Department of Statistics at Florida State University. Education and career Bradley was born in Smith Falls, Ontario, Canada and grew up in Wellington, Ontario. He studied mathematics and physics at Queen's University, receiving an honors degree in 1944. After a stint at the Canadian Army, he returned to Queen's University and obtained an MA in 1946. He received his PhD in theoretical statistics in 1949 at the University of North Carolina at Chapel Hill under the supervision of Harold Hotelling. Bradley was employed at McGill University from 1949 to 1950 and as a faculty member at Virginia Tech from 1950 to 1958. In 1959, he moved to Florida State University to found the Department of Statistics there and stayed on until 1978 as the head of the department. In 1982, he moved to the University of Georgia as Research Professor of Statistics. Bradley retired from the University of Georgia in 1992. He continued to participate in research activities afterwards and was named professor emeriti at both Florida State University and University of Georgia. Honors and awards Bradley was editor of the journal Biometrics from 1957 to 1962. He was Vice-President from 1975 to 1978 and President in 1981 of the American Statistical Association. Bradley was a Fellow of the American Statistical Association and a Fellow of the Institute of Mathematical Statistics. The statistics department of the University of Georgia organizes an annual lecture in his name.
https://en.wikipedia.org/wiki/End%20%28category%20theory%29
In category theory, an end of a functor is a universal dinatural transformation from an object e of X to S. More explicitly, this is a pair , where e is an object of X and is an extranatural transformation such that for every extranatural transformation there exists a unique morphism of X with for every object a of C. By abuse of language the object e is often called the end of the functor S (forgetting ) and is written Characterization as limit: If X is complete and C is small, the end can be described as the equalizer in the diagram where the first morphism being equalized is induced by and the second is induced by . Coend The definition of the coend of a functor is the dual of the definition of an end. Thus, a coend of S consists of a pair , where d is an object of X and is an extranatural transformation, such that for every extranatural transformation there exists a unique morphism of X with for every object a of C. The coend d of the functor S is written Characterization as colimit: Dually, if X is cocomplete and C is small, then the coend can be described as the coequalizer in the diagram Examples Natural transformations: Suppose we have functors then . In this case, the category of sets is complete, so we need only form the equalizer and in this case the natural transformations from to . Intuitively, a natural transformation from to is a morphism from to for every in the category with compatibility conditions. Looking at the equalizer diagram defining the end makes the equivalence clear. Geometric realizations: Let be a simplicial set. That is, is a functor . The discrete topology gives a functor , where is the category of topological spaces. Moreover, there is a map sending the object of to the standard -simplex inside . Finally there is a functor that takes the product of two topological spaces. Define to be the composition of this product functor with . The coend of is the geometric realization of . Not