source
stringlengths 31
227
| text
stringlengths 9
2k
|
|---|---|
https://en.wikipedia.org/wiki/Electronic%20speed%20control
|
An electronic speed control (ESC) is an electronic circuit that controls and regulates the speed of an electric motor. It may also provide reversing of the motor and dynamic braking.
Miniature electronic speed controls are used in electrically powered radio controlled models. Full-size electric vehicles also have systems to control the speed of their drive motors.
Function
An electronic speed control follows a speed reference signal (derived from a throttle lever, joystick, or other manual input) and varies the switching rate of a network of field effect transistors (FETs). By adjusting the duty cycle or switching frequency of the transistors, the speed of the motor is changed. The rapid switching of the current flowing through the motor is what causes the motor itself to emit its characteristic high-pitched whine, especially noticeable at lower speeds.
Different types of speed controls are required for brushed DC motors and brushless DC motors. A brushed motor can have its speed controlled by varying the voltage on its armature. (Industrially, motors with electromagnet field windings instead of permanent magnets can also have their speed controlled by adjusting the strength of the motor field current.) A brushless motor requires a different operating principle. The speed of the motor is varied by adjusting the timing of pulses of current delivered to the several windings of the motor.
Brushless ESC systems basically create three-phase AC power, like a variable frequency drive, to run brushless motors. Brushless motors are popular with radio controlled airplane hobbyists because of their efficiency, power, longevity and light weight in comparison to traditional brushed motors. Brushless DC motor controllers are much more complicated than brushed motor controllers.
The correct phase of the current fed to the motor varies with the motor rotation, which is to be taken into account by the ESC: Usually, back EMF from the motor windings is used to detect this
|
https://en.wikipedia.org/wiki/Minimal%20residual%20method
|
The Minimal Residual Method or MINRES is a Krylov subspace method for the iterative solution of symmetric linear equation systems. It was proposed by mathematicians Christopher Conway Paige and Michael Alan Saunders in 1975.
In contrast to the popular CG method, the MINRES method does not assume that the matrix is positive definite, only the symmetry of the matrix is mandatory. The popular GMRES method is an improved generalization of MINRES but requires much more memory.
GMRES vs. MINRES
The GMRES method is essentially a generalization of MINRES for arbitrary matrices. Both minimize the 2-norm of the residual and do the same calculations in exact arithmetic when the matrix is symmetric. MINRES is a short-recurrence method with a constant memory requirement, whereas GMRES requires storing the whole Krylov space, so its memory requirement is roughly proportional to the number of iterations. On the other hand, GMRES tends to suffer less from loss of orthogonality.
Therefore, MINRES tends to be used only when there is not enough memory for GMRES and the matrix is symmetric. Even then, sometimes other methods are preferred, particularly CG for positive-definite matrices.
Properties of the MINRES method
The MINRES method iteratively calculates an approximate solution of a linear system of equations of the form
where is a symmetric matrix and a vector.
For this, the norm of the residual in a -dimensional Krylov subspace
is minimized. Here is an initial value (often zero) and .
More precisely, we define the approximate solutions through
where is the standard Euclidean norm on .
Because of the symmetry of , unlike in the GMRES method, it is possible to carry out this minimization process recursively, storing only two previous steps (short recurrence). This saves memory.
MINRES algorithm
Note: The MINRES method is more complicated than the algebraically equivalent Conjugate Residual method. The Conjugate Residual (CR) method was therefore produced below
|
https://en.wikipedia.org/wiki/N-terminal%20prohormone%20of%20brain%20natriuretic%20peptide
|
The N-terminal prohormone of brain natriuretic peptide (NT-proBNP or BNPT) is a prohormone with a 76 amino acid N-terminal inactive protein that is cleaved from the molecule to release brain natriuretic peptide 32 (BNP, also known as B-type natriuretic peptide).
Both BNP and NT-proBNP levels in the blood are used for screening, diagnosis of acute congestive heart failure (CHF) and may be useful to establish prognosis in heart failure, as both markers are typically higher in patients with worse outcome. The plasma concentrations of both BNP and NT-proBNP are also typically increased in patients with asymptomatic or symptomatic left ventricular dysfunction and is associated with coronary artery disease and myocardial ischemia.
Blood levels
There is no level of BNP that perfectly separates patients with and without heart failure.
In screening for congenital heart disease in pediatric patients, an NT-proBNP cut-off value of 91 pg/mL could differentiate an acyanotic heart disease (ACNHD) patient from a healthy patient with a sensitivity of 84% and specificity of 42%. On the other hand, an NT-proBNP cut-off value of 318 pg/mL is more appropriate in differing patients with congenital nonspherocytic hemolytic disease (CNHD) from healthy patients, with 94% sensitivity and 97% specificity. An NT-proBNP value of 408 pg/mL has been estimated to be 83% sensitive and 57% specific in differentiating patients with ACNHD from patients with CNHD.
Test usage in a clinical setting
Canada
While discussed in Canadian medical journals in the mid to late 2000s, the test is not widely used. It was only approved for use in Alberta in February 2012.
Test usage in the life insurance industry
The test has been widely used in the life insurance industry to screen applicants as part of the routine requirements when applying for a life insurance policy. It is also inexpensive and can be measured from blood samples routinely drawn as part of the application process. The test can be used to e
|
https://en.wikipedia.org/wiki/ENO%20methods
|
ENO (essentially non-oscillatory) methods are classes of high-resolution schemes in numerical solution of differential equations.
History
The first ENO scheme was developed by Harten, Engquist, Osher and Chakravarthy in 1987. In 1994, the first weighted version of ENO was developed.
See also
High-resolution scheme
WENO methods
Shock-capturing method
|
https://en.wikipedia.org/wiki/Anti-lock%20braking%20system
|
An anti-lock braking system (ABS) is a safety anti-skid braking system used on aircraft and on land vehicles, such as cars, motorcycles, trucks, and buses. ABS operates by preventing the wheels from locking up during braking, thereby maintaining tractive contact with the road surface and allowing the driver to maintain more control over the vehicle.
ABS is an automated system that uses the principles of threshold braking and cadence braking, techniques which were once practiced by skillful drivers before ABS was widespread. ABS operates at a much faster rate and more effectively than most drivers could manage. Although ABS generally offers improved vehicle control and decreases stopping distances on dry and some slippery surfaces, on loose gravel or snow-covered surfaces ABS may significantly increase braking distance, while still improving steering control. Since ABS was introduced in production vehicles, such systems have become increasingly sophisticated and effective. Modern versions may not only prevent wheel lock under braking, but may also alter the front-to-rear brake bias. This latter function, depending on its specific capabilities and implementation, is known variously as electronic brakeforce distribution, traction control system, emergency brake assist, or electronic stability control (ESC).
History
Early systems
The concept for ABS predates the modern systems that were introduced in the 1950s. In 1908, for example, J.E. Francis introduced his 'Slip Prevention Regulator for Rail Vehicles'.
In 1920 the French automobile and aircraft pioneer Gabriel Voisin experimented with systems that modulated the hydraulic braking pressure on his aircraft brakes to reduce the risk of tire slippage, as threshold braking on aircraft is nearly impossible. These systems used a flywheel and valve attached to a hydraulic line that feeds the brake cylinders. The flywheel is attached to a drum that runs at the same speed as the wheel. In normal braking, the drum and flyw
|
https://en.wikipedia.org/wiki/List%20of%20chess%20software
|
Chess software comes in different forms. A chess playing program provides a graphical chessboard on which one can play a chess game against a computer. Such programs are available for personal computers, video game consoles, smartphones/tablet computers or mainframes/supercomputers. A chess engine generates moves, but is accessed via a command-line interface with no graphics. A dedicated chess computer has been purpose built solely to play chess. A graphical user interface (GUI) allows one to import and load an engine, and play against it. A chess database allows one to import, edit, and analyze a large archive of past games.
Chess-playing programs for personal computers
Chess engines
This list contains only chess engines for which Wikipedia articles exist yet and therefore is very incomplete. It does not reflect or imply current or historic play strength as this characteristic in itself usually does not warrant an entry on Wikipedia.
Chess graphical user interfaces
Chess-playing programs for video game consoles
Chess apps for phones/tablets
Chess computers
The following are special-purpose hardware/software combinations that are inextricably connected:
Programs for reading and editing chess databases
Chess Assistant
Chess Informant Expert
ChessBase
Shane's Chess Information Database
Chess-playing programs for mainframes/supercomputers
AlphaZero
Chess (Northwestern University)
CilkChess
Kaissa
Kotok-McCarthy
Mac Hack
Tiniest chess programs
1K ZX Chess
Microchess
Toledo Nanochess
See also
|
https://en.wikipedia.org/wiki/Square%20gyrobicupola
|
In geometry, the square gyrobicupola is one of the Johnson solids (). Like the square orthobicupola (), it can be obtained by joining two square cupolae () along their bases. The difference is that in this solid, the two halves are rotated 45 degrees with respect to one another.
The square gyrobicupola is the second in an infinite set of gyrobicupolae.
Related to the square gyrobicupola is the elongated square gyrobicupola. This polyhedron is created when an octagonal prism is inserted between the two halves of the square gyrobicupola. It is argued whether or not the elongated square gyrobicupola is an Archimedean solid because, although it meets every other standard necessary to be an Archimedean solid, it is not highly symmetric.
Formulae
The following formulae for volume and surface area can be used if all faces are regular, with edge length a:
Related polyhedra and honeycombs
The square gyrobicupola forms space-filling honeycombs with tetrahedra, cubes and cuboctahedra; and with tetrahedra, square pyramids, and elongated square bipyramids. (The latter unit can be decomposed into elongated square pyramids, cubes, and/or square pyramids).
|
https://en.wikipedia.org/wiki/Gravity%20Probe%20A
|
Gravity Probe A (GP-A) was a space-based experiment to test the equivalence principle, a feature of Einstein's theory of relativity. It was performed jointly by the Smithsonian Astrophysical Observatory and the National Aeronautics and Space Administration. The experiment sent a hydrogen masera highly accurate frequency standardinto space to measure with high precision the rate at which time passes in a weaker gravitational field. Masses cause distortions in spacetime, which leads to the effects of length contraction and time dilation, both predicted results of Albert Einstein's theory of general relativity. Because of the bending of spacetime, an observer on Earth (in a lower gravitational potential) should measure a slower rate at which time passes than an observer that is higher in altitude (at higher gravitational potential). This effect is known as gravitational time dilation.
The experiment was a test of a major consequence of Einstein's general relativity, the equivalence principle. The equivalence principle states that a reference frame in a uniform gravitational field is indistinguishable from a reference frame that is under uniform acceleration. Further, the equivalence principle predicts that phenomenon of different time flow rates, present in a uniformly accelerating reference frame, will also be present in a stationary reference frame that is in a uniform gravitational field.
The probe was launched on June 18, 1976 from the NASA-Wallops Flight Center in Wallops Island, Virginia. The probe was carried via a Scout rocket, and attained a height of , while remaining in space for 1 hour and 55 minutes, as intended. It returned to Earth by splashing down into the Atlantic Ocean.
Background
The objective of the Gravity Probe A experiment was to test the validity of the equivalence principle. The equivalence principle is a key component of Albert Einstein's theory of general relativity, and states that the laws of physics are the same in an accelerating refe
|
https://en.wikipedia.org/wiki/Homotopy%20type%20theory
|
In mathematical logic and computer science, homotopy type theory (HoTT ) refers to various lines of development of intuitionistic type theory, based on the interpretation of types as objects to which the intuition of (abstract) homotopy theory applies.
This includes, among other lines of work, the construction of homotopical and higher-categorical models for such type theories; the use of type theory as a logic (or internal language) for abstract homotopy theory and higher category theory; the development of mathematics within a type-theoretic foundation (including both previously existing mathematics and new mathematics that homotopical types make possible); and the formalization of each of these in computer proof assistants.
There is a large overlap between the work referred to as homotopy type theory, and as the univalent foundations project. Although neither is precisely delineated, and the terms are sometimes used interchangeably, the choice of usage also sometimes corresponds to differences in viewpoint and emphasis. As such, this article may not represent the views of all researchers in the fields equally. This kind of variability is unavoidable when a field is in rapid flux.
History
The groupoid model
At one time the idea that types in intensional type theory with their identity types could be regarded as groupoids was mathematical folklore. It was first made precise semantically in the 1994 paper of Martin Hofmann and Thomas Streicher called "The groupoid model refutes uniqueness of identity proofs", in which they showed that intensional type theory had a model in the category of groupoids. This was the first truly "homotopical" model of type theory, albeit only "1-dimensional" (the traditional models in the category of sets being homotopically 0-dimensional).
Their follow-up paper foreshadowed several later developments in homotopy type theory. For instance, they noted that the groupoid model satisfies a rule they called "universe extensionalit
|
https://en.wikipedia.org/wiki/Security%20Assertion%20Markup%20Language
|
Security Assertion Markup Language (SAML, pronounced SAM-el, ) is an open standard for exchanging authentication and authorization data between parties, in particular, between an identity provider and a service provider. SAML is an XML-based markup language for security assertions (statements that service providers use to make access-control decisions). SAML is also:
A set of XML-based protocol messages
A set of protocol message bindings
A set of profiles (utilizing all of the above)
An important use case that SAML addresses is web-browser single sign-on (SSO). Single sign-on is relatively easy to accomplish within a security domain (using cookies, for example) but extending SSO across security domains is more difficult and resulted in the proliferation of non-interoperable proprietary technologies. The SAML Web Browser SSO profile was specified and standardized to promote interoperability.
Overview
The SAML specification defines three roles: the principal (typically a human user), the identity provider (IdP) and the service provider (SP). In the primary use case addressed by SAML, the principal requests a service from the service provider. The service provider requests and obtains an authentication assertion from the identity provider. On the basis of this assertion, the service provider can make an access control decision, that is, it can decide whether to perform the service for the connected principal.
At the heart of the SAML assertion is a subject (a principal within the context of a particular security domain) about which something is being asserted. The subject is usually (but not necessarily) a human. As in the SAML 2.0 Technical Overview, the terms subject and principal are used interchangeably in this document.
Before delivering the subject-based assertion from IdP to the SP, the IdP may request some information from the principal—such as a user name and password—in order to authenticate the principal. SAML specifies the content of the assertion
|
https://en.wikipedia.org/wiki/Density%20dependence
|
In population ecology, density-dependent processes occur when population growth rates are regulated by the density of a population. This article will focus on density dependence in the context of macroparasite life cycles.
Positive density-dependence
Positive density-dependence, density-dependent facilitation, or the Allee effect describes a situation in which population growth is facilitated by increased population density.
Examples
For dioecious (separate sex) obligatory parasites, mated female worms are required to complete a transmission cycle. At low parasite densities, the probability of a female worm encountering a male worm and forming a mating pair can become so low that reproduction is restricted due to single sex infections. At higher parasite densities, the probability of mating pairs forming and successful reproduction increases. This has been observed in the population dynamics of Schistosomes.
Positive density-dependence processes occur in macroparasite life cycles that rely on vectors with a cibarial armature, such as Anopheles or Culex mosquitoes. For Wuchereria bancrofti, a filarial nematode, well-developed cibarial armatures in vectors can damage ingested microfilariae and impede the development of infective L3 larvae. At low microfilariae densities, most microfilariae can be ruptured by teeth, preventing successful development of infective L3 larvae. As more larvae are ingested, the ones that become entangled in the teeth may protect the remaining larvae, which are then left undamaged during ingestion.
Positive density-dependence processes may also occur in macroparasite infections that lead to immunosuppression. Onchocerca volvulus infection promotes immunosuppressive processes within the human host that suppress immunity against incoming infective L3 larvae. This suppression of anti-parasite immunity causes parasite establishment rates to increase with higher parasite burden.
Negative density-dependence
Negative density-dependence, or dens
|
https://en.wikipedia.org/wiki/Orbital%20gyri
|
The inferior or orbital surface of the frontal lobe is concave, and rests on the orbital plate of the frontal bone. It is divided into four orbital gyri by a well-marked H-shaped orbital sulcus. These are named, from their position, the medial, anterior, lateral, and posterior, orbital gyri. The medial orbital gyrus presents a well-marked antero-posterior sulcus, the olfactory sulcus, for the olfactory tract; the portion medial to this is named the straight gyrus, and is continuous with the superior frontal gyrus on the medial surface.
Function
Bailey and Bremer reported that stimulation to the central end of the vagus nerve caused electrical activity in the inferior orbital surface (http://brain.oxfordjournals.org/cgi/pdf_extract/75/2/244)
Additional images
|
https://en.wikipedia.org/wiki/Flood%20pulse%20concept
|
The flood pulse concept explains how the periodic inundation and drought (flood pulse) control the lateral exchange of water, nutrients and organisms between the main river channel (or lake) and the connected floodplain. The annual flood pulse is the most important aspect and the most biologically productive feature of a river's ecosystem. describing the movement, distribution and quality of water in river ecosystems and the dynamic interaction in the transition zone between water and land (aquatic/terrestrial transition zones - ATTZ). It contrasts with previous ecological theories which considered floods to be catastrophic events.
Background
River floodplain systems consist of an area surrounding a river that is periodically flooded by the overflow of the river as well as by precipitation, called the aquatic/terrestrial transition zone (ATTZ). The ATTZ is the area covered by water only during the flooding. This flooding in turn creates unique habitat that is essential to the survival of many different species. The flood pulse concept is unique because it incorporates the outlying rivers and streams which add a lateral aspect to previous concepts, e.g. the River Continuum Concept (RCC) that failed in explain processes that happen in big rivers and their floodplains. From this lateral perspective, rivers can be seen as a collection of width-based water systems.
Flooding consists of multiple stages. First, at the start of the flooding, nutrients rush in from the area where the flood begins. During flood periods, the most important element is called the moving littoral. As flooding begins and water levels increase nutrients that have been mineralized in the dry phase are suspended with sediments in the flood waters and main river. The moving littoral consists of the water from the shoreline to a few meters deep in the river. This pulse of water is the primary driver of high productivity and decomposition rates as it moves nutrients in and out of the system and
|
https://en.wikipedia.org/wiki/Siliceous%20sponge
|
The siliceous sponges form a major group of the phylum Porifera, consisting of classes Demospongiae and Hexactinellida. They are characterized by spicules made out of silicon dioxide, unlike calcareous sponges.
Individual siliachoates (silica skeleton scaffolding) can be arranged tightly within the sponginocyte or crosshatched and fused together. Siliceous spicules come in two sizes called megascleres and microscleres.
Systematics
Most studies support the monophyly of siliceous sponges.
The group, as a part of the phylum Porifera, has been named Silicispongia Schmidt, 1862 and Silicea Bowerbank, 1864. Silicarea is a proposed new phylum based on molecular studies of the phylum Porifera. It consists of the Poriferan classes Demospongiae and Hexactinellida. Some scientists believe that Porifera is polyphyletic/paraphyletic, and that some sponges, the Calcarea, are a separate phylum which was the first to diverge from the main line of kingdom Animalia. Silicarea is considered the next phylum to diverge from the primary animal lineage.
Ecology
Siliceous sponges are usually found in the marine ecosystem but they are occasionally found in freshwater.
During the Triassic, siliceous sponges grew reefs similar to calcarea of the modern era. During the Cretaceous period, diatoms became so successful that they significantly decreased the amount of silica present in sea water, after which "siliceous sponges could never again form reefs."
|
https://en.wikipedia.org/wiki/Snub%20disphenoid
|
In geometry, the snub disphenoid, Siamese dodecahedron, triangular dodecahedron, trigonal dodecahedron, or dodecadeltahedron is a convex polyhedron with twelve equilateral triangles as its faces. It is not a regular polyhedron because some vertices have four faces and others have five. It is a dodecahedron, one of the eight deltahedra (convex polyhedra with equilateral triangle faces), and is the 84th Johnson solid (non-uniform convex polyhedra with regular faces). It can be thought of as a square antiprism where both squares are replaced with two equilateral triangles.
The snub disphenoid is also the vertex figure of the isogonal 13-5 step prism, a polychoron constructed from a 13-13 duoprism by selecting a vertex on a tridecagon, then selecting the 5th vertex on the next tridecagon, doing so until reaching the original tridecagon. It cannot be made uniform, however, because the snub disphenoid has no circumscribed sphere.
History and naming
This shape was called a Siamese dodecahedron in the paper by Hans Freudenthal and B. L. van der Waerden (1947) which first described the set of eight convex deltahedra. The dodecadeltahedron name was given to the same shape by , referring to the fact that it is a 12-sided deltahedron. There are other simplicial dodecahedra, such as the hexagonal bipyramid, but this is the only one that can be realized with equilateral faces. Bernal was interested in the shapes of holes left in irregular close-packed arrangements of spheres, so he used a restrictive definition of deltahedra, in which a deltahedron is a convex polyhedron with triangular faces that can be formed by the centers of a collection of congruent spheres, whose tangencies represent polyhedron edges, and such that there is no room to pack another sphere inside the cage created by this system of spheres. This restrictive definition disallows the triangular bipyramid (as forming two tetrahedral holes rather than a single hole), pentagonal bipyramid (because the spheres for
|
https://en.wikipedia.org/wiki/Oracle%20Fusion%20Applications
|
Oracle Fusion Applications (OFA) are a suite of applications built on Oracle Cloud that include cloud-based applications for enterprise resource planning (ERP), enterprise performance management (EPM), supply chain management and manufacturing (SCM), human capital management (HCM), and customer experience (CX).
Oracle Fusion Applications were launched in September 2010 and released one year later at OpenWorld 2011. The name refers to the fact that each application runs in the cloud and shares one underlying platform, which means that data can be shared between them. This is different from many cloud applications that exist in isolation and are not interoperable. Oracle also developed "coexistence" processes that extend across all Oracle Applications so that users could adopt what they needed from Fusion while still using existing on-premise applications.
Oracle Fusion Applications were originally envisioned and pitched as an enterprise resource planning suite—a combination of features and functionalities taken from the Oracle E-Business Suite, JD Edwards, PeopleSoft, and Siebel product lines. The suite was built on top of the Oracle Fusion Middleware technology stack; both layers implement the Oracle Fusion Architecture, which leverages the capabilities of service-oriented architecture. Companies can deploy Fusion cloud applications in a private cloud, public cloud, hybrid cloud, or a private cloud built and managed by Oracle Cloud Services.
In September 2021, Oracle launched Fusion Marketing as part of Oracle Advertising and CX. The system uses artificial intelligence to automate digital marketing campaigns and identify qualified leads.
See also
Oracle Applications
Oracle Advertising and Customer Experience (CX)
Oracle Cloud ERP
Oracle Fusion Architecture
Oracle Fusion Middleware
Oracle Cloud HCM
Oracle SOA Suite
|
https://en.wikipedia.org/wiki/Initial
|
In a written or published work, an initial capital (also referred to as a drop capital or simply an initial cap, initial, initcapital, initcap or init or a drop cap or drop) is a letter at the beginning of a word, a chapter, or a paragraph that is larger than the rest of the text. The word is derived from the Latin initialis, which means standing at the beginning. An initial is often several lines in height and in older books or manuscripts are known as "inhabited" initials. Certain important initials, such as the Beatus initial or "B" of Beatus vir... at the opening of Psalm 1 at the start of a vulgate Latin. These specific initials in an illuminated manuscript were also called initiums.
In the present, the word "initial" commonly refers to the first letter of any word or name, the latter normally capitalized in English usage and is generally that of a first given name or a middle one or ones.
History
The classical tradition was slow to use capital letters for initials at all; in surviving Roman texts it often is difficult even to separate the words as spacing was not used either. In late antiquity (–6th century) both came into common use in Italy, the initials usually were set in the left margin (as in the third example below), as though to cut them off from the rest of the text, and about twice as tall as the other letters. The radical innovation of insular manuscripts was to make initials much larger, not indented, and for the letters immediately following the initial also to be larger, but diminishing in size (called the "diminuendo" effect, after the musical term). Subsequently, they became larger still, coloured, and penetrated farther and farther into the rest of the text, until the whole page might be taken over. The decoration of insular initials, especially large ones, was generally abstract and geometrical, or featured animals in patterns. Historiated initials were an Insular invention, but did not come into their own until the later developments
|
https://en.wikipedia.org/wiki/Chris%20Wallace%20%28computer%20scientist%29
|
Christopher Stewart Wallace (26 October 1933 – 7 August 2004) was an Australian computer scientist and physicist.
Wallace is notable for having devised:
The minimum message length principle — an information-theoretic principle in statistics, econometrics, machine learning, inductive inference and knowledge discovery which can be seen both as a mathematical formalisation of Occam's Razor and as an invariant Bayesian method of model selection and point estimation,
The Wallace tree form of binary multiplier (1964),
a variety of random number generators,
a theory in physics and philosophy that entropy is not the arrow of time,
a refrigeration system (from the 1950s, whose design is still in use in 2010),
hardware for detecting and counting cosmic rays,
design of computer operating systems,
the notion of universality probability in mathematical logic,
and a vast range of other works - see, e.g., and its Foreword re C. S. Wallace , pp 523-560.
He was appointed Foundation Chair of Information Science at Monash University in 1968 at the age of 34 (before the Department was re-named Computer Science), and Professor Emeritus in 1996. Wallace was a fellow of the Australian Computer Society and in 1995 he was appointed a fellow of the ACM "For research in a number of areas in Computer Science including fast multiplication algorithm, minimum message length principle and its applications, random number generation, computer architecture, numerical solution of ODE's, and contribution to Australian Computer Science."
Wallace received his PhD (in Physics) from the University of Sydney in 1959. He was married to Judy Ogilvie, the first secretary and programme librarian of SILLIAC, which was launched on the 12 of September 1956 at the University of Sydney and which was one of Australia's first computers. He also engineered one of the world's first Local Area Networks in the mid-1960s.
|
https://en.wikipedia.org/wiki/Solasodine
|
Solasodine is a poisonous alkaloid chemical compound that occurs in plants of the family Solanaceae such as potatoes and tomatoes. Solasonine and solamargine are glycoalkaloid derivatives of solasodine. Solasodine is teratogenic to hamster fetuses in a dose of 1200 to 1600 mg/kg.
Literature survey reveals that solasodine has diuretic, anticancer, antifungal, cardiotonic, antispermatogenetic, antiandrogenic, immunomodulatory, antipyretic and various effects on central nervous system.
Uses
It is commercially used as a precursor for the production of complex steroidal compounds such as contraceptive pills, via a 16-DPA intermediate.
See also
Solanum mauritianum
|
https://en.wikipedia.org/wiki/Yanoconodon
|
Yanoconodon is a monotypic genus of extinct early mammal whose representative species Yanoconodon allini lived during the Mesozoic in what is now China. The holotype fossil of Yanoconodon was excavated in the Yan Mountains about 300 kilometres from Beijing in the Qiaotou member of the Huajiying Formation (which the original authors considered part of the Yixian Formation) of Hebei Province, China, and is therefore of uncertain age. The Qiaotou Member may correlate with the more well-known Early Cretaceous Yixian Formation, and so probably dates to around 122 Ma ago.
Yanoconodon was a eutriconodont, a group composing most taxa once classified as "triconodonts" which lived during the time of the dinosaurs. These were a highly ecologically diverse group, including large sized taxa such as Repenomamus that were able to eat small dinosaurs, the arboreal Jeholodens, the aerial volaticotherines and the spined Spinolestes. Yanoconodon is inferred to be a generalized terrestrial mammal, capable of multiple forms of locomotion.
Yanoconodon'''s name is composed of two elements: 'Yan' is taken from the Yan Mountains in the north of the Hebei Province near where the holotype of Yanoconodon was found; 'Conodon' is an often used as a mammalian taxonomic suffix meaning 'cuspate tooth'. Its species name, "allini," is derived from mammalian researcher Edgar Allin, who was notable for his research on the mammalian middle ear.Paleontologists Discover New Mammal from Mesozoic Era at www.physorg.com - Retrieved 25/6/2007.
Description Yanoconodon was a small mammal, barely 5 inches (13 centimetres) long. It had a sprawling posture, and although previously inferred to be semi-aquatic, direct study of its postcrania indicates that Yanoconodon was likely a terrestrial mammal, and that it has features in common with digging, arboreal, and semiaquatic mammals. Yanocodon had lumbar ribs, a feature not seen in modern mammals. The closely related eutriconodont Jeholodens lacks these lumbar ri
|
https://en.wikipedia.org/wiki/Candi%20sugar
|
Candi sugar is a Belgian sugar product commonly used in brewing beer. It is particularly associated with stronger Belgian style ales such as dubbel and tripel. Chemically, it is an unrefined sugar beet derived sugar which has been subjected to Maillard reaction and caramelization. A common misconception is to consider this is the same as invert sugar, while actual candi sugar is subjected to multiple complex chemical reactions during the Maillard process.
Also used as a priming sugar, to aid in bottle-conditioning, and carbonation, with the same benefits as listed above.
See also
Adjuncts
Inverted sugar syrup
Rock candy
|
https://en.wikipedia.org/wiki/Protein%20precipitation
|
Protein precipitation is widely used in downstream processing of biological products in order to concentrate proteins and purify them from various contaminants. For example, in the biotechnology industry protein precipitation is used to eliminate contaminants commonly contained in blood. The underlying mechanism of precipitation is to alter the solvation potential of the solvent, more specifically, by lowering the solubility of the solute by addition of a reagent.
General principles
The solubility of proteins in aqueous buffers depends on the distribution of hydrophilic and hydrophobic amino acid residues on the protein's surface. Hydrophobic residues predominantly occur in the globular protein core, but some exist in patches on the surface. Proteins that have high hydrophobic amino acid content on the surface have low solubility in an aqueous solvent. Charged and polar surface residues interact with ionic groups in the solvent and increase the solubility of a protein. Knowledge of a protein's amino acid composition will aid in determining an ideal precipitation solvent and methods.
Repulsive electrostatic force
Repulsive electrostatic forces form when proteins are dissolved in an electrolyte solution. These repulsive forces between proteins prevent aggregation and facilitate dissolution. Upon dissolution in an electrolyte solution, solvent counterions migrate towards charged surface residues on the protein, forming a rigid matrix of counterions on the protein's surface. Next to this layer is another solvation layer that is less rigid and, as one moves away from the protein surface, contains a decreasing concentration of counterions and an increasing concentration of co-ions. The presence of these solvation layers cause the protein to have fewer ionic interactions with other proteins and decreases the likelihood of aggregation.
Repulsive electrostatic forces also form when proteins are dissolved in water. Water forms a solvation layer around the hydrophilic sur
|
https://en.wikipedia.org/wiki/Transcriptome%20instability
|
Transcriptome instability is a genome-wide, pre-mRNA splicing-related characteristic of certain cancers. In general, pre-mRNA splicing is dysregulated in a high proportion of cancerous cells. For certain types of cancer, like in colorectal and prostate, the number of splicing errors per cancer has been shown to vary greatly between individual cancers, a phenomenon referred to as transcriptome instability. Transcriptome instability correlates significantly with reduced expression level of splicing factor genes. Mutation of DNMT3A contributes to development of hematologic malignancies, and DNMT3A-mutated cell lines exhibit transcriptome instability as compared to their isogenic wildtype counterparts.
|
https://en.wikipedia.org/wiki/21st%20Century%20Medicine
|
21st Century Medicine (21CM) is a California cryobiological research company which has as its primary focus the development of perfusates and protocols for viable long-term cryopreservation of human organs, tissues and cells at temperatures below −100 °C through the use of vitrification. 21CM was founded in 1993.
In 2004 21CM received a $900,000 grant from the U.S. National Institutes of Health (NIH) to study a preservation solution developed by the University of Rochester in New York for extending simple cold storage time of human hearts removed for transplant.
At the July 2005 annual conference of the Society for Cryobiology, 21st Century Medicine announced the vitrification of a rabbit kidney to −135 °C with their vitrification mixture. The kidney was successfully transplanted upon rewarming to a rabbit, the rabbit being euthanized on the 48th day for histological follow-up.
On February 9, 2016, 21st Century Medicine won the Small Mammal Brain Preservation Prize. On March 13, 2018, they won the Large Mammal Brain Preservation Prize.
|
https://en.wikipedia.org/wiki/Composite%20methods%20for%20structural%20dynamics
|
Composite methods are an approach applied in structural dynamics and related fields. They combine various methods in each time step, in order to acquire the advantages of different methods. The existing composite methods show satisfactory accuracy and powerful numerical dissipation, which is particularly useful for solving stiff problems and differential-algebraic equations.
Definitions
After spatial discretization, structural dynamics problems are generally described by the second-order ordinary differential equation:
.
Here , and denote the displacement, velocity and acceleration vectors respectively, is the mass matrix, is the damping matrix, collects the internal force, and is the external load. At the initial time , the initial displacement and velocity are supposed to be given as and , respectively, and the initial acceleration can be solved as
.
For numerical analysis, the overall time domain is divided into a series of time steps by , , , , , . Taking the step
( is the step size), the main concept of composite methods is to subdivide the current step to several sub-steps , , , and to use different numerical methods in each sub-step.
Although there are lots of available methods, see the review, the existing composite methods basically employ the combination of the trapezoidal rule and linear multistep methods. However, to acquire at least second-order accuracy and unconditional stability, the scalar parameters of each method and the division of sub-steps need to be determined carefully.
Two examples of composite method
Two-sub-step Bathe method
The Bathe method is a two-sub-step method. In the first sub-step (, ), the trapezoidal rule is used as:
In the second sub-step (), the 3-point Euler backward method is employed as
For nonlinear dynamics, that is, the internal force is a nonlinear function with respect to , the Newton-Raphson iterations can be used to solve the nonlinear equations per step. The parameter is usually set as
|
https://en.wikipedia.org/wiki/Clinical%20Anatomy
|
Clinical Anatomy is a peer-reviewed medical journal that covers anatomy in all its aspects—gross, histologic, developmental, and neurologic—as applied to medical practice. It is the official publication of the American Association of Clinical Anatomists, the British Association of Clinical Anatomists, the Australian and New Zealand Association of Clinical Anatomists, and the Anatomical Society of Southern Africa.
According to the Journal Citation Reports, the journal has a 2022 impact factor of 2.4, ranking it 7th out of 20 journals in the category "Anatomy & Morphology".
|
https://en.wikipedia.org/wiki/Higgins%20project
|
Higgins is an open-source project dedicated to giving individuals more control over their personal identity, profile and social network data.
The project is organized into three main areas:
Active Clients — An active client integrates with a browser and runs on a computer or mobile device.
Higgins 1. X: the active client supports the OASIS IMI protocol and performs the functions of an Information Card selector.
Higgins 2.0: the plan is to move beyond selector functionality to add support for managing passwords and Higgins relationship cards, as well other protocols such as OpenID. It also becomes a client for the Personal Data Store (see below) and thereby provides a kind of dashboard for personal information and a place to manage "permissioning" — deciding who gets access to what slice of the user's data.
Personal Data Store (PDS) is a new work area under development for Higgins 2.0. A PDS stores local personal data, controls access to remotely hosted personal data, synchronizes personal data to other devices and computers, accessed directly or via a PDS client it allows the user to share selected aspects of their information with people and organizations that they trust.
Identity Services — Code for (i) an IMI and SAML compatible Identity Provider, and (ii) enabling websites to be IMI and OpenID compatible.
History
The initial code for the Higgins Project was written by Paul Trevithick in the summer of 2003. In 2004 the effort became part of SocialPhysics.org, a collaboration between Paul and Mary Ruddy, of Azigo, (formerly Parity Communications, Inc.), and Meristic, and John Clippinger, at the Berkman Center for Internet & Society. Higgins, under its original name Eclipse Trust Framework, was accepted into the Eclipse Foundation in early 2005. Mary and Paul are the project co-leads. IBM and Novell's participation in the project was announced in early 2006. Higgins has received technology contributions from IBM, Novell, Oracle, CA, Serena, Google, eperi GmbH a
|
https://en.wikipedia.org/wiki/Four-sides%20model
|
The four-sides model (also known as communication square or four-ears model) is a communication model postulated in 1981 by German psychologist Friedemann Schulz von Thun. According to this model every message has four facets though not the same emphasis might be put on each. The four sides of the message are fact, self-disclosure, Social relationship between sender and receiver, and wish or want.
Background
The four-sides model also known as communication square or four-ears model is a communication model described in 1981 by German psychologist Friedemann Schulz von Thun. It describes the multi-layered structure of human utterances. In it von Thun combined the idea of a postulate (the second axiom) from psychologist Paul Watzlawick, that every message contains content and relational facets, with the three sides of the Organon model by Karl Bühler, that every message might reveal something about the sender, the receiver, and the request at hand. These models are part of the linguistic speech act theory.
The four sides of communication
Per Schulz von Thun
The Factual Level contains statements which are matter of fact like data and facts, which are part of the message.
In the self-revealing or self-disclosure the speaker - deliberately or unintentionally - reveals something about him or herself, their motives, values, emotions etc.
In the relationship-layer the speaker expresses, how the sender gets along with the receiver and what they think of each other.
The wish or want contains the plea or desire, the advice, instruction and possibly the effects which the speaker is seeking.
Every layer of a message can be misunderstood by itself.
The classic example of Schulz von Thun is the front-seat passenger who tells the driver: "Hey, the traffic lights are green". The driver will understand something different, depending on the ear with which he will hear, and will react differently. (On the matter layer he will understand the "fact" "the traffic lights are gree
|
https://en.wikipedia.org/wiki/List%20of%20large%20cardinal%20properties
|
This page includes a list of cardinals with large cardinal properties. It is arranged roughly in order of the consistency strength of the axiom asserting the existence of cardinals with the given property. Existence of a cardinal number κ of a given type implies the existence of cardinals of most of the types listed above that type, and for most listed cardinal descriptions φ of lesser consistency strength, Vκ satisfies "there is an unbounded class of cardinals satisfying φ".
The following table usually arranges cardinals in order of consistency strength, with size of the cardinal used as a tiebreaker. In a few cases (such as strongly compact cardinals) the exact consistency strength is not known and the table uses the current best guess.
"Small" cardinals: 0, 1, 2, ..., ,..., , ... (see Aleph number)
worldly cardinals
weakly and strongly inaccessible, α-inaccessible, and hyper inaccessible cardinals
weakly and strongly Mahlo, α-Mahlo, and hyper Mahlo cardinals.
reflecting cardinals
weakly compact (= Π-indescribable), Π-indescribable, totally indescribable cardinals
λ-unfoldable, unfoldable cardinals, ν-indescribable cardinals and λ-shrewd, shrewd cardinals (not clear how these relate to each other).
ethereal cardinals, subtle cardinals
almost ineffable, ineffable, n-ineffable, totally ineffable cardinals
remarkable cardinals
α-Erdős cardinals (for countable α), 0# (not a cardinal), γ-iterable, γ-Erdős cardinals (for uncountable γ)
almost Ramsey, Jónsson, Rowbottom, Ramsey, ineffably Ramsey, completely Ramsey, strongly Ramsey, super Ramsey cardinals
measurable cardinals, 0†
λ-strong, strong cardinals, tall cardinals
Woodin, weakly hyper-Woodin, Shelah, hyper-Woodin cardinals
superstrong cardinals (=1-superstrong; for n-superstrong for n≥2 see further down.)
subcompact, strongly compact (Woodin< strongly compact≤supercompact), supercompact, hypercompact cardinals
η-extendible, extendible cardinals
Vopěnka cardinals, Shelah for supercompactness,
|
https://en.wikipedia.org/wiki/National%20Microbiology%20Laboratory
|
The National Microbiology Laboratory (NML) is part of the Public Health Agency of Canada (PHAC), the agency of the Government of Canada that is responsible for public health, health emergency preparedness and response, and infectious and chronic disease control and prevention.
NML is located in several sites across the country including the Canadian Science Centre for Human and Animal Health (CSCHAH) in Winnipeg, Manitoba. NML has a second site in Winnipeg, the JC Wilt Infectious Disease Research Centre on Logan Avenue which serves as a hub for HIV research and diagnostics in Canada. The three other primary sites include locations in Guelph, St. Hyacinthe and Lethbridge.
The CSCHAH is a biosafety level 4 infectious disease laboratory facility, the only one of its kind in Canada. With maximum containment, scientists are able to work with pathogens including Ebola, Marburg and Lassa fever.
The NML's CSCHAH is also home to the Canadian Food Inspection Agency's National Centre for Foreign Animal Disease, and thus the scientists at the NML share their premises with animal virologists.
History
The National Microbiology Laboratory was preceded by the Bureau of Microbiology which was originally part of the Laboratory Centre for Disease Control of Health Canada in Ottawa. In the 1980s, Health Canada identified both the need to replace existing laboratory space that was reaching the end of its lifespan and the need for Containment Level 4 space in the country.
Around the same time, Agriculture Canada (prior to the Canadian Food Inspection Agency being formed) also identified the need for new laboratory space including high-containment. Numerous benefits were identified for housing both laboratories in one building and Winnipeg was chosen as the site; an announcement to that effect was made in October 1987.
After some debate, the spot chosen for the site was a city works yard near to the Health Sciences Centre (a major teaching hospital), the University of Manitoba's me
|
https://en.wikipedia.org/wiki/Information%20ratio
|
The information ratio measures and compares the active return of an investment (e.g., a security or portfolio) compared to a benchmark index relative to the volatility of the active return (also known as active risk or benchmark tracking risk). It is defined as the active return (the difference between the returns of the investment and the returns of the benchmark) divided by the tracking error (the standard deviation of the active return, i.e., the additional risk). It represents the additional amount of return that an investor receives per unit of increase in risk.
The information ratio is simply the ratio of the active return of the portfolio divided by the tracking error of its return, with both components measured relative to the performance of the agreed-on benchmark.
It is often used to gauge the skill of managers of mutual funds, hedge funds, etc. It measures the active return of the manager's portfolio divided by the amount of risk that the manager takes relative to the benchmark. The higher the information ratio, the higher the active return of the portfolio, given the amount of risk taken, and the better the manager.
The information ratio is similar to the Sharpe ratio, the main difference being that the Sharpe ratio uses a risk-free return as benchmark (such as a U.S. Treasury security) whereas the information ratio uses a risky index as benchmark (such as the S&P500). The Sharpe ratio is useful for an attribution of the absolute returns of a portfolio, and the information ratio is useful for an attribution of the relative returns of a portfolio.
Definition
The information ratio is defined as:
,
where is the portfolio return, is the benchmark return, is the expected value of the active return, and is the standard deviation of the active return, which is an alternate definition of the aforementioned tracking error.
Note in this case, is defined as excess return, not the risk-adjusted excess return or Jensen's alpha calculated using regress
|
https://en.wikipedia.org/wiki/Magnapinna%20sp.%20B
|
Magnapinna'' sp. B is an undescribed species of bigfin squid known only from a single immature specimen collected in the northern Atlantic Ocean.
Description
It is characterised by its dark epidermal pigmentation, which is epithelial, as opposed to the chromatophoral pigmentation found in other Magnapinna species.
Discovery
The only known specimen of Magnapinna sp. B is a juvenile male of mantle length (ML) held in the Bergen Museum. It was caught by the R/V G.O. SARS (MAR-ECO cruise super station 46, local station 374) on July 11, 2004, at .
|
https://en.wikipedia.org/wiki/Bitwise%20operations%20in%20C
|
In the C programming language, operations can be performed on a bit level using bitwise operators.
Bitwise operations are contrasted by byte-level operations which characterize the bitwise operators' logical counterparts, the AND, OR, NOT operators. Instead of performing on individual bits, byte-level operators perform on strings of eight bits (known as bytes) at a time. The reason for this is that a byte is normally the smallest unit of addressable memory (i.e. data with a unique memory address).
This applies to bitwise operators as well, which means that even though they operate on only one bit at a time they cannot accept anything smaller than a byte as their input.
All of these operators are also available in C++, and many C-family languages.
Bitwise operators
C provides six operators for bit manipulation.
Bitwise AND &
The bitwise AND operator is a single ampersand: &. It is just a representation of AND which does its work on the bits of the operands rather than the truth value of the operands. Bitwise binary AND performs logical conjunction (shown in the table above) of the bits in each position of a number in its binary form.
For instance, working with a byte (the char type):
11001000
& 10111000
--------
= 10001000
The most significant bit of the first number is 1 and that of the second number is also 1 so the most significant bit of the result is 1; in the second most significant bit, the bit of second number is zero, so we have the result as 0.
Bitwise OR |
Similar to bitwise AND, bitwise OR performs logical disjunction at the bit level. Its result is a 1 if either of the bits is 1 and zero only when both bits are 0. Its symbol is | which can be called a pipe.
11001000
| 10111000
--------
= 11111000
Bitwise XOR ^
The bitwise XOR (exclusive or) performs an exclusive disjunction, which is equivalent to adding two bits and discarding the carry. The result is zero only when we have two zeroes or two
|
https://en.wikipedia.org/wiki/Landau%20Gold%20Medal
|
The Landau Gold Medal () is the highest award in theoretical physics awarded by the Russian Academy of Sciences and its predecessor the Soviet Academy of Sciences. It was established in 1971 and is named after Soviet physicist and Nobel Laureate Lev Landau. When awarded by the Soviet Academy of Sciences the award was the "Landau Prize"; the name was changed to the "Landau Gold Medal" in 1992.
Prize laureates
1971 - Vladimir Gribov
1974 - Evgeny Lifshitz, Vladimir Belinski, and Isaak Khalatnikov
1977 - Arkady Migdal
1980 - Aleksandr Gurevich and Lev Pitaevskii
1981- Eva Jablonka
1983 - Alexander Patashinski and Valery Pokrovsky
1986 - Boris Shklovskii and Alexei L. Efros
1989 - Alexei Abrikosov, Lev Gor'kov, and Igor Dzyaloshinskii
1992 - Grigoriy Volovik and Vladimir P. Mineev
1998 - Spartak Belyaev
2002 - Lev Okun
2008 - Lev Pitaevskii
2013 - Semyon Gershtein
2018 - Valery Pokrovsky
See also
List of physics awards
Prizes named after people
|
https://en.wikipedia.org/wiki/Internal%20iliac%20lymph%20nodes
|
The internal iliac lymph nodes (or hypogastric) surround the internal iliac artery and its branches (the hypogastric vessels), and receive the lymphatics corresponding to the distribution of the branches of it, i. e., they receive lymphatics from all the pelvic viscera, from the deeper parts of the perineum, including the membranous and cavernous portions of the urethra, and from the buttock and back of the thigh. The internal iliac lymph nodes also drain the superior half of the rectum, above the pectinate line.
It does not receive lymph from the ovary or testis, which drain to the paraaortic lymph nodes.
Additional images
See also
External iliac lymph nodes
pararectal lymph nodes
|
https://en.wikipedia.org/wiki/Electric-field%20screening
|
In physics, screening is the damping of electric fields caused by the presence of mobile charge carriers. It is an important part of the behavior of charge-carrying fluids, such as ionized gases (classical plasmas), electrolytes, and charge carriers in electronic conductors (semiconductors, metals).
In a fluid, with a given permittivity , composed of electrically charged constituent particles, each pair of particles (with charges and ) interact through the Coulomb force as
where the vector is the relative position between the charges. This interaction complicates the theoretical treatment of the fluid. For example, a naive quantum mechanical calculation of the ground-state energy density yields infinity, which is unreasonable. The difficulty lies in the fact that even though the Coulomb force diminishes with distance as , the average number of particles at each distance is proportional to , assuming the fluid is fairly isotropic. As a result, a charge fluctuation at any one point has non-negligible effects at large distances.
In reality, these long-range effects are suppressed by the flow of particles in response to electric fields. This flow reduces the effective interaction between particles to a short-range "screened" Coulomb interaction. This system corresponds to the simplest example of a renormalized interaction.
In solid-state physics, especially for metals and semiconductors, the screening effect describes the electrostatic field and Coulomb potential of an ion inside the solid. Like the electric field of the nucleus is reduced inside an atom or ion due to the shielding effect, the electric fields of ions in conducting solids are further reduced by the cloud of conduction electrons.
Description
Consider a fluid composed of electrons moving in a uniform background of positive charge (one-component plasma). Each electron possesses a negative charge. According to Coulomb's interaction, negative charges repel each other. Consequently, this electron will
|
https://en.wikipedia.org/wiki/Phase-locked%20loop%20range
|
The terms hold-in range, pull-in range (acquisition range), and lock-in range are widely used by engineers for the concepts of frequency deviation ranges within which phase-locked loop-based circuits can achieve lock under various additional conditions.
History
In the classic books on phase-locked loops, published in 1966, such concepts as hold-in, pull-in, lock-in, and other frequency ranges for which PLL can achieve lock, were introduced. They are widely used nowadays (see, e.g. contemporary engineering literature and other publications). Usually in engineering literature only non-strict definitions are given for these concepts.
Many years of using definitions based on the above concepts has led to the advice given in a handbook on synchronization and communications, namely to check the definitions carefully before using them. Later some rigorous mathematical definitions were given in.
Gardner problem on the lock-in range definition
In the 1st edition of his well-known work, Phaselock Techniques, Floyd M. Gardner introduced a lock-in concept: If, for some reason, the frequency difference between input and VCO is less than the loop bandwidth, the loop will lock up almost instantaneously without slipping cycles. The maximum frequency difference for which this fast acquisition is possible is called the lock-in frequency. His notion of the lock-in frequency and corresponding definition of the lock-in range have become popular and nowadays are given in various engineering publications. However, since even for zero frequency difference there may exist initial states of loop such that cycle slipping may take place during the acquisition process, the consideration of initial state of the loop is of utmost importance for the cycle slip analysis and, therefore, Gardner’s concept of lock-in frequency lacked rigor and required clarification.
In the 2nd edition of his book, Gardner stated: "there is no natural way to define exactly any unique lock-in frequency", and he wro
|
https://en.wikipedia.org/wiki/Systemness
|
Systemness is the state, quality, or condition of a complex system, that is, of a set of interconnected elements that behave as, or appear to be, a whole, exhibiting behavior distinct from the behavior of the parts. The term is new and has been applied to large social phenomena and organizations (healthcare and higher education) by advocates of higher degrees of system-like, coherent behavior for delivering value to stakeholders.
In sociology, Montreal-based Polish academic Szymon Chodak (1973) used "societal systemness" in English to describe the empirical reality that inspired Emile Durkheim.
The healthcare-related usage of the term was as early as 1986 in a Dutch psychiatric research paper. It has recently been adapted to describe the sustainability efforts of healthcare institutions amidst budget cuts stemming from the 2008–2012 global recession.
The higher educational use appears to have featured in professional discussions between sociologist Neil Smelser and University of California Chancellor and President Clark Kerr in the 1950s or 60s; in the foreword to Kerr's 2001 memoir, Smelser uses the term in inverted commas in recalling such discussions.
The term's overt operationalization, however, was instituted by The State University of New York's (SUNY) Chancellor Nancy L. Zimpher in the State of the University Address on January 9, 2012. Zimpher noted systemness as "the coordination of multiple components that, when working together, create a network of activity that is more powerful than any action of individual parts on their own." The concept was later explored in the volume, Higher Education Systems 3.0, edited by Jason E. Lane and D. Bruce Johnston.
Use in higher education
The term "systemness" has received widespread adoption in discussions within and among the leaders of multi-campus university systems to discuss the evolution of multi-campus collaboration and coordination in a range of different programmatic areas. The term was first coined b
|
https://en.wikipedia.org/wiki/List%20of%20human%20cell%20types%20derived%20from%20the%20germ%20layers
|
This is a list of cells in humans derived from the three embryonic germ layers – ectoderm, mesoderm, and endoderm.
Cells derived from ectoderm
Surface ectoderm
Skin
Trichocyte
Keratinocyte
Anterior pituitary
Gonadotrope
Corticotrope
Thyrotrope
Somatotrope
Lactotroph
Tooth enamel
Ameloblast
Neural crest
Peripheral nervous system
Neuron
Glia
Schwann cell
Satellite glial cell
Neuroendocrine system
Chromaffin cell
Glomus cell
Skin
Melanocyte
Nevus cell
Merkel cell
Teeth
Odontoblast
Cementoblast
Eyes
Corneal keratocyte
Neural tube
Central nervous system
Neuron
Glia
Astrocyte
Ependymocytes
Muller glia (retina)
Oligodendrocyte
Oligodendrocyte progenitor cell
Pituicyte (posterior pituitary)
Pineal gland
Pinealocyte
Cells derived from mesoderm
Paraxial mesoderm
Mesenchymal stem cell
Osteochondroprogenitor cell
Bone (Osteoblast → Osteocyte)
Cartilage (Chondroblast → Chondrocyte)
Myofibroblast
Fat
Lipoblast → Adipocyte
Muscle
Myoblast → Myocyte
Myosatellite cell
Tendon cell
Cardiac muscle cell
Other
Fibroblast → Fibrocyte
Other
Digestive system
Interstitial cell of Cajal
Intermediate mesoderm
Renal stem cell
Angioblast → Endothelial cell
Mesangial cell
Intraglomerular
Extraglomerular
Juxtaglomerular cell
Macula densa cell
Stromal cell → Interstitial cell → Telocytes
Simple epithelial cell → Podocyte
Kidney proximal tubule brush border cell
Reproductive system
Sertoli cell
Leydig cell
Granulosa cell
Peg cell
Germ cells (which migrate here primordially)
spermatozoon
ovum
Lateral plate mesoderm
Hematopoietic stem cell
Lymphoid
Lymphoblast
see lymphocytes
Myeloid
CFU-GEMM
see myeloid cells
Circulatory system
Endothelial progenitor cell
Endothelial colony forming cell
Endothelial stem cell
Angioblast/Mesoangioblast
Pericyte
Mural cell
Body cavities
Mesothelial cell
Cells derived from endoderm
Foregut
Respiratory system
Pneumocyte
Type I cell
Type II cell
Club cell
Goblet cell
|
https://en.wikipedia.org/wiki/In-target%20probe
|
In-target probe, or ITP is a device used in computer hardware and microprocessor design, to control a target microprocessor or similar ASIC at the register level. It generally allows full control of the target device and allows the computer engineer access to individual processor registers, program counter, and instructions within the device. It allows the processor to be single-stepped or for breakpoints to be set. Unlike an in-circuit emulator (ICE), an In-Target Probe uses the target device to execute, rather than substituting for the target device.
See also
Hardware-assisted virtualization
In-circuit emulator
Joint Test Action Group
External links
ITP700 Debug Port Design Guide - Intel
Embedded systems
Debugging
|
https://en.wikipedia.org/wiki/Van%20Gieson%27s%20stain
|
Van Gieson's stain is a mixture of picric acid and acid fuchsin. It is the simplest method of differential staining of collagen and other connective tissue. It was introduced to histology by American neuropsychiatrist and pathologist Ira Van Gieson.
HvG stain generally refers to the combination of hematoxylin and Van Gieson's stain, but can possibly refer to a combination of hibiscus extract-iron solution and Van Gieson's stain.
Other dyes
Other dyes used in connection with Van Gieson staining include:
Alcian blue
Amido black 10B
Verhoeff's stain
|
https://en.wikipedia.org/wiki/Camidanlumab%20tesirine
|
Camidanlumab tesirine (Cami-T or ADCT-301) is an antibody-drug conjugate (ADC) composed of a human antibody that binds to the protein CD25, conjugated to a pyrrolobenzodiazepine dimer toxin.
The experimental drug, developed by ADC Therapeutics is being tested in clinical trials for the treatment of B-cell Hodgkin's lymphoma (HL) and non-Hodgkin lymphoma (NHL), and for the treatment of B-cell acute lymphoblastic leukemia (ALL) and acute myeloid leukemia (AML).
Technology
The human monoclonal antibody is conjugated via a cleavable linker to a cytotoxic (anticancer) pyrrolobenzodiazepine (PBD) dimer. The antibody binds to CD25, which is the alpha chain of the interleukin 2 receptor IL2RA. This molecule is expressed mainly on activated T- and B-cells, both of which are types of white blood cells that play a role in the human immune system. CD25 is over-expressed in a wide range of hematological malignancies, such as leukemias and lymphomas. After binding to a CD25-expressing cell, the antibody is internalized into the cell where enzymes release the cytotoxic drug. PBD dimers work by crosslinking specific sites of the DNA, blocking the cancer cells’ division that cause the cells to die. As a class of DNA-crosslinking agents they are significantly more potent than systemic chemotherapeutic drugs.
Clinical trials
Two phase I trials are evaluating the drug in patients with relapsed or refractory Hodgkin’s and non-Hodgkin’s lymphoma and relapsed or refractory CD25-positive acute myeloid leukemia or acute lymphoblastic leukemia. At the 59th American Society of Hematology (ASH) Annual Meeting, interim results from a Phase I, open-label, single agent dose-escalating study designed to evaluate the treatment of camidanlumab tesirine in relapsed or refractory Hodgkin’s or non-Hodgkin’s lymphoma were presented. Among the patients enrolled at the time of the data cutoff the overall response rate was 78% (including a 44% complete response rate) in patients with relapsing or r
|
https://en.wikipedia.org/wiki/Soil%20thermal%20properties
|
The thermal properties of soil are a component of soil physics that has found important uses in engineering, climatology and agriculture. These properties influence how energy is partitioned in the soil profile. While related to soil temperature, it is more accurately associated with the transfer of energy (mostly in the form of heat) throughout the soil, by radiation, conduction and convection.
The main soil thermal properties are
Volumetric heat capacity, SI Units: J∙m−3∙K−1
Thermal conductivity, SI Units: W∙m−1∙K−1
Thermal diffusivity, SI Units: m2∙s−1
Measurement
It is hard to say something general about the soil thermal properties at a certain location because these are in a constant state of flux from diurnal and seasonal variations. Apart from the basic soil composition, which is constant at one location, soil thermal properties are strongly influenced by the soil volumetric water content, volume fraction of solids and volume fraction of air. Air is a poor thermal conductor and reduces the effectiveness of the solid and liquid phases to conduct heat. While the solid phase has the highest conductivity it is the variability of soil moisture that largely determines thermal conductivity. As such soil moisture properties and soil thermal properties are very closely linked and are often measured and reported together. Temperature variations are most extreme at the surface of the soil and these variations are transferred to sub surface layers but at reduced rates as depth increases. Additionally there is a time delay as to when maximum and minimum temperatures are achieved at increasing soil depth (sometimes referred to as thermal lag).
One possible way of assessing soil thermal properties is the analysis of soil temperature variations versus depth Fourier's law,
where Q is heat flux or rate of heat transfer per unit area J·m−2∙s−1 or W·m−2,
λ is thermal conductivity W·m−1∙K−1;
dT/dz is the gradient of temperature (change in temp/change in depth) K·m−1.
The mo
|
https://en.wikipedia.org/wiki/Conformal%20group
|
In mathematics, the conformal group of an inner product space is the group of transformations from the space to itself that preserve angles. More formally, it is the group of transformations that preserve the conformal geometry of the space.
Several specific conformal groups are particularly important:
The conformal orthogonal group. If V is a vector space with a quadratic form Q, then the conformal orthogonal group is the group of linear transformations T of V for which there exists a scalar λ such that for all x in V
For a definite quadratic form, the conformal orthogonal group is equal to the orthogonal group times the group of dilations.
The conformal group of the sphere is generated by the inversions in circles. This group is also known as the Möbius group.
In Euclidean space En, , the conformal group is generated by inversions in hyperspheres.
In a pseudo-Euclidean space Ep,q, the conformal group is .
All conformal groups are Lie groups.
Angle analysis
In Euclidean geometry one can expect the standard circular angle to be characteristic, but in pseudo-Euclidean space there is also the hyperbolic angle. In the study of special relativity the various frames of reference, for varying velocity with respect to a rest frame, are related by rapidity, a hyperbolic angle. One way to describe a Lorentz boost is as a hyperbolic rotation which preserves the differential angle between rapidities. Thus, they are conformal transformations with respect to the hyperbolic angle.
A method to generate an appropriate conformal group is to mimic the steps of the Möbius group as the conformal group of the ordinary complex plane. Pseudo-Euclidean geometry is supported by alternative complex planes where points are split-complex numbers or dual numbers. Just as the Möbius group requires the Riemann sphere, a compact space, for a complete description, so the alternative complex planes require compactification for complete description of conformal mapping. Nevertheless,
|
https://en.wikipedia.org/wiki/Boron%20aluminum%20titanium%20hydride
|
Boron Aluminum Titanium Hydride (BATH) was developed as a radiation shielding material in the NERVA project for space nuclear thermal propulsion applications.
|
https://en.wikipedia.org/wiki/Inertial%20fusion%20power%20plant
|
Inertial Fusion Energy is a proposed approach to building a nuclear fusion power plant based on performing inertial confinement fusion at industrial scale. This approach to fusion power is still in a research phase. ICF first developed shortly after the development of the laser in 1960, but was a classified US research program during its earliest years. In 1972, John Nuckolls wrote a paper predicting that compressing a target could create conditions where fusion reactions are chained together, a process known as fusion ignition or a burning plasma. On August 8, 2021, the NIF at Livermore National Laboratory became the first ICF facility in the world to demonstrate this (see plot). This breakthrough drove the US Department of Energy to create an Inertial Fusion Energy program in 2022 with a budget of 3 million dollars in its first year.
Design of a IFE power plant
This kind of fusion reactor would consist of two parts:
Targets which can be small capsules (<7 millimeter diameter) that contain fusion fuel. Although many kinds of targets have been tested including: cylinders, shells coated with nanotubes, solid blocks, hohlraum, glass shells filled with fusion fuel, cryogenically frozen targets, plastic shells, foam shells and materials suspended on spider silk.
Drivers which are used to compress and create a shock wave that squeezes the target. This compression wave pushes the material down to the temperature and pressure where fusion occurs. Drivers that have been explored are solid-state lasers, excimer lasers, high velocity solid objects, X-rays, beams of ions (heavy ion fusion (HIF)) and beams of electrons.
Net energy in ICF comes from getting fusion reactions to chain together in a process known as ignition. To get there we need to squeeze material to hot and dense conditions for long enough. But a key problem is that after a plasma becomes hot - it becomes hard to compress. The goal then is to avoid getting material hot until after it is compress
|
https://en.wikipedia.org/wiki/Flyback%20converter
|
The flyback converter is used in both AC/DC, and DC/DC conversion with galvanic isolation between the input and any outputs. The flyback converter is a buck-boost converter with the inductor split to form a transformer, so that the voltage ratios are multiplied with an additional advantage of isolation. When driving, for example, a plasma lamp or a voltage multiplier, the rectifying diode of the boost converter is left out and the device is called a flyback transformer.
Structure and principle
The schematic of a flyback converter can be seen in Fig. 1. It is equivalent to that of a buck-boost converter, with the inductor split to form a transformer. Therefore, the operating principle of both converters is very similar:
When the switch is closed (top of Fig. 2), the primary of the transformer is directly connected to the input voltage source. The primary current and magnetic flux in the transformer increases, storing energy in the transformer. The voltage induced in the secondary winding is negative, so the diode is reverse-biased (i.e., blocked). The output capacitor supplies energy to the output load.
When the switch is opened (bottom of Fig. 2), the primary current and magnetic flux drops. The secondary voltage is positive, forward-biasing the diode, allowing current to flow from the transformer. The energy from the transformer core recharges the capacitor and supplies the load.
The operation of storing energy in the transformer before transferring to the output of the converter allows the topology to easily generate multiple outputs with little additional circuitry, although the output voltages have to be able to match each other through the turns ratio. Also there is a need for a controlling rail which has to be loaded before load is applied to the uncontrolled rails, this is to allow the PWM to open up and supply enough energy to the transformer.
Operations
The flyback converter is an isolated power converter. The two prevailing control schemes are vo
|
https://en.wikipedia.org/wiki/Founder%20effect
|
In population genetics, the founder effect is the loss of genetic variation that occurs when a new population is established by a very small number of individuals from a larger population. It was first fully outlined by Ernst Mayr in 1942, using existing theoretical work by those such as Sewall Wright. As a result of the loss of genetic variation, the new population may be distinctively different, both genotypically and phenotypically, from the parent population from which it is derived. In extreme cases, the founder effect is thought to lead to the speciation and subsequent evolution of new species.
In the figure shown, the original population has nearly equal numbers of blue and red individuals. The three smaller founder populations show that one or the other color may predominate (founder effect), due to random sampling of the original population. A population bottleneck may also cause a founder effect, though it is not strictly a new population.
The founder effect occurs when a small group of migrants—not genetically representative of the population from which they came—establish in a new area. In addition to founder effects, the new population is often very small, so it shows increased sensitivity to genetic drift, an increase in inbreeding, and relatively low genetic variation.
Founder mutation
In genetics, a founder mutation is a mutation that appears in the DNA of one or more individuals which are founders of a distinct population. Founder mutations initiate with changes that occur in the DNA and can be passed down to other generations. Any organism—from a simple virus to something complex like a mammal—whose progeny carry its mutation has the potential to express the founder effect, for instance a goat or a human.
Founder mutations originate in long stretches of DNA on a single chromosome; indeed, the original haplotype is the whole chromosome. As the generations progress, the proportion of the haplotype that is common to all carriers of the mutation is
|
https://en.wikipedia.org/wiki/Helmholtz%20Centre%20for%20Environmental%20Research
|
The work of the Helmholtz Centre for Environmental Research – UFZ (prior to 28 November 2006 UFZ-Umweltforschungszentrum Leipzig-Halle GmbH) covers both basic research and applied research.
The UFZ was established on 12 December 1991. The Centre commenced its research activities on 2 January 1992. The UFZ has locations in Leipzig, Halle and Magdeburg in Germany. In addition, it operates a experimental station in Bad Lauchstädt. The UFZ employs a total of about 1,200 employees (as of 2021).
The UFZ has been operating KUBUS, a modern communications and event venue in Leipzig, since 2004. KUBUS has flexible event spaces in a range of sizes and accommodates up to 550 people.
Research areas
As an international competence centre for environmental sciences, the UFZ investigates interrelationships between humans and nature under the influence of global change. The research activities of UFZ scientists focus on the terrestrial environment – on densely populated urban and industrial conurbations, on agricultural landscapes and near-natural landscapes. They examine issues relating to future land use, the preservation of biological diversity and of ecosystem services, the sustainable management of soil and water resources and the effect of chemicals on humans and the environment – from the level of single cells and organisms up to the scale of regions.
The work of the UFZ is characterised by integrated environmental research that overcomes disciplinary boundaries between the natural and social sciences (interdisciplinarity) and brings together decision-makers from business, government and society (transdisciplinarity). Major scientific infrastructures such as climate and land-use experiments (e.g., GCEF Global Change Experimental Facility, ProVis Centre for the visualisation of biochemical processes at cellular level), platforms and technologies for environmental monitoring (e.g. TERENO terrestrial environmental observatories ), modelling and visualisation (e.g., TESSIN/V
|
https://en.wikipedia.org/wiki/PyLadies
|
PyLadies is an international mentorship group which focuses on helping more women become active participants in the Python open-source community. It is part of the Python Software Foundation. It was started in Los Angeles in 2011. The mission of the group is to create a diverse Python community through outreach, education, conferences and social gatherings. PyLadies also provides funding for women to attend open source conferences. The aim of PyLadies is increasing the participation of women in computing. PyLadies became a multi-chapter organization with the founding of the Washington, D.C., chapter in August 2011. The group currently has more than 40 chapters around the world.
History
The organization was created in Los Angeles in April 2011 by seven women: Audrey Roy Greenfeld, Christine Cheung, Esther Nam, Jessica Venticinque (Stanton at the time), Katharine Jarmul, Sandy Strong, and Sophia Viklund. Around 2012, the organization filed for nonprofit status.
As of November 2022, PyLadies has over 100 chapters.
About
PyLadies has conducted outreach events for both beginners and experienced users. PyLadies has conducted hackathons, social nights and workshops for Python enthusiasts.
Each chapter is free to run themselves as they wish as long as they are focused on the goal of empowering women and other marginalized genders in tech. Women make up the majority of the group, but membership is not limited to women and the group is open to helping people who identify as other gender identities as well.
|
https://en.wikipedia.org/wiki/Language%20equation
|
Language equations are mathematical statements that resemble numerical equations, but the variables assume values of formal languages rather than numbers. Instead of arithmetic operations in numerical equations, the variables are joined by language operations. Among the most common operations on two languages A and B are the set union A ∪ B, the set intersection A ∩ B, and the concatenation A⋅B. Finally, as an operation taking a single operand, the set A* denotes the Kleene star of the language A. Therefore language equations can be used to represent formal grammars, since the languages generated by the grammar must be the solution of a system of language equations.
Language equations and context-free grammars
Ginsburg and Rice
gave an alternative definition of context-free grammars by language equations. To every context-free grammar , is associated a system of equations in variables . Each variable is an unknown language over and is defined by the equation where , ..., are all productions for . Ginsburg and Rice used a fixed-point iteration argument to show that a solution always exists, and proved that i.e. any other solution must be a of this one.
For example, the grammar
corresponds to the equation system
which has as solution every superset of .
Language equations with added intersection analogously correspond to conjunctive grammars.
Language equations and finite automata
Brzozowski and Leiss studied left language equations where every concatenation is with a singleton constant language on the left, e.g. with variable , but not nor . Each equation is of the form with one variable on the right-hand side. Every nondeterministic finite automaton has such corresponding equation using left-concatenation and union, see Fig. 1. If intersection operation is allowed, equations correspond to alternating finite automata.
Baader and Narendran studied equations using left-concatenation and union and proved that their satisfiability problem is EXPTIME-c
|
https://en.wikipedia.org/wiki/Markus%20Ralser
|
Markus Ralser (born 3 April 1980 in Vipiteno, Italy) is an Italian biologist. His main research interest is metabolism of microorganisms. He is also known for his work on the origin of metabolism during the origin of life, and proteomics.
Life and career
Prof. Ralser serves since 2019 as head of the Institute of Biochemistry at the Charité – Universitätsmedizin Berlin, Germany; as well as since 2022 as group leader at the University of Oxford, UK.
He studied genetics and molecular biology in Salzburg, Austria. He completed his PhD in 2006 at the Max Planck Institute for Molecular Genetics in Berlin, Germany, studying neurodegenerative diseases. This was followed by a postdoctoral fellowship at the Vrije Universiteit Amsterdam, Netherlands, where he started to explore mass spectrometry. He returned to the MPI for Molecular Genetics in 2007 to become junior group leader, but in 2011 relocated his group to the University of Cambridge, UK. He then relocated again, becoming group leader at the newly opened Francis Crick Institute in London in 2013 (senior group leader since 2019). His group moved to Oxford in 2022.
Research
Ralser's two research groups use LC–MS to analyze the proteomes and metabolomes of microorganisms. The main model organism is the baking yeast (Saccharomyces cerevisiae), but other species, such as pathogenic fungus Candida albicans and the fission yeast Schizosaccharomyces pombe, are used too.
The Ralser lab not only uses LC–MS, but also develops novel LC–MS methods and protocols that improve detection accuracy, speed, and throughput. Specializing in data-independent acquisition, the group has developed scanning SWATH MS and Zeno SWATH MS in collaboration with MS manufacturer SCIEX. Both methods greatly improve upon SWATH MS, which was developed in Switzerland in 2012. The group additionally developed an acquisition method—DIA-NN—that uses neural networks. But proteins and metabolites are not the only focus: in 2022 the lab developed a protoco
|
https://en.wikipedia.org/wiki/Ethics%20committee
|
An ethics committee is a body responsible for ensuring that medical experimentation and human subject research are carried out in an ethical manner in accordance with national and international law.
Specific regions
An ethics committee in the European Union is a body responsible for oversight of medical or human research studies in EU member states. Local terms for a European ethics committee include:
A Research Ethics Committee (REC) in the United Kingdom
A Medical Research Ethics Committee (MREC) in the Netherlands.
A Comité de Protection des Personnes (CPP) in France.
In the United States, an ethics committee is usually known as an institutional review board (IRB) or research ethics board (REB) and is dedicated to overseeing the rights and well-being of research subjects participating in scientific studies in the US. Similarly in Canada, the committee is called a Research Ethics Board (REB).
In Australia, an ethics committee in medical research refers to a Human Research Ethics Committee (HREC).
Since 1977 for the purposes of its subsidies to university research the Government of Canada, under the GOSA Act in the person (since 2015) of its Minister of Innovation, Science and Industry, donates annually to several of the federal funder agencies; these in turn disburse the funds into person-sized chunks. These persons typically are university professors, who are selected according to success in the wielding of soft power as measured by track record. In Canada, the Interagency Advisory Panel on Research Ethics (IAPRE) promotes "the ethical conduct of research involving human participants" under a document sometimes referred to as TCPS2. The panel was jointly started in 2001 by three of the federal university research-funding agencies CIHR, NSERC, and SSHRC. The IAPRE FAQ says that "Failure to comply with the requirements of the TCPS2 by researchers or their institution may result in a recourse by the Agencies." Other organizations have opted to adhere to the
|
https://en.wikipedia.org/wiki/Query%20expansion
|
Query expansion (QE) is the process of reformulating a given query to improve retrieval performance in information retrieval operations, particularly in the context of query understanding.
In the context of search engines, query expansion involves evaluating a user's input (what words were typed into the search query area, and sometimes other types of data) and expanding the search query to match additional documents. Query expansion involves techniques such as:
Finding synonyms of words, and searching for the synonyms as well
Finding semantically related words (e.g. antonyms, meronyms, hyponyms, hypernyms)
Finding all the various morphological forms of words by stemming each word in the search query
Fixing spelling errors and automatically searching for the corrected form or suggesting it in the results
Re-weighting the terms in the original query
Query expansion is a methodology studied in the field of computer science, particularly within the realm of natural language processing and information retrieval.
Precision and recall trade-offs
Search engines invoke query expansion to increase the quality of user search results. It is assumed that users do not always formulate search queries using the best terms. Best in this case may be because the database does not contain the user entered terms.
By stemming a user-entered term, more documents are matched, as the alternate word forms for a user entered term are matched as well, increasing the total recall. This comes at the expense of reducing the precision. By expanding a search query to search for the synonyms of a user entered term, the recall is also increased at the expense of precision. This is due to the nature of the equation of how precision is calculated, in that a larger recall implicitly causes a decrease in precision, given that factors of recall are part of the denominator. It is also inferred that a larger recall negatively impacts overall search result quality, given that many users do no
|
https://en.wikipedia.org/wiki/NH41
|
The NR4A1 gene is a transcription factor important in the development of cells that secrete the hormone insulin-like 3 (INSL3). In general, the NR4A gene family regulates cell growth and differentiation.
In humans, INSL3 aids in the regulation of testicular descent during fetal development. In an adult INSL3 helps in keeping germ cells alive, in both males and females.
In the male reproductive system the NR4A1 gene is a promoter of insulin-like 3 in Leydig cells. Leydig cells are located in the testis and known for the function as a support cell in spermatogenesis. Along with other things they secrete testosterone.
In females, INSL3 is also involved in the reproductive system. It is secreted by thecal and luteal cells in the ovary and thus is important in the maturation of developing oocytes. In the rabbit model, disorders observed involving NR4A1 expression include testicular dysgenesis, and cryptorchidism.
Immunoprecipitation assays show that INSL3 does bind to NR4A1; however, much is still not known about INSL3 regulation and the direct involvement of the NR4A1 gene.
|
https://en.wikipedia.org/wiki/Logarithmic%20integral%20function
|
In mathematics, the logarithmic integral function or integral logarithm li(x) is a special function. It is relevant in problems of physics and has number theoretic significance. In particular, according to the prime number theorem, it is a very good approximation to the prime-counting function, which is defined as the number of prime numbers less than or equal to a given value .
Integral representation
The logarithmic integral has an integral representation defined for all positive real numbers ≠ 1 by the definite integral
Here, denotes the natural logarithm. The function has a singularity at , and the integral for is interpreted as a Cauchy principal value,
Offset logarithmic integral
The offset logarithmic integral or Eulerian logarithmic integral is defined as
As such, the integral representation has the advantage of avoiding the singularity in the domain of integration.
Equivalently,
Special values
The function li(x) has a single positive zero; it occurs at x ≈ 1.45136 92348 83381 05028 39684 85892 02744 94930... ; this number is known as the Ramanujan–Soldner constant.
li(Li^-1(0)) = li(2) ≈ 1.045163 780117 492784 844588 889194 613136 522615 578151...
This is where is the incomplete gamma function. It must be understood as the Cauchy principal value of the function.
Series representation
The function li(x) is related to the exponential integral Ei(x) via the equation
which is valid for x > 0. This identity provides a series representation of li(x) as
where γ ≈ 0.57721 56649 01532 ... is the Euler–Mascheroni constant. A more rapidly convergent series by Ramanujan is
Asymptotic expansion
The asymptotic behavior for x → ∞ is
where is the big O notation. The full asymptotic expansion is
or
This gives the following more accurate asymptotic behaviour:
As an asymptotic expansion, this series is not convergent: it is a reasonable approximation only if the series is truncated at a finite number of terms, and only large values of x are employed
|
https://en.wikipedia.org/wiki/Clockwise
|
Two-dimensional rotation can occur in two possible directions or senses of rotation. Clockwise motion (abbreviated CW) proceeds in the same direction as a clock's hands: from the top to the right, then down and then to the left, and back up to the top. The opposite sense of rotation or revolution is (in Commonwealth English) anticlockwise (ACW) or (in North American English) counterclockwise (CCW).
Three-dimensional rotation can have similarly defined senses when considering the corresponding angular velocity vector.
Terminology
Before clocks were commonplace, the terms "sunwise" and "deasil", "deiseil" and even "deocil" from the Scottish Gaelic language and from the same root as the Latin "dexter" ("right") were used for clockwise. "Widdershins" or "withershins" (from Middle Low German "weddersinnes", "opposite course") was used for counterclockwise.
The terms clockwise and counterclockwise can only be applied to a rotational motion once a side of the rotational plane is specified, from which the rotation is observed. For example, the daily rotation of the Earth is clockwise when viewed from above the South Pole, and counterclockwise when viewed from above the North Pole (considering "above a point" to be defined as "farther away from the center of earth and on the same ray").
Clocks traditionally follow this sense of rotation because of the clock's predecessor: the sundial. Clocks with hands were first built in the Northern Hemisphere (see Clock), and they were made to work like horizontal sundials. In order for such a sundial to work north of the equator during spring and summer, and north of the Tropic of Cancer the whole year, the noon-mark of the dial must be placed northward of the pole casting the shadow. Then, when the Sun moves in the sky (from east to south to west), the shadow, which is cast on the sundial in the opposite direction, moves with the same sense of rotation (from west to north to east). This is why hours must be drawn in horizontal sundi
|
https://en.wikipedia.org/wiki/Dactylitis
|
Dactylitis or sausage digit is inflammation of an entire digit (a finger or toe), and can be painful.
The word dactyl comes from the Greek word "daktylos" meaning "finger". In its medical term, it refers to both the fingers and the toes.
Associated conditions
Dactylitis can occur in seronegative arthropathies, such as psoriatic arthritis and ankylosing spondylitis, and in sickle-cell disease as result of a vasoocclusive crisis with bone infarcts, and in infectious conditions including tuberculosis, syphilis, and leprosy. In reactive arthritis, sausage fingers occur due to synovitis. Dactylitis may also be seen with sarcoidosis.
In sickle-cell disease it is manifested for the first time between 6–9 month old infants (as their protective fetal hemoglobin, HbF, is replaced with adult hemoglobin and the disease manifests) and is very often the presenting sign of the disorder.
|
https://en.wikipedia.org/wiki/Boolean%20algebras%20canonically%20defined
|
Boolean algebras are models of the equational theory of two values; this definition is equivalent to the lattice and ring definitions.
Boolean algebra is a mathematically rich branch of abstract algebra. Stanford Encyclopaedia of Philosophy defines Boolean algebra as 'the algebra of two-valued logic with only sentential connectives, or equivalently of algebras of sets under union and complementation.' Just as group theory deals with groups, and linear algebra with vector spaces, Boolean algebras are models of the equational theory of the two values 0 and 1 (whose interpretation need not be numerical). Common to Boolean algebras, groups, and vector spaces is the notion of an algebraic structure, a set closed under some operations satisfying certain equations.
Just as there are basic examples of groups, such as the group of integers and the symmetric group of permutations of objects, there are also basic examples of Boolean algebras such as the following.
The algebra of binary digits or bits 0 and 1 under the logical operations including disjunction, conjunction, and negation. Applications include the propositional calculus and the theory of digital circuits.
The algebra of sets under the set operations including union, intersection, and complement. Applications are far-reaching because set theory is the standard foundations of mathematics.
Boolean algebra thus permits applying the methods of abstract algebra to mathematical logic and digital logic.
Unlike groups of finite order, which exhibit complexity and diversity and whose first-order theory is decidable only in special cases, all finite Boolean algebras share the same theorems and have a decidable first-order theory. Instead, the intricacies of Boolean algebra are divided between the structure of infinite algebras and the algorithmic complexity of their syntactic structure.
Definition
Boolean algebra treats the equational theory of the maximal two-element finitary algebra, called the Boolean prototype,
|
https://en.wikipedia.org/wiki/Architectural%20drawing
|
An architectural drawing or architect's drawing is a technical drawing of a building (or building project) that falls within the definition of architecture. Architectural drawings are used by architects and others for a number of purposes: to develop a design idea into a coherent proposal, to communicate ideas and concepts, to convince clients of the merits of a design, to assist a building contractor to construct it based on design intent, as a record of the design and planned development, or to make a record of a building that already exists.
Architectural drawings are made according to a set of conventions, which include particular views (floor plan, section etc.), sheet sizes, units of measurement and scales, annotation and cross referencing.
Historically, drawings were made in ink on paper or similar material, and any copies required had to be laboriously made by hand. The twentieth century saw a shift to drawing on tracing paper so that mechanical copies could be run off efficiently. The development of the computer had a major impact on the methods used to design and create technical drawings, making manual drawing almost obsolete, and opening up new possibilities of form using organic shapes and complex geometry. Today the vast majority of drawings are created using CAD software.
Size and scale
The size of drawings reflects the materials available and the size that is convenient to transport – rolled up or folded, laid out on a table, or pinned up on a wall. The drafting process may impose limitations on the size that is realistically workable. Sizes are determined by a consistent paper size system, according to local usage. Normally the largest paper size used in modern architectural practice is ISO A0 () or in the USA Arch E () or Large E size ().
Architectural drawings are drawn to scale so that relative sizes are correctly represented. The scale is chosen both to ensure the whole building will fit on the chosen sheet size and to show the required amo
|
https://en.wikipedia.org/wiki/Neutral%20mutation
|
Neutral mutations are changes in DNA sequence that are neither beneficial nor detrimental to the ability of an organism to survive and reproduce. In population genetics, mutations in which natural selection does not affect the spread of the mutation in a species are termed neutral mutations. Neutral mutations that are inheritable and not linked to any genes under selection will be lost or will replace all other alleles of the gene. That loss or fixation of the gene proceeds based on random sampling known as genetic drift. A neutral mutation that is in linkage disequilibrium with other alleles that are under selection may proceed to loss or fixation via genetic hitchhiking and/or background selection.
While many mutations in a genome may decrease an organism’s ability to survive and reproduce, also known as fitness, those mutations are selected against and are not passed on to future generations. The most commonly-observed mutations that are detectable as variation in the genetic makeup of organisms and populations appear to have no visible effect on the fitness of individuals and are therefore neutral. The identification and study of neutral mutations has led to the development of the neutral theory of molecular evolution, which is an important and often-controversial theory that proposes that most molecular variation within and among species is essentially neutral and not acted on by selection. Neutral mutations are also the basis for using molecular clocks to identify such evolutionary events as speciation and adaptive or evolutionary radiations.
History
Charles Darwin commented on the idea of neutral mutation in his work, hypothesizing that mutations that do not give an advantage or disadvantage may fluctuate or become fixed apart from natural selection. "Variations neither useful nor injurious would not be affected by natural selection, and would be left either a fluctuating element, as perhaps we see in certain polymorphic species, or would ultimately become
|
https://en.wikipedia.org/wiki/Loren%20Kohnfelder
|
Loren Kohnfelder invented what is today called public key infrastructure (PKI) in his May 1978 MIT S.B. (BSCSE) thesis, which described a practical means of using public key cryptography to secure network communications.
The Kohnfelder thesis introduced the terms 'certificate' and 'certificate revocation list' as well as introducing numerous other concepts now established as important parts of PKI. The X.509 certificate specification that provides the basis for SSL, S/MIME and most modern PKI implementations are based on the Kohnfelder thesis.
He was also the co-creator, with Praerit Garg, of the STRIDE model of security threats, widely used in threat modeling.
In 2021 he published Designing Secure Software with No Starch Press. He maintains a medium blog.
|
https://en.wikipedia.org/wiki/Residence%20time
|
The residence time of a fluid parcel is the total time that the parcel has spent inside a control volume (e.g.: a chemical reactor, a lake, a human body). The residence time of a set of parcels is quantified in terms of the frequency distribution of the residence time in the set, which is known as residence time distribution (RTD), or in terms of its average, known as mean residence time.
Residence time plays an important role in chemistry and especially in environmental science and pharmacology. Under the name lead time or waiting time it plays a central role respectively in supply chain management and queueing theory, where the material that flows is usually discrete instead of continuous.
History
The concept of residence time originated in models of chemical reactors. The first such model was an axial dispersion model by Irving Langmuir in 1908. This received little attention for 45 years; other models were developed such as the plug flow reactor model and the continuous stirred-tank reactor, and the concept of a washout function (representing the response to a sudden change in the input) was introduced. Then, in 1953, Peter Danckwerts resurrected the axial dispersion model and formulated the modern concept of residence time.
Distributions
The time that a particle of fluid has been in a control volume (e.g. a reservoir) is known as its age. In general, each particle has a different age. The frequency of occurrence of the age in the set of all the particles that are located inside the control volume at time is quantified by means of the (internal) age distribution .
At the moment a particle leaves the control volume, its age is the total time that the particle has spent inside the control volume, which is known as its residence time. The frequency of occurrence of the age in the set of all the particles that are leaving the control volume at time is quantified by means of the residence time distribution, also known as exit age distribution .
Both distr
|
https://en.wikipedia.org/wiki/Motronic
|
Motronic is the trade name given to a range of digital engine control units developed by Robert Bosch GmbH (commonly known as Bosch) which combined control of fuel injection and ignition in a single unit. By controlling both major systems in a single unit, many aspects of the engine's characteristics (such as power, fuel economy, drivability, and emissions) can be improved.
Motronic 1.x
Motronic M1.x is powered by various i8051 derivatives made by Siemens, usually SAB80C515 or SAB80C535. Code/data is stored in DIL or PLCC EPROM and ranges from 32k to 128k.
1.0
Often known as "Motronic basic", Motronic ML1.x was one of the first digital engine-management systems developed by Bosch. These early Motronic systems integrated the spark timing element with then-existing Jetronic fuel injection technology. It was originally developed and first used in the BMW 7 Series, before being implemented on several Volvo and Porsche engines throughout the 1980s.
The components of the Motronic ML1.x systems for the most part remained unchanged during production, although there are some differences in certain situations. The engine control module (ECM) receives information regarding engine speed, crankshaft angle, coolant temperature and throttle position. An air flow meter also measures the volume of air entering the induction system.
If the engine is naturally aspirated, an air temperature sensor is located in the air flow meter to work out the air mass. However, if the engine is turbocharged, an additional charge air temperature sensor is used to monitor the temperature of the inducted air after it has passed through the turbocharger and intercooler, in order to accurately and dynamically calculate the overall air mass.
Main system characteristics
Fuel delivery, ignition timing, and dwell angle incorporated into the same control unit.
Crank position and engine speed is determined by a pair of sensors reading from the flywheel.
Separate constant idle speed system monitors and re
|
https://en.wikipedia.org/wiki/Max%20Planck%20Institute%20for%20Solid%20State%20Research
|
The Max Planck Institute for Solid State Research (German: Max-Planck-Institut für Festkörperforschung) was founded in 1969 and is one of the 82 Max Planck Institutes of the Max Planck Society. It is located on a campus in Stuttgart, together with the Max Planck Institute for Intelligent Systems.
Research focus
Research at the Max Planck Institute for Solid State Research is focused on the physics and chemistry of condensed matter, including especially complex materials and nanoscale science. In both of these fields, electronic and ionic transport phenomena are of particular interest.
Organization
The institute currently has eight departments.
Electronic Structure Theory
Led by Ali Alavi, the Department of Electronic Structure Theory is concerned with the development of ab initio methods for treating correlated electronic systems, using Quantum Monte Carlo, quantum chemical and many-body methodologies. Ab initio methods (including density functional theory) will be applied to problems of interest in heterogeneous catalysis, surface chemistry, electrochemistry, and photochemistry.
Solid State Spectroscopy
The Department of Solid State Spectroscopy is headed by Bernhard Keimer. Collective quantum phenomena in highly correlated electronic materials are studied by spectroscopic and scattering techniques. Topics of particular current interest include the interplay between charge, orbital, and spin degrees of freedom in transition metal oxides, the mechanism of high-temperature superconductivity, and the control of electronic phase behavior in metal-oxide superlattices. The department also develops new spectroscopic methods such as high-resolution neutron spectroscopy and spectral ellipsometry.
Nanoscale Science
Research efforts in the Department of Nanoscale Science, directed by Klaus Kern, are centered on nanometer-scale science and technology with a focus on the bottom-up paradigm. The aim of the interdisciplinary research at the interface between physics, chemi
|
https://en.wikipedia.org/wiki/Bruhathkayosaurus
|
Bruhathkayosaurus (; meaning "huge-bodied lizard") is an extinct genus of sauropod dinosaur found in the Kallamedu Formation of India. The fragmentary remains were originally described as a theropod, but it was later determined to be a titanosaurian sauropod. Length estimates by researchers exceed those of the titanosaur Argentinosaurus, as longer than and weighing over 80 tonnes. A 2023 estimate placed Bruhathkayosaurus as potentially weighing approximately . If the upper estimates of the 2023 records are accurate, Bruhathkayosaurus may have rivalled the blue whale as one of the largest animals to ever exist. However, all of the estimates are based on the dimensions of the fossils described in Yadagiri and Ayyasami (1987), and in 2017, it was reported that the holotype fossils had disintegrated and no longer exist.
Discovery and naming
The holotype of Bruhathkayosaurus, GSI PAL/SR/20, was discovered around 1978 near the southern tip of India, specifically in the Tiruchirappalli district of Tamil Nadu, northeast of Kallamedu village. It was recovered from rocks of the Kallamedu Formation, which are dated to the Maastrichtian stage of the Late Cretaceous, about 70 million years ago. The fossilized remains include hip bones (the ilium and ischium), partial leg bones (femur and tibia), a forearm (radius) and a tail bone (part of a vertebra, specifically a platycoelous caudal centrum). The remains were originally classified as belonging to a carnosaurian theropod by Yadagiri and Ayyasami in 1987 (not 1989, as some sources indicate). The generic name chosen, "Bruhathkayosaurus", is derived from a combination of the Sanskrit word Bruhathkaya (bṛhat , 'huge, heavy' and kāya, काय 'body'), and the Greek sauros (lizard). The specific epithet, "matleyi", honours British palaeontologist Charles Alfred Matley, who discovered many fossils in India.
The monsoon season, combined with the sands and clays of the Kallamedu Formation, creates water-saturated fossils which are very
|
https://en.wikipedia.org/wiki/Neural%20tissue%20engineering
|
Neural tissue engineering is a specific sub-field of tissue engineering. Neural tissue engineering is primarily a search for strategies to eliminate inflammation and fibrosis upon implantation of foreign substances. Often foreign substances in the form of grafts and scaffolds are implanted to promote nerve regeneration and to repair damage caused to nerves of both the central nervous system (CNS) and peripheral nervous system (PNS) by an injury.
Introduction
The nervous system is divided into two sections: the CNS and the PNS. The CNS consists of the brain and the spinal cord, while the PNS consists of nerves that originate from the brain and spinal cord and innervate the rest of the body.
The need for neural tissue engineering arises from the difficulty of the nerve cells and neural tissues to regenerate on their own after neural damage has occurred. The PNS has some, but limited, regeneration of neural cells. Adult stem cell neurogenesis in the CNS has been found to occur in the hippocampus, the subventricular zone (SVZ), and spinal cord. CNS injuries can be caused by stroke, neurodegenerative disorders, trauma, or encephalopathy. A few methods currently being investigated to treat CNS injuries are: implanting stem cells directly into the injury site, delivering morphogens to the injury site, or growing neural tissue in vitro with neural stem or progenitor cells in a 3D scaffold. Proposed use of electrospun polymeric fibrous scaffolds for neural repair substrates dates back to at least 1986 in a NIH SBIR application from Simon. For the PNS, a severed nerve can be reconnected and reinnervated using grafts or guidance of the existing nerve through a channel.
Recent research into creating miniature cortexes, known as corticopoiesis, and brain models, known as cerebral organoids, are techniques that could further the field of neural tissue regeneration. The native cortical progenitors in corticopoiesis are neural tissues that could be effectively embedded into the
|
https://en.wikipedia.org/wiki/Automatic%20group
|
In mathematics, an automatic group is a finitely generated group equipped with several finite-state automata. These automata represent the Cayley graph of the group. That is, they can tell if a given word representation of a group element is in a "canonical form" and can tell if two elements given in canonical words differ by a generator.
More precisely, let G be a group and A be a finite set of generators. Then an automatic structure of G with respect to A is a set of finite-state automata:
the word-acceptor, which accepts for every element of G at least one word in representing it;
multipliers, one for each , which accept a pair (w1, w2), for words wi accepted by the word-acceptor, precisely when in G.
The property of being automatic does not depend on the set of generators.
Properties
Automatic groups have word problem solvable in quadratic time. More strongly, a given word can actually be put into canonical form in quadratic time, based on which the word problem may be solved by testing whether the canonical forms of two words represent the same element (using the multiplier for ).
Automatic groups are characterized by the fellow traveler property. Let denote the distance between in the Cayley graph of . Then, G is automatic with respect to a word acceptor L if and only if there is a constant such that for all words which differ by at most one generator, the distance between the respective prefixes of u and v is bounded by C. In other words, where for the k-th prefix of (or itself if ). This means that when reading the words synchronously, it is possible to keep track of the difference between both elements with a finite number of states (the neighborhood of the identity with diameter C in the Cayley graph).
Examples of automatic groups
The automatic groups include:
Finite groups. To see this take the regular language to be the set of all words in the finite group.
Euclidean groups
All finitely generated Coxeter groups
Geometrically finite grou
|
https://en.wikipedia.org/wiki/Similarities%20between%20Wiener%20and%20LMS
|
The Least mean squares filter solution converges to the Wiener filter solution, assuming that the unknown system is LTI and the noise is stationary. Both filters can be used to identify the impulse response of an unknown system, knowing only the original input signal and the output of the unknown system. By relaxing the error criterion to reduce current sample error instead of minimizing the total error over all of n, the LMS algorithm can be derived from the Wiener filter.
Derivation of the Wiener filter for system identification
Given a known input signal , the output of an unknown LTI system can be expressed as:
where is an unknown filter tap coefficients and is noise.
The model system , using a Wiener filter solution with an order N, can be expressed as:
where are the filter tap coefficients to be determined.
The error between the model and the unknown system can be expressed as:
The total squared error can be expressed as:
Use the Minimum mean-square error criterion over all of by setting its gradient to zero:
which is
for all
Substitute the definition of :
Distribute the partial derivative:
Using the definition of discrete cross-correlation:
Rearrange the terms:
for all
This system of N equations with N unknowns can be determined.
The resulting coefficients of the Wiener filter can be determined by: , where is the cross-correlation vector between and .
Derivation of the LMS algorithm
By relaxing the infinite sum of the Wiener filter to just the error at time , the LMS algorithm can be derived.
The squared error can be expressed as:
Using the Minimum mean-square error criterion, take the gradient:
Apply chain rule and substitute definition of y[n]
Using gradient descent and a step size :
which becomes, for i = 0, 1, ..., N-1,
This is the LMS update equation.
See also
Wiener filter
Least mean squares filter
|
https://en.wikipedia.org/wiki/Paratransgenesis
|
Paratransgenesis is a technique that attempts to eliminate a pathogen from vector populations through transgenesis of a symbiont of the vector. The goal of this technique is to control vector-borne diseases. The first step is to identify proteins that prevent the vector species from transmitting the pathogen. The genes coding for these proteins are then introduced into the symbiont, so that they can be expressed in the vector. The final step in the strategy is to introduce these transgenic symbionts into vector populations in the wild. One use of this technique is to prevent mortality for humans from insect-borne diseases. Preventive methods and current controls against vector-borne diseases depend on insecticides, even though some mosquito breeds may be resistant to them. There are other ways to fully eliminate them. “Paratransgenesis focuses on utilizing genetically modified insect symbionts to express molecules within the vector that are deleterious to pathogens they transmit.” The acidic bacteria Asaia symbionts are beneficial in the normal development of mosquito larvae; however, it is unknown what Asais symbionts do to adult mosquitoes.
The first example of this technique used Rhodnius prolixus which is associated with the symbiont Rhodococcus rhodnii. R. prolixus is an important insect vector of Chagas disease that is caused by Trypanosoma cruzi. The strategy was to engineer R. rhodnii to express proteins such as Cecropin A that are toxic to T. cruzi or that block the transmission of T. cruzi.
Attempts are also made in Tse-tse flies using bacteria and in malaria mosquitoes using fungi, viruses, or bacteria.
Uses
Although the use of paratransgenesis can serve many different purposes, one of the main purposes is “breaking the disease cycle”. This study focuses on the experiments with tsetse flies and trypanosomes, which cause sleeping sickness in Subsaharan Africa. The tsetse fly’s transmission biology was studied to learn how it transmits the disease. This
|
https://en.wikipedia.org/wiki/Hahn%20decomposition%20theorem
|
In mathematics, the Hahn decomposition theorem, named after the Austrian mathematician Hans Hahn, states that for any measurable space and any signed measure defined on the -algebra , there exist two -measurable sets, and , of such that:
and .
For every such that , one has , i.e., is a positive set for .
For every such that , one has , i.e., is a negative set for .
Moreover, this decomposition is essentially unique, meaning that for any other pair of -measurable subsets of fulfilling the three conditions above, the symmetric differences and are -null sets in the strong sense that every -measurable subset of them has zero measure. The pair is then called a Hahn decomposition of the signed measure .
Jordan measure decomposition
A consequence of the Hahn decomposition theorem is the , which states that every signed measure defined on has a unique decomposition into a difference of two positive measures, and , at least one of which is finite, such that for every -measurable subset and for every -measurable subset , for any Hahn decomposition of . We call and the positive and negative part of , respectively. The pair is called a Jordan decomposition (or sometimes Hahn–Jordan decomposition) of . The two measures can be defined as
for every and any Hahn decomposition of .
Note that the Jordan decomposition is unique, while the Hahn decomposition is only essentially unique.
The Jordan decomposition has the following corollary: Given a Jordan decomposition of a finite signed measure , one has
for any in . Furthermore, if for a pair of finite non-negative measures on , then
The last expression means that the Jordan decomposition is the minimal decomposition of into a difference of non-negative measures. This is the minimality property of the Jordan decomposition.
Proof of the Jordan decomposition: For an elementary proof of the existence, uniqueness, and minimality of the Jordan measure decomposition see Fischer (2012).
Proof of
|
https://en.wikipedia.org/wiki/L%C3%A9vy%20hierarchy
|
In set theory and mathematical logic, the Lévy hierarchy, introduced by Azriel Lévy in 1965, is a hierarchy of formulas in the formal language of the Zermelo–Fraenkel set theory, which is typically called just the language of set theory. This is analogous to the arithmetical hierarchy, which provides a similar classification for sentences of the language of arithmetic.
Definitions
In the language of set theory, atomic formulas are of the form x = y or x ∈ y, standing for equality and set membership predicates, respectively.
The first level of the Lévy hierarchy is defined as containing only formulas with no unbounded quantifiers and is denoted by . The next levels are given by finding a formula in prenex normal form which is provably equivalent over ZFC, and counting the number of changes of quantifiers:p. 184
A formula is called:
if is equivalent to in ZFC, where is
if is equivalent to in ZFC, where is
If a formula has both a form and a form, it is called .
As a formula might have several different equivalent formulas in prenex normal form, it might belong to several different levels of the hierarchy. In this case, the lowest possible level is the level of the formula.
Lévy's original notation was (resp. ) due to the provable logical equivalence, strictly speaking the above levels should be referred to as (resp. ) to specify the theory in which the equivalence is carried out, however it is usually clear from context.pp. 441–442 Pohlers has defined in particular semantically, in which a formula is " in a structure ".
The Lévy hierarchy is sometimes defined for other theories S. In this case and by themselves refer only to formulas that start with a sequence of quantifiers with at most i−1 alternations, and and refer to formulas equivalent to and formulas in the language of the theory S. So strictly speaking the levels and of the Lévy hierarchy for ZFC defined above should be denoted by and .
Examples
Σ0=Π0=Δ0 formulas and con
|
https://en.wikipedia.org/wiki/Steve%20Furber
|
Stephen Byram Furber (born 21 March 1953) is a British computer scientist, mathematician and hardware engineer, currently the ICL Professor of Computer Engineering in the Department of Computer Science at the University of Manchester, UK. After completing his education at the University of Cambridge (BA, MMath, PhD), he spent the 1980s at Acorn Computers, where he was a principal designer of the BBC Micro and the ARM 32-bit RISC microprocessor. , over 100 billion copies of the ARM processor have been manufactured, powering much of the world's mobile computing and embedded systems.
In 1990, he moved to Manchester to lead research into asynchronous systems, low-power electronics and neural engineering, where the Spiking Neural Network Architecture (SpiNNaker) project is delivering a computer incorporating a million ARM processors optimised for computational neuroscience.
Education
Furber was educated at Manchester Grammar School and represented the UK in the International Mathematical Olympiad in Hungary in 1970 winning a bronze medal. He went on to study the Mathematical Tripos as an undergraduate student of St John's College, Cambridge, receiving a Bachelor of Arts (BA) and Master of Mathematics (MMath - Part III of the Mathematical Tripos) degrees. In 1978, he was appointed a Rolls-Royce research fellow in aerodynamics at Emmanuel College, Cambridge and was awarded a PhD in 1980 for research on the fluid dynamics of the Weis-Fogh principle supervised by John Ffowcs Williams. During his PhD in the late 1970s, Furber worked on a voluntary basis for Hermann Hauser and Chris Curry within the fledging Acorn Computers (originally the Cambridge Processor Unit), on a number of projects; notably a microprocessor based fruit machine controller, and the Proton - the initial prototype version of what was to become the BBC Micro, in support of Acorn's tender for the BBC Computer Literacy Project.
Career and research
In 1981, following the completion of his PhD and the award
|
https://en.wikipedia.org/wiki/Wicksellian%20Differential
|
The Wicksellian Differential is derived from Knut Wicksell's theory of interest and is an approximation of the extent of disequilibrium in an economy.
Formula: Wicksellian Differential = Natural Rate of Interest - Money Rate of Interest
Wicksell argued in Interest and Prices that the equilibrium of a credit economy could be ascertained by comparing the money rate of interest to the natural rate of interest. In modern terminology this equates to comparing the cost of capital with the return on capital. In economies where the natural rate is higher than the money rate, credit growth will drive a positive disequilibrium in an economy. When the natural rate of interest is lower than the money rate, the demand for credit dries up leading to a negative disequilibrium and capital destruction.
Wicksell also argued that prices would be stable when the two rates of interest were in equilibrium, an idea that was subsequently demonstrated to have been flawed by Gunnar Myrdal and Friedrich Hayek in the late 1920s and early 1930s.
The Wicksellian Differential therefore provides an alternative estimation of the extent of disequilibrium in a credit-based economy to more standard Taylor Rule approaches which attempt to estimate deviations from a general equilibrium based on the fluctuations in output and the general price level.
|
https://en.wikipedia.org/wiki/Heat%20capacity%20ratio
|
In thermal physics and thermodynamics, the heat capacity ratio, also known as the adiabatic index, the ratio of specific heats, or Laplace's coefficient, is the ratio of the heat capacity at constant pressure () to heat capacity at constant volume (). It is sometimes also known as the isentropic expansion factor and is denoted by (gamma) for an ideal gas or (kappa), the isentropic exponent for a real gas. The symbol is used by aerospace and chemical engineers.
where is the heat capacity, the molar heat capacity (heat capacity per mole), and the specific heat capacity (heat capacity per unit mass) of a gas. The suffixes and refer to constant-pressure and constant-volume conditions respectively.
The heat capacity ratio is important for its applications in thermodynamical reversible processes, especially involving ideal gases; the speed of sound depends on this factor.
Thought experiment
To understand this relation, consider the following thought experiment. A closed pneumatic cylinder contains air. The piston is locked. The pressure inside is equal to atmospheric pressure. This cylinder is heated to a certain target temperature. Since the piston cannot move, the volume is constant. The temperature and pressure will rise. When the target temperature is reached, the heating is stopped. The amount of energy added equals , with representing the change in temperature.
The piston is now freed and moves outwards, stopping as the pressure inside the chamber reaches atmospheric pressure. We assume the expansion occurs without exchange of heat (adiabatic expansion). Doing this work, air inside the cylinder will cool to below the target temperature.
To return to the target temperature (still with a free piston), the air must be heated, but is no longer under constant volume, since the piston is free to move as the gas is reheated. This extra heat amounts to about 40% more than the previous amount added. In this example, the amount of heat added with a locked
|
https://en.wikipedia.org/wiki/Thomson%20%28unit%29
|
The thomson (symbol: Th) is a unit that has appeared infrequently in scientific literature relating to the field of mass spectrometry as a unit of mass-to-charge ratio. The unit was proposed by Cooks and Rockwood naming it in honour of J. J. Thomson who measured the mass-to-charge ratio of electrons and ions.
Definition
The thomson is defined as
where Da is the symbol for the unit dalton (also called the unified atomic mass unit, symbol u), and e is the elementary charge which is the unit of electric charge in the system of Hartree atomic units.
For example, the ion C7H72+ has a mass of 91 Da. Its charge number is +2, and hence its charge is 2e. The ion will be observed at 45.5 Th in a mass spectrum.
The thomson allows for negative values for negatively charged ions. For example, the benzoate anion would be observed at −121 Th since the charge is −e.
Use
The thomson has been used by some mass spectrometrists, for example Alexander Makarov—the inventor of the Orbitrap—in a scientific poster, and a 2015 presentation. Other uses of the thomson include papers,
and (notably) one book. The journal Rapid Communications in Mass Spectrometry (in which the original article appeared) states that "the thomson (Th) may be used for such purposes as a unit of mass-to-charge ratio although it is not currently approved by IUPAP or IUPAC." Even so, the term has been called "controversial" by RCM's former Editor-in Chief (in a review the Hoffman text cited above). The book, Mass Spectrometry Desk Reference, argues against the use of the thomson. However, the editor-in-chief of the Journal of the Mass Spectrometry Society of Japan has written an editorial in support of the thomson unit.
The thomson is not an SI unit, nor has it been defined by IUPAC.
Since 2013, the thomson is deprecated by IUPAC (Definitions of Terms Relating to Mass Spectrometry). Since 2014, Rapid Communications in Mass Spectrometry regards the thomson as a "term that should be avoided in mass spectrometry p
|
https://en.wikipedia.org/wiki/Tetter
|
Tetter refers to any skin condition characterized by reddish vesicular eruptions and intense itching. Common diseases called tetter include:
Eczema and Duhring's disease
Herpes
Porphyria cutanea tarda (PCT)
Psoriasis
Ringworm and jock itch
Cutaneous conditions
Inflammations
|
https://en.wikipedia.org/wiki/Direct-coupled%20amplifier
|
A direct-coupled amplifier or DC amplifier is a type of amplifier in which the output of one stage of the amplifier is coupled to the input of the next stage in such a way as to permit signals with zero frequency, also referred to as direct current, to pass from input to output. This is an application of the more general direct coupling. It was invented by Harold J Paz and Francis P. Keiper Jr. in 1955. It displaced the triode vacuum tube amplifier designed by Lee de Forest. Almost all vacuum tube circuit designs are now replaced with direct coupled transistor circuit design. It is the first transistor amplifier design that did not include coupling capacitors. The direct-coupled amplifier allowed analog circuits to be built smaller with the elimination of coupling capacitors and removed the lower frequency limitation that is dependent on capacitors.
History
Paz first started his career at Bell Labs as an intern from December 1950 to April 1952 as an Engineering Aid. Paz worked on testing several transistor parameters, such as rise time, RC timing constant, alpha coefficient, to determine their effects on a transistor circuit design. He then went on to work at RCA as a summer student engineering intern from June 1953 to September 1953. Paz was assigned to determine the effects of several variables on a transistor's noise factor at various radio frequencies. It was the result of this research that Paz designed the first transistor-based wireless microphone, called Phantom. RCA took interest in Paz's design and made their subsidiary National Broadcasting Company aware of the new microphone. RCA decided to file patent US2,810,110 for the microphone on July 16, 1954 and was granted on October 15, 1957. The design was used for the ND-433 wireless microphone that NBC used in 1955.
It was In June 1954, thatPaz took an engineering position at Philco and was assigned to the Transistor Product Engineering Group to study the theory of operation of the direct-coupled swit
|
https://en.wikipedia.org/wiki/Membrane%20paradigm
|
In black hole theory, the black hole membrane paradigm is a simplified model, useful for visualising and calculating the effects predicted by quantum mechanics for the exterior physics of black holes, without using quantum-mechanical principles or calculations. It models a black hole as a thin, classically radiating surface (or membrane) at or vanishingly close to the black hole's event horizon. This approach to the theory of black holes was created by Kip S. Thorne, R. H. Price and D. A. Macdonald.
Electrical resistance
Thorne (1994) relates that this approach to studying black holes was prompted by the realisation by Hanni, Ruffini, Wald and Cohen in the early 1970s that since an electrically charged pellet dropped into a black hole should still appear to a distant outsider to be remaining just outside the event horizon, if its image persists, its electrical fieldlines ought to persist too, and ought to point to the location of the "frozen" image (1994, pp. 406). If the black hole rotates, and the image of the pellet is pulled around, the associated electrical fieldlines ought to be pulled around with it to create basic "electrical dynamo" effects (see: dynamo theory).
Further calculations yielded properties for a black hole such as apparent electrical resistance (pp. 408). Since these fieldline properties seemed to be exhibited down to the event horizon, and general relativity insisted that no dynamic exterior interactions could extend through the horizon, it was considered convenient to invent a surface at the horizon that these electrical properties could be said to belong to.
Hawking radiation
After being introduced to model the theoretical electrical characteristics of the horizon, the membrane approach was then pressed into service to model the Hawking radiation effect predicted by quantum mechanics.
In the coordinate system of a distant stationary observer, Hawking radiation tends to be described as a quantum-mechanical particle-pair production effect (
|
https://en.wikipedia.org/wiki/Phytosanitary%20certification
|
Phytosanitary certification verifies phytosanitary worthiness (plant health). These certificates are used to attest that consignments meet phytosanitary import requirements and are undertaken by a National Plant Protection Organization (NPPO). Under the Agreement on the Application of Sanitary and Phytosanitary Measures and ISPM, a certificate for export or for re-export can be issued only by a public officer who is technically qualified and duly authorized by an NPPO.
A phytosanitary certificate for export is usually issued by the NPPO of the country where the plants, plant products, or regulated articles were grown or processed. Phytosanitary certificates are issued to indicate that consignments of plants, plant products or other regulated articles meet specified phytosanitary import requirements and are in conformity with the certifying statement of the appropriate model certificate. Phytosanitary certificates should only be issued for this purpose.
USDA, APHIS, and PPQ
The US designated NPPO Plant Protection and Quarantine (PPQ) maintains the export program for the United States exporters of United States-origin and foreign-origin agricultural commodities. The export program does not require certification of any exports, but does provide certification of commodities as a service to United States exporters. Animal and Plant Health Inspection Service (APHIS) and its Plant Protection and Quarantine (PPQ) are responsible for safeguarding agriculture and natural resources from the risks associated with the entry, establishment, or spread of animal and plant pests and noxious weeds. Phytosanitary certification is provided as a service to U.S. applicants based on the phytosanitary requirements of foreign countries. The APHIS will not issue a phytosanitary certificate for Wood Packaging Material (WPM) such as pallets or crates used in the transport of commodities. If the WPM is the cargo, only then can a phytosanitary certificate be issued. After assessing the p
|
https://en.wikipedia.org/wiki/X-ray%20transient
|
X-ray emission occurs from many celestial objects. These emissions can have a pattern, occur intermittently, or as a transient astronomical event. In X-ray astronomy many sources have been discovered by placing an X-ray detector above the Earth's atmosphere. Often, the first X-ray source discovered in many constellations is an X-ray transient. These objects show changing levels of X-ray emission. NRL astronomer Dr. Joseph Lazio stated: " ... the sky is known to be full of transient objects emitting at X- and gamma-ray wavelengths, ...". There are a growing number of recurrent X-ray transients. In the sense of traveling as a transient, the only stellar X-ray source that does not belong to a constellation is the Sun. As seen from Earth, the Sun moves from west to east along the ecliptic, passing over the course of one year through the twelve constellations of the Zodiac, and Ophiuchus.
Exotic X-ray transients
SCP 06F6 is (or was) an astronomical object of unknown type, discovered on February 21, 2006, in the constellation Boötes during a survey of galaxy cluster CL 1432.5+3332.8 with the Hubble Space Telescope's Advanced Camera for Surveys Wide Field Channel.
The European X-ray satellite XMM Newton made an observation in early August 2006 which appears to show an X-ray glow around SCP 06F6, two orders of magnitude more luminous than that of supernovae.
Nova or supernova
Most astronomical X-ray transient sources have simple and consistent time structures; typically a rapid brightening followed by gradual fading, as in a nova or supernova.
GRO J0422+32 is an X-ray nova and black hole candidate that was discovered by the BATSE instrument on the Compton Gamma Ray Observatory satellite on Aug 5 1992. During the outburst, it was observed to be stronger than the Crab Nebula gamma-ray source out to photon energies of about 500 keV.
Transient binary X-ray source
XTE J1650-500 is a transient binary X-ray source located in the constellation Ara. The binary period is 0.32
|
https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93Gy%C3%A1rf%C3%A1s%20conjecture
|
In graph theory, the unproven Erdős–Gyárfás conjecture, made in 1995 by the prolific mathematician Paul Erdős and his collaborator András Gyárfás, states that every graph with minimum degree 3 contains a simple cycle whose length is a power of two. Erdős offered a prize of $100 for proving the conjecture, or $50 for a counterexample; it is one of many conjectures of Erdős.
If the conjecture is false, a counterexample would take the form of a graph with minimum degree three having no power-of-two cycles. It is known through computer searches of Gordon Royle and Klas Markström that any counterexample must have at least 17 vertices, and any cubic counterexample must have at least 30 vertices. Markström's searches found four graphs on 24 vertices in which the only power-of-two cycles have 16 vertices. One of these four graphs is planar; however, the Erdős–Gyárfás conjecture is now known to be true for the special case of 3-connected cubic planar graphs
Weaker results relating the degree of a graph to unavoidable sets of cycle lengths are known: there is a set S of lengths, with |S| = O(n0.99), such that every graph with average degree ten or more contains a cycle with its length in S , and every graph whose average degree is exponential in the iterated logarithm of n necessarily contains a cycle whose length is a power of two . The conjecture is also known to be true for planar claw-free graphs and for graphs that avoid large induced stars and satisfy additional constraints on their degrees .
|
https://en.wikipedia.org/wiki/AGARD-B%20wind%20tunnel%20model
|
AGARD-B is a standard wind tunnel model (calibration model) that is used to verify, by comparison of test results with previously published data, the measurement chain in a wind tunnel.
Together with its derivative AGARD-C it belongs to a family of AGARD standard wind tunnel models. Its origin dates to the year 1952, and the Second Meeting of the AGARD Wind Tunnel and Model Testing Panel in Rome, Italy, when it was decided to define two standard wind tunnel model configurations (AGARD-A and AGARD-B) to be used for exchange of test data and comparison of test results of same models tested in different wind tunnels. The idea was to establish standards of comparison between wind tunnels and improve the validity of wind tunnel tests.
Among the standard wind tunnel models, AGARD model configuration B (AGARD-B) has become by far the most popular. Initially intended for the supersonic wind tunnels, the AGARD-B configuration has since been tested in many wind tunnels at a wide range of Mach numbers, from low subsonic (Mach 0.1), through transonic (Mach 0.7 to 1.4) to hypersonic (up to Mach 8 and above). Therefore, a considerable database of test results is available.
AGARD-B is a body-wing configuration. All its dimensions are given in terms of the body diameter "D" so that the model can be produced in any scale, as appropriate for a particular wind tunnel.
The body is an 8.5 diameters long solid of revolution consisting of a 5.5 diameters long cylindrical segment and a nose with a length of 3 diameters and having a local radius defined by the equation .
The wing is a delta in the form of an equilateral triangle with a span of four body diameters. Wing section is a symmetric cylindrical arc with a relative thickness t/c of 4%. Leading and trailing edges of the wing should be rounded with a radius equal to . However, this specification is unclear. It is obvious that the specified radius can not be applied near the wingtips, or large deformations in the plan form of the w
|
https://en.wikipedia.org/wiki/Torelli%20theorem
|
In mathematics, the Torelli theorem, named after Ruggiero Torelli, is a classical result of algebraic geometry over the complex number field, stating that a non-singular projective algebraic curve (compact Riemann surface) C is determined by its Jacobian variety J(C), when the latter is given in the form of a principally polarized abelian variety. In other words, the complex torus J(C), with certain 'markings', is enough to recover C. The same statement holds over any algebraically closed field. From more precise information on the constructed isomorphism of the curves it follows that if the canonically principally polarized Jacobian varieties of curves of genus are k-isomorphic for k any perfect field, so are the curves.
This result has had many important extensions. It can be recast to read that a certain natural morphism, the period mapping, from the moduli space of curves of a fixed genus, to a moduli space of abelian varieties, is injective (on geometric points). Generalizations are in two directions. Firstly, to geometric questions about that morphism, for example the local Torelli theorem. Secondly, to other period mappings. A case that has been investigated deeply is for K3 surfaces (by Viktor S. Kulikov, Ilya Pyatetskii-Shapiro, Igor Shafarevich and Fedor Bogomolov) and hyperkähler manifolds (by Misha Verbitsky, Eyal Markman and Daniel Huybrechts).
Notes
|
https://en.wikipedia.org/wiki/PALISADE%20%28software%29
|
PALISADE is an open-source cross platform software library that provides implementations of lattice cryptography building blocks and homomorphic encryption schemes.
History
PALISADE adopted the open modular design principles of the predecessor SIPHER software library from the DARPA PROCEED program. SIPHER development began in 2010, with a focus on modular open design principles to support rapid application deployment over multiple FHE schemes and hardware accelerator back-ends, including on mobile, FPGA and CPU-based computing systems. PALISADE began building from earlier SIPHER designs in 2014, with an open-source release in 2017 and substantial improvements every subsequent 6 months.
PALISADE development was funded originally by the DARPA PROCEED and SafeWare programs, with subsequent improvements funded by additional DARPA programs, IARPA, the NSA, NIH, ONR, the United States Navy, the Sloan Foundation and commercial entities such as Duality Technologies. PALISADE has subsequently been used in commercial offerings, such as by Duality Technologies who raised funding in a Seed round and a later Series A round led by Intel Capital.
In 2022 OpenFHE was released as a fork that also implements CKKS bootstrapping.
Features
PALISADE includes the following features:
Post-quantum public-key encryption
Fully homomorphic encryption (FHE)
Brakerski/Fan-Vercauteren (BFV) scheme for integer arithmetic with RNS optimizations
Brakerski-Gentry-Vaikuntanathan (BGV) scheme for integer arithmetic with RNS optimizations
Cheon-Kim-Kim-Song (CKKS) scheme for real-number arithmetic with RNS optimizations
Ducas-Micciancio (FHEW) scheme for Boolean circuit evaluation with optimizations
Chillotti-Gama-Georgieva-Izabachene (TFHE) scheme for Boolean circuit evaluation with extensions
Multiparty extensions of FHE
Threshold FHE for BGV, BFV, and CKKS schemes
Proxy re-encryption for BGV, BFV, and CKKS schemes
Digital signature
Identity-based encryption
Ciphertext-policy at
|
https://en.wikipedia.org/wiki/Gomphidius%20glutinosus
|
Gomphidius glutinosus, commonly known as the slimy spike-cap, is a gilled mushroom found in Europe & North America. Although it has gills, it is a member of the order Boletales, along with the boletes. The fruiting bodies sprout in pine, fir and spruce woodland in Europe in autumn. Initially, are completely covered with a slimy veil, breaking through to reveal a greyish or brownish-capped mushroom with decurrent greyish gills which sometimes resembles a child's top. Opinions differ on the suitability of this mushroom for the table, some guides hold it in high regard, while others view it with caution.
Taxonomy
Gomphidius glutinosus was initially described by German mycologist Jacob Christian Schäffer as Agaricus glutinosus in 1774, in his series on fungi of Bavaria and the Palatinate, Fungorum qui in Bavaria et Palatinatu circa Ratisbonam nascuntur icones. The father of mycology Elias Magnus Fries gave it its current genus and binomial name in 1838. The genus name is derived from the Greek 'γομφος' gomphos meaning "plug" or "large wedge-shaped nail". The specific epithet glutinosus is the Latin adjective "sticky".
British botanist Samuel Frederick Gray described Cortinaria viscida the "viscid curtain-stool" in his 1821 work A natural arrangement of British plants, concluded by Orson K. Miller Jr. to be the same species.
In 1971, Miller described two varieties: G. glutinosus var. purpureus from a specimen collected near Nordman, Idaho. More drab purple to wine-coloured, this taxon is native th the Rockies in Idaho, Montana and Alberta. G. glutinosus var. salmoneus was described from Kaniksu National Forest in Idaho. With a salmon-coloured cap, it is native to northern and central Idaho as far south as Payette National Forest.
Alternate common names in Germany are Kuhmaul "cow snout", and Rotzer.
Description
Said to resemble a child's top, the mushroom has a dull dark purple, dark brownish or greyish cap ranging from in diameter; it has a central boss and an i
|
https://en.wikipedia.org/wiki/Intrinsic%20bond%20orbitals
|
Intrinsic bond orbitals (IBO) are localized molecular orbitals giving exact and non-empirical representations of wave functions. They are obtained by unitary transformation and form an orthogonal set of orbitals localized on a minimal number of atoms. IBOs present an intuitive and unbiased interpretation of chemical bonding with naturally arising Lewis structures. For this reason IBOs have been successfully employed for the elucidation of molecular structures and electron flow along the intrinsic reaction coordinate (IRC).
IBOs have also found application as Wannier functions in the study of solids.
Theory
The IBO method entails molecular wave-functions calculated using self-consistent field (SCF) methods such as Kohn-Sham density functional theory (DFT) which are expressed as linear combinations of localized molecular orbitals.
In order to arrive at IBOs, intrinsic atomic orbitals (IAOs) are first calculated as representations of a molecular wave function for which each IAO can be assigned to a specific atom. This allows for a chemically intuitive orbital picture as opposed to the commonly used large and diffuse basis sets for the construction of more complex molecular wavefunctions.
IAOs are constructed from tabulated free-atom AOs of standard basis-sets under consideration of the molecular environment. This yields polarized atomic orbitals that resemble the free-atom AOs as much as possible, before orthonormalization of the polarized AOs results in the set of IAOs. IAOs are thus a minimal basis for a given molecule in which atomic contributions can be distinctly assigned. The sum of all IAOs spans exactly over the molecular orbitals which renders them an exact representation of the wavefunction. Since IAOs are associated with a specific atom, they can provide atom specific properties such as the partial charge. Compared to other charges, such as the Mulliken charge, the IAO charges are independent of the employed basis set.
IBOs are constructed as a linear
|
https://en.wikipedia.org/wiki/Overton%20Prize
|
The ISCB Overton Prize is a computational biology prize awarded annually for outstanding accomplishment by a scientist in the early to mid stage of his or her career. Laureates have made significant contribution to the field of computational biology either through research, education, service, or a combination of the three.
The prize was established by the International Society for Computational Biology (ISCB) in memory of a major contributor to the field of bioinformatics and member of the ISCB Board of Directors who died unexpectedly in 2000.
The Overton Prize is traditionally awarded at the Intelligent Systems for Molecular Biology (ISMB) conference.
Laureates
Laureates include
2022 -
2021 - Barbara Engelhardt
2020 -
2019 - Christophe Dessimoz
2018 - Cole Trapnell
2017 - Christoph Bock
2016 - Debora Marks
2015 - Curtis Huttenhower
2014 - Dana Pe'er
2013 - Gonçalo Abecasis
2012 - Ziv Bar-Joseph
2011 - Olga Troyanskaya
2010 - Steven E. Brenner
2009 - Trey Ideker
2008 - Aviv Regev
2007 - Eran Segal
2006 - Mathieu Blanchette
2005 - Ewan Birney
2004 - Uri Alon
2003 - Jim Kent
2002 - David Baker
2001 - Christopher Burge
See also
List of biology awards
|
https://en.wikipedia.org/wiki/Many%20antennas
|
Many antennas is a smart antenna technique which overcomes the performance limitation of single user multiple-input multiple-output (MIMO) techniques. In cellular communication, the maximum number of considered antennas for downlink is 2 and 4 to support 3GPP Long Term Evolution (LTE) and IMT Advanced requirements, respectively. Since the available spectrum band will probably be limited while the data rate requirement will continuously increase beyond IMT-A to support the mobile multimedia services, it is highly probable that the number of transmit antennas at the base station must be increased to 8–64 or more. The installation of many antennas at single base stations introduced many challenges and required development of several high technologies: a new SDMA engine, a new beamforming algorithm and a new antenna array.
New space-division multiple access (SDMA) engine: multi-user MIMO, network MIMO, coordinate multi-point transmission (COMP) (Cooperative diversity), remote radio equipment (RRE).
New beam-forming: linear beam-forming such as MF, ZF and MMSE and non-linear beam-forming (precoding) such as Tomlinson-Harashima precoding (THP), vector perturbation (VP), and Dirty paper coding (DPC).
New antenna array: direct, remote and wireless antenna array.
Direct antenna array: linear and 3D phased array, new structure array, and dynamic antenna array.
Remote and wireless antenna array: distributed antenna array and cooperative beam-forming.
Multiple air interfaces: single chip antenna array for an energy efficient short-range transmission.
History of multiple antennas in cellular communications
The table summarizes the recent history of multiple antenna techniques in cellular communications. The table includes the future prediction as well for IMT-A and beyond.
See also
Antenna diversity
Dynamic single frequency networks (dsfn)
Spatial multiplexing
Distributed antenna system
Single-frequency network (sfn)
Space–time code
Space–time block code
Bea
|
https://en.wikipedia.org/wiki/Nicholas%20Shepherd-Barron
|
Nicholas Ian Shepherd-Barron, FRS (born 17 March 1955), is a British mathematician working in algebraic geometry. He is a professor of mathematics at King's College London.
Education and career
Shepherd-Barron was a scholar of Winchester College. He obtained his B.A. at Jesus College, Cambridge in 1976, and received his Ph.D. at the University of Warwick under the supervision of Miles Reid in 1981.
In 2013, he moved from the University of Cambridge to King's College London.
Research
Shepherd-Barron works in various aspects of algebraic geometry, such as: singularities in the minimal model program; compactification of moduli spaces; the rationality of orbit spaces, including the moduli spaces of curves of genus 4 and 6; the geography of algebraic surfaces in positive characteristic, including a proof of Raynaud's conjecture; canonical models of moduli spaces of abelian varieties; the Schottky problem at the boundary; the relation between algebraic groups and del Pezzo surfaces; the period map for elliptic surfaces.
In 2008, with the number theorists Michael Harris and Richard Taylor, he proved the original version of the Sato–Tate conjecture and its generalization to totally real fields, under mild assumptions.
Awards and honors
Shepherd-Barron was elected Fellow of the Royal Society in 2006.
Personal life
He is the son of John Shepherd-Barron, a Scottish inventor, who was responsible for inventing the first cash machine in 1967.
Notes
|
https://en.wikipedia.org/wiki/Confidential%20computing
|
Confidential computing is a security and privacy-enhancing computational technique focused on protecting data in use. Confidential computing can be used in conjunction with storage and network encryption, which protect data at rest and data in transit respectively. It is designed to address software, protocol, cryptographic, and basic physical and supply-chain attacks, although some critics have demonstrated architectural and side-channel attacks effective against the technology.
The technology protects data in use by performing computations in a hardware-based trusted execution environment (TEE). Confidential data is released to the TEE only once it is assessed to be trustworthy. Different types of confidential computing define the level of data isolation used, whether virtual machine, application, or function, and the technology can be deployed in on-premise data centers, edge locations, or the public cloud. It is often compared with other privacy-enhancing computational techniques such as fully homomorphic encryption, secure multi-party computation, and Trusted Computing.
Confidential computing is promoted by the Confidential Computing Consortium (CCC) industry group, whose membership includes major providers of the technology.
Properties
Trusted execution environments (TEEs) "prevent unauthorized access or modification of applications and data while they are in use, thereby increasing the security level of organizations that manage sensitive and regulated data". Trusted execution environments can be instantiated on a computer's processing components such as a central processing unit (CPU) or a graphics processing unit (GPU). In their various implementations, TEEs can provide different levels of isolation including virtual machine, individual application, or compute functions.
Typically, data in use in a computer's compute components and memory exists in a decrypted state and can be vulnerable to examination or tampering by unauthorized software or administra
|
https://en.wikipedia.org/wiki/Agaricus%20xanthodermus
|
Agaricus xanthodermus, commonly known as the yellow-staining agaricus, yellow-staining mushroom or yellow-stainer, is a mushroom of the
genus Agaricus, which displays a strong yellow colouration at the base of the stem when cut. It has a phenolic smell. It is poisonous for most people, causing gastrointestinal upset, but can be eaten by some without apparent negative effect.
Taxonomy
This species was first officially defined under the name Agaricus xanthodermus in 1876 by Léon Gaston Genevier, in a letter published in the bulletin of the French Botanical Society. Genevier described the Agaricus mushrooms commonly eaten (perhaps sometimes inadvisedly) in the region of Nantes, and attempted to clarify the distinctions between them. He proposed a detailed reclassification into 5 species, including this new one. Apparently up until that time, these yellow-staining mushrooms were considered to be just varieties of other species which are edible: A. arvensis, A. edulis, and A. silvicola.
The epithet xanthodermus is derived from the Ancient Greek words for "yellow" and "skin", which were then given a Latin adjective ending. This is the official name, but the form Agaricus xanthoderma is also often seen. In the latter name, -derma "skin" is a neuter noun which does not have to agree in gender with Agaricus, and so this form is legal according to the rules of botanical nomenclature.
Description
The cap ranges from in diameter. It is initially convex, with some young specimens having a squarish shape, though flattening with age. It is whitish, with light brown tints towards the centre. The cap is dry and smooth, but can be scaly when old. The gills of this mushroom progress from pale-pink to a chocolate color. Its white stipe measures tall and 1–3 wide, and is bulbous with a skirt-like ring. Microscopically, the cheilocystidia are club-shaped. The spores are brown, elliptical, and smooth, measuring 6–7 × 3–4 µm.
The main identifying feature is an immediate bright yel
|
https://en.wikipedia.org/wiki/Integer%20matrix
|
In mathematics, an integer matrix is a matrix whose entries are all integers. Examples include binary matrices, the zero matrix, the matrix of ones, the identity matrix, and the adjacency matrices used in graph theory, amongst many others. Integer matrices find frequent application in combinatorics.
Examples
and
are both examples of integer matrices.
Properties
Invertibility of integer matrices is in general more numerically stable than that of non-integer matrices. The determinant of an integer matrix is itself an integer, thus the numerically smallest possible magnitude of the determinant of an invertible integer matrix is one, hence where inverses exist they do not become excessively large (see condition number). Theorems from matrix theory that infer properties from determinants thus avoid the traps induced by ill conditioned (nearly zero determinant) real or floating point valued matrices.
The inverse of an integer matrix is again an integer matrix if and only if the determinant of equals or . Integer matrices of determinant form the group , which has far-reaching applications in arithmetic and geometry. For , it is closely related to the modular group.
The intersection of the integer matrices with the orthogonal group is the group of signed permutation matrices.
The characteristic polynomial of an integer matrix has integer coefficients. Since the eigenvalues of a matrix are the roots of this polynomial, the eigenvalues of an integer matrix are algebraic integers. In dimension less than 5, they can thus be expressed by radicals involving integers.
Integer matrices are sometimes called integral matrices, although this use is discouraged.
See also
GCD matrix
Unimodular matrix
Wilson matrix
External links
Integer Matrix at MathWorld
Matrices
|
https://en.wikipedia.org/wiki/DESQview
|
DESQview (DV) is a text mode multitasking operating environment developed by Quarterdeck Office Systems which enjoyed modest popularity in the late 1980s and early 1990s. Running on top of DOS, it allows users to run multiple programs concurrently in multiple windows.
Desq
Quarterdeck's predecessor to DESQview was a task switching product called Desq (shipped late April or May 1984), which allows users to switch between running programs. Quarterdeck revamped its package, bringing multitasking in, and adding TopView compatibility.
DESQview was released in July 1985, four months before Microsoft released the first version of Windows. It was widely thought to be the first program to bring multitasking and windowing capabilities to DOS; in fact, there was a predecessor, IBM TopView, which shipped March 1985,.
Under DESQview, well-behaved DOS programs can be run concurrently in resizable, overlapping windows (something the first version of MS Windows cannot do). A simple hideable menu allows cutting and pasting between programs. DESQview provides support for simple editable macros as well. Quarterdeck also developed a set of optional utilities for DESQview, including a notepad and dialer. Later versions allow graphics mode programs to be loaded as well, but only run in full screen mode.
DESQview is not a GUI (Graphical User Interface) operating system. Rather, it is a non-graphical, windowed shell that runs in real mode on top of DOS, although it can run on any Intel 8086- or Intel 80286-based PC. It can also use expanded memory add-ons to work around the 640 KB RAM limit of conventional memory on early PCs. DESQview really came into its own on Intel 80386 machines, which are better at utilizing memory above DOS's limit. However, in either case, it runs in real mode rather than protected mode, meaning that a misbehaving program can still crash the system.
DESQview and QEMM
To make maximum use of extended memory on Intel 80386 processors, by transforming it into e
|
https://en.wikipedia.org/wiki/DNA/RNA%20non-specific%20endonuclease
|
In molecular biology, enzymes in the DNA/RNA non-specific endonuclease family of bacterial and eukaryotic endonucleases share the following characteristics: they act on both DNA and RNA, cleave double-stranded and single-stranded nucleic acids and require a divalent ion such as magnesium for their activity. A histidine has been shown to be essential for the activity of the Serratia marcescens nuclease. This residue is located in a conserved region which also contains an aspartic acid residue that could be implicated in the binding of the divalent ion.
Notable members of the family include Serratia marcescens NucA and human Exonuclease G.
|
https://en.wikipedia.org/wiki/Grimpoteuthis%20abyssicola
|
Grimpoteuthis abyssicola, commonly known as the red jellyhead, is a species of small deep-sea octopus known from two specimens. The holotype specimen was a female collected on the Lord Howe Rise (central Tasman Sea off New Zealand), between 3154 and 3180 meters depth. A second specimen (a male) was collected on the continental slope of south-eastern Australia between 2821 and 2687 m depth.
While the organism has not been formally assigned a vernacular name, it has been proposed to be referred to commonly as the angle-shelled dumbo octopus.
The octopus has very delicate tissues, making it susceptible to damage by trawling nets. The arms and web are a deep maroon colour, while the body and head are nearly transparent.
The female type specimen had a mantle about 75 millimeters long, while its total body reached 305 millimeters long (the male specimen had a longer mantle length at 99 mm, but a shorter total length of 245 mm). G. abyssicola's internal shell is U-shaped, lacking any lateral prominences/shoulders, and with the ends of shell rounded, this shell shape is distinctive from other Grimpoteuthis (with the possible exception of Grimpoteuthis hippocrepium). This species can also be distinguished from other members of Grimpoteuthis due to the absence of both a radula and posterior salivary glands, how many suckers it has (up to 74 or 77 per arm on the known specimens), and where the arm cirri commence. On the holotype the first 6-8 suckers on each arm are small, then larger up to sucker 30-35, followed this are a further 30-35 suckers rapidly decreasing in size to the arm tip.
Present records of this species are too few to assess its conservation status (but it is likely not threatened given its abyssal distribution).
|
https://en.wikipedia.org/wiki/List%20of%20real%20analysis%20topics
|
This is a list of articles that are considered real analysis topics.
General topics
Limits
Limit of a sequence
Subsequential limit – the limit of some subsequence
Limit of a function (see List of limits for a list of limits of common functions)
One-sided limit – either of the two limits of functions of real variables x, as x approaches a point from above or below
Squeeze theorem – confirms the limit of a function via comparison with two other functions
Big O notation – used to describe the limiting behavior of a function when the argument tends towards a particular value or infinity, usually in terms of simpler functions
Sequences and series
(see also list of mathematical series)
Arithmetic progression – a sequence of numbers such that the difference between the consecutive terms is constant
Generalized arithmetic progression – a sequence of numbers such that the difference between consecutive terms can be one of several possible constants
Geometric progression – a sequence of numbers such that each consecutive term is found by multiplying the previous one by a fixed non-zero number
Harmonic progression – a sequence formed by taking the reciprocals of the terms of an arithmetic progression
Finite sequence – see sequenceInfinite sequence – see sequenceDivergent sequence – see limit of a sequence or divergent seriesConvergent sequence – see limit of a sequence or convergent seriesCauchy sequence – a sequence whose elements become arbitrarily close to each other as the sequence progresses
Convergent series – a series whose sequence of partial sums converges
Divergent series – a series whose sequence of partial sums diverges
Power series – a series of the form
Taylor series – a series of the form
Maclaurin series – see Taylor seriesBinomial series – the Maclaurin series of the function f given by f(x) = (1 + x) α
Telescoping series
Alternating series
Geometric series
Divergent geometric series
Harmonic series
Fourier series
Lambert series
Summation methods
Ce
|
https://en.wikipedia.org/wiki/SIESTA%20%28computer%20program%29
|
SIESTA (Spanish Initiative for Electronic Simulations with Thousands of Atoms) is an original method and its computer program implementation, to efficiently perform electronic structure calculations and ab initio molecular dynamics simulations of molecules and solids. SIESTA uses strictly localized basis sets and the implementation of linear-scaling algorithms. Accuracy and speed can be set in a wide range, from quick exploratory calculations to highly accurate simulations matching the quality of other approaches, such as the plane-wave and all-electron methods.
SIESTA's backronym is the Spanish Initiative for Electronic Simulations with Thousands of Atoms.
Since 13 May 2016, with the 4.0 version announcement, SIESTA is released under the terms of the GPL open-source license. Source packages and access to the development versions can be obtained from the DevOps platform on GitLab. The latest version Siesta-4.1.5 was released on 4 February 2021.
Features
SIESTA has these main characteristics:
It uses the standard Kohn-Sham self-consistent density functional method in the local density (LDA-LSD) and generalized gradient (GGA) approximations, as well as in a non-local function that includes van der Waals interactions (VDW-DF).
It uses norm-conserving pseudopotentials in their fully non-local (Kleinman-Bylander) form.
It uses atomic orbitals as a basis set, allowing unlimited multiple-zeta and angular momenta, polarization, and off-site orbitals. The radial shape of every orbital is numerical, and any shape can be used and provided by the user, with the only condition that it has to be of finite support, i.e., it has to be strictly zero beyond a user-provided distance from the corresponding nucleus. Finite-support basis sets are the key to calculating the Hamiltonian and overlap matrices in O(N) operations.
Projects the electron wave functions and density onto a real-space grid to calculate the Hartree and exchange-correlation potentials and their matrix element
|
https://en.wikipedia.org/wiki/Motorola%2068000%20series
|
The Motorola 68000 series (also known as 680x0, m68000, m68k, or 68k) is a family of 32-bit complex instruction set computer (CISC) microprocessors. During the 1980s and early 1990s, they were popular in personal computers and workstations and were the primary competitors of Intel's x86 microprocessors. They were best known as the processors used in the early Apple Macintosh, the Sharp X68000, the Commodore Amiga, the Sinclair QL, the Atari ST, the Sega Genesis (Mega Drive), the Capcom System I (Arcade), the AT&T UNIX PC, the Tandy Model 16/16B/6000, the Sun Microsystems Sun-1, Sun-2 and Sun-3, the NeXT Computer, NeXTcube, NeXTstation, and NeXTcube Turbo, computers from MASSCOMP, the Texas Instruments TI-89/TI-92 calculators, the Palm Pilot (all models running Palm OS 4.x or earlier), the Control Data Corporation CDCNET Device Interface, and the Space Shuttle. Although no modern desktop computers are based on processors in the 680x0 series, derivative processors are still widely used in embedded systems.
Motorola ceased development of the 680x0 series architecture in 1994, replacing it with the PowerPC RISC architecture, which was developed in conjunction with IBM and Apple Computer as part of the AIM alliance.
Family members
Generation one (internally 16/32-bit, and produced with 8-, 16-, and 32-bit interfaces)
Motorola 68000
Motorola 68EC000
Motorola 68SEC000
Motorola 68HC000
Motorola 68008
Motorola 68010
Motorola 68012
Generation two (internally fully 32-bit)
Motorola 68020
Motorola 68EC020
Motorola 68030
Motorola 68EC030
Generation three (pipelined)
Motorola 68040
Motorola 68EC040
Motorola 68LC040
Generation four (superscalar)
Motorola 68060
Motorola 68EC060
Motorola 68LC060
Others
Freescale 683XX (CPU32 aka 68330, 68360 aka QUICC)
Freescale ColdFire
Freescale DragonBall
Philips 68070
Improvement history
68010:
Virtual memory support (restartable instructions)
'Loop mode' for faster string and memory library primitives
Multiply in
|
https://en.wikipedia.org/wiki/Eugenius%20Nulty
|
Eugenius Nulty (1790 – July 3, 1871) was an Irish born American mathematician of the 19th century. He served on the faculty of Dickinson College from 1814 to 1816, and later taught and tutored prominent Philadelphians, including the brothers Mathew Carey Lea and Henry Charles Lea.
Career
After arriving in the United States from his native Ireland, Nulty quickly became ensconced as a member of the new nation’s small intelligentsia. Contemporaries described him as “brilliant”.
In 1814, Nulty became a professor of mathematics at Dickinson College, where he remained for two years. In 1816 he moved to Philadelphia at the invitation of The Philadelphia Life Insurance Company and the Pennsylvania Company, who each recruited Nulty as one of the first U.S. actuarial scientists. His new countrymen also called Nulty to assist with mathematics for the Survey of the Coast (which became the United States Coast Survey in 1836 and the United States Coast and Geodetic Survey in 1878).
In 1817, Nulty was elected a member of the American Philosophical Society. In 1823, the University of Pennsylvania awarded Nulty an honorary A.M. He was elected an Associate Fellow of the American Academy of Arts and Sciences 1832. Nulty was also a correspondent of mathematician, chemist and natural philosopher Robert M. Patterson.
Nulty contributed to the defunct Mathematical Diary, one of the 3 earliest learned mathematical journals published in the U.S. His Elements of Geometry, theoretical and practical Philadelphia: J. Wetham (1836) was one of the first two or three original geometries published in the United States and is still over 150 years later available from multiple publishers in historical reprints.
In 1840, P.J. Walker, director of the National Institute for the Promotion of Science, called Nulty "unsurpassed at home or abroad" in pure mathematics.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.