source
stringlengths 31
227
| text
stringlengths 9
2k
|
|---|---|
https://en.wikipedia.org/wiki/Extension%20by%20new%20constant%20and%20function%20names
|
In mathematical logic, a theory can be extended with
new constants or function names under certain conditions with assurance that the extension will introduce
no contradiction. Extension by definitions is perhaps the best-known approach, but it requires
unique existence of an object with the desired property. Addition of new names can also be done
safely without uniqueness.
Suppose that a closed formula
is a theorem of a first-order theory . Let be a theory obtained from by extending its language with new constants
and adding a new axiom
.
Then is a conservative extension of , which means that the theory has the same set of theorems in the original language (i.e., without constants ) as the theory .
Such a theory can also be conservatively extended by introducing a new functional symbol:
Suppose that a closed formula is a theorem of a first-order theory , where we denote . Let be a theory obtained from by extending its language with a new functional symbol (of arity ) and adding a new axiom . Then is a conservative extension of , i.e. the theories and prove the same theorems not involving the functional symbol ).
Shoenfield states the theorem in the form for a new function name, and constants are the same as functions
of zero arguments. In formal systems that admit ordered tuples, extension by multiple constants as shown here
can be accomplished by addition of a new constant tuple and the new constant names
having the values of elements of the tuple.
See also
Conservative extension
Extension by definition
|
https://en.wikipedia.org/wiki/Cry34Ab1
|
Cry34Ab1 is one member of a binary Bacillus thuringiensis (Bt) crystal protein set isolated from Bt strain PS149B1. The protein exists as a 14 kDa aegerolysin that, in presence of Cry35Ab1, exhibits insecticidal activity towards Western Corn Rootworm. The protein has been transformed into maize plants under the commercialized events 4114 (DP-ØØ4114-3) by Pioneer Hi-Bred and 59122 (DAS-59122-7) by Dow AgroSciences. These events have, in turn, been bred into multiple trait stacks in additional products.
Cry34/35Ab1 binary toxins bind to the insect's brush border membrane vesicles (BBMVs) of cells in the epithelial lining of midgut, where they form pores; this leads to necrosis and, eventually, the insect's death. The Cry35Ab1 (45 kDa) protein does not convey specificity in the absence of Cry34Ab1, indicating that the smaller 14 kDa Cry34Ab1 protein is critical for BBMV binding and recruitment of Cry35Ab1 to induce insecticidal effect.
Cry34Ab1 is unrelated to Bt delta-endotoxins. It is an aegerolysin () composed of two beta sheets in a beta-sandwich structure; the total protein is composed of 117 amino acid residues and contains a hydrophobic core. Its family placement shows that interaction with cell membranes is consistent with its role in the binary Cry34Ab1/Cry35Ab1 toxin complex. The partner toxin, Cry35Ab1, is a prototypical member of its own group (). Its structure is similar to that of aerolysin, Cry45Aa1, and Cry46Aa1.
|
https://en.wikipedia.org/wiki/Forest%20restoration
|
Forest restoration is defined as “actions to re-instate ecological processes, which accelerate recovery of forest structure, ecological functioning and biodiversity levels towards those typical of climax forest” i.e. the end-stage of natural forest succession. Climax forests are relatively stable ecosystems that have developed the maximum biomass, structural complexity and species diversity that are possible within the limits imposed by climate and soil and without continued disturbance from humans (more explanation here). Climax forest is therefore the target ecosystem, which defines the ultimate aim of forest restoration. Since climate is a major factor that determines climax forest composition, global climate change may result in changing restoration aims. Additionally, the potential impacts of climate change on restoration goals must be taken into account, as changes in temperature and precipitation patterns may alter the composition and distribution of climax forests.
Forest restoration is a specialized form of reforestation, but it differs from conventional tree plantations in that its primary goals are biodiversity recovery and environmental protection.
Forest and landscape restoration (FLR) is defined as a process that aims to regain ecological functionality and enhance human well-being in deforested or degraded landscapes. FLR has been developed as a response to the growing degradation and loss of forest and land, which resulted in declined biodiversity and ecosystem services. Effective FLR will support the achievement of the Sustainable Development Goals. The United Nations Decade on Ecosystem Restoration (2021–2030) provides the opportunity to restore hundreds of millions of hectares of degraded forests and other ecosystems. Successful ecosystem restoration requires a fundamental understanding of the ecological characteristics of the component species, together with knowledge of how they assemble, interact and function as communities
Scope
Forest rest
|
https://en.wikipedia.org/wiki/Trident%20laser
|
The Trident Laser was a high power, sub-petawatt class, solid-state laser facility located at Los Alamos National Laboratory (LANL website), in Los Alamos, New Mexico, originally built in the late 1980s for Inertial confinement fusion (ICF) research by KMS Fusion, founded by Kip Siegel, in Ann Arbor, Michigan, it was later moved to Los Alamos in the early 1990s to be used in ICF and materials research. The Trident Laser has been decommissioned, with final experiments in 2017, and is now in storage at the University of Texas at Austin.
The Trident Laser consisted of three main laser chains (A,B, and C) of neodymium glass amplifiers (or Nd:glass), two identical longpulse beams lines, A&B, and a third beamline, C, that could be operated either in longpulse or in chirped pulse amplification (CPA) shortpulse mode.
Longpulse beams A and B, were laser chains capable of delivering up to ~500 J at 1054 nm, which were frequency doubled to 527 nm and ~200 J depending on pulse duration; the pulse duration could be varied from 100 ps to 1 μs, and was a unique capability of any large laser in the US (and possibly the world). The third laser chain, beamline C, could produce up to ~200 J at 1054 nm, or could be frequency doubled to 527 nm at ~100 J in the longpulse mode with the same pulse duration variability as beams A and B; or could be used in the Trident enhancement configuration allowing the ~200 J beam to be compressed via CPA to ~600 fs and ~100 J, producing powers on the scale of a quarter petawatt(~200 TW) with a host of laser and plasma diagnostics. A 100 mJ 500 fs probe beamline is also available.
The 200TW shortpulse ultra high-intensity laser system is currently a world record holder in ion acceleration energy with Target Normal Sheath Acceleration mechanism, producing protons at 58.5 MeV from a flat-foil, beating the record of the NOVA Petawatt laser back in 1999; and 67.5 MeV protons from micro-cone targets. Trident delivers Petawatt performance at a fifth of
|
https://en.wikipedia.org/wiki/Copper%20naphthenate
|
Copper naphthenate is the copper salt of naphthenic acid. Naphthenic acid is a term commonly used in the petroleum industry to collectively refer to all of the carboxylic acids naturally occurring in crude oil. Naphthenic acids are primarily cycloaliphatic carboxylic acids with 10 to 24 or more carbons, although substantial quantities of non-cyclic, aromatic and heteroatom- containing carboxylic acids are also present. Copper naphthenate is most widely used in wood preservation and for protecting other cellulosic materials such as textiles and cordage from damage by decay fungi and insects. Other metal naphthenates are used as paint driers, rubber adhesion promoters, lubricant additives, and catalysts where oil solubility is required.
Composition
Naphthenic acid is a complex group of carboxylic acids with the general formula CnH2n-zO2 where n indicates the carbon number and z specifies the hydrogen deficiency resulting from ring formation. Naphthenic acids occur naturally in crude oil at varying concentrations, and their composition varies with the crude oil composition and the conditions during refining. Crude oils with total acid number (TAN) as little as 0.5 mg KOH/g acid or petroleum fractions greater than about 1.0 mg KOH/g oil usually qualify as a high acid crude or oil. At the 1.0 mg/g TAN level, acidic crude oils begin to be heavily discounted in value and so are referred to as opportunity crudes. Commercial grades of naphthenic acid are most often recovered from kerosene/jet fuel and diesel fractions, where their corrosivity and negative impact on burning qualities require their removal. Naphthenic acids are also a major contaminant in water produced during the extraction of oil from Athabasca oil sands.
Metal naphthenates
Metal salts of naphthenic acids, which are called naphthenates, are widely used as hydrophobic sources of metal ions in diverse applications. They have the formula M(naphthenate)2, or M3O(naphthenate)6 for basic oxides. Metal na
|
https://en.wikipedia.org/wiki/Ravi%20Vakil
|
Ravi D. Vakil (born February 22, 1970) is a Canadian-American mathematician working in algebraic geometry.
Education and career
Vakil attended high school at Martingrove Collegiate Institute in Etobicoke, Ontario, where he won several mathematical contests and olympiads. After earning a BSc and MSc from the University of Toronto in 1992, he completed a PhD in mathematics at Harvard University in 1997 under Joe Harris. He has since been an instructor at both Princeton University and MIT. Since the fall of 2001, he has taught at Stanford University, becoming a full professor in 2007.
Contributions
Vakil is an algebraic geometer and his research work spans over enumerative geometry, topology, Gromov–Witten theory, and classical algebraic geometry. He has solved several old problems in Schubert calculus. Among other results, he proved that all Schubert problems are enumerative over the real numbers, a result that resolves an issue mathematicians have worked on for at least two decades.
Awards and honors
Vakil has received many awards, including an NSF CAREER Fellowship, a Sloan Research Fellowship, an American Mathematical Society Centennial Fellowship, a G. de B. Robinson prize for the best paper published (2000) in the Canadian Journal of Mathematics and the Canadian Mathematical Bulletin, and the André-Aisenstadt Prize from the Centre de Recherches Mathématiques at the Université de Montréal (2005), and the Chauvenet Prize (2014)..
In 2012 he became a fellow of the American Mathematical Society.
Mathematics contests
He was a member of the Canadian team in three International Mathematical Olympiads, winning silver, gold (perfect score), and gold in 1986, 1987, and 1988 respectively. He was also the fourth person to be a four-time Putnam Fellow in the history of the contest. Also, he has been the coordinator of weekly Putnam preparation seminars at Stanford.
|
https://en.wikipedia.org/wiki/Riemann%E2%80%93Hurwitz%20formula
|
In mathematics, the Riemann–Hurwitz formula, named after Bernhard Riemann and Adolf Hurwitz, describes the relationship of the Euler characteristics of two surfaces when one is a ramified covering of the other. It therefore connects ramification with algebraic topology, in this case. It is a prototype result for many others, and is often applied in the theory of Riemann surfaces (which is its origin) and algebraic curves.
Statement
For a compact, connected, orientable surface , the Euler characteristic is
,
where g is the genus (the number of handles), since the Betti numbers are . In the case of an (unramified) covering map of surfaces
that is surjective and of degree , we have the formula
That is because each simplex of should be covered by exactly in , at least if we use a fine enough triangulation of , as we are entitled to do since the Euler characteristic is a topological invariant. What the Riemann–Hurwitz formula does is to add in a correction to allow for ramification (sheets coming together).
Now assume that and are Riemann surfaces, and that the map is complex analytic. The map is said to be ramified at a point P in S′ if there exist analytic coordinates near P and π(P) such that π takes the form π(z) = zn, and n > 1. An equivalent way of thinking about this is that there exists a small neighborhood U of P such that π(P) has exactly one preimage in U, but the image of any other point in U has exactly n preimages in U. The number n is called the ramification index at P and also denoted by eP. In calculating the Euler characteristic of S′ we notice the loss of eP − 1 copies of P above π(P) (that is, in the inverse image of π(P)). Now let us choose triangulations of S and S′ with vertices at the branch and ramification points, respectively, and use these to compute the Euler characteristics. Then S′ will have the same number of d-dimensional faces for d different from zero, but fewer than expected vertices. Therefore, we find a "corrected"
|
https://en.wikipedia.org/wiki/Thurston%20elliptization%20conjecture
|
William Thurston's elliptization conjecture states that a closed 3-manifold with finite fundamental group is spherical, i.e. has a Riemannian metric of constant positive sectional curvature.
Relation to other conjectures
A 3-manifold with a Riemannian metric of constant positive sectional curvature is covered by the 3-sphere, moreover the group of covering transformations are isometries of the 3-sphere.
If the original 3-manifold had in fact a trivial fundamental group, then it is homeomorphic to the 3-sphere (via the covering map). Thus, proving the elliptization conjecture would prove the Poincaré conjecture as a corollary. In fact, the elliptization conjecture is logically equivalent to two simpler conjectures: the Poincaré conjecture and the spherical space form conjecture.
The elliptization conjecture is a special case of Thurston's geometrization conjecture, which was proved in 2003 by G. Perelman.
|
https://en.wikipedia.org/wiki/MAX%20%28operating%20system%29
|
MaX, also known as Madrid_linux is a linux distribution created with funds from the Conserjería de Educación, Juventud y Deporte of the Comunidad de Madrid adapted for use in educational environments. The main features of this operating system are simplicity, stability and a huge collection of software.
History
Start
MaX started out simply as an educational distribution made for lower performance computers, since most of the computers in the schools of the Comunidad de Madrid had only 32 bits and very little RAM, so they had problems with the main operating systems. Later, more members were joining the Max Group and were receiving suggestions for improvements, additions. In 2013 they already had 60 members, including professors of all kinds.
Desktop and Support
During the first versions of MaX it was used with the KDE desktop environment and only supported 32-bit computers of the i486 architecture, until MaX version 3, where the processor architecture was changed to i386, also 32-bit. In MaX version 4, the desktop environment was changed to GNOME, one of the most widely used in Linux, and in version 5 support for 64-bit computers with x86_64 architecture was added. In version 7, the desktop environment was changed to Xfce. In version 9 the desktop environment was changed again to Mate, and in version 11 support for 32-bit computers ended.
Activities performed
The MaX team has participated in events of schools, such as 'Program-me at IES Clara del Rey, with the purpose of supporting new talents; or with IES Príncipe Felipe at MediaLab Prado, to encourage diffusion of MaX.
Conferences
MaX has held many conferencess over the years, almost annually, to present future versions of MaX, create workshops, or hold install parties. The conference was held online in 2021.
Characteristics
Community
The community of users actively participates, testing the system, reporting problems, contributing with improvement proposals, helping other users or sharing with their peers
|
https://en.wikipedia.org/wiki/The%20Activision%20Decathlon
|
The Activision Decathlon is a sports game written by David Crane for the Atari 2600 and published by Activision in 1983. It was ported to the Atari 8-bit family, Atari 5200, Commodore 64, ColecoVision, and MSX. Up to four players compete in the ten different events of a real-life decathlon, either in sequence or individually.
Gameplay
The events are:
100-Meter Dash
Long Jump
Shot Put
High Jump
400-Meter Race
110-Meter Hurdles
Discus Throw
Pole Vault
Javelin Throw
1500-Meter Race
Activision Decathlon Club patches
A player whose score met or exceeded the values below could send photo documentation to receive an Activision Decathlon Club patch in one of the colors of an Olympic medal:
8,600 points: Bronze
9,000 points: Silver
10,000 points: Gold
Reception
Shortly after release, Activision's Decathlon drew comparisons to Konami's popular arcade game Hyper Olympic (Track & Field), which was introduced at the Amusement Machine Show a month later in September 1983. According to Cash Box magazine, several people claimed there were "cursory similarities" between the two games.
The Atari 2600 version of Decathlon was reviewed by Video magazine in its "Arcade Alley" column where it was described as "an absolute triumph of imaginative programming" and as "a masterwork". Computer and Video Games rated the VCS version 92% while giving the ColecoVision version a 93% score.
In 1985, the game appeared at number-two on the Atari 8-bit chart in the United Kingdom.
Legacy
The game was later reissued simply as Decathlon by the UK budget label Firebird. It was included in the 2002 PlayStation 2 compilation Activision Anthology.
See also
List of Atari 2600 games
List of Activision games: 1980–1999
Olympic Decathlon, 1980 computer game with similar concept and controls
|
https://en.wikipedia.org/wiki/ADAM12
|
Disintegrin and metalloproteinase domain-containing protein 12 (previously Meltrin) is an enzyme that in humans is encoded by the ADAM12 gene. ADAM12 has two splice variants: ADAM12-L, the long form, has a transmembrane region and ADAM12-S, a shorter variant, is soluble and lacks the transmembrane and cytoplasmic domains.
Function
This gene encodes a member of the ADAM (a disintegrin and metalloprotease) protein family. Members of this family are membrane-anchored proteins structurally related to snake venom disintegrins, and have been implicated in a variety of biological processes involving cell-cell and cell-matrix interactions, including fertilization, muscle development, and neurogenesis. This gene has two alternatively spliced transcripts: a shorter secreted form and a longer membrane-bound form. The shorter form is found to stimulate myogenesis.
Clinical Significance
ADAM 12, a metalloprotease that binds insulin growth factor binding protein-3 (IGFBP-3), appears to be an effective early Down syndrome marker. Decreased levels of ADAM 12 may be detected in cases of trisomy 21 as early as 8 to 10 weeks gestation. Maternal serum ADAM 12 and PAPP-A levels at 8 to 9 weeks gestation in combination with maternal age yielded a 91% detection rate for Down syndrome at a 5% false-positive rate. When nuchal translucency data from approximately 12 weeks gestation was added, this increased the detection rate to 97%.
ADAM12 has also been implicated in the development of pathology in various cancers, hypertension, liver fibrogenesis, and asthma. In asthma, ADAM12 is upregulated in lung epithelium in response to TNF-alpha.
In a study of about 1200 persons with extremely high intelligence (IQ about 170), variants of the gene were associated with IQ compared with a general population.
Interactions
ADAM12 has been shown to interact with:
ACTN2,
IGFBP3, and
PIK3R1.
|
https://en.wikipedia.org/wiki/Friendship%20paradox
|
The friendship paradox is the phenomenon first observed by the sociologist Scott L. Feld in 1991 that on average, an individual's friends have more friends than that individual. It can be explained as a form of sampling bias in which people with more friends are more likely to be in one's own friend group. In other words, one is less likely to be friends with someone who has very few friends. In contradiction to this, most people believe that they have more friends than their friends have.
The same observation can be applied more generally to social networks defined by other relations than friendship: for instance, most people's sexual partners have had (on the average) a greater number of sexual partners than they have.
The friendship paradox is an example of how network structure can significantly distort an individual's local observations.
Mathematical explanation
In spite of its apparently paradoxical nature, the phenomenon is real, and can be explained as a consequence of the general mathematical properties of social networks. The mathematics behind this are directly related to the arithmetic-geometric mean inequality and the Cauchy–Schwarz inequality.
Formally, Feld assumes that a social network is represented by an undirected graph , where the set of vertices corresponds to the people in the social network, and the set of edges corresponds to the friendship relation between pairs of people. That is, he assumes that friendship is a symmetric relation: if is a friend of , then is a friend of . The friendship between and is therefore modeled by the edge and the number of friends an individual has corresponds to a vertex's degree. The average number of friends of a person in the social network is therefore given by the average of the degrees of the vertices in the graph. That is, if vertex has edges touching it (representing a person who has friends), then the average number of friends of a random person in the graph is
The average number of frie
|
https://en.wikipedia.org/wiki/Buster%20Capp
|
Buster Capp is a British comic strip series which debuted on 28 May 1960 in the magazine Buster and ran until January 2000. The character was the mascot of the magazine too. The series is a spin-off of Andy Capp, starring Andy's young son Buster, despite not being drawn by the original artist of that comic, Reg Smythe. The first artist was Bill Titcombe, but Hugh McNeill took over as artist after less than a year. After a few years Àngel Nadal took over and drew the strip until 1974, when he was followed by Reg Parlett. Tom Paterson in turn took over from Parlett in 1985 and drew the strip until 1990. Jimmy Hansen then became the strip's artist until the magazine folded in 2000; the magazine started using reprints of Hansen's earlier strips in 1998, but he still drew a complete new strip once a month, along with covers for every magazine until the end of its life.
Swedish version
There also existed a Swedish version of Buster Capp, done directly for the Swedish edition by local Swedish artists, this version was a heavily sports-orientated humour strip.
|
https://en.wikipedia.org/wiki/Tensor%20representation
|
In mathematics, the tensor representations of the general linear group are those that are obtained by taking finitely many tensor products of the fundamental representation and its dual. The irreducible factors of such a representation are also called tensor representations, and can be obtained by applying Schur functors (associated to Young tableaux). These coincide with the rational representations of the general linear group.
More generally, a matrix group is any subgroup of the general linear group. A tensor representation of a matrix group is any representation that is contained in a tensor representation of the general linear group. For example, the orthogonal group O(n) admits a tensor representation on the space of all trace-free symmetric tensors of order two. For orthogonal groups, the tensor representations are contrasted with the spin representations.
The classical groups, like the symplectic group, have the property that all finite-dimensional representations are tensor representations (by Weyl's construction), while other representations (like the metaplectic representation) exist in infinite dimensions.
|
https://en.wikipedia.org/wiki/India-based%20Neutrino%20Observatory
|
India-based Neutrino Observatory (INO) is a particle physics research project under construction to primarily study atmospheric neutrinos in a deep cave under INO Peak near Theni, Tamil Nadu, India. This project is notable in that it is anticipated to provide a precise measurement of neutrino mixing parameters. The project is a multi-institute collaboration and one of the biggest experimental particle physics projects undertaken in India.
The project was originally to be completed in 2015 at an estimated cost of 1,500 crores(15 billion or 209.7 million), has been cleared by the Ministry of Environment (India) for construction in the Bodi West Hills Reserved Forest in the Theni district of Tamil Nadu. Although delayed, the project was underway .
When completed, the main magnetised iron calorimeter (ICAL) experiment will include the world's largest magnet, four times larger than the 12,500-tonne magnet in the Compact Muon Solenoid detector at CERN in Geneva, Switzerland.
Iron Calorimeter (ICAL) Detector
The main experiment proposed at INO is the Iron-Calorimeter Detector which aims to probe the Earth matter effects on the propagation of atmospheric neutrinos and to determine neutrino oscillation parameters in the 2-3 oscillation sector. ICAL will be a 50000 tonne magnetised detector with iron as the passive detector element and resistive plate chambers (RPCs) as the active detector elements. i.e., the neutrinos will interact with the iron to produce final state particles. The RPCs will detect those final state particles which have charge and will record the signals and these signals which have position and timing information will help us reconstruct the tracks and/or showers and thus the energy and directions of the final state particles and also the incident neutrino.
The ICAL design is mostly based on the Monolith detector .
ICAL detector will have three modules, each module will have 151 layers of iron and 150 layers of RPCs, stacked one over the other. T
|
https://en.wikipedia.org/wiki/Pandiagonal%20magic%20cube
|
In recreational mathematics, a pandiagonal magic cube is a magic cube with the additional property that all broken diagonals (parallel to exactly two of the three coordinate axes) have the same sum as each other. Pandiagonal magic cubes are extensions of diagonal magic cubes (in which only the unbroken diagonals need to have the same sum as the rows of the cube) and generalize pandiagonal magic squares to three dimensions.
In a pandiagonal magic cube, all 3m planar arrays must be panmagic squares. The 6 oblique squares are always magic. Several of them may be panmagic squares.
A proper pandiagonal magic cube has exactly 9m2 lines plus the 4 main space diagonals summing correctly (no broken space diagonals have the correct sum.)
The smallest pandiagonal magic cube has order 7.
See also
Magic cube classes
|
https://en.wikipedia.org/wiki/ADE%20model
|
An ADE model is a genetic model for twin studies which includes dominance genetic effects.
A stands for additive genetic effects, D for non-additive genetic (or dominance) effects, and E for nonshared environment effects.
See also
ACE model
|
https://en.wikipedia.org/wiki/LimeSurvey
|
LimeSurvey (formerly PHPSurveyor) is a free and open source on-line statistical survey web app written in PHP based on a MySQL, SQLite, PostgreSQL or MSSQL database, distributed under the GNU General Public License. As a web server-based software it enables users using a web interface to develop and publish on-line surveys, collect responses, create statistics, and export the resulting data to other applications.
History
LimeSurvey was registered as a SourceForge.net project called PHPSurveyor on February 20, 2003 and was originally written by the Australian software developer Jason Cleeland. The first public release, version 0.93, was published on March 5, 2003. The project quickly developed a large audience of users after the development of advanced features such as branching (conditions), token control and templating.
In 2004, during the 2004 U.S. presidential election, PHPSurveyor was used to gather data about voting irregularities. It identified over 13500 incidents in the first 10 hours of voting and was selected as part of their Election Incident Reporting System.
Starting in early 2005, Carsten Schmitz, a German IT project manager, started taking on some of the lead developer responsibilities, with the full project being transferred to him in 2006. On May 17, 2007 the project name was changed from PHPSurveyor to LimeSurvey in order to make software licensing easier by not including PHP in the name.
In late 2008, a LimeSurvey hosting service named LimeService was created by LimeSurvey project leader Carsten Schmitz. It hosts LimeSurvey for users for a small fee per response.
As of June 4, 2008, LimeSurvey was ranked on SourceForge.net with an overall rank of 99 out of over 100,000 projects as of June 4, 2008. It has been downloaded more than 200,000 times and its development status is listed as "5 - Production/Stable, 6 - Mature".
In 2009, LimeSurvey participated in the Google Summer of Code, a program encouraging students older than 18 years old to wo
|
https://en.wikipedia.org/wiki/X-No-Archive
|
X-No-Archive, also known colloquially as xna, is a newsgroup message header field used to prevent a Usenet message from being archived in various servers.
Origin
The need for X-No-Archive began when DejaNews debuted in 1995. DejaNews was the first large-scale commercial attempt to archive the Usenet news feed, and several newsgroup participants were concerned about privacy rights and about the possibility that their messages could be re-posted through DejaNews in the future. DejaNews addressed these concerns by announcing that it would not archive Usenet messages containing the X-No-Archive header field.
How it works
X-No-Archive was designed to follow the standard message header protocol, RFC 1036 and RFC 977, used in existing newsgroups. In addition to the standard header fields used in all newsgroup messages (including Path:, From:, Subject:, and Date:), news reader software allows a user to add optional fields to a header. According to RFC 822, these additional fields are prefixed with the label X- so that they can be ignored by news servers and newsreaders. The phrase "No Archive" was coined as a way to state "Do not archive this message," and the X- prefix was added to complete the term X-No-Archive.
The proper field to prevent a message from being archived is: X-No-Archive: Yes (abbreviated as "XNAY").
Some software systems also do not archive if the first line in the body of the message contains this text. This is useful for those users who cannot change the header of messages they send out. If the X-No-Archive field is set to "No", or the field is absent, a Usenet archive will not recognize a prohibition on archiving the message.
Newsreader software programs
When DejaNews was purchased by Google, Google continued to honor the X-No-Archive directive. Other newsgroup archiving services have also followed in DejaNews' footsteps, though the decision not to archive X-No-Archive messages has been entirely voluntary.
Many popular newsreader and postin
|
https://en.wikipedia.org/wiki/Ada%20%28programming%20language%29
|
Ada is a structured, statically typed, imperative, and object-oriented high-level programming language, inspired by Pascal and other languages. It has built-in language support for design by contract (DbC), extremely strong typing, explicit concurrency, tasks, synchronous message passing, protected objects, and non-determinism. Ada improves code safety and maintainability by using the compiler to find errors in favor of runtime errors. Ada is an international technical standard, jointly defined by the International Organization for Standardization (ISO), and the International Electrotechnical Commission (IEC). , the standard, called Ada 2012 informally, is ISO/IEC 8652:2012.
Ada was originally designed by a team led by French computer scientist Jean Ichbiah of Honeywell under contract to the United States Department of Defense (DoD) from 1977 to 1983 to supersede over 450 programming languages used by the DoD at that time. Ada was named after Ada Lovelace (1815–1852), who has been credited as the first computer programmer.
Features
Ada was originally designed for embedded and real-time systems. The Ada 95 revision, designed by S. Tucker Taft of Intermetrics between 1992 and 1995, improved support for systems, numerical, financial, and object-oriented programming (OOP).
Features of Ada include: strong typing, modular programming mechanisms (packages), run-time checking, parallel processing (tasks, synchronous message passing, protected objects, and nondeterministic select statements), exception handling, and generics. Ada 95 added support for object-oriented programming, including dynamic dispatch.
The syntax of Ada minimizes choices of ways to perform basic operations, and prefers English keywords (such as "or else" and "and then") to symbols (such as "||" and "&&"). Ada uses the basic arithmetical operators "+", "-", "*", and "/", but avoids using other symbols. Code blocks are delimited by words such as "declare", "begin", and "end", where the "end" (in most
|
https://en.wikipedia.org/wiki/Generative%20artificial%20intelligence
|
Generative artificial intelligence (also generative AI or GenAI) is artificial intelligence capable of generating text, images, or other media, using generative models. Generative AI models learn the patterns and structure of their input training data and then generate new data that has similar characteristics.
In the early 2020s, advances in transformer-based deep neural networks enabled a number of generative AI systems notable for accepting natural language prompts as input. These include large language model chatbots such as ChatGPT, Bing Chat, Bard, and LLaMA, and text-to-image artificial intelligence art systems such as Stable Diffusion, Midjourney, and DALL-E.
Generative AI has uses across a wide range of industries, including art, writing, script writing, software development, product design, healthcare, finance, gaming, marketing, and fashion.
Investment in generative AI surged during the early 2020s, with large companies such as Microsoft, Google, and Baidu as well as numerous smaller firms developing generative AI models. However, there are also concerns about the potential misuse of generative AI, including cybercrime or creating fake news or deepfakes which can be used to deceive or manipulate people.
History
The academic discipline of artificial intelligence was founded at a research workshop at Dartmouth College in 1956, and has experienced several waves of advancement and optimism in the decades since. Since its founding, researchers in the field have raised philosophical and ethical arguments about the nature of the human mind and the consequences of creating artificial beings with human-like intelligence; these issues have previously been explored by myth, fiction and philosophy since antiquity. These concepts of automated art date back at least to the automata of ancient Greek civilization, where inventors such as Daedalus and Hero of Alexandria were described as having designed machines capable of writing text, generating sounds, and playing
|
https://en.wikipedia.org/wiki/TorGuard
|
Torguard is a VPN and proxy software producer which also sells private email service.
Background
Torguard VPN comes with Linux and Windows and Android apps and also browser extensions. Torguard does not keep logs and/or throttle internet speed. Torguard billing uses random hash as to keep VPN secure. VPN software uses Stealth protocol good for tight internet censorship scenario.
|
https://en.wikipedia.org/wiki/Immunoprecipitation
|
Immunoprecipitation (IP) is the technique of precipitating a protein antigen out of solution using an antibody that specifically binds to that particular protein. This process can be used to isolate and concentrate a particular protein from a sample containing many thousands of different proteins. Immunoprecipitation requires that the antibody be coupled to a solid substrate at some point in the procedure.
Types
Individual protein immunoprecipitation (IP)
Involves using an antibody that is specific for a known protein to isolate that particular protein out of a solution containing many different proteins. These solutions will often be in the form of a crude lysate of a plant or animal tissue. Other sample types could be body fluids or other samples of biological origin.
Protein complex immunoprecipitation (Co-IP)
Immunoprecipitation of intact protein complexes (i.e. antigen along with any proteins or ligands that are bound to it) is known as co-immunoprecipitation (Co-IP). Co-IP works by selecting an antibody that targets a known protein that is believed to be a member of a larger complex of proteins. By targeting this known member with an antibody it may become possible to pull the entire protein complex out of solution and thereby identify unknown members of the complex.
This works when the proteins involved in the complex bind to each other tightly, making it possible to pull multiple members of the complex out of the solution by latching onto one member with an antibody. This concept of pulling protein complexes out of solution is sometimes referred to as a "pull-down". Co-IP is a powerful technique that is used regularly by molecular biologists to analyze protein–protein interactions.
A particular antibody often selects for a subpopulation of its target protein that has the epitope exposed, thus failing to identify any proteins in complexes that hide the epitope. This can be seen in that it is rarely possible to precipitate even half of a given pro
|
https://en.wikipedia.org/wiki/Hitachi-LG%20Data%20Storage
|
Hitachi-LG Data Storage (HLDS, HL-DT-ST or H-L Data Storage), a joint venture between Hitachi, Ltd. and LG Electronics, is a manufacturer of DVD and Blu-ray optical disc drives for desktop computers and laptops. Founded in late 2000, the company began operation in January 2001. In 2006, HLDS began developing Blu-ray Disc drives. The company claims that it has led the disk drive industry in market share since its founding, with a 20% share for fiscal year 2001, 29% for fiscal year 2012, and 60% for fiscal year 2016. Building upon its core optical technology and extensive experience in production management, the company is currently expanding its product lineup to include interactive digital signage, TOF (Time of Flight) 3D sensors, unmanned store solutions, and in-vehicle air purifiers.
|
https://en.wikipedia.org/wiki/John%20Mudge
|
John Mudge (1721 – 26 March 1793) was a British physician and amateur creator of telescope mirrors. He won the Copley Medal in 1777 for a paper on reflecting telescopes.
Life
He was the fourth and youngest son of the Rev. Zachariah Mudge, by his first wife, Mary Fox, and was born at Bideford, Devon. He was educated at Bideford and Plympton grammar schools, and studied medicine at Plymouth Hospital.
Several invitations were made to Mudge to try his fortunes in London. But he preferred to remain at Plymouth, where he practised for the remainder of his life, first as surgeon, and, after 1784, when he received the degree of M.D. from King's College, Aberdeen, as a physician.
Mudge inherited a friendship with the family of Sir Joshua Reynolds, and when in 1762 Samuel Johnson accompanied Reynolds on his visit to Plymouth, Johnson became a friend and consulted Mudge as a physician. Another intimate friend was John Smeaton. Other allies and guests of Mudge were James Ferguson, the astronomer, and James Northcote, originally a chemist's assistant, who owed him his position in Reynolds's studio.
Telescopes
On 29 May 1777 Mudge was elected a Fellow of the Royal Society, and in the same year was awarded the Copley medal for his ‘Directions for making the best Composition for the Metals for reflecting Telescopes; together with a Description of the Process for Grinding, Polishing, and giving the great Speculum the true Parabolic Curve,’ which were communicated by the author to the society, and printed in the Philosophical Transactions (1777, lxvii. 296). The ‘Directions’ were also issued separately by Bowyer (London, 1778). Sir John Pringle, the president, in making the presentation, remarked that Isaac Newton had predicted the role of mechanical devices in making parabolic mirrors.
The manufacture of telescopes continued to occupy much of his spare time. He made two large ones with a magnifying power of two hundred times; one of these he gave to Hans Moritz von Brühl, and i
|
https://en.wikipedia.org/wiki/Sausage%20making
|
The origins of meat preservation are lost to the ages but probably began when humans began to realize the preservative value of salt.
Sausage making originally developed as a means to preserve and transport meat. Primitive societies learned that dried berries and spices could be added to dried meat.
The procedure of stuffing meat into casings remains basically the same today, but sausage recipes have been greatly refined and sausage making has become a highly respected culinary art.
Sausages come in two main types: fresh and cured. Cured sausages may be either cooked or dried. Many cured sausages are smoked, but this is not mandatory. The curing process itself changes the meat and imparts its own flavors. An example is the difference in taste between a pork roast and a ham.
All smoked sausages are cured. The reason is the threat of botulism. The bacterium responsible, Clostridium botulinum, is ubiquitous in the environment, grows in the anaerobic conditions created in the interior of the sausage, and thrives in the to temperature range common in the smoke house and subsequent ambient storage. Thus, for safety reasons, sausages are cured before smoking.
Types of sausages and their storage
Fresh sausages
Fresh sausages are simply seasoned ground meats that are cooked before serving. Fresh sausages normally do not use cure (Prague powder #1) although cure can be used if desired. In addition fresh sausages typically do not use smoke flavors, although liquid smoke can be used. Fresh sausages are never smoked in a cold smoker because of the danger of botulism.
The primary seasoning agents in fresh sausages are salt and sugar along with various savory herbs and spices, and often vegetables, including onion and garlic.
A British fresh sausage typically contains around 10% butcher's rusk, 10% water, 2.5% seasoning, and 77.5% meat. At the point of sale, British sausages will often be labelled as "actual meat content X%". As meat can be fatty or lean, the X% is calcul
|
https://en.wikipedia.org/wiki/The%20Future%20Is%20Wild
|
The Future Is Wild (also referred to by the acronym FIW) is a 2002 speculative evolution docufiction miniseries and an accompanying multimedia entertainment franchise. The Future Is Wild explores the ecosystems and wildlife of three future time periods: 5, 100, and 200 million years in the future, in the format of a nature documentary. Though the settings and animals are fictional, the series has an educational purpose, serving as an informative and entertaining way to explore concepts such as evolution and climate change.
The Future Is Wild was first conceived by independent producer Joanna Adams in 1996 and developed together with various scientists, including Dougal Dixon, best known as the author of the 1981 book After Man, which also explored future wildlife. The 2002 series was an international co-production, involving the British BBC, the Franco-German channel Arte, the German ZDF, the Austrian ORF, the Italian MFE - MediaForEurope (via their Mediaset division), and the American Animal Planet and Discovery Channel. Wildly successful, The Future Is Wild continues to be broadcast to this day and has been shown on TV in more than 60 countries.
The success of The Future Is Wild spawned a large multimedia franchise, including books, children's entertainment, exhibitions, theme park rides, educational material, and toys. There have also been cancelled projects, such as a potential movie adaptation, as well as a sequel series, The Future Is Wild 2. From 2016 onwards, there has been talk of "relaunching" the franchise through various projects, such as an action-adventure TV series and The Future is Wild VR (a virtual reality videogame), though no new media has yet materialized.
Premise
The Future Is Wild explores twelve different future ecosystems across three future time periods: 5 million years in the future, 100 million years in the future, and 200 million years in the future. Four ecosystems from each period are explored and described.
Ice World: 5 million y
|
https://en.wikipedia.org/wiki/Arrhenius%20equation
|
In physical chemistry, the Arrhenius equation is a formula for the temperature dependence of reaction rates. The equation was proposed by Svante Arrhenius in 1889, based on the work of Dutch chemist Jacobus Henricus van 't Hoff who had noted in 1884 that the van 't Hoff equation for the temperature dependence of equilibrium constants suggests such a formula for the rates of both forward and reverse reactions. This equation has a vast and important application in determining the rate of chemical reactions and for calculation of energy of activation. Arrhenius provided a physical justification and interpretation for the formula. Currently, it is best seen as an empirical relationship. It can be used to model the temperature variation of diffusion coefficients, population of crystal vacancies, creep rates, and many other thermally-induced processes/reactions. The Eyring equation, developed in 1935, also expresses the relationship between rate and energy.
Equation
The Arrhenius equation gives the dependence of the rate constant of a chemical reaction on the absolute temperature as
where
is the rate constant (frequency of collisions resulting in a reaction),
is the absolute temperature (in Kelvin or degree Rankine),
is the pre-exponential factor or Arrhenius factor or frequency factor. Arrhenius originally considered A to be a temperature-independent constant for each chemical reaction. However more recent treatments include some temperature dependence - see Modified Arrhenius equation below.
is the activation energy for the reaction (in the same units as RT),
is the universal gas constant.
Alternatively, the equation may be expressed as
where
is the activation energy for the reaction (in the same units as kBT),
is the Boltzmann constant.
The only difference is the energy units of : the former form uses energy per mole, which is common in chemistry, while the latter form uses energy per molecule directly, which is common in physics.
The different units
|
https://en.wikipedia.org/wiki/Electronic%20anticoincidence
|
Electronic anticoincidence is a method (and its associated hardware) widely used to suppress unwanted, "background" events in high energy physics, experimental particle physics, gamma-ray spectroscopy, gamma-ray astronomy, experimental nuclear physics, and related fields.
In the typical case, a desired high-energy interaction or event occurs and is detected by some kind of detector, creating a fast electronic pulse in the associated nuclear electronics. But the desired events are mixed up with a significant number of other events, produced by other particles or processes, which create indistinguishable events in the detector. Very often it is possible to arrange other physical photon or particle detectors to intercept the unwanted background events, producing essentially simultaneous pulses that can be used with fast electronics to reject the unwanted background.
Gamma-ray astronomy
Early experimenters in X-ray and gamma-ray astronomy found that their detectors, flown on balloons or sounding rockets, were corrupted by the large fluxes of high-energy photon and cosmic-ray charged-particle events. Gamma-rays, in particular, could be collimated by surrounding the detectors with heavy shielding materials made of lead or other such elements, but it was quickly discovered that the high fluxes of very penetrating high-energy radiation present in the near-space environment created showers of secondary particles that could not be stopped by reasonable shielding masses. To solve this problem, detectors operating above 10 or 100 keV were often surrounded by an active anticoincidence shield made of some other detector, which could be used to reject the unwanted background events.
An early example of such a system, first proposed by Kenneth John Frost in 1962, is shown in the figure. It has an active CsI(Tl) scintillation shield around the X-ray/gamma-ray detector, also of CsI(Tl), with the two connected in electronic anticoincidence to reject unwanted charged particl
|
https://en.wikipedia.org/wiki/Interactionism%20%28nature%20versus%20nurture%29
|
In the context of the nature-nurture debate, interactionism is the view that all human behavioral traits develop from the interaction of both "nature" and "nurture", that is, from both genetic and environmental factors. This view further holds that genetic and environmental influences on organismal development are so closely interdependent that they are inseparable from one another. Historically, it has often been confused with the statistical concept of gene-environment interaction. Historically, interactionism has presented a limited view of the manner in which behavioral traits develop, and has simply demonstrated that "nature" and "nurture" are both necessary. Among the first biologists to propose an interactionist theory of development was Daniel Lehrman. Since then, numerous interactionist perspectives have been proposed, and the contradictions between many of these perspectives has led to much controversy in evolutionary psychology and behavioral genetics. Proponents of various forms of interactionist perspectives include Philip Kitcher, who refers to his view as "causal democracy", and Susan Oyama, who describes her perspective as "constructive interactionism". Critics of interactionism include major figures in behavioral genetics such as Arthur Jensen, Robert Plomin, and philosopher Neven Sesardic.
Interactionist perspective to depression
Depression is not dependent entirely on genetic nor environmental influences alone to trigger its onset. Both genetic and environmental factors work accompanied to transform a vulnerability to depression to be expressed in its actuality. Research has demonstrated the synchrony of polygenic scores of major depressive disorders (MDD) with stressful life events and social support to increase the probability of developing depression. Although Monroe and Simons criticize interactionism for a lack of precise measurement to grasp its ‘conceptual essence’, there’s been numerous studies surrounding gene by environment interactio
|
https://en.wikipedia.org/wiki/Resource%20fragmentation%20hypothesis
|
The resource fragmentation hypothesis was first proposed by Janzen & Pond (1975), and says that as species richness becomes large there is not a linear increase in the number of parasitoid species that can be supported. The mechanism for this hyperbolic relationship is suggested to be that each of the new host species are too rare to support the evolution of specialist parasitoids (Janzen & Pond, 1975). The resource fragmentation hypothesis is one of two hypotheses that seek to explain the distribution of the Ichneumonidae.
|
https://en.wikipedia.org/wiki/Gajski%E2%80%93Kuhn%20chart
|
The Gajski–Kuhn chart (or Y diagram) depicts the different perspectives in VLSI hardware design. Mostly, it is used for the development of integrated circuits. Daniel Gajski and Robert Kuhn developed it in 1983. In 1985, Robert Walker and Donald Thomas refined it.
According to this model, the development of hardware is perceived within three domains that are depicted as three axis and produce a Y. Along these axis, the abstraction levels that describe the degree of abstraction. The outer shells are generalisations, the inner ones refinements of the same subject.
The issue in hardware development is most often a top-down design problem. This is perceived by the three domains of behaviour, structure, and the layout that goes top-down to more detailed abstraction levels. The designer can select one of the perspectives and then switch from one view to another. Generally, the design process is not following a specific sequence in this diagram.
On the system level, basic properties of an electronic system are determined. For the behavioural description, block diagrams are used by making abstractions of signals and their time response. Blocks used in the structure domain are CPUs, memory chip, etc.
The algorithmic level is defined by the definition of concurrent algorithms (signals, loops, variables, assignments). In the structural domain, blocks like ALUs are in use.
The register-transfer level (RTL) is a more detailed abstraction level on which the behaviour between communicating registers and logic units is described. Here, data structures and data flows are defined. In the geometric view, the design step of the floorplan is located.
The logical level is described in the behaviour perspective by boolean equations. In the structural view, this is displayed with gates and flip-flops. In the geometric domain, the logical level is described by standard cells.
The behaviour of the circuit level is described by mathematics using differential equations or logical equa
|
https://en.wikipedia.org/wiki/Wolf%20%28video%20game%29
|
Wolf is a 1994 life simulation and role-playing video game. The player takes the role of a wolf. It was followed by Lion in 1995.
It is unrelated to the 1994 film of the same name.
Gameplay
The gameplay is divided into two parts. The first is a sandbox mode, where the player has no predetermined goal. The second is a scenario mode, where the player has to complete specific actions; this is comparable to quests given in RPGs.
Reception
In Electronic Entertainment, Joel Enos wrote that Wolfs "unique take on role-playing/simulation games puts it in a class by itself". He concluded, "You'll not only spend many happy hours playing, you might even learn something." The magazine later named Wolf its 1994 "Best Simulation Game". The editors reiterated that the game is "great fun to play, and you also can't help but learn about these intriguing creatures as you step into their skin."
Vince DeNardo of Computer Gaming World called Wolf "a role-playing simulation that is both worthwhile for your children, and for the child that lies within each of us." He believed that it is "a novel concept backed up by solid execution", and that it "redefines the genre of Role-Playing as we gamers know it". Computer Gaming World went on to nominate Wolf as its 1994 "Role-Playing Game of the Year", with the editors calling it an innovative product that "skillfully mixes role-playing elements and scientific fact".
The reviewer for PC Gamer US remarked that "hours pass like minutes in this fascinating RPG for nature lovers", and summarized it as an "unusual, entertaining game that gives genuine insight into one of nature's most magnificent and misunderstood creatures." The magazine's editors later awarded Wolf their 1994 "Special Achievement in Innovative Design". They wrote that the game is "both entertaining and enlightening — and a breath of fresh air in a genre [role-playing games] that some say has run its course."
Computer Players Peter Suciu summarized it as "a nice novelty game
|
https://en.wikipedia.org/wiki/Algebra%20i%20Logika
|
Algebra i Logika (English: Algebra and Logic) is a peer-reviewed Russian mathematical journal founded in 1962 by Anatoly Ivanovich Malcev, published by the Siberian Fund for Algebra and Logic at Novosibirsk State University. An English translation of the journal is published by Springer-Verlag as Algebra and Logic since 1968. It published papers presented at the meetings of the "Algebra and Logic" seminar at the Novosibirsk State University. The journal is edited by academician Yury Yershov.
The journal is reviewed cover-to-cover in Mathematical Reviews and Zentralblatt MATH.
Abstracting and Indexing
Algebra i Logika is indexed and abstracted in the following databases:
According to the Journal Citation Reports, the journal had a 2020 impact factor of 0.753.
|
https://en.wikipedia.org/wiki/Nvidia%20DGX
|
Nvidia DGX is a line of Nvidia-produced servers and workstations which specialize in using GPGPU to accelerate deep learning applications. The typical design of a DGX system is based upon a rackmount chassis with motherboard that carries high performance x86 server CPUs (Typically Intel Xeons, with the exception DGX A100 and DGX Station A100, which both utilize AMD EPYC CPUs). The main component of a DGX system is a set of 4 to 16 Nvidia Tesla GPU modules on an independent system board. DGX systems have large heatsinks and powerful fans to adequately cool thousands of watts of thermal output. The GPU modules are typically integrated into the system using a version of the SXM socket or by a PCIe x16 slot.
Models
Pascal - Volta
DGX-1
DGX-1 servers feature 8 GPUs based on the Pascal or Volta daughter cards with 128GB of total HBM2 memory, connected by an NVLink mesh network. The DGX-1 was announced on the 6th of April in 2016. All models are based on a dual socket configuration of Intel Xeon E5 CPUs, and are equipped with the following features.
512 GB of DDR4-2133
Dual 10Gb networking
4 x 1.92 TB SSDs
3200W of combined power supply capability
3U Rackmount Chassis
The product line is intended to bridge the gap between GPUs and AI accelerators in that the device has specific features specializing it for deep learning workloads.
The initial Pascal based DGX-1 delivered 170 teraflops of half precision processing, while the Volta-based upgrade increased this to 960 teraflops.
The DGX-1 was first available only with the Pascal based configuration, with the first generation SXM socket. The later revision of the DGX-1 offered support for first generation Volta cards via the SXM-2 socket. Nvidia offered upgrade kits that allowed users with a Pascal based DGX-1 to upgrade to a Volta based DGX-1.
The Pascal based DGX-1 has two variants, one with a 16 core Intel Xeon E5-2698 V3, and one with a 20 core E5-2698 V4. Pricing for the variant equipped with an E5-2698 V4
|
https://en.wikipedia.org/wiki/Hikonyan
|
is a mascot created by the city government of Hikone, Japan. He was created in 2007 to mark the 400th anniversary of the founding of Hikone Castle. The character design is derived from a legend concerning maneki-neko and Ii Naotaka, the 3rd Lord of Hikone. The daimyō was beckoned by a white cat to seek shelter from a storm in a temple, and thus saved from a lightning strike. In Japanese, "nyan" is an onomatopoeia for a cat's meow. Hikonyan's samurai helmet is based on a Ii family helmet currently in the Hikone Castle museum. Hikonyan's popularity increased tourist visitation of Hikone by over 200,000 annually. The estimate of Hikonyan's effect on the tourist industry is 17.4 billion yen (approximately US$218 million) and the overall economic effects total 33.8 billion yen (US$425 million). Total merchandise sales reached about 1.7 billion yen (US$21 million) as of 2008. In 2010, Hikonyan won first place in the open vote competition of other mascots, collectively known as .
See also
Kumamon
Funassyi
Choruru
Kigurumi
|
https://en.wikipedia.org/wiki/Flatland
|
Flatland: A Romance of Many Dimensions is a satirical novella by the English schoolmaster Edwin Abbott Abbott, first published in 1884 by Seeley & Co. of London. Written pseudonymously by "A Square", the book used the fictional two-dimensional world of Flatland to comment on the hierarchy of Victorian culture, but the novella's more enduring contribution is its examination of dimensions.
Several films have been made from the story, including the feature film Flatland (2007). Other efforts have been short or experimental films, including one narrated by Dudley Moore and the short films Flatland: The Movie (2007) and Flatland 2: Sphereland (2012).
Plot
The story describes a two-dimensional world inhabited by geometric figures (flatlanders); women are line segments, while men are polygons with various numbers of sides. The narrator is a square, a member of the caste of gentlemen and professionals, who guides the readers through some of the implications of life in two dimensions. The first half of the story goes through the practicalities of existing in a two-dimensional universe, as well as a history leading up to the year 1999 on the eve of the 3rd Millennium.
On New Year's Eve, the Square dreams of a visit to a one-dimensional world, "Lineland", inhabited by men, consisting of lines, while the women consisted of "lustrous points". These points and lines are unable to see the Square as anything other than a set of points on a line. Thus, the Square attempts to convince the realm's monarch of a second dimension but cannot do so. In the end, the monarch of Lineland tries to kill the Square rather than tolerate him any further.
Following this vision, the Square is visited by a sphere. Similar to the "points" in Lineland, he is unable to see the three-dimensional object as anything other than a circle (more precisely, a disk). The Sphere then levitates up and down through Flatland, allowing the Square to see the circle expand and contract between great circle and sm
|
https://en.wikipedia.org/wiki/NORDUnet
|
NORDUnet is an international collaboration between the National research and education networks in the Nordic countries.
Members
The members of NORDUnet are:
SUNET of Sweden
UNINETT of Norway
FUNET of Finland
Forskningsnettet of Denmark
RHnet of Iceland
Network
NORDUnet interconnects the Nordic national research and education networks and connects them to the worldwide network for research and education and to the general purpose Internet. NORDUnet provides its services by a combination of leased lines and Internet services provided by other international operators. NORDUnet has peering in multiple important internet exchange sites outside the Nordics, such as Amsterdam, Chicago, Frankfurt, London, Miami and New York.
In addition to the basic Internet service NORDUnet operates information services and provides USENET NetNews and Multicast connectivity to the Nordic national networks. NORDUnet also coordinates the national networks' Computer Emergency Response Team (CERT) activities and the Nordic national networks' IPv6 activities - an area where NORDUnet has been active for years.
NORDUnet is one of the members, alongside Internet2, ESnet, SURFnet, CANARIE and GÉANT, to pilot a 100G intercontinental connection between Europe and North America.
History
NORDUnet is the result of the NORDUNET programme (1986 to 1992) financed by the Nordic Council of Ministers, officially beginning operations 1989. It was the first European NREN to embrace the TCP/IP technology and to connect to the National Science Foundation Network in the United States providing open access for university students in member countries. Along with other early adopters of TCP/IP, particularly CERN, it encouraged the adoption of TCP/IP in Europe (see Protocol Wars).
NORDUnet has only few permanent employees. Most of the work is contracted to appropriate organisations in the Nordic area.
Distinction
The web site for NORDUNet, nordu.net, is the oldest active domain name. It was registered
|
https://en.wikipedia.org/wiki/The%20Jepson%20Manual
|
The Jepson Manual is a flora of the vascular plants that are either native to or naturalized in California. Botanists often refer to the book simply as Jepson. It is produced by the University and Jepson Herbaria, of the University of California, Berkeley. Its second edition is the basis of the online Jepson eFlora.
History
1923: Willis Linn Jepson – Manual of the Flowering Plants of California
1958, 1968: Philip Alexander Munz – A California Flora and Supplement
1993: James Craig Hickman (editor) – The Jepson Manual, Higher Plants of California (TJM93)
2012: Bruce Gregg Baldwin – The Jepson Manual: Vascular Plants of California, 2nd edition (TJM2)
2010−ongoing: The Jepson Online Interchange for California Floristics − Jepson eFlora (TJM2) – online.
Preceding works
The Jepson Manual also follows Philip A. Munz and David D. Keck in their A California Flora and Supplement of 1958 and 1968.
Like other florae, The Jepson Manual builds upon these prior publications. Except for the number of line drawings, it has more in common with Munz's 1968 book than with Jepson's 1923 book.
Editions
The first edition of The Jepson Manual was published in 1993 as The Jepson Manual: Higher Plants of California (TJM93), and was edited by James C. Hickman.
The second edition was published in 2012, as The Jepson Manual: Vascular Plants of California, Thoroughly Revised and Expanded (TJM2), and was edited by Bruce Gregg Baldwin, Douglas H. Goldman, David John Keil, Robert Patterson, and Thomas James Rosatti. The second edition features 7,601 California plant species, subspecies and varieties.
While the book is named in honor of Jepson, The Jepson Manual is not simply a new edition of Jepson's 1923 book, but a new work which Baldwin calls "the most time-consuming undertaking of my career."
Jepson eFlora
The Jepson eFlora is a taxonomic database that builds on and expands the second edition of The Jepson Manual. It describes itself as "the foremost authority on the native and natura
|
https://en.wikipedia.org/wiki/Insecticide
|
Insecticides are pesticides used to kill insects. They include ovicides and larvicides used against insect eggs and larvae, respectively. Insecticides are used in agriculture, medicine, industry and by consumers. Insecticides are claimed to be a major factor behind the increase in the 20th-century's agricultural productivity. Nearly all insecticides have the potential to significantly alter ecosystems; many are toxic to humans and/or animals; some become concentrated as they spread along the food chain.
Insecticides can be classified into two major groups: systemic insecticides, which have residual or long-term activity; and contact insecticides, which have no residual activity.
The mode of action describes how the pesticide kills or inactivates a pest. It provides another way of classifying insecticides. Mode of action can be important in understanding whether an insecticide will be toxic to unrelated species, such as fish, birds and mammals.
Insecticides may be repellent or non-repellent. Social insects such as ants cannot detect non-repellents and readily crawl through them. As they return to the nest they take insecticide with them and transfer it to their nestmates. Over time, this eliminates all of the ants including the queen. This is slower than some other methods, but usually completely eradicates the ant colony.
Insecticides are distinct from non-insecticidal repellents, which repel but do not kill.
Type of activity
Systemic insecticides
Systemic insecticides become incorporated and distributed systemically throughout the whole plant. When insects feed on the plant, they ingest the insecticide. Systemic insecticides produced by transgenic plants are called plant-incorporated protectants (PIPs). For instance, a gene that codes for a specific Bacillus thuringiensis biocidal protein was introduced into corn (maize) and other species. The plant manufactures the protein, which kills the insect when consumed.
Contact insecticides
Contact insecticides are
|
https://en.wikipedia.org/wiki/Janny%20Leung
|
Janny May-yee Leung () is a Chinese operations researcher and academic administrator, the master of Choi Kai Yau College at the University of Macau and courtesy professor in the State Key Laboratory of Internet of Things for Smart City of the University of Macau. Topics in her research have included transportation scheduling, logistics, facility location, and polyhedral combinatorics.
Education and career
Leung studied applied mathematics in Radcliffe College at Harvard University, graduating in 1979. After studying for a second bachelor's degree at Magdalen College, Oxford, from 1979 to 1981, she earned a Ph.D. in operations research at the Massachusetts Institute of Technology in 1986, under the supervision of Thomas L. Magnanti. Her dissertation was Polyhedral Structure of Capacitated Fixed Charge Problems and A Problem in Delivery Route Planning.
She became an assistant professor of operations research in the Yale University School of Organization and Management, and in 1991 moved to the Management Information Systems Department at the University of Arizona, earning tenure there.
In 1996 she moved to the City University of Hong Kong as an associate professor of management sciences, and in 1998 she moved to the Chinese University of Hong Kong, in systems engineering and engineering management. At the Chinese University of Hong Kong, she helped found Morningside College, becoming its first dean of students and deputy master in 2009, and later becoming its warden.
Leung became professor of science and engineering, and founding master of Shaw College of the Chinese University of Hong Kong, Shenzhen (CUHK-Shenzhen), in 2016. Morningside College continues to list her as an emerita fellow.
Since 2019, she has resigned from CUHK-Shenzhen and become master of Choi Kai Yau College at the University of Macau.
Recognition
Leung was president of the Forum for Women in OR/MS of the Institute for Operations Research and the Management Sciences (INFORMS) for 2001–2002. S
|
https://en.wikipedia.org/wiki/Welt%20%28bruise%29
|
A welt is a bloodshot stripe on the skin. They are haematoma, a special form of a bruise.
Welts occur when blunt force is applied to the body with elongated objects without sharp edges. Like other haematomas, welts change their colors as they heal, which usually takes 2 to 4 weeks. The colors include purplish black, reddish blue, brown or yellowish green. That makes it possible to determine their approximate age.
A special form of welts—when the impact tool is not flat—are double welts. They occur after blows with stick-like, long objects, such as a cane for caning, a riding crop or a garden hose. The energy of the striking tool hitting the skin forces the tissue fluid or tissue matrix to the outside of the instrument. This leaves a rather anaemic, brighter area of skin, so the area of impact of the striking tool on the skin is usually free of injury. However, capillary vessels located under the skin at the edges of the impacting object rupture, resulting in two parallel, striated skin lacerations. Such double straps in a child indicate abuse and require clarification.
|
https://en.wikipedia.org/wiki/Avenger%20%281986%20video%20game%29
|
Way of the Tiger II: Avenger is a video game made by Gremlin Graphics in 1986, for the computers Commodore 64, Amstrad CPC, ZX Spectrum and MSX.
Gameplay
In the game's story, Yaemon the Grand Master of Flame has killed the player's foster-father Naijish and stolen the Scrolls of Kettsuin. To recover the scrolls, the player has to find enough keys to penetrate the Quench Heart Keep, and then kill each of the 3 guards. The game is viewed from above and superficially resembles Gauntlet.
Reception
Guy Langley for Computer and Video Games compared the game to Gauntlet, noting that Avenger was more of an adventure and has "smarter graphics".
Rich Pelley for Your Sinclair gave the game a rating of 86 and praised its variety, with elements of maze games, beat-em ups, and puzzlers.
Zzap! rated the game at 86%, calling it "highly competent".
Mike Roberts for Computer Gamer rated the game at 85% overall and complimented the life energy system.
Amtix rated the game at 73% overall, noting that it was too similar to the original Way of the Tiger.
Commodore User scored the game 7 out of 10 overall, noting its "tough arcade adventure-style".
Reviews
Jeux & Stratégie #43
Aktueller Software Markt
Happy Computer
|
https://en.wikipedia.org/wiki/Joseph%20A.%20Thas
|
Joseph Adolphe François Thas (born 13 October 1944, Dilbeek, Belgium) is a Belgian mathematician, who works on combinatorics, incidence geometry and finite geometries.
Thas received in 1969 his PhD from Ghent University under Julien Bilo with thesis Een studie betreffende de projective rechte over de totale matrix algebra der 3x3-matrices met elementen in een algebraïsch afgesloten veld K. Thas showed how to extend projective geometry and cross-ratios with the concept of a projective line over a ring.
Thas is an emeritus professor at Ghent University.
Awards and honors
In 1994 Thas received the Euler medal. In 1998 he gave an invited address at the International Congress of Mathematicians in Berlin with lecture Finite geometries, varieties and codes. He received in 1969 the prize of the Royal Academy of Sciences, Letters and Fine Arts of Belgium, in 1970 the Scientific Louis Empain Award and in the same year the François Deruyts prize of the Royal Academy of Belgium.
In 1988 he became a member of the Royal Flemish Academy of Belgium for Science and the Arts; he was vice-director of the Class of Sciences in 1998 and director in 1999. In 1999 he was awarded an Erskine Fellowship of the University of Canterbury, New Zealand, in 2008 he was Platinum Jubilee Lecturer at the Indian Statistical Institute, and in 2012 he became one of the inaugural fellows of the American Mathematical Society.
Selected works
with Koen Thas, H. Van Maldeghem Translation generalized quadrangles, World Scientific 2006
with Stanley E. Payne Finite generalized quadrangles, Pitman 1984, 2nd edition, European Mathematical Society 2009
with J. W. P. Hirschfeld General Galois Geometries, Oxford University Press 1991
Projective geometry over a finite field and Generalized Polygons in F. Buekenhout Handbook of incidence geometry, North Holland 1995
with J. Bilo Enkele aspecten van de theorie der axiomatische projectieve vlakken, Simon Stevin, Supplement, Vol. 55, 1981
(For a complete list of pa
|
https://en.wikipedia.org/wiki/Divisia%20monetary%20aggregates%20index
|
In econometrics and official statistics, and particularly in banking, the Divisia monetary aggregates index is an index of money supply. It uses Divisia index methods.
Background
The monetary aggregates used by most central banks (notably the US Federal Reserve) are simple-sum indexes in which all monetary components are assigned the same weight:
in which is one of the monetary components of the monetary aggregate . The summation index implies that all monetary components contribute equally to the money total, and it views all components as dollar-for-dollar perfect substitutes. It has been argued that such an index does not weight such components in a way that properly summarizes the services of the quantities of money.
There have been many attempts at weighting monetary components within a simple-sum aggregate. An index can rigorously apply microeconomic- and aggregation-theoretic foundations in the construction of monetary aggregates. That approach to monetary aggregation was derived and advocated by William A. Barnett (1980) and has led to the construction of monetary aggregates based on Diewert's (1976) class of superlative quantity index numbers. The new aggregates are called the Divisia aggregates or Monetary Services Indexes. Salam Fayyad's 1986 PhD dissertation did early research with those aggregates using U.S. data.
This index is a discrete-time approximation with this definition:
Here, the growth rate of the aggregate is the weighted average of the growth rates of the component quantities. The discrete time Divisia weights are defined as the expenditure shares averaged over the two periods of the change
for , where
is the expenditure share of asset during period , and is the user cost of asset , derived by Barnett (1978),
Which is the opportunity cost of holding a dollar's worth of the th asset. In the last equation, is the market yield on the th asset, and is the yield available on a benchmark asset, held only to carry wealth between diff
|
https://en.wikipedia.org/wiki/Monoboard
|
A monoboard is a device or product that consists of a single printed circuit board (PCB).
Benefits
The primary benefit of a monoboard solution is cost savings. There are a number of ways that incorporating all parts on a single board can reduce costs. The primary reason is that solutions with multiple boards require connections between the boards via edge connectors, and sometimes include a ribbon cable. By connecting devices directly together on the same PCB, there is no need for these additional connectors and cables. Additionally, PCB space may be optimized using electronic design automation (EDA) tools, resulting in a smaller device as well as further cost savings. Aesthetically enhanced micro branding is used as well.
Disadvantages
The disadvantages of using a monoboard solution is that they are inflexible to upgrades. For example, take a personal computer; in most personal computers, the video card is an external device that plugs into a dedicated slot in the motherboard and may be swapped out with a higher performing card. However, in small form-factor designs, a graphics processing unit (GPU) may be placed directly on the board. The typical reason is to reduce size and cost, but this removed the ability to later upgrade the device.
Another common example is with data acquisition hardware. Many DAQ solutions involve the use of DAQ modules, which are cards that plug into a backplane. Different DAQ modules can be purchased with different functionalities depending on the speed and resolution of signals being acquired. Examples of this can be seen in products ranging from Fluke DAQs to Tektronix Logic Analyzer modules (such as the TLA7000 series frames which support more than 4 different series of acquisition cards).
Examples
Personal computers, specifically small form-factor versions such as the UMPC or MacBook Air
Data acquisition hardware
Cameras
Mobile phones
Single-board computer
Motherboard
Printed circuit board manufacturing
|
https://en.wikipedia.org/wiki/Enterprise%20Distributed%20Object%20Computing
|
The UML profile for Enterprise Distributed Object Computing (EDOC) is a standard of the Object Management Group in support of open distributed computing using model-driven architecture and service-oriented architecture. Its aim is to simplify the development of component based (EDOC) systems by providing a UML-based modeling framework conforming to the MDA of the OMG.
The basis of EDOC is the Enterprise Collaboration Architecture, ECA, meta model that defines how roles interact within communities in the performance of collaborative business processes.
The seven EDOC specifications
EDOC is composed of seven specifications:
The Enterprise Collaboration Architecture, ECA
The Metamodel and UML Profile for Java and EJB
The Flow Composition Model, FCM
The UML Profile for Patterns
The UML Profile for ECA
The UML Profile for Meta Object Facility
The UML Profile for Relationships
See also
Model Driven Engineering (MDE)
Model-driven architecture (MDA)
Meta-model
Meta-modeling
Meta-Object Facility (MOF)
Unified Modeling Language (UML)
External links
OMG EDOC Standard at the Internet Archive
Unified Modeling Language
Year of introduction missing
|
https://en.wikipedia.org/wiki/ScREC
|
ScREC is a supercomputer developed by the Research Centre for Modeling and Simulation (RCMS) at the National University of Sciences and Technology, Pakistan (NUST) in Islamabad, Pakistan. With a 132 teraflops performance, it is currently the fastest supercomputer in Pakistan.
System specifications
ScREC is able to perform parallel computing and has a performance speed of 132 teraFLOPS (trillion operations per second). It is the fastest running graphics processing unit (GPU) parallel computing system to have been developed in Pakistan. The supercomputer has multi-core processors and graphics co-processors, with an inter-process communication speed of 40 gigabits per second. According to system specifications, the computer cluster consists of 66 nodes equipped with 30,992 processor cores. Additional component details include:
32 dual quad core computer nodes (256 processor cores)
32 Nvidia graphics processing units
QDR InfiniBand interconnection
21.6 TB storage
See also
Information technology in Pakistan
Supercomputing in Pakistan
|
https://en.wikipedia.org/wiki/Electronic%20signature
|
An electronic signature, or e-signature, is data that is logically associated with other data and which is used by the signatory to sign the associated data. This type of signature has the same legal standing as a handwritten signature as long as it adheres to the requirements of the specific regulation under which it was created (e.g., eIDAS in the European Union, NIST-DSS in the USA or ZertES in Switzerland).
Electronic signatures are a legal concept distinct from digital signatures, a cryptographic mechanism often used to implement electronic signatures. While an electronic signature can be as simple as a name entered in an electronic document, digital signatures are increasingly used in e-commerce and in regulatory filings to implement electronic signatures in a cryptographically protected way. Standardization agencies like NIST or ETSI provide standards for their implementation (e.g., NIST-DSS, XAdES or PAdES). The concept itself is not new, with common law jurisdictions having recognized telegraph signatures as far back as the mid-19th century and faxed signatures since the 1980s.
Description
An electronic signature is intended to provide a secure and accurate identification method for the signatory during a transaction.
Definitions of electronic signatures vary depending on the applicable jurisdiction. A common denominator in most countries is the level of an advanced electronic signature requiring that:
The signatory can be uniquely identified and linked to the signature
The signatory must have sole control of the private key that was used to create the electronic signature
The signature must be capable of identifying if its accompanying data has been tampered with after the message was signed
In the event that the accompanying data has been changed, the signature must be invalidated
Electronic signatures may be created with increasing levels of security, with each having its own set of requirements and means of creation on various levels that prove t
|
https://en.wikipedia.org/wiki/Arrested%20development
|
The term "arrested development" has had multiple meanings for over 200 years. In the field of medicine, the term "arrested development" was first used, circa 1835–1836, to mean a stoppage of physical development; the term continues to be used in the same way. In literature, Ernest Hemingway used the term in The Sun Also Rises, published in 1926: On page 51, Harvey tells Cohn, "I misjudged you [...] You're not a moron. You're only a case of arrested development."
In contrast, the UK's Mental Health Act 1983 used the term "arrested development" to characterize a form of mental disorder comprising severe mental impairment, resulting in a lack of intelligence. However, some researchers have objected to the notion that mental development can be "arrested" or stopped, preferring to consider mental status as developing in other ways in psychological terminology. Consequently, the term "arrested development" is no longer used when referring to a developmental disorder in mental health.
In anthropology and archaeology, the term "arrested development" means that a plateau of development in some sphere has been reached. Often it is a technological plateau such as the development of high temperature ceramics, but without glaze because of a lack of materials, or copper smelting without development of bronze because of a lack of tin. Arrested development is key in the insight of self-domestication in the evolution of hominidae where it involves being in an environment that favors reduction in aggression, including interspecific and intraspecific antagonism, for survival, in favor of attitudes that favor living together in a group, social behavior, traits that favor the group as a whole to come to the front stage, elimination of bullies - individuals with an antisocial personality disorder.
|
https://en.wikipedia.org/wiki/Multidrug-resistant%20tuberculosis
|
Multidrug-resistant tuberculosis (MDR-TB) is a form of tuberculosis (TB) infection caused by bacteria that are resistant to treatment with at least two of the most powerful first-line anti-TB medications (drugs): isoniazid and rifampin. Some forms of TB are also resistant to second-line medications, and are called extensively drug-resistant TB (XDR-TB).
Tuberculosis is caused by infection with the bacterium Mycobacterium tuberculosis. Almost one in four people in the world are infected with TB bacteria. Only when the bacteria become active do people become ill with TB. Bacteria become active as a result of anything that can reduce the person's immunity, such as HIV, advancing age, diabetes or other immunocompromising illnesses. TB can usually be treated with a course of four standard, or first-line, anti-TB drugs (i.e., isoniazid, rifampin, pyrazinamide and ethambutol).
However, beginning with the first antibiotic treatment for TB in 1943, some strains of the TB bacteria developed resistance to the standard drugs through genetic changes (see mechanisms.) Currently the majority of multidrug-resistant cases of TB are due to one strain of TB bacteria called the Beijing lineage. This process accelerates if incorrect or inadequate treatments are used, leading to the development and spread of multidrug-resistant TB (MDR-TB). Incorrect or inadequate treatment may be due to use of the wrong medications, use of only one medication (standard treatment is at least two drugs), or not taking medication consistently or for the full treatment period (treatment is required for several months). Treatment of MDR-TB requires second-line drugs (i.e., fluoroquinolones, aminoglycosides, and others), which in general are less effective, more toxic and much more expensive than first-line drugs. Treatment schedules for MDR-TB involving fluoroquinolones and aminoglycosides can run for two years, compared to the six months of first-line drug treatment, and cost over US$100,000. If these sec
|
https://en.wikipedia.org/wiki/White%20stag
|
A white stag (or white hind for the female) is a white-colored red deer, wapiti, sika deer, chital, fallow deer, roe deer, white-tailed deer, black-tailed deer, reindeer, moose, or rusa, explained by a condition known as leucism that causes its hair and skin to lose its natural colour. The white deer has played a prominent role in many cultures' mythology.
Biology
Leucism is a rare genetic pattern that causes a reduction in the pigment of an animal's hair and skin. The natural colour of the red deer ranges from dark red to brown. They are often thought to be albinos; however, unlike albinos, who have characteristically red eyes, deer with leucism have normal colouring in their eyes. It is distinguished biologically from albinism in that it causes a reduced pigmentation in all skin types, and not just melanin. There is also a black color morph of the white-tailed deer which exhibits melanism. While rare in their natural distribution, there is an incidence of up to 8.5% occurrence of black deer in certain regions of central Texas.
Symbolism
Folklore
White deer hold a place in the traditions of many cultures. They are considered to be messengers from the otherworld in some Celtic mythology; they also played an important role in other pre-Indo-European cultures, especially in the north. The Celts believed that the white stag would appear when one was transgressing a taboo, such as when Pwyll trespassed into Arawn's hunting grounds. In English folklore, the white hart is associated with Herne the Hunter.
Arthurian legend states that the creature has a perennial ability to evade capture, and that the pursuit of the animal represents mankind's spiritual quest. It also signalled that the time was high for the knights of the kingdom to pursue a quest.
In Catholicism and Eastern Orthodox, the white stag was partly responsible for the conversion of the martyr Saint Eustace. Eustace saw a vision of Christ between the stag's antlers and was told that he would suffer for Ch
|
https://en.wikipedia.org/wiki/2007%20Vietnam%20food%20scare
|
The 2007 Vietnam food scare was a food scandal which exposed contaminated food. Among the issues were formaldehyde in noodles of the national dish, phở, banned pesticides in vegetables and fruit, and toxic soy sauce.
|
https://en.wikipedia.org/wiki/Boolean%20circuit
|
In computational complexity theory and circuit complexity, a Boolean circuit is a mathematical model for combinational digital logic circuits. A formal language can be decided by a family of Boolean circuits, one circuit for each possible input length.
Boolean circuits are defined in terms of the logic gates they contain. For example, a circuit might contain binary AND and OR gates and unary NOT gates, or be entirely described by binary NAND gates. Each gate corresponds to some Boolean function that takes a fixed number of bits as input and outputs a single bit.
Boolean circuits provide a model for many digital components used in computer engineering, including multiplexers, adders, and arithmetic logic units, but they exclude sequential logic. They are an abstraction that omits many aspects relevant to designing real digital logic circuits, such as metastability, fanout, glitches, power consumption, and propagation delay variability.
Formal definition
In giving a formal definition of Boolean circuits, Vollmer starts by defining a basis as set B of Boolean functions, corresponding to the gates allowable in the circuit model. A Boolean circuit over a basis B, with n inputs and m outputs, is then defined as a finite directed acyclic graph. Each vertex corresponds to either a basis function or one of the inputs, and there is a set of exactly m nodes which are labeled as the outputs. The edges must also have some ordering, to distinguish between different arguments to the same Boolean function.
As a special case, a propositional formula or Boolean expression is a Boolean circuit with a single output node in which every other node has fan-out of 1. Thus, a Boolean circuit can be regarded as a generalization that allows shared subformulas and multiple outputs.
A common basis for Boolean circuits is the set {AND, OR, NOT}, which is functionally complete, i. e. from which all other Boolean functions can be constructed.
Computational complexity
Background
A partic
|
https://en.wikipedia.org/wiki/Pterygopalatine%20nerves
|
The two pterygopalatine nerves (or sphenopalatine branches) descend to the pterygopalatine ganglion.
Although it is closely related to the pterygopalatine ganglion, it is still considered a branch of the maxillary nerve and does not synapse in the ganglion.
It is found in the pterygopalatine fossa.
Additional images
|
https://en.wikipedia.org/wiki/Aditya%20Birla%20Payments%20Bank
|
Aditya Birla Payments Bank Limited (ABPB) was a payments bank started as a joint venture by Aditya Birla Nuvo Ltd. and Idea Cellular. Launched on 22 February 2018, it is the fourth payments bank to begin operations since issuance of licenses to 11 firms by the Reserve Bank of India in August 2015. Payments Banks are a special category of banks that can accept deposits of up to 2 lakh but cannot give loans or credit cards.
On 20 July 2019, Aditya Birla Payments Bank announced that it would be shutting down operations subject to the receipt of requisite regulatory consents and approval.
History
Aditya Birla Nuvo (now Grasim Industries Limited) was one of the 11 entities to receive an in-principle approval by Reserve Bank of India (RBI) to set-up payments banks in India, in August 2015. Post the in-principle approval, RBI had issued a license to Aditya Birla Payments Bank under Section 22 (1) of the Banking Regulation Act, 1949 to commence with the business of payments bank in April 2017.
Aditya Birla Payments Bank earlier operated as IMCSL (Idea Mobile Commerce Services Ltd) as a brand of Idea Cellular Ltd.
Aditya Birla Nuvo Limited holds 51 percent shares while the remaining 49 percent are with Idea Cellular.
Aditya Birla Payments Bank, which was a joint venture between Aditya Birla Nuvo and Idea Cellular, discontinued its banking business within 2 years of its inceptions due to lack of funds.
|
https://en.wikipedia.org/wiki/Checkout.com
|
Checkout.com, (with legal name of main entity as Checkout Ltd), is an international financial technology company which processes payments for other companies. Founded as Opus Payments in 2009, it is headquartered in London, United Kingdom. It had a valuation of $40 billion in 2022, making it the most valuable European fintech startup. Customers include Netflix, Pizza Hut, and digital asset exchange Coinbase.
History
Checkout.com was founded in 2009 by Swiss national Guillaume Pousaz in Singapore under the name Opus Payments, which processed payments for merchants in Hong Kong. The company became profitable in 2011 through a deal with Chinese tech gadget trading website Dealextreme. In 2012, Opus Payments was renamed Checkout.com in and registered in the U.K. In 2013, Checkout.com was granted membership with Visa and Mastercard, and Checkout.com subsequently focused on partnerships with Alipay and WeChat.
In 2018, the company joined London & Partners' Mayor's International Business Programme. In 2019, the company received a $230 million Series A funding round led by Insight Partners and DST Global. After additional funding rounds, the company's valuation had increased to $15 billion by June 2020. In May 2020, it acquired Australian company Pin Payments, allowing Checkout.com to expand into the Australian and New Zealand markets.
In January 2022, the company announced a $1 billion funding round, surpassing the value of competitors such as Revolut and Wise. Investors included the Qatar Investment Authority and Tiger Global Management, among others. The company announced it would use the capital to invest in Web3 applications. In May 2022, Checkout.com announced it was acquiring French startup Ubble which provides a remote identity verification service.
|
https://en.wikipedia.org/wiki/Fermat%20primality%20test
|
The Fermat primality test is a probabilistic test to determine whether a number is a probable prime.
Concept
Fermat's little theorem states that if p is prime and a is not divisible by p, then
If one wants to test whether p is prime, then we can pick random integers a not divisible by p and see whether the congruence holds. If it does not hold for a value of a, then p is composite. This congruence is unlikely to hold for a random a if p is composite. Therefore, if the equality does hold for one or more values of a, then we say that p is probably prime.
However, note that the above congruence holds trivially for , because the congruence relation is compatible with exponentiation. It also holds trivially for if p is odd, for the same reason. That is why one usually chooses a random a in the interval .
Any a such that
when n is composite is known as a Fermat liar. In this case n is called Fermat pseudoprime to base a.
If we do pick an a such that
then a is known as a Fermat witness for the compositeness of n.
Example
Suppose we wish to determine whether n = 221 is prime. Randomly pick 1 < a < 220, say a = 38. We check the above congruence and find that it holds:
Either 221 is prime, or 38 is a Fermat liar, so we take another a, say 24:
So 221 is composite and 38 was indeed a Fermat liar. Furthermore, 24 is a Fermat witness for the compositeness of 221.
Algorithm
The algorithm can be written as follows:
Inputs: n: a value to test for primality, n>3; k: a parameter that determines the number of times to test for primality
Output: composite if n is composite, otherwise probably prime
Repeat k times:
Pick a randomly in the range [2, n − 2]
If , then return composite
If composite is never returned: return probably prime
The a values 1 and n-1 are not used as the equality holds for all n and all odd n respectively, hence testing them adds no value.
Complexity
Using fast algorithms for modular exponentiation and multiprecision multiplication, the running ti
|
https://en.wikipedia.org/wiki/Square-integrable%20function
|
In mathematics, a square-integrable function, also called a quadratically integrable function or function or square-summable function, is a real- or complex-valued measurable function for which the integral of the square of the absolute value is finite. Thus, square-integrability on the real line is defined as follows.
One may also speak of quadratic integrability over bounded intervals such as for .
An equivalent definition is to say that the square of the function itself (rather than of its absolute value) is Lebesgue integrable. For this to be true, the integrals of the positive and negative portions of the real part must both be finite, as well as those for the imaginary part.
The vector space of (equivalence classes of) square integrable functions (with respect to Lebesgue measure) forms the space with Among the spaces, the class of square integrable functions is unique in being compatible with an inner product, which allows notions like angle and orthogonality to be defined. Along with this inner product, the square integrable functions form a Hilbert space, since all of the spaces are complete under their respective -norms.
Often the term is used not to refer to a specific function, but to equivalence classes of functions that are equal almost everywhere.
Properties
The square integrable functions (in the sense mentioned in which a "function" actually means an equivalence class of functions that are equal almost everywhere) form an inner product space with inner product given by
where
and are square integrable functions,
is the complex conjugate of
is the set over which one integrates—in the first definition (given in the introduction above), is , in the second, is .
Since , square integrability is the same as saying
It can be shown that square integrable functions form a complete metric space under the metric induced by the inner product defined above.
A complete metric space is also called a Cauchy space, because sequences in such
|
https://en.wikipedia.org/wiki/Interosseous%20membrane%20of%20leg
|
The interosseous membrane of the leg (middle tibiofibular ligament) extends between the interosseous crests of the tibia and fibula, helps stabilize the Tib-Fib relationship and separates the muscles on the front from those on the back of the leg.
It consists of a thin, aponeurotic joint lamina composed of oblique fibers, which for the most part run downward and lateralward; some few fibers, however, pass in the opposite direction.
It is broader above than below. Its upper margin does not quite reach the tibiofibular joint, but presents a free concave border, above which is a large, oval aperture for the passage of the anterior tibial vessels to the front of the leg.
In its lower part is an opening for the passage of the anterior peroneal vessels.
It is continuous below with the interosseous ligament of the tibiofibular syndesmosis, and presents numerous perforations for the passage of small vessels.
It is in relation, in front, with the Tibialis anterior, Extensor digitorum longus, Extensor hallucis proprius, Peronæus tertius, and the anterior tibial vessels and deep peroneal nerve; behind, with the Tibialis posterior and Flexor hallucis longus.
Additional images
|
https://en.wikipedia.org/wiki/Birgit%20Penzenstadler
|
Birgit Penzenstadler (born September 9, 1981 in Erding, Germany) is a German assistant professor of Software Engineering at Chalmers University of Technology and adjunct docent at Lappeenranta University of Technology. She is well known for her work on environmental sustainability in software engineering and for being one of the founders of the sustainability design initiative, which seeks to advance the research on sustainability in technical disciplines such as computer science and software engineering.
Work
Prior to Chalmers University of Technology, Birgit was a professor at California State University, Long Beach. Also she has completed a postdoctoral fellowship at the University of California, Irvine with Prof. Debra J. Richardson and Prof. Bill Tomlinson. They developed framework called SE4S that supports the infusion of sustainability in the requirements engineering (RE) and quality assurance (QA) stages of software engineering processes.
Penzenstadler coined the term "Software Engineering for Sustainability" in 2013. Also, she is the main organizer of the workshop series “Requirements Engineering for Sustainable Systems” since 2012 at the International Requirements Engineering Conferences. She led the Resilience Lab at California State University, Long Beach during 2015- 2019 which focused on research that evaluated the properties of a software system in relation to sustainability.
|
https://en.wikipedia.org/wiki/Tree-Puzzle
|
TREE-PUZZLE is a computer program used to construct phylogenetic trees from sequence data by maximum likelihood analysis. Branch lengths can be calculated with and without the molecular clock hypothesis.
The software also implemented likelihood mapping, a method to visualize phylogenetic information in datasets, as well as several tests to assess if the likelihoods of trees are significantly worse than those of other trees.
The program's successor is IQ-TREE.
See also
Computational phylogenetics
|
https://en.wikipedia.org/wiki/Hyena%20butter
|
Hyena butter is a secretion from the anal gland of hyenas used to mark territory and to identify individuals by odor. The gooey substance is spread onto objects within the territory of the hyena by rubbing their posterior against the object they mark.
Folk beliefs in some regions of East Africa state that witches would ride hyenas and use a gourd full of hyena butter as fuel for the torches that they carried through the night.
See also
Deer musk
Dog odor
Territorial marking
|
https://en.wikipedia.org/wiki/A-site
|
The A-site (A for aminoacyl) of a ribosome is a binding site for charged t-RNA molecules during protein synthesis. One of three such binding sites, the A-site is the first location the t-RNA binds during the protein synthesis process, the other two sites being P-site (peptidyl) and E-site (exit).
|
https://en.wikipedia.org/wiki/Kazakhstani%20tenge
|
The tenge ( or ; , ; sign: ₸ ; code: KZT) is the currency of Kazakhstan. It is divided into 100 tiyn ( also transliterated as tiyin).
History
After the breakup of the Soviet Union in December 1991, most of the formerly Soviet republics attempted to maintain a common currency. Some politicians hoped to at least maintain "special relations" among former Soviet republics (the "near abroad"). Other reasons were the economic considerations for maintaining the ruble zone. The wish to preserve strong trade relations between former Soviet republics was considered the most important goal.
The break-up of the Soviet Union was not accompanied by any formal changes in monetary arrangements. The Central Bank of Russia was authorized to take over the State Bank of the USSR (Gosbank) on 1 January 1992. It continued to ship Soviet notes and coins to the central banks of the eleven newly independent countries, which had formerly been the main branches of Gosbank in the republics.
The political situation, however, was not favourable for maintaining a common currency. Maintaining a common currency requires a strong political consensus in respect to monetary and fiscal targets, a common institution in charge of implementing these targets, and some minimum of common legislation (concerning the banking and foreign-exchange regulations). These conditions were far from being met amidst the turbulent economic and political situation.
During the first half of 1992, there existed a monetary union where 15 independent states all used the ruble. Since it was clear that the situation would not last, each of them was using its position as "free-riders" to issue huge amounts of money in the form of credit. As a result, some countries were issuing coupons in order to "protect" their markets from buyers from other states. The Russian central bank responded in July 1992 by setting up restrictions to the flow of credit between Russia and other states. The final collapse of the ruble zone began whe
|
https://en.wikipedia.org/wiki/Equatorial%20plasma%20bubble
|
Equatorial plasma bubbles are an ionospheric phenomenon near the Earth's geomagnetic equator at night time. They affect radio waves by causing varying delays. They degrade the performance of GPS.
Different times of the year and locations have different frequencies of occurrence. In Northern Australia, the most common times are February to April and August to October, when a plasma bubble is expected every night. Plasma bubbles have dimensions around 100 km. Plasma bubbles form after dark when the sun stops ionising the ionosphere. The ions recombine, forming a lower density layer. This layer can rise through the more ionized layers above via convection, which makes a plasma bubble. The bubbles are turbulent with irregular edges.
An equatorial plasma bubble could have affected the Battle of Shah-i-Kot by disabling communications from a communications satellite to a helicopter.
|
https://en.wikipedia.org/wiki/N%C3%A9ron%E2%80%93Ogg%E2%80%93Shafarevich%20criterion
|
In mathematics, the Néron–Ogg–Shafarevich criterion states that if A is an elliptic curve or abelian variety over a local field K and ℓ is a prime not dividing the characteristic of the residue field of K then A has good reduction if and only if the ℓ-adic Tate module Tℓ of A is unramified. introduced the criterion for elliptic curves. used the results of to extend it to abelian varieties,
and named the criterion after Ogg, Néron and Igor Shafarevich (commenting that Ogg's result seems to have been known to Shafarevich).
|
https://en.wikipedia.org/wiki/Gears%20of%20War%204
|
Gears of War 4 is a 2016 third-person shooter video game developed by The Coalition and published by Microsoft Studios for Windows and Xbox One. It is the fourth main installment in the Gears of War series, and the first mainline entry not to be developed by Epic Games. The game was released worldwide on October 11, 2016. The sequel, Gears 5, was released on September 10, 2019.
Upon release, the game received generally positive reviews, with critics praising the game's combat, visuals and set pieces, while criticism was directed towards the campaign and writing.
The game was also the first Gears of War title since the original to be released for Windows.
Gameplay
Many gameplay elements from the previous games are reintroduced, along with new elements such as the Dropshot weapon that fires explosive drills and the Buzzkill that fires ricocheting sawblades at enemies. Players can also perform a short distance shoulder charge, knocking enemies off balance and enemies can also pull players out of cover and perform their own takedowns. The game features 4 categories of weather: from heavy breeze (trees, leaves, dust kicking up) and can grow into violent category 3 windstorms that can affect combat and weapon usage.
Like previous entries in the series, the story campaign can be played through co-operatively with a second player in a local or online co-operative mode.
Story
Setting
Gears of War 4 takes place 25 years after the events of Gears of War 3, where the Imulsion Countermeasure weapon destroyed all Imulsion on the planet Sera, killing the Locust and the Lambent in the process, the weapon forced humanity to adapt to new ways of survival after knocking out the use of fossil fuels, and bringing across the planet powerful windstorms called "Windflares". To protect the surviving human population from decline, estimated to be hundreds of thousands, the reformed Coalition of Ordered Governments (COG) established walled-off cities to protect its citizens from the d
|
https://en.wikipedia.org/wiki/Two-dimensional%20quantum%20turbulence
|
Quantum turbulence involves the chaotic dynamics of many interacting quantum vortices. In highly excited bulk superfluid, many vortex lines interact with each other forming quantum turbulent states. By introducing tight confinement along one direction the Kelvin wave excitations can be strongly suppressed, favouring vortex alignment with the axis of tight confinement. Vortex dynamics enters a regime of effective 2D motion, equivalent to point vortices moving on a plane. In general, 2D quantum turbulence (2DQT) can exhibit complex phenomenology involving coupled vortices and sound in compressible superfluids. The quantum vortex dynamics can exhibit signatures of turbulence including a Kolmogorov −5/3 power law, a quantum manifestation of the inertial transport of energy to large scales observed in classical fluids, known as an inverse energy cascade.
Point vortices
The point vortex model, introduced by Helmholtz and Kirchhoff, describes the motion of ideal point vortices confined to a plane, with direct mapping to planar electrodynamics. The model plays a central role in the study of planar Navier-Stokes flows, and can be realized in compressible superfluids such as those in ultracold gas Bose-Einstein condensates, when the healing length setting the vortex core size is very small compared to the system size.
Negative temperature
Point vortices confined to finite area were predicted by Onsager to exhibit states of negative temperature. This possibility of negative absolute temperature can be traced to the finite phase space of the point vortex system: in contrast to a massive particle moving on a plane, each point vortex only has two degrees of freedom. Specifying the spatial coordinates of the vortex also completely determines the superfluid velocity. At leading order a quantum vortex is massless, with each filament moving with the net background flow and obeying a form of the Biot–Savart law. Guiding-centre plasmas exhibit a symmetry breaking transition at high e
|
https://en.wikipedia.org/wiki/Optical%20square
|
The optical square uses a pentaprism to reflect and refract a beam or sighting 90 degrees, it is used in pairs in surveying and in a singular block in metrology.
In optical square
A Horizon glass is placed at an angle of 120° with the horizon sight.
The Index glass is placed at an angle of 105° with the Index sight.
Angle between Index glass and Horizon glass is 45°.
Metrology
Used with an autocollimator or angle dekkor and mirror it can be used for machine tool axis squareness checking and for measuring the squareness of surfaces. It has two mirrors at 45 degree to each other. One is half-silvered, called horizon glass, and other is fully silvered, called index glass. It measures angle by reflection. Two prisms can be used as an optical square.
Optical square in surveying
In surveying it is used both as a hand held tool for sighting between two poles (often with a plumb bob hung from the handle) and also mounted on a Jacob's staff.
|
https://en.wikipedia.org/wiki/Sequence%20labeling
|
In machine learning, sequence labeling is a type of pattern recognition task that involves the algorithmic assignment of a categorical label to each member of a sequence of observed values. A common example of a sequence labeling task is part of speech tagging, which seeks to assign a part of speech to each word in an input sentence or document. Sequence labeling can be treated as a set of independent classification tasks, one per member of the sequence. However, accuracy is generally improved by making the optimal label for a given element dependent on the choices of nearby elements, using special algorithms to choose the globally best set of labels for the entire sequence at once.
As an example of why finding the globally best label sequence might produce better results than labeling one item at a time, consider the part-of-speech tagging task just described. Frequently, many words are members of multiple parts of speech, and the correct label of such a word can often be deduced from the correct label of the word to the immediate left or right. For example, the word "sets" can be either a noun or verb. In a phrase like "he sets the books down", the word "he" is unambiguously a pronoun, and "the" unambiguously a determiner, and using either of these labels, "sets" can be deduced to be a verb, since nouns very rarely follow pronouns and are less likely to precede determiners than verbs are. But in other cases, only one of the adjacent words is similarly helpful. In "he sets and then knocks over the table", only the word "he" to the left is helpful (cf. "...picks up the sets and then knocks over..."). Conversely, in "... and also sets the table" only the word "the" to the right is helpful (cf. "... and also sets of books were ..."). An algorithm that proceeds from left to right, labeling one word at a time, can only use the tags of left-adjacent words and might fail in the second example above; vice versa for an algorithm that proceeds from right to left.
|
https://en.wikipedia.org/wiki/Metal%20Gear%20Solid%202%3A%20Sons%20of%20Liberty
|
is a 2001 action-adventure stealth video game developed by Konami Computer Entertainment Japan and published by Konami for the PlayStation 2. Originally released on November 13, 2001, it is the fourth Metal Gear game produced by Hideo Kojima, the seventh overall game in the series and is a sequel to Metal Gear Solid (1998). An expanded edition, titled was released the following year for Xbox and Windows in addition to the PlayStation 2. A remastered version of the game, Metal Gear Solid 2: Sons of Liberty - HD Edition, was later included in the Metal Gear Solid HD Collection for the PlayStation 3, Xbox 360, and PlayStation Vita. The HD Edition of the game was included in the Metal Gear Solid Master Collection Vol. 1 compilation for Nintendo Switch, PlayStation 4, PlayStation 5, Windows, and Xbox Series X/S, which was released on October 24, 2023.
The story revolves around the Big Shell, a massive offshore clean-up facility seized by a group of terrorists who call themselves the Sons of Liberty. They demand an enormous ransom in exchange for the life of the President of the United States and threaten to destroy the facility and create a cataclysmic environmental disaster if their demands are not met. The motives and identities of many of the antagonists and allies change throughout the game, as the protagonists discover a world-shaking conspiracy constructed by a powerful organization known as the Patriots.
Metal Gear Solid 2 received acclaim for its gameplay, graphics, and attention to detail. However, critics were initially divided on the protagonist and the philosophical nature and execution of the game's storyline, which explores themes such as the Information Age, memetics, social engineering, political conspiracies, censorship, artificial intelligence, existentialism, postmodernism, virtual reality, and the internal struggle of freedom of thought. The game was a commercial success, selling seven million copies by 2004. It has since been considered to be one
|
https://en.wikipedia.org/wiki/Exeter%20Mathematics%20School
|
Exeter Mathematics School is a maths school located in Exeter in the English county of Devon.
It opened in September 2014 under the free schools initiative and is sponsored by Exeter College and the University of Exeter. It is intended to be a regional centre of excellence in mathematics for Cornwall, Devon, Dorset and Somerset. As a result, the school offers boarding facilities for pupils who live more than an hours drive away from the school. A total of 120 students are catered for at the school with some boarding from Monday to Friday during term time.
The school is highly selective, with prospective students expected to have GCSE qualifications at grade 8-9 in Mathematics and Physics or Computer Science. Prospective students must also have five GCSEs in total at grade 5 or above including English at grade 6. The course structure of Exeter Mathematics School requires all students to study A-level Mathematics and Further Mathematics and either A-level Physics or Computer Science. Students may choose to study both, but one may be chosen and an additional A-level from a wider range of options, which are taught at Exeter College, may be taken as an alternative.
|
https://en.wikipedia.org/wiki/Liquid-hydrogen%20tanktainer
|
A liquid hydrogen tank-tainer also known as a liquid hydrogen tank container is a specialized type of container designed to carry cryogenic liquid hydrogen (LH2) on standard intermodal equipment. The tank is held within a box-shaped frame the same size and shape as a container.
Size and volume
Liquid hydrogen tanktainers are referenced by their size or volume capacity, generally an ISO container.
See also
Compressed hydrogen tube trailer
Containerization
Hydrogen economy
Hydrogen infrastructure
Liquid hydrogen tank car
Liquid hydrogen trailer
Tank chassis
Tank container
|
https://en.wikipedia.org/wiki/Helium%20atom%20scattering
|
Helium atom scattering (HAS) is a surface analysis technique used in materials science. It provides information about the surface structure and lattice dynamics of a material by measuring the diffracted atoms from a monochromatic helium beam incident on the sample.
History
The first recorded helium diffraction experiment was completed in 1930 by Immanuel Estermann and Otto Stern on the (100) crystal face of lithium fluoride. This experimentally established the feasibility of atom diffraction when the de Broglie wavelength, λ, of the impinging atoms is on the order of the interatomic spacing of the material. At the time, the major limit to the experimental resolution of this method was due to the large velocity spread of the helium beam. It wasn't until the development of high pressure nozzle sources capable of producing intense and strongly monochromatic beams in the 1970s that HAS gained popularity for probing surface structure. Interest in studying the collision of rarefied gases with solid surfaces was helped by a connection with aeronautics and space problems of the time. Plenty of studies showing the fine structures in the diffraction pattern of materials using helium atom scattering were published in the 1970s. However, it wasn't until a third generation of nozzle beam sources was developed, around 1980, that studies of surface phonons could be made by helium atom scattering. These nozzle beam sources were capable of producing helium atom beams with an energy resolution of less than 1meV, making it possible to explicitly resolve the very small energy changes resulting from the inelastic collision of a helium atom with the vibrational modes of a solid surface, so HAS could now be used to probe lattice dynamics. The first measurement of such a surface phonon dispersion curve was reported in 1981, leading to a renewed interest in helium atom scattering applications, particularly for the study of surface dynamics.
Basic principles
Surface sensitivity
|
https://en.wikipedia.org/wiki/Isogeometric%20analysis
|
Isogeometric analysis is a computational approach that offers the possibility of integrating finite element analysis (FEA) into conventional NURBS-based CAD design tools. Currently, it is necessary to convert data between CAD and FEA packages to analyse new designs during development, a difficult task since the two computational geometric approaches are different. Isogeometric analysis employs complex NURBS geometry (the basis of most CAD packages) in the FEA application directly. This allows models to be designed, tested and adjusted in one go, using a common data set.
The pioneers of this technique are Tom Hughes and his group at The University of Texas at Austin. A reference free software implementation of some isogeometric analysis methods is GeoPDEs. Likewise, other implementations can be found online. For instance, PetIGA is an open framework for high performance isogeometric analysis heavily based on PETSc. In addition, MIGFEM is another IGA code which is implemented in Matlab and supports Partition of Unity enrichment IGA for 2D and 3D fracture. Furthermore, G+Smo is an open C++ library for isogeometric analysis. In particular, FEAP is a finite element analysis program which includes an Isogeometric analysis library FEAP IsoGeometric (Version FEAP84 & Version FEAP85). An account of the developments leading up to IGA has been documented in.
Advantages of IGA with respect to FEA
Isogeometric analysis presents two main advantages with respect to the finite element method:
There is no geometric approximation error, due to the fact the domain is represented exactly
Wave propagation problems, arising for instance in cardiac electrophysiology, acoustics and elastodynamics, are better described, thanks to the reduction of numerical dispersion and dissipation errors.
Meshes
In the framework of IGA, the notions of both control mesh and physical mesh are defined.
A control mesh is made by the so called control points and it is obtained by a piecewise linear interp
|
https://en.wikipedia.org/wiki/Spark%20%28mathematics%29
|
In mathematics, more specifically in linear algebra, the spark of a matrix is the smallest integer such that there exists a set of columns in which are linearly dependent. If all the columns are linearly independent, is usually defined to be 1 more than the number of rows. The concept of matrix spark finds applications in error-correction codes, compressive sensing, and matroid theory, and provides a simple criterion for maximal sparsity of solutions to a system of linear equations.
The spark of a matrix is NP-hard to compute.
Definition
Formally, the spark of a matrix is defined as follows:
where is a nonzero vector and denotes its number of nonzero coefficients ( is also referred to as the size of the support of a vector). Equivalently, the spark of a matrix is the size of its smallest circuit (a subset of column indices such that has a nonzero solution, but every subset of it does not).
If all the columns are linearly independent, is usually defined to be (if has m rows).
By contrast, the rank of a matrix is the largest number such that some set of columns of is linearly independent.
Example
Consider the following matrix .
The spark of this matrix equals 3 because:
There is no set of 1 column of which are linearly dependent.
There is no set of 2 columns of which are linearly dependent.
But there is a set of 3 columns of which are linearly dependent. The first three columns are linearly dependent because .
Properties
If , the following simple properties hold for the spark of a matrix :
(If the spark equals , then the matrix has full rank.)
if and only if the matrix has a zero column.
.
Criterion for uniqueness of sparse solutions
The spark yields a simple criterion for uniqueness of sparse solutions of linear equation systems.
Given a linear equation system . If this system has a solution that satisfies , then this solution is the sparsest possible solution. Here denotes the number of nonzero entries of the vector .
Lower bo
|
https://en.wikipedia.org/wiki/Solaris%20Containers
|
Solaris Containers (including Solaris Zones) is an implementation of operating system-level virtualization technology for x86 and SPARC systems, first released publicly in February 2004 in build 51 beta of Solaris 10, and subsequently in the first full release of Solaris 10, 2005. It is present in illumos (formerly OpenSolaris) distributions, such as OpenIndiana, SmartOS, Tribblix and OmniOS, as well as in the official Oracle Solaris 11 release.
A Solaris Container is the combination of system resource controls and the boundary separation provided by zones. Zones act as completely isolated virtual servers within a single operating system instance. By consolidating multiple sets of application services onto one system and by placing each into isolated virtual server containers, system administrators can reduce cost and provide most of the same protections of separate machines on a single machine.
Terminology
The name of this technology changed during development and the pre-launch public events. Before the launch of Solaris Zones in 2005, a Solaris Container was any type of workload constrained by Solaris resource management features. The latter had been a separate software package in earlier history. By 2007 the term Solaris Containers came to mean a Solaris Zone combined with resource management controls.
Later, there was a gradual move such that Solaris Containers specifically referred to non-global zones, with or without additional Resource Management. Zones hosted by a global zone are known as "non-global zones" but are sometimes just called "zones". The term "local zone" is specifically discouraged, since in this usage "local" is not an antonym of "global". The global zone has visibility of all resource on the system, whether these are associated with the global zone or a non-global zone. Unless otherwise noted, "zone" will refer to non-global zones in this article.
To simplify terminology, Oracle dropped the use of the term Container in Solaris 11, and has
|
https://en.wikipedia.org/wiki/Saltatory%20conduction
|
In neuroscience, saltatory conduction () is the propagation of action potentials along myelinated axons from one node of Ranvier to the next node, increasing the conduction velocity of action potentials. The uninsulated nodes of Ranvier are the only places along the axon where ions are exchanged across the axon membrane, regenerating the action potential between regions of the axon that are insulated by myelin, unlike electrical conduction in a simple circuit.
Mechanism
Myelinated axons only allow action potentials to occur at the unmyelinated nodes of Ranvier that occur between the myelinated internodes. It is by this restriction that saltatory conduction propagates an action potential along the axon of a neuron at rates significantly higher than would be possible in unmyelinated axons (150 m/s compared to 0.5 to 10 m/s). As sodium rushes into the node it creates an electrical force which pushes on the ions already inside the axon. This rapid conduction of electrical signal reaches the next node and creates another action potential, thus refreshing the signal. In this manner, saltatory conduction allows electrical nerve signals to be propagated long distances at high rates without any degradation of the signal. Although the action potential appears to jump along the axon, this phenomenon is actually just the rapid conduction of the signal inside the myelinated portion of the axon.
If the entire surface of an axon were insulated, action potentials could not be regenerated along the axon resulting in signal degradation.
Energy efficiency
In addition to increasing the speed of the nerve impulse, the myelin sheath helps in reducing energy expenditure over the axon membrane as a whole, because the amount of sodium and potassium ions that need to be pumped to bring the concentrations back to the resting state following each action potential is decreased.
Distribution
Saltatory conduction occurs widely in the myelinated nerve fibers of vertebrates, but was later disc
|
https://en.wikipedia.org/wiki/Blue%20Plate%20Building
|
The Blue Plate Building, is a building in the Gert Town section of New Orleans, Louisiana, at 1315 S. Norman C. Francis Parkway at the corner of Earhart Boulevard. It was listed on the National Register of Historic Places on October 16, 2008. Its NRIS reference number is 08000989.
Built in 1941 in the Streamline Moderne style, it was originally a mayonnaise factory. It was designed by architect August Perez Jr in the Streamline Moderne style, a close cousin of Art Deco known for its curving forms, long horizontal lines and "streamlined" appearance.
In 2012 it was redeveloped by HRI Properties into 72 loft-style apartments with a designed leasing preference for artists, and was renamed "Blue Plate Artists Lofts."
|
https://en.wikipedia.org/wiki/Mackey%E2%80%93Arens%20theorem
|
The Mackey–Arens theorem is an important theorem in functional analysis that characterizes those locally convex vector topologies that have some given space of linear functionals as their continuous dual space.
According to Narici (2011), this profound result is central to duality theory; a theory that is "the central part of the modern theory of topological vector spaces."
Prerequisites
Let be a vector space and let be a vector subspace of the algebraic dual of that separates points on .
If is any other locally convex Hausdorff topological vector space topology on , then we say that is compatible with duality between and if when is equipped with , then it has as its continuous dual space.
If we give the weak topology then is a Hausdorff locally convex topological vector space (TVS) and is compatible with duality between and (i.e. ).
We can now ask the question: what are all of the locally convex Hausdorff TVS topologies that we can place on that are compatible with duality between and ?
The answer to this question is called the Mackey–Arens theorem.
Mackey–Arens theorem
See also
Dual system
Mackey topology
Polar topology
|
https://en.wikipedia.org/wiki/IAPWS
|
The International Association for the Properties of Water and Steam (IAPWS) is an international non-profit association of national organizations concerned with the properties of water and steam, particularly thermophysical properties and other aspects of high-temperature steam, water and aqueous mixtures that are relevant to thermal power cycles and other industrial applications.
The organization publishes a range of 'releases.' Specifically, these relate to the thermal and expansion properties of steam.
Both free software and commercial software implementations of the IAPWS correlations are available.
Thermodynamics
|
https://en.wikipedia.org/wiki/Computational%20and%20Mathematical%20Organization%20Theory
|
Computational and Mathematical Organization Theory is a quarterly double-blind peer-reviewed scientific journal covering the field of organization theory. The journal is published by Springer Science+Business Media. It was established in 1995 and initially published by Kluwer. The founding editors-in-chief were William A. Wallace (Rensselaer Polytechnic Institute) and Kathleen Carley (Carnegie Mellon University). Carley has continued as co-editor-in-chief, a role she currently shares with Terrill L. Frantz (Harrisburg University of Science and Technology).
Abstracting and indexing
Computational and Mathematical Organization Theory is abstracted and indexed in ACM Digital Library, CSA databases, Current Contents, Current Index to Statistics, Digital Bibliography & Library Project, EBSCO databases, Engineered Materials Abstracts, Inspec, InfoTrac, International Bibliography of the Social Sciences, ProQuest, Science Citation Index, Scopus, Social Sciences Citation Index, VINITI Database RAS, and Zentralblatt MATH. According to the Journal Citation Reports, the journal has a 2020 impact factor of 2.023.
|
https://en.wikipedia.org/wiki/Nutcracker%20doll
|
Nutcracker dolls, also known as Christmas nutcrackers, are decorative nutcracker figurines most commonly made to resemble a toy soldier. In German tradition, the dolls are symbols of good luck, frightening away malevolent spirits.
While nearly all nutcrackers from before the first half of the 20th century are functional, a significant proportion of modern nutcrackers are primarily decorative, and not able to crack nuts. Nutcrackers are also a part of German folklore, serving as protectors of a house.
History
Nutcracker dolls originate from late-17th century Germany, particularly the Ore Mountains (German: Erzgebirge) region. One origin story attributes the creation of the first nutcracker doll to a craftsman from Seiffen. They were often given as gifts, and at some point they became associated with Christmas season. They grew in popularity around the 19th century and spread to nearby European countries. As the demand grew, nutcracker doll production also began on a mass scale in factories. (1844–1923), commonly known in Germany as "father of the nutcracker", began the first mass production of the design (using a lathe) at his workshop in Seiffen in Saxony during 1872.
Decorative Nutcracker dolls began being popularized outside of Europe after the Second World War, when numerous American soldiers stationed in Germany came home to the United States with German nutcrackers as souvenirs. Further popularization came from Pyotr Ilyich Tchaikovsky's The Nutcracker, an 1892 ballet adaptation of E. T. A. Hoffmann's 1816 story "The Nutcracker and the Mouse King", which featured a toy soldier nutcracker. The ballet, introduced to America during the mid-20th century, became a favorite holiday tradition across the United States and helped make nutcracker dolls a Christmas decoration and a seasonal icon across Western culture.
Design
An average handcrafted nutcracker doll is made out of about 60 separate pieces. Nutcracker dolls traditionally resemble toy soldiers, and are
|
https://en.wikipedia.org/wiki/Ornamental%20bulbous%20plant
|
Ornamental bulbous plants, often called ornamental bulbs or just bulbs in gardening and horticulture, are herbaceous perennials grown for ornamental purposes, which have underground or near ground storage organs. Botanists distinguish between true bulbs, corms, rhizomes, tubers and tuberous roots, any of which may be termed "bulbs" in horticulture. Bulb species usually lose their upper parts during adverse conditions such as summer drought and heat or winter cold. The bulb's storage organs contain moisture and nutrients that are used to survive these adverse conditions in a dormant state. When conditions become favourable the reserves sustain a new growth cycle. In addition, bulbs permit vegetative or asexual multiplication in these species. Ornamental bulbs are used in parks and gardens and as cut flowers.
Terminology
The word "bulb" has a somewhat different meaning to botanists than it does to gardeners and horticulturalists. In gardening, a "bulb" is a plant's underground or ground-level storage organ that can be dried, stored and sold in this state, and then planted to grow again. Many bulbs in this sense are produced by geophytes – plants whose growing point is below ground level. However, not all bulbs in the gardening sense are produced by geophytes. For example, rhizomatous irises are included in books on ornamental bulbs, but their growing points are above ground. Many bulbs are produced by lilioid monocots, but not all lilioid monocots have bulbs. Brian Mathew says that "we just have to accept that there is no accurate term which we can use for this group of plants and we are left with 'bulbs' as the snappiest and most convenient."
Botanically, gardeners' "bulbs" may be true bulbs, corms, rhizomes or tubers, or combinations of these.
Types
True bulbs
A true bulb (i.e. a bulb in the botanical sense) is an underground vertical shoot that has modified leaves (or thickened leaf bases) that are used as food storage organs by the plant. The bottom of the b
|
https://en.wikipedia.org/wiki/Fiske%20Planetarium
|
Fiske Planetarium and Science Center is a public museum dedicated to educating the public on astronomy and astrophysics. It is a constituent of the Department of Astrophysical and Planetary Sciences at the University of Colorado Boulder, located in Boulder, Colorado. Fiske was founded in 1975, by a donation from Wallace Franz Fiske. It is located at the Southeast tip of the University of Colorado Boulder.
In 2013, Fiske underwent a major upgrade where the facility retired their 38-year-old, Zeiss Mark VI Star projector to a Megastar Star Ball. They also installed a 6-projector system capable of projecting fulldome 8K films and live renderings of a digital universe.
|
https://en.wikipedia.org/wiki/Matrix%20isolation
|
Matrix isolation is an experimental technique used in chemistry and physics. It generally involves a material being trapped within an unreactive matrix. A host matrix is a continuous solid phase in which guest particles (atoms, molecules, ions, etc.) are embedded. The guest is said to be isolated within the host matrix. Initially the term matrix-isolation was used to describe the placing of a chemical species in any unreactive material, often polymers or resins, but more recently has referred specifically to gases in low-temperature solids. A typical matrix isolation experiment involves a guest sample being diluted in the gas phase with the host material, usually a noble gas or nitrogen. This mixture is then deposited on a window that is cooled to below the melting point of the host gas. The sample may then be studied using various spectroscopic procedures.
Experimental setup
The transparent window, on to which the sample is deposited, is usually cooled using a compressed helium or similar refrigerant. Experiments must be performed under a high vacuum to prevent contaminants from unwanted gases freezing to the cold window. Lower temperatures are preferred, due to the improved rigidity and "glassiness" of the matrix material. Noble gases such as argon are used not just because of their unreactivity but also because of their broad optical transparency in the solid state. Mono-atomic gases have relatively simple face-centered cubic (fcc) crystal structure, which can make interpretations of the site occupancy and crystal-field splitting of the guest easier. In some cases a reactive material, for example, methane, hydrogen or ammonia, may be used as the host material so that the reaction of the host with the guest species may be studied.
Using the matrix isolation technique, short-lived, highly-reactive species such as radical ions and reaction intermediates may be observed and identified by spectroscopic means. For example, the solid noble gas krypton can be used to
|
https://en.wikipedia.org/wiki/Mathematical%20problem
|
A mathematical problem is a problem that can be represented, analyzed, and possibly solved, with the methods of mathematics. This can be a real-world problem, such as computing the orbits of the planets in the solar system, or a problem of a more abstract nature, such as Hilbert's problems. It can also be a problem referring to the nature of mathematics itself, such as Russell's Paradox.
Real-world problems
Informal "real-world" mathematical problems are questions related to a concrete setting, such as "Adam has five apples and gives John three. How many has he left?". Such questions are usually more difficult to solve than regular mathematical exercises like "5 − 3", even if one knows the mathematics required to solve the problem. Known as word problems, they are used in mathematics education to teach students to connect real-world situations to the abstract language of mathematics.
In general, to use mathematics for solving a real-world problem, the first step is to construct a mathematical model of the problem. This involves abstraction from the details of the problem, and the modeller has to be careful not to lose essential aspects in translating the original problem into a mathematical one. After the problem has been solved in the world of mathematics, the solution must be translated back into the context of the original problem.
Abstract problems
Abstract mathematical problems arise in all fields of mathematics. While mathematicians usually study them for their own sake, by doing so, results may be obtained that find application outside the realm of mathematics. Theoretical physics has historically been a rich source of inspiration.
Some abstract problems have been rigorously proved to be unsolvable, such as squaring the circle and trisecting the angle using only the compass and straightedge constructions of classical geometry, and solving the general quintic equation algebraically. Also provably unsolvable are so-called undecidable problems, such as the
|
https://en.wikipedia.org/wiki/Hypergeusia
|
Hypergeusia is a taste disorder where the sense is abnormally heightened. It can be associated with a lesion of the posterior fossa and Addison's disease; where a patient will crave for salty and sour taste due to the abnormal loss of ions with urine.
|
https://en.wikipedia.org/wiki/List%20of%20definite%20integrals
|
In mathematics, the definite integral
is the area of the region in the xy-plane bounded by the graph of f, the x-axis, and the lines x = a and x = b, such that area above the x-axis adds to the total, and that below the x-axis subtracts from the total.
The fundamental theorem of calculus establishes the relationship between indefinite and definite integrals and introduces a technique for evaluating definite integrals.
If the interval is infinite the definite integral is called an improper integral and defined by using appropriate limiting procedures. for example:
A constant, such pi, that may be defined by the integral of an algebraic function over an algebraic domain is known as a period.
The following is a list of some of the most common or interesting definite integrals. For a list of indefinite integrals see List of indefinite integrals.
Definite integrals involving rational or irrational expressions
Definite integrals involving trigonometric functions
(see Dirichlet integral)
Definite integrals involving exponential functions
(see also Gamma function)
(the Gaussian integral)
(where !! is the double factorial)
(where is Euler–Mascheroni constant)
Definite integrals involving logarithmic functions
Definite integrals involving hyperbolic functions
Frullani integrals
holds if the integral exists and is continuous.
See also
List of integrals
Indefinite sum
Gamma function
List of limits
|
https://en.wikipedia.org/wiki/Falcon%20Northwest
|
Falcon Northwest is a private company headquartered in Medford, Oregon. It designs, assembles, and markets high-end custom computers. The company was founded in 1992 and was one of the first to specialize in PCs built specifically for gaming.
History
Falcon Northwest was founded in April 1992 by gamer hobbyist and former pilot Kelt Reeves. Falcon Northwest released the first pre-built computer model intended specifically for gaming, the Mach V, in 1993, starting the "gaming PC" category of computer products. The company was founded in Florida, but later moved to Coos Bay, Oregon, then Ashland, Oregon, and finally Medford, Oregon.
In the late 1990s, Falcon grew to $3 million in annual revenues and opened a new manufacturing facility in Oregon. Later on, the company collaborated with Intel on early liquid cooling components. Intel worked with Falcon Northwest in secret, in order to avoid the appearance of endorsing overclocking by selling liquid cooling products under its own brand.
Products
Falcon Northwest sells high-end computers that are optimized for gaming, scientific, or military applications. As of 2013, about half of its sales were from gamers. Falcon's computers are consistently highly-ranked in benchmark tests, but cost $1,500 to over $10,000 depending on the user's configuration. Many Falcon PCs are sold with custom paint jobs, high-end graphics cards, and special low-latency components. Though it was originally known for tower desktops like the Mach V, and also sells laptops, as of 2017 Falcon is best-known for its smaller, portable mini-PCs.
Their products include:
Mach V - Desktop tower PC
Talon - Desktop tower PC
FragBox - Small Form Factor (SFF) PC
Tiki - Micro-tower PC
TLX - Thin & light class laptop PC
DRX - Desktop replacement class laptop PC
Reception
In benchmark tests by Maximum PC in 2018, Falcon Northwest's Tiki mini-PC performed better than a tower computer with a high-end graphics card, but was also the most expensive computer the rev
|
https://en.wikipedia.org/wiki/Open%20%28system%20call%29
|
For most file systems, a program initializes access to a file in a file system using the open system call. This allocates resources associated to the file (the file descriptor), and returns a handle that the process will use to refer to that file. In some cases the open is performed by the first access.
The same file may be opened simultaneously by several processes, and even by the same process, resulting in several file descriptors for the same file; depending on the file organization and filesystem. Operations on the descriptors such as moving the file pointer or closing it are independentthey do not affect other descriptors for the same file. Operations on the file, such as a write, can be seen by operations on the other descriptors: a later read can read the newly written data.
During the open, the filesystem may allocate memory for buffers, or it may wait until the first operation.
The absolute file path is resolved. This may include connecting to a remote host and notifying an operator that a removable medium is required. It may include the initialization of a communication device. At this point an error may be returned if the host or medium is not available. The first access to at least the directory within the filesystem is performed. An error will usually be returned if the higher level components of the path (directories) cannot be located or accessed. An error will be returned if the file is expected to exist and it does not or if the file should not already exist and it does.
If the file is expected to exist and it does, the file access, as restricted by permission flags within the file meta data or access control list, is validated against the requested type of operations. This usually requires an additional filesystem access although in some filesystems meta-flags may be part of the directory structure.
If the file is being created, the filesystem may allocate the default initial amount of storage or a specified amount depending on the file
|
https://en.wikipedia.org/wiki/Swim%20bladder
|
The swim bladder, gas bladder, fish maw, or air bladder is an internal gas-filled organ that contributes to the ability of many bony fish (but not cartilaginous fish) to control their buoyancy, and thus to stay at their current water depth without having to expend energy in swimming. Also, the dorsal position of the swim bladder means the center of mass is below the center of volume, allowing it to act as a stabilizing agent. Additionally, the swim bladder functions as a resonating chamber, to produce or receive sound.
The swim bladder is evolutionarily homologous to the lungs of tetrapods and lungfish. Charles Darwin remarked upon this in On the Origin of Species. Darwin reasoned that the lung in air-breathing vertebrates had derived from a more primitive swim bladder as a specialized form of enteral respiration.
In the embryonic stages, some species, such as redlip blenny, have lost the swim bladder again, mostly bottom dwellers like the weather fish. Other fish—like the opah and the pomfret—use their pectoral fins to swim and balance the weight of the head to keep a horizontal position. The normally bottom dwelling sea robin can use their pectoral fins to produce lift while swimming.
The gas/tissue interface at the swim bladder produces a strong reflection of sound, which is used in sonar equipment to find fish.
Cartilaginous fish, such as sharks and rays, do not have swim bladders. Some of them can control their depth only by swimming (using dynamic lift); others store fats or oils with density less than that of seawater to produce a neutral or near neutral buoyancy, which does not change with depth.
Structure and function
The swim bladder normally consists of two gas-filled sacs located in the dorsal portion of the fish, although in a few primitive species, there is only a single sac. It has flexible walls that contract or expand according to the ambient pressure. The walls of the bladder contain very few blood vessels and are lined with guanine crystals,
|
https://en.wikipedia.org/wiki/Boiled%20fish
|
Boiled fish, or more precisely salt-boiled fish, is fish boiled with salt and thus preserved for later consumption. Although this method is used in other parts of the world, it is of major commercial significance only in Southeast Asia. The shelf life of products so treated can range from as little as one or two days, up to several months. In Indonesia, this fish preservation method is known as pindang.
Preservation method
The technique works to preserve fish through both exposure to high temperatures and the high temperature of boiling water kills microbes that might otherwise decompose the fish flesh while the application of salt directly promotes preservation.
This technique is especially prevalent in the tropics during monsoon season, since the torrential rains hinder the simpler and traditional salting and sun-drying method of preservation. This salted fish method is considered 'dry preservation', while the Pindang method is often called 'wet preservation'.
After being covered in coarse salt, the fish are boiled on a low flame until the liquids are evaporated and the salt seasoning is well absorbed into fish. The wet boiling method requires less salt than dry preservation, and thus the taste is not as salty as that of sun-dried salted fish.
Although the basic ingredients often involve only fish, water, and salt, other ingredients, especially spices or herbs that contains tannin, can be added to boost preservation effectiveness. Examples of sources of tannin used include turmeric, tamarind, shallot skin, teak leaves, guava leaves, tea, and soy sauce, as well as other spices common in Southeast Asia. Including tannins gives the food a yellowish to brown color and fish so treated will last longer than fish preserved via the plain boiled method.
Regional variation
In Indonesia, various boiled fish products are generally known as pindang, and the method of preparation is often described as 'Indonesian salt-boiled fish'.
See also
Cured fish
Fish processing
Fi
|
https://en.wikipedia.org/wiki/Nikola%20Kalabi%C4%87
|
Nikola Kalabić (; 20 December 1906 – 19 January 1946) was a Serb and Yugoslav surveyor and Chetnik commander during World War II.
Early life
He was born to father Milan and mother Joka in Podnovlje (municipality of Doboj) in present-day Bosnia and Herzegovina. He had a sister named Angelina (1912–1999).
Nikola's father separated from Joka after World War I and proceeded to get married three times. Nikola lived with his father at first so he attended school in places where his father served with the Serbian army. He eventually finished six grades of gymnasium before becoming a student of geodesy in Belgrade. During his studies, he met Borka (a year younger than him) who was born in Rajkovići near Valjevo in present-day Serbia into a family of old supporters of the People's Radical Party and Nikola Pašić. Borka and Nikola married in 1929 and on 3 August 1930 they had twins Mirjana and Milan. Their first service was in Belgrade. They then moved to Aranđelovac and finally back to Valjevo (where there are Kalabićs today). Nikola Kalabić (until the start of World War II) worked in Land-registry management in Valjevo.
World War II
During World War II, Nikola Kalabić was commander of Draža Mihailović's formation named the Mountain Guard Corps (). On 26 November 1943, together with the General Inspector of Chetnik Troops Colonel Simić, he concluded a formal collaboration agreement () with the representative of the German Military Commander in Southeast Europe, General der Infanterie (Lieutenant General) Hans Felber.
In the night of 25th of December 1943, Chetniks under command of Nikola Kalabić killed a large number of civilians in village of Kopljare, near Aranđelovac. Out of 22 killed, 19 were Romani. In a letter to Draža Mihailović, Kalabić wrote: In Kopljere we caught while sleeping and slayed 24 active communists, out of which 20 were Gypsies, who admitted to be so-called jarugaši, at day they do their housework, but at night in action. I killed them all. There is n
|
https://en.wikipedia.org/wiki/Schwarz%20lemma
|
In mathematics, the Schwarz lemma, named after Hermann Amandus Schwarz, is a result in complex analysis about holomorphic functions from the open unit disk to itself. The lemma is less celebrated than deeper theorems, such as the Riemann mapping theorem, which it helps to prove. It is, however, one of the simplest results capturing the rigidity of holomorphic functions.
Statement
Let be the open unit disk in the complex plane centered at the origin, and let be a holomorphic map such that and on .
Then for all , and .
Moreover, if for some non-zero or , then for some with .
Proof
The proof is a straightforward application of the maximum modulus principle on the function
which is holomorphic on the whole of , including at the origin (because is differentiable at the origin and fixes zero). Now if denotes the closed disk of radius centered at the origin, then the maximum modulus principle implies that, for , given any , there exists on the boundary of such that
As we get .
Moreover, suppose that for some non-zero , or . Then, at some point of . So by the maximum modulus principle, is equal to a constant such that . Therefore, , as desired.
Schwarz–Pick theorem
A variant of the Schwarz lemma, known as the Schwarz–Pick theorem (after Georg Pick), characterizes the analytic automorphisms of the unit disc, i.e. bijective holomorphic mappings of the unit disc to itself:
Let be holomorphic. Then, for all ,
and, for all ,
The expression
is the distance of the points , in the Poincaré metric, i.e. the metric in the Poincaré disc model for hyperbolic geometry in dimension two. The Schwarz–Pick theorem then essentially states that a holomorphic map of the unit disk into itself decreases the distance of points in the Poincaré metric. If equality holds throughout in one of the two inequalities above (which is equivalent to saying that the holomorphic map preserves the distance in the Poincaré metric), then must be an analytic automorphism of th
|
https://en.wikipedia.org/wiki/Parallel%20all-pairs%20shortest%20path%20algorithm
|
A central problem in algorithmic graph theory is the shortest path problem. Hereby, the problem of finding the shortest path between every pair of nodes is known as all-pair-shortest-paths (APSP) problem. As sequential algorithms for this problem often yield long runtimes, parallelization has shown to be beneficial in this field. In this article two efficient algorithms solving this problem are introduced.
Another variation of the problem is the single-source-shortest-paths (SSSP) problem, which also has parallel approaches: Parallel single-source shortest path algorithm.
Problem definition
Let be a directed Graph with the set of nodes and the set of edges . Each edge has a weight assigned. The goal of the all-pair-shortest-paths problem is to find the shortest path between all pairs of nodes of the graph. For this path to be unique it is required that the graph does not contain cycles with a negative weight.
In the remainder of the article it is assumed that the graph is represented using an adjacency matrix. We expect the output of the algorithm to be a distancematrix . In , every entry is the weight of the shortest path in from node to node .
The Floyd algorithm presented later can handle negative edge weights, whereas the Dijkstra algorithm requires all edges to have a positive weight.
Dijkstra algorithm
The Dijkstra algorithm originally was proposed as a solver for the single-source-shortest-paths problem. However, the algorithm can easily be used for solving the All-Pair-Shortest-Paths problem by executing the Single-Source variant with each node in the role of the root node.
In pseudocode such an implementation could look as follows:
1 func DijkstraSSSP(G,v) {
2 ... //standard SSSP-implementation here
3 return dv;
4 }
5
6 func DijkstraAPSP(G) {
7 D := |V|x|V|-Matrix
8 for i from 1 to |V| {
9 //D[v] denotes the v-th row of D
10 D[v] := DijkstraSSP(G,i)
11 }
|
https://en.wikipedia.org/wiki/Direct%20coupling%20analysis
|
Direct coupling analysis or DCA is an umbrella term comprising several methods for analyzing sequence data in computational biology. The common idea of these methods is to use statistical modeling to quantify the strength of the direct relationship between two positions of a biological sequence, excluding effects from other positions. This contrasts usual measures of correlation, which can be large even if there is no direct relationship between the positions (hence the name direct coupling analysis). Such a direct relationship can for example be the evolutionary pressure for two positions to maintain mutual compatibility in the biomolecular structure of the sequence, leading to molecular coevolution between the two positions.
DCA has been used in the inference of protein residue contacts, RNA structure prediction, the inference of protein-protein interaction networks, the modeling of fitness landscapes, the generation of novel function proteins, and the modeling of protein evolution.
Mathematical Model and Inference
Mathematical Model
The basis of DCA is a statistical model for the variability within a set of phylogenetically related biological sequences. When fitted to a multiple sequence alignment (MSA) of sequences of length , the model defines a probability for all possible sequences of the same length. This probability can be interpreted as the probability that the sequence in question belongs to the same class of sequences as the ones in the MSA, for example the class of all protein sequences belonging to a specific protein family.
We denote a sequence by , with the being categorical variables representing the monomers of the sequence (if the sequences are for example aligned amino acid sequences of proteins of a protein family, the take as values any of the 20 standard amino acids). The probability of a sequence within a model is then defined as
where
are sets of real numbers representing the parameters of the model (more below)
is a normaliza
|
https://en.wikipedia.org/wiki/List%20of%20BSD%20operating%20systems
|
There are a number of Unix-like operating systems under active development, descended from the Berkeley Software Distribution (BSD) series of UNIX variants developed (originally by Bill Joy) at the University of California, Berkeley, Department of Electrical Engineering and Computer Science. there were four major BSD operating systems, and an increasing number of other OSs derived from these, that add or remove certain features but generally remain compatible with their originating OS—and so are not really forks of them. This is a list of those that have been active since 2014, and their websites.
FreeBSD-based
FreeBSD is a free Unix-like operating system descended from AT&T UNIX via the Berkeley Software Distribution (BSD). FreeBSD currently has more than 200 active developers and thousands of contributors. Other notable derivatives include DragonFly BSD, which was forked from FreeBSD 4.8, and Apple Inc.'s macOS, with its Darwin base including a large amount of code derived from FreeBSD.
Active
Discontinued
DragonFly BSD-based
NetBSD-based
NetBSD is a freely redistributable, open source version of the Unix-derivative Berkeley Software Distribution (BSD) computer operating system. It was the second open source BSD descendant to be formally released, after 386BSD, and continues to be actively developed. Noted for its portability and quality of design and implementation, it is often used in embedded systems and as a starting point for the porting of other operating systems to new computer architectures.
OpenBSD-based
OpenBSD is a Unix-like computer operating system descended from Berkeley Software Distribution (BSD), a Unix derivative developed at the University of California, Berkeley. It was forked from NetBSD in 1995. OpenBSD includes a number of security features absent or optional in other operating systems and has a tradition of developers auditing the source code for software bugs and security problems.
Historic BSD
BSD was originally derived from Unix
|
https://en.wikipedia.org/wiki/Composite%20epoxy%20material
|
Composite epoxy materials (CEM) are a group of composite materials typically made from woven glass fabric surfaces and non-woven glass core combined with epoxy synthetic resin. They are typically used in printed circuit boards.
There are different types of CEMs:
CEM-1 is low-cost, flame-retardant, cellulose-paper-based laminate with only one layer of woven glass fabric.
CEM-2 has cellulose paper core and woven glass fabric surface.
CEM-3 is very similar to the most commonly used PCB material, FR-4. Its color is white, and it is flame-retardant.
CEM-4 quite similar as CEM-3 but not flame-retardant.
CEM-5 (also called CRM-5) has polyester woven glass core.
See also
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.