source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Packet%20analyzer
A packet analyzer, also known as packet sniffer, protocol analyzer, or network analyzer, is a computer program or computer hardware such as a packet capture appliance that can analyze and log traffic that passes over a computer network or part of a network. Packet capture is the process of intercepting and logging traffic. As data streams flow across the network, the analyzer captures each packet and, if needed, decodes the packet's raw data, showing the values of various fields in the packet, and analyzes its content according to the appropriate RFC or other specifications. A packet analyzer used for intercepting traffic on wireless networks is known as a wireless analyzer or WiFi analyzer. While a packet analyzer can also be referred to as a network analyzer or protocol analyzer these terms can also have other meanings. Protocol analyzer can technically be a broader, more general class that includes packet analyzers/sniffers. However, the terms are frequently used interchangeably. Capabilities On wired shared-medium networks, such as Ethernet, Token Ring, and FDDI, depending on the network structure (hub or switch), it may be possible to capture all traffic on the network from a single machine. On modern networks, traffic can be captured using a network switch using port mirroring, which mirrors all packets that pass through designated ports of the switch to another port, if the switch supports port mirroring. A network tap is an even more reliable solution than to use a monitoring port since taps are less likely to drop packets during high traffic load. On wireless LANs, traffic can be captured on one channel at a time, or by using multiple adapters, on several channels simultaneously. On wired broadcast and wireless LANs, to capture unicast traffic between other machines, the network adapter capturing the traffic must be in promiscuous mode. On wireless LANs, even if the adapter is in promiscuous mode, packets not for the service set the adapter is configure
https://en.wikipedia.org/wiki/Write-once%20%28cache%20coherence%29
In cache coherency protocol literature, Write-Once was the first MESI protocol defined. It has the optimization of executing write-through on the first write and a write-back on all subsequent writes, reducing the overall bus traffic in consecutive writes to the computer memory. It was first described by James R. Goodman in (1983). Cache coherence protocols are an important issue in Symmetric multiprocessing systems, where each CPU maintains a cache of the memory. States In this protocol, each block in the local cache is in one of these four states: Invalid: This block has an incoherent copy of the memory. Valid: This block has a coherent copy of the memory. The data may be possibly shared, but its content is not modified. Reserved: The block is the only copy of the memory, but it is still coherent. No write-back is needed if the block is replaced. Dirty: The block is the only copy of the memory and it is incoherent. This copy was written one or more times. This is the only state that generates a write-back when the block is replaced in the cache. These states have exactly the same meanings as the four states of the MESI protocol (they are simply listed in reverse order), but this is a simplified form of it that avoids the Read for Ownership operation. Instead, all invalidation is done by writes to main memory. For any given pair of caches, the permitted states of a given cache line are as follows (abbreviated in the order above): Transitions The protocol follows some transition rules for each event: Read hit: The information is supplied by the current cache. No state change. Read miss: The data is read from main memory. The read is snooped by other caches; if any of them have the line in the Dirty state, the read is interrupted long enough to write the data back to memory before it is allowed to continue. Any copies in the Dirty or Reserved states are set to the Valid state. Write hit: If the information in the cache is in Dirty or Reserved state, the
https://en.wikipedia.org/wiki/Switching%20Kalman%20filter
The switching Kalman filtering (SKF) method is a variant of the Kalman filter. In its generalised form, it is often attributed to Kevin P. Murphy, but related switching state-space models have been in use. Applications Applications of the switching Kalman filter include: Brain–computer interfaces and neural decoding, real-time decoding for continuous neural-prosthetic control, and sensorimotor learning in humans. It also has application in econometrics, signal processing, tracking, computer vision, etc. It is an alternative to the Kalman filter when the system's state has a discrete component. The additional error when using a Kalman filter instead of a Switching Kalman filter may be quantified in terms of the switching system's parameters. For example, when an industrial plant has "multiple discrete modes of behaviour, each of which having a linear (Gaussian) dynamics". Model There are several variants of SKF discussed in. Special case In the simpler case, switching state-space models are defined based on a switching variable which evolves independent of the hidden variable. The probabilistic model of such variant of SKF is as the following: [This section is badly written: It does not explain the notation used below.] The hidden variables include not only the continuous , but also a discrete *switch* (or switching) variable . The dynamics of the switch variable are defined by the term . The probability model of and can depend on . The switch variable can take its values from a set . This changes the joint distribution which is a separate multivariate Gaussian distribution in case of each value of . General case In more generalised variants, the switch variable affects the dynamics of , e.g. through . The filtering and smoothing procedure for general cases is discussed in.
https://en.wikipedia.org/wiki/Meltwater%20pulse%201A
Meltwater pulse 1A (MWP1a) is the name used by Quaternary geologists, paleoclimatologists, and oceanographers for a period of rapid post-glacial sea level rise, between 13,500 and 14,700 calendar years ago, during which the global sea level rose between and in about 400–500 years, giving mean rates of roughly /yr. Meltwater pulse 1A is also known as catastrophic rise event 1 (CRE1) in the Caribbean Sea. The rates of sea level rise associated with meltwater pulse 1A are the highest known rates of post-glacial, eustatic sea level rise. Meltwater pulse 1A is also the most widely recognized and least disputed of the named, postglacial meltwater pulses. Other named, postglacial meltwater pulses are known most commonly as meltwater pulse 1A0 (meltwater pulse 19ka), meltwater pulse 1B, meltwater pulse 1C, meltwater pulse 1D, and meltwater pulse 2. It and these other periods of rapid sea level rise are known as meltwater pulses because the inferred cause of them was the rapid release of meltwater into the oceans from the collapse of continental ice sheets. Sea level and timing Meltwater pulse 1A occurred in a period of rising sea level and rapid climate change, known as Termination I, when the retreat of continental ice sheets was going on during the end of the last ice age. Several researchers have narrowed the period of the pulse to between 13,500 and 14,700 calendar years ago with its peak at about 13,800 calendar years ago. The start of this meltwater event coincides with or closely follows the abrupt onset of the Bølling-Allerød (B-A) interstadial and warming in the NorthGRIP ice core in Greenland at 14,600 calendar years ago. During meltwater pulse 1A, sea level is estimated to have risen at a rate of /yr. This rate of sea level rise was much larger than the rate of current sea level rise, which has been estimated to be in the region of /yr. Source(s) of meltwater pulse 1A The source of meltwaters for meltwater pulse 1A and the path they took remains a matter of
https://en.wikipedia.org/wiki/Finite%20element%20limit%20analysis
A finite element limit analysis (FELA) uses optimisation techniques to directly compute the upper or lower bound plastic collapse load (or limit load) for a mechanical system rather than time stepping to a collapse load, as might be undertaken with conventional non-linear finite element techniques. The problem may be formulated in either a kinematic or equilibrium form. The technique has been used most significantly in the field of soil mechanics for the determination of collapse loads for geotechnical problems (e.g. slope stability analysis). An alternative technique which may be used to undertake similar direct plastic collapse computations using optimization is Discontinuity layout optimization. Software for finite element limit analysis OptumG2 (2014-) General purpose software for 2D geotechnical applications. OptumG3 (2017-) General purpose software for 3D geotechnical applications. See also Limit analysis
https://en.wikipedia.org/wiki/Bump-in-the-wire
Bump-in-the-wire (BITW) is a class of communications devices which can be inserted into existing (legacy) systems to enhance the integrity, confidentiality, or reliability of communications across an existing logical link without altering the communications endpoints. The term was originally used to indicate that the device should introduce only a relatively small increased latency in communications compared to the original, unsecured, approach. An example of such a device might be a security appliance which applies IPsec protection to communications between existing devices which themselves lack IPsec implementation protocol stack. Such a device might also be called a security gateway or could be implemented as part of a network firewall to implement a tunneling protocol.
https://en.wikipedia.org/wiki/Dimensional%20operator
In mathematics, specifically set theory, a dimensional operator on a set E is a function from the subsets of E to the subsets of E. Definition If the power set of E is denoted P(E) then a dimensional operator on E is a map that satisfies the following properties for S,T ∈ P(E): S ⊆ d(S); d(S) = d(d(S)) (d is idempotent); if S ⊆ T then d(S) ⊆ d(T); if Ω is the set of finite subsets of S then d(S) = ∪A∈Ωd(A); if x ∈ E and y ∈ d(S ∪ {x}) \ d(S), then x ∈ d(S ∪ {y}). The final property is known as the exchange axiom. Examples For any set E the identity map on P(E) is a dimensional operator. The map which takes any subset S of E to E itself is a dimensional operator on E.
https://en.wikipedia.org/wiki/Interface%20Message%20Processor
The Interface Message Processor (IMP) was the packet switching node used to interconnect participant networks to the ARPANET from the late 1960s to 1989. It was the first generation of gateways, which are known today as routers. An IMP was a ruggedized Honeywell DDP-516 minicomputer with special-purpose interfaces and software. In later years the IMPs were made from the non-ruggedized Honeywell 316 which could handle two-thirds of the communication traffic at approximately one-half the cost. An IMP requires the connection to a host computer via a special bit-serial interface, defined in BBN Report 1822. The IMP software and the ARPA network communications protocol running on the IMPs was discussed in , the first of a series of standardization documents published by what later became the Internet Engineering Task Force (IETF). History The concept of an "Interface computer" was first proposed in 1966 by Donald Davies for the NPL network in England. The same idea was independently developed in early 1967 at a meeting of principal investigators for the Department of Defense's Advanced Research Projects Agency (ARPA) to discuss interconnecting machines across the country. Larry Roberts, who led the ARPANET implementation, initially proposed a network of host computers. Wes Clark suggested inserting "a small computer between each host computer and the network of transmission lines", i.e. making the IMP a separate computer. The IMPs were built by the Massachusetts-based company Bolt Beranek and Newman (BBN) in 1969. BBN was contracted to build four IMPs, the first being due at UCLA by Labor Day; the remaining three were to be delivered in one-month intervals thereafter, completing the entire network in a total of twelve months. When Massachusetts Senator Edward Kennedy learned of BBN's accomplishment in signing this million-dollar agreement, he sent a telegram congratulating the company for being contracted to build the "Interfaith Message Processor". The team worki
https://en.wikipedia.org/wiki/Baldwin%20effect%20%28astronomy%29
The Baldwin effect in astronomy describes a relationship between continuum and emission-line fluxes observed in the electromagnetic spectra of quasars and active galactic nuclei, namely an anticorrelation between the equivalent width, Wλ, of a spectral line and the continuum luminosity, L, in broad UV optical emission lines. This means that the ratio of brightness of the emission line to the brightness of the nearby continuum decreases with increasing luminosity of the continuum. Discovery and observations The effect was discovered in observations of high-redshift quasars reported in 1977, and has been commonly named after the discoverer Jack Allen Baldwin, an astronomer at Lick Observatory. It was found that the equivalent width of the broad ultraviolet C IV (λ 1549) emission line, caused by the presence of thrice ionised carbon C3+, decreased with increasing continuum luminosity of the source. For flat-spectrum, radio-loud quasars the relation was found. The same effect has been observed for other spectral lines such as the Ly-α hydrogen line and the C III] (λ 1909) line of twice ionised carbon. It has also been found for spectral lines in the infrared range, e.g. in several Seyfert galaxies. Explanation More or less satisfactory explanations of the phenomenon have been attempted using photoionization models, continuum beaming, variability, or continuum shape. An adequate, but not exclusive explanation is offered as a consequence of the presence of massive accretion disks in the centres of active galaxies and quasars. However, a complete explanation is still outstanding.
https://en.wikipedia.org/wiki/Neurobiological%20effects%20of%20physical%20exercise
The neurobiological effects of physical exercise are numerous and involve a wide range of interrelated effects on brain structure, brain function, and cognition. A large body of research in humans has demonstrated that consistent aerobic exercise (e.g., 30 minutes every day) induces persistent improvements in certain cognitive functions, healthy alterations in gene expression in the brain, and beneficial forms of neuroplasticity and behavioral plasticity; some of these long-term effects include: increased neuron growth, increased neurological activity (e.g., and BDNF signaling), improved stress coping, enhanced cognitive control of behavior, improved declarative, spatial, and working memory, and structural and functional improvements in brain structures and pathways associated with cognitive control and memory. The effects of exercise on cognition have important implications for improving academic performance in children and college students, improving adult productivity, preserving cognitive function in old age, preventing or treating certain neurological disorders, and improving overall quality of life. In healthy adults, aerobic exercise has been shown to induce transient effects on cognition after a single exercise session and persistent effects on cognition following regular exercise over the course of several months. People who regularly perform an aerobic exercise (e.g., running, jogging, brisk walking, swimming, and cycling) have greater scores on neuropsychological function and performance tests that measure certain cognitive functions, such as attentional control, inhibitory control, cognitive flexibility, working memory updating and capacity, declarative memory, spatial memory, and information processing speed. The transient effects of exercise on cognition include improvements in most executive functions (e.g., attention, working memory, cognitive flexibility, inhibitory control, problem solving, and decision making) and information processing speed fo
https://en.wikipedia.org/wiki/Tarski%27s%20theorem%20about%20choice
In mathematics, Tarski's theorem, proved by , states that in ZF the theorem "For every infinite set , there is a bijective map between the sets and " implies the axiom of choice. The opposite direction was already known, thus the theorem and axiom of choice are equivalent. Tarski told that when he tried to publish the theorem in Comptes Rendus de l'Académie des Sciences de Paris, Fréchet and Lebesgue refused to present it. Fréchet wrote that an implication between two well known propositions is not a new result. Lebesgue wrote that an implication between two false propositions is of no interest. Proof The goal is to prove that the axiom of choice is implied by the statement "for every infinite set ". It is known that the well-ordering theorem is equivalent to the axiom of choice; thus it is enough to show that the statement implies that for every set there exists a well-order. Since the collection of all ordinals such that there exists a surjective function from to the ordinal is a set, there exists an infinite ordinal, such that there is no surjective function from to We assume without loss of generality that the sets and are disjoint. By the initial assumption, thus there exists a bijection For every it is impossible that because otherwise we could define a surjective function from to Therefore, there exists at least one ordinal such that so the set is not empty. We can define a new function: This function is well defined since is a non-empty set of ordinals, and so has a minimum. For every the sets and are disjoint. Therefore, we can define a well order on for every we define since the image of that is, is a set of ordinals and therefore well ordered.
https://en.wikipedia.org/wiki/List%20of%20international%20call%20prefixes
This is a list of international dialing prefixes used in various countries for direct dialing of international telephone calls. These prefixes are typically required only when dialling from a landline, while in GSM-compliant mobile phone (cell phone) systems, only the symbol + before the country code may be used irrespective of where the telephone is used at that moment; the network operator provides the access codes automatically. Countries by international prefix Countries using carrier selection codes The following is a non-exhaustive list of countries that optionally allow for carrier selection in addition to using the standard prefix listed in the preceding section. Historic international prefixes The following are international call prefixes that were used in various countries sometime in the past but are no longer used. See also List of country calling codes (International telephone dialing codes) List of mobile telephone prefixes by country List of North American Numbering Plan area codes Public switched telephone network Notes
https://en.wikipedia.org/wiki/Minimum%20audibility%20curve
Minimum audibility curve is a standardized graph of the threshold of hearing frequency for an average human, and is used as the reference level when measuring hearing loss with an audiometer as shown on an audiogram. Audiograms are produced using a piece of test equipment called an audiometer, and this allows different frequencies to be presented to the subject, usually over calibrated headphones, at any specified level. The levels are, however, not absolute, but weighted with frequency relative to a standard graph known as the minimum audibility curve which is intended to represent 'normal' hearing. This is not the best threshold found for all subjects, under ideal test conditions, which is represented by around 0 phon or the threshold of hearing on the equal-loudness contours, but is standardised in an ANSI standard to a level somewhat higher at 1 kHz . There are several definitions of the minimal audibility curve, defined in different international standards, and they differ significantly, giving rise to differences in audiograms according to the audiometer used. The ASA-1951 standard for example used a level of 16.5 dB SPL at 1 kHz whereas the later ANSI-1969/ISO-1963 standard uses 6.5 dB SPL, and it is common to allow a 10 dB correction for the older standard. See also Articulation index Audiogram Audiology Audiometry A-weighting Equal-loudness contour Hearing range Hearing (sense) Psychoacoustics Pure tone audiometry External links Hearing Loss by Robert Thayer Sataloff Otology Acoustics
https://en.wikipedia.org/wiki/RC%20algorithm
The RC algorithms are a set of symmetric-key encryption algorithms invented by Ron Rivest. The "RC" may stand for either Rivest's cipher or, more informally, Ron's code. Despite the similarity in their names, the algorithms are for the most part unrelated. There have been six RC algorithms so far: RC1 was never published. RC2 was a 64-bit block cipher developed in 1987. RC3 was broken before ever being used. RC4 is a stream cipher. RC5 is a 32/64/128-bit block cipher developed in 1994. RC6, a 128-bit block cipher based heavily on RC5, was an AES finalist developed in 1997.
https://en.wikipedia.org/wiki/King%27s%20Knight
is a scrolling shooter video game developed and published by Square for the Nintendo Entertainment System and MSX. The game was released in Japan on September 18, 1986 and in North America in 1989. It was later re-released for the Wii's Virtual Console in Japan on November 27, 2007 and in North America on March 24, 2008. This would be followed by a release on the Virtual Console in Japan on February 4, 2015, for 3DS and July 6, 2016, for Wii U. The game became Square's first North American release under their Redmond subsidiary Squaresoft, and their first release as an independent company. The 1986 release's title screen credits Workss for programming. King's Knight saw a second release in 1987 on the NEC PC-8801mkII SR and the Sharp X1. These versions of the game were retitled King's Knight Special and released exclusively in Japan. It was the first game designed by Hironobu Sakaguchi for the Famicom. Nobuo Uematsu provided the musical score for King's Knight. It was Uematsu's third work of video game music composition. Plot King's Knight follows a basic storyline similar to many NES-era role-playing video games: Princess Claire of Olthea has been kidnapped in the Kingdom of Izander, and the player must choose one of the four heroes (the knight/warrior "Ray Jack", the wizard "Kaliva", the monster/gigant "Barusa" and the (kid) thief "Toby") to train and set forth to attack Gargatua Castle, defeat the evil dragon Tolfida and rescue the princess. Gameplay King's Knight is a vertically scrolling shooter, where the main objective is to dodge or destroy all onscreen enemies and obstacles. Various items, however, add depth to the game. As any character, the player can collect various power-ups to increase a character's level (maximum of twenty levels per character): as many as seven Jump Increases, seven Speed Increases, three Weapon Increases, and three Shield Increases. There are also Life Ups, which are collected to increase the character's life meter. There are
https://en.wikipedia.org/wiki/Palatal%20hook
The palatal hook () is a type of hook diacritic formerly used in the International Phonetic Alphabet to represent palatalized consonants. It is a small, leftwards-facing hook joined to the bottom-right side of a letter, and is distinguished from various other hooks indicating retroflexion, etc. It was withdrawn by the IPA in 1989, in favour of a superscript j following the consonant (i.e., ⟨⟩ becomes ⟨⟩). The IPA recommended that esh ⟨⟩ and ezh ⟨⟩) not use the palatal hook, but instead get special curled symbols: ⟨⟩ and ⟨⟩. However, versions with the hook have also been used and are supported by Unicode. Palatal hooks are also used in Lithuanian dialectology by the Lithuanian Phonetic Transcription System (or Lithuanian Phonetic Alphabet) and in the orthography of Nez Perce. Computer encoding Unicode includes both a combining character for the palatal hook, as well as several precomposed characters, including superscript letters with palatal hooks. While LATIN SMALL LETTER T WITH PALATAL HOOK has been in Unicode since 1991, the rest were not added until 2005 or later. As such, font support for the latter characters is much less than for the former.
https://en.wikipedia.org/wiki/Drinfeld%20reciprocity
In mathematics, Drinfeld reciprocity, introduced by , is a correspondence between eigenforms of the moduli space of Drinfeld modules and factors of the corresponding Jacobian variety, such that all twisted L-functions are the same.
https://en.wikipedia.org/wiki/Ion%20beam
An ion beam is a type of charged particle beam consisting of ions. Ion beams have many uses in electronics manufacturing (principally ion implantation) and other industries. A variety of ion beam sources exists, some derived from the mercury vapor thrusters developed by NASA in the 1960s. The most common ion beams are of singly-charged ions. Units Ion current density is typically measured in mA/cm^2, and ion energy in eV. The use of eV is convenient for converting between voltage and energy, especially when dealing with singly-charged ion beams, as well as converting between energy and temperature (1 eV = 11600 K). Broad-beam ion sources Most commercial applications use two popular types of ion source, gridded and gridless, which differ in current and power characteristics and the ability to control ion trajectories. In both cases electrons are needed to generate an ion beam. The most common electron emitters are hot filament and hollow cathode. Gridded ion source In a gridded ion source, DC or RF discharge are used to generate ions, which are then accelerated and decimated using grids and apertures. Here, the DC discharge current or the RF discharge power are used to control the beam current. The ion current density that can be accelerated using a gridded ion source is limited by the space charge effect, which is described by Child's law: where is the voltage between the grids, is the distance between the grids, and is the ion mass. The grids are placed as closely as possible to increase the current density, typically . The ions used have a significant impact on the maximum ion beam current, since . Everything else being equal, the maximum ion beam current with krypton is only 69% the maximum ion current of an argon beam, and with xenon the ratio drops to 55%. Gridless ion sources In a gridless ion source, ions are generated by a flow of electrons (no grids). The most common gridless ion source is the end-Hall ion source. Here, the discharge current
https://en.wikipedia.org/wiki/Beltanelliformis
Beltanelliformis is a genus of discoid fossil from the Ediacaran period containing the two species B. brunsae and B. minutae, sometimes ascribed to the Ediacaran Biota. The chemical signature obtained from organically preserved specimens points to a cyanobacterial affinity (cf. Nostoc). Depending on its preservation, it is sometimes referred to as Nemiana or Beltanelloides.
https://en.wikipedia.org/wiki/Inverse%20problem%20in%20optics
The inverse problem in optics (or the inverse optics problem) refers to the fundamentally ambiguous mapping between sources of retinal stimulation and the retinal images that are caused by those sources. For example, the size of an object, the orientation of the object, and its distance from the observer are conflated in the retinal image. For any given projection on the retina there are an infinite number of pairings of object size, orientation and distance that could have given rise to that projection on the retina. Because the image on the retina does not specify which pairing did in fact cause the image, this and other aspects of vision qualify as an inverse problem.
https://en.wikipedia.org/wiki/Algebraic%20specification
Algebraic specification is a software engineering technique for formally specifying system behavior. It was a very active subject of computer science research around 1980. Overview Algebraic specification seeks to systematically develop more efficient programs by: formally defining types of data, and mathematical operations on those data types abstracting implementation details, such as the size of representations (in memory) and the efficiency of obtaining outcome of computations formalizing the computations and operations on data types allowing for automation by formally restricting operations to this limited set of behaviors and data types. An algebraic specification achieves these goals by defining one or more data types, and specifying a collection of functions that operate on those data types. These functions can be divided into two classes: Constructor functions: Functions that create or initialize the data elements, or construct complex elements from simpler ones. The set of available constructor functions is implied by the specification's signature. Additionally, a specification can contain equations defining equivalences between the objects constructed by these functions. Whether the underlying representation is identical for different but equivalent constructions is implementation-dependent. Additional functions: Functions that operate on the data types, and are defined in terms of the constructor functions. Examples Consider a formal algebraic specification for the boolean data type. One possible algebraic specification may provide two constructor functions for the data-element: a true constructor and a false constructor. Thus, a boolean data element could be declared, constructed, and initialized to a value. In this scenario, all other connective elements, such as XOR and AND, would be additional functions. Thus, a data element could be instantiated with either "true" or "false" value, and additional functions could be used to perform an
https://en.wikipedia.org/wiki/Differentially%20private%20analysis%20of%20graphs
Differentially private analysis of graphs studies algorithms for computing accurate graph statistics while preserving differential privacy. Such algorithms are used for data represented in the form of a graph where nodes correspond to individuals and edges correspond to relationships between them. For examples, edges could correspond to friendships, sexual relationships, or communication patterns. A party that collected sensitive graph data can process it using a differentially private algorithm and publish the output of the algorithm. The goal of differentially private analysis of graphs is to design algorithms that compute accurate global information about graphs while preserving privacy of individuals whose data is stored in the graph. Variants Differential privacy imposes a restriction on the algorithm. Intuitively, it requires that the algorithm has roughly the same output distribution on neighboring inputs. If the input is a graph, there are two natural notions of neighboring inputs, edge neighbors and node neighbors, which yield two natural variants of differential privacy for graph data. Let ε be a positive real number and be a randomized algorithm that takes a graph as input and returns an output from a set . The algorithm is -differentially private if, for all neighboring graphs and and all subsets of , where the probability is taken over the randomness used by the algorithm. Edge differential privacy Two graphs are edge neighbors if they differ in one edge. An algorithm is -edge-differentially private if, in the definition above, the notion of edge neighbors is used. Intuitively, an edge differentially private algorithm has similar output distributions on any pair of graphs that differ in one edge, thus protecting changes to graph edges. Node differential privacy Two graphs are node neighbors if one can be obtained from the other by deleting a node and its adjacent edges. An algorithm is -node-differentially private if, in the definition ab
https://en.wikipedia.org/wiki/Information%20ecology
Information ecology is the application of ecological concepts for modeling the information society. It considers the dynamics and properties of the increasingly dense, complex and important digital informational environment. "Information ecology" often is used as metaphor, viewing the information space as an ecosystem, the information ecosystem. Information ecology also makes a connection to the concept of collective intelligence and knowledge ecology . Eddy et al. (2014) use information ecology for science-policy integration in ecosystems-based management (EBM). Networked information economy In The Wealth of Networks: How Social Production Transforms Markets and Freedom, a book published in 2006 and available under a Creative Commons license on its own wikispace, Yochai Benkler provides an analytic framework for the emergence of the networked information economy that draws deeply on the language and perspectives of information ecology together with observations and analyses of high-visibility examples of successful peer production processes, citing Wikipedia as a prime example. Bonnie Nardi and Vicki O'Day in their book "Information Ecologies: Using Technology with Heart," apply the ecology metaphor to local environments, such as libraries and schools, in preference to the more common metaphors for technology as tool, text, or system. In different domains / disciplines Anthropology Nardi and O’Day's book represents the first specific treatment of information ecology by anthropologists. H.E. Kuchka situates information within socially-distributed cognition of cultural systems. Casagrande and Peters use information ecology for an anthropological critique of Southwest US water policy. Stepp (1999) published a prospectus for the anthropological study of information ecology. Knowledge management Information ecology was used as book title by Thomas H. Davenport and Laurence Prusak , with a focus on the organization dimensions of information ecology. There was
https://en.wikipedia.org/wiki/Digital%20Interface%20for%20Video%20and%20Audio
The Digital Interface for Video and Audio (DIVA or DiiVA) was a proposal for a bi-directional audio/video interface for transmitting both compressed and uncompressed digital streams. It was developed by Synerchip Company, Limited, based in Guangzhou and Sunnyvale, California. DIVA supports a downstream data rate (from source to display) of 13.5 Gbit/s which is capable of deep color at resolutions higher than 1080p. DIVA also supports a 2.25 Gbit/s bi-directional communication data channel that can carry multiple sub-channels (audio, control, compressed video, etc.). This gives DIVA a raw bi-directional data rate of 18 Gbit/s or a usable bi-directional data rate of 14.4 Gbit/s (because of 8b/10b encoding). DIVA was demonstrated at the China Digital Living Forum & Showcase 2008 using a single Category 6A cable. The DIVA Promoters Group was formed by Changhong, Haier, Hisense, Konka, Panda, Skyworth, SVA, TCL Corporation and Synerchip. In February 2008, the DIVA Promoters Group hoped to finish the specs for DIVA by the end of 2008 and have DIVA chips released in 2009 or 2010. See also Digital Visual Interface DisplayPort HDMI List of video connectors
https://en.wikipedia.org/wiki/Iodine-125
Iodine-125 (125I) is a radioisotope of iodine which has uses in biological assays, nuclear medicine imaging and in radiation therapy as brachytherapy to treat a number of conditions, including prostate cancer, uveal melanomas, and brain tumors. It is the second longest-lived radioisotope of iodine, after iodine-129. Its half-life is 59.49 days and it decays by electron capture to an excited state of tellurium-125. This state is not the metastable 125mTe, but rather a lower energy state that decays immediately by gamma decay with a maximum energy of 35 keV. Some of the excess energy of the excited 125Te may be internally converted ejected electrons (also at 35 keV), or to x-rays (from electron bremsstrahlung), and also a total of 21 Auger electrons, which are produced at the low energies of 50 to 500 electron volts. Eventually, stable ground state 125Te is produced as the final decay product. In medical applications, the internal conversion and Auger electrons cause little damage outside the cell which contains the isotope atom. The X-rays and gamma rays are of low enough energy to deliver a higher radiation dose selectively to nearby tissues, in "permanent" brachytherapy where the isotope capsules are left in place (125I competes with palladium-103 in such uses). Because of its relatively long half-life and emission of low-energy photons which can be detected by gamma-counter crystal detectors, 125I is a preferred isotope for tagging antibodies in radioimmunoassay and other gamma-counting procedures involving proteins outside the body. The same properties of the isotope make it useful for brachytherapy, and for certain nuclear medicine scanning procedures, in which it is attached to proteins (albumin or fibrinogen), and where a half-life longer than that provided by 123I is required for diagnostic or lab tests lasting several days. Iodine-125 can be used in scanning/imaging the thyroid, but iodine-123 is preferred for this purpose, due to better radiation penetr
https://en.wikipedia.org/wiki/NSD
In Internet computing, NSD (for "name server daemon") is an open-source Domain Name System (DNS) server. It was developed by NLnet Labs of Amsterdam in cooperation with the RIPE NCC, from scratch as an authoritative name server (i.e., not implementing the recursive caching function by design). The intention of this development is to add variance to the "gene pool" of DNS implementations used by higher level name servers and thus increase the resilience of DNS against software flaws or exploits. NSD uses BIND-style zone-files (zone-files used under BIND can usually be used unmodified in NSD, once entered into the NSD configuration). NSD uses zone information compiled via zonec into a binary database file (nsd.db) which allows fast startup of the NSD name-service daemon, and allows syntax-structural errors in Zone-Files to be flagged at compile-time (before being made available to NSD service itself). The collection of programs/processes that make-up NSD are designed so that the NSD daemon itself runs as a non-privileged user and can be easily configured to run in a Chroot jail, such that security flaws in the NSD daemon are not so likely to result in system-wide compromise as without such measures. As of May, 2018, four of the Internet root nameservers are using NSD: k.root-servers.net was switched to NSD on February 19, 2003. One of the 2 load-balanced servers for h.root-servers.net (called "H1", "H2") was switched to NSD, and now there are 3 servers all running NSD (called "H1", "H2", "H3"). l.root-servers.net switched to NSD on February 6, 2007. d.root-servers.net was switched to NSD in May 2018. Several other TLDs use NSD for part of their servers. See also Unbound, a recursive DNS server, also developed by NLnet Labs Comparison of DNS server software
https://en.wikipedia.org/wiki/Bothrioneodermata
Bothrioneodermata is a clade of flatworms containing the Bothrioplanida and the Neodermata.
https://en.wikipedia.org/wiki/Atromentin
Atromentin is a natural chemical compound found in Agaricomycetes fungi in the orders Agaricales and Thelephorales. It can also be prepared by laboratory synthesis. Chemically, it is a polyphenol and a benzoquinone. Occurrences Atromentin has been found in cultures of Clitocybe subilludens and in extracts of Hydnellum peckii. The first enzymes in its biosynthesis have been characterized in Tapinella panuoides. One of those is called atromentin synthetase. Biological activities A number of potential biological activities of atromentin have been studied in vitro. Atromentin possesses in vitro antibacterial activity, inhibiting the enzyme enoyl-acyl carrier protein reductase (essential for the biosynthesis of fatty acids) in the bacteria Streptococcus pneumoniae. Atromentin has been shown to be a smooth muscle stimulant. It also induces apoptosis in isolated human leukemia U937 cells. It is also an anticoagulant. Genetic and enzymatic basis of atromentin Atromentin is biosynthesized from two units of 4-hydroxyphenylpyruvic acid (4-HPP) via a nonribosomal peptide synthetase-like enzyme (atromentin synthetase), containing the domain architecture adenylation-thiolation-thioesterase (A-T-TE). 4-HPP is produced from a deamination via an aminotransferase. The genetic basis of these two genes is clustered (i.e., adjacent to one another). These enzymes were first characterized in Tapinella panuoides by overexpressing the respective genes (AtrA and AtrD) in E. coli and incubating the holo-enzyme with 4-HPP to observe the formation of atromentin. This was followed by characterization of the enzyme GreA in Suillus grevillei, six (InvA1-6, of which InvA1, 2 and 5 were functional) in Paxillus involutus, and NPS3 from Serpula lacrymans. In addition, there is another adjacent and conserved gene encoding for an alcohol dehydrogenase/oxidoreductase whose function is unclear. In most cases the clustered biosynthetic genes are found orthologous in basidiomycetes. A common promo
https://en.wikipedia.org/wiki/Raynaud%20syndrome
Raynaud syndrome, also known as Raynaud's phenomenon, is a medical condition in which the spasm of small arteries causes episodes of reduced blood flow to end arterioles. Typically, the fingers, and less commonly, the toes, are involved. Rarely, the nose, ears, or lips are affected. The episodes classically result in the affected part turning white and then blue. Often, numbness or pain occurs. As blood flow returns, the area turns red and burns. The episodes typically last minutes but can last several hours. The condition is named after the physician Auguste Gabriel Maurice Raynaud, who first described it in his doctoral thesis in 1862. Episodes are typically triggered by cold or emotional stress. Primary Raynaud's is idiopathic (spontaneous and of unknown cause) and not correlated with another disease. Secondary Raynaud's occurs as a result of some other condition and has an older age at onset; episodes are intensely painful and can be asymmetric and associated with skin lesions. Secondary Raynaud's can occur due to a connective-tissue disorder such as scleroderma or lupus, injuries to the hands, prolonged vibration, smoking, thyroid problems, and certain medications, such as birth control pills and stimulants. Diagnosis is typically based on the symptoms. The primary treatment is avoiding the cold. Other measures include the discontinuation of nicotine or stimulant use. Medications for treatment of cases that do not improve include calcium channel blockers and iloprost. Little evidence supports alternative medicine. Severe disease may in rare cases lead to complications, specifically skin sores or gangrene. About 4% of people have the condition. Onset of the primary form is typically between ages 15 and 30 and occurs more frequently in females. The secondary form usually affects older people. Both forms are more common in cold climates. Signs and symptoms The condition can cause localized pain, discoloration (paleness), and sensations of cold and/or numbnes
https://en.wikipedia.org/wiki/Union%20of%20Food%20and%20Drink%20Workers
The Union of Food and Drink Workers (, VNG) was a trade union representing workers in the food and drink processing industry in Germany. The union was founded in 1927, when the Central Union of Bakers and Confectioners merged with the Union of Brewery and Mill Workers, the Central Union of Butchers, and the Union of Coopers, Cellar Managers, and Helpers in Germany. On formation, the union had 159,636 members and, like its predecessors, it affiliated to the General German Trade Union Confederation. It was based in Berlin, and led by Eduard Backert. In 1933, it was banned by the Nazis, and after World War II, workers in the industry were represented by the Food, Beverages and Catering Union.
https://en.wikipedia.org/wiki/Dirichlet%27s%20approximation%20theorem
In number theory, Dirichlet's theorem on Diophantine approximation, also called Dirichlet's approximation theorem, states that for any real numbers and , with , there exist integers and such that and Here represents the integer part of . This is a fundamental result in Diophantine approximation, showing that any real number has a sequence of good rational approximations: in fact an immediate consequence is that for a given irrational α, the inequality is satisfied by infinitely many integers p and q. This shows that any irrational number has irrationality measure at least 2. This corollary also shows that the Thue–Siegel–Roth theorem, a result in the other direction, provides essentially the tightest possible bound, in the sense that the bound on rational approximation of algebraic numbers cannot be improved by increasing the exponent beyond 2. The Thue–Siegel–Roth theorem uses advanced techniques of number theory, but many simpler numbers such as the golden ratio can be much more easily verified to be inapproximable beyond exponent 2. This exponent is referred to as the irrationality measure. Simultaneous version The simultaneous version of the Dirichlet's approximation theorem states that given real numbers and a natural number then there are integers such that Method of proof Proof by the pigeonhole principle This theorem is a consequence of the pigeonhole principle. Peter Gustav Lejeune Dirichlet who proved the result used the same principle in other contexts (for example, the Pell equation) and by naming the principle (in German) popularized its use, though its status in textbook terms comes later. The method extends to simultaneous approximation. Proof outline: Let be an irrational number and be an integer. For every we can write such that is an integer and . One can divide the interval into smaller intervals of measure . Now, we have numbers and intervals. Therefore, by the pigeonhole principle, at least two of them are in the same i
https://en.wikipedia.org/wiki/Acoustic%20seabed%20classification
Acoustic seabed classification is the partitioning of a seabed acoustic image into discrete physical entities or classes. This is a particularly active area of development in the field of seabed mapping, marine geophysics, underwater acoustics and benthic habitat mapping. Seabed classification is one route to characterizing the seabed and its habitats. Seabed characterization makes the link between the classified regions and the seabed physical, geological, chemical or biological properties. Acoustic seabed classification is possible using a wide range of acoustic imaging systems including multibeam echosounders, sidescan sonar, single-beam echosounders, interferometric systems and sub-bottom profilers. Seabed classification based on acoustic properties can be divided into two main categories; surficial seabed classification and sub-surface seabed classification. Sub-surface imaging technologies use lower frequency sound to provide higher penetration, whereas surficial imaging technologies provide higher resolution imagery by utilizing higher frequencies (especially in shallow water). Surficial seabed classification Classification methods Surficial seabed classification is concerned primarily with distinguishing marine benthic habitat characteristics (e.g. hard, soft, rough, smooth, mud, sand, clay, cobble) of the surveyed area. Multibeam echosounders, sidescan sonar systems and acoustic ground discrimination systems (AGDS) are the most commonly used technologies. The use of optical sensors has been restricted to depths less than 40 m due to absorption of electromagnetic radiation by water. Despite this limitation, processing tools have been developed to classify data acquired using airborne bathymetric LiDAR systems. Nevertheless, acoustics remain the preferred method of imaging the seafloor because data can be acquired over a much larger area (than in-situ sampling) from almost any depth. Multibeam systems acquire both bathymetry (depth) and backscatter (inten
https://en.wikipedia.org/wiki/Galegine
Galegine is a toxic chemical compound that has been isolated from Goat's rue (Galega officinalis). It has also been found to be the principal cause of the toxicity of poison sedge (Schoenus asperocarpus). Galegine was used in the 1920s as a pharmaceutical treatment for diabetes; however, because of its toxicity, its use was soon supplanted by superior alternatives. Research on galegine eventually led to the development of metformin which is used today for treatment of type 2 diabetes. See also Nitensidine D
https://en.wikipedia.org/wiki/Distributed%20control%20system
A distributed control system (DCS) is a computerised control system for a process or plant usually with many control loops, in which autonomous controllers are distributed throughout the system, but there is no central operator supervisory control. This is in contrast to systems that use centralized controllers; either discrete controllers located at a central control room or within a central computer. The DCS concept increases reliability and reduces installation costs by localising control functions near the process plant, with remote monitoring and supervision. Distributed control systems first emerged in large, high value, safety critical process industries, and were attractive because the DCS manufacturer would supply both the local control level and central supervisory equipment as an integrated package, thus reducing design integration risk. Today the functionality of Supervisory control and data acquisition (SCADA) and DCS systems are very similar, but DCS tends to be used on large continuous process plants where high reliability and security is important, and the control room is not geographically remote. Many machine control systems exhibit similar properties as plant and process control systems do. Structure The key attribute of a DCS is its reliability due to the distribution of the control processing around nodes in the system. This mitigates a single processor failure. If a processor fails, it will only affect one section of the plant process, as opposed to a failure of a central computer which would affect the whole process. This distribution of computing power local to the field Input/Output (I/O) connection racks also ensures fast controller processing times by removing possible network and central processing delays. The accompanying diagram is a general model which shows functional manufacturing levels using computerised control. Referring to the diagram; Level 0 contains the field devices such as flow and temperature sensors, and final cont
https://en.wikipedia.org/wiki/Helper/suppressor%20ratio
The T-Lymphocyte Helper/Suppressor Profile (Helper/Suppressor ratio, T4:T8 ratio, CD4:CD8 ratio) is a basic laboratory test in which the percentage of CD3-positive lymphocytes in the blood positive for CD4 (T helper cells) and CD8 (a class of regulatory T cells) are counted and compared. Normal values (95% confidence intervals) are approximately 30-60% CD4 and 10-30% CD8 depending on age (ratio 0.9 to 3.7 in adults). One reason for abnormal results is the loss of CD4-positive cells to the human immunodeficiency virus (HIV) infection. The loss of CD4-positive cells to HIV infection can result in various distortions in the ratio, as in the initial period, production of HIV specific CD8 positive cells will cause a large fall in the ratio, but subsequent immunosuppression over time may lead to overall non production of immune cells and inversion of the ratio. It has been shown that the degree of inversion of this ratio in individuals on antiretroviral therapy is indicative of the age of the infection and independently predictive of mortality associated with non HIV events.
https://en.wikipedia.org/wiki/Yasm%C3%ADn%20R%C3%ADos-Sol%C3%ADs
Yasmín Águeda Ríos-Solís is a Mexican computer scientist and operations researcher who studies problems of scheduling, timetabling, and synchronization of public transport. She is a professor and researcher in the School of Engineering and Sciences at the Monterrey Institute of Technology and Higher Education. Education and career Ríos-Solís studied applied mathematics at the Instituto Tecnológico Autónomo de México, graduating in 2002. Next, she went to Pierre and Marie Curie University in France for graduate study in computer science and operations research, funded by the Consejo Nacional de Ciencia y Tecnología (Mexico). She earned a master's degree in 2003 and completed her PhD there in 2007. Her doctoral dissertation, Earliness and Tardiness Parallel Machine Scheduling, was jointly supervised by Francis Sourd and Philippe Chrétienne. After postdoctoral research in bioinformatics at Bielefeld University in Germany, she returned to Mexico in 2008 as an assistant professor at the Autonomous University of Nuevo León, in the Graduate Program in Systems Engineering. There, she was promoted to associate professor in 2011. She moved to her present position at the Monterrey Institute of Technology and Higher Education in 2020. Recognition Ríos-Solís was elected to the Mexican Academy of Sciences in 2020.
https://en.wikipedia.org/wiki/Smart%20fixture
A smart fixture is an advanced test fixture which combines a fixture with sensors to collect data and provide feedback. They are a subset of cyber-physical systems. Sensors and/or instrumentation embedded in the fixture are connected to a programmable logic controller or computer which apply algorithms to determine if required criteria are met i.e. functional testing. A smart fixture can be a manufacturing fixture or simply a test fixture. The smart aspect is fully realised in the use of the data collected via IIoT/ Industry 4.0. machine learning, artificial intelligence and other big data tools can be applied to perform real time analysis and predictive decision making to improve operational efficiencies. Electronics Circuit boards, are held in place and subjected to controlled electronic test signals. Examples of fixtures: Uploading a software into a microcontroller and testing the functionality of the PCBA Checking the correct parameters of a LED light See also Device under test Bed of nails tester
https://en.wikipedia.org/wiki/Medical%20Devices%20Park%2C%20Hyderabad
Medical Devices Park, Hyderabad is a medical devices industrial estate located in Hyderabad, Telangana, India. The largest such Park in India spread over 250 acres. The dedicated park's ecosystem supports medical technology innovation and manufacturing. History The Park was inaugurated on 17 June 2017 near Hyderabad at Sultanpur in Patancheru of Sangareddy district by the Minister for Industries, K. T. Rama Rao.
https://en.wikipedia.org/wiki/Biozone
In biostratigraphy, biostratigraphic units or biozones are intervals of geological strata that are defined on the basis of their characteristic fossil taxa, as opposed to a lithostratigraphic unit which is defined by the lithological properties of the surrounding rock. A biostratigraphic unit is defined by the zone fossils it contains. These may be a single taxon or combinations of taxa if the taxa are relatively abundant, or variations in features related to the distribution of fossils. The same strata may be zoned differently depending on the diagnostic criteria or fossil group chosen, so there may be several, sometimes overlapping, biostratigraphic units in the same interval. Like lithostratigraphic units, biozones must have a type section designated as a stratotype. These stratotypes are named according to the typical taxon (or taxa) that are found in that particular biozone. The boundary of two distinct biostratigraphic units is called a biohorizon. Biozones can be further subdivided into subbiozones, and multiple biozones can be grouped together in a superbiozone in which the grouped biozones usually have a related characteristic. A succession of biozones is called biozonation. The length of time represented by a biostratigraphic zone is called a biochron. History The concept of a biozone was first established by the 19th century paleontologist Albert Oppel, who characterized rock strata by the species of the fossilized animals found in them, which he called zone fossils. Oppel's biozonation was mainly based on Jurassic ammonites he found throughout Europe, which he used to classify the period into 33 zones (now 60). Alcide d'Orbigny would further reinforce the concept in his Prodrome de Paléontologie Stratigraphique, in which he established comparisons between geological stages and their biostratigraphy. Types of biozone The International Commission on Stratigraphy defines the following types of biozones: Range zones Range zones are biozones defined b
https://en.wikipedia.org/wiki/Indentation%20%28typesetting%29
In the written form of many languages, an indentation or indent is an empty space at the beginning of a line to signal the start of a new paragraph. Many computer languages have adopted this technique to designate "paragraphs" or other logical blocks in the program. For example, the following lines are indented, using between one and six spaces:  This paragraph is indented by 1 space.    This paragraph is indented by 3 spaces.       This paragraph is indented by 6 spaces. In computer programming, the neologisms outdent, unindent and dedent are used to describe the reversal of the indentation process, realigning text with the page margin (or with previous, lesser, levels of indentation). In right-to-left languages (e.g. Hebrew and Arabic), indentation is used just the same, but from the right margin of the paper, where the line begins. Indentation in typesetting There are three main types of indentation, illustrated below in relation to borders representing the page dimensions. Indentation in programming In computer programming languages, indentation is used to format program source code to improve readability. Indentation is generally only of use to programmers; compilers and interpreters rarely care how much whitespace is present in between programming statements. However, certain programming languages rely on the use of indentation to demarcate programming structure, often using a variation of the off-side rule. The Haskell, Occam, Python, MoonScript, and Ya programming languages rely on indentation in this way. Opinions about where to indent, whether to use spaces or tabs, and how many spaces to use are often hotly debated among programmers, leading some to describe indentation disputes as akin to a religious war. In 2006 a third method of indentation was proposed, called elastic tabstops. In addition to general indentation of statements, different bracket indentation styles are commonly used.
https://en.wikipedia.org/wiki/InnoLux%20Corporation
Innolux Corporation () is a company producing TFT LCD panels, established in 2003 and located in Taiwan. Overview Innolux Display Corp., following its merger with Chi Mei Optoelectronics and TPO Displays Corp., began operating under the name Innolux Corporation (INX) in March 2010. With products spanning the full range of TFT-LCD panel modules and touch panels, including TV panels, desktop monitors and notebook computer panels, AV & mobile panels, Innolux is a TFT-LCD supplier to information technology and consumer electronics product makers worldwide. Innolux's largest customers include Toshiba, Samsung, Philips, LG, Sony, Panasonic, Sharp, Lenovo, HP, Broteko, Dell & HDMIPI. The 17in touch panel in Tesla, Inc. automobiles is an Innolux product. In 2018 Innolux had the third highest R&D spending of any listed Taiwanese company. Innolux's investment of NT$11.3 billion represented 4.1 percent of its total sales. Innolux also investing NT$41.7 billion in fixed assets, the second most of any listed Taiwanese company. History Chi Mei Optoelectronics is established on 6 August 1998. The department of TPO Displays is established on 24 December 1999. Innolux Display Corp. is originally established on 14 January 2003. Innolux Display Corp. publicly listed its shares on the Taiwan Stock Exchange on 24 October 2006. 10/05/2009 Innolux Display Corp. and TPO Displays carry out merger via a share swap arrangement 11/14/2009 Innolux Display Corp. and Chi Mei Optoelectronics carry out merger via a share swap arrangement, with Innolux Display Corp. 03/18/2010 Innolux Display Corp. announces the completion of merger with Chi Mei Optoelectronics and TPO Displays. 03/30/2010 Innolux Display Corp. officially changes its name to Chimei Innolux Corporation. 03/16/2012 The board of directors elected Dr. Hsing Chien Tuan as new chairman of the board. 12/26/2012 Chimei Innolux Corporation changes its name to “Innolux Corporation”. 01/14/2013 Innolux Corp. to hold the 10-ye
https://en.wikipedia.org/wiki/Journal%20of%20Algebraic%20Combinatorics
Journal of Algebraic Combinatorics is a peer-reviewed scientific journal covering algebraic combinatorics. It was established in 1992 and is published by Springer Science+Business Media. The editor-in-chief is Ilias S. Kotsireas (Wilfrid Laurier University). In 2017, the journal's four editors-in-chief and editorial board resigned to protest the publisher's high prices and limited accessibility. They criticized Springer for "double-dipping", that is, charging large subscription fees to libraries in addition to high fees for authors who wished to make their publications open access. The board subsequently started their own open access journal, Algebraic Combinatorics. Abstracting and indexing The journal is abstracted and indexed in: According to the Journal Citation Reports, the journal has a 2020 impact factor of 0.875.
https://en.wikipedia.org/wiki/Chemosis
Chemosis is the swelling (or edema) of the conjunctiva. The term derives from the Greek words cheme and -osis, cheme meaning cockleshell due to the swollen conjunctiva resembling it, and -osis meaning condition. The swelling is due to the oozing of exudate from abnormally permeable capillaries. In general, chemosis is a nonspecific sign of eye irritation. The outer surface covering appears to have fluid in it. The conjunctiva becomes swollen and gelatinous in appearance. Often, the eye area swells so much that the eyes become difficult or impossible to close fully. Sometimes, it may also appear as if the eyeball has moved slightly backwards from the white part of the eye due to the fluid filled in the conjunctiva all over the eyes except the iris. The iris is not covered by this fluid and so it appears to be moved slightly inwards. Causes It is usually caused by allergies or viral infections, often inciting excessive eye rubbing. Chemosis is also included in the Chandler Classification system of orbital infections. If chemosis has occurred due to excessive rubbing of the eye, the first aid to be given is a cold water wash for eyes. Other causes of chemosis include: Superior vena cava obstruction, accompanied by facial oedema Hyperthyroidism, associated with exophthalmos, periorbital puffiness, lid retraction, and lid lag Cavernous sinus thrombosis, associated with infection of the paranasal sinuses, proptosis, periorbital oedema, retinal haemorrhages, papilledema, extraocular movement abnormalities, and trigeminal nerve sensory loss Carotid-cavernous fistula - classic triad of chemosis, pulsatile proptosis, and ocular bruit Cluster headache Trichinellosis Systemic lupus erythematosus (SLE) Angioedema Acute glaucoma Panophthalmitis Orbital cellulitis Gonorrheal conjunctivitis Dacryocystitis Spitting cobra venom to the eye High concentrations of phenacyl chloride in chemical mace spray Urticaria Trauma HSV Keratitis Post surgical Mucor Rhabdomyosarcoma of the or
https://en.wikipedia.org/wiki/TSQ
6-Methoxy-(8-p-toluenesulfonamido)quinoline (TSQ) is one of the most efficient fluorescent stains for zinc(II). It was introduced by Soviet biochemists Toroptsev and Eshchenko in the early 1970s. The popularity of TSQ as physiological stain rose after seminal works by Christopher Frederickson two decades later. TSQ forms a 2:1 (ligand-metal) complex with zinc and emits blue light upon excitation at 365 nanometers. TSQ has been extensively applied for determination of extracellular or intracellular levels of Zn2+ in biological systems, also to study Zn2+ in mossy fibers of the hippocampus.
https://en.wikipedia.org/wiki/Y%CE%94-%20and%20%CE%94Y-transformation
In graph theory, ΔY- and YΔ-transformations (also written delta-wye and wye-delta) are a pair of operations on graphs. A ΔY-transformation replaces a triangle by a vertex of degree three; and conversely, a YΔ-transformation replaces a vertex of degree three by a triangle. The name for the operations derives from the shapes of the involved subgraphs, which look respectively like the letter Y and the Greek capital letter Δ. A YΔ-transformation might create parallel edges, even if applied to a simple graph. For this reason ΔY- and YΔ-transformations are often considered as operations on multigraphs. On multigraphs both operations preserve the edge count and are exact inverses of each other. In the context of simple graphs it is common to combine a YΔ-transformation with a subsequent normalization step that reduces any parallel edges to a single edge. This might no longer preserve the number of edges, nor be exactly reversible via a ΔY-transformation. Formal definition Let be a graph (potentially a multigraph). If the edges form a triangle with vertices , then a ΔY-transformation of at removes and adds a new vertex adjacent to all of . If is a vertex of degree three with neighbors , then a YΔ-transformation of at deletes and adds three new edges so that connects and . If the resulting graph is supposed to be a simple graph, then any resulting parallel edges are to be replaced by a single edge. Relevance ΔY- and YΔ-transformations are studied as a tool both in pure graph theory as well as applications. Both operations preserve a number of natural topological properties of graphs, such as being planar or linkless. They are therefore used to compactly describe the families of forbidden minors as well as other graph families. For example: the Petersen family (the six forbidden minors for the linkless graphs) is generated from the complete graph using ΔY- and YΔ-transformations. the Heawood family (consisting of 78 graphs) is generated from an
https://en.wikipedia.org/wiki/Encoding%20law
In digital communications, an encoding law is a (typically non-uniform) allocation of signal quantization levels across the possible analog signal levels in an analog-to-digital converter system. They can be viewed as a simple form of instantaneous companding. The best-known encoding laws are the μ-law and A-law encoding laws defined in the ITU-T standard G.711 for use in digital telephony, and still used to the present day.
https://en.wikipedia.org/wiki/Laboratory%20rat
Laboratory rats or lab rats are strains of the subspecies Rattus norvegicus domestica which are bred and kept for scientific research. While less commonly used for research than laboratory mice, rats have served as an important animal model for research in psychology and biomedical science. Origins In 18th-century Europe, wild brown rats (Rattus norvegicus) ran rampant and this infestation fueled the industry of rat-catching. Rat-catchers would not only make money by trapping the rodents, but also by selling them for food or, more commonly, for rat-baiting. Rat-baiting was a popular sport, which involved filling a pit with rats and timing how long it took for a terrier to kill them all. Over time, breeding the rats for these contests may have produced variations in color, notably the albino and hooded varieties. The first time one of these albino mutants was brought into a laboratory for a study was in 1828 for an experiment on fasting. Over the next 30 years, rats were used for several more experiments and eventually the laboratory rat became the first animal domesticated for purely scientific reasons. In Japan, there was a widespread practice of keeping rats as a domesticated pet during the Edo period and in the 18th century guidebooks on keeping domestic rats were published by Youso Tamanokakehashi (1775) and Chingan Sodategusa (1787). Genetic analysis of 117 albino rat strains collected from all parts of the world carried out by a team led by Takashi Kuramoto at Kyoto University in 2012 showed that the albinos descended from hooded rats and all the albinos descended from a single ancestor. As there is evidence that the hooded rat was known as the "Japanese rat" in the early 20th century, Kuramoto concluded that one or more Japanese hooded rats might have been brought to Europe or the Americas and an albino rat that emerged as a product of the breeding of these hooded rats was the common ancestor of all the albino laboratory rats in use today. Use in resear
https://en.wikipedia.org/wiki/List%20of%20food%20safety%20organisations
This is a list of organizations and associated posts which are related to food safety, either as a primary interest or through statutory responsibility. National organizations are grouped by the UN geoscheme. Africa Africa Food Safety Forum (AFSF) African Food Safety Network (AFoSaN) The Food Safety and Quality Authority of The Gambia (FSQA) Tanzanian Food and Drugs Authority (TFDA) Ethiopian Food and Drug Authority (EFDA) Moroccan National Office of Food Safety (ONSSA) South African National Regulator for Compulsory Specifications (NRCS) Egyptian Food Safety Authority Ghana Food and Drugs Authority National Agency for Food and Drug Administration and Control Nigerian Food and Drugs Authority National Agency For Food And Drug Administration and Control [Kaduna State Livestock Regulatory Authority] [www.kadlra.kdsg.gov.ng] Americas Argentina Minister of Economy Secretariat of Agriculture, Livestock, Fishing and Food National Food Safety and Quality Service (SENASA) Servicio Nacional de Sanidad y Calidad Agroalimentaria Canada Minister of Agriculture Agriculture and Agri-Food Canada (AAFC) Canadian Food Inspection Agency (CFIA) Minister of Health Health Canada Health Products and Food Branch Guelph Food Technology Centre (Canada) United States of America U.S. Food and Drug Administration (FDA) Center for Food Safety and Applied Nutrition (CFSAN) International Food Protection Training Institute (IFPTI) Joint Expert Committee on Food Additives (JECFA) United States Department of Agriculture (USDA) Under Secretary for Food Safety Food Safety and Inspection Service (FSIS) Asia Saudi Arabia Saudi Food and Drug Authority Bangladesh Bangladesh Food Safety Authority (BFSA) Burma Food and Druhg Administration (Burma) China General Administration of Quality Supervision, Inspection and Quarantine State Food and Drug Administration Hong Kong SAR Centre for Food Safety India Food Safety and Standards Authority of India Indone
https://en.wikipedia.org/wiki/Foramen%20spinosum
The foramen spinosum is a small open hole in the greater wing of the sphenoid bone that gives passage to the middle meningeal artery and vein, and the meningeal branch of the mandibular nerve (sometimes it passes through the foramen ovale instead). The foramen spinosum is often used as a landmark in neurosurgery due to its close relations with other cranial foramina. It was first described by Jakob Benignus Winslow in the 18th century. Structure The foramen spinosum is a small foramen in the greater wing of the sphenoid bone of the skull. It connects the middle cranial fossa (superiorly), and infratemporal fossa (inferiorly). Contents The foramen transmits the middle meningeal artery and vein, and sometimes the meningeal branch of the mandibular nerve (it may pass through the foramen ovale instead). Relations The foramen is situated just anterior to the sphenopetrosal suture. It is located posterolateral to the foramen ovale, and anterior to the sphenoidal spine. A groove for the middle meningeal artery and vein extends anterolaterally from the foramen. Variation The foramen spinosum varies in size and location. The foramen is rarely absent, usually unilaterally, in which case the middle meningeal artery enters the cranial cavity through the foramen ovale. It may be incomplete, which may occur in almost half of the population. Conversely, in a minority of cases (less than 1%), it may also be duplicated, particularly when the middle meningeal artery is also duplicated. The foramen may pass through the sphenoid bone at the apex of the spinous process, or along its medial surface. Development In the newborn, the foramen spinosum is about 2.25 mm long and in adults about 2.56 mm. The width of the foramen variesfrom 1.05 mm to about 2.1 mm in adults. The average diameter of the foramen spinosum is 2.63 mm in adults. It is usually between 3 and 4 mm away from the foramen ovale in adults. The earliest perfect ring-shaped formation of the foramen spinosum w
https://en.wikipedia.org/wiki/Comparison%20of%20programming%20languages%20%28strings%29
This comparison of programming languages (strings) compares the features of string data structures or text-string processing for over 52 various computer programming languages. Concatenation Different languages use different symbols for the concatenation operator. Many languages use the "+" symbol, though several deviate from this. Common variants Unique variants AWK uses the empty string: two expressions adjacent to each other are concatenated. This is called juxtaposition. Unix shells have a similar syntax. Rexx uses this syntax for concatenation including an intervening space. C (along with Python) allows juxtaposition for string literals, however, for strings stored as character arrays, the strcat function must be used. COBOL uses the STRING statement to concatenate string variables. MATLAB and Octave use the syntax "[x y]" to concatenate x and y. Visual Basic and Visual Basic .NET can also use the "+" sign but at the risk of ambiguity if a string representing a number and a number are together. Microsoft Excel allows both "&" and the function "=CONCATENATE(X,Y)". Rust has the concat! macro and the format! macro, of which the latter is the most prevalent throughout the documentation and examples. String literals This section compares styles for declaring a string literal. Quoted interpolated An expression is "interpolated" into a string when the compiler/interpreter evaluates it and inserts the result in its place. Escaped quotes "Escaped" quotes means that a 'flag' symbol is used to warn that the character after the flag is used in the string rather than ending the string. Dual quoting "Dual quoting" means that whenever a quote is used in a string, it is used twice, and one of them is discarded and the single quote is then used within the string. Quoted raw "Raw" means the compiler treats every character within the literal exactly as written, without processing any escapes or interpolations. Multiline string Many languages have a syntax
https://en.wikipedia.org/wiki/Regression%20discontinuity%20design
In statistics, econometrics, political science, epidemiology, and related disciplines, a regression discontinuity design (RDD) is a quasi-experimental pretest-posttest design that aims to determine the causal effects of interventions by assigning a cutoff or threshold above or below which an intervention is assigned. By comparing observations lying closely on either side of the threshold, it is possible to estimate the average treatment effect in environments in which randomisation is unfeasible. However, it remains impossible to make true causal inference with this method alone, as it does not automatically reject causal effects by any potential confounding variable. First applied by Donald Thistlethwaite and Donald Campbell (1960) to the evaluation of scholarship programs, the RDD has become increasingly popular in recent years. Recent study comparisons of randomised controlled trials (RCTs) and RDDs have empirically demonstrated the internal validity of the design. Example The intuition behind the RDD is well illustrated using the evaluation of merit-based scholarships. The main problem with estimating the causal effect of such an intervention is the homogeneity of performance to the assignment of treatment (e.g. scholarship award). Since high-performing students are more likely to be awarded the merit scholarship and continue performing well at the same time, comparing the outcomes of awardees and non-recipients would lead to an upward bias of the estimates. Even if the scholarship did not improve grades at all, awardees would have performed better than non-recipients, simply because scholarships were given to students who were performing well before. Despite the absence of an experimental design, an RDD can exploit exogenous characteristics of the intervention to elicit causal effects. If all students above a given grade — for example 80% — are given the scholarship, it is possible to elicit the local treatment effect by comparing students around the 80% cut
https://en.wikipedia.org/wiki/Video%20Core%20Next
Video Core Next is AMD's brand for its dedicated video encoding and decoding hardware core. It is a family of hardware accelerator designs for encoding and decoding video, and is built into AMD's GPUs and APUs since AMD Raven Ridge, released January 2018. Background Video Core Next is AMD's successor to both the Unified Video Decoder and Video Coding Engine designs, which are hardware accelerators for video decoding and encoding, respectively. It can be used to decode, encode and transcode ("sync") video streams, for example, a DVD or Blu-ray Disc to a format appropriate to, for example, a smartphone. Unlike video encoding on a CPU or a general-purpose GPU, Video Core Next is a dedicated hardware core on the processor die. This application-specific integrated circuit (ASIC) allows for more power-efficient video processing. Support Video Core Next supports: MPEG-2 Decode, MPEG-4 Decode, VC-1 Decode, H.264/MPEG-4 AVC Encode/Decode, HEVC Encode/Decode, and VP9 Decode. VCN 2.0 is implemented with Navi products and the Renoir APU. The feature set remains the same as VCN 1.0. VCN 3.0 is implemented with Navi 2 products. See also Video hardware technologies Nvidia PureVideo - Nvidia GeForce 256's Motion Compensation High-Definition Video Processor Video Processing Engine Nvidia NVENC Nvidia NVDEC AMD Video Core Next - AMD Video Coding Engine - AMD Unified Video Decoder - AMD Video Shader - ATI Intel Quick Sync Video - Intel Clear Video - Intel Qualcomm Qualcomm Hexagon
https://en.wikipedia.org/wiki/Sepal
A sepal () is a part of the flower of angiosperms (flowering plants). Usually green, sepals typically function as protection for the flower in bud, and often as support for the petals when in bloom. Etymology The term sepalum was coined by Noël Martin Joseph de Necker in 1790, and derived . Collectively, the sepals are called the calyx (plural: calyces), the outermost whorl of parts that form a flower. The word calyx was adopted from the Latin , not to be confused with 'cup, goblet'. Calyx is derived from Greek 'bud, calyx, husk, wrapping' ( Sanskrit 'bud'), while is derived from Greek 'cup, goblet', and the words have been used interchangeably in botanical Latin. Description The term tepal is usually applied when the parts of the perianth are difficult to distinguish, e.g. the petals and sepals share the same color or the petals are absent and the sepals are colorful. When the undifferentiated tepals resemble petals, they are referred to as "petaloid", as in petaloid monocots, orders of monocots with brightly colored tepals. Since they include Liliales, an alternative name is lilioid monocots. Examples of plants in which the term tepal is appropriate include genera such as Aloe and Tulipa. In contrast, genera such as Rosa and Phaseolus have well-distinguished sepals and petals. The number of sepals in a flower is its merosity. Flower merosity is indicative of a plant's classification. The merosity of a eudicot flower is typically four or five. The merosity of a monocot or palaeodicot flower is three, or a multiple of three. The development and form of the sepals vary considerably among flowering plants. They may be free (polysepalous) or fused together (gamosepalous). Often, the sepals are much reduced, appearing somewhat awn-like, or as scales, teeth, or ridges. Most often such structures protrude until the fruit is mature and falls off. Examples of flowers with much-reduced perianths are found among the grasses. In some flowers, the sepals are f
https://en.wikipedia.org/wiki/Python%20syntax%20and%20semantics
The syntax of the Python programming language is the set of rules that defines how a Python program will be written and interpreted (by both the runtime system and by human readers). The Python language has many similarities to Perl, C, and Java. However, there are some definite differences between the languages. It supports multiple programming paradigms, including structured, object-oriented programming, and functional programming, and boasts a dynamic type system and automatic memory management. Python's syntax is simple and consistent, adhering to the principle that "There should be one— and preferably only one —obvious way to do it." The language incorporates built-in data types and structures, control flow mechanisms, first-class functions, and modules for better code reusability and organization. Python also uses English keywords where other languages use punctuation, contributing to its uncluttered visual layout. The language provides robust error handling through exceptions, and includes a debugger in the standard library for efficient problem-solving. Python's syntax, designed for readability and ease of use, makes it a popular choice among beginners and professionals alike. Design philosophy Python was designed to be a highly readable language. It has a relatively uncluttered visual layout and uses English keywords frequently where other languages use punctuation. Python aims to be simple and consistent in the design of its syntax, encapsulated in the mantra , from the Zen of Python. This mantra is deliberately opposed to the Perl and Ruby mantra, "there's more than one way to do it". Keywords Python has 35 keywords or reserved words; they cannot be used as identifiers. and as assert async await break class continue def del elif else except False finally for from global if import in is lambda None nonlocal not or pass raise return True try while with yield In addition, Python also has 3 soft keywords. Unlike regular hard keywords, soft keyword
https://en.wikipedia.org/wiki/Bernhard%20Riemann
Georg Friedrich Bernhard Riemann (; 17 September 1826 – 20 July 1866) was a German mathematician who made profound contributions to analysis, number theory, and differential geometry. In the field of real analysis, he is mostly known for the first rigorous formulation of the integral, the Riemann integral, and his work on Fourier series. His contributions to complex analysis include most notably the introduction of Riemann surfaces, breaking new ground in a natural, geometric treatment of complex analysis. His 1859 paper on the prime-counting function, containing the original statement of the Riemann hypothesis, is regarded as a foundational paper of analytic number theory. Through his pioneering contributions to differential geometry, Riemann laid the foundations of the mathematics of general relativity. He is considered by many to be one of the greatest mathematicians of all time. Biography Early years Riemann was born on 17 September 1826 in Breselenz, a village near Dannenberg in the Kingdom of Hanover. His father, Friedrich Bernhard Riemann, was a poor Lutheran pastor in Breselenz who fought in the Napoleonic Wars. His mother, Charlotte Ebell, died before her children had reached adulthood. Riemann was the second of six children, shy and suffering from numerous nervous breakdowns. Riemann exhibited exceptional mathematical talent, such as calculation abilities, from an early age but suffered from timidity and a fear of speaking in public. Education During 1840, Riemann went to Hanover to live with his grandmother and attend lyceum (middle school years). After the death of his grandmother in 1842, he attended high school at the Johanneum Lüneburg. In high school, Riemann studied the Bible intensively, but he was often distracted by mathematics. His teachers were amazed by his ability to perform complicated mathematical operations, in which he often outstripped his instructor's knowledge. In 1846, at the age of 19, he started studying philology and Christian
https://en.wikipedia.org/wiki/Clone%20manager
Clone Manager is a commercial bioinformatics software work suite of Sci-Ed, that supports molecular biologists with data management and allows them to perform certain in silico preanalysis. This type of bioinformatics software is used for managing, analyzing and visualizing DNA and protein sequence data essential for molecular biology. For enzyme read control, sequence processing of identical individuals, cloning simulation, graphic map drawing, primer design and analysis, global and local sequence alignment, similarity search, laboratory scale sequence assembly projects A comprehensive set of tools.
https://en.wikipedia.org/wiki/Axiom%20of%20pairing
In axiomatic set theory and the branches of logic, mathematics, and computer science that use it, the axiom of pairing is one of the axioms of Zermelo–Fraenkel set theory. It was introduced by as a special case of his axiom of elementary sets. Formal statement In the formal language of the Zermelo–Fraenkel axioms, the axiom reads: In words: Given any object A and any object B, there is a set C such that, given any object D, D is a member of C if and only if D is equal to A or D is equal to B. Or in simpler words: Given two objects, there is a set whose members are exactly the two given objects. Consequences As noted, what the axiom is saying is that, given two objects A and B, we can find a set C whose members are exactly A and B. We can use the axiom of extensionality to show that this set C is unique. We call the set C the pair of A and B, and denote it {A,B}. Thus the essence of the axiom is: Any two objects have a pair. The set {A,A} is abbreviated {A}, called the singleton containing A. Note that a singleton is a special case of a pair. Being able to construct a singleton is necessary, for example, to show the non-existence of the infinitely descending chains from the Axiom of regularity. The axiom of pairing also allows for the definition of ordered pairs. For any objects and , the ordered pair is defined by the following: Note that this definition satisfies the condition Ordered n-tuples can be defined recursively as follows: Alternatives Non-independence The axiom of pairing is generally considered uncontroversial, and it or an equivalent appears in just about any axiomatization of set theory. Nevertheless, in the standard formulation of the Zermelo–Fraenkel set theory, the axiom of pairing follows from the axiom schema of replacement applied to any given set with two or more elements, and thus it is sometimes omitted. The existence of such a set with two elements, such as { {}, { {} } }, can be deduced either from the axiom of empty set an
https://en.wikipedia.org/wiki/Biotrauma
Although the term has occasionally been used in other ways, in medical literature biotrauma is usually defined as a severe inflammatory response produced in the lungs of patients who breathe by means of a mechanical ventilator for a long period of time. The term was coined in a 1998 paper by L. N. Tremblay and A. S. Slutsky, titled Ventilator-induced injury: from barotrauma to biotrauma. The message of that paper was that barotrauma caused by pressure differentials is only one of several types of lung damage that a ventilator can produce.
https://en.wikipedia.org/wiki/Georg%20Friedrich%20Brander
Georg Friedrich Brander (* 28. November 1713 in Regensburg; † 1. April 1783 in Augsburg) was an important maker of scientific instruments. Early years Brander was born 1713 to Georg Brandner, an apothecary with roots in Nuremberg, and Sibylla Katharina Brandner (widowed Pfaffreuter, 1678–1756) from a Regensburg goldsmith family. He studied in Altdorf bei Nürnberg from 1731 to 1734 mathematics and physics under the supervision of Johann Gabriel Doppelmayr. Professional work In 1734 he moved to Augsburg (Germany), where he founded a workshop for fine mechanics. He was supported by local financier Josef von Halder with capital and in establishing a business network. Brander was the first one in Germany to build a mirror telescope in 1737. Since 1754 the workshop has produced also microscopes, and the high-resolution micrometer glass rulers were its speciality. A telescope, coupled in 1776 with a map of the starry sky, is named "starfinder". In 1778 he invented the coincidence telemeter, a device used to measure distances to remote objects. In those years, surveying was an important task and Brander delivered a substantial number of optical instruments, e.g. sextant, goniometer, leveling instrument and a predecessor of today's theodolite. He also published several writings on mechanics. He gave a precise description and manual with each of his instruments, which was unusual at that time. Brander won moderate fame by building precision machines for many European courts and academies. A collection of his instruments can be found at Deutsches Museum, München. In today's Augsburg, a street bears his name (Branderstraße). Family life He married Sabina Barbara Thennin ( 1754-02-14) and had a daughter Barbara Eurphrosina who married his business partner Christoph Caspar Hoeschel (1744-1820). After Brander's death, Hoeschel continued the fine mechanics business and could keep the high reputation for quality instruments. After Hoeschel died, his son took over and the
https://en.wikipedia.org/wiki/Vitis%20International%20Variety%20Catalogue
The Vitis International Variety Catalogue (VIVC) is a database of various species and varieties/cultivars of grapevine, the genus Vitis. VIVC is administered by the Geilweilerhof Institute for Grape Breeding (Institut für Rebenzüchtung Geilweilerhof) in Siebeldingen, Germany, and contains information from grapevine collections existing in various institutes of viticulture around the world. As of April 2009, the information in the database brought together information from 130 institutions located in 45 countries, and contains about 18,000 entries. The database was started in 1983, and has been available online since 1996. Its initial creation was supported by the International Organisation of Vine and Wine and the International Board for Plant Genetic Resources, a forerunner of Bioversity International. The purpose of the VIVC database is to provide documentation on available grapevine genetic resources, and to be a source of information to grape breeders, viticultural researchers and others. The information on grape cultivars in VIVC includes basic characteristics of the cultivars, holding institutes, passport data, and all known synonyms, which are quite numerous for many grape cultivars. In some cases, photos and genetic information (microsatellite information used for DNA profiling) are included.
https://en.wikipedia.org/wiki/Megakaryocyte-associated%20tyrosine%20kinase
Megakaryocyte-associated tyrosine-protein kinase is an enzyme that in humans is encoded by the MATK gene. The protein encoded by this gene has amino acid sequence similarity to Csk tyrosine kinase and has the structural features of the CSK subfamily: SRC homology SH2 and SH3 domains, a catalytic domain, a unique N terminus, lack of myristylation signals, lack of a negative regulatory phosphorylation site, and lack of an autophosphorylation site. This protein is thought to play a significant role in the signal transduction of hematopoietic cells. It is able to phosphorylate and inactivate Src family kinases, and may play an inhibitory role in the control of T-cell proliferation. This protein might be involved in signaling in some cases of breast cancer. Three alternatively spliced transcript variants that encode different isoforms have been described for this gene. Interactions Megakaryocyte-associated tyrosine kinase has been shown to interact with CD117 and TrkA.
https://en.wikipedia.org/wiki/Dry%20quicksand
Dry quicksand is loose sand whose bulk density is reduced by blowing air through it and which yields easily to weight or pressure. It acts similarly to normal quicksand, but it does not contain any water and does not operate on the same principle. Dry quicksand can also be a resulting phenomenon of contractive dilatancy. Historically, the existence of dry quicksand was doubted, and the reports of humans and complete caravans being lost in dry quicksand were considered to be folklore. In 2004, it was created in the laboratory, but it is still not clear what its actual prevalence in nature is. Scientific research Writing in Nature, physicist Detlef Lohse and coworkers of University of Twente in Enschede, Netherlands allowed air to flow through very fine sand (typical grain diameter was about 40 micrometers) in a container with a perforated base. They then turned the air stream off before the start of the experiment and allowed the sand to settle: the packing fraction of this sand was only 41% (compared to 55–60% for untreated sand). Lohse found that a weighted table tennis ball (radius 2 cm, mass 133 g), when released from just above the surface of the sand, would sink to about five diameters. Lohse also observed a "straight jet of sand [shooting] violently into the air after about 100 ms". Objects are known to make a splash when they hit sand, but this type of jet had never been described before. Lohse concluded that In nature, dry quicksands may evolve from the sedimentation of very fine sand after it has been blown into the air and, if large enough, might be a threat to humans. Indeed, reports that travellers and whole vehicles have been swallowed instantly may even turn out to be credible in the light of our results. During the planning of the Project Apollo Moon missions, dry quicksand on the Moon was considered as a potential danger to the missions. The successful landings of the unmanned Surveyor probes a few years earlier and their observations of a soli
https://en.wikipedia.org/wiki/Aspergillosis
Aspergillosis is a fungal infection of usually the lungs, caused by the genus Aspergillus, a common mould that is breathed in frequently from the air, but does not usually affect most people. It generally occurs in people with lung diseases such as asthma, cystic fibrosis or tuberculosis, or those who are immunocompromized such as those who have had a stem cell or organ transplant or those who take medications such as steroids and some cancer treatments which suppress the immune system. Rarely, it can affect skin. Aspergillosis occurs in humans, birds and other animals. Aspergillosis occurs in chronic or acute forms which are clinically very distinct. Most cases of acute aspergillosis occur in people with severely compromised immune systems such as those undergoing bone marrow transplantation. Chronic colonization or infection can cause complications in people with underlying respiratory illnesses, such as asthma, cystic fibrosis, sarcoidosis, tuberculosis, or chronic obstructive pulmonary disease. Most commonly, aspergillosis occurs in the form of chronic pulmonary aspergillosis (CPA), aspergilloma, or allergic bronchopulmonary aspergillosis (ABPA). Some forms are intertwined; for example ABPA and simple aspergilloma can progress to CPA. Other, noninvasive manifestations include fungal sinusitis (both allergic in nature and with established fungal balls), otomycosis (ear infection), keratitis (eye infection), and onychomycosis (nail infection). In most instances, these are less severe, and curable with effective antifungal treatment. The most frequently identified pathogens are Aspergillus fumigatus and Aspergillus flavus, ubiquitous organisms capable of living under extensive environmental stress. Most people are thought to inhale thousands of Aspergillus spores daily but without effect due to an efficient immune response. Invasive aspergillosis has a 20% mortality at 6 months. The major chronic, invasive, and allergic forms of aspergillosis account for around
https://en.wikipedia.org/wiki/Signal%20chain
Signal chain, or signal-processing chain is a term used in signal processing and mixed-signal system design to describe a series of signal-conditioning electronic components that receive input (data acquired from sampling either real-time phenomena or from stored data) sequentially, with the output of one portion of the chain supplying input to the next. Signal chains are often used in signal processing applications to gather and process data or to apply system controls based on analysis of real-time phenomena. Definition This definition comes from common usage in the electronics industry and can be derived from definitions of its parts: Signal: "The event, phenomenon, or electrical quantity, that conveys information from one point to another". Chain: "1. Any series of items linked together. 2. Pertaining to a routine consisting of segments which are run through the computer in tandem, only one segment being within the computer at any one time and each segment using the output from the previous program as its input". The concept of a signal chain is familiar to electrical engineers, but the term has many synonyms such as circuit topology. The goal of any signal chain is to process a variety of signals to monitor or control an analog-, digital-, or analog-digital system. See also Audio signal flow Daisy chain (electrical engineering) Feedback
https://en.wikipedia.org/wiki/1%2C000%2C000
One million (1,000,000), or one thousand thousand, is the natural number following 999,999 and preceding 1,000,001. The word is derived from the early Italian millione (milione in modern Italian), from mille, "thousand", plus the augmentative suffix -one. It is commonly abbreviated in British English as m (not to be confused with the metric prefix "m", milli, for , or with metre), M, MM ("thousand thousands", from Latin "Mille"; not to be confused with the Roman numeral = 2,000), mm (not to be confused with millimetre), or mn in financial contexts. In scientific notation, it is written as or 106. Physical quantities can also be expressed using the SI prefix mega (M), when dealing with SI units; for example, 1 megawatt (1 MW) equals 1,000,000 watts. The meaning of the word "million" is common to the short scale and long scale numbering systems, unlike the larger numbers, which have different names in the two systems. The million is sometimes used in the English language as a metaphor for a very large number, as in "Not in a million years" and "You're one in a million", or a hyperbole, as in "I've walked a million miles" and "You've asked a million-dollar question". 1,000,000 is also the square of 1000 and also the cube of 100. Visualizing one million Even though it is often stressed that counting to precisely a million would be an exceedingly tedious task due to the time and concentration required, there are many ways to bring the number "down to size" in approximate quantities, ignoring irregularities or packing effects. Information: Not counting spaces, the text printed on 136 pages of an Encyclopædia Britannica, or 600 pages of pulp paperback fiction contains approximately one million characters. Length: There are one million millimetres in a kilometre, and roughly a million sixteenths of an inch in a mile (1 sixteenth = 0.0625). A typical car tire might rotate a million times in a trip, while the engine would do several times that number of revolution
https://en.wikipedia.org/wiki/Wave%20tank
A wave tank is a laboratory setup for observing the behavior of surface waves. The typical wave tank is a box filled with liquid, usually water, leaving open or air-filled space on top. At one end of the tank, an actuator generates waves; the other end usually has a wave-absorbing surface. A similar device is the ripple tank, which is flat and shallow and used for observing patterns of surface waves from above. Wave basin A wave basin is a wave tank which has a width and length of comparable magnitude, often used for testing ships, offshore structures and three-dimensional models of harbors (and their breakwaters). Wave flume A wave flume (or wave channel) is a special sort of wave tank: the width of the flume is much less than its length. The generated waves are therefore – more or less – two-dimensional in a vertical plane (2DV), meaning that the orbital flow velocity component in the direction perpendicular to the flume side wall is much smaller than the other two components of the three-dimensional velocity vector. This makes a wave flume a well-suited facility to study near-2DV structures, like cross-sections of a breakwater. Also (3D) constructions providing little blockage to the flow may be tested, e.g. measuring wave forces on vertical cylinders with a diameter much less than the flume width. Wave flumes may be used to study the effects of water waves on coastal structures, offshore structures, sediment transport and other transport phenomena. The waves are most often generated with a mechanical wavemaker, although there are also wind–wave flumes with (additional) wave generation by an air flow over the water – with the flume closed above by a roof above the free surface. The wavemaker frequently consists of a translating or rotating rigid wave board. Modern wavemakers are computer controlled, and can generate besides periodic waves also random waves, solitary waves, wave groups or even tsunami-like wave motion. The wavemaker is at one end of the w
https://en.wikipedia.org/wiki/Roseimaritima
Roseimaritima is a genus of bacteria from the family of Planctomycetaceae with three known species. Roseimaritima ulvae has been isolated from an Ulva from Carreço in Portugal. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI) See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/Solid-state%20lighting
Solid-state lighting (SSL) is a type of lighting that uses semiconductor light-emitting diodes (LEDs), organic light-emitting diodes (OLED), or polymer light-emitting diodes (PLED) as sources of illumination rather than electrical filaments, plasma (used in arc lamps such as fluorescent lamps), or gas. Solid state electroluminescence is used in SSL, as opposed to incandescent bulbs (which use thermal radiation) or fluorescent tubes. Compared to incandescent lighting, SSL creates visible light with reduced heat generation and less energy dissipation. Most common "white LEDs” convert blue light from a solid-state device to an (approximate) white light spectrum using photoluminescence, the same principle used in conventional fluorescent tubes. The typically small mass of a solid-state electronic lighting device provides for greater resistance to shock and vibration compared to brittle glass tubes/bulbs and long, thin filament wires. They also eliminate filament evaporation, potentially increasing the life span of the illumination device. Solid-state lighting is often used in traffic lights and is also used in modern vehicle lights, street and parking lot lights, train marker lights, building exteriors, remote controls etc. Controlling the light emission of LEDs may be done most effectively by using the principles of nonimaging optics. Solid-state lighting has made significant advances in industry. In the entertainment lighting industry, standard incandescent tungsten-halogen lamps are being replaced by solid-state lighting fixtures. See also L Prize LED lamp List of light sources Smart lighting
https://en.wikipedia.org/wiki/Complex%20adaptive%20system
A complex adaptive system is a system that is complex in that it is a dynamic network of interactions, but the behavior of the ensemble may not be predictable according to the behavior of the components. It is adaptive in that the individual and collective behavior mutate and self-organize corresponding to the change-initiating micro-event or collection of events. It is a "complex macroscopic collection" of relatively "similar and partially connected micro-structures" formed in order to adapt to the changing environment and increase their survivability as a macro-structure. The Complex Adaptive Systems approach builds on replicator dynamics. The study of complex adaptive systems, a subset of nonlinear dynamical systems, is an interdisciplinary matter that attempts to blend insights from the natural and social sciences to develop system-level models and insights that allow for heterogeneous agents, phase transition, and emergent behavior. Overview The term complex adaptive systems, or complexity science, is often used to describe the loosely organized academic field that has grown up around the study of such systems. Complexity science is not a single theory—it encompasses more than one theoretical framework and is interdisciplinary, seeking the answers to some fundamental questions about living, adaptable, changeable systems. Complex adaptive systems may adopt hard or softer approaches. Hard theories use formal language that is precise, tend to see agents as having tangible properties, and usually see objects in a behavioral system that can be manipulated in some way. Softer theories use natural language and narratives that may be imprecise, and agents are subjects having both tangible and intangible properties. Examples of hard complexity theories include Complex Adaptive Systems (CAS) and Viability Theory, and a class of softer theory is Viable System Theory. Many of the propositional consideration made in hard theory are also of relevance to softer theory. Fr
https://en.wikipedia.org/wiki/Linear%20inequality
In mathematics a linear inequality is an inequality which involves a linear function. A linear inequality contains one of the symbols of inequality: < less than > greater than ≤ less than or equal to ≥ greater than or equal to ≠ not equal to A linear inequality looks exactly like a linear equation, with the inequality sign replacing the equality sign. Linear inequalities of real numbers Two-dimensional linear inequalities Two-dimensional linear inequalities are expressions in two variables of the form: where the inequalities may either be strict or not. The solution set of such an inequality can be graphically represented by a half-plane (all the points on one "side" of a fixed line) in the Euclidean plane. The line that determines the half-planes (ax + by = c) is not included in the solution set when the inequality is strict. A simple procedure to determine which half-plane is in the solution set is to calculate the value of ax + by at a point (x0, y0) which is not on the line and observe whether or not the inequality is satisfied. For example, to draw the solution set of x + 3y < 9, one first draws the line with equation x + 3y = 9 as a dotted line, to indicate that the line is not included in the solution set since the inequality is strict. Then, pick a convenient point not on the line, such as (0,0). Since 0 + 3(0) = 0 < 9, this point is in the solution set, so the half-plane containing this point (the half-plane "below" the line) is the solution set of this linear inequality. Linear inequalities in general dimensions In Rn linear inequalities are the expressions that may be written in the form or where f is a linear form (also called a linear functional), and b a constant real number. More concretely, this may be written out as or Here are called the unknowns, and are called the coefficients. Alternatively, these may be written as or where g is an affine function. That is or Note that any inequality containing a "greater than
https://en.wikipedia.org/wiki/Structural%20synthesis%20of%20programs
Structural synthesis of programs (SSP) is a special form of (automatic) program synthesis that is based on propositional calculus. More precisely, it uses intuitionistic logic for describing the structure of a program in such a detail that the program can be automatically composed from pieces like subroutines or even computer commands. It is assumed that these pieces have been implemented correctly, hence no correctness verification of these pieces is needed. SSP is well suited for automatic composition of services for service-oriented architectures and for synthesis of large simulation programs. History Automatic program synthesis began in the artificial intelligence field, with software intended for automatic problem solving. The first program synthesizer was developed by Cordell Green in 1969. At about the same time, mathematicians including R. Constable, Z. Manna, and R. Waldinger explained the possible use of formal logic for automatic program synthesis. Practically applicable program synthesizers appeared considerably later. The idea of structural synthesis of programs was introduced at a conference on algorithms in modern mathematics and computer science organized by Andrey Ershov and Donald Knuth in 1979. The idea originated from G. Pólya’s well-known book on problem solving. The method for devising a plan for solving a problem in SSP was presented as a formal system. The inference rules of the system were restructured and justified in logic by G. Mints and E. Tyugu in 1982. A programming tool PRIZ that uses SSP was developed in the 1980s. A recent Integrated development environment that supports SSP is CoCoViLa — a model-based software development platform for implementing domain specific languages and developing large Java programs. The logic of SSP Structural synthesis of programs is a method for composing programs from already implemented components (e.g. from computer commands or software object methods) that can be considered as functions.
https://en.wikipedia.org/wiki/Acidulated%20water
Acidulated water is water where some sort of acids is added—often lemon juice, lime juice, or vinegar—to prevent cut or skinned fruits or vegetables from browning so as to maintain their appearance. Some vegetables and fruits often placed in acidulated water are apples, avocados, celeriac, potatoes and pears. When the fruit or vegetable is removed from the mixture, it will usually resist browning for at least an hour or two, even though it is being exposed to oxygen. An added benefit of placing items in acidulated water is that the food item acquires a taste of the acid used, which can be very pleasant on the palate. Acidulated water, most often made with the use of vinegar, can be used on an aged, hanging beef carcass (butchered) to help clean it. The hanging primals / sub-primals can be wiped down with a cloth that has been submerged in the acidulated solution to help remove the "slick" surface that can build up during the aging process. See also
https://en.wikipedia.org/wiki/Fastest%20animals
This is a list of the fastest animals in the world, by types of animal. Fastest organism The peregrine falcon is the fastest bird, and the fastest member of the animal kingdom, with a diving speed of over . The fastest land animal is the cheetah. Among the fastest animals in the sea is the black marlin, with uncertain and conflicting reports of recorded speeds. When drawing comparisons between different classes of animals, an alternative unit is sometimes used for organisms: body length per second. On this basis the 'fastest' organism on earth, relative to its body length, is the Southern Californian mite, Paratarsotomus macropalpis, which has a speed of 322 body lengths per second. The equivalent speed for a human, running as fast as this mite, would be , or approximately Mach 1.7. The speed of the P. macropalpis is far in excess of the previous record holder, the Australian tiger beetle Cicindela eburneola, which is the fastest insect in the world relative to body size, with a recorded speed of , or 171 body lengths per second. The cheetah, the fastest land mammal, scores at only 16 body lengths per second, while Anna's hummingbird has the highest known length-specific velocity attained by any vertebrate. Invertebrates Fish Due to physical constraints, fish may be incapable of exceeding swim speeds of 36 km/h (22 mph). The larger reported figures below are therefore highly questionable: Amphibians Reptiles Birds Mammals See also Speed records Notes
https://en.wikipedia.org/wiki/Bohr%20compactification
In mathematics, the Bohr compactification of a topological group G is a compact Hausdorff topological group H that may be canonically associated to G. Its importance lies in the reduction of the theory of uniformly almost periodic functions on G to the theory of continuous functions on H. The concept is named after Harald Bohr who pioneered the study of almost periodic functions, on the real line. Definitions and basic properties Given a topological group G, the Bohr compactification of G is a compact Hausdorff topological group Bohr(G) and a continuous homomorphism b: G → Bohr(G) which is universal with respect to homomorphisms into compact Hausdorff groups; this means that if K is another compact Hausdorff topological group and f: G → K is a continuous homomorphism, then there is a unique continuous homomorphism Bohr(f): Bohr(G) → K such that f = Bohr(f) ∘ b. Theorem. The Bohr compactification exists and is unique up to isomorphism. We will denote the Bohr compactification of G by Bohr(G) and the canonical map by The correspondence G ↦ Bohr(G) defines a covariant functor on the category of topological groups and continuous homomorphisms. The Bohr compactification is intimately connected to the finite-dimensional unitary representation theory of a topological group. The kernel of b consists exactly of those elements of G which cannot be separated from the identity of G by finite-dimensional unitary representations. The Bohr compactification also reduces many problems in the theory of almost periodic functions on topological groups to that of functions on compact groups. A bounded continuous complex-valued function f on a topological group G is uniformly almost periodic if and only if the set of right translates gf where is relatively compact in the uniform topology as g varies through G. Theorem. A bounded continuous complex-valued function f on G is uniformly almost periodic if and only if there is a continuous function f1 on Bohr(G) (which i
https://en.wikipedia.org/wiki/Neuropilin%201
Neuropilin-1 is a protein that in humans is encoded by the NRP1 gene. In humans, the neuropilin 1 gene is located at 10p11.22. This is one of two human neuropilins. Function NRP1 is a membrane-bound coreceptor to a tyrosine kinase receptor for both vascular endothelial growth factor (for example, VEGFA) and semaphorin (for example, SEMA3A) family members. NRP1 plays versatile roles in angiogenesis, axon guidance, cell survival, migration, and invasion.[supplied by OMIM] Interactions Neuropilin 1 has been shown to interact with Vascular endothelial growth factor A. Role in COVID-19 Research has shown that neuropilin 1 facilitates entry of SARS-CoV-2 into cells, making it a possible target for future antiviral drugs. Implication in cancer Neuropilin 1 has been implicated in the vascularization and progression of cancers. NRP1 expression has been shown to be elevated in a number of human patient tumor samples, including brain, prostate, breast, colon, and lung cancers and NRP1 levels are positively correlated with metastasis. In prostate cancer NRP1 has been demonstrated to be an androgen-suppressed gene, upregulated during the adaptive response of prostate tumors to androgen-targeted therapies and a prognostic biomarker of clinical metastasis and lethal PCa. In vitro and in vivo mouse studies have shown membrane bound NRP1 to be proangiogenic and that NRP1 promotes the vascularization of prostate tumors. Elevated NRP1 expression is also correlated with the invasiveness of non-small cell lung cancer both in vitro and in vivo. Target for cancer therapies As a co-receptor for VEGF, NRP1 is a potential target for cancer therapies. A synthetic peptide, EG3287, was generated in 2005 and has been shown to block NRP1 activity. EG3287 has been shown to induce apoptosis in tumor cells with elevated NRP1 expression. A patent for EG3287 was filed in 2002 and approved in 2003. As of 2015 there were no clinical trials ongoing or completed for EG3287 as a human canc
https://en.wikipedia.org/wiki/Centre%20for%20Internet%20and%20Society%20%28India%29
The Centre for Internet and Society (CIS) is a Bengaluru-based non-profit multidisciplinary research organization. CIS works on digital pluralism, public accountability and pedagogic practices, in the field of the Internet and Society. Wikimedia Projects The Wikimedia Foundation granted a project to CIS to promote and support the Indic language Wikimedia’s Indic language free knowledge projects, including Wikipedia in Indic languages and English. The grant is also aimed to support wider distribution of Wikimedia’s free knowledge within India. The award amount for the first year of the two-year project was Rs. 11 million (US $200,000). Swatantra 2014 CIS supported and participated in Swatantra 2014, the fifth international free software conference event, organized at Thiruvananthapuram, Kerala from 18 to 20 December 2014.
https://en.wikipedia.org/wiki/Terminally%20redundant%20DNA
Terminally redundant DNA is DNA that contains repeated sequences at each end called terminal repeats. These ends are used (e.g. in virus T4) to join the ends of the linear DNA to form a cyclic DNA. The term was first coined by Dr. Michael London in 1964.
https://en.wikipedia.org/wiki/Breakthrough%20infection
A breakthrough infection is a case of illness in which a vaccinated individual becomes infected with the illness, because the vaccine has failed to provide complete immunity against the pathogen (currently only viruses). Breakthrough infections have been identified in individuals immunized against a variety of diseases including mumps, varicella (Chickenpox), influenza, and COVID-19. The characteristics of the breakthrough infection are dependent on the virus itself. Often, infection of the vaccinated individual results in milder symptoms and shorter duration than if the infection were contracted naturally. Causes of breakthrough infections include biological factors in the recipient, improper administration or storage of vaccines, mutations in viruses, blocking antibody formation, and other factors. For these reasons, vaccines are rarely 100% effective. A 2021 study found the common flu vaccine provided immunity to the flu in 58% of recipients. The measles vaccine fails to provide immunity to 2% of children that receive the vaccine. However, if herd immunity exists, it typically prevents individuals who are ineffectively vaccinated from contracting the disease. Accordingly, herd immunity reduces the number of breakthrough infections in a population. By disease Varicella The varicella vaccine is 85% effective at preventing varicella (chickenpox) infection. However, 75% of individuals that are diagnosed with breakthrough varicella exhibit milder symptoms than individuals that are not vaccinated. These individuals with mild varicella have low fevers, fewer than 50 lesions on their skin, and a maculopapular rash. In contrast, unvaccinated individuals typically have a fever of 102, 200-500 skin lesions, and macules (lesions that are not elevated) evolve to papules and vesicular lesions. Additionally, infection in unvaccinated individuals tends to last for a longer period of time than in individuals who have been vaccinated. The majority of cases of breakthrough v
https://en.wikipedia.org/wiki/Design%20rule%20checking
In electronic design automation, a design rule is a geometric constraint imposed on circuit board, semiconductor device, and integrated circuit (IC) designers to ensure their designs function properly, reliably, and can be produced with acceptable yield. Design rules for production are developed by process engineers based on the capability of their processes to realize design intent. Electronic design automation is used extensively to ensure that designers do not violate design rules; a process called design rule checking (DRC). DRC is a major step during physical verification signoff on the design, which also involves LVS (layout versus schematic) checks, XOR checks, ERC (electrical rule check), and antenna checks. The importance of design rules and DRC is greatest for ICs, which have micro- or nano-scale geometries; for advanced processes, some fabs also insist upon the use of more restricted rules to improve yield. Design rules Design rules are a series of parameters provided by semiconductor manufacturers that enable the designer to verify the correctness of a mask set. Design rules are specific to a particular semiconductor manufacturing process. A design rule set specifies certain geometric and connectivity restrictions to ensure sufficient margins to account for variability in semiconductor manufacturing processes, so as to ensure that most of the parts work correctly. The most basic design rules are shown in the diagram on the right. The first are single layer rules. A width rule specifies the minimum width of any shape in the design. A spacing rule specifies the minimum distance between two adjacent objects. These rules will exist for each layer of semiconductor manufacturing process, with the lowest layers having the smallest rules (typically 100 nm as of 2007) and the highest metal layers having larger rules (perhaps 400 nm as of 2007). A two layer rule specifies a relationship that must exist between two layers. For example, an enclosure rule might s
https://en.wikipedia.org/wiki/Field%20Hygiene%20and%20Sanitation
Field hygiene and sanitation are two facets of military medicine that seek to ensure reduction of casualties through avoidance of non-combat related health issues among military personnel, particularly in the prevention of disease. As such, it encompasses prevention of communicable diseases; promotes personal hygiene; ensures adequate field water supply; supervises food sanitation; administers waste disposal; and controls, prevents, and combats insect-borne diseases transmitted by mosquitoes, louses, flies, fleas, ticks, mites, and other insects. Field hygiene also includes knowledge, avoidance, and control of venomous animals and rodents, as well as mitigation of health problems related to extreme temperature environments. Lack of field hygiene and sanitation were major contributors to non-combat casualties and deaths in pre-modern field armies, and these remained serious threats to soldier health in modern warfare during the First World War, on the Eastern Front during the Second World War, in the Korean War, the Vietnam War and the Soviet–Afghan War. Inadequate field hygiene and sanitation are also major medical problems and causes of death among refugee populations around the world. External links http://www.marines.mil/Portals/59/Publications/MCRP%204-11.1D%20Field%20Hygiene%20and%20Sanitation.pdf Military medicine
https://en.wikipedia.org/wiki/DCache
dCache is a system for storing and retrieving huge amounts of data, distributed among a large number of heterogeneous server nodes, under a single virtual filesystem tree with a variety of standard access methods. dCache is open source software built in Java and is used by, among others, ten out of fourteen Tier1 sites to CERN to store data from the Large Hadron Collider. dCache provides methods for exchanging data with backend (tertiary) storage systems as well as space management, pool attraction, dataset replication, hot spot determination and recovery from disk or node failures. Connected to a tertiary storage system, the cache simulates unlimited direct access storage space. Data exchanges to and from the underlying hierarchical storage management system are performed automatically and invisibly to the user. Beside through protocols specific to high-energy physics, data in dCache can be accessed via NFSv4.1 as well as through WebDAV.
https://en.wikipedia.org/wiki/Vieussens%20valve%20of%20the%20coronary%20sinus
The Vieussens valve of the coronary sinus is a prominent valve at the end of the great cardiac vein, marking the commencement of the coronary sinus. It is often a flimsy valve composed of one to three leaflets. It is present in 80-90% of individuals. It serves as an anatomical landmark. It is clinically important because it is often an obstruction to catheters in 20% of patients.
https://en.wikipedia.org/wiki/Five%20lemma
In mathematics, especially homological algebra and other applications of abelian category theory, the five lemma is an important and widely used lemma about commutative diagrams. The five lemma is not only valid for abelian categories but also works in the category of groups, for example. The five lemma can be thought of as a combination of two other theorems, the four lemmas, which are dual to each other. Statements Consider the following commutative diagram in any abelian category (such as the category of abelian groups or the category of vector spaces over a given field) or in the category of groups. The five lemma states that, if the rows are exact, m and p are isomorphisms, l is an epimorphism, and q is a monomorphism, then n is also an isomorphism. The two four-lemmas state: Proof The method of proof we shall use is commonly referred to as diagram chasing. We shall prove the five lemma by individually proving each of the two four lemmas. To perform diagram chasing, we assume that we are in a category of modules over some ring, so that we may speak of elements of the objects in the diagram and think of the morphisms of the diagram as functions (in fact, homomorphisms) acting on those elements. Then a morphism is a monomorphism if and only if it is injective, and it is an epimorphism if and only if it is surjective. Similarly, to deal with exactness, we can think of kernels and images in a function-theoretic sense. The proof will still apply to any (small) abelian category because of Mitchell's embedding theorem, which states that any small abelian category can be represented as a category of modules over some ring. For the category of groups, just turn all additive notation below into multiplicative notation, and note that commutativity of abelian group is never used. So, to prove (1), assume that m and p are surjective and q is injective. Let c′ be an element of C′. Since p is surjective, there exists an element d in D with p(d) = t(c′). By co
https://en.wikipedia.org/wiki/Nature%20Human%20Behaviour
Nature Human Behaviour is a monthly multidisciplinary online-only peer-reviewed scientific journal covering all aspects of human behaviour. It was established in January 2017 and is published by Nature Portfolio. The editor-in-chief is Stavroula Kousta. According to the Journal Citation Reports, the journal has a 2022 impact factor of 29.9. Potential Harms Policy In August 2022, the journal published an editorial article called "Science must respect the dignity and rights of all humans", to counter structural inequalities and discrimination in society in science. It stated that editors have the right to amend, refuse, or retract content which can be "disparag[ing]", "exclusionary" or content that "undermines the dignity or rights of specific groups".
https://en.wikipedia.org/wiki/Coherence%20%28fairness%29
Coherence, also called uniformity or consistency, is a criterion for evaluating rules for fair division. Coherence requires that the outcome of a fairness rule is fair not only for the overall problem, but also for each sub-problem. Every part of a fair division should be fair. The coherence requirement was first studied in the context of apportionment. In this context, failure to satisfy coherence is called the new states paradox: when a new state enters the union, and the house size is enlarged to accommodate the number of seats allocated to this new state, some other unrelated states are affected. Coherence is also relevant to other fair division problems, such as bankruptcy problems. Definition There is a resource to allocate, denoted by . For example, it can be an integer representing the number of seats in a house of representatives. The resource should be allocated between some agents. For example, these can be federal states or political parties. The agents have different entitlements, denoted by a vector . For example, ti can be the fraction of votes won by party i. An allocation is a vector with . An allocation rule is a rule that, for any and entitlement vector , returns an allocation vector . An allocation rule is called coherent (or uniform) if, for every subset S of agents, if the rule is activated on the subset of the resource , and on the entitlement vector , then the result is the allocation vector . That is: when the rule is activated on a subset of the agents, with the subset of resources they received, the result for them is the same. Handling ties In general, an allocation rule may return more than one allocation (in case of a tie). In this case, the definition should be updated. Denote the allocation rule by , and Denote by the set of allocation vectors returned by on the resource and entitlement vector . The rule is called coherent if the following holds for every allocation vector and any subset S of agents: . That is, every pa
https://en.wikipedia.org/wiki/Integration%20by%20parts%20operator
In mathematics, an integration by parts operator is a linear operator used to formulate integration by parts formulae; the most interesting examples of integration by parts operators occur in infinite-dimensional settings and find uses in stochastic analysis and its applications. Definition Let E be a Banach space such that both E and its continuous dual space E∗ are separable spaces; let μ be a Borel measure on E. Let S be any (fixed) subset of the class of functions defined on E. A linear operator A : S → L2(E, μ; R) is said to be an integration by parts operator for μ if for every C1 function φ : E → R and all h ∈ S for which either side of the above equality makes sense. In the above, Dφ(x) denotes the Fréchet derivative of φ at x. Examples Consider an abstract Wiener space i : H → E with abstract Wiener measure γ. Take S to be the set of all C1 functions from E into E∗; E∗ can be thought of as a subspace of E in view of the inclusions For h ∈ S, define Ah by This operator A is an integration by parts operator, also known as the divergence operator; a proof can be found in Elworthy (1974). The classical Wiener space C0 of continuous paths in Rn starting at zero and defined on the unit interval [0, 1] has another integration by parts operator. Let S be the collection i.e., all bounded, adapted processes with absolutely continuous sample paths. Let φ : C0 → R be any C1 function such that both φ and Dφ are bounded. For h ∈ S and λ ∈ R, the Girsanov theorem implies that Differentiating with respect to λ and setting λ = 0 gives where (Ah)(x) is the Itō integral The same relation holds for more general φ by an approximation argument; thus, the Itō integral is an integration by parts operator and can be seen as an infinite-dimensional divergence operator. This is the same result as the integration by parts formula derived from the Clark-Ocone theorem.
https://en.wikipedia.org/wiki/Organization%20and%20expression%20of%20immunoglobulin%20genes
Antibody (or immunoglobulin) structure is made up of two heavy-chains and two light-chains. These chains are held together by disulfide bonds. The arrangement or processes that put together different parts of this antibody molecule play important role in antibody diversity and production of different subclasses or classes of antibodies. The organization and processes take place during the development and differentiation of B cells. That is, the controlled gene expression during transcription and translation coupled with the rearrangements of immunoglobulin gene segments result in the generation of antibody repertoire during development and maturation of B cells. B-Cell development During the development of B cells, the immunoglobulin gene undergoes sequences of rearrangements that lead to formation of the antibody repertoire. For example, in the lymphoid cell, a partial rearrangement of the heavy-chain gene occurs which is followed by complete rearrangement of heavy-chain gene. Here at this stage, Pre-B cell, mμ heavy chain and surrogate light chain are formed. The final rearrangement of the light chain gene generates immature B cell and mIgM. The process explained here occurs only in the absence of the antigen. The mature B cell formed as RNA processing changes leaves the bone marrow and is stimulated by the antigen then differentiated into IgM -secreted plasma cells. Also at first, the mature B cell expresses membrane-bound IgD and IgM. These two classes could switch to secretory IgD and IgM during the processing of mRNAs. Finally, further class switching follows as the cell keep dividing and differentiating. For instance, IgM switches to IgG which switches to IgA that eventually switches to IgE The multigene organization of immunoglobulin genes From studies and predictions such as Dreyer and Bennett's, it shows that the light chains and heavy chains are encoded by separate multigene families on different chromosomes. They are referred to as gene segm
https://en.wikipedia.org/wiki/Muirfield%20Seamount
The Muirfield Seamount is a submarine mountain located in the Indian Ocean approximately 130 kilometres (70 nautical miles) southwest of the Cocos (Keeling) Islands. The Cocos Islands are an Australian territory, and therefore the Muirfield Seamount is within Australia's Exclusive Economic Zone (EEZ). The Muirfield Seamount is a submerged archipelago, approximately in diameter and below the surface of the sea. A 1999 biological survey of the seamount performed by the Australian Commonwealth Scientific and Industrial Research Organisation (CSIRO) revealed that the area is depauperate. The Muirfield Seamount was discovered accidentally in 1973 when the cargo ship MV Muirfield (a merchant vessel named after Muirfield, Scotland) was underway in waters charted at a depth of greater than , when she suddenly struck an unknown object, resulting in extensive damage to her keel. In 1983, , a Royal Australian Navy survey ship, surveyed the area where Muirfield was damaged, and charted in detail this previously unsuspected hazard to navigation. The dramatic accidental discovery of the Muirfield Seamount is often cited as an example of limitations in the vertical datum accuracy of some offshore areas as represented on nautical chart especially on small-scale charts. More recently, in 2005 the submarine ran into an uncharted seamount about 560 kilometers (350 statute miles) south of Guam at a speed of , sustaining serious damage and killing one seaman. See also Graveyard Seamounts Jasper Seamount Muirfield Reef Mud volcano Sedlo Seamount South Chamorro Seamount
https://en.wikipedia.org/wiki/Infinitism
Infinitism is the view that knowledge may be justified by an infinite chain of reasons. It belongs to epistemology, the branch of philosophy that considers the possibility, nature, and means of knowledge. Epistemological infinitism Since Gettier, "knowledge" is no longer widely accepted as meaning "justified true belief" only. However, some epistemologists still consider knowledge to have a justification condition. Traditional theories of justification (foundationalism and coherentism) and indeed some philosophers consider an infinite regress not to be a valid justification. In their view, if A is justified by B, B by C, and so forth, then either The chain must end with a link that requires no independent justification (a foundation), The chain must come around in a circle in some finite number of steps (the belief may be justified by its coherence), or Our beliefs must not be justified after all (as is posited by philosophical skeptics). Infinitism, the view, for example, of Peter D. Klein, challenges this consensus, referring back to work of Paul Moser (1984) and John Post (1987). In this view, the evidential ancestry of a justified belief must be infinite and non-repeating, which follows from the conjunction of two principles that Klein sees as having straightforward intuitive appeal: "The Principle of Avoiding Circularity" and "The Principle of Avoiding Arbitrariness." The Principle of Avoiding Circularity (PAC) is stated as follows: "For all x, if a person, S, has a justification for x, then for all y, if y is in the evidential ancestry of x for S, then x is not in the evidential ancestry of y for S." PAC says that the proposition to be justified cannot be a member of its own evidential ancestry, which is violated by coherence theories of justification. The Principle of Avoiding Arbitrariness (PAA) is stated as follows: "For all x, if a person, S, has a justification for x, then there is some reason, r1, available to S for x; and there is some reaso
https://en.wikipedia.org/wiki/List%20of%20Portuguese%20flags
This is a list of flags used in Portugal. National flag Autonomous regions Municipalities Government flags Military flags Army staff Navy staff Historical flags County of Portugal Kingdom of Portugal Portuguese Macau Military Flags Merchant marine flags Navy flags Royal banners Governmental flags Chartered trading company flags Colonial authorities flag Proposed flags Overseas provinces Capitals of former overseas territories Political flags Ethnic groups flags See also A Portuguesa Coat of arms of Portugal List of personal standards of the Kings of Portugal List of Portuguese municipal flags List of Macanese flags Portuguese vexillology External links Portugal Flags of Portugal Flags
https://en.wikipedia.org/wiki/Vesicle-associated%20membrane%20protein
Vesicle associated membrane proteins (VAMPs) are a family of SNARE proteins with similar structure, and are mostly involved in vesicle fusion. VAMP1 and VAMP2 proteins known as synaptobrevins are expressed in brain and are constituents of the synaptic vesicles, where they participate in neurotransmitter release. VAMP3 (known as cellubrevin) is ubiquitously expressed and participates in regulated and constitutive exocytosis as a constituent of secretory granules and secretory vesicles. VAMP5 and VAMP7 participate in constitutive exocytosis. VAMP5 is a constituent of secretory vesicles, myotubes and tubulovesicular structures. VAMP7 is found both in secretory granules and endosomes. VAMP8 (known as endobrevin) participates in endocytosis and is found in early endosomes. VAMP8 also participates the regulated exocytosis in pancreatic acinar cells. VAMP4 is involved in transport from the Golgi.
https://en.wikipedia.org/wiki/Compression%20artifact
A compression artifact (or artefact) is a noticeable distortion of media (including images, audio, and video) caused by the application of lossy compression. Lossy data compression involves discarding some of the media's data so that it becomes small enough to be stored within the desired disk space or transmitted (streamed) within the available bandwidth (known as the data rate or bit rate). If the compressor cannot store enough data in the compressed version, the result is a loss of quality, or introduction of artifacts. The compression algorithm may not be intelligent enough to discriminate between distortions of little subjective importance and those objectionable to the user. The most common digital compression artifacts are DCT blocks, caused by the discrete cosine transform (DCT) compression algorithm used in many digital media standards, such as JPEG, MP3, and MPEG video file formats. These compression artifacts appear when heavy compression is applied, and occur often in common digital media, such as DVDs, common computer file formats such as JPEG, MP3 and MPEG files, and some alternatives to the compact disc, such as Sony's MiniDisc format. Uncompressed media (such as on Laserdiscs, Audio CDs, and WAV files) or losslessly compressed media (such as FLAC or PNG) do not suffer from compression artifacts. The minimization of perceivable artifacts is a key goal in implementing a lossy compression algorithm. However, artifacts are occasionally intentionally produced for artistic purposes, a style known as glitch art or datamoshing. Technically speaking, a compression artifact is a particular class of data error that is usually the consequence of quantization in lossy data compression. Where transform coding is used, it typically assumes the form of one of the basis functions of the coder's transform space. Images When performing block-based discrete cosine transform (DCT) coding for quantization, as in JPEG-compressed images, several types of artifacts can
https://en.wikipedia.org/wiki/Negative%20thermal%20expansion
Negative thermal expansion (NTE) is an unusual physicochemical process in which some materials contract upon heating, rather than expand as most other materials do. The most well-known material with NTE is water at 0 to 3.98 °C. Also, the density of water ice is smaller than the density of liquid water. Water's NTE is the reason why water ice floats, rather than sinks, in liquid water. Materials which undergo NTE have a range of potential engineering, photonic, electronic, and structural applications. For example, if one were to mix a negative thermal expansion material with a "normal" material which expands on heating, it could be possible to use it as a thermal expansion compensator that might allow for forming composites with tailored or even close to zero thermal expansion. Origin of negative thermal expansion There are a number of physical processes which may cause contraction with increasing temperature, including transverse vibrational modes, rigid unit modes and phase transitions. In 2011, Liu et al. showed that the NTE phenomenon originates from the existence of high pressure, small volume configurations with higher entropy, with their configurations present in the stable phase matrix through thermal fluctuations. They were able to predict both the colossal positive thermal expansion (In cerium) and zero and infinite negative thermal expansion (in ). Alternatively, large negative and positive thermal expansion may result from the design of internal microstructure. Negative thermal expansion in close-packed systems Negative thermal expansion is usually observed in non-close-packed systems with directional interactions (e.g. ice, graphene, etc.) and complex compounds (e.g. , , beta-quartz, some zeolites, etc.). However, in a paper, it was shown that negative thermal expansion (NTE) is also realized in single-component close-packed lattices with pair central force interactions. The following sufficient condition for potential giving rise to NTE behavio
https://en.wikipedia.org/wiki/Isotropic%20quadratic%20form
In mathematics, a quadratic form over a field F is said to be isotropic if there is a non-zero vector on which the form evaluates to zero. Otherwise the quadratic form is anisotropic. More explicitly, if q is a quadratic form on a vector space V over F, then a non-zero vector v in V is said to be isotropic if . A quadratic form is isotropic if and only if there exists a non-zero isotropic vector (or null vector) for that quadratic form. Suppose that is quadratic space and W is a subspace of V. Then W is called an isotropic subspace of V if some vector in it is isotropic, a totally isotropic subspace if all vectors in it are isotropic, and an anisotropic subspace if it does not contain any (non-zero) isotropic vectors. The of a quadratic space is the maximum of the dimensions of the totally isotropic subspaces. A quadratic form q on a finite-dimensional real vector space V is anisotropic if and only if q is a definite form: either q is positive definite, i.e. for all non-zero v in V ; or q is negative definite, i.e. for all non-zero v in V. More generally, if the quadratic form is non-degenerate and has the signature , then its isotropy index is the minimum of a and b. An important example of an isotropic form over the reals occurs in pseudo-Euclidean space. Hyperbolic plane Let F be a field of characteristic not 2 and . If we consider the general element of V, then the quadratic forms and are equivalent since there is a linear transformation on V that makes q look like r, and vice versa. Evidently, and are isotropic. This example is called the hyperbolic plane in the theory of quadratic forms. A common instance has F = real numbers in which case and are hyperbolas. In particular, is the unit hyperbola. The notation has been used by Milnor and Husemoller for the hyperbolic plane as the signs of the terms of the bivariate polynomial r are exhibited. The affine hyperbolic plane was described by Emil Artin as a quadratic space with basis satisf
https://en.wikipedia.org/wiki/Infratemporal%20crest
The lateral surface of the greater wing of the sphenoid is convex, and divided by a transverse ridge, the infratemporal crest, into two portions. The superior or temporal portion, convex from above downward, concave from before backward, forms a part of the temporal fossa, and gives attachment to the Temporalis; the inferior or infratemporal, smaller in size and concave, enters into the formation of the infratemporal fossa, and, together with the infratemporal crest, affords attachment to the Pterygoideus externus. Additional images
https://en.wikipedia.org/wiki/Anthropogenic%20biome
Anthropogenic biomes, also known as anthromes, human biomes or intensive land-use biome, describe the terrestrial biosphere (biomes) in its contemporary, human-altered form using global ecosystem units defined by global patterns of sustained direct human interaction with ecosystems. Anthromes are generally composed of heterogeneous mosaics of different land uses and land covers, including significant areas of fallow or regenerating habitats. Origin and evolution of the concept Anthromes were first named and mapped by Erle Ellis and Navin Ramankutty in their 2008 paper, "Putting People in the Map: Anthropogenic Biomes of the World". Anthrome maps now appear in numerous textbooks. and in the National Geographic World Atlas. The most recent version of anthrome maps were published in 2021. In a recent global ecosystem classification, anthropogenic biomes have been incorporated into several distinct functional biomes in the terrestrial and freshwater realms, and additional units have been described for the freshwater, marine, subterranean and transitional realms to create a more comprehensive description of all ecosystems created and maintained by human activities. The intensive land-use biome comprises five distinct terrestrial ecosystem functional groups: pastures, crops, plantations, urban and semi-natural ecosystem functional group. The artificial wetlands biome in the freshwater realm includes large reservoirs and other constructed wetlands, rice paddies, aquafarms and networks of canals and ditches. The anthropogenic marine biome in the marine realm includes submerged artificial structures and marine aquafarms. The anthropogenic subterranean voids biome includes industrial excavations or artificial cave-like systems. There are two additional biomes in transitions between realms: the anthropogenic shoreline biome includes artificial shorelines; the anthropogenic subterranean freshwaters biome includes water pipes, subterranean canals and flooded mines. Anthropo
https://en.wikipedia.org/wiki/Absorptance
In the study of heat transfer, absorptance of the surface of a material is its effectiveness in absorbing radiant energy. It is the ratio of the absorbed to the incident radiant power. Mathematical definitions Hemispherical absorptance Hemispherical absorptance of a surface, denoted is defined as where is the radiant flux absorbed by that surface; is the radiant flux received by that surface. Spectral hemispherical absorptance Spectral hemispherical absorptance in frequency and spectral hemispherical absorptance in wavelength of a surface, denoted and respectively, are defined as where is the spectral radiant flux in frequency absorbed by that surface; is the spectral radiant flux in frequency received by that surface; is the spectral radiant flux in wavelength absorbed by that surface; is the spectral radiant flux in wavelength received by that surface. Directional absorptance Directional absorptance of a surface, denoted , is defined as where is the radiance absorbed by that surface; is the radiance received by that surface. Spectral directional absorptance Spectral directional absorptance in frequency and spectral directional absorptance in wavelength of a surface, denoted and respectively, are defined as where is the spectral radiance in frequency absorbed by that surface; is the spectral radiance received by that surface; is the spectral radiance in wavelength absorbed by that surface; is the spectral radiance in wavelength received by that surface. Other radiometric coefficients
https://en.wikipedia.org/wiki/Bioprocess
A bioprocess is a specific process that uses complete living cells or their components (e.g., bacteria, enzymes, chloroplasts) to obtain desired products. Transport of energy and mass is fundamental to many biological and environmental processes. Areas, from food processing (including brewing beer) to thermal design of buildings to biomedical devices, manufacture of monoclonal antibodies to pollution control and global warming, require knowledge of how energy and mass can be transported through materials (momentum, heat transfer, etc.). Cell bioprocessing Cell therapy bioprocessing is a discipline that bridges the fields of cell therapy and bioprocessing (i.e., biopharmaceutical manufacturing), and is a sub-field of bioprocess engineering. The goals of cell therapy bioprocessing are to establish reproducible and robust manufacturing processes for the production of therapeutic cells. Commercially relevant bioprocesses will: Produce products that maintain all of the quality standards of biopharmaceutical drugs Supply both clinical and commercial quantities of therapeutic cells throughout the various stages of development. The processes and production technologies must be scalable, and Control the cost of goods (CoGs) of the final drug product. This aspect is critical to building the foundation for a commercially viable industry. Upstream bioprocessing Therapeutic cell manufacturing processes can be separated into upstream processes and downstream processes. The upstream process is defined as the entire process from early cell isolation and cultivation, to cell banking and culture expansion of the cells until final harvest (termination of the culture and collection of the live cell batch). Aside from technology challenges, concerning the scalability of culture apparatus, a number of raw material supply risks have emerged in recent years, including the availability of GMP grade fetal bovine serum. The upstream part of a bioprocess refers to the first step
https://en.wikipedia.org/wiki/Lothar%20Seegers
Lothar Seegers (3 October 1947 – 6 August 2018) was a German ichthyologist. Seegers authored 9 species within the family of Rivulidae. Publications (selection) The Fishes of the Lake Rukwa Drainage. Koninklijk Museum voor Midden-Afrika, 1996. Killifishes of the world: Old World Killis II. A.C.S. GmbH, 1997. Killifishes of the World: New World Killis. A.C.S., 2000. The catfishes of Africa: A Handbook for Identification and Maintenance. Aqualog Verlag, 2008. See also :Category:Taxa named by Lothar Seegers
https://en.wikipedia.org/wiki/Foundations%20of%20geometry
Foundations of geometry is the study of geometries as axiomatic systems. There are several sets of axioms which give rise to Euclidean geometry or to non-Euclidean geometries. These are fundamental to the study and of historical importance, but there are a great many modern geometries that are not Euclidean which can be studied from this viewpoint. The term axiomatic geometry can be applied to any geometry that is developed from an axiom system, but is often used to mean Euclidean geometry studied from this point of view. The completeness and independence of general axiomatic systems are important mathematical considerations, but there are also issues to do with the teaching of geometry which come into play. Axiomatic systems Based on ancient Greek methods, an axiomatic system is a formal description of a way to establish the mathematical truth that flows from a fixed set of assumptions. Although applicable to any area of mathematics, geometry is the branch of elementary mathematics in which this method has most extensively been successfully applied. There are several components of an axiomatic system. Primitives (undefined terms) are the most basic ideas. Typically they include objects and relationships. In geometry, the objects are things like points, lines and planes while a fundamental relationship is that of incidence – of one object meeting or joining with another. The terms themselves are undefined. Hilbert once remarked that instead of points, lines and planes one might just as well talk of tables, chairs and beer mugs. His point being that the primitive terms are just empty shells, place holders if you will, and have no intrinsic properties. Axioms (or postulates) are statements about these primitives; for example, any two points are together incident with just one line (i.e. that for any two points, there is just one line which passes through both of them). Axioms are assumed true, and not proven. They are the building blocks of geometric concepts,