source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Condenser%20%28heat%20transfer%29 | In systems involving heat transfer, a condenser is a heat exchanger used to condense a gaseous substance into a liquid state through cooling. In so doing, the latent heat is released by the substance and transferred to the surrounding environment. Condensers are used for efficient heat rejection in many industrial systems. Condensers can be made according to numerous designs and come in many sizes ranging from rather small (hand-held) to very large (industrial-scale units used in plant processes). For example, a refrigerator uses a condenser to get rid of heat extracted from the interior of the unit to the outside air.
Condensers are used in air conditioning, industrial chemical processes such as distillation, steam power plants, and other heat-exchange systems. The use of cooling water or surrounding air as the coolant is common in many condensers.
History
The earliest laboratory condenser, a "Gegenstromkühler" (counter-flow condenser), was invented in 1771 by the Swedish-German chemist Christian Weigel. By the mid-19th century, German chemist Justus von Liebig would provide his own improvements on the preceding designs of Weigel and Johann Friedrich August Göttling, with the device becoming known as the Liebig condenser.
Principle of operation
A condenser is designed to transfer heat from a working fluid (e.g. water in a steam power plant) to a secondary fluid or the surrounding air. The condenser relies on the efficient heat transfer that occurs during phase changes, in this case during the condensation of a vapor into a liquid. The vapor typically enters the condenser at a temperature above that of the secondary fluid. As the vapor cools, it reaches the saturation temperature, condenses into liquid, and releases large quantities of latent heat. As this process occurs along the condenser, the quantity of vapor decreases and the quantity of liquid increases; at the outlet of the condenser, only liquid remains. Some condenser designs contain an additional lengt |
https://en.wikipedia.org/wiki/Chiemgau%20impact%20hypothesis | Chiemgau impact hypothesis is an obsolete scientific theory that claimed the Tüttensee lake in southern Bavaria, Germany, to be the result of a Holocene meteorite impact. This claim has been refuted by geological research and the finding of a soil horizon of undisturbed peat and sedimentation since the end of the last glaciation period. The lake is in fact one of many kettles under the foothills of the Bavarian alps.
The claims of an impact crater had been raised by a team of hobby-archaeologists, calling themselves the CIRT (Chiemgau impact research team), and have resulted in some media reports in Germany and discussions in the local tourism industry, but are not accepted beyond the CIRT team today. |
https://en.wikipedia.org/wiki/New%20Zealand%20Food%20Safety |
New Zealand Food Safety (NZFS), or Haumaru Kai Aotearoa, is the New Zealand government body responsible for food safety, and is the controlling authority for imports and exports of food and food-related products. In April 2012 it was merged into the Ministry for Primary Industries.
The NZFSA administered legislation covering:
food for sale in New Zealand
primary processing of animal products and official assurances related to their export
exports of plant products and the controls surrounding registration, and
use of agricultural compounds and veterinary medicines.
In July 2007 the NZFSA was separated from the Ministry of Agriculture and Forestry to form a new Public Service Department. On 1 July 2010, the New Zealand Food Safety Authority (NZFSA) was amalgamated back into the Ministry of Agriculture and Forestry.
Food Bill 160-2
Food Bill 160-2 was introduced on 26 May 2010 to make some fundamental changes to New Zealand's domestic food regulatory regime. Significantly, for an export-led economic recovery for New Zealand, the domestic food regulatory regime is the platform for exports. The New Zealand domestic standard is used as the basis for negotiating equivalence arrangements with trading partners. This minimizes the excessive importing country requirements that may be imposed but which do not go to food safety. If passed into law and fully implemented, it would replace the Food Act 1981 and the Food Hygiene Regulations 1974. Food Bill will also make consequential amendments to the Animal Products Act 1999 and the Wine Act 2003 to improve the interface of regulatory processes across food sectors.
MAF (through NZFSA) managed New Zealand's participation in Codex Alimentarius and set strategic priorities which ensure that Codex standards have the widest possible application.
See also
Food safety in New Zealand
Food Standards Australia New Zealand
Food Act 1981 - New Zealand |
https://en.wikipedia.org/wiki/Lists%20of%20problems | The following articles contain lists of problems:
List of undecidable problems
Lists of unsolved problems
List of NP-complete problems
List of PSPACE-complete problems |
https://en.wikipedia.org/wiki/Helmut%20Schreyer | Helmut Theodor Schreyer (4 July 1912 – 12 December 1984) was a German inventor. He is mostly known for his work on the Z3, one of the first personal computers (PC).
Early life
Helmut Schreyer was the son of the minister Paul Schreyer and Martha. When his father started to work in a parish in Mosbach, the young Schreyer went to a school there. He earned his Abitur in 1933.
Career
Schreyer started to study electronic and telecommunications engineering at the Technical University of Berlin in 1934. He got to know Konrad Zuse at the company AV Motiv in 1935. In 1938 Schreyer earned his diploma and then worked as a graduate assistant for Prof Wilhelm Stäblein. Another assistant of Stäblein was Herbert Raabe, who had worked at AEG's research division until 1936.
World War II
In 1939, when World War II started, Schreyer applied for exemption from the drafting for military service, on the basis that his work was important for the war efforts of Nazi Germany. Schreyer submitted to the German government a plan to build a large electronic computer. This plan was rejected by the Nazi German military, because the war was expected to only last a couple of years and building the electronic computer Schreyer envisaged, would have taken much longer. Among others, Schreyer worked on detection technology for unexploded ordnance. He then worked on the accelerometer for the V-2-rocket. Schreyer's prototype of this accelerometer was destroyed, when he fled to Vienna on a train, during the last days of World War II. Schreyer also worked on technology to convert the radar signal into an audio signal which the pilot of a fighter aircraft might recognize.
Konrad Zuse invented and built the so called Z-series of personal computers between 1936 and 1945. Zuse was a schoolmate and co-worker of Schreyer, who advised Zuse on relays. Subsequently, Zuse built the Z3 computer, integrating relays as arithmetic logic unit. The Z3 computer was completed in 1941 and used 2,600 relays, with the di |
https://en.wikipedia.org/wiki/Regional%20Red%20List | A Regional Red List is a report of the threatened status of species within a certain country or region. It is based on the IUCN Red List of Threatened Species, an inventory of the conservation status of species on a global scale. Regional Red Lists assess the risk of extinction to species within a political management unit and therefore may feed directly into national and regional planning. This project is coordinated by the Zoological Society of London, the World Conservation Union (IUCN) and partners in national governments, universities and organizations throughout the world.
Regional Red Lists may assist countries or regions in:
Determining the conservation status and trends of species
Identifying species or ecosystems under greatest threat
Informing conservation planning and priority setting
Raising awareness of threatened species
Assessing extinction risk on a regional scale
The IUCN Categories and Criteria were initially designed to assess the conservation status of species globally, however there was a demand for guidelines to apply the system at the regional level. In 2003, IUCN developed a set of transparent, quantitative criteria to assess the conservation status of species at the regional and national level. This approach is now being applied in many countries throughout the world.
Recently, Regional Red Lists have been completed for Mongolian Mammals and Fishes. These have also been accompanied by Summary Conservation Action Plans, detailing recommended conservation measures for each threatened species.
Creating a Regional Red List
A Regional Red List may be created by any country or organisation by following the clear, repeatable protocol. The process is as follows:
All information relevant to a species conservation status is collected, including species distribution, population trend information, habitat, ecology and life history information, threats to the species and conservation measures currently in place.
A conservation assess |
https://en.wikipedia.org/wiki/Z-Push | Z-Push (presumably Z is for Zarafa) is a FOSS implementation of the Microsoft Exchange ActiveSync protocol which is used to synchronize email, personal contacts and other items between a central server and a mobile device.
Z-Push enables any PHP and non-PHP based groupware package to become fully sync-able with any ActiveSync compliant device with an appropriate backend.
Currently, Z-Push includes four backends: the IMAP and the Maildir backends for e-mail synchronization, the vCard and the CardDAV backends for contact synchronization, CalDAV for calendar synchronization, stickynotes for Sticky Notes Synchronization and one for the Kopano and Zarafa package which is sold by allowing full synchronization of E-mail, Calendar, Contacts, and Tasks.
There is also a 3rd party project that implements a Zimbra Backend, allowing Z-Push to be used with a ZCS server (Including opensource edition).
Since version 2.3.0, released in July 2016, significant performance improvements have been achieved, as well as significantly lower memory usage. Connecting to Outlook 2013 and 2016 via EAS is also officially supported. With the optional Kopano Outlook Extension (available only for paid subscribers of Zarafa/Kopano), additional Outlook features are enabled such as Out of Office replies and opening of shared and public folders.
The future sustainability of the Z-Push project was in question after Kopano (formerly Zarafa) announced that they would no longer be supporting the project. Z-Push now has a new maintainer from the community for future support of the project.
Technical background and architecture
The ActiveSync protocol is a binary XML (WBXML) protocol across HTTP. The protocol is specifically designed with efficient use from mobile devices in mind. As such the protocol is optimized for low bandwidth, high latency connections. Also the protocol is designed for minimum number of whole request round trips. This means that the protocol can use many of the same techniques |
https://en.wikipedia.org/wiki/Acoustic%20enhancement | Acoustic enhancement is a subtle type of sound reinforcement system used to augment direct, reflected, or reverberant sound. While sound reinforcement systems are usually used to increase the sound level of the sound source (like a person speaking into a microphone, or musical instruments in a pop ensemble), acoustic enhancement systems are typically used to increase the acoustic energy in the venue in a manner that is not noticed by the audience. The correctly installed systems replicate the desired acoustics of early reflections and reverberation from a room that is properly designed for Acoustic Music. An additional benefit of these systems is that the room acoustics can be changed or adjusted to be matched to the type of performance. The use of Acoustic Enhancement as Electronic Architecture offers a good solution for multi-use performance halls that need to be "dead" for amplified music , and are used occasionally for acoustic performances. These systems are often associated with acoustic sound sources like a chamber orchestra, symphony orchestra, or opera, but have also found acceptance in a variety of applications and venues that include rehearsal rooms, recording facilities conference rooms, sound stages, sports arenas, and outdoor venues.
Design and application
Acoustic enhancement systems use microphones, amplifiers, and loudspeakers interconnected with some form of processing. The number, type, and placement of microphones and loudspeakers varies according to both the application, as well as the physics limitations that are imposed by the inherent operating principles associated with each manufacturer's equipment. In most instances, however, these systems employ at least one array of loudspeakers that are distributed throughout the venue.
As concertgoers have become aware of the use of these systems, debates have arisen, because "...purists maintain that the natural acoustic sound of [Classical] voices [or] instruments in a given hall should n |
https://en.wikipedia.org/wiki/Law%20of%20continuity | The law of continuity is a heuristic principle introduced by Gottfried Leibniz based on earlier work by Nicholas of Cusa and Johannes Kepler. It is the principle that "whatever succeeds for the finite, also succeeds for the infinite". Kepler used the law of continuity to calculate the area of the circle by representing it as an infinite-sided polygon with infinitesimal sides, and adding the areas of infinitely many triangles with infinitesimal bases. Leibniz used the principle to extend concepts such as arithmetic operations from ordinary numbers to infinitesimals, laying the groundwork for infinitesimal calculus. The transfer principle provides a mathematical implementation of the law of continuity in the context of the hyperreal numbers.
A related law of continuity concerning intersection numbers in geometry was promoted by Jean-Victor Poncelet in his "Traité des propriétés projectives des figures".
Leibniz's formulation
Leibniz expressed the law in the following terms in 1701:
In any supposed continuous transition, ending in any terminus, it is permissible to institute a general reasoning, in which the final terminus may also be included (Cum Prodiisset).
In a 1702 letter to French mathematician Pierre Varignon subtitled “Justification of the Infinitesimal Calculus by that of Ordinary Algebra," Leibniz adequately summed up the true meaning of his law, stating that "the rules of the finite are found to succeed in the infinite."
The law of continuity became important to Leibniz's justification and conceptualization of the infinitesimal calculus.
See also
Transcendental law of homogeneity |
https://en.wikipedia.org/wiki/Journal%20of%20Radiological%20Protection | Journal of Radiological Protection is a quarterly peer-reviewed scientific journal covering radiobiological research on all aspects of radiological protection, including non-ionizing as well as ionizing radiations. It is the official journal of the Society for Radiological Protection and published on their behalf by IOP Publishing. It was established in 1981 as the Journal of the Society for Radiological Protection, before obtaining its current name in 1988. The editor-in-chief is Mike THorne.
Abstracting and indexing
The journal is abstracted and indexed in:
According to the Journal Citation Reports, the journal had a 2020 impact factor of 1.394.
See also
Centre for Radiation, Chemical and Environmental Hazards (CRCE) in Oxfordshire |
https://en.wikipedia.org/wiki/Direct%20image%20with%20compact%20support | In mathematics, the direct image with compact (or proper) support is an image functor for sheaves that extends the compactly supported global sections functor to the relative setting. It is one of Grothendieck's six operations.
Definition
Let f: X → Y be a continuous mapping of locally compact Hausdorff topological spaces, and let Sh(–) denote the category of sheaves of abelian groups on a topological space. The direct image with compact (or proper) support is the functor
f!: Sh(X) → Sh(Y)
that sends a sheaf F on X to the sheaf f!(F) given by the formula
f!(F)(U) := {s ∈ F(f −1(U)) | f|supp(s): supp(s) → U is proper}
for every open subset U of Y. Here, the notion of a proper map of spaces is unambiguous since the spaces in question are locally compact Hausdorff. This defines f!(F) as a subsheaf of the direct image sheaf f∗(F), and the functoriality of this construction then follows from basic properties of the support and the definition of sheaves.
The assumption that the spaces be locally compact Hausdorff is imposed in most sources (e.g., Iversen or Kashiwara–Schapira). In slightly greater generality, Olaf Schnürer and Wolfgang Soergel have introduced the notion of a "locally proper" map of spaces and shown that the functor of direct image with compact support remains well-behaved when defined for separated and locally proper continuous maps between arbitrary spaces.
Properties
If f is proper, then f! equals f∗.
If f is an open embedding, then f! identifies with the extension by zero functor. |
https://en.wikipedia.org/wiki/Katz%27s%20back-off%20model | Katz back-off is a generative n-gram language model that estimates the conditional probability of a word given its history in the n-gram. It accomplishes this estimation by backing off through progressively shorter history models under certain conditions. By doing so, the model with the most reliable information about a given history is used to provide the better results.
The model was introduced in 1987 by Slava M. Katz. Prior to that, n-gram language models were constructed by training individual models for different n-gram orders using maximum likelihood estimation and then interpolating them together.
Method
The equation for Katz's back-off model is:
where
C(x) = number of times x appears in training
wi = ith word in the given context
Essentially, this means that if the n-gram has been seen more than k times in training, the conditional probability of a word given its history is proportional to the maximum likelihood estimate of that n-gram. Otherwise, the conditional probability is equal to the back-off conditional probability of the (n − 1)-gram.
The more difficult part is determining the values for k, d and α.
is the least important of the parameters. It is usually chosen to be 0. However, empirical testing may find better values for k.
is typically the amount of discounting found by Good–Turing estimation. In other words, if Good–Turing estimates as , then
To compute , it is useful to first define a quantity β, which is the left-over probability mass for the (n − 1)-gram:
Then the back-off weight, α, is computed as follows:
The above formula only applies if there is data for the "(n − 1)-gram". If not, the algorithm skips n-1 entirely and uses the Katz estimate for n-2. (and so on until an n-gram with data is found)
Discussion
This model generally works well in practice, but fails in some circumstances. For example, suppose that the bigram "a b" and the unigram "c" are very common, but the trigram "a b c" is never seen. Since "a b" and "c" |
https://en.wikipedia.org/wiki/Friction%20motor | A friction motor is a simple mechanism to propel toy cars, trucks, trains, action figures and similar toys. The motor consists of a large flywheel which is connected to the drive wheels of the toy via a very low gear ratio, so that the flywheel revolves faster. The flywheel's axis is perpendicular to the direction in which the toy faces and moves. When the toy is pushed forward, the drive wheels engage the flywheel. Pushing the vehicle forward repeatedly spins this flywheel up to speed. When let go, the flywheel drives the vehicle forward. The flywheel stores the kinetic energy of the initial acceleration and propels the toy after it is released, by forcing the perpetual motor that revolves the kinetic energy.
As the flywheel, unlike the spring of a pullback motor, is continuously rotating and not held, the motor may be "pumped up" by pushing the car repeatedly forward. The cars also typically work in forward and reverse. Some used a zip cord pulled from the vehicle body to accelerate the flywheel directly. Another system was the Turbo Tower of Power (TTP) in which air expelled from a hand-operated pump pushed turbine blades on the flywheel's rim.
These toys were especially popular in the 1960s to 1980s though they continue to be available today. |
https://en.wikipedia.org/wiki/Attributional%20calculus | Attributional calculus is a logic and representation system defined by Ryszard S. Michalski. It combines elements of predicate logic, propositional calculus, and multi-valued logic. Attributional calculus provides a formal language for natural induction, which is an inductive learning process whose outcomes are in human-readable forms. |
https://en.wikipedia.org/wiki/Per%20cent%20mille | A per cent mille or pcm is one one-thousandth of a percent. It can be thought of as a "milli-percent". It is commonly used in epidemiology, and in nuclear reactor engineering as a unit of reactivity.
Epidemiology
Statistics of crime rates, mortality and disease prevalence in a population are often given in
Nuclear Reactivity
In nuclear reactor engineering, a per cent mille is equal to one-thousandth of a percent of the reactivity, denoted by Greek lowercase letter rho. Reactivity is a dimensionless unit representing a departure from criticality, calculated by:
where keff denotes the effective multiplication factor for the reaction. Therefore, one pcm is equal to:
This unit is commonly used in the operation of light-water reactor sites because reactivity values tend to be small, so measuring in pcm allows reactivity to be expressed using whole numbers.
Related units
Percentage point difference of 1 part in 100
Percentage (%) 1 part in 100
Per mille (‰) 1 part in 1,000
Basis point (bp) difference of 1 part in 10,000
Permyriad (‱) 1 part in 10,000
Parts-per notation including parts-per million, parts-per billion etc
See also
InHour (another unit of reactivity)
Dollar (reactivity)
Parts-per notation
Per-unit system
Percent point function
Notes
Units of measurement |
https://en.wikipedia.org/wiki/Recording%20at%20the%20edge | Recording at the edge is the recording of video and storing it either in camera or to a directly connected storage device instead of transporting it across the network to a centralized recording facility such as a digital video recorder.
Recording at the edge is a distributed or de-centralized approach to storage—the video is spread across a number of edge-storage devices as opposed to centralized on one.
Network bandwidth-friendly
Recording at the edge is a network bandwidth-friendly approach because it does not use any network bandwidth in order to record video—instead the network is only used to play back audio or video from the edge at some review station. When the network includes a WAN or, worse still, the Internet, by definition Recording at the Edge becomes the only way to ensure high quality recordings.
Reliability
Recording at the edge is more reliable because recording is independent of the network's health and degree of congestion. Even if the network grinds to a halt recording continues unaffected. However, you have to be very cautious while only recording at the edge. SD cards in these cameras are continuously filled with data, written over, and over-worked -They burn out very quickly, compared to a normal Hard Drive.
Simple Deployment
Recording at the Edge can be achieved by deploying PC-based DVRs at the edge, however with vulnerable operating systems these are more challenging to maintain than a simple purpose built embedded appliance. Embedded DVRs are better suited to this role, although conventional DVRs traditionally focus on recording, searching and playback than scalable live streaming over the network, which is a primary function of an IP-based CCTV architecture.
Surveillance System Fault Tolerance
One of the concerns about IP-based CCTV is the dependence on the network. In a network outage in a centralized storage model both live video and the ability to record is lost. With Recording at the edge this is less of an issue because networ |
https://en.wikipedia.org/wiki/Focal%20cloud | A focal cloud is the collection of focal points of an imperfect lens or parabolic reflector whether optical, electrostatic or electromagnetic. This includes parabolic antennas and lens-type reflective antennas of all kinds. The effect is analogous to the circle of confusion in photography.
In a perfect lens or parabolic reflector, rays parallel to the device's axis striking the lens or reflector all pass through a single point, the focal point. In an imperfectly constructed lens or reflector, rays passing through different parts of the element do not converge to a single point but have different focal points. The set of these focal points forms a region called the focal cloud. The diameter of the focal cloud determines the maximum resolution of the optical system. Lens-reflector artifacts, geometry and other imperfections determine the actual diameter of the focal cloud.
Satellite dish effects
A focal cloud can render a satellite dish less efficient, resulting in a reduced gain. The imperfections of the antenna lead to two problems: On the one hand the more the radio waves are directed away from the phase center of the feed horn, the more attenuated they get coupled into the waveguide. On the other hand the more the distance between antenna surface and feed horn varies, the bigger the phase shift gets - the closer the phase shift gets to 180 degrees (half a wave length) between different parts of the antenna surface, the more the signal gets extinguished (attenuated). Feed horn design might be able mitigate some of these losses.
For countering the effect, there are several techniques, either in construction of the reflectors or lenses, or in the way signal beams are concentrated.
Satellite-based dish antennas may be deformed intentionally to selectively distribute radiated power over a desired "footprint", in order to increase received power in the desired reception area (e.g. one selected nation), and reduce power outside of it (e.g. a neighboring nation) |
https://en.wikipedia.org/wiki/Redalyc | The Redalyc project (Red de Revistas Científicas de América Latina y El Caribe, España y Portugal) is a bibliographic database and a digital library of Open Access journals, supported by the Universidad Autónoma del Estado de México with the help of numerous other higher education institutions and information systems.
The project started in October 2002 with the general aim of building a scientific information system made up by the leading journals of all the knowledge areas edited in and about Latin America. Since its creation, its goal is: to give visibility to the scientific production generated in Ibero-America, that is underestimated worldwide due to various factors like low investment in science and technology, low participation of Latin American scientists in some of the main currents of science, as measured by percentage of articles by Latin American authors in established electronic databases e.g., MEDLINE, and the low impact of that production. Participation, measured by percentage of articles by authors of Latin America in such databases was very low in the dominant repositories e.g., 2.7% in the Science Citation Index (SCI).
As of 2015, Redalyc is an information system that also evaluates the scientific and editorial quality of knowledge in Ibero-America. A research group generates bibliometric indicators about the impact of the journals, authors and countries included in the journal electronic library. Redalyc has been consolidated as an important repository of knowledge with over 1,000 journals and more than 425,000 full-text articles.
Scientific Journal Electronic Library
Organized in two main areas (social and natural sciences) and many specialised sub-sections, Redalyc gathers journals published in 15 countries, with over 550 journals and 16,000 articles available in PDF format, along with abstracts in Spanish and English languages, reference information, and other metadata. Similar to parallel projects such as Latindex, Redalyc fully embraces |
https://en.wikipedia.org/wiki/Symplectization | In mathematics, the symplectization of a contact manifold is a symplectic manifold which naturally corresponds to it.
Definition
Let be a contact manifold, and let . Consider the set
of all nonzero 1-forms at , which have the contact plane as their kernel. The union
is a symplectic submanifold of the cotangent bundle of , and thus possesses a natural symplectic structure.
The projection supplies the symplectization with the structure of a principal bundle over with structure group .
The coorientable case
When the contact structure is cooriented by means of a contact form , there is another version of symplectization, in which only forms giving the same coorientation to as are considered:
Note that is coorientable if and only if the bundle is trivial. Any section of this bundle is a coorienting form for the contact structure.
Differential topology
Structures on manifolds
Symplectic geometry |
https://en.wikipedia.org/wiki/Location%20routing%20number | A location routing number (LRN) is an identification for a telephone switch for the purpose of routing telephone calls through the public switched telephone network (PSTN) in the United States. This identification has the format of a telephone number, in accordance with the North American Numbering Plan (NANP). The association of a location routing number with a telephone number is required for local number portability.
Function
In the US, the location routing number is a ten-digit number following the specifications of the North American Numbering Plan. The LRN is stored in a database called a Service Control Point (SCP) that identifies a switching port for a local telephone exchange. Using LRN, when a telephone number has been dialed, the local telephone exchange queries or "dips" a routing database, usually the SCP, for the LRN associated with the subscriber number. The LRN removes the need for the public telephone number to identify the local exchange carrier. If a subscriber changes to another telephone service provider, the current telephone number can be retained, and only the LRN needs to be changed.
In addition to supporting service provider telephone number portability, an LRN also supports the possibility of two other types of number portability: service portability (for example, ordinary service to ISDN) and geographic portability.
History
In 1996, the US Congress mandated a change in local telephone service that allows any carrier to enter a local market. The new regulation provided for local number portability (LNP), which permitted the servicing of telephone numbers from other wire centers than that the given by the NPA-NXX prefixes of each number. In practice, a subscriber can keep a telephone number when moving to another exchange area by a process called porting a telephone number. Every ported telephone number has an LRN assigned.
Virginia-based NeuStar has been contracted with developing and maintaining the Number Portability Administration Ce |
https://en.wikipedia.org/wiki/Wolf%20summation | The Wolf summation is a method for computing the electrostatic interactions of systems (e.g. crystals). This method is generally more computationally efficient than the Ewald summation. It was proposed by Dieter Wolf. |
https://en.wikipedia.org/wiki/Additron%20tube | The Additron was an electron tube designed by Dr. Josef Kates, circa 1950, to replace the several individual electron tubes and support components required to perform the function of a single bit digital full adder. Dr. Kates developed the Additron with the intention of increasing the likelihood of success and reliability while reducing the size, power consumption and complexity of the University of Toronto Electronic Computer, (UTEC)
The Additron neither went into production at the Canadian Rogers Vacuum Tube Company, where the prototypes were built, nor was it used in the UTEC machine. It did make a widely publicized appearance at the 1950 Canadian National Exhibition operating an electronic game of Tic-Tac-Toe, dubbed Bertie the Brain, to show the marvels of electronic computing.
The tube was registered with the Radio Television Manufacturing Association on 20 March 1951 as type 6047.
Patents |
https://en.wikipedia.org/wiki/Confronted%20animals | Confronted animals, or confronted-animal as an adjective, where two animals face each other in a symmetrical pose, is an ancient bilateral motif in art and artifacts studied in archaeology and art history. The "anti-confronted animals" is the opposing motif, with the animals back to back.
Bilateral symmetry is a dominant aspect of our world and strong representation of it with matching figures often creates a balance that is appealing in artwork.
In ancient art, confronted-animal motifs often involve the Master of Animals, a central human figure between two confronted animals, often grasping them, and are probably part of a unified socio-cultural motif. A related motif in ancient art is known as the Mistress of Animals.
It is thought that the iconography sometimes has ritual and religious associations; for example, the Lion Gate of Mycenae has a column between the protective, surmounted and confronted lionesses standing with two feet on the ground and two on the same base on which the column rests. The column is thought to represent a goddess, abstracted to avoid tabooed direct representation. Alternatively, the column has also been interpreted as symbolizing the entrance to the palace. The lions are thus guarding the entrance to the palace and the walled fortification simultaneously. The motif called the Tree of Life, where two confronted animals graze on a shrub or tree, is also very ancient.
Gebel el-Arak Knife
The reverse side of the Gebel el-Arak Knife handle, displayed at the top of the article, shows two powerful confronted lions, separated by a figure who is grasping them. Of note, the knob on the reverse shows the top end of the hole for a cord or rope, on the suspension lug; this knife may have been worn around the neck ceremonially.
Examples from archaeology
Cylinder seals
Examples of confronted animals exist on Cylinder seals from Mesopotamia. Deities, or heroes grasping lionesses, cattle, griffins, or other, imaginary creatures are sometimes f |
https://en.wikipedia.org/wiki/Brain%20%26%20Behavior%20Research%20Foundation | The Brain & Behavior Research Foundation (BBRF) is a nonprofit 501(c)(3) organization that funds mental health research. It was originally called the National Alliance for Research on Schizophrenia & Depression or the acronym for that, NARSAD. It received its nonprofit ruling in 1981.
Mission and work
BBRF states that it is "committed to alleviating the suffering caused by mental illness by awarding grants that will lead to advances and breakthroughs in scientific research." The Foundation focuses its research grants in the following areas: addiction, ADHD, anxiety, autism, bipolar disorder, borderline personality disorder, depression, eating disorders, OCD, PTSD, schizophrenia, as well as research in suicide prevention. Grant applications (943 in 2019) are assessed by the BBRF Scientific Council. This group of 187 prominent mental health researchers, led by Herbert Pardes, M.D., reviews each grant application and selects those deemed most likely to lead to breakthroughs. A total of 150 Young Investigator grants were made in 2022.
Research funding awarded
From its inception in 1987 through calendar year 2022, BBRF has awarded over $440 million to fund more than 6,400 grants to more than 5,300 scientists around the world. The Foundation states that 100% of every dollar donated for research goes to research. BBRF is able to do this thanks to the support of two family foundations which fully cover its operating expenses.
Awards |
https://en.wikipedia.org/wiki/International%20Phytogeographic%20Excursion | The International Phytogeographic Excursions was a series of international meetings in plant geography that significantly contributed to exchange of scientific ideas across national and linguistic barriers and also to the rise of Anglo-American plant ecology. The initiative was taken by the British botanist Arthur Tansley at the International Geographic Congress in Geneva in 1908. Tansley and another early key figure, Henry C. Cowles, were both much-inspired by the new 'ecological plant geography' introduced by Eugenius Warming and its quest for answering why-questions about plant distribution, as opposed to the traditional, merely descriptive 'floristic plant geography'.
The First International Phytogeographic Excursion was held in the British Isles in 1911. It was organized by Arthur Tansley and went through parts of England, Scotland and Ireland.
The participants were:
Eduard Rübel, Switzerland
Carl Schroeter, Switzerland
Oscar Drude, Germany
Paul Graebner, Germany
C.A.M. Lindman, Sweden
G. Claridge Druce, England
Jean Massart, Belgium
C.H. Ostenfeld, Denmark
Frederic Clements, U.S.A.
Henry C. Cowles, U.S.A., who gave a brief report in Science in 1913.
The Second International Phytogeographic Excursion was a travel across North America from July to September 1913. It was hosted by a number of American ecologists led by Henry C. Cowles. The participants were:
Henry C. Cowles, U.S.A.
Frederic Clements, U.S.A.
Edith S. Clements, U.S.A.
Alfred Dachnowsky, U.S.A.
George Fuller, U.S.A.
George E. Nichols, U.S.A.
Willis Linn Jepson, U.S.A.
Heinrich Brockmann-Jerosch, Switzerland
Marie Charlotte Brockmann-Jerosch, Switzerland
Ove Paulsen, Denmark
Carl Skottsberg, Sweden
Eduard Rübel, Switzerland
Karl von Tubeuf, Germany
Carl Schroeter, Switzerland
Theodoor J. Stomps, Netherlands
Arthur Tansley, England
Adolf Engler, Germany
Cecil Crampton, Scotland.
The Third International Phytogeographic Excursion was proposed in 1915, but postponed due to the First World War. It |
https://en.wikipedia.org/wiki/Inversion%20%28music%29 | In music theory, an inversion is a type of change to intervals, chords, voices (in counterpoint), and melodies. In each of these cases, "inversion" has a distinct but related meaning. The concept of inversion also plays an important role in musical set theory.
Intervals
An interval is inverted by raising or lowering either of the notes by one or more octaves so that the positions of the notes reverse (i.e. the higher note becomes the lower note and vice versa). For example, the inversion of an interval consisting of a C with an E above it (the third measure below) is an E with a C above it – to work this out, the C may be moved up, the E may be lowered, or both may be moved.
The tables to the right show the changes in interval quality and interval number under inversion. Thus, perfect intervals remain perfect, major intervals become minor and vice versa, and augmented intervals become diminished and vice versa. (Doubly diminished intervals become doubly augmented intervals, and vice versa.).
Traditional interval numbers add up to nine: seconds become sevenths and vice versa, thirds become sixths and vice versa, and so on. Thus, a perfect fourth becomes a perfect fifth, an augmented fourth becomes a diminished fifth, and a simple interval (that is, one that is narrower than an octave) and its inversion, when added together, equal an octave. See also complement (music).
Chords
A chord's inversion describes the relationship of its lowest notes to the other notes in the chord. For instance, a C major triad contains the tones C, E and G; its inversion is determined by which of these tones is the lowest note (or bass note) in the chord.
The term inversion often categorically refers to the different possibilities, though it may also be restricted to only those chords where the lowest note is not also the root of the chord. Texts that follow this restriction may use the term position instead, to refer to all of the possibilities as a category.
Root position and inver |
https://en.wikipedia.org/wiki/Molecular%20epidemiology | Molecular epidemiology is a branch of epidemiology and medical science that focuses on the contribution of potential genetic and environmental risk factors, identified at the molecular level, to the etiology, distribution and prevention of disease within families and across populations. This field has emerged from the integration of molecular biology into traditional epidemiological research. Molecular epidemiology improves our understanding of the pathogenesis of disease by identifying specific pathways, molecules and genes that influence the risk of developing disease. More broadly, it seeks to establish understanding of how the interactions between genetic traits and environmental exposures result in disease.
History
The term "molecular epidemiology" was first coined by Edwin D. Kilbourne in a 1973 article entitled "The molecular epidemiology of influenza". The term became more formalized with the formulation of the first book on molecular epidemiology titled Molecular Epidemiology: Principles and Practice by Paul A. Schulte and Frederica Perera. At the heart of this book is the impact of advances in molecular research that have given rise to and enabled the measurement and exploitation of the biomarker as a vital tool to link traditional molecular and epidemiological research strategies to understand the underlying mechanisms of disease in populations.
Modern use
While most molecular epidemiology studies are using conventional disease designation system for an outcome (with the use of exposures at the molecular level), compelling evidence indicates that disease evolution represents inherently heterogeneous process differing from person to person. Conceptually, each individual has a unique disease process different from any other individual ("the unique disease principle"), considering uniqueness of the exposome and its unique influence on molecular pathologic process in each individual. Studies to examine the relationship between an exposure and molecular pa |
https://en.wikipedia.org/wiki/Heteromer | A heteromer is something that consists of different parts; the antonym of homomeric. Examples are:
Biology
Spinal neurons that pass over to the opposite side of the spinal cord.
A protein complex that contains two or more different polypeptides.
Pharmacology
Ligand-gated ion channels such as the nicotinic acetylcholine receptor and GABAA receptor are composed of five subunits arranged around a central pore that opens to allow ions to pass through. There are many different subunits available that can come together in a wide variety of combinations to form different subtypes of the ion channel. Sometimes the channel can be made from only one type of subunit, such as the α7 nicotinic receptor, which is made up from five α7 subunits, and so is a homomer rather than a heteromer, but more commonly several different types of subunit will come together to form a heteromeric complex (e.g., the α4β2 nicotinic receptor, which is made up from two α4 subunits and three β2 subunits). Because the different ion channel subtypes are expressed to different extents in different tissues, this allows selective modulation of ion transport and means that a single neurotransmitter can produce varying effects depending on where in the body it is released.
G protein-coupled receptors are composed of seven membrane-spanning alpha-helical segments that are usually linked together into a single folded chain to form the receptor complex. However, research has demonstrated that a number of GPCRs are also capable of forming heteromers from a combination of two or more individual GPCR subunits under some circumstances, especially where several different GPCRs are densely expressed in the same neuron. Such heteromers may be between receptors from the same family (e.g., adenosine A1/A2A heteromers and dopamine D1/D2 and D1/D3 heteromers) or between entirely unrelated receptors such as CB1/A2A, glutamate mGluR5 / adenosine A2A heteromers, cannabinoid CB1 / dopamine D2 heteromers, and even CB1/A2 |
https://en.wikipedia.org/wiki/Negro%20of%20Banyoles | The Negro of Banyoles (, or ) was a controversial piece of taxidermy of a San individual, which used to be a major attraction in the Darder Museum of Banyoles (Catalonia, Spain). In 2000, the remains of the man were sent to Botswana for burial.
History
In 1830, the Verreaux brothers stuffed the corpse of a member of the San. Analysis of the teeth shows that man was approximately 27 years old, having features typical of the African bushman. In 1916, it was acquired by the Darder Museum of Banyoles. The body remained in the museum without controversy until 29 October 1991. At this date Alphonse Arcelin, a doctor of Haitian origin who lived in Cambrils where he was a PSC councillor, wrote a letter to the mayor of Banyoles, Joan Solana, asking him to stop displaying the San's remains. This request attracted the attention of the press, which widely published the story.
The first step towards the return of the "negre" to Botswana was made in 1991, when the then-secretary of UNESCO, Federico Mayor Zaragoza, held the meeting with Joan Solana. Later, when Kofi Annan became Secretary-General of the United Nations, he expressed interest in the issue and spoke with Solana.
By that time, the "negre" had become so notorious that it was quite usual to hear references made to the displayed San in diplomatic communications. Some African governments showed their support for Arcelín, who had sent several letters to the press and various heads of government. The issue worried many international museum associations because it made them fear that human remains kept in museums might have to be returned to their place of origin.
In 1997, the issue was repeatedly discussed by both the UN and the Organisation of African Unity. Later, in March of that year, the body was removed from the Darder Museum. It was described in El Mundo as a relic of colonialism. Many people in Banyoles and the surrounding area were unhappy with the removal because the San was seen as "a member of the community |
https://en.wikipedia.org/wiki/Solemyidae | Solemyidae is a family of saltwater clams, marine protobranch bivalve mollusks in the order Solemyida.
Biology
Solemyids are remarkable in that their digestive tract is either extremely small or non-existent, and their feeding appendages are too short to reach outside the shell.
It has been shown that these clams host sulphur-oxidizing bacteria intracellularly within their gill filaments. As chemoautotrophs, these bacterial symbionts synthesize organic matter from CO2 and are the primary source of nutrition for the whole organism. In turn, the animal host provides its symbionts a habitat in which they have access to the substrates of chemoautotrophy (O2, CO2, and reduced inorganic compounds such as H2S). Together, these partners create "animals" with novel metabolic capabilities.
Genera and species
The family Solemyidae includes two genera and the following species:
Acharax Dall, 1908
Acharax alinae Métivier & Cosel, 1993
Acharax bartschii (Dall, 1908)
Acharax caribbaea (Vokes, 1970)
Acharax clarificata Dell, 1995
Acharax gadirae Oliver, Rodrigues & Cunha, 2011
Acharax grandis (Verrill & Bush, 1898)
Acharax japonica (Dunker, 1882)
Acharax johnsoni (Dall, 1891)
Acharax patagonica (E. A. Smith, 1885)
Acharax prashadi (Vokes, 1955)
Solemya Lamarck, 1818
Solemya africana Martens, 1880
Solemya atacama (Kuznetzov & Schileyko, 1984)
Solemya australis Lamarck, 1818
Solemya borealis Totten, 1834
Solemya elarraichensis Oliver, Rodrigues & Cunha, 2011
Solemya flava Sato, Sasaki & Watanabe, 2013
Solemya moretonensis Taylor, Glover & Williams, 2008
Solemya notialis Simone, 2009
Solemya occidentalis Deshayes, 1857
Solemya panamensis Dall, 1908
Solemya parkinsonii E. A. Smith, 1874
Solemya pervernicosa (Kuroda, 1948)
Solemya pusilla Gould, 1861
Solemya tagiri Okutani, Hashimoto & Miura, 2004
Solemya terraereginae Iredale, 1929
Solemya togata (Poli, 1791)
Solemya valvulus Carpenter, 1864
Solemya velesiana Iredale, 1931 |
https://en.wikipedia.org/wiki/Bramble%E2%80%93Hilbert%20lemma | In mathematics, particularly numerical analysis, the Bramble–Hilbert lemma, named after James H. Bramble and Stephen Hilbert, bounds the error of an approximation of a function by a polynomial of order at most in terms of derivatives of of order . Both the error of the approximation and the derivatives of are measured by norms on a bounded domain in . This is similar to classical numerical analysis, where, for example, the error of linear interpolation can be bounded using the second derivative of . However, the Bramble–Hilbert lemma applies in any number of dimensions, not just one dimension, and the approximation error and the derivatives of are measured by more general norms involving averages, not just the maximum norm.
Additional assumptions on the domain are needed for the Bramble–Hilbert lemma to hold. Essentially, the boundary of the domain must be "reasonable". For example, domains that have a spike or a slit with zero angle at the tip are excluded. Lipschitz domains are reasonable enough, which includes convex domains and domains with continuously differentiable boundary.
The main use of the Bramble–Hilbert lemma is to prove bounds on the error of interpolation of function by an operator that preserves polynomials of order up to , in terms of the derivatives of of order . This is an essential step in error estimates for the finite element method. The Bramble–Hilbert lemma is applied there on the domain consisting of one element (or, in some superconvergence results, a small number of elements).
The one-dimensional case
Before stating the lemma in full generality, it is useful to look at some simple special cases. In one dimension and for a function that has derivatives on interval , the lemma reduces to
where is the space of all polynomials of degree at most and indicates
the th derivative of a function .
In the case when , , , and is twice differentiable, this means that there exists a polynomial of degree one such that for all ,
T |
https://en.wikipedia.org/wiki/NEC%20SX-9 | The SX-9 is a NEC SX supercomputer built by NEC Corporation. The SX-9 Series implements an SMP system in a compact node module and uses an enhanced version of the single chip vector processor that was introduced with the SX-6. The NEC SX-9 processors run at 3.2 GHz, with eight-way replicated vector pipes, each having two multiply units and two addition units; this results in a peak vector performance of 102.4 gigaFLOPS. For non-vectorized code, there is a scalar processor that runs at half the speed of the vector unit, i.e. 1.6 GHz. Up to 16 CPUs and 1 terabyte of memory may be used in a single node. Each node is packaged in an air-cooled cabinet, similar in size to a standard 42U computer rack. The SX-9 series ranges from the single-node SX-9/B system with 4 CPUs to the maximum expansion stage with 512 nodes, 8,192 CPUs, and 970 TFLOPS peak performance. There is up to 4 TB/s shared memory bandwidth per node and 2×128 GB/s node interconnect bandwidth. The operating system is NEC's SUPER-UX, a Unix-like OS.
The SX-9 had the world's fastest vector CPU core. A fully equipped system with 512 nodes would have been the world's fastest supercomputer at the time of release in the first quarter of 2008, with a performance of 819 TFLOPS. The SX-9 was discontinued in 2015.
The German national meteorological service (DWD) operated two independent SX-9 clusters, with 976 processors, 31,232 GB of RAM and 98 TFLOPS performance in total.
NEC Published Product Highlights
1.6 TFLOPS max. peak performance per node
350 million transistors per CPU, 1.0 V, 8,960 pins (1,791 signal pins)
Up to 16 CPUs per node, manufactured in 65 nm CMOS cu technology, 11 copper layers
Up to 64 GB of memory per CPU, 1 TB in a single node
Up to 4 TB/s bandwidth per node, 256 GB/s per CPU
IXS Super-Switch between nodes, up to 512 nodes supported, 256 GB/s per node (128 GB/s for each direction)
50% less power consumption compared to the NEC SX-8R
See also
SUPER-UX
SX architecture
External links
spec |
https://en.wikipedia.org/wiki/Irish%20Nutrition%20and%20Dietetic%20Institute | The Irish Nutrition & Dietetic Institute (INDI) is the national organisation in Ireland for nutrition and dietetic professionals. The INDI's mission is to encourage, foster and maintain the highest possible standards in the science and practice of human nutrition and dietetics, to positively influence the nutrition status and health of the individual and the population in general. This incorporates clinical nutrition, community nutrition, business and industry, research, education and private practice.
External links
Official web site
Food science institutes
Nutrition organizations
Government agencies of the Republic of Ireland |
https://en.wikipedia.org/wiki/Secondary%20cell%20wall | The secondary cell wall is a structure found in many plant cells, located between the primary cell wall and the plasma membrane. The cell starts producing the secondary cell wall after the primary cell wall is complete and the cell has stopped expanding.
Secondary cell walls provide additional protection to cells and rigidity and strength to the larger plant. These walls are constructed of layered sheaths of cellulose microfibrils, wherein the fibers are in parallel within each layer. The inclusion of lignin makes the secondary cell wall less flexible and less permeable to water than the primary cell wall. In addition to making the walls more resistant to degradation, the hydrophobic nature of lignin within these tissues is essential for containing water within the vascular tissues that carry it throughout the plant.
The secondary cell wall consists primarily of cellulose, along with other polysaccharides, lignin, and glycoprotein. It sometimes consists of three distinct layers - S1, S2 and S3 - where the direction of the cellulose microfibrils differs between the layers. The direction of the microfibrils is called microfibril angle (MFA). In the secondary cell wall of fibres of trees a low microfibril angle is found in the S2-layer, while S1 and S3-layers show a higher MFA . However, the MFA can also change depending on the loads on the tissue. It has been shown that in reaction wood the MFA in S2-layer can vary. Tension wood has a low MFA, meaning that the microfibril is oriented parallel to the axis of the fibre. In compression wood the MFA is high and reaches up to 45°. These variations influence the mechanical properties of the cell wall.
The secondary cell wall has different ratios of constituents compared to the primary wall. An example of this is that secondary wall in wood contains polysaccharides called xylan, whereas the primary wall contains the polysaccharide xyloglucan.
The cells fraction in secondary walls is also higher. Pectins may also be a |
https://en.wikipedia.org/wiki/Multimedia%20extension%20connector | A multimedia extension connector (MXC) is a method of connecting video cameras and other video inputs to video capture cards and the like. MXC is based on the 8-pin Mini-DIN connector. It is used by Winnov's range of Videum capture cards.
External links
Videum 1000 Plus Family Quick Start Guide listing on page 9 the pinout of the MXC connector.
Computer connectors
Electrical connectors
Audiovisual connectors
RF connectors
Deutsches Institut für Normung |
https://en.wikipedia.org/wiki/Ionic%20conductivity%20%28solid%20state%29 | Ionic conductivity (denoted by ) is a measure of a substance's tendency towards ionic conduction. Ionic conduction is the movement of ions. The phenomenon is observed in solids and solutions. Ionic conduction is one mechanism of current.
In crystalline solids
In most solids, ions rigidly occupy fixed positions, strongly embraced by neighboring atoms or ions. In some solids, selected ions are highly mobile allowing ionic conduction. The mobility increases with temperature. Materials exhibiting this property are used in batteries. A well-known ion conductive solid is β''-alumina ("BASE"), a form of aluminium oxide that has channels through which sodium cations can hop. When this ceramic is complexed with a mobile ion, such as Na+, it behaves as so-called fast ion conductor. BASE is used as a membrane in several types of molten salt electrochemical cell.
In glasses
Ion conduction in disordered solids like glasses, polymers, nanocomposites, defective crystals and other disordered solids plays an important role in technology.
History
Ionic conduction in solids has been a subject of interest since the beginning of the 19th century. Michael Faraday established in 1839 that the laws of electrolysis are also obeyed in ionic solids like lead(II) fluoride () and silver sulfide (). In 1921, solid silver iodide () was found to have had extraordinary high ionic conductivity at temperatures above 147 °C, AgI changes into a phase that has an ionic conductivity of ~ 1 –1 cm−1. This high temperature phase of AgI is an example of a superionic conductor. The disordered structure of this solid allows the Ag+ ions to move easily. The present record holder for ionic conductivity is the related material Ag2HgI4. β''-alumina was developed at the Ford Motor Company in the search for a storage device for electric vehicles while developing the sodium–sulfur battery.
See also
Lattice energy
Fast ion conductor
NASICON |
https://en.wikipedia.org/wiki/Dynamic%20voltage%20scaling | In computer architecture, dynamic voltage scaling is a power management technique in which the voltage used in a component is increased or decreased, depending upon circumstances. Dynamic voltage scaling to increase voltage is known as overvolting; dynamic voltage scaling to decrease voltage is known as undervolting. Undervolting is done in order to conserve power, particularly in laptops and other mobile devices, where energy comes from a battery and thus is limited, or in rare cases, to increase reliability. Overvolting is done in order to support higher frequencies for performance.
The term "overvolting" is also used to refer to increasing static operating voltage of computer components to allow operation at higher speed (overclocking).
Background
MOSFET-based digital circuits operate using voltages at circuit nodes to represent logical state. The voltage at these nodes switches between a high voltage and a low voltage during normal operation—when the inputs to a logic gate transition, the transistors making up that gate may toggle the gate's output.
Toggling a MOSFET's state requires changing its gate voltage from below the transistor's threshold voltage to above it (or from above it to below it). However, changing the gate's voltage requires charging or discharging the capacitance at its node. This capacitance is the sum of capacitances from various sources: primarily transistor gate capacitance, diffusion capacitance, and wires (coupling capacitance).
Higher supply voltages result in faster slew rate (rate of change of voltage per unit of time) when charging and discharging, which allows for quicker transitioning through the MOSFET's threshold voltage. Additionally, the more the gate voltage exceeds the threshold voltage, the lower the resistance of the transistor's conducting channel. This results in a lower RC time constant for quicker charging and discharging of the capacitance of the subsequent logic stage. Quicker transitioning afforded by higher sup |
https://en.wikipedia.org/wiki/Hut%207 | Hut 7 was a wartime section of the Government Code and Cypher School (GC&CS) at Bletchley Park tasked with the solution of Japanese naval codes such as JN4, JN11, JN40, and JN-25. The hut was headed by Hugh Foss who reported to Frank Birch, the head of Bletchley's Naval section.
Hut 7 supplied cryptanalysts and linguists to Bletchley's front line station the Far East Combined Bureau (FECB) at Hong Kong, then Singapore, then Anderson Station (Colombo, Ceylon, now Sri Lanka), then Allidina School in Kilindini, Kenya before moving back to Colombo.
Bletchley co-operated with the US Navy Code and Signals Section known as OP-20-G in Washington D.C., and with FRUMEL in Melbourne (although the reciprocal cooperation from Fabian at FRUMEL was limited and reluctant);
see Central Bureau and FRUMEL.
Pre-War origins
Among the first challenges was overcoming the special kana and romaji Morse code system used by the Japanese. GC&CS began breaking Japanese diplomatic traffic in the early 1920s. GC&CS subsequently attacked the Naval Reporting Code, and General Operations Code.
The section was headed by William "Nobby" Clarke with Harry Shaw and Ernest Hobart-Hampden. These were later joined by Eric Nave, seconded from the Royal Australian Navy, John Tiltman, and Hugh Foss.
GC&CS operated the Far East Combined Bureau, the codebreaking and intercept station in Hong Kong prewar, which during the war moved to Singapore, Colombo and Kilindini.
WWII expansion
Potential linguists and cryptographers were recruited from the University of Cambridge and the University of Oxford by referral through people like the Revd Martin Charlesworth, A.D. Lindsay, Dr C. P. Snow, and Theodore Chaundy. Candidates were interviewed and approved by a board that included Colonel Tiltman. Successful candidates received a final interview at Bletchley by a senior representative of their section.
At the outset of World War II, Britain had few Japanese linguists and conventional wisdom held that it would |
https://en.wikipedia.org/wiki/Docosanoid | In biochemistry, docosanoids are signaling molecules made by the metabolism of twenty-two-carbon fatty acids (EFAs), especially the omega-3 fatty acid, Docosahexaenoic acid (DHA) (i.e. 4Z,7Z,10Z,13Z,16Z,19Z-docosahexaenoic acid) by lipoxygenase, cyclooxygenase, and cytochrome P450 enzymes. Other docosanoids are metabolites of n-3 docosapentaenoic acid (i.e. 7Z,10Z,13Z,16Z,19Z-docosahexaenoic acid), n-6 DHA (i.e. 4Z,7Z,10Z,13Z,16Z-docosahexaenoic acid, and docosatetraenoic acid (i.e. 7Z,10Z,13Z,16Z-docosatetraenoic acid, DTA, or adrenic acid). Prominent docosanoid metabolites of DHA and n-3 DHA are members of the specialized proresolving mediator class of polyunsaturated fatty acid metabolites that possess potent anti-inflammation, tissue healing, and other activities (see specialized proresolving mediators).
Prominent docosanoids
Specialized proresolving mediator docosanoids
Potently bioactive agents of the specialized proresolving mediator class include:
DHA-derived Resolvins (Rv's) of the D series: RvD1, RvD2, RvD3, RvD4, RvD5, RvD6, AT-RvD1, AT-RvD2, AT-RvD3, AT-RvD4, AT-RvD5, and AT-RvD6 (see specialized proresolving mediators#DHA-derived Resolvins).
n-3 DPA-derived Rvs of the D series (RvD1n-3, RvD2n-3, and RvDD1n-3) and the T series (RvT1, TvT2, RvT3, and RvT4) (see specialized proresolving mediators#n-3 DPA-derived resolvins).
DHA-derived Neuroprotectins, also termed protectins: PD1, PDX, 17-epi PD1, and 10-epi-DHA1 (see specialized proresolving mediators#DHA-derived protectins/neuroprotectins).
n-3 DPA derived protectins: RD1n-3 and RvD1n-3 (see specialized proresolving mediators#n-3 DPA-derived resolvins)(see DPA-derived protectins/neuroprotectins.
DHA derived Maresins: MaR1, MaR2, 7-epi-Mar1, Mar-L1, and Mar-L2 (see specialized proresolving mediators#DHA-derived Maresins).
n-3 DPA-derived maresins: Mar1n-3, Mar2n-3, and Mar3n-3 (see specialized proresolving mediators#n-3 DPA-derived maresins).
These DHA metabolites possess anti-inflammation and tissue- |
https://en.wikipedia.org/wiki/ERB1 | Erb1 also known as the eukaryotic ribosome biogenesis protein 1 is a yeast protein required for maturation of the 25S and 5.8S ribosomal RNAs. It is a component of 66S pre-ribosomal particles and is homologous to the human protein BOP1. |
https://en.wikipedia.org/wiki/Lamina%20affixa | Lamina affixa is a layer of epithelium growing on the surface of the thalamus and forming the floor of the central part of lateral ventricle, on whose medial margin is attached the choroid plexus of the lateral ventricle; it covers the superior thalamostriate vein and the superior choroid vein. The torn edge of this plexus is called the tela choroidea.
On the surface of the terminal vein is a narrow white band, named the lamina affixa.
GDF-15/MIC-1 has been observed in lamina affixa cells. |
https://en.wikipedia.org/wiki/Poikilohydry | Poikilohydry is the lack of ability (structural or functional mechanism) to maintain and/or regulate water content to achieve homeostasis of cells and tissue connected with quick equilibration of cell/tissue water content to that of the environment. The term is derived from Ancient Greek ποικίλος (poikílos, “spotted or variegate”).
Tolerance to desiccation has been utilized in the Archaea, Bacteria, and Eukaryote kingdoms to take advantage of ecological niches. The tolerance to desiccation is often combined with other abiotic stress factors such as temperature extremes, malnutrition, vitamin imbalances, salinity content, and ultraviolet radiation. Many plants control desiccation tolerance through non-specialized structures such as vegetative tissues or specialized structures such as spores, seeds, and tubers. Desiccation tolerance is distributed among Bryophytes that have no cuticle or stomata, nine Pteridophyte families and ten Angiosperm families, vascular plants that do have a cuticle and stomata.
Selaginella lepidophylla is a vascular lycophyte native to the Chihuahuan Desert in New Mexico, Texas and Mexico. It occurs in north-facing rock crevices and in open habitats. The notable leaf curling attributed to S. lepidophylla, tested by Lebkeucher and Eickmeier in 1991, occurs to prevent photoinhibition in the microphylls in response to UV radiation and gradual leaf uncurling when rehydrated, protects the plant from the same photoinhibition until photon fluxes are fully processed.
Hymenophyllaceae are poikilohydrous ferns that grow in high-humidity, high-shade areas like the rainforests of Trinidad, Venezuela and New Zealand. These ferns face the unique challenges of having a high-resistance for gas exchange and limited water loss. This provided strong selective evolutionary pressure towards individuals that were the most energy efficient. Two major components of efficiency have been focused on by these ferns: reducing wasteful structures and increasing light |
https://en.wikipedia.org/wiki/Kepler%20triangle | A Kepler triangle is a special right triangle with edge lengths in geometric progression. The ratio of the progression is where is the golden ratio, and the progression can be written: or approximately . Squares on the edges of this triangle have areas in another geometric progression, . Alternative definitions of the same triangle characterize it in terms of the three Pythagorean means of two numbers, or via the inradius of isosceles triangles.
This triangle is named after Johannes Kepler, but can be found in earlier sources. Although some sources claim that ancient Egyptian pyramids had proportions based on a Kepler triangle, most scholars believe that the golden ratio was not known to Egyptian mathematics and architecture.
History
The Kepler triangle is named after the German mathematician and astronomer Johannes Kepler (1571–1630), who wrote about this shape in a 1597 letter. Two concepts that can be used to analyze this triangle, the Pythagorean theorem and the golden ratio, were both of interest to Kepler, as he wrote elsewhere:
However, Kepler was not the first to describe this triangle. Kepler himself credited it to "a music professor named Magirus". The same triangle appears earlier in a book of Arabic mathematics, the Liber mensurationum of Abû Bekr, known from a 12th-century translation by Gerard of Cremona into Latin, and in the of Fibonacci (published in 1220–1221), who defined it in a similar way to Kepler. A little earlier than Kepler, Pedro Nunes wrote about it in 1567, and it is "likely to have been widespread in late medieval and Renaissance manuscript traditions". It has also been independently rediscovered several times, later than Kepler.
According to some authors, a "golden pyramid" with a doubled Kepler triangle as its cross-section accurately describes the design of Egyptian pyramids such as the Great Pyramid of Giza; one source of this theory is a 19th-century misreading of Herodotus by pyramidologist John Taylor. Many other theories |
https://en.wikipedia.org/wiki/Freedup | freedup is a program to scan directories or file lists for duplicate files. The file lists may be provided to an input pipe or internally generated using find with provided options. There are more options to specify the search conditions more detailed. Other options influence the performed actions, i.e. whether to display only or to specify what kind of link under which circumstances. freedup first compares file sizes, then on equal sizes the MD5 signatures, and before taking actions a byte-by-byte check for verification is performed. An interactive mode allows to decide individually which files to link soft or hard or to delete.
The comparison by ignoring metadata tags and comments is a unique feature of freedup. Filesize, start and end of unique content is kept for later processing. Comparing sound files you may ignore the tags, e.g. whether one is tagged with an ID3v1-tag while another sound file with identical music is tagged with ID3v2. It also works, if you copied and retagged the copy to fit into another album. This works for JPEG files (Exif) and mp4-Movies as well. An auto-Mode is supported to instruct freedup to ignore all tags that are recognized. The author will extend this function on demand, if there is sufficient documentation how to strip the tags.
freedup is written in POSIX compliant C and is released under the GNU General Public License. Its complexity is O(n log n) for full file comparison. This is done for equally long files after sorting according to filesize using qsort().
See also
List of Unix commands
Duplicate code |
https://en.wikipedia.org/wiki/74181 | The 74181 is a 4-bit slice arithmetic logic unit (ALU), implemented as a 7400 series TTL integrated circuit. Introduced by Texas Instruments in February 1970, it was the first complete ALU on a single chip. It was used as the arithmetic/logic core in the CPUs of many historically significant minicomputers and other devices.
The 74181 represents an evolutionary step between the CPUs of the 1960s, which were constructed using discrete logic gates, and today's single-chip microprocessor CPUs. Although no longer used in commercial products, the 74181 is still referenced in computer organization textbooks and technical papers. It is also sometimes used in "hands-on" college courses to train future computer architects.
Specifications
The 74181 is a 7400 series medium-scale integration (MSI) TTL integrated circuit, containing the equivalent of 75 logic gates
and most commonly packaged as a 24-pin DIP. The 4-bit wide ALU can perform all the traditional add / subtract / decrement operations with or without carry, as well as AND / NAND, OR / NOR, XOR, and shift. Many variations of these basic functions are available, for a total of 16 arithmetic and 16 logical operations on two four-bit words. Multiply and divide functions are not provided but can be performed in multiple steps using the shift and add or subtract functions.
Shift is not an explicit function but can be derived from several available functions; e.g., selecting function "A plus A" with carry (M=0) will give an arithmetic left shift of the A input.
The 74181 performs these operations on two four-bit operands generating a four-bit result with carry in 22 nanoseconds (45 MHz). The 74S181 performs the same operations in 11 nanoseconds (90 MHz), while the 74F181 performs the operations in 7 nanoseconds (143 MHz) (typical).
Multiple 'slices' can be combined for arbitrarily large word sizes. For example, sixteen 74S181s and five 74S182 look ahead carry generators can be combined to perform the same operations on |
https://en.wikipedia.org/wiki/Tunicamycin | Tunicamycin is a mixture of homologous nucleoside antibiotics that inhibits the UDP-HexNAc: polyprenol-P HexNAc-1-P family of enzymes. In eukaryotes, this includes the enzyme GlcNAc phosphotransferase (GPT), which catalyzes the transfer of N-acetylglucosamine-1-phosphate from UDP-N-acetylglucosamine to dolichol phosphate in the first step of glycoprotein synthesis. Tunicamycin blocks N-linked glycosylation (N-glycans) and treatment of cultured human cells with tunicamycin causes cell cycle arrest in G1 phase. It is used as an experimental tool in biology, e.g. to induce unfolded protein response. Tunicamycin is produced by several bacteria, including Streptomyces clavuligerus and Streptomyces lysosuperificus.
Tunicamycin homologues have varying molecular weights owing to the variability in fatty acid side chain conjugates.
Biosynthesis
The biosynthesis of tunicamycins was studied in Streptomyces chartreusis and a proposed biosynthetic pathway was characterized. The bacteria utilize the enzymes in the tun gene cluster (TunA-N) to make tunicamycins.
TunA uses the starter unit uridine diphosphate-N-acetyl-glucosamine (UDP-GlcNAc) and catalyzes the dehydration of the 6’ hydroxyl group. First, a Tyr residue in TunA abstracts a proton from the 4’ hydroxyl group, forming a ketone at that position. A hydride is subsequently abstracted from the 4’ carbon by NAD+, forming NADH. The ketone is stabilized by hydrogen bonding from the Tyr residue, and a nearby Thr residue. A glutamate residue then abstracts a proton from the 5’ carbon, pushing the electrons up to form a double bond between the 5’ and 6’ carbon. A nearby cysteine donates a proton to the hydroxyl group as it leaves as water. NADH donates a hydride to the 4’ carbon, reforming a hydroxide in that position and forming UDP-6’-deoxy-5-6-ene-GlcNAc. TunF then catalyzes the epimerization of the intermediate to UDP-6’-deoxy-5-6-ene-GalNAc, changing the 4’ hydroxyl from the equatorial to axial position.
The ot |
https://en.wikipedia.org/wiki/Uridine%20diphosphate%20N-acetylglucosamine | Uridine diphosphate N-acetylglucosamine or UDP-GlcNAc is a nucleotide sugar and a coenzyme in metabolism. It is used by glycosyltransferases to transfer N-acetylglucosamine residues to substrates. D-Glucosamine is made naturally in the form of glucosamine-6-phosphate, and is the biochemical precursor of all nitrogen-containing sugars. To be specific, glucosamine-6-phosphate is synthesized from fructose 6-phosphate and glutamine as the first step of the hexosamine biosynthesis pathway. The end-product of this pathway is UDP-GlcNAc, which is then used for making glycosaminoglycans, proteoglycans, and glycolipids.
UDP-GlcNAc is extensively involved in intracellular signaling as a substrate for O-linked N-acetylglucosamine transferases (OGTs) to install the O-GlcNAc post-translational modification in a wide range of species. It is also involved in nuclear pore formation and nuclear signalling. OGTs and OG-ases play an important role in the structure of the cytoskeleton. In mammals, there is enrichment of OGT transcripts in the pancreas beta-cells, and UDP-GlcNAc is thought to be part of the glucose sensing mechanism. There is also evidence that it plays a part in insulin sensitivity in other cells. In plants, it is involved in the control of gibberellin production.
Clostridium novyi type A alpha-toxin is an O-linked N-actetylglucosamine transferase acting on Rho proteins and causing the collapse of the cytoskeleton. |
https://en.wikipedia.org/wiki/Japanese%20theorem%20for%20cyclic%20quadrilaterals | In geometry, the Japanese theorem states that the centers of the incircles of certain triangles inside a cyclic quadrilateral are vertices of a rectangle.
Triangulating an arbitrary cyclic quadrilateral by its diagonals yields four overlapping triangles (each diagonal creates two triangles). The centers of the incircles of those triangles form a rectangle.
Specifically, let be an arbitrary cyclic quadrilateral and let , , , be the incenters of the triangles , , , . Then the quadrilateral formed by , , , is a rectangle.
Note that this theorem is easily extended to prove the Japanese theorem for cyclic polygons. To prove the quadrilateral case, simply construct the parallelogram tangent to the corners of the constructed rectangle, with sides parallel to the diagonals of the quadrilateral. The construction shows that the parallelogram is a rhombus, which is equivalent to showing that the sums of the radii of the incircles tangent to each diagonal are equal.
The quadrilateral case immediately proves the general case by induction on the set of triangulating partitions of a general polygon.
See also
Carnot's theorem
Sangaku
Japanese mathematics |
https://en.wikipedia.org/wiki/Anguimorpha | The Anguimorpha is a suborder of squamates. The group was named by Fürbringer in 1900 to include all autarchoglossans closer to Varanus and Anguis than Scincus. These lizards, along with iguanians and snakes, constitute the proposed "venom clade" Toxicofera of all venomous reptiles.
Evolution
The oldest widely accepted member of Anguimorpha is Dorsetisaurus from the Late Jurassic of Europe and North America. In 2022, the genus Cryptovaranoides was described from the latest Triassic (Rhaetian) of England. Cryptovaranoides was recovered in the study as a crown-group Anguimorph, and therefore the oldest crown group-squamate, 35 million years older than any previously known crown-group squamate. However, a 2023 study found that Cryptovaranoides most likely represents an archosauromorph with no apparent squamate affinities.
Families
Anguidae
There are 9 genera found within the Anguidae family. They are characterized by being heavily armored with non-overlapping scales, and almost all having well-developed ventrolateral folds (excluding Anguis). Anguidae members can, however, be somewhat difficult to identify in their family, as members can be limbed or limbless, and can be both viviparous and oviparous. Anguidae members have pterygoid teeth. Many members of this family have tail autotomy.
Anniellidae
There is only one genus within the Anniellidae family, comprising 6 species of American legless lizards. They are characterized as having no limbs, and can be found in California and Baja California. They have wedge-shaped heads and a countersunk jaw that allows them to bury themselves in sand or loose soil, which they can drink water from if the soil has a water quantity greater than 7%. They give live birth and usually have two offspring.
Diploglossidae
There are three genera in the Diploglossidae family. They are characterized by having very long, automatized tails, small and well-developed limbs, and no ventrolateral fold. They do have bicuspid posterior teeth. |
https://en.wikipedia.org/wiki/Discoid%20lupus%20erythematosus | Discoid lupus erythematosus is the most common type of chronic cutaneous lupus (CCLE), an autoimmune skin condition on the lupus erythematosus spectrum of illnesses. It presents with red, painful, inflamed and coin-shaped patches of skin with a scaly and crusty appearance, most often on the scalp, cheeks, and ears. Hair loss may occur if the lesions are on the scalp. The lesions can then develop severe scarring, and the centre areas may appear lighter in color with a rim darker than the normal skin. These lesions can last for years without treatment.
Patients with systemic lupus erythematous develop discoid lupus lesions with some frequency. However, patients who present initially with discoid lupus infrequently develop systemic lupus. Discoid lupus can be divided into localized, generalized, and childhood discoid lupus.
The lesions are diagnosed by biopsy. Patients are first treated with sunscreen and topical steroids. If this does not work, an oral medication—most likely hydroxychloroquine or a related medication—can be tried.
Signs and symptoms
Morphology of lesions
Discoid lupus erythematosus (DLE) skin lesions first present as dull or purplish red, disc-shaped flat or raised and firm areas of skin. These lesions then develop increasing amounts of white, adherent scale. Finally, the lesions develop extensive scarring and/or atrophy, as well as pigment changes. They may also have overlying dried fluid, known as crust. On darker skin, the lesions often lose skin pigmentation in the center and develop increased, dark skin pigmentation around the rim. On lighter skin, the lesions often develop a gray color or have very little color change. More rarely, the lesions may be bright red and look like hives.
Location of lesions
The skin lesions are most often in sun-exposed areas localized above the neck, with favored sites being the scalp, bridge of the nose, upper cheeks, lower lip, and ear and hands 24% of patients also have lesions in the mouth (most often the |
https://en.wikipedia.org/wiki/DS80C390 | The DS80C390 is a microcontroller, introduced by Dallas Semiconductor (now part of Maxim Integrated Products), whose architecture is derived from that of the Intel 8051 processor series. It contains a code memory address space of twenty-two bits. It also contains two Controller Area Network (CAN) controllers and a 32-bit integer coprocessor. The open-source Small Device C Compiler (SDCC) supports the processor. It was used in the initial version of the Tiny Internet Interface (TINI) processor module where it was superseded by the DS80C400, a processor that also incorporates an Ethernet port. |
https://en.wikipedia.org/wiki/History%20of%20robots | The history of robots has its origins in the ancient world. During the industrial revolution, humans developed the structural engineering capability to control electricity so that machines could be powered with small motors. In the early 20th century, the notion of a humanoid machine was developed.
The first uses of modern robots were in factories as industrial robots. These industrial robots were fixed machines capable of manufacturing tasks which allowed production with less human work. Digitally programmed industrial robots with artificial intelligence have been built since the 2000s.
Early legends
Concepts of artificial servants and companions date at least as far back as the ancient legends of Cadmus, who is said to have sown dragon teeth that turned into soldiers and Pygmalion whose statue of Galatea came to life. Many ancient mythologies included artificial people, such as the talking mechanical handmaidens (Ancient Greek: (Kourai Khryseai); "Golden Maidens") built by the Greek god Hephaestus (Vulcan to the Romans) out of gold.
The Buddhist scholar Daoxuan (596-667 AD) described humanoid automata crafted from metals that recite sacred texts in a cloister which housed a fabulous clock. The "precious metal-people" weeped when Buddha Shakyamuni died. Humanoid automations also feature in the Epic of King Gesar, a Central Asian cultural hero.
Early Chinese lore on the legendary carpenter Lu Ban and the philosopher Mozi described mechanical imitations of animals and demons. The implications of humanoid automatons were discussed in Liezi (4th century CE), a compilation of Daoist texts which went on to become a classic. In chapter 5 King Mu of Zhou is on tour of the West and upon asking the craftsman Master Yan Shi "What can you do?" the royal court is presented with an artificial man. The automation was indistinguishable from a human and performed various tricks for the king and his entourage. But the king flew into a rage when apparently the automation starte |
https://en.wikipedia.org/wiki/Rhode%20Island%20Math%20League | The Rhode Island Mathematics League (RIML) competition consists of four meets spanning the entire year. It culminates at the state championship held at Bishop Hendricken High School. Top schools from the state championship are invited to the New England Association of Math Leagues (NEAML) championship.
Format
Each meet consists of five rounds and a team round. Each team consists of five students, and each school may have as many as six teams. However, each team may have a maximum of two seniors and four sophomores/juniors. At least one sophomore or freshman must be on each team (or the team may compete with an empty slot). Three students from each team participate in a round. Therefore, each student participates in three rounds and the team round. The first five rounds consist of three questions each. Beginning in 2007, one of the five rounds is designated as "calculator-free", in 2008, this number was increased to two, and in 2018, calculators were banned from all meets. The first question in each round is worth one point, the second two points, and the third three points. Each student works on the questions independently in the ten minutes allotted. All answers must be presented in simplified and rationalized form unless specified otherwise. After the completion of the first five rounds, there is a team round. All five players from each team collaborate on five questions worth two points each. The maximum score for one team is 100 points, and the maximum score for one student is 18 points.
Rounds
At the first meet the rounds are as follows:
Round 1: Arithmetic, Number Theory, and Matrices
Round 2: Algebra I
Round 3: Geometry
Round 4: Algebra II
Round 5: Miscellaneous Math
Team Round
At the second meet the rounds are as follows:
Round 1: Arithmetic, Number Theory, and Matrices
Round 2: Algebra I
Round 3: Geometry
Round 4: Algebra II
Round 5: Miscellaneous Math
Team Round
At the third meet the rounds are as follows:
Round 1: S |
https://en.wikipedia.org/wiki/Nasal%20placode | The nasal placode (or olfactory placode) gives rise to the olfactory epithelium of the nose. Two nasal placodes arise as thickened ectoderm from the frontonasal process. They give rise to the nose, the philtrum of the upper lip, and the primary palate.
Development
During the fifth week of human embryonic development the placodes increase in size. In the sixth week of development the centre of each placode grows inwards to form the two nasal pits. The invaginations will give rise to the olfactory epithelium that lines the roof of the nasal cavity.
The nasal pits are oval shaped and they leave a raised margin which is divided into a medial nasal process and a lateral nasal process.
The medial and lateral nasal process of each placode gives rise to the nose, the philtrum of the upper lip and the primary palate.
See also
Placode |
https://en.wikipedia.org/wiki/Laryngotracheal%20groove | The laryngotracheal groove is a precursor for the larynx and trachea.
The rudiment of the respiratory organs appears as a median longitudinal groove in the ventral wall of the pharynx. The groove deepens, and its lips fuse to form a septum, which grows from below upward and converts the groove into a tube, the laryngotracheal tube. The cephalic end opens into the pharynx through a slit-like aperture formed by the persistent anterior part of the groove. Initially, the cephalic end is in open communication with the foregut, but eventually it becomes separated by the indentations of the mesoderm, the tracheoesophageal folds.
When the tracheoesophageal folds fuse in the midline to form the tracheoesophageal septum, the foregut is divided into the trachea ventrally and the esophagus dorsally. The tube is lined by an endoderm, from which the epithelial lining of the respiratory tract is developed. The cephalic part of the tube becomes the larynx, and its next succeeding part is the trachea, while from its caudal end, a respiratory diverticulum appears as the lung bud. The lung bud branches into two lateral outgrowths known as the bronchial buds, one on each side of the trachea. The right and left bronchial buds branch into main (primary), lobar (secondary), segmental (tertiary), and subsegmental bronchi and lead to the development of the lungs. The Hox complex, FGF-10 (fibroblast growth factor), BMP-4 (bone morphogenetic protein), N-myc (a proto-oncogene), syndecan (a proteglycan), tenascin (an extracellular matrix protein), and epimorphin (a protein) appear to play a role in the development of the respiratory system. |
https://en.wikipedia.org/wiki/Studentized%20range | In statistics, the studentized range, denoted q, is the difference between the largest and smallest data in a sample normalized by the sample standard deviation.
It is named after William Sealy Gosset (who wrote under the pseudonym "Student"), and was introduced by him in 1927.
The concept was later discussed by Newman (1939), Keuls (1952), and John Tukey in some unpublished notes.
Its statistical distribution is the studentized range distribution, which is used for multiple comparison procedures, such as the single step procedure Tukey's range test, the Newman–Keuls method, and the Duncan's step down procedure, and establishing confidence intervals that are still valid after data snooping has occurred.
Description
The value of the studentized range, most often represented by the variable q, can be defined based on a random sample x1, ..., xn from the N(0, 1) distribution of numbers, and another random variable s that is independent of all the xi, and νs2 has a χ2 distribution with ν degrees of freedom. Then
has the Studentized range distribution for n groups and ν degrees of freedom. In applications, the xi are typically the means of samples each of size m, s2 is the pooled variance, and the degrees of freedom are ν = n(m − 1).
The critical value of q is based on three factors:
α (the probability of rejecting a true null hypothesis)
n (the number of observations or groups)
ν (the degrees of freedom used to estimate the sample variance)
Distribution
If X1, ..., Xn are independent identically distributed random variables that are normally distributed, the probability distribution of their studentized range is what is usually called the studentized range distribution. Note that the definition of q does not depend on the expected value or the standard deviation of the distribution from which the sample is drawn, and therefore its probability distribution is the same regardless of those parameters.
Studentization
Generally, the term studentized means that |
https://en.wikipedia.org/wiki/Fibrin%20degradation%20product | Fibrin degradation products (FDPs), also known as fibrin split products, are components of the blood produced by clot degeneration.
Clotting, also called coagulation, at the wound site produces a mass of fibrin threads called a net that remains in place until the cut is healed. As a cut heals, the clotting slows down. Eventually the clot is broken down and dissolved by plasmin. When the clot and fibrin net dissolve, fragments of protein are released into the body. These fragments are fibrin degradation products or FDPs. If your body is unable to dissolve a clot, you may have abnormal levels of FDPs. The most notable subtype of fibrin degradation products is D-dimer.
The levels of these FDPs rise after any thrombotic event.
Fibrin and fibrinogen degradation product (FDP) testing is commonly used to diagnose disseminated intravascular coagulation.
As tumor marker
A tumor marker known as AMDL-ELISA DR-70 (FDP), and now as Onko-Sure, was approved by the US FDA on July 1, 2008 for in vitro diagnostic only and serial testing for monitoring colorectal cancer with more effective by 50 percent than carcinoembryonic antigen (CEA) when CEA values is low. The Onko-Sure blood test can detect also tumors/cancers: of lung, breast, stomach, liver, colon; rectal, ovarian, esophageal, cervical, trophoblastic, thyroid, malignant lymphoma, and pancreatic.
See also
Fibrinolysis |
https://en.wikipedia.org/wiki/Lazard%27s%20universal%20ring | In mathematics, Lazard's universal ring is a ring introduced by Michel Lazard in over which the universal commutative one-dimensional formal group law is defined.
There is a universal commutative one-dimensional formal group law over a universal commutative ring defined as follows. We let
be
for indeterminates , and we define the universal ring R to be the commutative ring generated by the elements , with the relations that are forced by the associativity and commutativity laws for formal group laws. More or less by definition, the ring R has the following universal property:
For every commutative ring S, one-dimensional formal group laws over S correspond to ring homomorphisms from R to S.
The commutative ring R constructed above is known as Lazard's universal ring. At first sight it seems to be incredibly complicated: the relations between its generators are very messy. However Lazard proved that it has a very simple structure: it is just a polynomial ring (over the integers) on generators of degree 1, 2, 3, ..., where has degree . proved that the coefficient ring of complex cobordism is naturally isomorphic as a graded ring to Lazard's universal ring. Hence, topologists commonly regrade the Lazard ring so that has degree , because the coefficient ring of complex cobordism is evenly graded. |
https://en.wikipedia.org/wiki/Edison%20Wetlands%20Association | The Edison Wetlands Association was founded by noted activist Robert Spiegel in 1989 as a nonprofit environmental organization devoted to the cleanup of hazardous waste sites and the preservation of open space in densely populated central New Jersey.
EWA advocates for the cleanup of over 60 toxic waste sites, working to protect public health and the environment. EWA founded its Community Assistance Remediation Program in 2002 to empower disadvantaged grassroots citizens groups in the remediations and balanced redevelopments of local contaminated sites in their communities. Spiegel has testified three times on the Superfund before the U.S. Senate, as well as the United States National Academy of Sciences.
EWA's Conservation Program pursues the preservation of the little remaining open space in densely populated Middlesex County, as well as conservation measures. EWA developed the first-ever Middlesex County Birds report and checklist, as well as the Bald Eagle Sightings report, documenting the comeback of an endangered American icon in an urban and suburban crossroads.
EWA's leadership has been profiled in publications as diverse as the late Molly Ivins' bestselling 2003 book, Bushwhacked, and Tony Hiss's Geraldine R. Dodge publication, H2O: Highlands To Ocean. The organization is also regularly featured in United States national and regional media coverage.
Current Campaigns
American Cyanamid
History-
The American Cyanamid site is approximately 575 acres in size, located adjacent to the Raritan River and lies above the Brunswick Aquifer, New Jersey's second largest source for drinking water. Located in Bound Brook, NJ approximately 14,000 people live within a three-mile radius of the site. During approximately 64 years of operation, the company American Cyanamid has produced over 800 chemicals including pharmaceutical, dyes and textile chemicals, organic pigments rubber compounds, and various intermediate chemicals. During this time plant operators routine |
https://en.wikipedia.org/wiki/Software%20quality%20control | Software quality control is the set of procedures used by organizations to ensure that a software product will meet its quality goals at the best value to the customer, and to continually improve the organization’s ability to produce software products in the future.
Software quality control refers to specified functional requirements as well as non-functional requirements such as supportability, performance and usability. It also refers to the ability for software to perform well in unforeseeable scenarios and to keep a relatively low defect rate.
These specified procedures and outlined requirements lead to the idea of Verification and Validation and software testing.
It is distinct from software quality assurance which encompasses processes and standards for ongoing maintenance of high quality of products, e.g. software deliverables, documentation and processes - avoiding defects. Whereas software quality control is a validation of artifacts compliance against established criteria - finding defects.
Definition
Software quality control is a function that checks whether a software component, or supporting artifact meets requirements, or is "fit for use". Software Quality Control is commonly referred to as Testing.
Quality Control Activities
Check that assumptions and criteria for the selection of data and the different factors related to data are documented.
Check for transcription errors in data input and reference.
Check the integrity of database files.
Check for consistency in data.
Check that the movement of inventory data among processing steps is correct.
Check for uncertainties in data, database files etc.
Undertake review of internal documentation.
Check methodological and data changes resulting in recalculations.
Undertake completeness checks.
Compare Results to previous Results.
Software Control Methods
Rome laboratory Software framework
Goal Question Metric Paradigm
Risk Management Model
The Plan-Do-Check-Action Model of Quality Control
Total Softw |
https://en.wikipedia.org/wiki/Satellite%20navigation%20software | Satellite navigation software or GNSS navigation software usually falls into one of the following two categories:
Navigation with route calculation and directions from the software to the user of the route to take, based on a vector-based map, normally for motorized vehicles with some motorized forms added on as an afterthought.
Navigation tracking, often with a map "picture" in the background, but showing where you have been, and allowing "routes" to be preprogrammed, giving a line you can follow on the screen. This type can also be used for geocaching.
Terminology
Track
A track is a trace of somewhere that you have actually been (often called a "breadcrumb trail"). The GNSS unit (external or internal) periodically sends details of the location which are recorded by the software, either by taking a reading based on a set time interval, based on a set distance, based on a change in direction by more than a certain angle, or a combination of these. Each point is stored together with its date and time. The resulting track can be displayed as a series of the recorded points or a line connecting them.
Retracing your steps is a simple matter of following the track back to the source.
Route
A route is a preset series of points that make up a set route to follow for your destination. Most software allows the route and the track to be displayed at the same time.
Waypoint
Waypoints are used to mark particular locations, typically used as markers along the "way" to somewhere. They are either key entered by users or downloaded from other sources, depending upon the sophistication of the device. Although not linked to tracks or routes, they can be used to simplify the construction of routes, by being able to be re-used. Frequently, waypoints serve a "safety" purpose, enabling a route to be taken around obstacles such as shallow water (marine navigation) or streams/cliffs/other hazards which may prevent a safe passage directly from point "A" to point "B".
Platforms
|
https://en.wikipedia.org/wiki/Health%20record%20trust | A health record trust (also independent health record trust or health record data bank) provides a secure and protected place for individuals to create, use, and maintain their lifetime electronic health record (EHR). The health record trust takes personal health records one step further by combining an individual's electronic health record with the personal health record. A health record trust protects patient privacy by establishing that the patient is the owner of their health care records. It gives patients the authority to access and review the entire document at any time. It allows healthcare professionals, facilities, and organizations to view all or a limited portion of the records.
The health record trust allows for all of the information to be in one central document. Patients cannot alter their health records but instead add notes and request corrections. They can also view every provider who downloads their EHR.
Public policy
The legislation was introduced in the 110th Congress to establish a regulatory framework for establishing health record trusts. The Independent Health Record Trust Act of 2007 (H.R. 2991) was introduced by Rep. Dennis Moore (D-KS) and Rep. Paul Ryan (R-WI) on July 11, 2007. The legislation seeks to give people control over their lifetime health records, with the broader goal of reducing healthcare costs that result from inefficiency, medical errors, inappropriate care, and incomplete information. This legislation provides standards for using health record trusts, including certifications and interoperability of independent health record trusts. HR 2991 was referred to the House Committee on Energy and Commerce and the House Committee on Ways and Means. The bill died in committee and has not been reintroduced.
Patients receive better quality of care with the availability of a longitudinal health record protected by a health record trust. They can pass along their medical records to future generations. Health record trusts promot |
https://en.wikipedia.org/wiki/Bicone | In geometry, a bicone or dicone (from , and Greek: di-, both meaning "two") is the three-dimensional surface of revolution of a rhombus around one of its axes of symmetry. Equivalently, a bicone is the surface created by joining two congruent, right, circular cones at their bases.
A bicone has circular symmetry and orthogonal bilateral symmetry.
Geometry
For a circular bicone with radius R and height center-to-top H, the formula for volume becomes
For a right circular cone, the surface area is
where is the slant height.
See also
Sphericon
Biconical antenna |
https://en.wikipedia.org/wiki/Hudson%27s%20equation | Hudson's equation, also known as Hudson formula, is an equation used by coastal engineers to calculate the minimum size of riprap (armourstone) required to provide satisfactory stability characteristics for rubble structures such as breakwaters under attack from storm wave conditions.
The equation was developed by the United States Army Corps of Engineers, Waterways Experiment Station (WES), following extensive investigations by Hudson (1953, 1959, 1961a, 1961b)
Initial equation
The equation itself is:
where:
W is the design weight of the riprap armor (Newton)
is the specific weight of the armor blocks (N/m3)
H is the design wave height at the toe of the structure (m)
KD is a dimensionless stability coefficient, deduced from laboratory experiments for different kinds of armour blocks and for very small damage (a few blocks removed from the armour layer) (-):
KD = around 3 for natural quarry rock
KD = around 10 for artificial interlocking concrete blocks
Δ is the dimensionless relative buoyant density of rock, i.e. (ρr / ρw - 1) = around 1.58 for granite in sea water
ρr and ρw are the densities of rock and (sea)water (-)
θ is the angle of revetment with the horizontal
Updated equation
This equation was rewritten as follows in the nineties:
where:
Hs is the design significant wave height at the toe of the structure (m)
Δ is the dimensionless relative buoyant density of rock, i.e. (ρr / ρw - 1) = around 1.58 for granite in sea water
ρr and ρw are the densities of rock and (sea)water (kg/m3)
Dn50 is the nominal median diameter of armor blocks = (W50/ρr)1/3 (m)
KD is a dimensionless stability coefficient, deduced from laboratory experiments for different kinds of armor blocks and for very small damage (a few blocks removed from the armor layer) (-):
KD = around 3 for natural quarry rock
KD = around 10 for artificial interlocking concrete blocks
θ is the angle of revetment with the horizontal
The armourstone may be considered stable if the stability numbe |
https://en.wikipedia.org/wiki/MusicStation | MusicStation is a music service platform developed by Omnifone. The cloud based platform works across a number of different digital device platforms and is tailored specifically for various partners including rara.com, RIM, Sony, Sony Ericsson, HP, BSkyB, Vodafone, Telenor, Hutchison Telecom and Vodacom.
Services
In February 2009 Omnifone introduced MusicStation Next Generation, a white label unlimited music audio and video download and streaming solution for ISPs and consumer electronics device vendors. The cloud based solution offers interoperable music on a wide range of connected devices, including home entertainment systems, gaming consoles, and car audio systems.
In April 2009 Hutchison Telecom became the first Omnifone partner to introduce MusicStation Desktop Edition to users in Hong Kong on its 3 network.
Omnifone's MusicStation was the first subscription based unlimited music download service available on mobiles in the UK launching with Vodafone in November 2007.
In June 2008, Omnifone announced that MusicStation had become the largest subscription based unlimited music download service in the UK within seven months of launching on Vodafone.
History
14 January 2003: British Entrepreneurs Rob Lewis, Phil Sant and Mark Knight form Omnifone
12 February 2007: Omnifone announces MusicStation
14 June 2007: MusicStation goes live with its first rollout in Sweden with Telenor
22 October 2007: MusicStation launches in Hong Kong with 3
1 November 2007: MusicStation launches in South Africa with Vodacom
1 November 2007: MusicStation launches in the UK with Vodafone
12 February 2008: Omnifone launches MusicStation Max, a pre-licensed unlimited music phone development programme with LG signed up as a handset manufacturer
15 September 2008: MusicStation launches in New Zealand with Vodafone
22 September 2008: MusicStation launches in Australia with Vodafone
24 September 2008: Omnifone announces partnership with Sony Ericsson to power PlayNow plus, its p |
https://en.wikipedia.org/wiki/Word%20addressing | In computer architecture, word addressing means that addresses of memory on a computer uniquely identify words of memory. It is usually used in contrast with byte addressing, where addresses uniquely identify bytes. Almost all modern computer architectures use byte addressing, and word addressing is largely only of historical interest. A computer that uses word addressing is sometimes called a word machine.
Basics
Consider a computer which provides 524,288 (219) bits of memory. If that memory is arranged in a byte-addressable flat address space using 8-bit bytes, then there are 65,536 (216) valid addresses, from 0 to 65,535, each denoting an independent 8 bits of memory. If instead it is arranged in a word-addressable flat address space using 32-bit words, then there are 16,384 (214) valid addresses, from 0 to 16,383, each denoting an independent 32 bits.
More generally, the minimum addressable unit (MAU) is a property of a specific memory abstraction. Different abstractions within a computer may use different MAUs, even when they are representing the same underlying memory. For example, a computer might use 32-bit addresses with byte addressing in its instruction set, but the CPU's cache coherence system might work with memory only at a granularity of 64-byte cache lines, allowing any particular cache line to be identified with only a 26-bit address and decreasing the overhead of the cache.
The address translation done by virtual memory often affects the structure and width of the address space, but it does not change the MAU.
Trade-offs of different minimum addressable units
The size of the minimum addressable unit of memory can have complex trade-offs. Using a larger MAU allows the same amount of memory to be covered with a smaller address, which can substantially decrease the memory requirements of a program. However, using a smaller MAU makes it easier to work efficiently with small items of data.
Suppose a program wishes to store one of the 12 |
https://en.wikipedia.org/wiki/Entropy%20%28journal%29 | Entropy is a monthly open access scientific journal covering research on all aspects of entropy and information theory. It was established in 1999 and is published by MDPI. The journal occasionally publishes special issues compiled by guest editors. The editor-in-chief is Kevin H. Knuth (University at Albany, SUNY).
Sections
Entropy consists of eight sections:
Thermodynamics Section
Statistical Mechanics
Information Theory
Quantum Information
Complexity
Astrophysics and Cosmology
Entropy Reviews
Entropy and Biology
Abstracting and indexing
The journal is abstracted and indexed in:
According to the Journal Citation Reports, the journal has a 2022 impact factor of 2.7.
2013 Paper on glyphosate
In 2013, Entropy published a review paper saying glyphosate may be the most important factor in the development of obesity, depression, attention deficit hyperactivity disorder, autism, Alzheimer's disease, Parkinson's disease, multiple sclerosis, cancer, and infertility. The paper does not contain any primary research results. It was criticized as pseudo-science by the science magazine Discover and Jeffrey Beall, founder of Beall's List of predatory open-access publishers, said "Will MDPI publish anything for money?". Beall removed MDPI from his list of predatory publishers in October 2015.
In response to the controversy, the editors of Entropy added an "Expression of Concern" to the article's frontmatter. In 2017 researchers Robin Mesnage and Michael N. Antoniou, both of whom are working to limit the use of glyphosate, said that "although evidence exists that glyphosate-based herbicides are toxic below regulatory set safety limits, the arguments of [authors] Samsel and Seneff largely serve to distract rather than to give a rational direction." |
https://en.wikipedia.org/wiki/Mechanostat | The Mechanostat is a term describing the way in which mechanical loading influences bone structure by changing the mass (amount of bone) and architecture (its arrangement) to provide a structure that resists habitual loads with an economical amount of material. As changes in the skeleton are accomplished by the processes of formation (bone growth) and resorption (bone loss), the mechanostat models the effect of influences on the skeleton by those processes, through their effector cells, osteocytes, osteoblasts, and osteoclasts. The term was invented by Harold Frost: an orthopaedic surgeon and researcher described extensively in articles referring to Frost and Webster Jee's Utah Paradigm of Skeletal Physiology in the 1960s. The Mechanostat is often defined as a practical description of Wolff's law described by Julius Wolff (1836–1902), but this is not completely accurate. Wolff wrote his treatises on bone after images of bone sections were described by Culmann and von Meyer, who suggested that the arrangement of the struts (trabeculae) at the ends of the bones were aligned with the stresses experienced by the bone. It has since been established that the static methods used for those calculations of lines of stress were inappropriate for work on what were, in effect, curved beams, a finding described by Lance Lanyon, a leading researcher in the area as "a triumph of a good idea over mathematics." While Wolff pulled together the work of Culmann and von Meyer, it was the French scientist Roux, who first used the term "functional adaptation" to describe the way that the skeleton optimized itself for its function, though Wolff is credited by many for that.
According to the Mechanostat, bone growth and bone loss is stimulated by the local, mechanical, elastic deformation of bone. The reason for the elastic deformation of bone is the peak forces caused by muscles (e.g. measurable using mechanography). The adaptation (feed-back control loop) of bone according to the maximu |
https://en.wikipedia.org/wiki/FAC%20System | FAC System is a mechanical construction set, invented by Swedish artist Mark Sylwan and launched in 1952. It uses metallic parts, and extensively uses steel rod. The company continued in business , supplying kits stated to give industry a basic construction system which gives the same freedom in 3D design as is available from the drawing board and CAD/CAM. Kits are suitable, and priced as, for laboratory work and teaching.
Shafts have a diameter of 6 mm, and have a groove along the shaft. Gears or other wheels have a bevel on either side of the hub of the wheel. A special sheet metal angle fits into this chamfer, and when the wheel is mounted on an axle, this sheet metal angle will fit into the groove on the axle. The design allows the wheel to be fixed on the axle and the stop screws on the wheel only have the task of ensuring that the wheel does not move along the axle.
See also
Meccano |
https://en.wikipedia.org/wiki/Genetic%20algorithm%20scheduling | The genetic algorithm is an operational research method that may be used to solve scheduling problems in production planning.
Importance of production scheduling
To be competitive, corporations must minimize inefficiencies and maximize productivity. In manufacturing, productivity is inherently linked to how well the firm can optimize the available resources, reduce waste and increase efficiency. Finding the best way to maximize efficiency in a manufacturing process can be extremely complex. Even on simple projects, there are multiple inputs, multiple steps, many constraints and limited resources. In general a resource constrained scheduling problem consists of:
A set of jobs that must be executed
A finite set of resources that can be used to complete each job
A set of constraints that must be satisfied
Temporal constraints – the time window to complete the task
Procedural constraints – the order each task must be completed
Resource constraints – is the resource available
A set of objectives to evaluate the scheduling performance
A typical factory floor setting is a good example of this, where it is necessary to schedule which jobs need to be completed on which machines, by which employees, in what order and at what time.
Use of algorithms in scheduling
In very complex problems such as scheduling there is no known way to get to a final answer, so we resort to searching for it trying to find a "good" answer. Scheduling problems most often use heuristic algorithms to search for the optimal solution. Heuristic search methods suffer as the inputs become more complex and varied. This type of problem is known in computer science as an NP-Hard problem. This means that there are no known algorithms for finding an optimal solution in polynomial time.
Genetic algorithms are well suited to solving production scheduling problems, because unlike heuristic methods genetic algorithms operate on a population of solutions rather than a single solution. In production schedu |
https://en.wikipedia.org/wiki/Nutshell | A nutshell is the outer shell of a nut. Most nutshells are inedible and are removed before eating the nut meat inside. It covers and protects the kernel, which may be edible.
Usage
Most nutshells are useful to some extent, depending on the circumstances. Walnut shells can be used for cleaning and polishing, as a filler in dynamite, and as a paint thickening agent. Shells from pecans, almonds, Brazil nuts, acorns, and most other nuts are useful in composting. Their high porosity makes them also ideal in the production of activated carbon by pyrolysis. Shells can also be used as loose-fill packing material, to protect fragile items in shipping.
Idiomatic usage
The expression "in a nutshell" (of a story, proof, etc.) means "in essence", metaphorically alluding to the fact that the essence of the nut - its edible part - is contained inside its shell. The expression further gave rise to the journalistic term nut graph, short for nutshell paragraph.
In Hamlet (Act 2, Scene 2) the title character exclaims: "O God, I could be bounded in a nutshell, and count myself a King of infinite space".
Pliny the Elder mentioned in the encyclopedic Naturalis historia a report by Cicero saying that a handwritten version of the Iliad by Homer would have fit in a nut[shell]: "in nuce inclusam Iliadem Homeri carmen in membrana scriptum tradit Cicero". |
https://en.wikipedia.org/wiki/Zeroshell | Zeroshell is a small open-source Linux distribution for servers and embedded systems which aims to provide network services. Its administration relies on a web-based graphical interface; no shell is needed to administer and configure it. Zeroshell is available as Live CD and CompactFlash images, and VMware virtual machines.
Zeroshell can be installed on any IA-32 computer with almost any Ethernet interface. It can also be installed on most embedded devices and single-board computers such as Raspberry Pi and Orange Pi.
The project reached EOL in April of 2021 with the version 3.9.5.
There are several known vulnerabilities for various versions of this software: V2, V3.6x up to V3.7, V3.9.0, V3.9.3 and last V3.9.5 for example, allowing an attacker to e.g. gain root access to the device easily. The main attack vector is the cgi script in use, 'kerbynet'.
Selected features
RADIUS server which is able to provide strong authentication for the Wireless clients by using IEEE 802.1X and Wi-Fi Protected Access (WPA/WPA2) protocols
Captive portal for network authentication in the HotSpots by using a web browser. The credentials can be verified against a Radius server, a Kerberos 5 KDC (such as Active Directory KDC)
Netfilter – Firewall, Packet Filter and Stateful Packet Inspection (SPI), Layer 7 filter to block or shape the connections generated by Peer to Peer clients
Linux network scheduler – control maximum bandwidth, the guaranteed bandwidth and the priority of some types of traffic such as VoIP and peer-to-peer
VPN host-to-LAN and LAN-to-LAN with the IPSec/L2TP and OpenVPN protocols
Routing and Bridging capabilities with VLAN IEEE 802.1Q support
Multizone DNS (Domain name system) server
Multi subnet DHCP server
PPPoE client for connection to the WAN (Wide area network) via ADSL, DSL and cable lines
Dynamic DNS client updater for DynDNS
NTP (Network Time Protocol) client and server
Syslog server for receiving and cataloging the system logs produced by the re |
https://en.wikipedia.org/wiki/Rotavirus%20translation | Rotavirus translation, the process of translating mRNA into proteins, occurs in a different way in Rotaviruses. Unlike the vast majority of cellular proteins in other organisms, in Rotaviruses the proteins are translated from capped but nonpolyadenylated mRNAs. The viral nonstructural protein NSP3 specifically binds the 3'-end consensus sequence of viral mRNAs and interacts with the eukaryotic translation initiation factor eIF4G. The Rotavirus replication cycle occurs entirely in the cytoplasm. Upon virus entry, the viral transcriptase synthesizes capped but nonpolyadenylated mRNA The viral mRNAs bear 5' and 3' untranslated regions (UTR) of variable length and are flanked by two different sequences common to all genes.
In the group A rotaviruses, the 3'-end consensus sequence UGACC is highly conserved among the 11 genes. Rotavirus NSP3 presents several similarities to PABP; in rotavirus-infected cells, NSP3 can be cross-linked to the 3' end of rotavirus mRNAs and is coimmunoprecipitated with eIF4G. The binding of NSP3A to eIF4G and its specific interaction with the 3' end of viral mRNA brings the viral mRNA and the translation initiation machinery into contact, thus favoring efficient translation of the viral mRNA. NSP3 interacts with the same region of eIF4G as PABP does. As a consequence, during rotavirus infection PABP is evicted from eIF4G, probably impairing the translation of polyadenylated mRNA and leading to the shutoff of cellular mRNA translation observed during rotavirus infection. |
https://en.wikipedia.org/wiki/Current%20Index%20to%20Statistics | The Current Index to Statistics is an online database published by the Institute of Mathematical Statistics and the American Statistical Association that contains bibliographic data of articles in statistics, probability, and related fields. It was shut down at the end of 2019.
See also
Web of Science
IEEE Xplore |
https://en.wikipedia.org/wiki/Geographical%20Names%20Board%20of%20Canada | The Geographical Names Board of Canada (GNBC) is a national committee with a secretariat in Natural Resources Canada, part of the Government of Canada, which authorizes the names used and name changes on official federal government maps of Canada.
History
It was created in December 1897, by Order in Council, as the Geographic Board of Canada. It consisted of one Board member from each of four Government of Canada departments, as well as the Surveyor General of Dominion Lands, while a secretariat was provided by the then-extant Department of the Interior. In December 1899, the Order in Council was amended to give the Canadian provinces and territories the right to nominate one official, each, to be a Board member.
Structure
, the Board consists of 27 members, one from each of the provinces and territories, and others from departments of the Government of Canada. The board also is involved with names of areas in the Antarctic through the Antarctic Treaty. The secretariat is provided by Natural Resources Canada.
In addition to the provincial and territorial members are members from the following federal government departments: Aboriginal Affairs and Northern Development Canada, Canada Post Corporation, Fisheries and Oceans Canada, Elections Canada, Library and Archives Canada, Department of National Defence, Natural Resources Canada (including Geological Survey of Canada and Canada Centre for Mapping and Earth Observation), Parks Canada, Statistics Canada, and the Translation Bureau. The Chair of the Geographical Names Board of Canada is Connie Wyatt Anderson from The Pas, Manitoba.
Process
In a two year period of 2019-2020, 750 names were added to the database with roughly 100 changes to names of already existing places. Citizens and government officials have the ability to write in with a form that is able to be filled out. The local naming authority then becomes involved on the place in question gathering suggestions from the local and indigenous communities. T |
https://en.wikipedia.org/wiki/Astronaut%20ranks%20and%20positions | Astronauts hold a variety of ranks and positions. Each of these roles carries responsibilities that are essential to the operation of a spacecraft. A spacecraft's cockpit, filled with sophisticated equipment, requires skills differing from those used to manage the scientific equipment on board, and so on.
NASA ranks and positions
Ranks
Members of the NASA Astronaut Corps hold one of two ranks. Astronaut Candidate is the rank of those training to be NASA astronauts.
Upon graduation, candidates are promoted to Astronaut and receive their Astronaut Pin. The pin is issued in two grades, silver and gold, with the silver pin awarded to candidates who have successfully completed astronaut training and the gold pin to astronauts who have flown in space.
Chief of the Astronaut Office is a position, not a rank.
Positions
Roscosmos and Soviet space program ranks and positions
Ranks
Cosmonauts are professional space travellers from Russia. After initial training, cosmonauts are assigned as either a test-cosmonaut (космонавт-испытатель, kosmonavt-ispytatel') or a research-cosmonaut (космонавт-исследователь, kosmonavt-issledovatel'). A test-cosmonaut has a more difficult preparation than a research-cosmonaut and can be the commander or the flight engineer of a spacecraft, while a research-cosmonaut cannot.
Higher ranks include pilot-cosmonaut, test-cosmonaut instructor, and research-cosmonaut instructor.
Pilot-Cosmonaut of the Russian Federation is a title that is presented to all cosmonauts who fly for the Russian space program.
Positions
China National Space Administration positions
Ranks
Similarly to NASA, members of the China National Space Administration (CNSA) hold one of two ranks. Astronaut Candidate is the rank of those training to be CNSA astronauts. The positions of Spacecraft Pilot, Flight Engineer, and Mission Payload Specialist were listed in the announcement for the Group 3 selection.
Upon graduation, candidates are promoted to Astronaut.
Position |
https://en.wikipedia.org/wiki/Jordan%27s%20rule | Jordan's rule (sense 1) is an ecogeographical rule that describes the inverse relationship between water temperature and meristic characteristics in various species of fish. The most commonly observed relationship is that fin ray, vertebrae, or scale numbers increase with decreasing temperature. The rule is named after David Starr Jordan (1851–1931), the father of American ichthyology.
Jordan's law (or rule) (sense 2) is also an ecogeographical rule (named after the same scientist) that states: "‘[g]iven any species in any region, the nearest related species is not likely to be found in the same region nor in a remote region, but in a neighbouring district separated from the first by a barrier of some sort’ This "rule" is frequently violated (see discussion in Fitzpatrick & Turelli 2007), but when patterns are consistent with Jordan's rule (sense 2), this suggests an important role for allopatric speciation in the diversification of the clade in question. Jordan himself wrote: "To this generalization Dr. Allen, in a late number of Science, gives the name of 'Jordan's Law.' The present writer makes no claim to the discovery of this law. The language above quoted is his, but the idea is familiar to all students of geographical distribution and goes back to the master in that field, Moritz Wagner." Thus, Jordan's law is an example of Stigler's law.
See also
Allen's rule
Allopatric speciation
Bergman's rule
Jordan's Principle |
https://en.wikipedia.org/wiki/Cultural%20materialism%20%28anthropology%29 | Cultural materialism is an anthropological research orientation first introduced by Marvin Harris in his 1968 book The Rise of Anthropological Theory, as a theoretical paradigm and research strategy. It is said to be the most enduring achievement of that work. Harris subsequently developed a full elaboration and defense of the paradigm in his 1979 book Cultural Materialism. To Harris social change is dependent of three factors: a society's infrastructure, structure, and superstructure.
Harris's concept of cultural materialism was influenced by the writings of Karl Marx and Friedrich Engels, as well as their theories as modified by Karl August Wittfogel and his 1957 book, Oriental Despotism: A Comparative Study of Total Power. Yet this materialism is distinct from Marxist dialectical materialism, as well as from philosophical materialism. Thomas Malthus's work encouraged Harris to consider reproduction of equal importance to production. The research strategy was also influenced by the work of earlier anthropologists including Herbert Spencer, Edward Tylor and Lewis Henry Morgan who, in the 19th century, first proposed that cultures evolved from the less complex to the more complex over time. Leslie White and Julian Steward and their theories of cultural evolution and cultural ecology were instrumental in the reemergence of evolutionist theories of culture in the 20th century and Harris took inspiration from them in formulating cultural materialism.
Epistemological principles
Cultural materialism is a scientific research strategy and as such utilizes the scientific method. Other important principles include operational definitions, Karl Popper's falsifiability, Thomas Kuhn's paradigms, and the positivism first proposed by Auguste Comte and popularized by the Vienna Circle. The primary question that arises in applying the techniques of science to understand the differences and similarities between cultures is how the research strategy "treats the relationship betwee |
https://en.wikipedia.org/wiki/Cultural%20materialism%20%28cultural%20studies%29 | Cultural materialism in literary theory and cultural studies traces its origin to the work of the left-wing literary critic Raymond Williams. Cultural materialism espouses analysis based in critical theory, in the tradition of the Frankfurt School.
Overview
Cultural materialism emerged as a theoretical movement in the early 1980s along with new historicism, an American approach to early modern literature, with which it shares common ground. The term was coined by Williams, who used it to describe a theoretical blending of leftist culturalism and Marxist analysis. Cultural materialists deal with specific historical documents and attempt to analyze and recreate the zeitgeist of a particular moment in history.
Williams viewed culture as a "productive process", part of the means of production, and cultural materialism often identifies what he called "residual", "emergent" and "oppositional" cultural elements. Following in the tradition of Herbert Marcuse, Antonio Gramsci and others, cultural materialists extend the class-based analysis of traditional Marxism (Neo-Marxism) by means of an additional focus on the marginalized.
Cultural materialists analyze the processes by which hegemonic forces in society appropriate canonical and historically important texts, such as Shakespeare and Austen, and utilize them in an attempt to validate or inscribe certain values on the cultural imaginary. Jonathan Dollimore and Alan Sinfield, authors of Political Shakespeare, have had considerable influence in the development of this movement and their book is considered to be a seminal text. They have identified four defining characteristics of cultural materialism as a theoretical device:
Historical context
Close textual analysis
Political commitment
Theoretical method
Cultural materialists seek to draw attention to the processes being employed by contemporary power structures, such as the church, the state or the academy, to disseminate ideology. To do this they explore a text’s |
https://en.wikipedia.org/wiki/Apioperdon | Apioperdon pyriforme commonly known as the pear-shaped puffball or stump puffball, is a saprobic fungus present throughout much of the world. Emerging in autumn, this puffball is common and abundant on decaying logs of both deciduous and coniferous wood. It is considered a choice edible when still immature and the inner flesh is white. It is often called Lycoperdon pyriforme, but was transferred to Apioperdon in 2017 based on phylogenetic and morphological differences. It is the only species in the genus.
Taxonomy
The fungus was first described in the scientific literature by Jacob Christian Schaeffer in 1774. In 2001, DNA evidence compiled by Dirk Krüger and a number of other mycologists suggested that the genus Lycoperdon was polyphyletic, with the pear-shaped puffball differing most significantly from the group. This finding was supported by several morphological differences including the presence of rhizomorphs and its preference for wood. A 2003 publication of further research moved the puffball to the genus Morganella with the specific epithet's ending changed for agreement. In 2008, however, Larsson and Jeppson revisited the phylogeny of the Lycoperdaceae, with a broader sampling of species, and retained the taxon in Lycoperdon. It was renamed Apioperdon pyriforme in 2017. The specific epithet pyriforme is Latin for "pear-shaped".
Description
The fruit body of the pear-shaped puffball measures 1.5 to 4.5 cm in width by 2 to 4.5 cm in height. They are often pear-shaped as the name suggests, but they may also be nearly spherical. When very young they are covered in small white spines that typically fall off before maturity. A small developing pore may be visible at the top, while the sterile base of the mushroom is small and appears to be pinched in. Colour ranges from nearly white to yellowish brown with the darker shades developing with age, though the base remains white. The central pore ruptures at late maturity to allow the wind and rain to disperse the |
https://en.wikipedia.org/wiki/MANET%20database | The Molecular Ancestry Network (MANET) database is a bioinformatics database that maps evolutionary relationships of protein architectures directly onto biological networks. It was originally developed by Hee Shin Kim, Jay E. Mittenthal and Gustavo Caetano-Anolles in the Department of Crop Sciences of the University of Illinois at Urbana-Champaign.
MANET traces for example the ancestry of individual metabolic enzymes in metabolism with bioinformatic, phylogenetic, and statistical methods. MANET currently links information in the Structural Classification of Proteins (SCOP) database, the metabolic pathways database of the Kyoto Encyclopedia of Genes and Genomes (KEGG), and phylogenetic reconstructions describing the evolution of protein fold architecture at a universal level. The database has been updated to reflect evolution of metabolism at the level of protein fold families. MANET literally "paints" the ancestries of enzymes derived from rooted phylogenetic trees directly onto over one hundred metabolic pathways representations, paying homage to one of the fathers of impressionism. It also provides numerous functionalities that enable searching specific protein folds with defined ancestry values, displaying the distribution of enzymes that are painted, and exploring quantitative details describing individual protein folds. This permits the study of global and local metabolic network architectures, and the extraction of evolutionary patterns at global and local levels.
A statistical analysis of the data in MANET showed for example a patchy distribution of ancestry values assigned to protein folds in each subnetwork, indicating that evolution of metabolism occurred globally by widespread recruitment of enzymes. MANET was used recently to sort out enzymatic recruitment processes in metabolic networks and propose that modern metabolism originated in the purine nucleotide metabolic subnetwork. The database is useful for the study of metabolic evolution.
External lin |
https://en.wikipedia.org/wiki/Header%20check%20sequence | A header check sequence (HCS) is an error checking feature for various header data structures, such as in the media access control (MAC) header of Ethernet. It may consist of a cyclic redundancy check (CRC) of the frame, obtained as the remainder of the division (modulo 2) by the generator polynomial multiplied by the content of the header excluding the HCS field.
The HCS can be one octet long, as in WiMAX, or a 16-bit value for cable modems.
See also
Checksum |
https://en.wikipedia.org/wiki/Dry%20lab | A dry lab is a laboratory where the nature of the experiments does not involve significant risk. This is in contrast to a wet lab where it is necessary to handle various types of chemicals and biological hazards. An example of a dry lab is one where computational or applied mathematical analyses are done on a computer-generated model to simulate a phenomenon in the physical realm. Examples of such phenomena include a molecule changing quantum states, the event horizon of a black hole or anything that otherwise might be impossible or too dangerous to observe under normal laboratory conditions. This term may also refer to a lab that uses primarily electronic equipment, for example, a robotics lab. A dry lab can also refer to a laboratory space for the storage of dry materials.
Dry labbing can also refer to supplying fictional (yet plausible) results in lieu of performing an assigned experiment, or carrying out a systematic review.
In silico chemistry
As computing power has grown exponentially this approach to research, often referred to as in silico (as opposed to in vitro and in vivo), has amassed more attention, especially in the area of bioinformatics. More specifically, within bioinformatics, is the study of proteins or proteomics, which is the elucidation of their unknown structures and folding patterns. The general approach in the elucidation of protein structure has been to first purify a protein, crystallize it and then send X-rays through such a purified protein crystal to observe how these x-rays diffract into specific pattern—a process referred to as X-ray crystallography. However, many proteins, especially those embedded in cellular membranes, are nearly impossible to crystallize due to their hydrophobic nature. Although other techniques exists, such as ramachandran plotting and mass spectrometry, these alone generally do not lead to the full elucidation of protein structure or folding mechanisms.
Distributed computing
As a means of surpassing the l |
https://en.wikipedia.org/wiki/Development%20of%20the%20nervous%20system%20in%20humans | The development of the nervous system in humans, or neural development or neurodevelopment involves the studies of embryology, developmental biology, and neuroscience to describe the cellular and molecular mechanisms by which the complex nervous system forms in humans, develops during prenatal development, and continues to develop postnatally.
Some landmarks of neural development in the embryo include the formation and differentiation of neurons from stem cell precursors (neurogenesis); the migration of immature neurons from their birthplaces in the embryo to their final positions; the outgrowth of axons from neurons and guidance of the motile growth cone through the embryo towards postsynaptic partners, the generation of synapses between these axons and their postsynaptic partners, the synaptic pruning that occurs in adolescence, and finally the lifelong changes in synapses which are thought to underlie learning and memory.
Typically, these neurodevelopmental processes can be broadly divided into two classes: activity-independent mechanisms and activity-dependent mechanisms. Activity-independent mechanisms are generally believed to occur as hardwired processes determined by genetic programs played out within individual neurons. These include differentiation, migration and axon guidance to their initial target areas. These processes are thought of as being independent of neural activity and sensory experience. Once axons reach their target areas, activity-dependent mechanisms come into play. Neural activity and sensory experience will mediate formation of new synapses, as well as synaptic plasticity, which will be responsible for refinement of the nascent neural circuits.
Development of the human brain
Overview
The central nervous system (CNS) is derived from the ectoderm—the outermost tissue layer of the embryo. In the third week of human embryonic development the neuroectoderm appears and forms the neural plate along the dorsal side of the embryo. The neural |
https://en.wikipedia.org/wiki/OpenFOAM | OpenFOAM (Open Field Operation And Manipulation) is a C++ toolbox for the development of customized numerical solvers, and pre-/post-processing utilities for the solution of continuum mechanics problems, most prominently including computational fluid dynamics (CFD).
The OpenFOAM software is used in research organisations, academic institutes and across many types of industries, for example, automotive, manufacturing, process engineering and environmental engineering.
OpenFOAM is open-source software which is freely available and licensed under the GNU General Public License Version 3, with the following variants:
OpenFOAM, released by OpenCFD Ltd. (with the name trademarked since 2007) first released as open-source in 2004. (Note: since 2012, OpenCFD Ltd is wholly-owned subsidiary of ESI Group)
FOAM-Extend, released by Wikki Ltd. (since 2009)
OpenFOAM, released by OpenFOAM Foundation. (since 2011)
History
The name FOAM has been claimed to appear for the first time as a post-processing tool written by Charlie Hill, in the early 90s in Prof. David Gosman's group in Imperial College London. As a counter argument , it has been claimed that Henry Weller created the FOAM library for field operation and manipulation which interfaced to the GUISE (Graphical User Interface Software Environment) which was created by Charlie Hill for interfacing to AVS.
As a continuum mechanics / computational fluid dynamics tool, the first development of FOAM (which became OpenFOAM later on) was virtually always presumed to be initiated by Henry Weller at the same institute by using the C++ programming language rather than the de facto standard programming language FORTRAN of the time to develop a powerful and flexible general simulation platform. From this initiation to the founding of a company called Nabla Ltd, (predominantly) Henry Weller and Hrvoje Jasak carried out the basic development of the software for almost a decade.
For a few years, FOAM was sold as a commercial code by |
https://en.wikipedia.org/wiki/Zariski%27s%20main%20theorem | In algebraic geometry, Zariski's main theorem, proved by , is a statement about the structure of birational morphisms stating roughly that there is only one branch at any normal point of a variety. It is the special case of Zariski's connectedness theorem when the two varieties are birational.
Zariski's main theorem can be stated in several ways which at first sight seem to be quite different, but are in fact deeply related. Some of the variations that have been called Zariski's main theorem are as follows:
The total transform of a normal fundamental point of a birational map has positive dimension. This is essentially Zariski's original form of his main theorem.
A birational morphism with finite fibers to a normal variety is an isomorphism to an open subset.
The total transform of a normal point under a proper birational morphism is connected.
A closely related theorem of Grothendieck describes the structure of quasi-finite morphisms of schemes, which implies Zariski's original main theorem.
Several results in commutative algebra that imply the geometric form of Zariski's main theorem.
A normal local ring is unibranch, which is a variation of the statement that the transform of a normal point is connected.
The local ring of a normal point of a variety is analytically normal. This is a strong form of the statement that it is unibranch.
The name "Zariski's main theorem" comes from the fact that Zariski labelled it as the "MAIN THEOREM" in .
Zariski's main theorem for birational morphisms
Let f be a birational mapping of algebraic varieties V and W. Recall that f is defined by a closed subvariety (a "graph" of f) such that the projection on the first factor induces an isomorphism between an open and , and such that is an isomorphism on U too. The complement of U in V is called a fundamental variety or indeterminacy locus, and the image of a subset of V under is called a total transform of it.
The original statement of the theorem in reads:
MAIN THEORE |
https://en.wikipedia.org/wiki/Lawrence%20E.%20Blume | Lawrence E. Blume is the Distinguished Arts and Sciences Professor of Economics and Professor of Information Science at Cornell University, US.
He is a visiting research professor at IHS Vienna and a member of the external faculty at the Santa Fe Institute, where he has served as co-director of the economics program and on the institute's steering committee. He teaches and conducts research in general equilibrium theory and game theory, and also has research projects on natural resource management, network design, and evolutionary processes in markets and games. A Fellow of the Econometric Society, he received a BA in economics from Washington University in St. Louis and a PhD in economics from the University of California, Berkeley.
Blume was one of the general editors of The New Palgrave Dictionary of Economics, to which he contributed several articles on mathematical economics: Convexity, convex programming, and duality. Currently he is the associate editor of the Journal of Economic Literature.
Selected publications |
https://en.wikipedia.org/wiki/Antagomir | Antagomirs, also known as anti-miRs, are a class of chemically engineered oligonucleotides designed to silence endogenous microRNAs (also known as miRNAs or miRs).
Antagomirs are a kind of antisense oligonucleotide, as their sequence is complementary to their specific miRNA target. Their structure has modifications so as to make them more resistant to degradation. These include 2'-methoxy groups on the ribose sugar, backbones with phosphorothioate bonds, and cholesterol conjugation on the 3' end.
Mechanism of action
Antagomirs are microRNA inhibitors that bind miRNAs and prevent them from binding to a target mRNA molecule, and the consequent degradation of that mRNA via the RNA-induced silencing complex (RISC). Due to the promiscuity of microRNAs, each of which regulate multiple mRNAs, antagomirs can potentially affect the expression of many different mRNA molecules besides the desired target.
Blockmirs are similarly engineered molecules which, on the other hand, are designed to have a sequence that is complementary to an mRNA sequence that is targeted by a microRNA. Upon binding to an untranslated region of an mRNA, blockmirs sterically block microRNAs from binding to the same site. Because blockmirs bind individual mRNAs and not miRNAs, their activity is more predictable than antagomirs' and less likely to cause off-target effects.
Applications
Antagomirs are used as a method to constitutively inhibit the activity of specific miRNAs associated with disease. For example, antagomirs against miR-21 have been successfully used to inhibit fibrosis of heart and lung.
HCV
The primary method for using microRNA technology to target the Hepatitis C Virus (HCV) is by knocking-out the liver-specific microRNA. miRNA-122 binds to the 5' untranslated region of HCV's mRNA strand and, contrary to miRNA's normal function of repressing mRNA, actually upregulates the expression of the Hepatitis C Virus. Thus, the therapeutic goal in such a case would be to keep miRNA-122 f |
https://en.wikipedia.org/wiki/Certified%20Senders%20Alliance | The Certified Senders Alliance is an email whitelist service provided by eco and the German Dialogmarketing Association (Deutscher Dialogmarketing Verband).
The CSA whitelist is a positive list for email bulk senders. With ISPs and email-providers participating in the whitelist, solicited email does not need to enter the ISPs' spamfilter any longer - the chances of false positives (legitimate email erroneously tagged as spam) are reduced. Bulk email senders can apply for certification for the whitelist. They have to comply with the admission criteria, such as Closed-loop authentication, providing Masthead, revocation and unsubscribe functions, a clear identity, having sole technological control of the servers used to send emails and more.
Applicants have to submit samples of their services, which prove that the CSA criteria have been implemented. They also have to pay a fee. Applications are reviewed by a certification committee. To date, more than 100 providers participate in the whitelist.
The CSA cooperates with eco's complaints office, where internet users can report breaches of the CSA criteria. The CSA frequently discloses, if members have breached the CSA criteria, or where excluded from the whitelist. Among the providers who have joined the CSA whitelist are: AOL, Arcor, Freenet, Host Europe Group, Microsoft, Vodafone Kabel Deutschland and Pironet, and the United Internet companies GMX, 1&1 Internet and web.de as well as Yahoo and Yandex.
Once a year, the CSA holds its congress "CSA Summit" where current legal requirements regarding email marketing are discussed, such as the General Data Protection Regulation as well as technical challenges and trends. |
https://en.wikipedia.org/wiki/Ultrasound%20attenuation%20spectroscopy | Ultrasound attenuation spectroscopy is a method for characterizing properties of fluids and dispersed particles. It is also known as acoustic spectroscopy.
There is an international standard for this method.
Measurement of attenuation coefficient versus ultrasound frequency yields raw data for further calculation of various system properties. Such raw data are often used in the calculation of the particle size distribution in heterogeneous systems such as emulsions and colloids. In the case of acoustic rheometers, the raw data are converted into extensional viscosity or volume viscosity.
Instruments that employ ultrasound attenuation spectroscopy are referred to as Acoustic spectrometers. |
https://en.wikipedia.org/wiki/Separating%20set | In mathematics, a set of functions with domain is called a and is said to (or just ) if for any two distinct elements and of there exists a function such that
Separating sets can be used to formulate a version of the Stone–Weierstrass theorem for real-valued functions on a compact Hausdorff space with the topology of uniform convergence. It states that any subalgebra of this space of functions is dense if and only if it separates points. This is the version of the theorem originally proved by Marshall H. Stone.
Examples
The singleton set consisting of the identity function on separates the points of
If is a T1 normal topological space, then Urysohn's lemma states that the set of continuous functions on with real (or complex) values separates points on
If is a locally convex Hausdorff topological vector space over or then the Hahn–Banach separation theorem implies that continuous linear functionals on separate points.
See also |
https://en.wikipedia.org/wiki/Materials%20science%20in%20science%20fiction | Materials science in science fiction is the study of how materials science is portrayed in works of science fiction. The accuracy of the materials science portrayed spans a wide range – sometimes it is an extrapolation of existing technology, sometimes it is a physically realistic portrayal of a far-out technology, and sometimes it is simply a plot device that looks scientific, but has no basis in science. Examples are:
Realistic: In 1944, the science fiction story "Deadline" by Cleve Cartmill depicted the atomic bomb. The properties of various radioactive isotopes are critical to the proposed device, and the plot. This technology was real, unknown to the author.
Extrapolation: In the 1979 novel The Fountains of Paradise, Arthur C. Clarke wrote about space elevators - basically long cables extending from the Earth's surface to geosynchronous orbit. These require a material with enormous tensile strength and light weight. Carbon nanotubes are strong enough in theory, so the idea is plausible; while one cannot be built today, it violates no physical principles.
Plot device: An example of an unsupported plot device is scrith, the material used to construct Ringworld, in the novels by Larry Niven. Scrith has unreasonable strength, and is unsupported by known physics, but needed for the plot.
Critical analysis of materials science in science fiction falls into the same general categories. The predictive aspects are emphasized, for example, in the motto of the Georgia Tech's department of materials science and engineering – Materials scientists lead the way in turning yesterday's science fiction into tomorrow's reality. This is also the theme of many technical articles, such as Material By Design: Future Science or Science Fiction?, found in IEEE Spectrum, the flagship magazine of the Institute of Electrical and Electronics Engineers.
On the other hand, there is criticism of the unrealistic materials science used in science fiction. In the professional materials science |
https://en.wikipedia.org/wiki/Rule%20of%20least%20power | In programming, the rule of least power is a design principle that
"suggests choosing the least powerful [computer] language suitable for a given purpose". Stated alternatively, given a choice among computer languages, classes of which range from descriptive (or declarative) to procedural, the less procedural, more descriptive the language one chooses, the more one can do with the data stored in that language.
This rule is an application of the principle of least privilege to protocol design. The Rule of Least Power is an example in context of the centuries older principle known as Occam's razor in philosophy. In particular, arguments for and against the Rule of Least Power are subject to the same analysis as for Occam's razor.
Rationale
Originally proposed as an axiom of good design, the term is an extension of the KISS principle applied to choosing among a range of languages ranging from
the plainly descriptive ones (such as the content of most databases, or progressive enhancement on the web),
logical languages of limited propositional logic (such as access control lists),
declarative languages on the verge of being Turing-complete,
those that are in fact Turing-complete though one is led not to use them that way (XSLT, SQL),
those that are functional and Turing-complete general-purpose programming languages, to
those that are "unashamedly imperative".
As explained by Tim Berners-Lee:
Computer Science in the 1960s to 80s spent a lot of effort making languages that were as powerful as possible. Nowadays we have to appreciate the reasons for picking not the most powerful solution but the least powerful. The reason for this is that the less powerful the language, the more you can do with the data stored in that language. If you write it in a simple declarative form, anyone can write a program to analyze it in many ways. The Semantic Web is an attempt, largely, to map large quantities of existing data onto a common language so that the data can be analy |
https://en.wikipedia.org/wiki/Boundaries%20between%20the%20continents | Determining the boundaries between the continents is generally a matter of geographical convention. Several slightly different conventions are in use. The number of continents is most commonly considered seven (in English-speaking countries) but may range as low as four when Afro-Eurasia and the Americas are both considered as single continents. An island can be considered to be associated with a given continent by either lying on the continent's adjacent continental shelf (e.g. Singapore, the British Isles) or being a part of a microcontinent on the same principal tectonic plate (e.g. Madagascar and Seychelles). An island can also be entirely oceanic while still being associated with a continent by geology (e.g. Bermuda, the Australian Indian Ocean Territories) or by common geopolitical convention (e.g. Ascension Island, the South Sandwich Islands). Another example is the grouping into Oceania of the Pacific Islands with Australia and Zealandia.
There are three overland boundaries subject to definition:
between Africa and Asia (dividing Afro-Eurasia into Africa and Eurasia): at the Isthmus of Suez;
between Asia and Europe (dividing Eurasia): along the Turkish straits, the Caucasus, and the Urals and the Ural River (historically also north of the Caucasus, along the Kuma–Manych Depression or along the Don River);
between North America and South America (dividing the Americas): at some point on the Isthmus of Panama, with the most common demarcation in atlases and other sources following the Darién Mountains watershed along the Colombia–Panama border where the isthmus meets the South American continent (see Darién Gap).
While today the isthmus between Asia and Africa is navigable via the Suez Canal, and that between North and South America via the Panama Canal, these artificial channels are not generally accepted as continent-defining boundaries in themselves. The Suez Canal happens to traverse the isthmus between the Mediterranean Sea and the Red Sea, dividin |
https://en.wikipedia.org/wiki/Programmable%20metallization%20cell | The programmable metallization cell, or PMC, is a non-volatile computer memory developed at Arizona State University. PMC, a technology developed to replace the widely used flash memory, providing a combination of longer lifetimes, lower power, and better memory density. Infineon Technologies, who licensed the technology in 2004, refers to it as conductive-bridging RAM, or CBRAM. CBRAM became a registered trademark of Adesto Technologies in 2011. NEC has a variant called "Nanobridge" and Sony calls their version "electrolytic memory".
Description
PMC is a two terminal resistive memory technology developed at Arizona State University. PMC is an electrochemical metallization memory that relies on redox reactions to form and dissolve a conductive filament. The state of the device is determined by the resistance across the two terminals. The existence of a filament between the terminals produces a low resistance state (LRS) while the absence of a filament results in a high resistance state (HRS). A PMC device is made of two solid metal electrodes, one relatively inert (e.g., tungsten or nickel) the other electrochemically active (e.g., silver or copper), with a thin film of solid electrolyte between them.
Device operation
The resistance state of a PMC is controlled by the formation (programming) or dissolution (erasing) of a metallic conductive filament between the two terminals of the cell. A formed filament is a fractal tree like structure.
Filament formation
PMC rely on the formation of a metallic conductive filament to transition to a low resistance state (LRS). The filament is created by applying a positive voltage bias (V) to the anode contact (active metal) while grounding the cathode contact (inert metal). The positive bias oxidizes the active metal (M):
M → M+ + e−
The applied bias generates an electric field between the two metal contacts. The ionized (oxidized) metal ions migrate along the electric field toward the cathode contact. At the ca |
https://en.wikipedia.org/wiki/Gene%E2%80%93environment%20correlation | Gene–environment correlation (or genotype–environment correlation) is said to occur when exposure to environmental conditions depends on an individual's genotype.
Definition
Gene–environment correlations (or rGE) is correlation of two traits, e.g. height and weight, which would mean that when one changes, so does the other. Gene–environment correlations can arise by both causal and non-causal mechanisms. Of principal interest are those causal mechanisms which indicate genetic control over environmental exposure. Genetic variants influence environmental exposure indirectly via behavior. Three causal mechanisms giving rise to gene–environment correlations have been described.
(i) Passive gene–environment correlation refers to the association between the genotype a child inherits from their parents and the environment in which the child is raised. Parents create a home environment that is influenced by their own heritable characteristics. Biological parents also pass on genetic material to their children. When the children's genotype also influences their behavioral or cognitive outcomes, the result can be a spurious relationship between environment and outcome. For example, because parents who have histories of antisocial behavior (which is moderately heritable) are at elevated risk of abusing their children, a case can be made for saying that maltreatment may be a marker for genetic risk that parents transmit to children rather than a causal risk factor for children’s conduct problems.
(ii) Evocative (or reactive) gene–environment correlation happens when an individual's (heritable) behavior evokes an environmental response. For example, the association between marital conflict and depression may reflect the tensions that arise when engaging with a depressed spouse rather than a causal effect of marital conflict on risk for depression.
(iii) Active gene–environment correlation occurs when an individual possesses a heritable inclination to select environmental |
https://en.wikipedia.org/wiki/List%20of%20prehistoric%20foraminifera%20genera | This is a list of fossil genera of foraminiferans.
A
Abadehella
Abathomphalus
Abditodentrix
Abdullaevia
Abrardia
Abyssamina
Acarinina
Accordiella
Acervoschwagerina
Acervulina
Aciculella
Acostina
Acruliammina
Actinocyclina
Actinosiphon
Adelosina
Adelungia
Adercotryma
Adhaerentia
Adraheutina
Aeolisaccus
Aeolostreptis
Affinetrina
Afghanella
Afrobolivina
Agathammina
Agathamminoides
Agglutinella
Agglutisolena
Akiyoshiella
Aktinorbitoides
Alabamina
Alabaminoides
Aleomorphella
Alfredina
Alliatina
Alliatinella
Allomorphina
Allomorphinella
Almaena
Alpillina
Alpinophragmium
Altasterella
Altinerina
Altistoma
Alveocyclammina
Alveolina
Alveolinella
Alveolophragmium
Alveosepta
Alveovalvulina
Alzonella
Ambitropus
Amijiella
Ammoastuta
Ammobacularia
Ammobaculites
Ammobaculoides
Ammocycloculina
Ammodiscella
Ammodiscoides
Ammodiscus
Ammoelphidiella
Ammoglobigerina
Ammogloborotalia
Ammolagena
Ammomarginulina
Ammonia
Ammopalmula
Ammoscalaria
Ammosiphonia
Ammosphaeroidina
Ammospirata
Ammotium
Ammotrochoides
Ammovertella
Ammovertellina
Ammovolummina
Amphicervicis
Amphicoryna
Amphimorphina
Amphisorus
Amphistegina
Amphitremoida
Amphoratheca
Amplectoproductina
Anaticinella
Anchispirocyclina
Andamookia
Andersenia
Andersenolina
Andrejella
Androsina
Angotia
Angulodiscus
Angulogavelinella
Angulogerina
Anictosphaera
Annectina
Annulocibicides
Annulofrondicularia
Annulopatellina
Anomalinella
Anomalinoides
Antalyna
Antarcticella
Apertauroria
Apterrinella
Aragonella
Aragonia
Arakaevella
Archaealveolina
Archaecyclus
Archaediscus
Archaeglobigerina
Archaeochintinia
Archaeochitosa
Archaeosepta
Archaesphaera
Archaias
Archiacina
Arenagula
Areniconulus
Arenobulimina
Arenodosaria
Arenogaudryina
Arenonionella
Arenoparrella
Arenosiphon
Arenoturrispirillina
Arenovidalina
Armenina
Arnaudiella
Articularia
Articulina
Asanoina
Asanonella
Asanospira
Aschemocella
Ascoliella
Asselodiscus
Assilina
Astacolus
Asterigerina
Asterigerinata
Asterigerinella
Asterigerinoides
Asteroammonia
Asteroarchaediscus
Asterocyclina
Ast |
https://en.wikipedia.org/wiki/Pair-rule%20gene | A pair-rule gene is a type of gene involved in the development of the segmented embryos of insects. Pair-rule genes are expressed as a result of differing concentrations of gap gene proteins, which encode transcription factors controlling pair-rule gene expression. Pair-rule genes are defined by the effect of a mutation in that gene, which causes the loss of the normal developmental pattern in alternating segments.
Pair-rule genes were first described by Christiane Nüsslein-Volhard and Eric Wieschaus in 1980. They used a genetic screen to identify genes required for embryonic development in the fruit fly Drosophila melanogaster. In normal unmutated Drosophila, each segment produces bristles called denticles in a band arranged on the side of the segment closer to the head (the anterior). They found five genes – even-skipped, hairy, odd-skipped, paired and runt – where mutations caused the deletion of a particular region of every alternate segment. For example, in even-skipped, the denticle bands of alternate segments are missing, which results in an embryo having half the number of denticle bands. Later work identified more pair-rule genes in the Drosophila early embryo – fushi tarazu, odd-paired and sloppy paired.
Once the pair-rule genes had been identified at the molecular level it was found that each gene is expressed in alternate parasegments – regions in the embryo that are closely related to segments, but are slightly out of register. Each parasegment includes the posterior part of one (future) segment, and an anterior part of the next (more posterior) segment. The bands of expression of the pair-rule genes correspond to the regions missing in the mutant. The expression of the pair-rule genes in bands is dependent both upon direct regulation by the gap genes and on regulatory interactions between the pair-rule genes themselves.
See also
Drosophila embryogenesis
Gap gene |
https://en.wikipedia.org/wiki/Harris%20functional | In density functional theory (DFT), the Harris energy functional is a non-self-consistent approximation to the Kohn–Sham density functional theory. It gives the energy of a combined system as a function of the electronic densities of the isolated parts. The energy of the Harris functional varies much less than the energy of the Kohn–Sham functional as the density moves away from the converged density.
Background
Kohn–Sham equations are the one-electron equations that must be solved in a self-consistent fashion in order to find the ground state density of a system of interacting electrons:
The density, is given by that of the Slater determinant formed by the spin-orbitals of the occupied states:
where the coefficients are the occupation numbers given by the Fermi–Dirac distribution at the temperature of the system with the restriction , where is the total number of electrons. In the equation above, is the Hartree potential and is the exchange–correlation potential, which are expressed in terms of the electronic density. Formally, one must solve these equations self-consistently, for which the usual strategy is to pick an initial guess for the density, , substitute in the Kohn–Sham equation, extract a new density and iterate the process until convergence is obtained. When the final self-consistent density is reached, the energy of the system is expressed as:
.
Definition
Assume that we have an approximate electron density , which is different from the exact electron density . We construct exchange-correlation potential and the Hartree potential based on the approximate electron density . Kohn–Sham equations are then solved with the XC and Hartree potentials and eigenvalues are then obtained; that is, we perform one single iteration of the self-consistency calculation. The sum of eigenvalues is often called the band structure energy:
where loops over all occupied Kohn–Sham orbitals. The Harris energy functional is defined as
Comments
It was discover |
https://en.wikipedia.org/wiki/Buffer%20strip | A buffer strip is an area of land maintained in permanent vegetation that helps to control air quality, soil quality, and water quality, along with other environmental problems, dealing primarily on land that is used in agriculture. Buffer strips trap sediment, and enhance filtration of nutrients and pesticides by slowing down surface runoff that could enter the local surface waters. The root systems of the planted vegetation in these buffers hold soil particles together which alleviate the soil of wind erosion and stabilize stream banks providing protection against substantial erosion and landslides. Farmers can also use buffer strips to square up existing crop fields to provide safety for equipment while also farming more efficiently.
Buffer strips can have several different configurations of vegetation found on them varying from simply grass to combinations of grass, trees, and shrubs. Areas with diverse vegetation provide more protection from nutrient and pesticide flow and at the same time provide better biodiversity amongst plants and animals.
Many country, state, and local governments provide financial incentives for conservation programs such as buffer strips because they help stabilize the environment, help reduce nitrogen emissions to water and soil loss by wind erosion, while simultaneously providing substantial environmental co-benefits, even when the land is being used. Buffer strips not only stabilize the land but can also provide a visual demonstration that land is under stewardship.
Types
Buffers within fields
A grassed waterway reduces soil erosion and captures most nutrients and pesticides that would normally wash out of crop fields and into major waters. These waterways help to carry surface water at a non-erosive velocity to an area where it will have a stable outlet. Outlets must be adequate enough to allow water to drain without ponding or flooding the area being protected, while also preventing erosion of the water into the outlet which ca |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.