text
stringlengths
5
10.5k
source
stringlengths
33
146
to disintegrate to form boron with the emission of an alpha particle. Feather was therefore the first to show that neutrons produce nuclear disintegrations. In Rome, Enrico Fermi and his team bombarded heavier elements with neutrons and found the products to be radioactive. By 1934 they had used neutrons to induce radioactivity in 22 different elements, many of these elements of high atomic number. Noticing that other experiments with neutrons at his laboratory seemed to work better on a wooden table than a marble table, Fermi suspected that the protons of the wood were slowing the neutrons and so increasing the chance for the neutron to interact with nuclei. Fermi therefore passed neutrons through paraffin wax to slow them and found that the radioactivity of some bombarded elements increased by a factor of tens to hundreds. The cross section for interaction with nuclei is much larger for slow neutrons than for fast neutrons. In 1938 Fermi received the Nobel Prize in Physics "for his demonstrations of the existence of new radioactive elements produced by neutron irradiation, and for his related discovery of nuclear reactions brought about by slow neutrons". Later, Fermi recounted to Chandrasekhar that he was originally planning to put a piece of lead there, but an inexplicable, intuitive feeling made him put a paraffin in the spot instead. In Berlin, the collaboration of Lise Meitner and Otto Hahn, together with their assistant Fritz Strassmann, furthered the research begun by Fermi and his team when they bombarded uranium with neutrons. Between 1934 and 1938, Hahn, Meitner, and Strassmann found a great number of radioactive transmutation products from these experiments, all of which they regarded as transuranic. Transuranic nuclides are those that have an atomic number greater than uranium (92), formed by neutron absorption; such nuclides are not naturally occurring.
{"page_id": 46190717, "title": "Discovery of the neutron"}
Nurse crops are a subtype of nurse plants, facilitating the growth of other species of plants. The term is used primarily in agriculture, but also in forestry. Cover crops are a type of nurse crop. == Agriculture == In agriculture, a nurse crop is an annual crop used to assist in establishment of a perennial crop. The widest use of nurse crops is in the establishment of legumaceous plants such as alfalfa, clover, and trefoil. Occasionally, nurse crops are used for establishment of perennial grasses. Nurse crops reduce the incidence of weeds, prevent erosion, and prevent excessive sunlight from reaching tender seedlings. Often, the nurse crop can be harvested for grain, straw, hay, or pasture. Oats are the most common nurse crop, though other annual grains are also used. Nurse cropping of tall or dense-canopied plants can protect more vulnerable species through shading or by providing a wind break. However, if ill-maintained, nurse crops can block sunlight from reaching seedlings. Trap crops prevent pests from affecting the desired plant. == Forestry == In forestry, 'nurse crop' can be applied to trees or shrubs that help the development of other species of trees. Wind breaking, frost protection, thermal insulation, and shade can all be provided by nurse crops in forests. Aspens especially provide partial shade, allowing understory growth. == See also == Companion planting Multiple cropping == References ==
{"page_id": 242786, "title": "Nurse crop"}
might be considered a temporary, working definition, and can only be disproved by showing a logical contradiction. In contrast, a "descriptive" definition can be shown to be "right" or "wrong" with reference to general usage. Swartz defines a precising definition as one that extends the descriptive dictionary definition (lexical definition) for a specific purpose by including additional criteria. A precising definition narrows the set of things that meet the definition. C.L. Stevenson has identified persuasive definition as a form of stipulative definition which purports to state the "true" or "commonly accepted" meaning of a term, while in reality stipulating an altered use (perhaps as an argument for some specific belief). Stevenson has also noted that some definitions are "legal" or "coercive" – their object is to create or alter rights, duties, or crimes. === Recursive definitions === A recursive definition, sometimes also called an inductive definition, is one that defines a word in terms of itself, so to speak, albeit in a useful way. Normally this consists of three steps: At least one thing is stated to be a member of the set being defined; this is sometimes called a "base set". All things bearing a certain relation to other members of the set are also to count as members of the set. It is this step that makes the definition recursive. All other things are excluded from the set For instance, we could define a natural number as follows (after Peano): "0" is a natural number. Each natural number has a unique successor, such that: the successor of a natural number is also a natural number; distinct natural numbers have distinct successors; no natural number is succeeded by "0". Nothing else is a natural number. So "0" will have exactly one successor, which for convenience can be called "1".
{"page_id": 7964, "title": "Definition"}
1965 following on from research on bats conducted at the New York Zoological Society's field station in Trinidad and Tobago, he married Jocelyn Crane. A resident of Lexington, Massachusetts, since his 1986 departure from Rockefeller University, Griffin died at his home there at age 88 on November 7, 2003. He was survived by two daughters and a son from his first marriage. == Publications == Listening in the Dark (1958) Echoes of Bats and Men (1959) Anchor Books (Doubleday). LCCN 59--12051 Animal Structure and Function (1962) Holt, Rinehart and Winston Bird Migration: The Biology and Physics of Orientation Behaviour (1965) London: Heinemann Animal Structure and Function (1970, 2nd ed. co-authored with Alvin Novick) Holt, Rinehart and Winston Question of Animal Awareness (1976) ISBN 0-86576-002-0 Animal Thinking (1985) Harvard University Press. ISBN 0-674-03713-8 Animal Minds (1992) Animal Minds: Beyond Cognition to Consciousness (2001) ISBN 0-226-30865-0 "Windows on nonhuman minds," in Michel Weber and Anderson Weekes (eds.), Process Approaches to Consciousness in Psychology, Neuroscience, and Philosophy of Mind (Whitehead Psychology Nexus Studies II), Albany, New York, State University of New York Press, 2009, pp. 219 sq. [This is one of the last publications of D. R. Griffin, received in March 2002.] == References == == External links == National Academy of Sciences Biographical Memoir
{"page_id": 297801, "title": "Donald Griffin"}
of them were unaware of Johannes Fabricius' earlier observation and publication of sunspots. ==== Milky Way and stars ==== Galileo observed the Milky Way, previously believed to be nebulous, and found it to be a multitude of stars packed so densely that they appeared from Earth to be clouds. He located many other stars too distant to be visible to the naked eye. He observed the double star Mizar in Ursa Major in 1617. In the Starry Messenger, Galileo reported that stars appeared as mere blazes of light, essentially unaltered in appearance by the telescope, and contrasted them to planets, which the telescope revealed to be discs. But shortly thereafter, in his Letters on Sunspots, he reported that the telescope revealed the shapes of both stars and planets to be "quite round". From that point forward, he continued to report that telescopes showed the roundness of stars, and that stars seen through the telescope measured a few seconds of arc in diameter. He also devised a method for measuring the apparent size of a star without a telescope. As described in his Dialogue Concerning the Two Chief World Systems, his method was to hang a thin rope in his line of sight to the star and measure the maximum distance from which it would wholly obscure the star. From his measurements of this distance and of the width of the rope, he could calculate the angle subtended by the star at his viewing point. In his Dialogue, he reported that he had found the apparent diameter of a star of first magnitude to be no more than 5 arcseconds, and that of one of sixth magnitude to be about 5/6 arcseconds. Like most astronomers of his day, Galileo did not recognise that the apparent sizes of stars that he measured
{"page_id": 29688374, "title": "Galileo Galilei"}
In sedimentary geology and geomorphology, the term progradation refers to the growth of a river delta farther out into the sea over time. This occurs when the volume of incoming sediment is greater than the volume of the delta that is lost through subsidence, sea-level rise, or erosion. Progradation can be caused by: Periods of sea-level fall which result in marine regression. This can occur during major continental glaciations within ice ages, be caused by changes in the rates of seafloor spreading that affects the volume of the ocean basins, or tectonic effects on the regional mantle density structure that can change the geoid elevation. Extremely high sediment input, such as by the Huang He (Yellow River) in China, which drains the Loess plateau, or from high sediment loads in proglacial rivers. == See also == Retrogradation – Movement of the front of a river delta inland over time Aggradation – Increase in land elevation due to the deposition of sediment Marine transgression – Geologic event in which sea level rises relative to the land Sedimentology – Study of natural sediments and their formation processes Stratigraphy – Study of rock layers and their formation Sequence stratigraphy – Study and analysis of groups of sedimentary deposits Sediment transport – Movement of solid particles, typically by gravity and fluid entrainment == References ==
{"page_id": 20776529, "title": "Progradation"}
Trimethylolpropane triacrylate (TMPTA) is a trifunctional acrylate ester monomer derived from trimethylolpropane, used in the manufacture of plastics, adhesives, acrylic glue, anaerobic sealants, and ink. It is useful for its low volatility and fast cure response. It has the properties of weather, chemical and water resistance, as well as good abrasion resistance. End products include alkyd coatings, compact discs, hardwood floors, concrete and cementitious applications, Dental composites, photolithography, letterpress, screen printing, elastomers, automobile headlamps, acrylics and plastic components for the medical industry. == Other uses == As the molecule has acrylic functionality, it is capable of doing the Michael reaction with an amine. This allows its use in epoxy chemistry where its use speeds up the cure time considerably TMPTA can be used as a vulcanization booster for specialty rubbers such as ethylene propylene rubber, EPDM rubber, silicone rubber, polyurethane, ethylene/vinyl acetate (EVA) copolymer and chlorinated polyethylene elastomer (CPE). Typically, the vulcanization process is catalyzed with organic peroxides (such as DCP or BPO) however some rubbers are difficult to fully vulcanize using only peroxide catalysts. This usually results in longer curing times and lower yields, compromising mechanical and physical properties. Adding TMPTA has greatly improved the kinetics of vulcanization: for example, when using DCP for curing, if 1 to 4% TMPTA additive is used as a curing agent, not only the curing time is greatly reduced, but the overall chemical and mechanical properties are improved. == Safety == TMPTA's classification got worse overtime; in November 2023 all mixtures containing ≥ 1% by weight of TMPTA had been classified as carcinogen category 2 (H351). To overcome this problem, ethoxylated versions (EOTMPTA) were launched on the market, registred on ECHA with a range going from 1 to 6.5 moles of ethoxylation. == See also == Pentaerythritol tetraacrylate 1,6-Hexanediol diacrylate == References ==
{"page_id": 3028468, "title": "Trimethylolpropane triacrylate"}
communication standards also define error correction codes and framing formats. Popular modulation types being used for satellite communications: Binary phase-shift keying (BPSK); Quadrature phase-shift keying (QPSK); Offset quadrature phase-shift keying (OQPSK); 8PSK; Quadrature amplitude modulation (QAM), especially 16QAM. The popular satellite error correction codes include: Convolutional codes: with constraint length less than 10, usually decoded using a Viterbi algorithm (see Viterbi decoder); with constraint length more than 10, usually decoded using a Fano algorithm (see Sequential decoder); Reed–Solomon codes usually concatenated with convolutional codes with an interleaving; New modems support superior error correction codes (turbo codes and LDPC codes). Frame formats that are supported by various satellite modems include: Intelsat business service (IBS) framing Intermediate data rate (IDR) framing MPEG-2 transport framing (used in DVB) E1 and T1 framing High-end modems also incorporate some additional features: Multiple data interfaces (like RS-232, RS-422, V.35, G.703, LVDS, Ethernet); Embedded Distant-end Monitor and Control (EDMAC), allowing to control the distant-end modem; Automatic Uplink Power Control (AUPC), that is, adjusting the output power to maintain a constant signal to noise ratio at the remote end; Drop and insert feature for a multiplexed stream, allowing to replace some channels in it. == Internal structure == Probably the best way of understanding how a modem works is to look at its internal structure. A block diagram of a generic satellite modem is shown on the image. === Analog tract === After a digital-to-analog conversion in the transmitter, the signal passes through a reconstruction filter. Then, if needed, frequency conversion is performed. The purpose of the analog tract in the receiver is to convert signal's frequency, to adjust its power via an automatic gain control circuit and to get its complex envelope components. The input signal for the analog tract is at the intermediate frequency, sometimes, in
{"page_id": 183241, "title": "Satellite modem"}
ViolaWWW and WorldWideWeb, made consistent site appearance difficult, and users had less control over how web content was displayed. The browser/editor developed by Tim Berners-Lee had style sheets that were hard-coded into the program. The style sheets could therefore not be linked to documents on the web. Robert Cailliau, also of CERN, wanted to separate the structure from the presentation so that different style sheets could describe different presentation for printing, screen-based presentations, and editors. Improving web presentation capabilities was a topic of interest to many in the web community and nine different style sheet languages were proposed on the www-style mailing list. Of these nine proposals, two were especially influential on what became CSS: Cascading HTML Style Sheets and Stream-based Style Sheet Proposal (SSP). Two browsers served as testbeds for the initial proposals; Lie worked with Yves Lafon to implement CSS in Dave Raggett's Arena browser. Bert Bos implemented his own SSP proposal in the Argo browser. Thereafter, Lie and Bos worked together to develop the CSS standard (the 'H' was removed from the name because these style sheets could also be applied to other markup languages besides HTML). Lie's proposal was presented at the "Mosaic and the Web" conference (later called WWW2) in Chicago, Illinois in 1994, and again with Bert Bos in 1995. Around this time the W3C was already being established and took an interest in the development of CSS. It organized a workshop toward that end chaired by Steven Pemberton. This resulted in W3C adding work on CSS to the deliverables of the HTML editorial review board (ERB). Lie and Bos were the primary technical staff on this aspect of the project, with additional members, including Thomas Reardon of Microsoft, participating as well. In August 1996, Netscape Communication Corporation presented an alternative style sheet language
{"page_id": 23290197, "title": "CSS"}
half a million dollars for RKO. John Ford's The Fugitive (1947) and Fort Apache (1948), which appeared right before studio ownership changed hands again, were followed by She Wore a Yellow Ribbon (1949) and Wagon Master (1950); all four were coproductions between RKO and Argosy, the company run by Ford and RKO alumnus Merian C. Cooper. Of the directors under long-term contract to RKO in the 1940s, the best known was Edward Dmytryk, who first came to notice with the remarkably profitable Hitler's Children (1943). Shot on a $205,000 budget, placing it in the bottom quartile of Big Five studio productions, it was one of the ten biggest Hollywood hits of the year.[k] Another low-cost war-themed film directed by Dmytryk, Behind the Rising Sun, released a few months later, was similarly profitable. === Focus on B movies === Much more than the other Big Five studios, RKO relied on B pictures to fill up its schedule. Of the thirty-one features released by RKO in 1944, for instance, ten were budgeted below $200,000, twelve were in the $200,000 to $500,000 range, and only nine cost more. In contrast, a clear majority of the features put out by the other top four studios were budgeted at over half a million dollars. A focus on B pictures limited the studio's financial risk; while it also limited the potential for reward (Dmytryk's extraordinary coups aside), RKO had a history of making better profits with its run-of-the-mill and low-cost product than with its A movies. The studio's low-budget films offered training opportunities for new directors, as well, among them Mark Robson, Robert Wise, and Anthony Mann. Film editors Robson and Wise received their first directing assignments with producer Val Lewton, whose specialized B horror unit also included the more experienced director Jacques Tourneur. The Lewton
{"page_id": 36437899, "title": "RKO Pictures"}
by **varname ,etc. Thus, for a given expression expr , 〈expr 〉 might contain variables of negative indirection level. How they are interpreted depends on whether expr represents a memory location to write to, or to read from: • If expr should be read from, then the result might depend on the values of the variables with negative indirection level. Therefore, in order to determine the read 32 CHAPTER 2. STANALYZER set for expr , variables of negative indirection level have their indirection level set to 0. The resulting set is denoted by 〈expr 〉R. • Conversely, when the memory location for a write operation should be computed, variables of negative indirection level should be removed . The resulting set is denoted by 〈expr 〉W .Applying these rules to the previous example where the expression was arr[index] , we get that 〈arr[index] 〉 = { arr , *index }. Therefore, 〈arr[index] 〉R = { arr , index } and 〈arr[index] 〉W = { arr }. ## 2.3.4 Instruction Interpretation The main algorithm visits each node of the AST and updates the dependency graph and set of leaking instructions according to the semantics defined in this section. Indirect dependencies are handled as explained in Section 2.2.4 and ignored in the description of the semantics for the sake of clarity. The pair comprised of the modified dependency graph and set of leakages, after analyzing the instruction inst , is denoted by φ(G, L ; inst ) where G and L are respectively the initial dependency graph and leakages. The dependency graph of this pair will be denoted by φG(G, L ; inst ) and the set of leakages by φL(G, L ; inst ). Analyzing simple instructions The modifications on the dependency graph G and leaking values L when analyzing simple instructions
{"source": 2294, "title": "from dpo"}
unique? Show that general position is a sufficient condition. Is it also necessary? 27. What can you say about the “quality” of a Delaunay triangulation? Prove that every Delaunay triangulation maximizes the smallest interior angle in the triangulation, among the set of all triangulations of the same point set. 77 Chapter 7 Delaunay Triangulation: Incremental Construction In the last lecture, we have learned about the Lawson flip algorithm that computes a Delaunay triangulation of a given n-point set P ⊆ R2 with O(n2) Lawson flips. One can actually implement this algorithm to run in O(n2) time, and there are point sets where it may take Ω(n2) flips. In this lecture, we will discuss a different algorithm. The final goal is to show that this algorithm can be implemented to run in O(n log n) time; this lecture, however, is concerned only with the correctness of the algorithm. Throughout the lecture we assume that P is in general position (no 3 points on a line, no 4 points on a common circle), so that the Delaunay triangulation is unique (Corollary 6.18). There are techniques to deal with non-general position, but we don’t discuss them here. 7.1 Incremental construction The idea is to build the Delaunay triangulation of P by inserting one point after another. We always maintain the Delaunay triangulation of the point set R inserted so far, and when the next point s comes along, we simply update the triangulation to the Delaunay triangulation of R ∪ {s}. Let DT (R) denote the Delaunay triangulation of R ⊆ P.To avoid special cases, we enhance the point set P with three artificial points “far out”. The convex hull of the resulting point set is a triangle; later, we can simply remove the extra points and their incident edges to obtain DT
{"source": 4191, "title": "from dpo"}
line, 3 planes compose to a point Vectors, Complex planes & lines, Dot and cross products of vectors **Read:** Lec 14 15 6 39 M 9/26 _Lecture:_ Gaussian elimination (intersection); Pivot matrices {$(\Pi_n)$}: {$U = \Pi_n^N P_n A$} gives upper-diagional {$U$} **Read:** Lec 15 **Homework 5 (AE-2):** Non-linear and linear systems of equations; Gaussian elimination; pdf; (the _extended plane_) (JS10, p. 298-312) Mobius Transformation (youtube drawing. **Read:** Lec 18; **AE-3:**pdf]( Notes on the Fourier series and transform from ECE 310 20 F 10/7 _Lecture:_ _Laplace transforms_ for systems The importance of _Causality_ Cauchy Riemann role in the acceptance of complex functions: Convolution of the step function: {$u(t) \leftrightarrow 1/s$} vs. {$2\tilde{u}(t) \equiv 1+ \mbox{sgn}(t) \leftrightarrow 2\pi \delta(\omega) + 2/j\omega$} **Read:** Lec 20; Laplace Transform pdf \leftrightarrow Z(s)$} A.E. Kennelly introduces complex impedance, 1893 pdf$} vs. {$u(t)$} **AE-3 Due** 22 W 10/12 No class
{"source": 6061, "title": "from dpo"}
and of boron on Mars are of interest as they are precursors for prebiotic chemistry. Such findings, along with previous discoveries that liquid water was clearly present on ancient Mars, further supports the possible early habitability of Gale Crater on Mars. Currently, the surface of Mars is bathed with ionizing radiation, and Martian soil is rich in perchlorates toxic to microorganisms. Therefore, the consensus is that if life exists—or existed—on Mars, it could be found or is best preserved in the subsurface, away from present-day harsh surface processes. In June 2018, NASA announced the detection of seasonal variation of methane levels on Mars. Methane could be produced by microorganisms or by geological means. The European ExoMars Trace Gas Orbiter started mapping the atmospheric methane in April 2018, and the 2022 ExoMars rover Rosalind Franklin was planned to drill and analyze subsurface samples before the programme's indefinite suspension, while the NASA Mars 2020 rover Perseverance, having landed successfully, will cache dozens of drill samples for their potential transport to Earth laboratories in the late 2020s or 2030s. As of February 8, 2021, an updated status of studies considering the possible detection of lifeforms on Venus (via phosphine) and Mars (via methane) was reported. In October 2024, NASA announced that it may be possible for photosynthesis to occur within dusty water ice exposed in the mid-latitude regions of Mars. == Early speculation == Mars's polar ice caps were discovered in the mid-17th century. In the late 18th century, William Herschel proved they grow and shrink alternately, in the summer and winter of each hemisphere. By the mid-19th century, astronomers knew that Mars had certain other similarities to Earth, for example that the length of a day on Mars was almost the same as a day on Earth. They also knew that its
{"page_id": 463835, "title": "Life on Mars"}
state. Other analyses indicate a spectrum or continuum of activity states. The nuclear and signaling landscapes of complex eukaryotic nuclei may favour more than two simple states- for example, there are over several dozen post-translational modifications of nucleosomes and perhaps a hundred different proteins involved in the average eukaryotic transcription reaction. What do the repressive and permissive states represent? An attractive idea is that the repressed state is a closed chromatin conformation whilst the permissive states are more open. Another hypothesis is that the fluctuations between states reflect reversible transitions in the binding and dissociation of pre-initiation complexes. Bursts may also result from bursty signalling, cell cycle effects or movement of chromatin to and from transcription factories. Bursting dynamics have been demonstrated to be influenced by cell size and the frequency of extracellular signalling. Recent data suggest different degrees of supercoiling distinguish the permissive and inactive states. The bursting phenomenon, as opposed to simple probabilistic models of transcription, can account for the high variability (see transcriptional noise) in gene expression occurring between cells in isogenic populations. This variability in turn can have tremendous consequences on cell behaviour, and must be mitigated or integrated. Suggested mechanisms by which noise can be dampened include strong extracellular signalling, diffusion of RNA and protein in cell syncitia, promoter proximal pausing, and nuclear retention of transcripts. In certain contexts, such as the survival of microbes in rapidly changing stressful environments, the expression variability may be essential. Variability also impacts upon the effectiveness of clinical treatment, with resistance of bacteria to antibiotics demonstrably caused by non-genetic differences. Similar phenomena may contribute to the resistance of sub-populations of cancer cells to chemotherapy. Spontaneous variability in gene expression is also proposed to act as a source of cell fate diversity in self-organizing differentiation processes, and may act as
{"page_id": 24277379, "title": "Transcriptional bursting"}
Compton scattering exists, Compton scattering usually refers to the interaction involving only the electrons of an atom. The Compton effect was observed by Arthur Holly Compton in 1923 at Washington University in St. Louis and further verified by his graduate student Y. H. Woo in the years following. Compton was awarded the 1927 Nobel Prize in Physics for the discovery. The effect is significant because it demonstrates that light cannot be explained purely as a wave phenomenon. Thomson scattering, the classical theory of an electromagnetic wave scattered by charged particles, cannot explain shifts in wavelength at low intensity: classically, light of sufficient intensity for the electric field to accelerate a charged particle to a relativistic speed will cause radiation-pressure recoil and an associated Doppler shift of the scattered light, but the effect would become arbitrarily small at sufficiently low light intensities regardless of wavelength. Thus, if we are to explain low-intensity Compton scattering, light must behave as if it consists of particles. Or the assumption that the electron can be treated as free is invalid resulting in the effectively infinite electron mass equal to the nuclear mass (see e.g. the comment below on elastic scattering of X-rays being from that effect). Compton's experiment convinced physicists that light can be treated as a stream of particle-like objects (quanta called photons), whose energy is proportional to the light wave's frequency. As shown in Fig. 2, the interaction between an electron and a photon results in the electron being given part of the energy (making it recoil), and a photon of the remaining energy being emitted in a different direction from the original, so that the overall momentum of the system is also conserved. If the scattered photon still has enough energy, the process may be repeated. In this scenario, the electron is
{"page_id": 55236, "title": "Compton scattering"}
+ ( 1 + ( 4ε / ( ρ2V2/ρ1V1 )1/2 ) ) ( S cos α2 - tTE )/h ] where ζ=total losses, ζp=blade profile losses, ζew=endwall losses. The expression for endwall losses in an axial-flow compressor is given by: η = ή ( 1 - ( δh* + δt*)/h ) / ( 1 - ( Fθh + Fθt ) / h ) where η=efficiency in absence of endwall boundary layer, where h refers to the hub and t refers to the tip. The values of Fθ and δ* are derived from the graph or chart. == Tip-leakage flow losses == The main points to consider are: The rotation of a rotor in turbomachinery induces a pressure differences between opposite sides of the blade profile, resulting in tip leakage. In a turbomachinery rotor, a gap between the annulus wall and the blade causes leakage, which also occurs in the gap between the rotating hub and stator. Direct loss through clearance volume, as no angular momentum is transferred to fluid. So, no work is done. Leakage, and its interaction with other losses in the flow field, is complex; and hence, at the tip, it has a more pronounced effect than secondary flow. Leakage-flow induced three-dimensionality, like the mixing of leakage flow with vortex formation, entrainment process, diffusion and convection. This results in aerodynamics losses and inefficiency. Tip leakage and clearance loss account for 20–40% of total losses. The effects of cooling in turbines causes vibration, noise, flutter, and high blade stress. Leakage flow causes low static pressure in the core area, increasing the risk of cavitation and blade damage. The leakage velocity is given as: QL = 2 ( ( Pp - Ps ) / ρ )1/2 The leakage flow sheet due to velocity induced by the vortex is given in
{"page_id": 39493834, "title": "Three-dimensional losses and correlation in turbomachinery"}
The Kyaukkyan Fault Zone is a large complex strike-slip fault that extends for about 510 km from Shan state, Myanmar to Thailand. It was the source of the 1912 Shan state earthquake when it ruptured for a length of 160 km along the northernmost segment. The fault is not very well studied, unlike the Sagaing Fault. The fault runs through the Shan hills nearly parallel to the Sagaing Fault. It is highly segmented, characterized by a broad array of splaying segments and basins, dominated by releasing bends and associated extensional fault systems. The slip rate for this fault is about 1 mm/yr. == Segments == === Myint Nge segment === The Kyaukkyan Fault Zone's northernmost segment runs 160 km just north of Taunggyi and terminates east of Mandalay. It gets its name from the Myint Nge River that is offset by the fault. In 1912, the entire Myint Nge segment ruptured and produced an earthquake of magnitude 7.7. === Taunggyi segment === This segment is a transtensional basin 50 km wide and 100 km long. It is bordered by the Taunggyi and Pindaya Faults, both normal faults. High fault scarps can be observed; the Pindaya Fault scarp faces east and is about 350 m high while the Taunggyi escarpment reaches 400 m. At Inle Lake, motion of the Taunggyi Fault has resulted in its eastern side is deeper than its west, associated with block tilting. === Salween segment === The Salween segment gets its namesake from the Salween River that is also offset by the fault by about 5.4 km. The fault makes a bend, like the "Big Bend" of the San Andreas Fault, creating a transtensional basin along the river. At this segment, the Mae Ping Fault connects with the Kayukkyan Fault. The Kyaukkan Fault continue its trace for
{"page_id": 65581604, "title": "Kyaukkyan Fault"}
Al Roth's game theory, experimental economics, and market design page. Video - Alvin E. Roth (2014) : Repugnant Markets and Prohibited Transactions at the Lindau Nobel Laureate Meeting, August 20, 2014 Appearances on C-SPAN Alvin E. Roth on Nobelprize.org
{"page_id": 18166140, "title": "Alvin E. Roth"}
comprise each beta-sandwich are joined by an alpha helix on one side. These alpha helices define the binding site for this protein, with the binding site incorporating one antiparallel beta strand on either side of the helix. In addition to the binding sites, the opposite side of the beta-sandwich is connected by a series of loops, which define a hypervariable loop system, that may have a role in determining the specificity of an interaction between IgA and an antigen. == Pathological Information == IGHA1 has been implicated in a chromosomal abnormality identified in multiple myeloma lines. The abnormality has been identified as a translocation event, where translocation between IGHA1 (found on chromosome 14), and FCRL4 (the gene sequence encoding for an inhibitory receptor, found on chromosome 1) leads to the production of a fusion protein. == References == == Further reading ==
{"page_id": 14797041, "title": "Immunoglobulin heavy constant alpha 1"}
2015–2019 Capital Program, as well as the entire line between Church Avenue and West Eighth Street–New York Aquarium, with three interlockings to be upgraded on that stretch. As of 2014, MTA projects that 355 miles (571 km) of track will receive CBTC signals by 2029, including most of the IND, as well as the IRT Lexington Avenue Line and the BMT Broadway Line. The MTA also is planning to install CBTC equipment on the IND Crosstown Line, the BMT Fourth Avenue Line and the BMT Brighton Line before 2025. Additionally, the New York City Subway uses a system known as Automatic Train Supervision (ATS) for dispatching and train routing on the A Division (the Flushing line, and the trains used on the 7 and <7>​ services, do not have ATS.) ATS allows dispatchers in the Operations Control Center (OCC) to see where trains are in real time, and whether each individual train is running early or late. Dispatchers can hold trains for connections, re-route trains, or short-turn trains to provide better service when a disruption causes delays. === Other ideas === In 2017, the MTA started testing ultra-wideband radio-enabled train signaling on the IND Culver Line. The ultra-wideband train signals would be able to carry more data wirelessly in a manner similar to CBTC, but can be installed faster. The ultra-wideband signals would have the added benefit of allowing passengers to use cellphones while between stations, instead of the current setup (see Technology of the New York City Subway § Cellular phone and wireless data) that only provides cellphone signals within the stations. == Platform screen doors == The MTA has long been reluctant to install platform screen doors in the subway system, though it had been considering such an idea since the 1980s. Originally, it was planned to install
{"page_id": 55208663, "title": "Technology of the New York City Subway"}
Strikes, lawsuits, and wars could arise from overplacement. If plaintiffs and defendants were prone to believe that they were more deserving, fair, and righteous than their legal opponents, that could help account for the persistence of inefficient enduring legal disputes. If corporations and unions were prone to believe that they were stronger and more justified than the other side, that could contribute to their willingness to endure labor strikes. If nations were prone to believe that their militaries were stronger than were those of other nations, that could explain their willingness to go to war. Overprecision could have important implications for investing behavior and stock market trading. Because Bayesians cannot agree to disagree, classical finance theory has trouble explaining why, if stock market traders are fully rational Bayesians, there is so much trading in the stock market. Overprecision might be one answer. If market actors are too sure their estimates of an asset's value is correct, they will be too willing to trade with others who have different information than they do. Oskamp tested groups of clinical psychologists and psychology students on a multiple-choice task in which they drew conclusions from a case study. Along with their answers, subjects gave a confidence rating in the form of a percentage likelihood of being correct. This allowed confidence to be compared against accuracy. As the subjects were given more information about the case study, their confidence increased from 33% to 53%. However their accuracy did not significantly improve, staying under 30%. Hence this experiment demonstrated overconfidence which increased as the subjects had more information to base their judgment on. Even if there is no general tendency toward overconfidence, social dynamics and adverse selection could conceivably promote it. For instance, those most likely to have the courage to start a new business are
{"page_id": 3080323, "title": "Overconfidence effect"}
supply chains will be disrupted in case of an AI Cold War. This prompted US technology companies to develop mitigation strategies including hoarding semiconductors and trying to set up local semiconductor production facilities, with the support of government subsidies. === Industrial policy initiatives === ==== United States ==== In June 2021, the US Senate approved the U.S. Innovation and Competition Act providing around 250 billion US dollars public money support to the US technological and manufacturing industry. The alleged Chinese threat in the area of technology helped secure a strong bipartisan support for the new legislation, amounting to the largest industrial policy move by the US in decades. Chinese authorities reproached to the US that the bill was “full of cold war zero-sum thinking”. The legislative bill is aimed at strengthening capabilities in the area of technology, such as quantum computing and AI specifically to face the competitive threat from China perceived as urgent. Senator Chuck Schumer, the leader of the Senate majority and one of the sponsors of the industrial policy bill invoked the threat of authoritarian regimes that want “grab the mantle of global economic leadership and own the innovations”. In 2022, U.S. Innovation and Competition Act was amended and turned into the Chips and Science Act with planned spending of 280 billion US dollars, 53 billion thereof are allocated directly to subsidies for semiconductors manufacturing. Commentators identified possible positive effects on innovation from the US attempts to compete with China in a perceived rivalry. Among the main beneficiaries of the US CHIPS Act are the semiconductor producers Intel, TSMC and Micron Technology. ==== European Chips Act ==== In February 2022, the European Union introduced its own European Chips Act initiative. The background of the initiative would be the objective of European strategic autonomy. The EU's initiative puts
{"page_id": 68130381, "title": "Artificial Intelligence Cold War"}
explicit lyrics. As intrigued as I was by Loveless’s voice, I was more impressed with the band as a whole. Lamb’s upright bass playing, May’s harmonies and guitar solos, and her fathers’ obvious fun on the drums earned my admiration. I ordered the LP Somewhere Else and also received a free digital download with the bonus track “Come Over” and soon found myself playing the album nonstop. Loveless writes songs with a sense of humor and literary knowledge. You can hear outlaw country in her music but there’s also a touch of punk. She’s sort of a modern day female David Allen Coe who sings about mistakes unapologetically. “Somewhere Else,” the title track from Loveless’s album, is all about indecision as she apologizes for what she’s done. While the tune includes explicit lyrics, like many of her songs, there’s also a sense of humor when she sings “I just wanted to 867-5309 you” as the steel guitar and May’s Telecaster play along. On “Wine Lips” Loveless is more decisive, knowing exactly where she wants to be and not apologizing for the “bad idea” she has for “wine lips.” Catchy and upbeat, the song is fun on its own but when you hear Loveless sing “tell your momma that my French has finally improved” your grin only grows bigger. Loveless shares her literary bent with “Verlaine Shot Rimbaud.” French poet Arthur Rimbaud’s affair wit Paul Verlaine turns tragic but Loveless sees their violent passion as the fuel for Rimbaud’s poetry. “I just like you so much better when we’re coming to blows” she sings as she tells her lover “I just wanna be the one you love.” There’s plenty of heartache, dysfunction, bad decisions and loss throughout most of Loveless’s songs but my favorite, “Everything’s Gone,” deals not with the loss
{"source": 990, "title": "from dpo"}
in classification tasks, especially in scenarios with a high cost of false positives. It provides insights into the model's ability to correctly predict positive instances while minimizing the risk of false alarms. Precision, often referred to as the positive predictive value, quantifies the proportion of true positive predictions among all positive predictions made by the model. It answers the question: _"Of all the instances predicted as positive, how many were positive?"_ ### Mathematical Formula !Image 3: Precision - Mathematical Formula | Encord can lead to unnecessary investigations, customer dissatisfaction, and potential loss of business. Here, high precision ensures that most flagged transactions are indeed fraudulent, minimizing the number of false alarms. ### Limitations Precision focuses solely on the correctly predicted positive cases, neglecting the false negatives. As a result, a model can achieve high precision by making very few positive predictions, potentially missing out on many actual positive cases. This narrow focus can be misleading, especially when false negatives have significant consequences. What is Recall? --------------- Recall, also known as sensitivity or [true positive rate]( is a crucial metric in classification that emphasizes the model's ability to identify all relevant instances. Recall measures the proportion of actual positive cases correctly identified by the model. It answers the question: "Of all the actual positive instances, how many were correctly predicted by the model?" Mathematical Formula: ![Image 4: Recall - Mathematical Formula | Encord]( Where: * TP = True Positives * FN = False Negatives ### Significance _Recall is
{"source": 978, "title": "from dpo"}
two-year revision cycle since 1986 when the first edition was pub-lished. They are among the best if not the best-selling strategic-management textbooks in the world and have been used at more than 500 colleges and universities, including Harvard University, Duke University, Carnegie-Mellon University, Johns Hopkins University, the University of Maryland, University of North Carolina, University of Georgia, San Francisco State University, University of South Carolina, and Wake Forest University. This textbook has been translated and published in Chinese, Japanese, Farsi, Spanish, Indonesian, Indian, Thai, and Arabic and is widely used across Asia and South America. It is the best-selling strategic-management textbook in Mexico, China, Peru, Chile, Japan, and number two in the United States. Approximately 90,000 students read Dr. David’s textbook annually as well as thousands of businesspersons. The book has led the field of strategic manage-ment for more than a decade in providing an applications/practitioner approach to the discipline. A native of Whiteville, North Carolina, Fred David received a BS degree in mathe-matics and an MBA from Wake Forest University before being employed as a bank manager with United Carolina Bank. He received a PhD in Business Administration from the University of South Carolina, where he majored in management. Currently the TranSouth Professor of Strategic Management at Francis Marion University (FMU) in Florence, South Carolina, Dr. David has also taught at Auburn University, Mississippi State University, East Carolina University, the University of South Carolina, and the University of North Carolina at Pembroke. He is the author of 152 referred publications, including 40 journal articles and 55 proceedings publications. David has articles pub-lished in such journals as Academy of Management Review , Academy of Management Executive , Journal of Applied Psychology , Long Range Planning , and Advanced Management Journal .Dr. David received a Lifetime Honorary Professorship Award from the
{"source": 4976, "title": "from dpo"}
to move this point to the origin by the change of variables x=π+u,y=v. Then the differential equations (2) become > du /dt =v,dv/ dt =gLsin u,(11) and the critical point is (0, 0 )in the uv-plane. Consider the function > V(u,v) =vsin u(12) and let Dbe the domain −π/ 4 ˙V=(v cos u)(v) +(sin u)[(g/L)sin u] = v2cos u+(g/L)sin 2u(13) is positive definite in D. The only remaining question is whether there are points in every neighborhood of the origin where Vitself is positive. From Eq. (12) we see that V(u,v) > 0in the first quadrant (where both sin uand vare positive) and in the third quadrant (where both are negative). Thus the conditions of Theorem 9.6.2 are satisfied, and the point (0, 0 )in the uv-plane, corresponding to the point (π , 0 )in the xy -plane, is unstable. The damped pendulum equations are discussed in Problem 7. From a practical point of view, we are often interested in the basin of attraction. The following theorem provides some information on this subject. # Theorem 9.6.3 Let the origin be an isolated critical point of the autonomous system (6). Let the function V be continuous and have continuous first partial derivatives. If there is a bounded domain DK containing the origin where V(x, y) < K for some positive K, V is positive definite, and ˙V is negative definite, then every solution of Eqs. (6) that starts at a point in DK approaches the origin as t approaches infinity. In other words, Theorem 9.6.3 says that if x = φ( t), y = ψ( t) is the solution of Eqs. (6) for initial data lying in DK , then (x, y) approaches the critical point (0, 0) as t → ∞ . Thus DK gives a region of asymptotic stability;
{"source": 6286, "title": "from dpo"}
would then have collapsed to form the planets. As captured planets would have initially eccentric orbits, Dormand and Woolfson proposed the possibility of a collision. They hypothesized that a filament was thrown out by a passing protostar and was captured by the Sun, resulting in the formation of planets. In this idea, there were 6 original planets, corresponding to 6 point-masses in the filament, with planets A and B, the two innermost, colliding. A, at 200 Earth masses, was ejected out of the Solar System, while B, estimated to be 25 Earth masses, split into Earth and Venus. Mercury is also a fragment of B. Mars and the Moon are escaped moons of A. Pluto is either a fragment from the collision or a former moon of Neptune sent into a heliocentric orbit by Triton, another of A's former satellites, which formed Charon from tidal disruption of Pluto. The collision also formed the asteroid belt and Oort's cloud. ==== Classification of the hypotheses ==== Jeans, in 1931, divided the various models into two groups: those where the material for planet formation came from the Sun, and those where it did not and may be concurrent or consecutive. In 1963, William McCrea divided them into another two groups: those that relate the formation of the planets to the formation of the Sun and those where it is independent of the formation of the Sun, where the planets form after the Sun becomes a normal star. Ter Haar and Cameron distinguished between those hypotheses that consider a closed system, which is a development of the Sun and possibly a solar envelope, that starts with a protosun rather than the Sun itself, and state that Belot calls these hypotheses monistic; and those that consider an open system, which is where there is an
{"page_id": 17052696, "title": "History of Solar System formation and evolution hypotheses"}
2 π ϵ 0 a 3 {\displaystyle {\frac {1}{2}}m_{e}\omega ^{2}={\frac {e^{2}}{\pi \epsilon _{0}a^{3}}}} or, combining this with the result from the quantum harmonic oscillator for the root-mean-square displacement ⟨ r 2 ⟩ a = 3 8 ( 1 r s ) 1 / 4 {\displaystyle {\frac {\sqrt {\langle r^{2}\rangle }}{a}}={\sqrt {\frac {3}{8}}}\left({\frac {1}{r_{s}}}\right)^{1/4}} The Lindemann criterion than gives us the estimate that rs > 40 is required to give a stable Wigner crystal. Quantum Monte Carlo simulations indicate that the uniform electron gas actually crystallizes at rs = 106 in 3D and rs = 31 in 2D. For classical systems at elevated temperatures one uses the average interparticle interaction in units of the temperature: G = e 2 / k B T a {\displaystyle G=e^{2}/k_{B}Ta} .. The Wigner transition occurs at G = 170 in 3D and G = 125 in 2D. It is believed that ions, such as those of iron, form a Wigner crystal in the interiors of white dwarf stars. == Experimental realisation == In practice, it is difficult to experimentally realize a Wigner crystal because quantum mechanical fluctuations overpower the Coulomb repulsion and quickly cause disorder. Low electron density is needed. One notable example occurs in quantum dots with low electron densities or high magnetic fields where electrons will spontaneously localize in some situations, forming a so-called rotating "Wigner molecule", a crystalline-like state adapted to the finite size of the quantum dot. Wigner crystallization in a two-dimensional electron gas under high magnetic fields was predicted (and was observed experimentally) to occur for small filling factors (less than ν = 1 / 5 {\displaystyle \nu =1/5} ) of the lowest Landau level. For larger fractional fillings, the Wigner crystal was thought to be unstable relative to the fractional quantum Hall effect (FQHE) liquid states. A Wigner crystal
{"page_id": 3707854, "title": "Wigner crystal"}
asset or a drawback. == Non-ideal CSTR == While the ideal CSTR model is useful for predicting the fate of constituents during a chemical or biological process, CSTRs rarely exhibit ideal behavior in reality. More commonly, the reactor hydraulics do not behave ideally or the system conditions do not obey the initial assumptions. Perfect mixing is a theoretical concept that is not achievable in practice. For engineering purposes, however, if the residence time is 5–10 times the mixing time, the perfect mixing assumption generally holds true. Non-ideal hydraulic behavior is commonly classified by either dead space or short-circuiting. These phenomena occur when some fluid spends less time in the reactor than the theoretical residence time, τ {\displaystyle \tau } . The presence of corners or baffles in a reactor often results in some dead space where the fluid is poorly mixed. Similarly, a jet of fluid in the reactor can cause short-circuiting, in which a portion of the flow exits the reactor much quicker than the bulk fluid. If dead space or short-circuiting occur in a CSTR, the relevant chemical or biological reactions may not finish before the fluid exits the reactor. Any deviation from ideal flow will result in a residence time distribution different from the ideal distribution, as seen at right. === Modeling non-ideal flow === Although ideal flow reactors are seldom found in practice, they are useful tools for modeling non-ideal flow reactors. Any flow regime can be achieved by modeling a reactor as a combination of ideal CSTRs and plug flow reactors (PFRs) either in series or in parallel. For examples, an infinite series of ideal CSTRs is hydraulically equivalent to an ideal PFR. Reactor models combining a number of CSTRs in series are often termed tanks-in-series (TIS) models. To model systems that do not obey
{"page_id": 2292624, "title": "Continuous stirred-tank reactor"}
US Cyber Challenge is a private program which recruits, trains, and places candidates in cybersecurity jobs in the United States. US Cyber Challenge was formerly a DHS S&T-funded not-for-profit organization, and is currently a program of the Center for Internet Security. == History == The founding director of US Cyber Challenge was Karen S. Evans. == Methodology == The US Cyber Challenge uses an online competition each April to select qualified candidates for a one-week cybersecurity training program the following summer. The summer program includes workshops, a job fair, and a capture the flag competition. Topics covered in the workshops include intrusion detection, penetration testing and forensics. Participation is limited to high school and college students who are US citizens. == References == == External links == Official web site US Cyber Challenge on Twitter
{"page_id": 45718216, "title": "US Cyber Challenge"}
Sorting describes the distribution of grain size of sediments, either in unconsolidated deposits or in sedimentary rocks. The degree of sorting is determined by the range of grain sizes in a sediment deposit and is the result of various transport processes (rivers, debris flow, wind, glaciers, etc.). This should not be confused with crystallite size, which refers to the individual size of a crystal in a solid. Crystallite is the building block of a grain. == Sorting parameters == The terms describing sorting in sediments – very poorly sorted, poorly sorted, moderately sorted, well sorted, very well sorted – have technical definitions and semi-quantitatively describe the amount of variance seen in particle sizes.Very poorly sorted indicates that the sediment sizes are mixed (large variance); whereas well sorted indicates that the sediment sizes are similar (low variance). In the field, sedimentologists use graphical charts to accurately describe the sorting of a sediment using one of these terms. Tangential sorting is the result of sediment being deposited in same direction as flow. Normal tangential sorting results in a gradient of sediment sizes deposited from largest to finest as they travel downstream. When sediments are deposited from smallest to largest as they travel downstream, this is referred to as reverse sorting. Rocks derived from well sorted sediments are commonly both porous and permeable, while poorly sorted rocks have low porosity and low permeability, particularly when fine grained. == Processes involved in sorting == Sediment sorting is influenced by: grain sizes of sediment, processes involved in grain transport, deposition, and post-deposition processes such as winnowing. As a result, studying the degree of sorting in deposits of sediment can give insight into the energy, rate, and/or duration of deposition, as well as the transport process responsible for laying down the sediment. === Aeolian processes ===
{"page_id": 7396036, "title": "Sorting (sediment)"}
TPC-5CN or Trans-Pacific Cable 5 Cable Network is a submarine telecommunications cable system linking Japan, Guam, Hawaii and mainland United States. It has landing points in: Ninomiya, Kanagawa Prefecture, Japan Bandon, Coos County, Oregon, United States San Luis Obispo, San Luis Obispo County, California, United States Keawaula/Yokohama Beach, Wai'anae, Honolulu County, Oahu, Hawaii, United States Tumon Bay, Tumon, Tamuning, Guam Miyazaki, Miyazaki Prefecture, Japan It has a transmission capacity of 5 Gbit/s, and a total cable length of 22,500 km (14,000 mi). It started operation on 31 December 1996. == External links == "AT&T Keawaula Cable Terminal Station, Waianae, Island of Oahu, Hawaii". Archived from the original on February 21, 2005. Retrieved February 12, 2006. == References == Barnett, W. Christopher; Takahira, Hitoshi; Baroni, James C.; Ogi, Yoshihiro (February 1996). "The TPC-5 Cable Network". IEEE Communications. 34 (2): 36–40. doi:10.1109/35.481242. "The Entire TPC-5CN to Begin Service" (Press release). KDDI. December 12, 1996. Retrieved February 11, 2006. Taga, Hidenori (1997). Global lightwave undersea cable networks: present and future. Asia Pacific Regional Internet Conference on Operational Technologies. Hong Kong. p. 17. Retrieved February 11, 2006. "Global Marine, OCC and NEC Announce Strategic Alliance To Deliver Integrated Submarine Cable Networks" (Press release). NEC. March 8, 2001. Retrieved February 11, 2006.
{"page_id": 4036904, "title": "TPC-5CN"}
returning to Earth. === China === China is not an ISS partner, and no Chinese nationals have been aboard. China has its own contemporary human space program, China Manned Space Program, and has carried out co-operation and exchanges with countries such as Russia and Germany in human and robotic space projects. China launched its first experimental space station, Tiangong 1, in September 2011, and has officially initiated the permanently crewed Chinese space station project since 2021. In 2007, Chinese vice-minister of science and technology Li Xueyong said that China would like to participate in the ISS. In 2010, ESA Director-General Jean-Jacques Dordain stated his agency was ready to propose to the other 4 partners that China be invited to join the partnership, but that this needs to be a collective decision by all the current partners. While ESA is open to China's inclusion, the US is against it. US concerns over the transfer of technology that could be used for military purposes echo similar concerns over Russia's participation prior to its membership. Concerns over Russian involvement were overcome and NASA became solely dependent upon Russian crew capsules when its shuttles were grounded after the Columbia accident in 2003, and again after its retirement in 2011. The Chinese government has voiced a belief that international exchanges and co-operation in the field of aerospace engineering should be intensified on the basis of mutual benefit, peaceful use and common development. China's crewed Shenzhou spacecraft use an APAS docking system, developed after a 1994–1995 deal for the transfer of Russian Soyuz spacecraft technology. Included in the agreement was training, provision of Soyuz capsules, life support systems, docking systems, and space suits. American observers comment that Shenzhou spacecraft could dock at the ISS if it became politically feasible, whilst Chinese engineers say work would still
{"page_id": 41550219, "title": "Politics of the International Space Station"}
to see dof preview), and autogain-locked or suspended during aperture stopped down (difficult to see dof, just like looking with a non-live view DSLR of optical viewfinder (OVF) preview through-the-lens (TTL). For ES-LV digicam users and ExpSim LV DSLR users, this advanced method of "Adjustable/Selectable Live Previewable Exposure Evaluation/Determination" shooting in Full Manual / M ISO mode completely eliminated light meter sensor reading TTL common from the film era, making optimally desired exposure selection with live preview rear monitor screens virtually "instant at a glance", and time was saved doing any spot/area evaluative light metering TTL, which non-ES-LV and totally non-LV DSLR users doing the TTL "light meter reading" routine, same as their film SLR counterparts. The time saved assessing instant optimal exposure choices for any new or changing lit scenario, freed the ExpSim LV DSLR users to spend all their time shooting primarily with their optical viewfinder (OVF). When interchangeable-lens DSLRs with STILLS only ExpSim LV (APS-C EOS 20Da, APS-H EOS 1DMkIII, Full Frame (FF) EOS 1DsMkIII), with its ExpSim LV having Full Manual / M/Auto-ISO mode and AE / M/Auto-ISO mode, ExpSim LV was extended for the first time in any interchangeable-lens DSLR to Canon's FF ExpSim LV (1080p) video in 5DMkIII, it was only then, did use of OVF shifted over, more to the rear lcd monitor to shoot Cinematic (larger FF sensor frame than pro digital Super35mm and 35mm Cine cameras). The lesser Autogain Framing/Focus Live preview, borrowed from framing/focus autogain live view digicams, in lesser autogain live view DSLRs does not typically serve as their principal means of framing and previewing before taking a photograph, with this function still being mainly performed with optical viewfinder. The first interchangeable-lens DSLR to feature framing live preview was the 2004 Fujifilm FinePix S3 Pro. While initially largely a
{"page_id": 7937506, "title": "Live preview"}
PCB assembly. In embedded PoP techniques, chips are embedded in a substrate on the bottom of the package. This PoP technology enables smaller packages with shorter electrical connections and is supported by companies such as Advanced Semiconductor Engineering (ASE). === Advantages over traditional isolated-chip packaging === The most obvious benefit is motherboard space savings. PoP uses much less PCB area, almost as little as stacked-die packages. Electrically, PoP offers benefits by minimizing track length between different interoperating parts, such as a controller and memory. This yields better electrical performance of devices, since shorter routing of interconnections between circuits yields faster signal propagation and reduced noise and cross-talk. === Advantages over chip stacking === There are several key differences between stacked-die and stacked-package products. The main financial benefit of package on a package is that the memory device is decoupled from the logic device. Therefore this gives PoP all the same advantages that traditional packaging has over stacked-die products: The memory package can be tested separately from the logic package Only "known good" packages are used in final assembly (if the memory is bad only the memory is discarded and so on). Compare this to stacked-die packages where the entire set is useless and rejected if either the memory or logic is bad. The end user (such as makers of mobile phones or digital cameras) controls the logistics. This means memory from different suppliers can be used at different times without changing the logic. The memory becomes a commodity to be sourced from the lowest cost supplier. This trait is also a benefit compared to PiP (package in package) which requires a specific memory device to be designed in and sourced upstream of the end user. Any mechanically mating top package can be used. For a low-end phone, a smaller memory
{"page_id": 12792117, "title": "Package on a package"}
line, it was questioned. However the high abundance of carbon was also observed in the ultraviolet spectrum of the nebula. A more detailed spectral analysis confirmed the chemical segregation of the knots. The polar knots (J1 and J3) have higher helium to hydrogen ratio, with 7 times more helium than hydrogen, than the other knots, where the He/H ratio is about 4. Also the abundances of oxygen and nitrogen in the polar knots is about an order of magnitude lower than those observed in planetary nebula while neon is about three times less. Contrary, knot 4 has the expected abundances of oxygen and nitrogen. The oxygen recombination lines indicate that the polar knots have cool, yet ionised carbon, oxygen and nitrogen in their core. The temperatures calculated from the oxygen recombination lines are 500 K for J1 and 2500 K for J3. Around the central star carbon dust has been discovered in the infrared, lying at the same location with the knots. The total dust mass is estimated to be 3.20+3.21−2.06×10−3 M☉. Accounting for the carbon embedding in the dust, there is more carbon than oxygen in the ejecta of the nebula, favoring the very late thermal pulse model. The carbon dust is expected to destroyed around the central star and be pussed away by radiation pressure. == See also == Abell 78 - a similar planetary nebula V605 Aquilae - the central star of Abell 58, that underwent a very late thermal pulse Sakurai's Object - a star in a faint planetary nebula undergoing a very late thermal pulse == References ==
{"page_id": 78367381, "title": "Abell 30"}
Diggavi HIGHLIGHT: In this work, we begin with a generative framework that could potentially unify several different algorithms as well as suggest new algorithms. 1034, TITLE: WiNeRT: Towards Neural Ray Tracing for Wireless Channel Modelling and Differentiable Simulations AUTHORS: Tribhuvanesh Orekondy, Pratik Kumar, Shreya Kadambi, Hao Ye, Joseph Soriaga, Arash Behboodi HIGHLIGHT: In this paper, we work towards a neural surrogate to model wireless electro-magnetic propagation effects in indoor environments. 1035, TITLE: Grounding Graph Network Simulators using Physical Sensor Observations AUTHORS: Jonas Linkerhägner, Niklas Freymuth, Paul Maria Scheikl, Franziska Mathis-Ullrich, Gerhard Neumann HIGHLIGHT: In this work, we integrate sensory information to ground Graph Network Simulators on real world observations. 1036, TITLE: Implicit Regularization for Group Sparsity AUTHORS: Jiangyuan Li, Thanh V Nguyen, Chinmay Hegde, Raymond K. W. Wong HIGHLIGHT: We study the implicit regularization of gradient descent towards structured sparsity via a novel neural reparameterization, which we call a diagonally grouped linear neural network. 1037, TITLE: Noise Injection Node Regularization for Robust Learning AUTHORS: Noam Itzhak Levi, Itay Mimouni Bloch, Marat Freytsis, Tomer Volansky HIGHLIGHT: We introduce Noise Injection Node Regularization (NINR), a method of injecting structured noise into Deep Neural Networks (DNN) during the training stage, resulting in an emergent regularizing effect. 1038, TITLE: Compositional Semantic Parsing with Large Language Models AUTHORS: Andrew Drozdov, Nathanael Schärli, Ekin Akyürek, Nathan Scales, Xinying Song, Xinyun Chen, Olivier Bousquet, Denny Zhou HIGHLIGHT: Previous research shows that appropriate prompting techniques enable large language models (LLMs) to solve artificial compositional generalization tasks such as SCAN. In this work, we identify additional challenges in more realistic semantic parsing tasks with larger vocabulary and refine these prompting techniques to address them. 1039, TITLE: Guess the Instruction! Making Language Models Stronger Zero-Shot Learners AUTHORS: Seonghyeon Ye, Doyoung Kim, Joel Jang, Joongbo
{"source": 1216, "title": "from dpo"}
combine these laws into one statement: The total quantity of matter and energy in the universe is fixed. # Thermal Energy, Temperature, and Heat Thermal energy is kinetic energy associated with the random motion of atoms and molecules. Temperature is a quantitative measure of “hot” or “cold.” When the atoms and molecules in an object are moving or vibrating quickly, they have a higher average kinetic energy (KE), and we say that the object is “hot.” When the atoms and molecules are moving slowly, they have lower KE, and we say that the object is “cold” ( Figure 5.4 ). Assuming that no chemical reaction or phase change (such as melting or vaporizing) occurs, increasing the amount of thermal energy in a sample of matter will cause its temperature to increase. And, assuming that no chemical reaction or phase change (such as condensation or freezing) occurs, decreasing the amount of thermal energy in a sample of matter will cause its temperature to decrease. Figure 5.4 (a) The molecules in a sample of hot water move more rapidly than (b) those in a sample of cold water. Click on this interactive simulation ( 16PHETtempFX) to view the effects of temperature on molecular motion. Most substances expand as their temperature increases and contract as their temperature decreases. This property can be used to measure temperature changes, as shown in Figure 5.5 . The operation of many thermometers depends on the expansion and contraction of substances in response to temperature changes. ## Link to Learning > 234 Chapter 5 | Thermochemistry This content is available for free at Figure 5.5 (a) In an alcohol or mercury thermometer, the liquid (dyed red for visibility) expands when heated and contracts when cooled, much more so than the glass tube that contains the liquid. (b)
{"source": 3700, "title": "from dpo"}
(x) : x ∈ N0} is a strictly increasing sequence of nonnegative integers for j = 1 , . . . , s .Define N1(F) by (12.16). Choose N2(F) sufficiently large that for N ≥ N2(F) and P = N 1/k , we have |aij | ≤ cP k−i for i = 0 , 1, . . . , k − 1, and so Theorem 12.3 applies to the polynomials in the sequence F.Let N (F) = max {N1(F), N 2(F)) and c1 = (2 k!) 1/k . By Lemma 12.6, if N ≥ N (F) and x1, . . . , x s are nonnegative integers such that f1(x1) + · · · + fs(xs) ≤ N, 12.4 Waring’s Problem for Sequences of Polynomials 397 then xj ≤ c1P for j = 1 , . . . , s. Therefore, if 0 ≤ n ≤ N , then rF (n) = NSE { f1(x1) + · · · + fs(xs) = n with xj ∈ N0 for j = 1 , . . . , s (k) } = NSE { f1(x1) + · · · + fs(xs) = n with 0 ≤ xj ≤ c1P for j = 1 , . . . , s (k) } > k,c P s−k by Theorem 12.3. Let B(n) be the counting function of the set B. We have RF (N ) = > N ∑ > n=0 rF (n) = > N ∑ > n=0 > n∈B rF (n) > k,c B(N )P s−k = B(N )P s N . By Lemma 12.5, RF (N ) > 12 ( 23cs )s/k P s. It follows that B(N )/N k,c 1. This completes the proof.  We say that sets of integers A and B eventually coincide if there exists a
{"source": 5734, "title": "from dpo"}
ignored in the argumen t were at play. For instance, President Ronald Reagan’s foreign policy was probably a cause— in part —of the fall of the Soviet Union’s political system. But to say that Re agan caused the fall of the Soviet Union is to oversimplify the situation. The USSR’s economic situation, internal political dissent, their foreign policy, and other factors contributed together to usher in the dramatic change. Other examples include: 61 Bob finally got an A in a college course. The textbook Bob used was written by Horace Kline. We can conclude that Kline’s writing produces student success. The relationship between big businesses and local communities is declining these days. Obviously the CEOs are not doing their jobs. The Backyard Nudist Association has acquired large contributions from both Democrats and Republicans in our state. Thanks must therefore go to the BNA’s chair, Senator Sunny Shine, as she must be a great fundraiser. # Slippery Slope Slippery Slope arguments can be strong or weak, but they all share a pattern i n common. They’re actually a form of False Cause, but have their own unique characteristics. A Slippery Slope line of reasoning argues that A will cause B, B will then cause C, C will then cause D, and D will then cause E. But we don’t want E! So we should not even allow ourselves to get started down the slope; we should thus reject A. This argument can be strong if we have good reason to affirm each causal link (e.g., between A and B). But if we have good reason to reject one or more of the causal links, then just because we allow A to take place, it does not mean that E will occur. Here is a pretty good causal chain argument: If straight-A
{"source": 6820, "title": "from dpo"}
to the different molecular mechanisms. These regions can be seen in Fig. 1, a typical stress vs. strain measurement for natural rubber. The three mechanisms (labelled Ia, Ib, and II) predominantly correspond to the regions shown on the plot. The concept of entropy comes to us from the area of mathematical physics called statistical mechanics which is concerned with the study of large thermal systems, e.g. rubber networks at room temperature. Although the detailed behavior of the constituent chains are random and far too complex to study individually, we can obtain very useful information about their "average" behavior from a statistical mechanics analysis of a large sample. There are no other examples of how entropy changes can produce a force in our everyday experience. One may regard the entropic forces in polymer chains as arising from the thermal collisions that their constituent atoms experience with the surrounding material. It is this constant jostling that produces a resisting (elastic) force in the chains as they are forced to become straight. While stretching a rubber sample is the most common example of elasticity, it also occurs when rubber is compressed. Compression may be thought of as a two dimensional expansion as when a balloon is inflated. The molecular mechanisms that produce the elastic force are the same for all types of strain. When these elastic force models are combined with the complex morphology of the network, it is not possible to obtain simple analytic formulae to predict the macroscopic stress. It is only via numerical simulations on computers that it is possible to capture the complex interaction between the molecular forces and the network morphology to predict the stress and ultimate failure of a rubber sample as it is strained. === The Molecular Kink Paradigm for rubber elasticity === The Molecular Kink
{"page_id": 7623862, "title": "Rubber elasticity"}
of the Underwater Association helped equivalent bodies to develop in South Africa and Australia. == Symposia and UA publications == The UA held its first symposium in 1965 to discuss the results of that year's expeditions. It took place in the Physics Department of Imperial College London, on 29 October. The resulting papers, and those of symposia from 1966 to 1969, were published in booklet form. The proceedings volumes were reviewed in maritime research journals for the information of other marine scientists. Woods and Lythgoe in 1971 and Drew, Lythgoe and Woods in 1976 also published books on underwater science with contributions from members of the UA and elsewhere. The Association also published a newsletter several times each year, circulated by post, which was an essential component of its success in the days before digital communications. Edward Drew and Helen Ross were newsletter editors for much of this time. Copies of most of the UA publications on the proceedings of symposia are held by the library of the National Oceanography Centre, Southampton. == Typical publications in different disciplines == The Underwater Association provided a vehicle for marine scientists who used diving to pool experiences, learn different techniques, experiment with different breathing gas mixtures and diving gear, and to publish their results in a way that emphasised the effectiveness of diving, as well as the skills needed to work efficiently underwater. This achievement coexisted with routine publishing in the mainstream single-discipline refereed journals that demonstrate the high quality of the research. Published papers by Members in the Proceedings volumes and other academic journals included research on diving medicine, psychology, marine biology, fisheries, marine geology, marine physics, archaeology, and oceanographic engineering. The Underwater Association enabled many young marine scientists to develop their careers at a time when diving was not a common
{"page_id": 61912781, "title": "Underwater Association"}
Erythema (Ancient Greek: ἐρύθημα, from Greek erythros 'red') is redness of the skin or mucous membranes, caused by hyperemia (increased blood flow) in superficial capillaries. It occurs with any skin injury, infection, or inflammation. Examples of erythema not associated with pathology include nervous blushes. == Types == == Causes == It can be caused by infection, massage, electrical treatment, acne medication, allergies, exercise, solar radiation (sunburn), photosensitization, acute radiation syndrome, mercury toxicity, blister agents, niacin administration, or waxing and tweezing of the hairs—any of which can cause the affected capillaries to dilate, resulting in redness. Erythema is a common side effect of radiotherapy treatment due to patient exposure to ionizing radiation. == Diagnosis == Erythema disappears on finger pressure (blanching), whereas purpura or bleeding in the skin and pigmentation do not. There is no temperature elevation, unless it is associated with the dilation of arteries in the deeper layer of the skin. == See also == Hyperemia Flushing (physiology) List of cutaneous conditions == References == == External links ==
{"page_id": 532849, "title": "Erythema"}
and height h. However, the left green parallelogram also has the same area as the left green parallelogram of the upper figure, because they have the same base (the upper left side of the triangle) and the same height normal to that side of the triangle. Repeating the argument for the right side of the figure, the bottom parallelogram has the same area as the sum of the two green parallelograms. === Solid geometry === In terms of solid geometry, Pythagoras' theorem can be applied to three dimensions as follows. Consider the cuboid shown in the figure. The length of face diagonal AC is found from Pythagoras' theorem as: A C ¯ 2 = A B ¯ 2 + B C ¯ 2 , {\displaystyle {\overline {AC}}^{\,2}={\overline {AB}}^{\,2}+{\overline {BC}}^{\,2}\,,} where these three sides form a right triangle. Using diagonal AC and the horizontal edge CD, the length of body diagonal AD then is found by a second application of Pythagoras' theorem as: A D ¯ 2 = A C ¯ 2 + C D ¯ 2 , {\displaystyle {\overline {AD}}^{\,2}={\overline {AC}}^{\,2}+{\overline {CD}}^{\,2}\,,} or, doing it all in one step: A D ¯ 2 = A B ¯ 2 + B C ¯ 2 + C D ¯ 2 . {\displaystyle {\overline {AD}}^{\,2}={\overline {AB}}^{\,2}+{\overline {BC}}^{\,2}+{\overline {CD}}^{\,2}\,.} This result is the three-dimensional expression for the magnitude of a vector v (the diagonal AD) in terms of its orthogonal components {vk} (the three mutually perpendicular sides): ‖ v ‖ 2 = ∑ k = 1 3 ‖ v k ‖ 2 . {\displaystyle \|\mathbf {v} \|^{2}=\sum _{k=1}^{3}\|\mathbf {v} _{k}\|^{2}.} This one-step formulation may be viewed as a generalization of Pythagoras' theorem to higher dimensions. However, this result is really just the repeated application of the original Pythagoras' theorem to a succession of right
{"page_id": 26513034, "title": "Pythagorean theorem"}
≤ lim inf n → ∞ t n {\displaystyle \limsup _{n\to \infty }t_{n}\leq e^{x}\leq \liminf _{n\to \infty }t_{n}} so that lim n → ∞ t n = e x . {\displaystyle \lim _{n\to \infty }t_{n}=e^{x}.} This equivalence can be extended to the negative real numbers by noting ( 1 − r n ) n ( 1 + r n ) n = ( 1 − r 2 n 2 ) n {\textstyle \left(1-{\frac {r}{n}}\right)^{n}\left(1+{\frac {r}{n}}\right)^{n}=\left(1-{\frac {r^{2}}{n^{2}}}\right)^{n}} and taking the limit as n goes to infinity. === Characterization 1 ⇔ characterization 3 === Here, the natural logarithm function is defined in terms of a definite integral as above. By the first part of fundamental theorem of calculus, d d x ln ⁡ x = d d x ∫ 1 x 1 t d t = 1 x . {\displaystyle {\frac {d}{dx}}\ln x={\frac {d}{dx}}\int _{1}^{x}{\frac {1}{t}}\,dt={\frac {1}{x}}.} Besides, ln ⁡ 1 = ∫ 1 1 d t t = 0 {\textstyle \ln 1=\int _{1}^{1}{\frac {dt}{t}}=0} Now, let x be any fixed real number, and let y = lim n → ∞ ( 1 + x n ) n . {\displaystyle y=\lim _{n\to \infty }\left(1+{\frac {x}{n}}\right)^{n}.} Ln(y) = x, which implies that y = ex, where ex is in the sense of definition 3. We have ln ⁡ y = ln ⁡ lim n → ∞ ( 1 + x n ) n = lim n → ∞ ln ⁡ ( 1 + x n ) n . {\displaystyle \ln y=\ln \lim _{n\to \infty }\left(1+{\frac {x}{n}}\right)^{n}=\lim _{n\to \infty }\ln \left(1+{\frac {x}{n}}\right)^{n}.} Here, the continuity of ln(y) is used, which follows from the continuity of 1/t: ln ⁡ y = lim n → ∞ n ln ⁡ ( 1 + x n ) = lim n → ∞ x ln ⁡ ( 1 +
{"page_id": 853175, "title": "Characterizations of the exponential function"}
of Märkischer Kreis Homert (519 m), county of Oberbergischer Kreis Horst (514.9 m), county of Märkischer Kreis Eulenberg (497.1 m), county of Märkischer Kreis Sonnenberg (458.5 m), county of Olpe Nordhelle (437.2), county of Märkischer Kreis Bracht (429.1 m), county of Olpe Rappelsberg (410.7 m), county of Olpe Gilberg (360.0 m), county of Olpe === Siegerland (natural region) === The natural region of Siegerland (not to be confused with the historic Siegerland) lies entirely within North Rhine-Westphalia: → see also section: Mountains in the article on Siegerland (natural region) Alte Burg (633.0 m), county of Siegen-Wittgenstein, (NW) Lipper Nürr (616.9 m), county of Altenkirchen (RP), county of Siegen-Wittgenstein (NW) Sanktkopf (ca. 606 m), county of Siegen-Wittgenstein, (NW) Die Höh (ca. 598 m), county of Siegen-Wittgenstein (NW) Die Burg (594.5 m), county of Siegen-Wittgenstein (NW) Hohenseelbachskopf (517.5 m), county of Siegen-Wittgenstein (NW), county of Altenkirchen (RP) Eisernhardt (482.3 m), borough of Siegen (NW) === North Sauerland Oberland === ==== Arnsberg Forest ==== The Arnsberg Forest lies in the eponymous nature park entirely within North Rhine-Westphalia. → see also section: Hills in the article on Arnsberg Forest Nature Park Plackweghöhe (actually an unnamed peak; 581.5), county of Hochsauerlandkreis second highest hill: unnamed peak (559.5), county of Hochsauerlandkreis third highest hill: unnamed peak (557.4), county of Hochsauerlandkreis Warsteiner Kopf (556.9 m), county of Hochsauerlandkreis Gemeinheitskopf (551.9 m), county of Hochsauerlandkreis Niekopf (550.4 m), county of Soest === Bergisches Land on the Upper Agger and Wiehl === Unnenberg (505.7 m), county of Oberbergischer Kreis == Taunus == The Taunus lies in Hesse with foothills in Rhineland-Palatinate. → main article: List of mountains and hills of the Taunus Großer Feldberg (881.5 m), county of Hochtaunuskreis Kleiner Feldberg (825.2 m), Hochtaunuskreis Altkönig (798.2 m), Hochtaunuskreis Weilsberg (700.7 m), Hochtaunuskreis Glaskopf (686.8 m), Hochtaunuskreis Pferdskopf (662.6 m),
{"page_id": 44405847, "title": "List of mountains and hills of the Rhenish Massif"}
the point. Amitai Etzioni – Professor at The George Washington University and an adviser to the US's Carter administration – countered in Humanity Would Be Better off Saving Earth, Rather Than Colonizing Mars that: "It is better to hold off disasters at home than to assume all is lost". Etzioni also pointed out the vast cost of colonization of extraterrestrial planets by citing that Elon Musk, an advocate of space exploration and colonization, had calculated the cost of sending the first 12 astronauts to Mars at £10 billion per person. The Mars Climate Orbiter is a good example of this argument, burning up—before returning any scientific data—at a cost of $328 million. Social critics say that the cost of space exploration cannot be justified when hunger and poverty are rampant. "As they see it, space exploration takes money, resources, and talent away from helping people in need and from improving the quality of life for everybody." In 1967, Martin Luther King Jr. said: "Without denying the value of scientific endeavor, there is a striking absurdity in committing billions to reach the moon where no people live, while only a fraction of that amount is appropriated to service the densely populated slums." Some critics have pointed out the hazards of space debris which affect satellites, spacecraft and the surface of the Earth. For example, in March 2009 debris believed to be a 10 cm (3.9 in) piece of the Kosmos 1275 satellite nearly hit the ISS. Although it is relatively rare for people on the ground to be hit by space debris, it does happen. In 1969 five sailors on a Japanese ship were injured by space debris. In 1997 an Oklahoma woman, Lottie Williams, was injured when she was hit in the shoulder by a 10 cm × 13 cm
{"page_id": 56768571, "title": "Benefits of space exploration"}
longevity was invaluable for the studies because while the monkeys were being trained at a task, neurons that were initially silent began firing as the task was learned, a phenomenon that would not have been observable if the electrode was not capable of long term implantation. == Components == === Glass cone === The glass cone is only 1–2 mm long, and is filled with trophic factors in order to encourage axons and dendrites to grow through its tip and hollow body. When the neurites reach the back end of the cone, they rejoin with the neuropil on that side, which anchors the glass cone in place. As a result, stable and robust long-term recording is attainable. The cone sits with its tip near layer five of the cortex, among corticospinal tract cell bodies, and is inserted at an angle of 45° from the surface, about 5 or 6 mm deep. === Gold wires === Three or four gold wires are glued to the inside of the glass cone and protrude out the back. They record the electrical activity of the axons that have grown through the cone, and are insulated with Teflon. The wires are coiled so as to relieve strain because they are embedded in the cortex on one end and attached to the amplifiers, which are fixed to the inside of the skull, on the other. Two wires are plugged into each amplifier to provide differential signalling. === Wireless transmitter === One of the greatest strengths of the neurotrophic electrode is its wireless capability, because without transdermal wiring, the risk of infection is significantly reduced. As neural signals are collected by the electrodes, they travel up the gold wires and through the cranium, where they are passed on to the bioamplifiers (usually implemented by differential amplifiers). The
{"page_id": 29806228, "title": "Neurotrophic electrode"}
Karen Henwood is a British social psychologist and Professor of Social Sciences at Cardiff University, and an expert on identity and risk, particularly socio-cultural and environmental change. Her research in recent years includes the ESRC project "Timescapes" on relationships and identities through the life course, and a project on men as fathers, as well as research on energy use, sustainable development, climate change policy and on living with nuclear risk. She was editor-in-chief of Qualitative Research from 2016 to 2019. == Biography == She holds a BSc in psychology and a PhD in social psychology from the University of Bristol, and worked in clinical psychology and health psychology at the School of Psychology, Bangor University (1995–1999) and the School of Medicine, Health Policy and Practice at the University of East Anglia (1999–2006), before her 2006 appointment as senior lecturer and later professor at the Cardiff School of Social Sciences. She is affiliated with the Collaborative Centre of Excellence in Understanding and Managing Natural and Environmental Risks, and with the Sustainable Places Research Institute. Henwood has authored several highly cited publications within social psychology. Her most cited work, "Qualitative research and psychological theorizing," was cited over 1100 times as of 2019. According to Google Scholar Henwood has been cited around 12,000 times in scientific literature and has an h-index of 45. == Selected bibliography == Henwood, Karen; Phoenix, Ann; Griffin, Christine (1998). Standpoints and differences: essays in the practice of feminist psychology. London Thousand Oaks, California: SAGE. ISBN 9780761954446. == References == == External links == Profile page: Karen Henwood, Cardiff University
{"page_id": 42427495, "title": "Karen Henwood"}
Stroke: A Randomized Control Trial Using Combinatory Approach Toward Reducing Workforce Demands. Front Neurol. 2021 Jun 2;12:622014. doi: 10.3389/fneur.2021.622014. PMID: 34149587; PMCID: PMC8206540)_. Treatments using rehabilitation robots enable delegating more manual and repetitive therapy components to robotic devices, allowing a clinician practitioner to take care of more patients in a given time and improving the accessibility of therapy for patients remotely from the comfort of their homes through telerehabilitation. The data collected can objectively assess performance and document compliance and progress using artificial intelligence (AI), promoting data-driven therapy. Virtual reality (VR), combined with haptics, offers therapists more customisable treatment options in a safe environment. * The benefits of the adoption of rehabilitation robots are clear: * A therapist can get an overview of the patient’s progress, * The technology is primarily automated and data-driven, and * Significant increase in accessibility to therapy by patients. There are still issues to be solved in rehabilitation robots and many questions to be answered, such as efficacy, costs, reimbursement, and regulatory challenges. McKinsey Global Institute concept
{"source": 10, "title": "from dpo"}
be considered as having been required to provide _certified cost or pricing data_. Consequently, award of any lower-tier _subcontract_ expected to exceed the _certified cost or pricing data_ threshold requires the submission of _certified cost or pricing data_ unless- (i) An exception otherwise applies to the _subcontract_; or (ii) The waiver specifically includes the _subcontract_ and the rationale supporting the waiver for that _subcontract_. #### 15.403-2 Other circumstances where certified cost or pricing data are not required. (a) The exercise of an _option_ at the _price_ established at contract award or initial negotiation does not require submission of _certified cost or pricing data_. (b)_Certified cost or pricing data_ are not required for proposals used solely for overrun funding or interim billing _price_ adjustments. #### 15.403-3 Requiring data other than certified cost or pricing data. (a) (1) In those _acquisitions_ that do not require _certified cost or pricing data_, the _contracting officer_ _shall_— (i) Obtain whatever data are available from Government or other secondary sources and use that data in determining a fair and reasonable _price_; (ii) Require submission of _data other than certified cost or pricing data_, as defined in 2.101, from the _offeror_ to the extent necessary to determine a fair and reasonable _price_ ( 10 U.S.C. 3705(a)") and 41 U.S.C.3505(a)) if the _contracting officer_ determines that adequate data from sources other than the _offeror_ are not available. This includes requiring data from an _offeror_ to support a _cost realism_ analysis; (iii) Consider whether cost data are necessary to determine a fair and reasonable _price_ when there is not adequate _price_ competition; (iv) Require that the data submitted by the _offeror_ include, at a minimum, appropriate data on the _prices_ at which the same item or similar items have previously
{"source": 2295, "title": "from dpo"}
6.5 Summary 32 6.6 Reviewers 32 6.7 References 33 6.8 Additional Information 6–1 CHAPTER 6 # AUTOMATED FINGERPRINT IDENTIFICATION SYSTEM (AFIS) # Kenneth R. Moses # Contributing authors Peter Higgins, Michael McCabe, Salil Prabhakar, Scott Swann # 6.1 Introduction Prior to the industrial revolution and the mass migrations to the cities, populations lived mostly in rural communities where everyone knew everyone else and there was little need for identification. Indeed, there were no police forces, no penitentiaries, and very few courts. As cities became crowded, crime rates soared and criminals flourished within a sea of anonymity. Newspapers feasted on stories of lawlessness, legislatures quickly responded with more laws and harsher penalties (especially for repeat offenders), and police departments were charged with identifying and arresting the miscreants. Identification systems—rogues’ galleries, anthropometry, Bertillon’s “portrait parlé” , and the Henry system—emerged and quickly spread worldwide at the end of the 19th and beginning of the 20th century. The late 1960s and early 1970s witnessed another era of civil turmoil and an unprecedented rise in crime rates, but this era happened to coincide with the development of the silicon chip. The challenges inherent in identification systems seemed ready-made for the solutions of auto - matic data processing, and AFIS—Automated Fingerprint Identification System—was born. During this same period, The RAND Corporation, working under a national grant, published The Criminal Investigative Process (Greenwood et al., 1975), a comprehensive study and critique of the process by which crimes get solved—or do not. Generally critical of traditional methods used by detectives, the study placed any hopes for improvement on physical evidence in general and latent prints in particular. In a companion study, Joan Petersilia concluded that: No matter how competent the evidence techni - cian is at performing his job, the gathering of physical evidence at a crime
{"source": 4211, "title": "from dpo"}
\in D , {\operatorname *{colim}}_{c \in C} Y_c(d) exists in \mathcal {E} , then it is functorial in d, and it is a colimit in \mathcal {E}^D . In particular, an object over {\operatorname *{colim}}_{c} Y_c is a D-indexed diagram X(d) \to {\operatorname *{colim}}_C Y_c(d) , which as these colimits are all van Kampen is the same as a (C \times D) -indexed diagram X_c(d) \to Y_c(d) that is Cartesian in the C-direction, which in turn is the same as a C-indexed diagram X_{\bullet } \in \mathcal {E}^D that is Cartesian over Y_{\bullet } , hence proving the lemma. We now recall the definition of various kinds of lextensive categories [ Reference Carboni, Lack and Walters 7]. Definition 2.4. Let \mathcal {E} be a category with finite limits. For a regular cardinal \alpha , we say that \mathcal {E} is \alpha -lextensive if \alpha -coproducts exist and are van Kampen colimits. Furthermore, we say that \mathcal {E} is (i) Lextensive if it is \omega -lextensive: that is, finite coproducts exist and are van Kampen colimits, (ii) Countably lextensive if it is \omega _1 -lextensive: that is, countable coproducts exist and are van Kampen colimits, (iii) Completely lextensive if it is \alpha -lextensive for all \alpha : that is, all small coproducts exist and are van Kampen colimits. Example 2.5. There are numerous examples of lextensive categories. (i) Any presheaf category is completely lextensive. In particular, for any group G, the category of G-sets is countably lextensive. (ii) More generally, any Grothendieck topos is completely lextensive. In fact, Giraud’s theorem characterises Grothendieck toposes as the locally presentable categories in which coproducts and (in an appropriate sense) quotients by equivalence relations are van Kampen colimits. (iii) The category of topological spaces is completely lextensive. The same is true for many of its
{"source": 6081, "title": "from dpo"}
of classical field theory == Notes == == References ==
{"page_id": 3114255, "title": "History of classical mechanics"}
act in which the Hall of Scientists of the Casa Rosada (seat of the executive power of the Argentine Republic) it was renamed as the Argentine Science Hall, making "the four of Melchior" part of the honored female figures. A photograph of Fontes was added to the gallery of images of the Hall together with those of 10 other women of science from Argentina: Elisa Bachofen, Eugenia Sacerdote de Lustig, Rebeca Guber, Sara Rietti, Elvira López, Telma Reca, Hetty Ladis Regina Bertoldi de Pomar, Irene Bernasconi, María Adela Caría and Carmen Pujals. == References ==
{"page_id": 72971108, "title": "Elena Martinez Fontes"}
therapy. Orphan indications include diffuse large B-cell lymphoma, chronic lymphocytic leukemia and mantle cell lymphoma. Lenalidomide is also approved for transfusion-dependent anemia due to low or intermediate-1-risk myelodysplastic syndromes associated with a deletion 5q cytogenetic abnormality with or without additional cytogenetic abnormalities in the U.S., Canada, Switzerland, Australia, New Zealand, Malaysia, Israel and several Latin American countries, while marketing authorization application is currently being evaluated in a number of other countries. Numerous clinical trials are already in the pipeline or being conducted to explore further use for lenalidomide, alone or in combination with other drugs. Some of these indications include acute myeloid leukemia, follicular lymphoma, MALT lymphoma, Waldenström macroglobulinemia, lupus erythematosus, Hodgkin's lymphoma, myelodysplastic syndrome and more. === Pomalidomide === Pomalidomide was submitted for FDA approval on April 26, 2012 and on 21 June it was announced that the drug would get standard FDA review. A marketing authorization application was filed to EMA 21 June 2012, where a decision could come as soon as early 2013. EMA has already granted pomalidomide an orphan designation for primary myelofibrosis, MM, systemic sclerosis, post-polycythaemia and post-essential thrombocythaemia myelofibrosis. == Adverse effects == The major toxicities of approved IMiDs are peripheral neuropathy, thrombocytopenia, anaemia and venous thromboembolism. There may be an increased risk of secondary malignancies, especially acute myeloid leukaemia in those receiving IMiDs. === Teratogenicity === Thalidomide's teratogenicity has been a subject of much debate and over the years numerous hypotheses have been proposed. Two of the best-known have been the anti-angiogenesis hypothesis and oxidative stress model hypothesis, with considerable experimental evidence supporting these two hypotheses regarding thalidomide's teratogenicity. Recently, new findings have emerged that suggest a novel mechanism of teratogenicity. Cereblon is a 51 kDa protein localized in the cytoplasm, nucleus and peripheral membrane of cells in numerous parts of the body.
{"page_id": 41933329, "title": "Cereblon E3 ligase modulator"}
Access Technology 3G/4G platform that is software-upgradable to 5G. === nanoCBRS === nanoCBRS is an OnGo certified solution for the deployment of private LTE networks in the 3.5 GHz band. Launched in October 2019, the solution incorporates Citizens Band Radio Devices, a suite of architectures – nanoViNE - to support use cases and nanoCBRSLab, a mobile network lab solution for testing. == Customers and Partners == Among ip.access' major customers are AT&T, Batelco, Bharti Airtel, Blue Ocean Wireless, Bouygues Telecom, Caribsat, Digicel, Globe, Jersey Telecom Monaco Telecom, O2 (UK), SFR, SPIE SA, T-Mobile, Tele2, Telefónica O2 Czech Republic, Telenet, Telia Sonera, T-Mobile, Vivacom and Vodafone. The company's technology partners include AeroMobile, Africa Mobile Networks, Altobridge, Blue Ocean Wireless, Cisco Systems, Druid Software, Intel, Pentonet, Private Mobile Networks, Qualcomm, Quortus, Setcom, and TriaGnoSys. == Awards == Corporate, product, and personnel awards won by ip.access include the following: === 2007 === Light Reading Leading Lights Awards - Best New Product (Private Company) Global Mobile Awards - Best Radio Access Product or Service The Sunday Times Fast Tech Track 100 World Communications Award - Best Technology Foresight === 2008 === The Sunday Times Fast Tech Track 100 === 2009 === Deloitte Fast 500 EMEA Femto Forum Awards - Femtocell or Femtocell Network Element Design and Technology Innovation Femto Forum Awards - Femtocell Service (Commercial, Prototype, or Demo) Global Telecoms Business Innovation Awards - Maritime Voice Services === 2010 === GTB Power100 World Vendor Awards - Best Specialist Vendor === 2011 === Femto Forum Awards - Femtocell Network Element Design and Technology Innovation === 2012 === World Vendor Awards - Small Cell Award Small Cell Forum Industry Awards - Judges Choice Award === 2013 === Small Cell Forum Industry Awards - Social Vision, promoting Small Cells for Social / Economical / Environmental development ===
{"page_id": 32261703, "title": "Ip.access"}
surface and structure. The nucleus is a conglomerate of ices and dust, often referred to as a "dirty snowball". Like all comets, as Halley nears the Sun, its volatile compounds (those with low boiling points, such as water, carbon monoxide, carbon dioxide and other ices) begin to sublimate from the surface. This causes the comet to develop a coma, or atmosphere, at distances up to 230,000 kilometres (140,000 mi) from the nucleus. Sublimation of this dirty ice releases dust particles, which travel with the gas away from the nucleus. Gas molecules in the coma absorb solar light and then re-radiate it at different wavelengths, a phenomenon known as fluorescence, whereas dust particles scatter the solar light. Both processes are responsible for making the coma visible. As a fraction of the gas molecules in the coma are ionised by the solar ultraviolet radiation, pressure from the solar wind, a stream of charged particles emitted by the Sun, pulls the coma's ions out into a long tail, which may extend more than 100 million kilometres into space. Changes in the flow of the solar wind can cause disconnection events, in which the tail completely breaks off from the nucleus. Despite the vast size of its coma, Halley's nucleus is relatively small: barely 15 kilometres (9.3 mi) long, 8 kilometres (5.0 mi) wide and perhaps 8 kilometres (5.0 mi) thick. Based on a reanalysis of images taken by the Giotto and Vega spacecraft, Lamy et al. determined an effective diameter of 11 kilometres (6.8 mi). Its shape has been variously compared to that of a peanut, a potato, or an avocado. Its mass is roughly 2.2 × 1014 kg, with an average density of about 0.55 grams per cubic centimetre (0.32 oz/cu in). The low density indicates that it is made of a
{"page_id": 46083, "title": "Halley's Comet"}
specific point in the signal's sweep; these replace all other pulse types when transmitted. The civilian VOR system differs from TACAN in utilizing a single continuous-wave 30 Hz modulation signal, using the phase difference between a fixed-phase and variable phase (rotating) component to derive bearing info. === Squitter function === TACAN stations transmit pulse-pairs at a composite rate of 3600 pairs/second: 900 of which are bearing reference bursts, and the other 2700 being composed of ranging and identification pulses. When insufficient interrogation pulses from aircraft are present, the station will use a squitter circuit to inject additional randomized pulse-pairs to maintain the desired pulse rate. This ensures that sufficient signal is present to support demodulating bearing signals. === Identification === TACAN stations are identified by Morse code. The transmitting station periodically replaces the randomized ranging pulse-pairs with regularly spaced pairs that de-modulate to a 1350 Hz tone, keying a three-letter identification code at approximately 6-7 wpm every 40 seconds. Ranging and squitter pulses are permitted during the gaps between dots and dashes. There is no capability for voice transmission in a TACAN-only system. === Operating modes === There are two basic channel configurations available: X (the original implementation) and Y (added in the 1960s to expand available channels and reduce mutual interference between closely-spaced stations). These configurations differ in pulse-pair width, fixed receiver response delay, and polarity of frequency offset from the interrogation channel. TACAN interrogators can operate in four modes: receive (for bearing/identification only), transmit/receive (for bearing, range, and ID), and air-to-air versions of the previous two. The typical TACAN onboard user panel has control switches for setting the channel (corresponding to the desired station's assigned frequency), the operation mode for either transmit/receive (T/R, to get both bearing and range) or receive only (REC, to get bearing but not
{"page_id": 1127304, "title": "Tactical air navigation system"}
HD 85512 is a solitary K-type main-sequence star 36.8 light-years (11.3 parsecs) away in the constellation Vela. It is about 1 billion years older than the Sun. It is extremely chromospherically inactive, only slightly more active than Tau Ceti. It exhibits a long-term variability and was thought to host one low-mass planet, although this is now doubtful. == Proposed planetary system == On August 19, 2011, a ≥3.6 Earth-mass planet was discovered using HARPS that is "just inside" the habitable zone, along with the inner planets of e (or 82 G.) Eridani, and HD 192310 c in Capricornus. These two other systems are closer to Earth than this system. Modelling at the time of the discovery announcement found that the planet could be cool enough to host liquid water if it has more than 50% cloud coverage, but with revised models of the habitable zone two years later it was found to be too hot to be potentially habitable. For a time it ranked fifth-best for habitability in the Planetary Habitability Laboratory's Habitable Exoplanets Catalog, which later listed it in an article about "false starts" in the search for potentially habitable exoplanets. In 2023, a study reassessed the radial velocity data of HD 85512. A signal was detected with a period of 51 days, inconsistent with the previously published 58-day orbital period of HD 85512 b, but consistent with previous estimates of the stellar rotation period. This indicates that the signal is very likely to be caused by the stellar rotation, rather than an orbiting planet.: 25–27 : 44 Due to this, the NASA Exoplanet Archive now considers this planet as a false positive detection. == References == == Notes == == External links == "ARICNS: 00766". ARICNS. Centre of Astronomy, Heidelberg University.
{"page_id": 15059909, "title": "HD 85512"}
| a k | + ∑ n = n 1 ∞ c n − n 1 | a n 1 | = ∑ k = 1 n 1 − 1 | a k | + | a n 1 | ∑ n = 0 ∞ c n . {\displaystyle \sum _{n=1}^{\infty }|a_{n}|=\sum _{k=1}^{n_{1}-1}|a_{k}|+\sum _{n=n_{1}}^{\infty }|a_{n}|\leq \sum _{k=1}^{n_{1}-1}|a_{k}|+\sum _{n=n_{1}}^{\infty }c^{n-n_{1}}|a_{n_{1}}|=\sum _{k=1}^{n_{1}-1}|a_{k}|+\left|a_{n_{1}}\right|\sum _{n=0}^{\infty }c^{n}.} The series ∑ n = 0 ∞ c n {\displaystyle \sum _{n=0}^{\infty }c^{n}} is the geometric series with common ratio c ∈ ( 0 ; 1 ) {\displaystyle c\in (0;1)} , hence ∑ n = 0 ∞ c n = 1 1 − c {\displaystyle \sum _{n=0}^{\infty }c^{n}={\frac {1}{1-c}}} which is finite. The sum ∑ k = 1 n 1 − 1 | a k | {\displaystyle \sum _{k=1}^{n_{1}-1}|a_{k}|} is a finite sum and hence it is bounded, this implies the series ∑ n = 1 ∞ | a n | {\displaystyle \sum _{n=1}^{\infty }|a_{n}|} converges by the monotone convergence theorem and the series ∑ n = 1 ∞ a n {\displaystyle \sum _{n=1}^{\infty }a_{n}} converges by the absolute convergence test. When the limit | a n + 1 a n | {\displaystyle \left|{\frac {a_{n+1}}{a_{n}}}\right|} exists and equals to L {\displaystyle L} then r = R = L {\displaystyle r=R=L} , this gives the original ratio test. == Extensions for L = 1 == As seen in the previous example, the ratio test may be inconclusive when the limit of the ratio is 1. Extensions to the ratio test, however, sometimes allow one to deal with this case. In all the tests below one assumes that Σan is a sum with positive an. These tests also may be applied to any series with a finite number of negative terms. Any such series may be written as: ∑ n
{"page_id": 421341, "title": "Ratio test"}
Vilnius College of Technologies and Design (Lithuanian: Vilniaus technologijų ir dizaino kolegija) or VTDK is a Lithuanian state institution of higher education VTDK is one of the largest universities of applied sciences in Lithuania, with approximately 2,000 students and 200 professors == History == The college was established in 1954. In 2008, the Vilnius Technical College (Lithuanian: Vilniaus statybos technikumas) merged with the Vilnius College of Construction and Design (Lithuanian: Vilniaus statybos ir dizaino kolegija), creating VTDK. == Overview == VTDK consists of three faculties: the Civil Engineering Faculty, the Technical Faculty and the Design Faculty. The college offers 15 full-time study programs, partial studies through exchange programs, as well as informal study courses and training. VTDK maintains close cooperation with many Lithuanian and foreign partners, takes part in various international projects, and is a member of the Universities of Applied Sciences network (UASnet) and the European Association of Institutions in Higher Education (EURASCHE). == Study programs == VTDK offers 15 full-time study programs, partial studies through exchange programs, as well as informal study courses and training. Program duration is 3 years (6 semesters) and students receive a professional bachelor's degree upon graduation. === Civil Engineering Faculty === Engineering Systems of Buildings Geodesy and Cadastre Transport Logistics Civil Engineering === Technical Faculty === Technical Maintenance of Automobiles Electrical and Automation Engineering Mechanical Technologies Engineering Railway Transport Engineering Transport Information Systems Renewable Energy Car Electronics === Design Faculty === Technology of Photography Multimedia Design Graphic Design Interior Design == Facilities == VTDK has the following facilities: 50 laboratories 9 libraries 5 drawing rooms The TOYOTA technical training center The language learning center The distance learning center Sport facilities Dormitories == External links == Official website
{"page_id": 40695272, "title": "Vilnius College of Technologies and Design"}
William Thurston's elliptization conjecture states that a closed 3-manifold with finite fundamental group is spherical, i.e. has a Riemannian metric of constant positive sectional curvature. == Relation to other conjectures == A 3-manifold with a Riemannian metric of constant positive sectional curvature is covered by the 3-sphere, moreover the group of covering transformations are isometries of the 3-sphere. If the original 3-manifold had in fact a trivial fundamental group, then it is homeomorphic to the 3-sphere (via the covering map). Thus, proving the elliptization conjecture would prove the Poincaré conjecture as a corollary. In fact, the elliptization conjecture is logically equivalent to two simpler conjectures: the Poincaré conjecture and the spherical space form conjecture. The elliptization conjecture is a special case of Thurston's geometrization conjecture, which was proved in 2003 by G. Perelman. == References == For the proof of the conjectures, see the references in the articles on geometrization conjecture or Poincaré conjecture. William Thurston. Three-dimensional geometry and topology. Vol. 1. Edited by Silvio Levy. Princeton Mathematical Series, 35. Princeton University Press, Princeton, NJ, 1997. x+311 pp. ISBN 0-691-08304-5. William Thurston. The Geometry and Topology of Three-Manifolds, 1980 Princeton lecture notes on geometric structures on 3-manifolds, that states his elliptization conjecture near the beginning of section 3.
{"page_id": 235954, "title": "Thurston elliptization conjecture"}
decision values given by the classifier.decision_function() method. In the case of providing the probability estimates, the probability of the class with the “greater label” should be provided. The “greater label” corresponds to classifier.classes_ and thus classifier.predict_proba(X)[:, 1]. Therefore, the y_score parameter is of size (n_samples,). >>> from sklearn.datasets import load_breast_cancer >>> from sklearn.linear_model import LogisticRegression >>> from sklearn.metrics import roc_auc_score >>> X, y = load_breast_cancer(return_X_y=True) >>> clf = LogisticRegression(solver="liblinear").fit(X, y) >>> clf.classes_ array([0, 1]) We can use the probability estimates corresponding to clf.classes_. >>> y_score = clf.predict_proba(X)[:, 1] >>> roc_auc_score(y, y_score) 0.99... Otherwise, we can use the non-thresholded decision values >>> roc_auc_score(y, clf.decision_function(X)) 0.99... 3.4.4.15.2. Multi-class case The roc_auc_score function can also be used in multi-class classification. Two averaging strategies are currently supported: the one-vs-one algorithm computes the average of the pairwise ROC AUC scores, and the one-vs-rest algorithm computes the average of the ROC AUC scores for each class against all other classes. In both cases, the predicted labels are provided in an array with values from 0 to n_classes, and the scores correspond to the probability estimates that a sample belongs to a particular class. The OvO and OvR algorithms support weighting uniformly (average='macro') and by prevalence (average='weighted'). One-vs-one Algorithm One-vs-rest Algorithm 3.4.4.15.3. Multi-label case In multi-label classification, the roc_auc_score function is extended by averaging over the labels as above. In this case, you should provide a y_score of shape (n_samples, n_classes). Thus, when using the probability estimates, one needs to select the probability of the class with the greater label for each output. >>> from sklearn.datasets import make_multilabel_classification >>> from sklearn.multioutput import MultiOutputClassifier >>> X, y = make_multilabel_classification(random_state=0) >>> inner_clf = LogisticRegression(solver="liblinear", random_state=0) >>> clf = MultiOutputClassifier(inner_clf).fit(X, y) >>> y_score = np.transpose([y_pred[:, 1] for y_pred in clf.predict_proba(X)]) >>> roc_auc_score(y, y_score, average=None) array([0.82..., 0.86..., 0.94..., 0.85... , 0.94...])
{"source": 992, "title": "from dpo"}
visualizations, such as binned scatterplots. Provide sensible defaults, but allow customization. Vega-Lite’s compiler automatically chooses default properties of a visualization based on a set of carefully designed rules. However, one can specify additional properties to customize the visualization. For example, the stacked bar chart on the left has a custom color palette. Vega-Lite uses a concise syntax, enabling rapid creation of visualizations without unduly restricting subsequent customization. Support programmatic generation, sharing, and reuse. Like Vega, Vega-Lite can serve as a standalone file format for visualizations. In particular, Vega-Lite is designed to be a convenient format for automatic visualization generation by visual analysis tools. Examples of applications that use Vega-Lite are Voyager, a recommendation-powered visualization browser for data exploration, and Polestar, a web-based visual specification interface inspired by Tableau. Moreover, one can reuse Vega-Lite specifications across datasets with similar schemas. Leverage Vega’s performance, flexibility across platforms, and expressivity. Vega-Lite specifications are compiled into Vega specifications and rendered using Vega’s runtime, which supports both browser-side and server-side rendering via SVG or Canvas. Vega-Lite directly benefits from Vega’s architecture. While Vega-Lite focuses on commonly-used charts, one can create more advanced designs by starting with Vega-Lite and then further customizing the resulting Vega specification. With the 1.0 release, Vega-Lite provides a useful tool for visualization on the web. That said, we are even more excited about what comes next. Composition and Interaction. A powerful aspect of modular approaches to visualization is the ability to create sophisticated graphics by composing simple ones. A static visualization typically provides at most a handful of insights into the data. The true power of visualization lies in the ability to interact with data and see it from multiple perspectives. So, we are building new methods for composite, interactive visualizations in Vega-Lite. In the coming months, Vega-Lite will add support
{"source": 2757, "title": "from dpo"}
2008 was 14 percent higher than the year before. Reese Witherspoon is now appearing in ads for Avon makeup, skin care products, and fragrances. In addition, she is also traveling worldwide as the Avon Foundation’s first global ambassador and honorary chairwoman. In December 2008, as a part of the company’s strategy to continue growing its universe of celebrity and designer beauty alliances, Avon announced a celebrity deal with Courteney Cox to be the face of the brand’s new women’s fragrance, Spotlight, launched in April 2009. The April introduction was followed by a global launch in summer via Avon’s 600,000 sales representatives in the United States and avon.com. Spotlight was described as a “fresh, oriental scent” by the company (Edgar, Michelle, December 5, 2008). The new campaigns may be coming at a good time. In a study of mass-market cosmetic brands, the research firm Brand Keys found that Avon lagged behind seven of their cosmetic companies in customer loyalty. Mary Kay Cosmetics (another direct marketer) was at the top of the list. Robert Passikoff, CEO of Brand Keys, suggested, “Avon’s problem is that it isn’t associated with anything in particular. It’s almost like a commodity.” 2 # Manufacturing and Logistics As of 2008, Avon’s largest manufacturing plants—in Brazil, China, and Poland—received ISO 14001 certification. This designate is the international standard for environmental management practices. The company’s plant in the Philippines has also received ISO 14001 certification. Avon’s manufacturing plant in Mexico received the Clean Industry Certificate from the federal Environmental Protection Agency in 2008. By the end of 2008, 60 percent of Avon’s production volume was being produced at sites where ISO 14001 certification has been achieved. Avon Brazil received an environmental award in 2008 from the Brazilian Benchmarking Environmental Program. Since 2003, this award has recognized environ-mental best practices among
{"source": 4976, "title": "from dpo"}
cost of 0.38 cpb. Associated data processing speed, which also determines MAC speed, is also 0.38 cpb for 1500 bytes. All of these figures assume 12-byte nonces and 16-byte tags. Advantages over GCM. AEZ has much stronger security properties than GCM. The later is not nonce-reuse secure, cannot safely generate short tags , and is not secure with respect to 17 Hoang, Krovetz, and Rogaway AEZ v3 disclosure of unverified plaintext. GCM does not achieve the RAE security definition. AEZ avoids GF (2128 ) multiplies (apart from the finite-field “doubling” that it uses). A closer match to AEZ in terms of high-level aims is SIV, which is at least nonce-reuse secure . But SIV has to output 128-bits more than its input; it is not RAE secure; and it is not parallelizable (although the last issue could easily be fixed). # 5 Design Rationale Enciphering-based AE. An old result had already shown that enciphering with a strong PRP provides a versatile route to AE . We recently came to understand just how attractive this route might be. On the one hand, we kept hearing requests for stronger AE security properties, like nonce-reuse misuse-resistance, authenticity without minimal ciphertext expansions, and security if unverified plaintexts are disclosed. Enciphering-based AE could deliver such aims. On the other hand, enciphering schemes that worked on either long or short strings were steadily becoming better-known objects. While they didn’t have the efficiency of OCB, say, neither were they computationally exorbitant. And there was the hope of doing better. Developing the enciphering scheme. With AES support increasingly embedded into devices, we wanted to base our enciphering scheme on the AES round function. A wide body of work had made abundantly clear that the best techniques for AES-based enciphering were going to depend on the length of the
{"source": 6303, "title": "from dpo"}
most of the mass comes from the large amount of energy associated with the strong interaction. Hadrons may also carry flavor quantum numbers such as isospin (G-parity), and strangeness. All quarks carry an additive, conserved quantum number called a baryon number (B), which is ++1⁄3 for quarks and −+1⁄3 for antiquarks. This means that baryons (composite particles made of three, five or a larger odd number of quarks) have B = 1 whereas mesons have B = 0. Hadrons have excited states known as resonances. Each ground state hadron may have several excited states; several hundred different resonances have been observed in experiments. Resonances decay extremely quickly (within about 10−24 seconds) via the strong nuclear force. In other phases of matter the hadrons may disappear. For example, at very high temperature and high pressure, unless there are sufficiently many flavors of quarks, the theory of quantum chromodynamics (QCD) predicts that quarks and gluons will no longer be confined within hadrons, "because the strength of the strong interaction diminishes with energy". This property, which is known as asymptotic freedom, has been experimentally confirmed in the energy range between 1 GeV (gigaelectronvolt) and 1 TeV (teraelectronvolt). All free hadrons except (possibly) the proton and antiproton are unstable. == Baryons == Baryons are hadrons containing an odd number of valence quarks (at least 3). Most well-known baryons such as the proton and neutron have three valence quarks, but pentaquarks with five quarks—three quarks of different colors, and also one extra quark-antiquark pair—have also been proven to exist. Because baryons have an odd number of quarks, they are also all fermions, i.e., they have half-integer spin. As quarks possess baryon number B = 1⁄3, baryons have baryon number B = 1. Pentaquarks also have B = 1, since the extra quark's and antiquark's baryon
{"page_id": 13821, "title": "Hadron"}
Buoyant density centrifugation (also isopycnic centrifugation or equilibrium density-gradient centrifugation) uses the concept of buoyancy to separate molecules in solution by their differences in density. == Implementation == Historically a cesium chloride (CsCl) solution was often used, but more commonly used density gradients are sucrose or Percoll. This application requires a solution with high density and yet relatively low viscosity, and CsCl suits it because of its high solubility in water, high density owing to the large mass of Cs, as well as low viscosity and high stability of CsCl solutions. The sample is put on top of the solution, and then the tube is spun at a very high speed for an extended time, at times lasting days. The CsCl molecules become densely packed toward the bottom, so a continuous gradient of layers of different densities (and CsCl concentrations) form. Since the original solution was approximately the same density, they go to a level where their density and the CsCl density are the same, to which they form a sharp, distinctive band. == Isotope separation == This method very sharply separates molecules, and is so sharp that it can even separate different molecular isotopes from one another. It has been utilized in the Meselson-Stahl experiment. == DNA separation == Buoyant density of the majority of DNA is 1.7g/cm3 which is equal to the density of 6M CsCl solution. Buoyant density of DNA changes with its GC content. The term "satellite DNA" refers to small bands of repetitive DNA sequences with distinct base composition floating above (A+T rich) or below (G+C rich) the main component DNA. == See also == Isopycnic Satellite DNA == References == == Further reading == Schildkraut, Carl L.; Marmur, Julius; Doty, Paul (1962). "Determination of the base composition of deoxyribonucleic acid from its buoyant density
{"page_id": 34248218, "title": "Buoyant density centrifugation"}
The σ-π model and equivalent-orbital model refer to two possible representations of molecules in valence bond theory. The σ-π model differentiates bonds and lone pairs of σ symmetry from those of π symmetry, while the equivalent-orbital model hybridizes them. The σ-π treatment takes into account molecular symmetry and is better suited to interpretation of aromatic molecules (Hückel's rule), although computational calculations of certain molecules tend to optimize better under the equivalent-orbital treatment. The two representations produce the same total electron density and are related by a unitary transformation of the occupied molecular orbitals; different localization procedures yield either of the two. Two equivalent orbitals h and h' can be constructed by taking linear combinations h = c1σ + c2π and h' = c1σ – c2π for an appropriate choice of coefficients c1 and c2. In a 1996 review, Kenneth B. Wiberg concluded that "although a conclusive statement cannot be made on the basis of the currently available information, it seems likely that we can continue to consider the σ/π and bent-bond descriptions of ethylene to be equivalent. Ian Fleming goes further in a 2010 textbook, noting that "the overall distribution of electrons [...] is exactly the same" in the two models. Nevertheless, as pointed out in Carroll's textbook, at lower levels of theory, the two models make different quantitative and qualitative predictions, and there has been considerable debate as to which model is most useful conceptually and pedagogically. == Multiple bonds == Two different explanations for the nature of double and triple covalent bonds in organic molecules were proposed in the 1930s. Linus Pauling proposed that the double bond in ethylene results from two equivalent tetrahedral orbitals from each atom, which later came to be called banana bonds or tau bonds. Erich Hückel proposed a representation of the double bond
{"page_id": 46398956, "title": "Sigma-pi and equivalent-orbital models"}
raised Roman Catholic, but he went to Jewish religious lessons during his school years. He later became an atheist "for moral reasons ... the world appears to me to be put together in such a painful way that I prefer to believe that it was not created ... intentionally". In later years he would call himself both an agnostic and an atheist. After the 1939 Soviet invasion of Poland's former eastern territory (now part of Ukraine and Belarus), he was not allowed to study at Lwów Polytechnic as he wished because of his "bourgeois origin", and only due to his father's connections he was accepted to study medicine at Lwów University in 1940. During the subsequent Nazi occupation (1941–1944), Lem's Jewish family avoided placement in the Nazi Lwów Ghetto, surviving with false papers. He would later recall: During that period, I learned in a very personal, practical way that I was no "Aryan". I knew that my ancestors were Jews, but I knew nothing of the Mosaic faith and, regrettably, nothing at all of Jewish culture. So it was, strictly speaking, only the Nazi legislation that brought home to me the realization that I had Jewish blood in my veins. During that time, Lem earned a living as a car mechanic and welder, and occasionally stole munitions from storehouses (to which he had access as an employee of a German company) to pass them on to the Polish resistance. In 1945, Lwów was annexed into the Soviet Ukraine, and the family, along with many other Polish citizens, was resettled to Kraków, where Lem, at his father's insistence, took up medical studies at the Jagiellonian University. He did not take his final examinations on purpose, to avoid the career of military doctor, which he suspected could have become lifelong. After receiving
{"page_id": 26790, "title": "Stanisław Lem"}
Donald in Mathmagic Land is an American live-action animated featurette produced by Walt Disney Productions and featuring Donald Duck. The short was directed by Hamilton Luske (with Wolfgang Reitherman, Les Clark, and Joshua Meador as sequence directors) and was released on June 26, 1959. It was nominated for an Academy Award for Best Documentary (Short Subject) at the 32nd Academy Awards, and became a widely viewed educational film in American schools of the 1960s and beyond. == Plot == The film begins with Donald Duck, holding a hunting rifle, as he passes through a doorway to find he has entered Mathmagic Land. This "mighty strange" fantasy land contains trees with square roots, a stream flowing with numbers, and a walking pencil that plays tic-tac-toe. A geometric bird recites (almost perfectly) the first 15 digits of pi. Donald soon hears the voice of the unseen "True Spirit of Adventure" (Paul Frees), who will guide him on his journey through "the wonderland of mathematics". Donald is initially not interested in exploring this foreign land, believing that mathematics is just for "eggheads" until "Mr. Spirit" suggests a fascinating connection between math and music. Intrigued, Donald discovers the relationships between octaves and string length which form the musical scale of today. Next, Donald finds himself in ancient Greece, where Pythagoras and his contemporaries are discovering these same relationships. Pythagoras (on the harp), a flute player, and a double bass player hold a "jam session" which Donald joins after a few moments using a vase as a bongo drum. Pythagoras' mathematical discoveries are, as the Spirit explains, the basis of today's music, and that no type of music could have ever existed without "eggheads." The segment ends with a sequence of live-action musicians playing both jazz and classical music and Pythagoras' acquaintances magically disappearing. After
{"page_id": 2153635, "title": "Donald in Mathmagic Land"}
waters. The oxygen content of these lakes is a function of their seasonally mixed hypolimnetic volume. Hypolimnetic volumes that are anoxic will result in fish congregating in areas where oxygen is sufficient for their needs. Anoxia is more common in the hypolimnion during the summer when mixing does not occur. In the absence of oxygen from the epilimnion, decomposition can cause hypoxia in the hypolimnion. === Mesotrophic === Mesotrophic lakes are lakes with an intermediate level of productivity. These lakes are commonly clear water lakes and ponds with beds of submerged aquatic plants and medium levels of nutrients. The term mesotrophic is also applied to terrestrial habitats. Mesotrophic soils have moderate nutrient levels. === Eutrophic and hypertrophic === ==== Eutrophic ==== A eutrophic water body, commonly a lake or pond, has high biological productivity. Due to excessive nutrients, especially nitrogen and phosphorus, these water bodies are able to support an abundance of aquatic plants. Usually, the water body will be dominated either by aquatic plants or algae. When aquatic plants dominate, the water tends to be clear. When algae dominate, the water tends to be darker. The algae engage in photosynthesis which supplies oxygen to the fish and biota which inhabit these waters. Occasionally, an excessive algal bloom will occur and can ultimately result in fish death, due to respiration by algae and bottom-living bacteria. The process of eutrophication can occur naturally and by human impact on the environment. Eutrophic comes from the Greek eutrophos meaning "well-nourished", from eu meaning good and trephein meaning "to nourish". ==== Hypertrophic ==== Hypertrophic or hypereutrophic lakes are very nutrient-rich lakes characterized by frequent and severe nuisance algal blooms and low transparency. Hypereutrophic lakes have a visibility depth of less than 90 centimetres (3 feet), they have greater than 40 micrograms/litre total chlorophyll and
{"page_id": 15819591, "title": "Trophic state index"}
Ultra-wideband (UWB, ultra wideband, ultra-wide band and ultraband) is a radio technology that can use a very low energy level for short-range, high-bandwidth communications over a large portion of the radio spectrum. The following is a list of devices that support the technology from various UWB silicon providers. == Smartphones, foldables, & tablets == == Smartwatches == == IoT devices == == References ==
{"page_id": 69076412, "title": "List of UWB-enabled mobile devices"}
in space. A relativistic solar neutron event was first reported for a 1982 event. == References ==
{"page_id": 23240438, "title": "Neutron monitor"}
of the Earth's mantle William Thomson, Lord Kelvin (Irish, 1824–1907) – influential estimate of the age of the Earth, ultimately proved incorrect == L == Kurt Lambeck (Dutch, 1941– ) – changed understanding of the ways post-glacial rebound affects ocean levels; awarded Wollaston Medal and Balzan Prize Johann von Lamont (Scottish, 1805–1879) – surveys of the Earth's magnetic field Louis J. Lanzerotti (American, 1938– ) – magnetosphere and ionosphere; awarded William Bowie Medal Joseph Larmor (Northern Irish, 1857–1942) – proton precession, dynamo theory Inge Lehmann (Danish, 1888–1993) – seismologist who discovered the Lehmann discontinuity and argued for a solid inner core Xavier Le Pichon (French, 1937–2025) – constructed history of plate motions Royden Charles Lilwall (British, 1944-2023) - Geophysicist / seismologist locating the epicenter of earthquakes Humphrey Lloyd (Irish, 1800–1881) – observational geomagnetism Cinna Lomnitz (Chilean–Mexican, 1925–2016) – creator of "Lomnitz Law", founder of Mexico's first seismic network and editor of Geofísica Internacional Andrew Long (Australian, 1965– ) – developed widely used instruments for marine exploration for oil and gas; Honorary Lecturer (Pacific South) for the Society of Exploration Geophysicists Augustus Edward Hough Love (English, 1863–1940) – developed theory of Love waves Bruce P. Luyendyk (American, 1943– ) – marine geophysics Leon Thomsen (American, 1942 ) - Seismic anisotropy Thomsen parameters (Awards: Maurice Ewing Medal, N.A.E.) == M == Gordon J. F. MacDonald (American, 1929–2002) – investigated rotation of the Earth and true polar wander James B. Macelwane (American, 1883–1956) – seismologist; awarded William Bowie Medal Jean-Jacques d'Ortous de Mairan (French, 1678 –1771) – shape of the Earth and aurora Robert Mallet (Irish, 1810–1881) – developed controlled source seismology; coined terms seismology and epicenter Syukuro Manabe (Japanese, 1931– ) – climate models; awarded William Bowie Medal Pierre de Maricourt (Petrus Peregrinus) (French, fl. 1269) – first extant treatise on properties
{"page_id": 209994, "title": "List of geophysicists"}
⁡ θ v s ) {\displaystyle \theta _{s}=\arccos \left({\frac {v\cos \theta }{v_{s}}}\right)} Final vs, θs and r must match the requirements of the target orbit as determined by orbital mechanics (see Orbital flight, above), where final vs is usually the required periapsis (or circular) velocity, and final θs is 90 degrees. A powered descent analysis would use the same procedure, with reverse boundary conditions. == Orbital flight == Orbital mechanics are used to calculate flight in orbit about a central body. For sufficiently high orbits (generally at least 190 kilometers (100 nautical miles) in the case of Earth), aerodynamic force may be assumed to be negligible for relatively short term missions (though a small amount of drag may be present which results in decay of orbital energy over longer periods of time.) When the central body's mass is much larger than the spacecraft, and other bodies are sufficiently far away, the solution of orbital trajectories can be treated as a two-body problem. This can be shown to result in the trajectory being ideally a conic section (circle, ellipse, parabola or hyperbola) with the central body located at one focus. Orbital trajectories are either circles or ellipses; the parabolic trajectory represents first escape of the vehicle from the central body's gravitational field. Hyperbolic trajectories are escape trajectories with excess velocity, and will be covered under Interplanetary flight below. Elliptical orbits are characterized by three elements. The semi-major axis a is the average of the radius at apoapsis and periapsis: a = r a + r p 2 {\displaystyle a={\frac {r_{a}+r_{p}}{2}}} The eccentricity e can then be calculated for an ellipse, knowing the apses: e = r a a − 1 {\displaystyle e={\frac {r_{a}}{a}}-1} The time period for a complete orbit is dependent only on the semi-major axis, and is independent of
{"page_id": 1559922, "title": "Spacecraft flight dynamics"}
; (21) Procedures for parents to review and discuss with the child -care center director any questions or concerns about the policies and procedures of the child -care center [Medium] ;Revised: April 2025 35 (22) Procedures for parents to participate in the child -care center’s operation and activities [Medium -Low] ; (23) Instructions on how a parent may access the [Medium] : (A) Minimum standards online; (B) Texas Abuse and Neglect Hotline; and (C) HHSC website. (24) Your emergency preparedness plan [High] ; (25) Your provisions to provide a comfortable place with an adult sized seat in your center or within a classroom that enables a mother to breastfeed her child. In addition, your policies must inform parents that they have the right to breastfeed or provide breast milk for their child while in care [Low] ; (26) Preventing and responding to abuse and neglect of children, including: (A) Required annual training for employees [Medium] ; (B) Methods for increasing employee and parent awareness of issues regarding child abuse and neglect, including warning signs that a child may be a victim of abuse or neglect and factors indicating a child is at risk for abuse or neglect [Medium] ; (C) Methods for increasing employee and parent awareness of prevention techniques for child abuse and neglect [Medium] ; (D) Strategies for coordination between the center and appropriate community organizations [Medium] ; and (E) Actions that the parent of a child who is a victim of abuse or neglect should take to obtain assistance and intervention, including procedures for reporting child abuse or neglect [Medium] ; (27) Procedures for conducting health checks, if applicable [Medium] ; and (28) Information on vaccine -preventable diseases for employees, unless your center is in the home of the permit holder. The policy must address the requirements
{"source": 1426, "title": "from dpo"}
following concentrations are found in mixtures of ions in equilibrium with slightly soluble solids. From the concentrations given, calculate Ksp for each of the slightly soluble solids indicated: (a) TlCl: [Tl +] = 1.21 × 10 –2 M, [Cl –] = 1.2 × 10 –2 M (b) Ce(IO 3)4: [Ce 4+ ] = 1.8 × 10 –4 M, [IO 3 −] = 2.6 × 10 –13 M (c) Gd 2(SO 4)3: [Gd 3+ ] = 0.132 M, [SO 4 2− ] = 0.198 M (d) Ag 2SO 4: [Ag +] = 2.40 × 10 –2 M, [SO 4 2− ] = 2.05 × 10 –2 M (e) BaSO 4: [Ba 2+ ] = 0.500 M, [SO 4 2− ] = 2.16 × 10 –10 M 30. Which of the following compounds precipitates from a solution that has the concentrations indicated? (See Appendix J for Ksp values.) (a) KClO 4: [K +] = 0.01 M, [ClO 4 −] = 0.01 M (b) K 2PtCl 6: [K +] = 0.01 M, [PtCl 6 2− ] = 0.01 M (c) PbI 2: [Pb 2+ ] = 0.003 M, [I –] = 1.3 × 10 –3 M (d) Ag 2S: [Ag +] = 1 × 10 –10 M, [S 2– ] = 1 × 10 –13 M 31. Which of the following compounds precipitates from a solution that has the concentrations indicated? (See Appendix J for Ksp values.) (a) CaCO 3: [Ca 2+ ] = 0.003 M, [CO 3 2− ] = 0.003 M (b) Co(OH) 2: [Co 2+ ] = 0.01 M, [OH –] = 1 × 10 –7 M (c) CaHPO 4: [Ca 2+ ] = 0.01 M, [HPO 4 2− ] = 2 × 10 –6 M (d) Pb 3(PO 4)2: [Pb 2+ ] = 0.01 M, [PO 4 3− ]
{"source": 3700, "title": "from dpo"}
module validation. TE02.20.01 shall explain how each CVL meets this IG and is used in an approved manner by the module. 1. An ECC CDH primitive from section 5.7.1.2 of SP 800-56Arev3 . The test performs the multiplication of a point on a NIST-recommended elliptic curve by an integer and verifies that the x-coordinate of the resulting point has been computed correctly. The integer in question is defined in SP 800-56Arev3 as the product of the tested party’s private key d A and the curve’s co-factor h. Usage Restriction: Shall only be used within the context of a SP 800-56Arev3 KAS (enforced by the module in that this CVL is not used internally for other purposes; but if external / operator-accessible enforced by operator guidance within the Security Policy, e.g., “the KAS-ECC CDH-Component (CVL) shall only be used within the context of an SP 800-56Arev3 KAS”). CAVP certificate and denotation: Within the context of the CAVP ACVTS, this component test is identified by the following algorithm/mode/revision capability combination: • “algorithm”: “KAS-ECC”, • “mode”: “CDH-Component”, • “revision”: ”Sp800-56Ar3”. This testing is denoted by “KAS-ECC CDH-Component SP800-56Ar3” on the CAVP webpage: .Implementation Guidance for FIPS PUB 140-3 and the Cryptographic Module Validation Program National Institute of Standards and Technology CMVP 28 04/18/2025 2. An RSA signature generation per FIPS 186-5 without the computation of a hash which is presumed to have already been computed. This RSA test verifies the ability of a module to perform the RSASP1 signature primitive described in PKCS#1 v2.1: RSA Cryptography Standard, RSA Laboratories, June 14, 2002, Section 5.2.1. RSASP1 takes as input an RSA private key and a message digest and outputs a signature representative or an error message. The test 1) verifies the ability of the module to detect an invalid message representative and 2),
{"source": 5751, "title": "from dpo"}
Rm×n is called an integer matrix if every entry of A is an integer. Suppose that A ∈ Rn×n is an invertible integer matrix. Then det( A) is a non-zero integer and (Cof( A)) T is an integer matrix. If A−1 is also an integer matrix then det( A−1) is also an integer. Now det( A) det( A−1) = 1 thus it must be the case that det( A) = ±1. Suppose on the other hand that det( A) = ±1. Then by the Cofactor method A−1 = 1 det( A)(Cof( A)) T = ±(Cof( A)) T and therefore A−1 is also an integer matrix. We have proved the following. Theorem 13.2: An invertible integer matrix A ∈ Rn×n has an integer inverse A−1 if and only if det A = ±1. We can use the previous theorem to generate integer matrices with an integer inverse as follows. Begin with an upper triangular matrix M0 having integer entries and whose diagonal entries are either 1 or −1. By construction, det( M0) = ±1. Perform any sequence of elementary row operations of Type 1 and Type 3. This generates a sequence of matrices M1, . . . , Mp whose entries are integers. Moreover, M0 ∼ M1 ∼ · · · ∼ Mp. Therefore, ±1 = det( M) = det( M1) = · · · = det( Mp). > 105 Applications of the Determinant # 13.2 Cramer’s Rule The Cofactor method can be used to give an explicit formula for the solution of a linear system where the coefficient matrix is invertible. The formula is known as Cramer’s Rule. To derive this formula, recall that if A is invertible then the solution to Ax = b is x = A−1b.Using the Cofactor method, A−1 = 1 > det A (Cof( A))
{"source": 6828, "title": "from dpo"}
for use as a shear-thinning lubricant. == See also == History of cell membrane theory Lipid polymorphism Lipid bilayer Micelle Unilamellar liposome == References == == External links == Bilayer formation through molecular self-assembly - YouTube Cell Membrane - The Lipid Bilayer - YouTube
{"page_id": 43275893, "title": "Lamellar phase"}
⟨ ψ b | ( q x ) | ψ a ⟩ = q ∫ ψ b ∗ ( r ) x ψ a ( r ) d 3 r {\displaystyle ({\text{x-component of t.d.m. }}a\rightarrow b)=\langle \psi _{b}|(qx)|\psi _{a}\rangle =q\int \psi _{b}^{*}(\mathbf {r} )\,x\,\psi _{a}(\mathbf {r} )\,d^{3}\mathbf {r} } In other words, the transition dipole moment can be viewed as an off-diagonal matrix element of the position operator, multiplied by the particle's charge. === Multiple charged particles === When the transition involves more than one charged particle, the transition dipole moment is defined in an analogous way to an electric dipole moment: The sum of the positions, weighted by charge. If the ith particle has charge qi and position operator ri, then the transition dipole moment is: ( t.d.m. a → b ) = ⟨ ψ b | ( q 1 r 1 + q 2 r 2 + ⋯ ) | ψ a ⟩ = ∫ ψ b ∗ ( r 1 , r 2 , … ) ( q 1 r 1 + q 2 r 2 + ⋯ ) ψ a ( r 1 , r 2 , … ) d 3 r 1 d 3 r 2 ⋯ {\displaystyle {\begin{aligned}({\text{t.d.m. }}a\rightarrow b)&=\langle \psi _{b}|(q_{1}\mathbf {r} _{1}+q_{2}\mathbf {r} _{2}+\cdots )|\psi _{a}\rangle \\&=\int \psi _{b}^{*}(\mathbf {r} _{1},\mathbf {r} _{2},\ldots )\,(q_{1}\mathbf {r} _{1}+q_{2}\mathbf {r} _{2}+\cdots )\,\psi _{a}(\mathbf {r} _{1},\mathbf {r} _{2},\ldots )\,d^{3}\mathbf {r} _{1}\,d^{3}\mathbf {r} _{2}\cdots \end{aligned}}} === In terms of momentum === For a single, nonrelativistic particle of mass m, in zero magnetic field, the transition dipole moment between two energy eigenstates ψa and ψb can alternatively be written in terms of the momentum operator, using the relationship ⟨ ψ a | r | ψ b ⟩ = i ℏ ( E b − E a ) m
{"page_id": 9812205, "title": "Transition dipole moment"}
The International Human Epigenome Consortium (IHEC) is a scientific organization, founded in 2010, that helps to coordinate global efforts in the field of Epigenomics. The initial goal was to generate at least 1,000 reference (baseline) human epigenomes from different types of normal and disease-related human cell types. == Structure and funding == IHEC's operations are funded by its full members (national and regional scientific funding agencies), and staffed largely on a volunteer basis by scientists and other experts from participating funding agencies and epigenome mapping projects. === Current IHEC Member Countries === Canadian Institutes for Health Research (Canada) European Commission (EU) Federal Ministry of Education and Research, Project Management Agency within the German Aerospace Centre (Germany) Genome Institute of Singapore (Singapore) Hong Kong Epigenomics Project (Hong Kong) Japan Agency for Medical Research and Development (Japan) National Institute of Health (South Korea) National Institutes of Health: Roadmap Epigenomics Program and Encyclopedia of DNA Elements (ENCODE) project (USA) In addition, countries and agencies supportive of IHEC goals are organizations that have not yet made a full financial contribution to the project, but whose members provide time and expertise: European Institute of Oncology, FIRC Institute of Molecular Oncology Foundation, Italian Institute of Technology, Centre for Genomic Science (Italy) National Agency of Research (France) National Health and Medical Research Council (Australia) UK Funders Alliance: Medical Research Council, Biotechnology and Biological Sciences Research Council, Cancer Research UK, and Wellcome Trust (UK) Oversight of IHEC is provided by an executive committee, whose members are nominated by Full Member organizations. This committee works closely with an International Scientific Steering Committee, whose members are the scientific leaders of participating projects and other leaders in the field of epigenetics, as well as a Data Coordination Center. Additional expertise is contributed by workgroups composed of members of participating research projects.
{"page_id": 40641489, "title": "International Human Epigenome Consortium"}
AppleTalk Remote Access (ARA) was a protocol stack that allowed AppleTalk to be run over modems. It became a fairly major product for Apple Computer in the early to mid-1990s when their first portable and laptop computers were available (and very popular). ARA slowly disappeared in the late 1990s when TCP/IP took over the vast majority of networking needs, notably remote access. Most networking protocols have strong "layering" that separates the various jobs inside the protocol into different pieces of software. This allows them to be run on top of any hardware by replacing the lowest level of the "stack", the hardware drivers. For instance, IP can be made to run on a variety of Ethernet cards or even Token Ring with little effort. For slower speeds, like on modems, things become somewhat more difficult, as the protocols often have invisible assumptions about timing and performance that make it inefficient on links with very limited bandwidth. AppleTalk included several protocols that made this even more difficult. In particular, AppleTalk had several internal tasks for discovery and naming that ran all the time and made the protocol "chatty". This added to the bandwidth problems, making it even less efficient in this case. Thus ARA was considerably more complex than similar solutions for IP, replacing many parts of the AppleTalk stack and seriously modifying others. As a result, ARA was quite large, larger than the basic AppleTalk stack, and somewhat memory hungry. It was also slow, a problem it shared with IP at similar speeds. Nevertheless, ARA was the only protocol available for remote access on the Mac, and also shared the typical Apple properties of being easy to install, set up and run. It became a fairly profitable product on its own, and was sold widely in stores. The introduction of
{"page_id": 1325856, "title": "AppleTalk Remote Access"}
contains a hidden Recorded TV library that appears in the Windows Explorer sidepane when TV is set up in Media Center for the first time. In addition to aggregating multiple storage locations, Libraries enable Arrangement Views and Search Filter Suggestions. Arrangement Views allow you to pivot your view of the library's contents based on metadata. For example, selecting the "By Month" view in the Pictures library will display photos in stacks, where each stack represents a month of photos based on the date they were taken. In the Music library, the "By Artist" view will display stacks of albums from the artists in your collection, and browsing into an artist stack will then display the relevant albums. Search Filter Suggestions are a new feature of the Windows 7 Explorer's search box. When the user clicks in the search box, a menu shows up below it showing recent searches as well as suggested Advanced Query Syntax filters that the user can type. When one is selected (or typed in manually), the menu will update to show the possible values to filter by for that property, and this list is based on the current location and other parts of the query already typed. For example, selecting the "tags" filter or typing "tags:" into the search box will display the list of possible tag values which will return search results. Arrangement Views and Search Filter Suggestions are database-backed features which require that all locations in the Library be indexed by the Windows Search service. Local disk locations must be indexed by the local indexer, and Windows Explorer will automatically add locations to the indexing scope when they are included in a library. Remote locations can be indexed by the indexer on another Windows 7 machine, on a Windows machine running Windows Search 4
{"page_id": 19990354, "title": "Features new to Windows 7"}
The commitment decreases when the expectation to achieve is decreased. From: Psychology and Work Today by Schultz and Schultz. Feedback while the employee or group is striving for the goal is seen as crucial. Feedback keeps employees on track and reinforces the importance of the goal as well as supporting the employees in adjusting their task strategies. Goal-setting Theory has strong empirical support dating back thirty years. However, there are some boundary conditions that indicate in some situations, goal-setting can be detrimental to performance on certain types of tasks. Goals require a narrowing of one's focus, so for more complex or creative tasks, goals can actually inhibit performance because they demand cognitive resources. Similarly, when someone is learning a new task, performance-related goals can distract from the learning process. During the learning process, it may be better to focus on mastering the task than achieving a particular result. Finally, too many goals can become distracting and counterproductive, especially if they conflict with one another. ==== Social cognitive theory ==== Bandura's Social Cognitive Theory is another cognitive process theory that offers the important concept of self-efficacy for explaining employee's level of motivation relative to workplace tasks or goals. Self-efficacy is an individual's belief in their ability to achieve results in a given scenario. Empirically, studies have shown a strong correlation between self-efficacy and performance. The concept has been extended to group efficacy, which is a group's belief that it can achieve success with a given task or project. Self-efficacy is seen to mediate important aspects of how an employee undertakes a given task, such as the level of effort and persistence. An employee with high self-efficacy is confident that effort they put forth has a high likelihood of resulting in success. In anticipation of success, an employee is willing to put
{"page_id": 35231573, "title": "Work motivation"}
habitability == Based on tidal heating models, scientists have defined zones in satellite systems similarly to those of planetary systems. One such zone is the circumplanetary habitable zone (or "habitable edge"). According to this theory, moons closer to their planet than the habitable edge cannot support liquid water at their surface. When effects of eclipses as well as constraints from a satellite's orbital stability are included into this concept, one finds that — depending on a moon's orbital eccentricity — there is a minimum mass of roughly 0.2 solar masses for stars to host habitable moons within the stellar HZ. The magnetic environment of exomoons, which is critically triggered by the intrinsic magnetic field of the host planet, has been identified as another effect on exomoon habitability. Most notably, it was found that moons at distances between about 5 and 20 planetary radii from a giant planet can be habitable from an illumination and tidal heating point of view, but still the planetary magnetosphere would critically influence their habitability. == See also == Minor-planet moon Subsatellite Double planet Quasi-satellite List of extraterrestrial orbiters == Notes == == References ==
{"page_id": 46326758, "title": "Satellite system (astronomy)"}
information content of the image, provided the image entropy does not exceed the threshold where 2:1 compression is reached. Due to camera problems and the need to reduce file size, there is a slight modification to the image coding scheme so that each compressed line is effectively bandwidth limited on the number of bits available to encode it. == References ==
{"page_id": 446876, "title": "Adaptive coding"}
to combine with other programs. This style has been called the use of software tools, and depends more on how the programs fit into the programming environment and how they can be used with other programs than on how they are designed internally. [...] This style was based on the use of tools: using programs separately or in combination to get a job done, rather than doing it by hand, by monolithic self-sufficient subsystems, or by special-purpose, one-time programs. The authors contrast Unix tools such as cat with larger program suites used by other systems. The design of cat is typical of most UNIX programs: it implements one simple but general function that can be used in many different applications (including many not envisioned by the original author). Other commands are used for other functions. For example, there are separate commands for file system tasks like renaming files, deleting them, or telling how big they are. Other systems instead lump these into a single "file system" command with an internal structure and command language of its own. (The PIP file copy program found on operating systems like CP/M or RSX-11 is an example.) That approach is not necessarily worse or better, but it is certainly against the UNIX philosophy. === Doug McIlroy on Unix programming === McIlroy, then head of the Bell Labs Computing Sciences Research Center, and inventor of the Unix pipe, summarized the Unix philosophy as follows: This is the Unix philosophy: Write programs that do one thing and do it well. Write programs to work together. Write programs to handle text streams, because that is a universal interface. Beyond these statements, he has also emphasized simplicity and minimalism in Unix programming: The notion of "intricate and beautiful complexities" is almost an oxymoron. Unix programmers vie with each
{"page_id": 223176, "title": "Unix philosophy"}
not dominated by tight loops (§2.2). However, the performance gains are limited, as observed in the RocksDB results. This is because compiler instrumentation only reduces the overhead of the preemp-tion mechanism and does not address the context switching overhead (§4.2). Overall, we conclude that user interrupts are likely a better choice for the majority of applications. ## 4.2 The Cost of Preemption In Figure 1 we measured the cost of preempting a thread, executing an empty handler function, and then immediately returning to the same thread ( preemption cost ). In this section, we also report the context-switch cost of preemption, which in-cludes the costs of traversing the scheduler, context switching to a different thread, and any resulting cache contention. To-gether, the preemption cost and context-switch cost represent the total overhead incurred per preemption. Figure 8 illustrates the preemption and context-switch costs for various preemption mechanisms, applications, and task combinations. Each application uses its optimal quantum as rmv ppo kmeans rmv + ppo > kmeans > ad + decay + rmv > ppo + kmeans > 0 > 1 > 2 > 3 > 4 > 5 > Average Cost ( μs) > Signals (context-switch cost) > UINTR (context-switch cost) > Compiler (context-switch cost) > Signals (preemption cost) > UINTR (preemption cost) > Compiler (preemption cost) > (a) Aspen-KB: C++ Dataframe (20 μs quantum) scan scan × 8 get × 8 + > scan × 8 > 0 > 1 > 2 > 3 > 4 > 5 > Average Cost ( μs) > (b) Aspen-KB: RocksDB (5 μs) rangescan 8 × rangescan 8 × get + > 8 × rangescan > 0 > 2 > 4 > 6(c) Aspen-Go: BadgerDB (50 μs) Figure 8: The costs of preemption ( preemption cost ) and context switching ( context-switch cost ),
{"source": 44, "title": "from dpo"}
the same result. For the following questions, assume the query–response time curve shown in Figure 6.24. The curve shows the latency of response, based on the number of queries per second, for a baseline server as well as a “small” server that uses a slower processor model. a. How many servers are required to achieve that SLA, assuming that the WSC receives 30,000 queries per second, and the query– response time curve shown in Figure 6.24? How many “small” servers are required to achieve that SLA, given this response-time probability curve? Looking only at server costs, how much cheaper must the “wimpy” servers be than the normal servers to achieve a cost advantage for the target SLA? b. Often “small” servers are also less reliable due to cheaper components. Using the numbers from Figure 6.1, assume that the number of events due to flaky machines and bad memories increases by 30%. How many “small” servers are required now? How much cheaper must those serv-ers be than the standard servers? 482 ■ Chapter Six Warehouse-Scale Computers to Exploit Request-Level and Data-Level Parallelism c. Now assume a batch processing environment. The “small” servers provide 30% of the overall performance of the regular serv-ers. Still assuming the reliability numbers from Exercise 6.15 part (b), how many “wimpy” nodes are required to provide the same expected throughput of a 2400-node array of standard servers, assuming perfect linear scaling of performance to node size and an average task length of 10 minutes per node? What if the scaling is 85%? 60%? d. Often the scaling is not a linear function, but instead a logarithmic function. A natural response may be instead to purchase larger nodes that have more computational
{"source": 2299, "title": "from dpo"}
every two of which are linked by a direct road. A matrix T = ( ti,j ) is given, where ti,j is time required to travel from i to j (not necessarily symmetric). Find a path between 1 and N which consumes minimum time. Bellman remarked: > Documenta Mathematica · Extra Volume ISMP (2012) 155–167 On the History of the Shortest Path Problem 161 Since there are only a finite number of paths available, the problem reduces to choosing the smallest from a finite set of numbers. This direct, or enumerative, approach is impossible to execute, however, for values of N of the order of magnitude of 20. He gave a ‘functional equation approach” The basic method is that of successive approximations. We choose an initial sequence {f (0) > i }, and then proceed iteratively, setting f (k+1) > i = Min > j6=i (tij + f (k) > j ], i = 1 , 2, · · · , N − 1,f (k+1) > N = 0 , for k = 0 , 1, 2 · · · , .As initial function f (0) > i Bellman proposed (upon a suggestion of F. Haight) to take f (0) > i = ti,N for all i. Bellman noticed that, for each fixed i, starting with this choice of f (0) > i gives that f (k) > i is monotonically nonincreasing in k, and stated: It is clear from the physical interpretation of this iterative scheme that at most ( N − 1) iterations are required for the sequence to converge to the solution. Since each iteration can be done in time O(N 2), the algorithm takes time O(N 3). As for the complexity, Bellman said: It is easily seen that the iterative scheme discussed above is a fea-sible
{"source": 4215, "title": "from dpo"}
(Followup to Olofsson & Lundberg, 1983) * Reitsma, P., & Wesseling, R. (1998). Effects of computer-assisted training of blending skills in kindergartners. Scientific Studies of Reading, 2 ,301-320. * Sanchez, E., & Rueda, M. (1991). Segmental awareness and dyslexia: Is it possible to learn to segment well and yet continue to read and write poorly? Reading and Writing: An Interdisciplinary Journal, 3 ,11-18. * Schneider, W., Kuspert, P., Roth, E., Vise, M., & Marx, H. (1997). Short- and long-term effects of training phonological awareness in kindergarten: Evidence from two German studies. Journal of Experimental Child Psychology, 66 , 311-340. * Solity, J. (1996). Phonological awareness: Learning disabilities revisited? Educational & Child Psychology, 13 , 103-113. * Tangel, D., & Blachman, B. (1992). Effect of phoneme awareness instruction on kindergarten children’s invented spelling. Journal of Reading Behavior, 24 , 233-261. * Torgesen, J., Morgan, S., & Davis, C. (1992). Effects of two types of phonological awareness training on word learning in kindergarten children. Journal of Educational Psychology, 84 , 364-370. * Treiman, R., & Baron, J. (1983). Phonemic-analysis training helps children benefit from spelling sound rules. Memory and Cognition, 11 ,382-389. * Uhry, J., & Shepherd, M. (1993). Segmentation/spelling instruction as part of a first-grade reading program: Effects on several measures of reading. Reading Research Quarterly, 28 , 218-233. * Vadasy, P., Jenkins, J., Antil, L., Wayne, S., & O’Connor, R. (1997). Community-based early reading intervention for at-risk first graders. Learning Disabilities Research and Practice, 12 , 29-39. * Vadasy, P., Jenkins, J., Antil, L., et al. (1997). The effectiveness of one-to-one tutoring by community tutors for at-risk beginning readers. Learning Disability Quarterly, 20 , 126-139. * Vellutino, F., & Scanlon, D. (1987). Phonological coding, phonological awareness, and reading ability: Evidence from a longitudinal and experimental study. Merrill-Palmer Quarterly, 33 ,321-363.
{"source": 6100, "title": "from dpo"}
analysis. Given two natural numbers, k>r>0, a training example is called a (k,r)NN class-outlier if its k nearest neighbors include more than r examples of other classes. === Condensed Nearest Neighbor for data reduction === Condensed nearest neighbor (CNN, the Hart algorithm) is an algorithm designed to reduce the data set for k-NN classification. It selects the set of prototypes U from the training data, such that 1NN with U can classify the examples almost as accurately as 1NN does with the whole data set. Given a training set X, CNN works iteratively: Scan all elements of X, looking for an element x whose nearest prototype from U has a different label than x. Remove x from X and add it to U Repeat the scan until no more prototypes are added to U. Use U instead of X for classification. The examples that are not prototypes are called "absorbed" points. It is efficient to scan the training examples in order of decreasing border ratio. The border ratio of a training example x is defined as where ‖x-y‖ is the distance to the closest example y having a different color than x, and ‖x'-y‖ is the distance from y to its closest example x' with the same label as x. The border ratio is in the interval [0,1] because ‖x'-y‖ never exceeds ‖x-y‖. This ordering gives preference to the borders of the classes for inclusion in the set of prototypes U. A point of a different label than x is called external to x. The calculation of the border ratio is illustrated by the figure on the right. The data points are labeled by colors: the initial point is x and its label is red. External points are blue and green. The closest to x external point is y. The
{"page_id": 1775388, "title": "K-nearest neighbors algorithm"}
to exist in the present. The relationship to the future is the state of anticipating a potential possibility, task, or engagement. It is related to the human propensity for caring and being concerned, which causes "being ahead of oneself" when thinking of a pending occurrence. Therefore, this concern for a potential occurrence also allows the future to exist in the present. The present becomes an experience, which is qualitative instead of quantitative. Heidegger seems to think this is the way that a linear relationship with time, or temporal existence, is broken or transcended. We are not stuck in sequential time. We are able to remember the past and project into the future; we have a kind of random access to our representation of temporal existence; we can, in our thoughts, step out of (ecstasis) sequential time. Modern era philosophers asked: is time real or unreal, is time happening all at once or a duration, is time tensed or tenseless, and is there a future to be? There is a theory called the tenseless or B-theory; this theory says that any tensed terminology can be replaced with tenseless terminology. For example, "we will win the game" can be replaced with "we do win the game", taking out the future tense. On the other hand, there is a theory called the tense or A-theory; this theory says that our language has tense verbs for a reason and that the future can not be determined. There is also something called imaginary time, this was from Stephen Hawking, who said that space and imaginary time are finite but have no boundaries. Imaginary time is not real or unreal, it is something that is hard to visualize. Philosophers can agree that physical time exists outside of the human mind and is objective, and psychological time
{"page_id": 30012, "title": "Time"}
power source, environmental effects such as biofouling, ingestion by a predator, its depth limitation and cost. Most PSATS have internal software designed to detect damaging or sub-optimal conditions that will trigger an early release and transmission of data. For example, PSATs can withstand pressures to depths of 2,000 to 2,500 metres (6,600 to 8,200 ft) depending on the model. If data indicate no change in pressure (depth) for a period of time, this could trigger an early release due to premature release (a tag pulling out of the fish early) or death of the animal to which it was attached. Such internal checks can alert researchers to unexpected or undesirable events. Ingestion by a predator is more difficult to detect in the sense of forcing a tag to report; however, in data processing it is indicated by an immediate loss of light and an increase in temperature that stabilizes while it is inside the predator. == Types == === Using light level === The most popular method of determining an animal's location underwater requires the tag to acquire light levels throughout the day. Observing the length of the day, from when the tag observed the first light until the last light, the tag can determine its latitudinal location (with accuracy exceeding 1 degree). From the length of day the tag computes the noon time which is converted to a longitude location (with accuracy averaging about 0.5 degree or 30–50 nautical miles). This method of geolocation is suitable for animals that inhabit clear waters near the surface. At depths or in turbid waters, light based geolocation does not work as well due to light attenuation. It also does not work well during the equinoxes when the length of day is globally uniform. Manufacturers of this technology include Wildlife Computers, Microwave Telemetry,
{"page_id": 24370288, "title": "Pop-up satellite archival tag"}
the original hydroxyapatite of the tooth enamel. Minerals include: Calcium Phosphate Fluoride === Buffering agents === Buffering agents in artificial saliva maintain the pH levels, ensuring the oral environment remains within the optimal range for enamel protection and microbial balance. A stable pH prevents tooth decay and maintains oral health. These buffering agents neutralise acidic substances in the oral cavity, safeguarding tooth enamel and soft tissues from acid-related damage. Buffering agents include: Sodium phosphate Potassium phosphate === Flavouring agents === Flavouring agents enhance the taste and acceptability of artificial saliva. These agents can mask any unpleasant flavours inherent in the other components and make the user's experience of using artificial saliva more pleasant. Flavouring agents encourage the user's compliance with the treatment regimen, mainly when long-term use is necessary. Flavouring agents include: Xylitol Sorbitol Citric acid == Types == Artificial saliva is available in different types and varies in chemical-physical properties, including viscosity, pH, buffering capacity, superficial tension, density and spinnbarkeit. Types of artificial saliva include: Oral liquid Oral spray Lozenge Gel Oil Mouthwash Toothpaste == Research == Research on artificial saliva focuses on replicating the biological components of natural saliva to substitute and enhance essential functions such as aiding digestion, performing antimicrobial action and protecting tissue layers. Approaches include enhancing or mimicking salivary proteins in natural saliva. === Enhancing salivary proteins === Salivary proteins such as histatin, statherin and mucin, which possess antimicrobial, lubrication and biomineralisation properties, are targets of enhancement in artificial saliva research as they are significant in maintaining oral health. Genetic engineering and recombinant DNA techniques produce recombinant salivary proteins for enhanced biological functions. Identification of recombinant histatin and statherin strains is more prevalent as they are smaller and simpler in structure. Therefore, a range of histatin and statherin variants with duplication of active protein domain
{"page_id": 76564427, "title": "Artificial saliva"}
code static checkers or bounded model checking for functional verification purposes, and also assist in determination of code timing properties. === High vs. low volume === For high-volume systems such as mobile phones, minimizing cost is usually the primary design consideration. Engineers typically select hardware that is just good enough to implement the necessary functions. For low-volume or prototype embedded systems, general-purpose computers may be adapted by limiting the programs or by replacing the operating system with an RTOS. == Embedded software architectures == In 1978 National Electrical Manufacturers Association released ICS 3-1978, a standard for programmable microcontrollers, including almost any computer-based controllers, such as single-board computers, numerical, and event-based controllers. There are several different types of software architecture in common use. === Simple control loop === In this design, the software simply has a loop which monitors the input devices. The loop calls subroutines, each of which manages a part of the hardware or software. Hence it is called a simple control loop or programmed input-output. === Interrupt-controlled system === Some embedded systems are predominantly controlled by interrupts. This means that tasks performed by the system are triggered by different kinds of events; an interrupt could be generated, for example, by a timer at a predefined interval, or by a serial port controller receiving data. This architecture is used if event handlers need low latency, and the event handlers are short and simple. These systems run a simple task in a main loop also, but this task is not very sensitive to unexpected delays. Sometimes the interrupt handler will add longer tasks to a queue structure. Later, after the interrupt handler has finished, these tasks are executed by the main loop. This method brings the system close to a multitasking kernel with discrete processes. === Cooperative multitasking === Cooperative
{"page_id": 46630, "title": "Embedded system"}