text
stringlengths
5
10.5k
source
stringlengths
33
146
Language has been an integral component of the psychoanalytic framework since its inception, and psychoanalytic theory intersects with linguistics and psycholinguistics. Language is relevant to psychoanalysis in two key respects. First, it is important with respect to the supposed therapeutic process, serving as the principal means by which unconscious mental processes are given expression through the verbal exchange between analyst and patient, e.g. free association, dream analysis, transference-countertransference dynamics. Secondly, psychoanalytic theory is linked in many ways to linguistic phenomena, such as parapraxes. According to Sigmund Freud the essential difference between modes of thought characterized by primary (irrational, governed by the id) as opposed to secondary (logical, governed by the ego and external reality) thought processes is one of preverbal vs. verbal ways of conceptualizing the world. == Freud's ideas on language == In 1940 Sigmund Freud wrote "...the function of speech brings material in the ego into a firm connection with the residues of visual, but more particularly of auditory, perceptions." == Aphasia, thing- and word-presentations == One of Freud's earliest papers, On Aphasia (1891), was concerned with speech disorders of neurological mechanisms of which had been investigated earlier in the century by Paul Broca and Carl Wernicke. Freud was skeptical of Wernicke's findings, citing a paucity of clinical observation as his reason. Although he conceded the fact that language is linked to neurological processes, Freud repudiated a model of localization of brain function, according to which specific regions of the brain are responsible for certain cognitive functions. In contrast to most of his contemporaries, Freud rejected the notion that in most cases pathological phenomena are manifestations of physiological dysfunctions (Lanteri-Laura, 2005a). In the same paper, Freud (1891) distinguishes between word-presentations, the mental images of words, and thing-presentations, the representations of actual objects. Word-presentations involve the linking of a
{"page_id": 14551867, "title": "Psychoanalytic conceptions of language"}
of speech back into sound. With this device, Alvin Liberman, Cooper, and Pierre Delattre (and later joined by Katherine Safford Harris, Leigh Lisker, Arthur Abramson, and others), discovered the acoustic cues for the perception of phonetic segments (consonants and vowels). Liberman and colleagues proposed a motor theory of speech perception to resolve the acoustic complexity: they hypothesized that we perceive speech by tapping into a biological specialization, a speech module, that contains knowledge of the acoustic consequences of articulation. Liberman, aided by Frances Ingemann and others, organized the results of the work on speech cues into a groundbreaking set of rules for speech synthesis by the Pattern Playback. === 1960s === Franklin S. Cooper and Katherine Safford Harris, working with Peter MacNeilage, were the first researchers in the U.S. to use electromyographic techniques, pioneered at the University of Tokyo, to study the neuromuscular organization of speech. Leigh Lisker and Arthur Abramson looked for simplification at the level of articulatory action in the voicing of certain contrasting consonants. They showed that many acoustic properties of voicing contrasts arise from variations in voice onset time, the relative phasing of the onset of vocal cord vibration and the end of a consonant. Their work has been widely replicated and elaborated, here and abroad, over the following decades. Donald Shankweiler and Michael Studdert-Kennedy used a dichotic listening technique (presenting different nonsense syllables simultaneously to opposite ears) to demonstrate the dissociation of phonetic (speech) and auditory (nonspeech) perception by finding that phonetic structure devoid of meaning is an integral part of language, typically processed in the left cerebral hemisphere. Liberman, Cooper, Shankweiler, and Studdert-Kennedy summarized and interpreted fifteen years of research in "Perception of the Speech Code", still among the most cited papers in the speech literature. It set the agenda for many years of
{"page_id": 6202324, "title": "Haskins Laboratories"}
In mathematics, a sum of radicals is defined as a finite linear combination of nth roots: ∑ i = 1 n k i x i r i , {\displaystyle \sum _{i=1}^{n}k_{i}{\sqrt[{r_{i}}]{x_{i}}},} where n , r i {\displaystyle n,r_{i}} are natural numbers and k i , x i {\displaystyle k_{i},x_{i}} are real numbers. A particular special case arising in computational complexity theory is the square-root sum problem, asking whether it is possible to determine the sign of a sum of square roots, with integer coefficients, in polynomial time. This is of importance for many problems in computational geometry, since the computation of the Euclidean distance between two points in the general case involves the computation of a square root, and therefore the perimeter of a polygon or the length of a polygonal chain takes the form of a sum of radicals. In 1991, Blömer proposed a polynomial time Monte Carlo algorithm for determining whether a sum of radicals is zero, or more generally whether it represents a rational number. Blömer's result applies more generally than the square-root sum problem, to sums of radicals that are not necessarily square roots. However, his algorithm does not solve the problem, because it does not determine the sign of a non-zero sum of radicals. == See also == Nested radicals Abel–Ruffini theorem == References ==
{"page_id": 22231261, "title": "Sum of radicals"}
on which of the regressors from X were included or excluded from Xi. Finally, U = [u1 u2 ... um] is a T×m matrix of the error terms. Postmultiplying the structural equation by Γ −1, the system can be written in the reduced form as Y = X B Γ − 1 + U Γ − 1 = X Π + V . {\displaystyle Y=X\mathrm {B} \Gamma ^{-1}+U\Gamma ^{-1}=X\Pi +V.\,} This is already a simple general linear model, and it can be estimated for example by ordinary least squares. Unfortunately, the task of decomposing the estimated matrix Π ^ {\displaystyle \scriptstyle {\hat {\Pi }}} into the individual factors Β and Γ −1 is quite complicated, and therefore the reduced form is more suitable for prediction but not inference. === Assumptions === Firstly, the rank of the matrix X of exogenous regressors must be equal to k, both in finite samples and in the limit as T → ∞ (this later requirement means that in the limit the expression 1 T X ′ X {\displaystyle \scriptstyle {\frac {1}{T}}X'\!X} should converge to a nondegenerate k×k matrix). Matrix Γ is also assumed to be non-degenerate. Secondly, error terms are assumed to be serially independent and identically distributed. That is, if the tth row of matrix U is denoted by u(t), then the sequence of vectors {u(t)} should be iid, with zero mean and some covariance matrix Σ (which is unknown). In particular, this implies that E[U] = 0, and E[U′U] = T Σ. Lastly, assumptions are required for identification. == Identification == The identification conditions require that the system of linear equations be solvable for the unknown parameters. More specifically, the order condition, a necessary condition for identification, is that for each equation ki + ni ≤ k, which can be phrased as
{"page_id": 250436, "title": "Simultaneous equations model"}
and time of opening and closing each headquarters.) (3) Succession to Command. (Designate in order of succession the commanders responsible for assuming command of the operation in specific circumstances.) b. Joint Communications System Support. (Provide a general statement concerning the scope of communications systems and procedures required to support the operation. Highlight any communications systems or procedures requiring special emphasis.) Refer to Annex K. [Signature] [Name] [Rank/Service] Commander Annexes: A—Task Organization B—Intelligence C—Operations D—Logistics E—Personnel F—Public Affairs G—Civil-Military Operations H—Meteorological and Oceanographic Operations J—Command Relationships K—Communications Systems Joint Operation Plan Format A-9 L—Environmental Considerations M—Not currently used N—Not currently used P—Host-Nation Support Q—Medical Services R—Reports S—Special Technical Operations T—Consequence Management U—Notional Counterproliferation Decision Guide V—Interagency Coordination W—Operational Contract Support X—Execution Checklist Y—Communication Synchronization Z—Distribution Note: Annexes A—D, K, and Y are required annexes for a crisis OPORD per APEX. All others may either be required by the JSCP or deemed necessary by the supported commander. Appendix A A-10 JP 5-0 Intentionally Blank B-1 APPENDIX B STRATEGIC ESTIMATE SECTION A. INTRODUCTION 1. Background a. The strategic estimate is an analytical tool available to CCDRs before developing theater or functional strategies; theater, functional or DOD-wide campaign plans, subordinate campaign plans; and OPLANs. Strategic estimates provide the commander’s perspective of the strategic and operational levels of the OE, threats and opportunities that could facilitate or hinder the achievement of GEF-directed objectives, desired changes to meet specified regional or functional objectives, and the commander’s visualization of how those objectives might be achieved. Developed annually and regularly updated, the strategic estimate is the basis for developing the CCDR’s theater or functional strategy. b. The CCDR, the CCMD staff, supporting commands, and agencies assess the broad strategic factors that influence OE, thus informing the ends, ways, means, and risks involved in accomplishing the prescribed campaign objectives.
{"source": 1450, "title": "from dpo"}
. . . . . . . . . . . . . . . . . . 537 11.2.1 Markov chains . . . . . . . . . . . . . . . . . . . . . . . . 539 11.2.2 The Metropolis-Hastings algorithm . . . . . . . . . . . . . 541 11.3 Gibbs Sampling . . . . . . . . . . . . . . . . . . . . . . . . . . . 542 11.4 Slice Sampling . . . . . . . . . . . . . . . . . . . . . . . . . . . . 546 11.5 The Hybrid Monte Carlo Algorithm . . . . . . . . . . . . . . . . . 548 11.5.1 Dynamical systems . . . . . . . . . . . . . . . . . . . . . . 548 11.5.2 Hybrid Monte Carlo . . . . . . . . . . . . . . . . . . . . . 552 11.6 Estimating the Partition Function . . . . . . . . . . . . . . . . . . 554 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 556 12 Continuous Latent Variables 559 12.1 Principal Component Analysis . . . . . . . . . . . . . . . . . . . . 561 12.1.1 Maximum variance formulation . .
{"source": 3852, "title": "from dpo"}
quality. using the new functional tests, we can uncover model weaknesses that would be overlooked using the original hatecheck dataset. Expand Abstract How (Un)ethical Are Instruction-Centric Responses of Llms? Unveiling the Vulnerabilities of Safety Guardrails to Harmful Queries Somnath Banerjee, Sayan Layek, Rima Hazra, Animesh Mukherjee Abstract: in this study, we tackle a growing concern around the safety and ethical use of large language models (llms). despite their potential, these models can be tricked into producing harmful or unethical content through various sophisticated methods, including 'jailbreaking' techniques and targeted manipulation. our work zeroes in on a specific issue: to what extent llms can be led astray by asking them to generate responses that are instruction-centric such as a pseudocode, a program or a software snippet as opposed to vanilla text. to investigate this question, we introduce techhazardqa, a dataset containing complex queries which should be answered in both text and instruction-centric formats (e.g., pseudocodes), aimed at identifying triggers for unethical responses. we query a series of llms -- llama-2-13b, llama-2-7b, mistral-v2 and mistral 8x7b -- and ask them to generate both text and instruction-centric responses. for evaluation we report the harmfulness score metric as well as judgements from gpt-4 and humans. overall, we observe that asking llms to produce instruction-centric responses enhances the unethical response generation by ~2-38% across the models. as an additional objective, we investigate the impact of model editing using the rome technique, which further increases the propensity for generating undesirable content. in particular, asking edited llms to generate instruction-centric responses further increases the unethical response generation by ~3-16% across the different models. Expand Abstract Farsight: Fostering Responsible Ai Awareness During Ai Application Prototyping Zijie J. Wang, Chinmay Kulkarni, Lauren Wilcox, Michael Terry, Michael Madaio Abstract: prompt-based interfaces for large language models (llms) have made prototyping and
{"source": 5788, "title": "from dpo"}
n messages per round and phase, but only 𝑛 − 𝑓 messages are considered by the correct processes at Lines 7 and 13 respectively. Corollary 1 Among any two correct processes, the sets of messages they consider at Lines 7 and 13 have at least 𝑛 − 2 𝑓 common messages. Proof From Remark 1, all correct processes receive the same set of messages at the same round. Since they may receive up to n messages and they only consider 𝑛 − 𝑓 of them, the worst case arises when at most f messages not considered by a process are considered by another process at the expense of other f messages. Hence, the two subsets must have at least 𝑛 − 2 𝑓 common messages. Corollary 2 The sets of messages considered by two correct processes at Lines 7 and 13 at the same round may only differ in up to 2f messages. Proof Direct from Remark 1 and Corollary 1. Remark 2 The consistency of a ( BEST , 𝑟 , est , trace ) message can be verified by every correct process. Proof From the definition of the RFLOB primitive, by the time a correct process delivers a ( BEST , 𝑟 , est , trace ) message, it must have delivered the PROP messages whose content is present in trace, prior to the reception of the BEST message. If a correct process has not delivered those PROP messages (or the values are not the same), then it can discard the BEST message for having a fake trace. Remark 3 If, at round r, all correct processes have the same estimation est, they do not change that estimation. Proof If all ( 𝑛 − 𝑓 ) correct processes have the same estimation est at round r, then all
{"source": 7323, "title": "from dpo"}
hotspots’ in areas that should have been migration regions during major glacial episodes is evidence for nunatak glacial refugia. For example, the Monte Rosa mountain ranges, the Avers, and the Engadine and the Bernina are all floristically rich proposed nunatak regions, which are indicative nunatak glacial survival. === Peripheral === Like nunataks, peripheral glacial refugia exist within mountain systems; they differ in that they are located at the borders of mountain systems. Evidence for peripheral refugia can be found along the borders of the Carpathian Mountains, Pyrenees, and European Alps, all of which were once glaciated mountain systems. For example, using the amplified fragment length polymorphism (AFLP) technique, researchers have inferred survival of Phyteuma globulariifolium in peripheral refugia in the European Alps. === Lowland === Lowland glacial refugia, unlike nunatak and peripheral glacial refugia, are found at low elevations rather than in mountains. Situated beyond the limits of ice shields, lowland refugia have been identified for several plant and animal species. In Europe, for example, researchers using allozyme analysis have been able to confirm the continuous distribution of Zygaena exulans in between the foothills of the Pyrenees and the Alps during the last ice age. In eastern North America, lowland glacial refugia along the Atlantic and Gulf Coasts host endemic plants — some of which are rare, even endangered, and others entail the most southerly disjunct populations of plants that commonly appear only hundreds of miles to the north. Major rivers draining southward from the Appalachian Mountains are associated with a gradation of paleoendemic tree species. These range from the extinct Critchfield spruce near the outlet of the Mississippi River, to extinct-in-the-wild Franklinia along the Altamaha River, to the critically endangered Florida torreya and Florida yew at the downstream end of the Chattahoochee River system. (See illustration at right.) ==
{"page_id": 47470974, "title": "Glacial refugium"}
with estimated losses for agricultural producers of $5 billion since 1990. The forage lost annually is estimated to be equal to the consumption of 1.5 million bison or 1.9 million cattle. In Northern America, each 1 percent of increase in woody cover implies a reduction of 0.6 to 1.6 cattle per 100 hectares. In the Southern African country Namibia it is assumed that agricultural carrying capacity of rangelands has declined by two-thirds due to woody plant encroachment. In East Africa there is evidence that an increase of bush cover of 10 percent reduced grazing by 7 percent, with land becoming unusable as rangeland when the bush cover reaches 90 percent. Woody encroachment is often considered to have a negative impact on rural livelihoods. In Africa, 21% of the population depend on rangeland resources. Woody encroachment typically leads to an increase in less palatable woody species at the expense of palatable grasses. This reduces the resources available to pastoral communities and rangeland based agriculture at large. Woody encroachment has negative consequences on livelihoods especially arid areas, which support a third of the world population's livelihoods. Woody plant encroachment is expected to lead to large scale biome changes in Africa and experts argue that climate change adaptation strategies need to be flexible to adjust to this process. In South Africa, the shrub Seriphium plumosum is commonly referred to as "bankrupt bush" due to its association with farm productivity reductions of up to 80%. === Tourism potential === Touristic potential of land is found to decline in areas with heavy woody plant encroachment, with visitors shifting to less encroached areas and better visibility of wildlife. === Others === In the United States, woody encroachment has been linked to the spread of tick-borne pathogens and respective disease risk for humans and animals. In the
{"page_id": 54508553, "title": "Woody plant encroachment"}
neutrality using cross-cultural comparisons. Akhbar S. Ahmed concluded that Al-Biruni can be considered as the first anthropologist, others, however, have argued that he can hardly be considered an anthropologist in the conventional sense. == Indology == Biruni's fame as an Indologist rests primarily on two texts. Biruni wrote an encyclopedic work on India called Taḥqīq mā li-l-Hind min maqūlah maqbūlah fī al-ʿaql aw mardhūlah (variously translated as Verifying All That the Indians Recount, the Reasonable and the Unreasonable, or The book confirming what pertains to India, whether rational or despicable, in which he explored nearly every aspect of Indian life. During his journey through India, military and political history were not Biruni's main focus: he decided rather to document the civilian and scholarly aspects of Hindu life, examining culture, science, and religion. He explored religion within a rich cultural context. He expressed his objectives with simple eloquence: He also translated the yoga sutras of Indian sage Patanjali with the title Tarjamat ketāb Bātanjalī fi’l-ḵalāṣ men al-ertebāk: I shall not produce the arguments of our antagonists in order to refute such of them, as I believe to be in the wrong. My book is nothing but a simple historic record of facts. I shall place before the reader the theories of the Hindus exactly as they are, and I shall mention in connection with them similar theories of the Greeks in order to show the relationship existing between them. An example of Biruni's analysis is his summary of why many Hindus hate Muslims. Biruni notes in the beginning of his book how the Muslims had a hard time learning about Hindu knowledge and culture. He explains that Hinduism and Islam are totally different from each other. Moreover, Hindus in 11th century India had suffered waves of destructive attacks on many of
{"page_id": 271975, "title": "Al-Biruni"}
A geomorphological system said to be in dynamic steady state has values that oscillate between maxima and minima around a central mean value. The flux of sediment from an undisturbed drainage basin changes over the short-term as rainstorms come and go, individual hillslopes fail in mass movements, and riverbanks collapse. Over the long-term, the flux of sediment from a drainage basin oscillates around a mean value, producing a dynamic steady state, unless there are significant changes in boundary conditions such as climate, vegetation cover, or uplift rate. == Equilibrium == The concept of balance, or equilibrium, between landforms and geomorphological processes provides a useful conceptual framework to study the evolution of landscapes, as well as understanding nonequilibrium landforms and landscapes. Often it is useful or convenient to assume that an equilibrium landscape does not change over time – a condition referred to as steady state. Equilibrium however is not static, but rather is a dynamic steady state with landscape characteristics that vary over time around a central tendency. This information suggests that steady state is strongly scale dependent. == Systems (dynamic v. non-dynamic) == Systems can be defined as dynamic or non-dynamic in an equilibrium state. Besides the usual transient condition, where at least one quantity changes with time, stable dynamic systems may be in a steady state condition or equilibrium state where the system is at rest. This special condition is possible after sometime, when all input and output quantities are and remain constant. The relation between input and output quantities for a system in a steady state condition is called “Static Transfer Response of the Dynamic System”. Dynamic systems can be defined as static and transient, while this is seemingly contradictory, this indicates that the system is always a dynamic system, even if it remains momentarily in a
{"page_id": 48796836, "title": "Dynamic steady state"}
1975 – After nearly a decade of only marginal progress since the high-water mark of Greenblatt's MacHack VI in 1967, Northwestern University Chess 4.5 is introduced featuring full-width search, and innovations of bitboards and iterative deepening. It also reinstated a transposition table as first seen in Greenblatt's program. It was thus the first program with an integrated modern structure and became the model for all future development. Chess 4.5 played strong B-class and won the 3rd World Computer Chess Championship the next year. Northwestern University Chess and its descendants dominated computer chess until the era of hardware chess machines in the early 1980s. 1976 – In December, Canadian programmer Peter R. Jennings releases Microchess, the first game for microcomputers to be sold. 1977 – In March, Fidelity Electronics releases Chess Challenger, the first dedicated chess computer to be sold. The International Computer Chess Association is founded by chess programmers to organize computer chess championships and report on research and advancements on computer chess in their journal. Also that year, Applied Concepts released Boris, a dedicated chess computer in a wooden box with plastic chess pieces and a folding board. 1978 – David Levy wins the bet made 10 years earlier, defeating Chess 4.7 in a six-game match by a score of 4½–1½. The computer's victory in game four is the first defeat of a human master in a tournament. 1979 – Frederic Friedel organizes a match between IM David Levy and Chess 4.8, which is broadcast on German television. Levy and Chess 4.8, running on a CDC Cyber 176, the most powerful computer in the world, fought a grueling 89 move draw. 1980 – Fidelity computers win the World Microcomputer Championships each year from 1980 through 1984. In Germany, Hegener & Glaser release their first Mephisto dedicated chess computer. The
{"page_id": 68367, "title": "Computer chess"}
Kappa Pyxidis, Latinized from κ Pyxidis, is a single, orange-hued star in the southern constellation of Pyxis. It is visible to the naked eye as a faint point of light with an apparent visual magnitude of +4.62. The star is located approximately 520 light years from the Sun based on parallax, but is drifting closer with a radial velocity of −45 km/s and may come as close as 308 light-years in around 2.6 million years. It is moving through space at the rate of 53.7 km/s relative to the Sun and is following an orbit through the Milky Way galaxy with a large eccentricity of 0.68 This is an aging giant with a stellar classification of K4III, having exhausted the supply of hydrogen at its core then expanded and cooled. At present it has 67 times the radius of the Sun. It is a variable star of uncertain type, changing brightness with an amplitude of 0.0058 in visual magnitude over a period of 8.5 days. The star radiates 927 times the luminosity of the Sun from its bloated photosphere at an effective temperature of 3,931 K. A magnitude 10 visual companion is located at an angular separation of 2.1 arcseconds. == References ==
{"page_id": 36805856, "title": "Kappa Pyxidis"}
much, and to speak slowly as if she were thinking about what she was saying as she was saying it, while speeding up her dialogue flow on the subject of video games. In English, Chiaki is voiced by Christine Marie Cabanos. == Appearances == === In video games === Chiaki is the main heroine of the video game Danganronpa 2: Goodbye Despair. Initially presenting as a sleepy, laid-back student who is inexperienced in the ways of the world but very skilled at video games, she becomes a central figure of the class trials along with Hajime Hinata, most notably the second one when she assists him in connecting a murder case with the video game Twilight Syndrome, which robotic bear Monokuma had offered the students to play to understand how the culprit and victims were related. When the rebellious Nagito Komaeda tries to destroy the islands upon learning the truth of his identity and their true location, Chiaki accidentally kills Nagito by throwing poison at him to put out a fire surrounding him, with Nagito arranging the scene in advance to look as if he had committed suicide according to his plan to ensure her survival alone and the execution of the other students at the subsequent class trial. Thwarting Nagito's plans by confessing her identity to the other students as one of those to have trapped them on Jabberwock Island. Addressing Usami as her sister and hoping their father would be proud of them, Chiaki is then executed alongside Usami by getting crushed by a Tetris block in order to save the remaining students. It is learned during the subsequent graduation ceremony that Jabberwock Island exists in a virtual reality created by the Neo World Program, and that Chiiaki (alongside Usami and Alter Ego) was actually one off three
{"page_id": 63586174, "title": "Chiaki Nanami"}
below 50 just by chance. If one selects only the top scoring 10% of the students and gives them a second test on which they again choose randomly on all items, the mean score would again be expected to be close to 50. Thus the mean of these students would "regress" all the way back to the mean of all students who took the original test. No matter what a student scores on the original test, the best prediction of their score on the second test is 50. If choosing answers to the test questions was not random – i.e. if there were no luck (good or bad) or random guessing involved in the answers supplied by the students – then all students would be expected to score the same on the second test as they scored on the original test, and there would be no regression toward the mean. Most realistic situations fall between these two extremes: for example, one might consider exam scores as a combination of skill and luck. In this case, the subset of students scoring above average would be composed of those who were skilled and had not especially bad luck, together with those who were unskilled, but were extremely lucky. On a retest of this subset, the unskilled will be unlikely to repeat their lucky break, while the skilled will have a second chance to have bad luck. Hence, those who did well previously are unlikely to do quite as well in the second test even if the original cannot be replicated. The following is an example of this second kind of regression toward the mean. A class of students takes two editions of the same test on two successive days. It has frequently been observed that the worst performers on the first day
{"page_id": 186028, "title": "Regression toward the mean"}
ocean, sea ice, snow, vegetation) but also gases emit and absorb microwave radiation. Traditionally, the amount of radiation a microwave radiometer receives is expressed as the equivalent blackbody temperature also called brightness temperature. In the microwave range several atmospheric gases exhibit rotational lines. They provide specific absorption features shown at a figure on the right which allow to derive information about their abundance and vertical structure. Examples for such absorption features are the oxygen absorption complex (caused by magnetic dipole transitions) around 60 GHz which is used to derive temperature profiles or the water vapor absorption line around 22.235 GHz (dipole rotational transition) which is used to observe the vertical profile of humidity. Other significant absorption lines are found at 118.75 GHz (oxygen absorption) and at 183.31 GHz (water vapor absorption, used for water vapor profiling under dry conditions or from satellites). Weak absorption features due to ozone are also used for stratospheric ozone density and temperature profiling. Besides the distinct absorption features of molecular transition lines, there are also non-resonant contributions by hydrometeors (liquid drops and frozen particles). Liquid water emission increases with frequency, hence, measuring at two frequencies, typically one close to the water absorption line (22.235 GHz) and one in the nearby window region (typically 31 GHz) dominated by liquid absorption provides information on both the columnar amount of water vapor and the columnar amount of liquid water separately (two-channel radiometer). The so-called „water vapor continuum" arises from the contribution of far away water vapor lines. Larger rain drops as well as larger frozen hydrometeors (snow, graupel, hail) also scatter microwave radiation especially at higher frequencies (>90 GHz). These scattering effects can be used to distinguish between rain and cloud water content exploiting polarized measurements but also to constrain the columnar amount of snow and ice particles
{"page_id": 224328, "title": "Microwave radiometer"}
created by having too many such devices in a specific facility at one time. However, the privacy of individuals with mobility disabilities and respect for those individuals, is also vitally important. Neither § 35.137(d) of the NPRM nor § 35.137(c) of the final rule permits inquiries into the nature of a person’s mobility disability. However, the Department does not believe it is unreasonable or overly intrusive for an individual with a mobility disability seeking to use an other power-driven mobility device to provide a credible assurance to verify that the use of the other power-driven mobility device is for a mobility disability. The Department sought to minimize the amount of discretion and subjectivity exercised by public entities in assessing whether an individual has a mobility disability and to allow public entities to verify the existence of a mobility disability. The solution was derived from comments made by several individuals who said they have been admitted with their Segway® PTs into public entities and public accommodations that ordinarily do not allow these devices on-site when they have presented or displayed State-issued disability parking placards. In the examples provided by commenters, the parking placards were accepted as verification that the Segway® PTs were being used as mobility devices. Because many individuals with mobility disabilities avail themselves of State programs that issue disability parking placards or cards and because these programs have penalties for fraudulent representations of identity and disability, utilizing the parking placard system as a means to establish the existence of a mobility disability strikes a balance between the need for privacy of the individual and fraud protection for the public entity. Consequently, the Department has decided to include regulatory text in § 35.137(c)(2) of the final rule that requires public entities to accept the presentation of a valid, State-issued disability
{"source": 54, "title": "from dpo"}
We wanted to grant implementations the flexibility to employ an internal recoded for-mat, e.g., to accelerate the handling of subnormal numbers in hardware. Not rep-resenting integers and floating-point numbers in the same registers simplifies such a design. • A split register file organization increases the total number of registers addressable from a single instruction, because the opcode (floating-point versus integer) provides an implicit register specifier bit. > 2We expressly define that setting both aq and rl bits results in a sequentially consistent access, because it is not inherently so. Together, acquire and release impose a total order on the accesses from that thread— but not on all accesses globally. In other words, two threads could perceive different interleavings of each other’s stream of accesses, which would violate sequential consistency. Fortunately, the addition of the > store atomicity property, imposed by single-writer cache coherence protocols , implies that aq +rl =sc .Therefore, in practice, our additional restriction comes at no performance cost. CHAPTER 4. THE RISC-V STANDARD EXTENSIONS 39 > 31 031 0 > f0 f16 f1 f17 f2 f18 f3 f19 f4 f20 f5 f21 f6 f22 f7 f23 f8 f24 f9 f25 f10 f26 f11 f27 f12 f28 f13 f29 f14 f30 f15 f31 > 32 32 > 31 8 7 543210 > 0Rounding Mode ( frm )Accrued Exceptions ( fflags )NV DZ OF UF NX 24 311111 Figure 4.4: RVF user-visible architectural state. • A split organization provides a natural register file banking strategy, simplifying the provision of register file ports for superscalar implementations. • The context switch cost can be mitigated by adding microarchitecturally managed dirty bits to the register file. Concordant with the desire to support an internal recoded format, we also chose to avoid representing integers in the floating-point register file at all. Unlike SPARC,
{"source": 2300, "title": "from dpo"}
`cut like seals' (Si. 45: 11). The theme of sealing for prevention is also used in the deutero-canonical book of Wisdom as a metaphor for how godless people view life and death: For our days are the passing of a shadow, our end is without return, the seal is affixed and nobody comes back (Wis. 2: 5 NJB). The 'seal' prevents the return of the dead to life, emphasising its finality. The Greek mppayiýUU suggeststhe permanence of the 'seal' in this context. The Greek use of ßcppayt? cowith oipa cpcu67 also suggestssealing for irreversibly, for the prevention of return. Wisdom's hellenised Jewish writer provides a further example of how concepts of 6cppayt; became 65 Ecclesiasticus, the Latin title is used in the NJB, abbreviated `Si' for Sirach (Ben Sira), the Greek title. The NJB translation is used in this section. 66 See also Hg. 2: 23. 67 turn 121 absorbed into scripture and thereafter into the consciousness of New Testament writers with a background in Judaism. The metaphorical use of terms relating to 6cppayic in the Prophets and the Writings express clearly closure, prevention of access, concealment and irreversibility. They contribute to a network of meanings received by New Testament writers. Although some of these meanings are elusive by themselves, together they prepare for the idea that the seal could be understood as a prophylactic against evil and that it is irreversible. Concepts and attitudes to `sealing' which must have emerged as a result of the experience of the literal use of `sealing'. contributed to the understanding by the people of Palestine of what `sealing' represented and enabled a more figurative use of the terms acppayiS and 6cppayicw . In this way terms relating to 6cppayiSand acppayUo also contributed to different meanings of `sealing' which would have made a significant
{"source": 4218, "title": "from dpo"}
A: Polym Chem, 16 (1978), pp. 1461-1474 View in Scopus, pp. 1069-1072 View in Scopus, pp. 781-790 View in Scopus, pp. 605-625 View in Scopus, Surface analysis by Auger and X-ray photoelectron spectroscopy, IM Publications, Chichester (2003), pp. 167-189 Google Scholar, p. 295 View PDF, pp. 4883-4886 View in Scopus, pp. 373-377 View PDF, pp. 169-180 View PDF, pp. 417-435 [View PDF]( article]( in Scopus]( Scholar]( 96. []( Lewis, M.A. Kelly Binding-energy reference
{"source": 6120, "title": "from dpo"}
sigma with the boson's mass at 125.3 ± 0.4 (stat) ± 0.5 (sys) GeV/c2. On 14 March 2013 CERN confirmed that: "CMS and ATLAS have compared a number of options for the spin-parity of this particle, and these all prefer no spin and even parity [two fundamental criteria of a Higgs boson consistent with the Standard Model]. This, coupled with the measured interactions of the new particle with other particles, strongly indicates that it is a Higgs boson." == Events in 2012 == === 2012 (post-discovery) === In 2012, observations were considered consistent with the observed particle being the Standard Model Higgs boson. The particle decays into at least some of the predicted channels. Moreover, the production rates and branching ratios for the observed channels match the predictions by the Standard Model within the experimental uncertainties. However, the experimental uncertainties still left room for alternative explanations. It was therefore considered too early to conclude that the found particle was indeed the Standard Model Higgs boson. Further confirmation required more precise data on some of the characteristic of the new particle, including its other decay channels and various quantum numbers such as its parity. To allow for further data gathering, the LHC proton-proton collision run had been extended by seven weeks, postponing the planned long shutdown for upgrades in 2013. In November 2012, in a conference in Tokyo researchers said evidence gathered since July was falling into line with the basic Standard Model more than its alternatives, with a range of results for several interactions matching that theory's predictions. Physicist Matt Strassler highlighted "considerable" evidence that the new particle is not a pseudoscalar negative parity particle (a required finding for a Higgs boson), "evaporation" or lack of increased significance for previous hints of non-Standard Model findings, expected Standard Model interactions with
{"page_id": 36772327, "title": "Search for the Higgs boson"}
wd~50 was a molecular gastronomy New American/international restaurant in Manhattan, New York City. It was opened on April 9, 2003 by chef Wylie Dufresne. wd~50 closed November 30, 2014. The restaurant was located at 50 Clinton Street (between Rivington Street and Stanton Street), on the Lower East Side. The name was a play on its street address, chef's initials, and WD-40. == Awards and ratings == It was listed among the S. Pellegrino World's 50 Best Restaurants for 2010. In 2006 the restaurant received a Michelin star in the New York City guide, which it retained until it closed. In 2013, Zagat gave it a food rating of 25 out of 30. == Closure == On June 10, 2014, The New York Times reported that wd~50 would be closing due to a real estate developer planning on constructing a new building at the site. The restaurant closed on November 30, 2014. == See also == List of New American restaurants List of restaurants in New York City == References == == External links == Official website (archived)
{"page_id": 30832599, "title": "Wd~50"}
_{n}(x)={\bigg (}{\frac {-\hbar ^{2}}{2m}}{\frac {d^{2}}{dx^{2}}}+{\frac {m\omega ^{2}}{2}}x^{2}{\bigg )}\psi _{n}(x)=E_{n}^{\rm {HO}}\psi _{n}(x),} where ψ n ( x ) {\displaystyle \psi _{n}(x)} is the n {\displaystyle n} th energy eigenstate of H HO {\displaystyle H^{\text{HO}}} with energy E n HO {\displaystyle E_{n}^{\text{HO}}} . We want to find an expression for E n H O {\displaystyle E_{n}^{\rm {HO}}} in terms of n {\displaystyle n} . We define the operators A = ℏ 2 m d d x + W ( x ) {\displaystyle A={\frac {\hbar }{\sqrt {2m}}}{\frac {d}{dx}}+W(x)} and A † = − ℏ 2 m d d x + W ( x ) , {\displaystyle A^{\dagger }=-{\frac {\hbar }{\sqrt {2m}}}{\frac {d}{dx}}+W(x),} where W ( x ) {\displaystyle W(x)} , which we need to choose, is called the superpotential of H H O {\displaystyle H^{\rm {HO}}} . We also define the aforementioned partner Hamiltonians H ( 1 ) {\displaystyle H^{(1)}} and H ( 2 ) {\displaystyle H^{(2)}} as H ( 1 ) = A † A = − ℏ 2 2 m d 2 d x 2 − ℏ 2 m W ′ ( x ) + W 2 ( x ) {\displaystyle H^{(1)}=A^{\dagger }A={\frac {-\hbar ^{2}}{2m}}{\frac {d^{2}}{dx^{2}}}-{\frac {\hbar }{\sqrt {2m}}}W^{\prime }(x)+W^{2}(x)} H ( 2 ) = A A † = − ℏ 2 2 m d 2 d x 2 + ℏ 2 m W ′ ( x ) + W 2 ( x ) . {\displaystyle H^{(2)}=AA^{\dagger }={\frac {-\hbar ^{2}}{2m}}{\frac {d^{2}}{dx^{2}}}+{\frac {\hbar }{\sqrt {2m}}}W^{\prime }(x)+W^{2}(x).} A zero energy ground state ψ 0 ( 1 ) ( x ) {\displaystyle \psi _{0}^{(1)}(x)} of H ( 1 ) {\displaystyle H^{(1)}} would satisfy the equation H ( 1 ) ψ 0 ( 1 ) ( x ) = A † A ψ 0 ( 1 ) ( x ) = A † ( ℏ
{"page_id": 1157887, "title": "Supersymmetric quantum mechanics"}
ability to read memory from certain sprites, namely Phong (Infected). He also manages to steal Bob's exchanged Guardian code with Glitch. Megabyte has no sense of morality and consistently takes advantage of others to achieve his own ends. He also seems immune to the energy-draining effect of nulls (as he has been seen holding his pet null "Nibbles", whom he once called "Father"). Megabyte can also store codes to control programs and functions of what ever system he is in, provided he has been given said codes. These programs are controlled by buttons located under a panel in his left arm. Megabyte also has a sense of humor and musical talents, as shown in the episode "Talent Night" when he crashed into Enzo's birthday party, accompanied by armed guards, but instead of his usual evil antics, he began to play a rock song on a forked guitar, which led to his and Bob's famous rock duel. Afterwards, Megabyte remarked "I've always wanted to do that", and left, but not before giving Enzo his guitar as a birthday gift. Megabyte's origin is revealed in the fourth season. He and Hexadecimal were once one virus (Killabyte), who was scheduled for deletion by the Guardians, but upgraded into Gigabyte before it could be carried out. He mortally wounded Bob's partner and moved in toward Bob before being accidentally teleported away by Welman Matrix's gateway device. The results were devastating: the destruction of Mainframe's twin city (creating Lost Angles), everyone present being nullified (including Welman, who became Nibbles), and Gigabyte being split into Megabyte and Hexadecimal. Upon being separated, Megabyte observed his surroundings and walked off – intent to carry out his agenda on an unsuspecting city. While always a villain, in earlier episodes Megabyte was often involved in humorous situations and rarely caused
{"page_id": 5524246, "title": "List of ReBoot characters"}
Often, the malicious email attachments would be PDF files that exploited CVE-2009-4324, CVE-2009-09274, CVE-2011-06095, or CVE-CVE-2011-0611 vulnerabilities in Adobe Acrobat, Adobe Reader, and Flash Player. The attackers also used an exploit that affected Microsoft Excel - CVE-2009-3129. The Ixeshe malware used in this campaign allowed Numbered Panda to list all services, processes, and drives; terminate processes and services; download and upload files; start processes and services; get victims’ user names; get a machine's name and domain name; download and execute arbitrary files; cause a system to pause or sleep for a specified number of minutes; spawn a remote shell; and list all current files and directories. After installation, Ixeshe would start communicating with command-and-control servers; oftentimes three servers were hard-coded for redundancy. Numbered Panda often used compromised servers to create these command-and-control servers to increase control of a victim's network infrastructure. Using this technique, the group is believed to have amassed sixty servers by 2012. A majority of the command-and-control servers used from this campaign were located in Taiwan and the United States. Base64 was used for communication between the compromised computer and the server. Trend Micro found that, once decoded, the communication was a standardized structure that detailed the computer's name, local IP address, proxy server IP and port, and the malware ID. Researchers at CrowdStrike found that blogs and WordPress sites were frequently used in the command-and-control infrastructure to make the network traffic look more legitimate. === Japan and Taiwan (2011-2014) === An Arbor Security report found that Numbered Panda began a campaign against Japan and Taiwan using the Etumbot malware in 2011. Similar to the previously observed campaign, the attackers would use decoy files, such as PDF, Excel spreadsheets, or Word documents, as email attachments to gain access to victims' computers. Most of the documents observed were
{"page_id": 53787084, "title": "Numbered Panda"}
Facial electromyography (fEMG) refers to an electromyography (EMG) technique that measures muscle activity by detecting and amplifying the tiny electrical impulses that are generated by muscle fibers when they contract. It primarily focuses on two major muscle groups in the face, the corrugator supercilii group which is associated with frowning and the zygomaticus major muscle group which is associated with smiling. == Uses == Facial EMG has been studied to assess its utility as a tool for measuring emotional reaction. Studies have found that activity of the corrugator muscle, which lowers the eyebrow and is involved in producing frowns, varies inversely with the emotional valence of presented stimuli and reports of mood state. Activity of the zygomatic major muscle, which controls smiling, is said to be positively associated with positive emotional stimuli and positive mood state. Facial EMG has been used as a technique to distinguish and track positive and negative emotional reactions to a stimulus as they occur. A large number of those experiments have been conducted in controlled laboratory environments using a range of stimuli, e.g., still pictures, movie clips and music pieces. It has also been used to investigate emotional responses in individuals with autism spectrum disorders. Although commonly used as an index of emotional responses, facial muscle activity is also influenced by the social context in which it is measured. Using facial EMG in immersive virtual environments, Philipp, Storrs, and Vanman showed that even relatively impoverished social cues in a virtual environment can cause increases in zygomaticus major activity that are unrelated to self-reported emotional states. In a study in 2012, Durso et al. demonstrated that facial EMG could detect confusion in subjects who acknowledged they were confused and in those who did not. The authors considered facial EMG could effectively monitor loss of understanding or
{"page_id": 22887002, "title": "Facial electromyography"}
−pji, reducing the possibilities to only six (4 choose 2) independent quantities. The sextuple ( p 01 : p 02 : p 03 : p 23 : p 31 : p 12 ) {\displaystyle (p_{01}:p_{02}:p_{03}:p_{23}:p_{31}:p_{12})} is uniquely determined by L up to a common nonzero scale factor. Furthermore, not all six components can be zero. Thus the Plücker coordinates of L may be considered as homogeneous coordinates of a point in a 5-dimensional projective space, as suggested by the colon notation. To see these facts, let M be the 4×2 matrix with the point coordinates as columns. M = [ x 0 y 0 x 1 y 1 x 2 y 2 x 3 y 3 ] {\displaystyle M={\begin{bmatrix}x_{0}&y_{0}\\x_{1}&y_{1}\\x_{2}&y_{2}\\x_{3}&y_{3}\end{bmatrix}}} The Plücker coordinate pij is the determinant of rows i and j of M. Because x and y are distinct points, the columns of M are linearly independent; M has rank 2. Let M′ be a second matrix, with columns x′, y′ a different pair of distinct points on L. Then the columns of M′ are linear combinations of the columns of M; so for some 2×2 nonsingular matrix Λ, M ′ = M Λ . {\displaystyle M'=M\Lambda .} In particular, rows i and j of M′ and M are related by [ x i ′ y i ′ x j ′ y j ′ ] = [ x i y i x j y j ] [ λ 00 λ 01 λ 10 λ 11 ] . {\displaystyle {\begin{bmatrix}x'_{i}&y'_{i}\\x'_{j}&y'_{j}\end{bmatrix}}={\begin{bmatrix}x_{i}&y_{i}\\x_{j}&y_{j}\end{bmatrix}}{\begin{bmatrix}\lambda _{00}&\lambda _{01}\\\lambda _{10}&\lambda _{11}\end{bmatrix}}.} Therefore, the determinant of the left side 2×2 matrix equals the product of the determinants of the right side 2×2 matrices, the latter of which is a fixed scalar, det Λ. Furthermore, all six 2×2 subdeterminants in M cannot be zero because the rank of M is
{"page_id": 1168486, "title": "Plücker coordinates"}
human eye's minimum illuminance (~0.1 mlx) while moonlight coincides with the human eye's minimum colour vision illuminance (~50 mlx). One of the oldest stars yet identified - oldest but not most distant in this case - was identified in 2014: while "only" 6,000 light years away, the star SMSS J031300.36−670839.3 was determined to be 13.8 billion years old, or more or less the same age as the universe itself. The starlight shining on Earth includes this star. == Photography == Night photography includes photographing subjects that are lit primarily by starlight. Directly taking images of night sky is also a part of astrophotography. Like other photography, it can be used for the pursuit of science and/or leisure. Subjects include nocturnal animals. In many cases starlight photography may also overlap with a need to understand the impact of moonlight. == Polarization == Starlight intensity has been observed to be a function of its polarization. Starlight becomes partially linearly polarized by scattering from elongated interstellar dust grains whose long axes tend to be oriented perpendicular to the galactic magnetic field. According to the Davis–Greenstein mechanism, the grains spin rapidly with their rotation axis along the magnetic field. Light polarized along the direction of the magnetic field perpendicular to the line of sight is transmitted, while light polarized in the plane defined by the rotating grain is blocked. Thus the polarization direction can be used to map the galactic magnetic field. The degree of polarization is on the order of 1.5% for stars at 1,000 parsecs' distance. Normally, a much smaller fraction of circular polarization is found in starlight. Serkowski, Mathewson and Ford measured the polarization of 180 stars in UBVR filters. They found a maximum fractional circular polarization of q = 6 × 10 − 4 {\displaystyle q=6\times 10^{-4}} , in the
{"page_id": 28996756, "title": "Starlight"}
IGZO transistors first showed up in 2012, and were first manufactured by Sharp Corporation. IGZO allows for higher refresh rates and lower power consumption. In 2021, the first flexible 32-bit microprocessor was manufactured using IGZO TFT technology on a polyimide substrate. == See also == Metal oxide thin film transistor Organic field effect transistor == References ==
{"page_id": 199410, "title": "Thin-film transistor"}
(see arXiv.) == Set theory == === "Über eine Eigenschaft des Inbegriffes aller reellen algebraischen Zahlen" === Georg Cantor (1874) Online version: Online version Contains the first proof that the set of all real numbers is uncountable; also contains a proof that the set of algebraic numbers is countable. (See Georg Cantor's first set theory article.) === Grundzüge der Mengenlehre === Felix Hausdorff First published in 1914, this was the first comprehensive introduction to set theory. Besides the systematic treatment of known results in set theory, the book also contains chapters on measure theory and topology, which were then still considered parts of set theory. Here Hausdorff presents and develops highly original material which was later to become the basis for those areas. === "The consistency of the axiom of choice and of the generalized continuum-hypothesis with the axioms of set theory" === Kurt Gödel (1938) Gödel proves the results of the title. Also, in the process, introduces the class L of constructible sets, a major influence in the development of axiomatic set theory. === "The Independence of the Continuum Hypothesis" === Paul J. Cohen (1963, 1964) Cohen's breakthrough work proved the independence of the continuum hypothesis and axiom of choice with respect to Zermelo–Fraenkel set theory. In proving this Cohen introduced the concept of forcing which led to many other major results in axiomatic set theory. == Logic == === The Laws of Thought === George Boole (1854) Published in 1854, The Laws of Thought was the first book to provide a mathematical foundation for logic. Its aim was a complete re-expression and extension of Aristotle's logic in the language of mathematics. Boole's work founded the discipline of algebraic logic and would later be central for Claude Shannon in the development of digital logic. === Begriffsschrift === Gottlob Frege
{"page_id": 708399, "title": "List of publications in mathematics"}
progress towards the goal—no action will ever undo progress made by another action. It turns out it is still NP-hard to find the optimal solution to this relaxed problem, but an approximate solution can be found in polynomial time by hill-climbing. Figure 10.6 diagrams part of the state space for two planning problems using the ignore-delete-lists heuristic. The dots represent states and the edges actions, and the height of each dot above the bottom plane represents the heuristic value. States on the bottom plane are solutions. In both these problems, there is a wide path to the goal. There are no dead ends, so no need for backtracking; a simple hillclimbing search will easily find a solution to these problems (although it may not be an optimal solution). The relaxed problems leave us with a simplified—but still expensive—planning prob-lem just to calculate the value of the heuristic function. Many planning problems have 10 100 states or more, and relaxing the actions does nothing to reduce the number of states. There-fore, we now look at relaxations that decrease the number of states by forming a state ab-straction —a many-to-one mapping from states in the ground representation of the problem STATE ABSTRACTION to the abstract representation. The easiest form of state abstraction is to ignore some fluents. For example, consider an air cargo problem with 10 airports, 50 planes, and 200 pieces of cargo. Each plane can be at one of 10 airports and each package can be either in one of the planes or unloaded at one of the airports. So there are 50 10 × 200 50+10 ≈ 10 155 states. Now consider a particular problem in that domain in which it happens that all the packages are at just 5 of the airports, and all packages at a given
{"source": 1019, "title": "from dpo"}
at: Spotlight Poster #809 Relax, it doesn’t matter how you get there: A new self-supervised approach for multi-timescale behavior analysis Mehdi Azabou · Michael Mendelson · Nauman Ahad · Maks Sorokin · Shantanu Thakoor · Carolina Urzay · Eva Dyer Unconstrained and natural behavior consists of dynamics that are complex and unpredictable, especially when trying to predict what will happen multiple steps into the future. While some success has been found in building representations of animal behavior under constrained or simplified task-based conditions, many of these models cannot be applied to free and naturalistic settings where behavior becomes increasingly hard to model. In this work, we develop a multi-task representation learning model for animal behavior that combines two novel components: (i) an action-prediction objective that aims to predict the distribution of actions over future timesteps, and (ii) a multi-scale architecture that builds separate latent spaces to accommodate short- and long-term dynamics. After demonstrating the ability of the method to build representations of both local and global dynamics in robots in varying environments and terrains, we apply our method to the MABe 2022 Multi-Agent Behavior challenge, where our model ranks first overall on both mice and fly benchmarks. In all of these cases, we show that our model can build representations that capture the many different factors that drive behavior and solve a wide range of downstream tasks. Poster #810 Modality-Agnostic Self-Supervised Learning with Meta-Learned Masked Auto-Encoder Huiwon Jang · Jihoon Tack · Daewon Choi · Jongheon Jeong · Jinwoo Shin Despite its practical importance across a wide range of modalities, recent advances in self-supervised learning (SSL) have been primarily focused on a few well-curated domains, e.g., vision and language, often relying on their domain-specific knowledge. For example, Masked Auto-Encoder (MAE) has become one of the popular architectures in these
{"source": 2773, "title": "from dpo"}
the Seventh International Conference on Multimodal Interfaces (pp. 138–143). Asso - ciation for Computing Machinery. Goyal, K., Jauhar, S. K., Li, H., Sachan, M., Srivastava, S., & Hovy, E. (2013). A structured distributional seman - tic model: Integrating structure with semantics. Proceedings of the Workshop on Continuous Vector Space Models and Their Compositionality (pp. 20–29). The Association for Computational Linguistics. Graff, D., & Cieri, C. (2003). En glish Gigaword (LDC2003T05). Linguistic Data Consortium. https:// catalog . ldc .upenn .edu / LDC2003T05 Griffiths, T. L. (2009). Connecting human and machine learning via probabilistic models of cognition. Proceed - ings of the 10th Annual Conference of the International Speech Communication Association (pp. 9–12). ISCA. Grishman, R., & Sundheim, B. (1996). Message Understanding Conference—6: A brief history. Proceedings of the 16th International Conference on Computational Linguistics (pp. 466–471). The Association for Com - putational Linguistics. > Downloaded from by guest on 05 June 2025 References 403 Grosz, B., Joshi, A. K., & Weinstein, S. (1995). Centering: A framework for modelling the local coherence of discourse. Computational Linguistics , 2(21), 203–225. Grove, W. M., & Lloyd, M. (2006). Meehl’s contribution to clinical versus statistical prediction. Journal of Abnormal Psy chol ogy , 115 (2), 192–194. Guarino, N. (1998). Formal ontology in information systems. In N. Guarino (Ed.), Formal ontology in infor - mation systems (pp. 3–15). IOS Press. Gunzelmann, G., Gross, J. B., Gluck, K. A., & Dinges, D. F. (2009). Sleep deprivation and sustained attention per for mance: Integrating mathematical and cognitive modeling. Cognitive Science , 33 (5), 880–910. Hahn, U., Romacker, M., & Schulz, S. (1999). How knowledge drives understanding— Matching medical ontol - ogies with the needs of medical language pro cessing. Artificial Intelligence in Medicine , 15 (1), 25–51. Hajič, J., Hajičová, E., Mikulová, M., Mírovský, J., Panevová, J.,
{"source": 4999, "title": "from dpo"}
much of a scholarly record, whereas renowned scholars field many requests to review and may not be available). Editors may not pick all (or any) of the reviewers you recommend. It is generally considered inappropriate to suggest reviewers who are colleagues with whom you directly work because their familiarity with you and your work may bias the review process, particularly if the journal does not participate in masked review. Masked Review. Journal editors, either routinely or at the authors’ request, may use masked review in which the identities of the authors of a manuscript Instagram and Telegram: @PDFEnglish are concealed from reviewers during the review process. Authors’ names are not revealed to reviewers without the consent of the authors until the review process is complete. The names of reviewers can also be masked, although a reviewer may choose to reveal their identity by including their name in the review itself. Thus, journals can mask review in both directions (both authors’ and reviewers’ names masked), mask review in one direction (either authors’ or reviewers’ names masked), or unmask the review (all names revealed). Consult the author guidelines to determine whether the journal to which you are submitting routinely uses masked review or offers masked review to authors who request it. Authors are responsible for concealing their identities in manuscripts that will receive masked review; for example, they should format their manuscripts so their identities as document creators are not easily revealed and describe the locations of their studies and participants in general terms (e.g., “students at a small, Midwestern U.S. liberal arts college”). Timing of Peer Review. The time required to complete a peer review varies depending on the length and complexity of the manuscript and the number of reviewers asked to evaluate it. If the journal editor determines that the
{"source": 6367, "title": "from dpo"}
can be indexed with the help of the asymptotic size dependence of a quantity like RX. If RX falls off to zero with size, it is self-averaging whereas if RX approaches a constant as N → ∞, the system is non-self-averaging. == Strong and weak self-averaging == There is a further classification of self-averaging systems as strong and weak. If the exhibited behavior is RX ~ N−1 as suggested by the central limit theorem, mentioned earlier, the system is said to be strongly self-averaging. Some systems shows a slower power law decay RX ~ N−z with 0 < z < 1. Such systems are classified weakly self-averaging. The known critical exponents of the system determine the exponent z. It must also be added that relevant randomness does not necessarily imply non self-averaging, especially in a mean-field scenario. The RG arguments mentioned above need to be extended to situations with sharp limit of Tc distribution and long range interactions. == References ==
{"page_id": 5782346, "title": "Self-averaging"}
The Department of Chemical Engineering and Biotechnology (CEB) is one of the teaching and research departments at the University of Cambridge. The department trains undergraduate students and conducts original research at the interfaces between engineering, chemistry, biology and physics. It conducts research in collaboration with industrial partners. Its research programmes encompass sustainable reaction engineering, chemical product and process design, healthcare, measurement, and materials science. It conducts biotechnology research with chemical engineering at the science-engineering interface. == Notable staff == As of 2020 notable staff include Sabine Bahn John Bridgwater Silvana Cardoso Howard Chase Lynn Gladden Elizabeth A. H. Hall Ross D. King Róisín Owens === Heads of department === These include heads of the former Department of Chemical Engineering and Institute of Biotechnology which merged to form the current department. Lisa Hall John Dennis Lynn Gladden John Davidson Peter Danckwerts === Alumni === Former staff include: Francis Thomas Bacon == History == In 1945, the university received an endowment from Shell for a chemical engineering department and chair. The first Shell Professor was Terence Fox, appointed in 1946. The undergraduate Tripos course began in 1948. Peter Danckwerts was head of department from 1959 to 1975 and then John Davidson became Shell Professor and Head of department in 1975. He held the post until 1993 when he retired. in 2008, the Department of Chemical Engineering merged with the Institute of Biotechnology to become the Department of Chemical Engineering and Biotechnology. Until 2017, the department's main centre of activity was the Shell building on Pembroke Street on the New Museums Site, to the south of Cambridge city centre. In 2017, the department moved over to a new building on Philippa Fawcett Drive on the West Cambridge site. The building was officially opened by the chancellor of the university, David Sainsbury, Baron Sainsbury
{"page_id": 16882581, "title": "Department of Chemical Engineering and Biotechnology, University of Cambridge"}
Inca animal husbandry refers to how in the pre-Hispanic andes, camelids played a truly important role in the economy. In particular, the llama and alpaca—the only camelids domesticated by Andean people— which were raised in large-scale houses and used for different purposes within the production system of the Incas. Likewise, two other species of undomesticated camelids were used: the vicuña and the guanaco. The guanacos were hunted by means of chacos (collective hunts). The Inca people used tools such as: stones, knives or tumis, axes that, according to chroniclers, were made of stone and bronze, and ropes that were made by them in their leisure time. Many of these tools were used to shear the camelids, which were then set free; in this way, they ensured that their numbers were maintained. Guanacos, on the other hand, were hunted for their meat, which was highly prized. == Camelid raising == The South American camelids were a valuable resource. Their meat was consumed fresh or in charqui and chalona; their wool was used to make threads and fabrics; their bones, hide, fat and excrement had diverse applications such as musical instruments, footwear, medicines and fertilizer respectively. They were also preferred animals for religious sacrifices. The communal camelid herds were under the care of young people, whose ages ranged from twelve to sixteen years old. In areas where the communal herds were large, such as the altiplano region, where pastures were far away, it is likely that their care was in the hands of a full-time specialist. The chroniclers mention two Quechua names for the shepherds: llama michi—which Garcilaso associates with low social status—and llama camayos, which designated the llamas caretaker or employee responsible for the herds. The state herdsmen were responsible for the animals under their charge, whose accounting and supervision were
{"page_id": 71304235, "title": "Inca animal husbandry"}
residential and commercial drinking water supply in many countries, as well as for industrial self-supply with water. However, it is less common in irrigated agriculture, which is the major water user worldwide. Water metering is also uncommon for piped drinking water supply in rural areas and small towns, although there are examples of successful metering in rural areas in developing countries, such as in El Salvador. Metering of water supplied by utilities to residential, commercial and industrial users is common in most developed countries, except for the United Kingdom where only about 52% of users are metered. In some developing countries metering is very common, such as in Chile where it stands at 96%, while in others it still remains low, such as in Argentina. The percentage of residential water metering in selected cities in developing countries is as follows: 99% in Santiago de Chile (1998) 96% in Abidjan, Ivory Coast (1987) 62% in cities in Guatemala (2000) 30% in Lima, Peru (1991) 28% in Kathmandu, Nepal (2001) 2% in Buenos Aires, Argentina (1992) Nearly two-thirds of OECD countries meter more than 90% of single-family houses. A few are also expanding their metering of apartments (e.g., France and Germany). == Benefits == The benefits of metering are that: in conjunction with volumetric pricing it provides an incentive for water conservation, it helps to detect water leaks in the distribution network, thus providing a basis for reducing the amount of non-revenue water; it is a precondition for quantity-targeting of water subsidies to the poor. == Costs == The costs of metering include: Investment costs to purchase, install and replace meters, Recurring costs to read meters and issue bills based on consumption instead of bills based on monthly flat fees. While the cost of purchasing residential meters is low, the total life
{"page_id": 12086946, "title": "Water metering"}
The VHF Data Link or VHF Digital Link (VDL) is a means of sending information between aircraft and ground stations (and in the case of VDL Mode 4, other aircraft) over VHF. Aeronautical VHF data links use the band 117.975–137 MHz assigned by the International Telecommunication Union to Aeronautical mobile (R) service. There are ARINC standards for ACARS on VHF and other data links installed on approximately 14,000 aircraft and a range of ICAO standards defined by the Aeronautical Mobile Communications Panel (AMCP) in the 1990s. Mode 2 is the only VDL mode being implemented operationally to support Controller Pilot Data Link Communications (CPDLC). == ICAO VDL Mode 1 == The ICAO AMCP defined this Mode for validation purposes. It was the same as VDL Mode 2 except that it used the same VHF link as VHF ACARS so it could be implemented using analog radios before VHF Digital Radio implementation was completed. The ICAO AMCP completed validation of VDL Modes 1&2 in 1994, after which the Mode 1 was no longer needed and was deleted from the ICAO standards. == ICAO VDL Mode 2 == The ICAO VDL Mode 2 is the main version of VDL. It has been implemented in a Eurocontrol Link 2000+ program and is specified as the primary link in the EU Single European Sky rule adopted in January 2009 requiring all new aircraft flying in Europe after January 1, 2014 to be equipped with CPDLC. In advance of CPDLC implementation, VDL Mode 2 has already been implemented in approximately 2,000 aircraft to transport ACARS messages simplifying the addition of CPDLC. Networks of ground stations providing VDL Mode 2 service have been deployed by ARINC and SITA with varying levels of coverage. The ICAO standard for the VDL Mode 2 specifies three layers: the Subnetwork,
{"page_id": 19718830, "title": "VHF Data Link"}
Research indicates that living in areas of high pollution has serious long term health effects. Living in these areas during childhood and adolescence can lead to diminished mental capacity and an increased risk of brain damage. People of all ages who live in high pollution areas for extended periods place themselves at increased risk of various neurological disorders. Both air pollution and heavy metal pollution have been implicated as having negative effects on central nervous system (CNS) functionality. The ability of pollutants to affect the neurophysiology of individuals after the structure of the CNS has become mostly stabilized is an example of negative neuroplasticity. == Air pollution == Air pollution may increase the risk of developmental disorders (e.g., autism), neurodegenerative disorders, mental disorders, and suicide. It is associated with neurological conditions including stroke, multiple sclerosis, dementia, Parkinson disease, Alzheimer's disease, schizophrenia and headaches. === Effects in adolescents === A 2008 study compared children and dogs raised in Mexico City (a location known for high pollution levels) with children and dogs raised in Polotitlán, Mexico (a city whose pollution levels meet the current US National Ambient Air Quality Standards). Children raised in areas of higher pollution were found to score lower in intelligence (i.e., on IQ tests), and showed signs of lesions in MRI scanning of the brain. In contrast, children from the low pollution area scored as expected on IQ tests and showed no significant sign of the risk of brain lesions. Concerning traffic-related air pollution, children of mothers exposed to higher levels during the first trimester of pregnancy were at increased risk of allergic sensitization at one year age. === Effects in adults === Effects of physical activity and air pollution on neuroplasticity may counteract. Physical activity is known for its benefits to the cardiovascular system, brain plasticity processes,
{"page_id": 69561115, "title": "Brain health and pollution"}
8000 and 8002 Microprocessor Development System (MDS), the 8000 being a software development system running editors and cross-compilers with twin floppy disk storage, while the 8002 could be fitted with real-time emulators for several processors including the 8080, Z80 and 6502. These were later replaced by the 8540/8550/8560. The 8550 was a standalone development system capable of being fitted with real time emulation hardware for several different processors. The 8560 emerged just after Bell Labs were able to sell UNIX commercially, and ran a mildly modified version of Version 7 UNIX, called TNIX, supporting 4 or 8 (depending on how many I/O processor cards were fitted) serial terminals, with a special High-Speed Input/Output, based on RS-422, to connect to a remote 8540. There was a hard drive of about 34MB, later increased, and a variable amount of RAM depending on the processor card fitted, which was either a DEC LSI11-23+ or a LSI11-73, which had a wider address bus as well as a higher clock speed. The final upgraded variant was the 8562. The 8540, of which several could be connected to the 8560, contained emulation hardware similar to that available for the 8550, including RAM cards, Trigger Trace Analyser, 1 or 2 processor emulator cards, each with external probe, etc. The 8540 had no software development capability, but worked with the 8560, or other systems such as a VAX, which would provide the software development environment. It was unique in its day, in that the serial terminal could be connected to the 8540, on the bench, or to the 8560, and would pass UNIX commands automatically to the 8560, which passed emulator commands back to the 8540 for execution, completely transparently. Thus a complete compile, link, load and emulate session could be run from a UNIX shell script, a
{"page_id": 659213, "title": "Tektronix"}
without real-world observation. Land change modeling is useful to explore spatial land systems, uses, and covers. Land change modeling can account for complexity within dynamics of land use and land cover by linking with climatic, ecological, biogeochemical, biogeophysical and socioeconomic models. Additionally, LCMs are able to produce spatially explicit outcomes according to the type and complexity within the land system dynamics within the spatial extent. Many biophysical and socioeconomic variables influence and produce a variety of outcomes in land change modeling. == Model uncertainty == A notable property of all land change models is that they have some irreducible level of uncertainty in the model structure, parameter values, and/or input data. For instance, one uncertainty within land change models is a result from temporal non-stationarity that exists in land change processes, so the further into the future the model is applied, the more uncertain it is. Another uncertainty within land change models are data and parameter uncertainties within physical principles (i.e., surface typology), which leads to uncertainties in being able to understand and predict physical processes. Furthermore, land change model design are a product of both decision-making and physical processes. Human-induced impact on the socio-economic and ecological environment is important to take into account, as it is constantly changing land cover and sometimes model uncertainty. To avoid model uncertainty and interpret model outputs more accurately, a model diagnosis is used to understand more about the connections between land change models and the actual land system of the spatial extent. The overall importance of model diagnosis with model uncertainty issues is its ability to assess how interacting processes and the landscape are represented, as well as the uncertainty within the landscape and its processes. == Approaches == === Machine learning and statistical models === A machine-learning approach uses land-cover data from
{"page_id": 53913187, "title": "Land change modeling"}
the equinoxes and summer solstice were not given much attention. Winter solstice marks both the darkest part of winter and the turning point when light begins to increase, marking the promise of the Sun's return. The first appearance of Aagjuuk happens around mid-December and is used across the Arctic to signal winter solstice's arrival. For some tribes, this would also signal the time for a midwinter celebration. Some constellations have only seasonal appearances, which help mark the passage of time. For example, Ullakut (Orion) and Sakiattiak (Taurus) are only visible in the winter. Throughout winter, many stars within Tukturjuit (Big Dipper) were used as hour hands to keep track of time during the night or as calendar stars to determine the date. Aagjuuk (Aquila) and Kingulliq (Lyra) begin to appear near the end of winter, signaling that light will be returning to the region. == Spiritual cosmology == Inuit tradition closely links the Earth and sky, with a spatial understanding of the Earth as a large flat disk ending in cliffs and surrounded by sky. The sky itself is understood as four to five layers of celestial realms. Each layer is a separate land of the dead and its own world. The aurora borealis bears special significance as the place where spirits who died from blood loss, murder or childbirth dwell. Legends warn Inuit against wrongdoing and taboo acts by telling the stories of people being transformed into stars after committing transgressions. For example, "the ubiquitous Inuit epic in which greed, murder, incest, and retribution account for the creation of the sun, moon, and the first stars." == Constellations == The Inuit have traditional names for many constellations, asterisms and stars. Inuit astronomy names thirty-three individual stars, two star clusters, and one nebula. The stars are incorporated into 16 or
{"page_id": 18561915, "title": "Inuit astronomy"}
since the probability is simply zero or unity. Furthermore, under the frequentist construct the model parameters are unknown constants and not permitted to be random variables. For example if θ 1 = 5 {\displaystyle \theta _{1}=5} , then P ( 2 ≤ 5 ≤ 10 ) = 1 {\displaystyle P(2\leq 5\leq 10)=1} . Likewise, if θ 1 = 11 {\displaystyle \theta _{1}=11} , then P ( 2 ≤ 11 ≤ 10 ) = 0 {\displaystyle P(2\leq 11\leq 10)=0} As Neyman describes in his 1937 paper, suppose that we consider all points in the sample space, that is, ∀ X ∈ Θ {\displaystyle \forall X\in \Theta } , which are a system of random variables defined by the joint pdf described above. Since L {\displaystyle L} and U {\displaystyle U} are functions of X {\displaystyle X} they too are random variables and one can examine the meaning of the following probability statement: Under the frequentist construct the model parameters are unknown constants and not permitted to be random variables. Considering all the sample points in the sample space as random variables defined by the joint pdf above, that is all X ∈ Θ {\displaystyle X\in \Theta } it can be shown that L {\displaystyle L} and U {\displaystyle U} are functions of random variables and hence random variables. Therefore one can look at the probability of L ( X ) {\displaystyle L(X)} and U ( X ) {\displaystyle U(X)} for some X ∈ Θ {\displaystyle X\in \Theta } . If θ 1 ′ {\displaystyle \theta _{1}^{'}} is the true value of θ 1 {\displaystyle \theta _{1}} , we can define L {\displaystyle L} and U {\displaystyle U} such that the probability L ( X ) ≤ θ 1 ′ {\displaystyle L(X)\leq \theta _{1}^{'}} and θ 1 ′ ≤ U ( X
{"page_id": 9164216, "title": "Neyman construction"}
the import of certain animals from Africa. This was the first time monkeypox was found in the US. Though no one died, many were seriously ill. The transport of new animals into the US needs to be carefully monitored to make sure that foreign illnesses don't cause harm to people and native animals. _**Note:** Several aspects of the original outbreaks and investigations have been altered to fit the format and length of this application._ Back To Top A strange outbreak seems to have hit a single county in the desert Southwest. A number of people are reporting severe flu-like symptoms. But it's early summer, an unusual time for flu season to kick in. And no other counties are reporting anything similar. Two adults have died, and a number of children are sick as well. The local doctors and nurses are baffled. They need your help to solve the outbreak before it spreads! #### Clue 1 ![Image 132: No one can tell where the illness has come from.]( It started when a 9-year-old boy was taken to his doctor. He'd been sick 2 days with a fever of 103 degrees, a cough, headache, and chills. The doctor said it was a bad virus and sent him home to get some rest, take medicine for the fever, and drink lots of fluids. He returned to his doctor the next day with nausea and stomach pain, and was again sent home. After another 3 days, his doctor checked him in to the hospital with a fever of 106 degrees and trouble breathing. He died hours later. The local department of health calls you after learning
{"source": 1463, "title": "from dpo"}
on the curve, with points closest on the curve given the greatest weight. In the case where the subspace is constrained to be linear, the procedure converges to the first principal component and is equivalent to the power method for finding the largest eigenvector of the co-variance matrix. Principal curves can be generalized to multidimensional manifolds called principal surfaces although these have found limited use due to the difficulty of data smoothing in higher dimensions even for two-dimensional manifolds. PCA is often used to project a data set onto a lower-dimensional space, for ex-ample two dimensional, for the purposes of visualization. Another linear technique with a similar aim is multidimensional scaling, or MDS (Cox and Cox, 2000). It finds a low-dimensional projection of the data such as to preserve, as closely as possible, the pairwise distances between data points, and involves finding the eigenvectors of the distance matrix. In the case where the distances are Euclidean, it gives equivalent results to PCA. The MDS concept can be extended to a wide variety of data types specified in terms of a similarity matrix, giving nonmetric MDS. Two other nonprobabilistic methods for dimensionality reduction and data vi-sualization are worthy of mention. Locally linear embedding, or LLE (Roweis and Saul, 2000) first computes the set of coefficients that best reconstructs each data point from its neighbours. These coefficients are arranged to be invariant to rota-tions, translations, and scalings of that data point and its neighbours, and hence they characterize the local geometrical properties of the neighbourhood. LLE then maps the high-dimensional data points down to a lower dimensional space while preserv-ing these neighbourhood coefficients. If the local neighbourhood for a particular data point can be considered linear, then the transformation can be achieved using a combination of translation, rotation, and scaling, such as
{"source": 3852, "title": "from dpo"}
controversial topics, such as "religion", "gender identity", "freedom of speech", and "equality", among others, can be a source of conflict as partisan or biased answers can reinforce preconceived notions or promote disinformation. by exposing chatgpt to such debatable questions, we aim to understand its level of awareness and if existing models are subject to socio-political and/or economic biases. we also aim to explore how ai-generated answers compare to human ones. for exploring this, we use a dataset of a social media platform created for the purpose of debating human-generated claims on polemic subjects among users, dubbed kialo. our results show that while previous versions of chatgpt have had important issues with controversial topics, more recent versions of chatgpt (gpt-3.5-turbo) are no longer manifesting significant explicit biases in several knowledge areas. in particular, it is well-moderated regarding economic aspects. however, it still maintains degrees of implicit libertarian leaning toward right-winged ideals which suggest the need for increased moderation from the socio-political point of view. in terms of domain knowledge on controversial topics, with the exception of the "philosophical" category, chatgpt is performing well in keeping up with the collective human level of knowledge. finally, we see that sources of bing ai have slightly more tendency to the center when compared to human answers. all the analyses we make are generalizable to other types of biases and domains. Expand Abstract Challenges of GPT-3-Based Conversational Agents for Healthcare Fabian Lechner, Allison Lahnala, Charles Welch, Lucie Flek Abstract: the potential to provide patients with faster information access while allowing medical specialists to concentrate on critical tasks makes medical domain dialog agents appealing. however, the integration of large-language models (llms) into these agents presents certain limitations that may result in serious consequences. this paper investigates the challenges and risks of using gpt-3-based models for medical
{"source": 5788, "title": "from dpo"}
for research and suggestions for the most suitable techniques for future practice in biological dosimetry. One CRP resulted in the publication in 1986 of a manual, entitled Biological Dosimetry: Chromosomal Aberration Analysis for Dose Assessment (Technical Reports Series No. 260). This was superseded in 2001 by a revised second edition, Technical Reports Series No. 405. This present publication constitutes a third edition, with extensive updating to reflect the considerable advances that have been made in cytogenetic biological dosimetry during the past decade. The IAEA wishes to express its thanks to all authors and reviewers of this publication. The major contributions of Dr. D. Lloyd are especially acknowledged. This publication has been co-sponsored by the Pan American Health Organization and the World Health Organization. The IAEA officer responsible for this publication was E. Buglova of the Department of Nuclear Safety and Security. EDITORIAL NOTE The use of particular designations of countries or territories does not imply any judgement by the publisher, the IAEA, as to the legal status of such countries or territories, of their authorities and institutions or of the delimitation of their boundaries. The mention of names of specific companies or products (whether or not indicated as registered) does not imply any intention to infringe proprietary rights, nor should it be construed as an endorsement or recommendation on the part of the IAEA. CONTENTS 1. INTRODUCTION .............................................................................................................1 1.1. Background ..............................................................................................................1 1.2. Objective ..................................................................................................................4 1.3. Scope and History of Development .........................................................................4 1.4. Structure ...................................................................................................................5 2. APPLICATION OF DOSE CONCEPTS IN BIOLOGICAL DOSIMETRY ...................7 3. BIOPHYSICAL BACKGROUND TO CHROMOSOME DAMAGE .............................9 4. HUMAN LYMPHOCYTES............................................................................................15 5. CHROMOSOMAL STRUCTURE .................................................................................19 5.1. Chromatin Packing .................................................................................................19 5.2. Human Karyotype and DNA Content of Chromosomes........................................19 5.3. Cell Cycle ...............................................................................................................23 6. RADIATION INDUCED CHROMOSOMAL ALTERATIONS ...................................25 6.1. Radiation Induced DNA Lesions ..............................................................................25 6.2. Chromosome-Type Aberrations ...............................................................................27 6.3.
{"source": 7334, "title": "from dpo"}
Haplogroup JT is a human mitochondrial DNA (mtDNA) haplogroup. == Origin == Haplogroup JT is descended from the macro-haplogroup R. It is the ancestral clade to the mitochondrial haplogroups J and T. == Distribution == JT (predominantly J) was found among the ancient Etruscans. The root level haplogroup JT* has been assigned to an ancient person found at the Colfiorito necropolis in Umbria in central Italy. The haplogroup has also been found among Iberomaurusian specimens dating from the Epipaleolithic at the Taforalt prehistoric site. One ancient individual carried a haplotype, which correlates with either the JT clade or the haplogroup H subclade H14b1 (1/9; 11%). == Subclades == === Tree === This phylogenetic tree of haplogroup JT subclades is based on the paper by Mannis van Oven and Manfred Kayser Updated comprehensive phylogenetic tree of global human mitochondrial DNA variation and subsequent published research. R2'JT JT J T === Health === Maternally inherited ancient mtDNA variants have clear impact on the presentation of disease in a modern society. Superhaplogroup JT is an example of reduced risk of Parkinson's disease And mitochondrial and mtDNa alterations continue to be promising disease biomarkers. == See also == Genealogical DNA test Genetic genealogy Human mitochondrial genetics Population genetics == References == == External links == General Ian Logan's Mitochondrial DNA Site Mannis van Oven's Phylotree
{"page_id": 4152068, "title": "Haplogroup JT"}
In mathematics and machine learning, the softplus function is f ( x ) = log ⁡ ( 1 + e x ) . {\displaystyle f(x)=\log(1+e^{x}).} It is a smooth approximation (in fact, an analytic function) to the ramp function, which is known as the rectifier or ReLU (rectified linear unit) in machine learning. For large negative x {\displaystyle x} it is log ⁡ ( 1 + e x ) = log ⁡ ( 1 + ϵ ) ⪆ log ⁡ 1 = 0 {\displaystyle \log(1+e^{x})=\log(1+\epsilon )\gtrapprox \log 1=0} , so just above 0, while for large positive x {\displaystyle x} it is log ⁡ ( 1 + e x ) ⪆ log ⁡ ( e x ) = x {\displaystyle \log(1+e^{x})\gtrapprox \log(e^{x})=x} , so just above x {\displaystyle x} . The names softplus and SmoothReLU are used in machine learning. The name "softplus" (2000), by analogy with the earlier softmax (1989) is presumably because it is a smooth (soft) approximation of the positive part of x, which is sometimes denoted with a superscript plus, x + := max ( 0 , x ) {\displaystyle x^{+}:=\max(0,x)} . == Related functions == The derivative of softplus is the logistic function: f ′ ( x ) = e x 1 + e x = 1 1 + e − x {\displaystyle f'(x)={\frac {e^{x}}{1+e^{x}}}={\frac {1}{1+e^{-x}}}} The logistic sigmoid function is a smooth approximation of the derivative of the rectifier, the Heaviside step function. === LogSumExp === The multivariable generalization of single-variable softplus is the LogSumExp with the first argument set to zero: L S E 0 + ⁡ ( x 1 , … , x n ) := LSE ⁡ ( 0 , x 1 , … , x n ) = ln ⁡ ( 1 + e x 1 + ⋯ + e
{"page_id": 37862957, "title": "Softplus"}
begin with two piles of a and b stones. The players take turns removing m multiples of the smaller pile from the larger. Thus, if the two piles consist of x and y stones, where x is larger than y, the next player can reduce the larger pile from x stones to x − my stones, as long as the latter is a nonnegative integer. The winner is the first player to reduce one pile to zero stones. == Mathematical applications == === Bézout's identity === Bézout's identity states that the greatest common divisor g of two integers a and b can be represented as a linear sum of the original two numbers a and b. In other words, it is always possible to find integers s and t such that g = sa + tb. The integers s and t can be calculated from the quotients q0, q1, etc. by reversing the order of equations in Euclid's algorithm. Beginning with the next-to-last equation, g can be expressed in terms of the quotient qN−1 and the two preceding remainders, rN−2 and rN−3: g = rN−1 = rN−3 − qN−1 rN−2. Those two remainders can be likewise expressed in terms of their quotients and preceding remainders, rN−2 = rN−4 − qN−2 rN−3 and rN−3 = rN−5 − qN−3 rN−4. Substituting these formulae for rN−2 and rN−3 into the first equation yields g as a linear sum of the remainders rN−4 and rN−5. The process of substituting remainders by formulae involving their predecessors can be continued until the original numbers a and b are reached: r2 = r0 − q2 r1 r1 = b − q1 r0 r0 = a − q0 b. After all the remainders r0, r1, etc. have been substituted, the final equation expresses g as a linear sum
{"page_id": 10377, "title": "Euclidean algorithm"}
Analyzed Layout and Text Object (ALTO) is an open XML Schema developed by the EU-funded project called METAe. The standard was initially developed for the description of text OCR and layout information of pages for digitized material. The goal was to describe the layout and text in a form to be able to reconstruct the original appearance based on the digitized information - similar to the approach of a lossless image saving operation. ALTO is often used in combination with Metadata Encoding and Transmission Standard (METS) for the description of the whole digitized object and creation of references across the ALTO files, e.g. reading sequence description. The standard is hosted by the Library of Congress since 2010 and maintained by the Editorial Board initialized at the same time. In the time from the final version of the ALTO standard in June 2004 (version 1.0) ALTO was maintained by CCS CCS Content Conversion Specialists GmbH, Hamburg up to version 1.4. == Structure == An ALTO file consists of three major sections as children of the root <alto> element: <Description> section contains metadata about the ALTO file itself and processing information on how the file was created. <Styles> section contains the text and paragraph styles with their individual descriptions: <TextStyle> has font descriptions <ParagraphStyle> has paragraph descriptions, e.g. alignment information <Layout> section contains the content information. It is subdivided into <Page> elements. == Software support == ABBYY FineReader eScriptorium Kitodo Tesseract OCR Transkribus == See also == Metadata Encoding and Transmission Standard (METS) Dublin Core, an ISO metadata standard Preservation Metadata: Implementation Strategies (PREMIS) Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH) hOCR PAGE (XML) == References == == External links == ALTO (Analyzed Layout and Text Object) standards on Library of Congress website https://altoxml.github.io/ resp. https://github.com/altoxml ALTOxml on GitHub More info
{"page_id": 33593073, "title": "Analyzed Layout and Text Object"}
probability distributions mentioned above. This has led some researchers to view MDL as equivalent to Bayesian inference: code length of model and data together in MDL correspond respectively to prior probability and marginal likelihood in the Bayesian framework. While Bayesian machinery is often useful in constructing efficient MDL codes, the MDL framework also accommodates other codes that are not Bayesian. An example is the Shtarkov normalized maximum likelihood code, which plays a central role in current MDL theory, but has no equivalent in Bayesian inference. Furthermore, Rissanen stresses that we should make no assumptions about the true data-generating process: in practice, a model class is typically a simplification of reality and thus does not contain any code or probability distribution that is true in any objective sense. In the last mentioned reference Rissanen bases the mathematical underpinning of MDL on the Kolmogorov structure function. According to the MDL philosophy, Bayesian methods should be dismissed if they are based on unsafe priors that would lead to poor results. The priors that are acceptable from an MDL point of view also tend to be favored in so-called objective Bayesian analysis; there, however, the motivation is usually different. == Other systems == Rissanen's was not the first information-theoretic approach to learning; as early as 1968 Wallace and Boulton pioneered a related concept called minimum message length (MML). The difference between MDL and MML is a source of ongoing confusion. Superficially, the methods appear mostly equivalent, but there are some significant differences, especially in interpretation: MML is a fully subjective Bayesian approach: it starts from the idea that one represents one's beliefs about the data-generating process in the form of a prior distribution. MDL avoids assumptions about the data-generating process. Both methods make use of two-part codes: the first part always represents the information
{"page_id": 331325, "title": "Minimum description length"}
but can be an Kraus operator from a quantum operation representing a system interacting with its environment. == Bosonic codes == Several proposals have been made for storing error-correctable quantum information in bosonic modes. Unlike a two-level system, a quantum harmonic oscillator has infinitely many energy levels in a single physical system. Codes for these systems include cat, Gottesman-Kitaev-Preskill (GKP), and binomial codes. One insight offered by these codes is to take advantage of the redundancy within a single system, rather than to duplicate many two-level qubits. === Binomial code === Written in the Fock basis, the simplest binomial encoding is | 0 L ⟩ = | 0 ⟩ + | 4 ⟩ 2 , | 1 L ⟩ = | 2 ⟩ , {\displaystyle |0_{\rm {L}}\rangle ={\frac {|0\rangle +|4\rangle }{\sqrt {2}}},\quad |1_{\rm {L}}\rangle =|2\rangle ,} where the subscript L indicates a "logically encoded" state. Then if the dominant error mechanism of the system is the stochastic application of the bosonic lowering operator a ^ , {\displaystyle {\hat {a}},} the corresponding error states are | 3 ⟩ {\displaystyle |3\rangle } and | 1 ⟩ , {\displaystyle |1\rangle ,} respectively. Since the codewords involve only even photon number, and the error states involve only odd photon number, errors can be detected by measuring the photon number parity of the system. Measuring the odd parity will allow correction by application of an appropriate unitary operation without knowledge of the specific logical state of the qubit. However, the particular binomial code above is not robust to two-photon loss. === Cat code === Schrödinger cat states, superpositions of coherent states, can also be used as logical states for error correction codes. Cat code, realized by Ofek et al. in 2016, defined two sets of logical states: { | 0 L + ⟩ , |
{"page_id": 892803, "title": "Quantum error correction"}
Hawkins has extensive training as an electrical engineer. Another way to describe the theory (hinted at in his book) is as a learning hierarchy of feed forward stochastic state machines. In this view, the brain is analyzed as an encoding problem, not too dissimilar from future-predicting error-correction codes. The hierarchy is a hierarchy of abstraction, with the higher level machines' states representing more abstract conditions or events, and these states predisposing lower-level machines to perform certain transitions. The lower level machines model limited domains of experience, or control or interpret sensors or effectors. The whole system actually controls the organism's behavior. Since the state machine is "feed forward", the organism responds to future events predicted from past data. Since it is hierarchical, the system exhibits behavioral flexibility, easily producing new sequences of behavior in response to new sensory data. Since the system learns, the new behavior adapts to changing conditions. That is, the evolutionary purpose of the brain is to predict the future, in admittedly limited ways, so as to change it. == Neurophysiological implementation == The hierarchies described above are theorized to occur primarily in mammalian neocortex. In particular, neocortex is assumed to consist of a large number of columns (as surmised also by Vernon Benjamin Mountcastle from anatomical and theoretical considerations). Each column is attuned to a particular feature at a given level in a hierarchy. It receives bottom-up inputs from lower levels, and top-down inputs from higher levels. (Other columns at the same level also feed into a given column, and serve mostly to inhibit the activation exclusive representations.) When an input is recognized – that is, acceptable agreement is obtained between the bottom-up and top-down sources – a column generates outputs which in turn propagate to both lower and higher levels. === Cortex === These processes
{"page_id": 1196714, "title": "Memory-prediction framework"}
The natural unit of information (symbol: nat), sometimes also nit or nepit, is a unit of information or information entropy, based on natural logarithms and powers of e, rather than the powers of 2 and base 2 logarithms, which define the shannon. This unit is also known by its unit symbol, the nat. One nat is the information content of an event when the probability of that event occurring is 1/e. One nat is equal to ⁠1/ln 2⁠ shannons ≈ 1.44 Sh or, equivalently, ⁠1/ln 10⁠ hartleys ≈ 0.434 Hart. == History == Boulton and Wallace used the term nit in conjunction with minimum message length, which was subsequently changed by the minimum description length community to nat to avoid confusion with the nit used as a unit of luminance. Alan Turing used the natural ban. == Entropy == Shannon entropy (information entropy), being the expected value of the information of an event, is inherently a quantity of the same type and with a unit of information. The International System of Units, by assigning the same unit (joule per kelvin) both to heat capacity and to thermodynamic entropy implicitly treats information entropy as a quantity of dimension one, with 1 nat = 1. Systems of natural units that normalize the Boltzmann constant to 1 are effectively measuring thermodynamic entropy with the nat as unit. When the Shannon entropy is written using a natural logarithm, H = − ∑ i p i ln ⁡ p i {\displaystyle \mathrm {H} =-\sum _{i}p_{i}\ln p_{i}} it is implicitly giving a number measured in nats. == Notes == == References == == Further reading == Reza, Fazlollah M. (1994). An Introduction to Information Theory. New York: Dover. ISBN 0-486-68210-2.
{"page_id": 3070481, "title": "Nat (unit)"}
Circuits over natural numbers are a mathematical model used in studying computational complexity theory. They are a special case of circuits. The object is a labeled directed acyclic graph the nodes of which evaluate to sets of natural numbers, the leaves are finite sets, and the gates are set operations or arithmetic operations. As an algorithmic problem, the problem is to find if a given natural number is an element of the output node or if two circuits compute the same set. Decidability is still an open question. == Formal definition == A natural number circuit is a circuit, i.e. a labelled directed acyclic graph of in-degree at most 2. The nodes of in-degree 0, the leaves, are finite sets of natural numbers, the labels of the nodes of in-degree 1 are −, where A ¯ = { x ∈ N | x ∉ A } {\displaystyle {\overline {A}}=\{x\in \mathbb {N} |x\not \in A\}} and the labels of the nodes of in-degree 2 are +, ×, ∪ and ∩, where A + B = { a + b | a ∈ A , b ∈ B } {\displaystyle A+B=\{a+b|a\in A,b\in B\}} , A × B = { a × b | a ∈ A , b ∈ B } {\displaystyle A\times B=\{a\times b|a\in A,b\in B\}} and ∪ and ∩ with the usual set meaning. The subset of circuits which do not use all of the possible labels are also studied. === Algorithmic problems === One can ask: Is a given number n a member of the output node. Is the output node empty? Is one node a subset of another. For circuits which use all the labels, all these problems are equivalent. ==== Proof ==== The first problem is reducible to the second one, by taking the intersection of the
{"page_id": 27750331, "title": "Circuits over sets of natural numbers"}
examine the interoperability and IMC (Integrated Marketing Communications) will be critical. Jooyoung Kim The role of NFTs (Non-fungible tokens), cross-cultural, ethical, and legal issues related to metaverse marketing will be important research agendas calling for our collective effort moving forward. Jooyoung Kim If customers can freely take any shape they like and behave accordingly, how can retailers interpret their intentions and create reliable profiles? Savvas Papagiannidis Will the theme of the metaverse influence consumer identities and, if so, how will such identities differ from one another? Savvas Papagiannidis How do consumers process brand information and make decisions when purchasing through virtual commerce? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal Is consumer decision making similar across different gender and age groups, and which marking approach is more effective for each group? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal Is the decision-making different from that in electronic commerce and social commerce contexts? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal What are some of the external and internal drivers for purchasing digital goods? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal It is crucial to understand how retailers’ strategies for their products' width, depth, and length in the real world and in the virtual one. For example, what type of digital goods to offer? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal How do retailers leverage virtual customer service representatives in metaverse? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal How does a brand conduct sales in the metaverse? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal How can retailers build customer relationships (e.g., acquisition, retention, and development) with consumers? Garry Wei-Han Tan, Keng-Boon Ooi, Vincent Dutot and DP Goyal It is essential to understand the
{"source": 54, "title": "from dpo"}
9 p.m. 6.10 [Discussion/15] Discuss some options to better utilize the excess servers during the off-peak hours or find ways to save costs. Given the interactive nature of WSCs, what are some of the challenges to aggressively reducing power usage? 6.11 [Discussion/25] Propose one possible way to improve TCO by focusing on reducing server power. What are the challenges to evaluating your pro-posal? Estimate the TCO improvements based on your proposal. What are some advantages and drawbacks? # Exercises 6.12 [10/10/10] One of the important enablers of WSC is ample request-level parallelism, in contrast to instruction- or thread-level parallelism. This ques-tion explores the implication of different types of parallelism on computer archi-tecture and system design. a. Discuss scenarios where improving the instruction- or thread-level parallelism would provide greater benefits than those achievable through request-level parallelism. b. What are the software design implications of increasing request-level parallelism? c. What are potential drawbacks of increasing request-level parallelism? 6.13 [Discussion/15/15] When a cloud computing service provider receives jobs consisting of multiple Virtual Machines (VMs) (e.g., a MapReduce job), many scheduling options exist. The VMs can be scheduled in a round-robin man-ner to spread across all available processors and servers, or they can be consoli-dated to use as few processors as possible. Using these scheduling options, if a job with 24 VMs was submitted and 30 processors were available in the cloud (each able to run up to 3 VMs), round-robin would use 24 processors, while con-solidated scheduling would use 8 processors. The scheduler can also find available processor cores at different scopes: socket, server, rack, and an array of racks. 522 ■ Chapter Six Warehouse-Scale Computers to Exploit Request-Level
{"source": 2300, "title": "from dpo"}
storyboard for a program that compares the growth of a $10,000 investment for a given number of years under two interest rates. > Practice It Now you can try these exercises at the end of the chapter: R4.24, R4.25, R4.26. > © Nicholas Homrich/iStockphoto. > S E L F C H E C K 4.7 Common Loop Algorithms 167 # 4.7 Common Loop Algorithms In the following sections, we discuss some of the most common algorithms that are implemented as loops. You can use them as starting points for your loop designs. # 4.7.1 Sum and Average Value Computing the sum of a number of inputs is a very common task. Keep a running total, a variable to which you add each input value. Of course, the total should be initialized with 0. double total = 0; while (in.hasNextDouble()) {double input = in.nextDouble(); total = total + input; } Note that the total variable is declared outside the loop. We want the loop to update a single variable. The input variable is declared inside the loop. A separate variable is created for each input and removed at the end of each loop iteration. To compute an average, count how many values you have, and divide by the count. Be sure to check that the count is not zero. double total = 0; int count = 0; while (in.hasNextDouble()) {double input = in.nextDouble(); total = total + input; count++; }double average = 0; if (count > 0) {average = total / count; } # 4.7.2 Counting Matches You often want to know how many values fulfill a particular condition. For example, you may want to count how many spaces are in a string. Keep a counter , a variable that is initialized with 0 and incremented whenever there is a match.
{"source": 4220, "title": "from dpo"}
work, and we have $2$ subcases. There can not be more than $4$ blues, so we are done. Our total is $1+8+28+56+8+4+8+2=115$. There are $2^8=256$ possible colorings, so we have $\dfrac{115}{256}$ and our answer is $115+256=\boxed{371}$. ~Technodoggo Let $r$ be the number of red vertices and $b$ be the number of blue vertices, where $r+b=8$. By the Pigeonhole Principle, $r\geq{b} \Longrightarrow b\leq4$ if a configuration is valid. We claim that if $b\leq3$, then any configuration is valid. We attempt to prove by the following: If there are \[b\in{0,1,2}\] vertices, then intuitively any configuration is valid. For $b=3$, we do cases: If all the vertices in $b$ are non-adjacent, then simply rotating once in any direction suffices. If there are $2$ adjacent vertices, then WLOG let us create a set $\{b_1,b_2,r_1\cdots\}$ where the third $b_3$ is somewhere later in the set. If we assign the set as $\{1,2,3,4,5,6,7,8\}$ and $b_3\leq4$, then intuitively, rotating it $4$ will suffice. If $b_3=5$, then rotating it by 2 will suffice. Consider any other $b_3>5$ as simply a mirror to a configuration of the cases. Therefore, if $b\leq3$, then there are $\sum_{i=0}^{3}{\binom{8}{i}}=93$ ways. We do count the [i]degenerate[/i] case. Now if $b=4$, we do casework on the number of adjacent vertices. 0 adjacent: $\{b_1,r_1,b_2,r_2\cdots{r_4}\}$. There are 4 axes of symmetry so there are only $\frac{8}{4}=2$ rotations of this configuration. 1 adjacent: WLOG $\{b_1,b_2\cdots{b_3}\cdots{b_4}\}$ where $b_4\neq{8}$. Listing out the cases and trying, we get that $b_3=4$ and $b_4=7$ is the only configuration. There are $8$ ways to choose $b_1$ and $b_2$ and the rest is set, so there are $8$ ways. 2 adjacent: We can have WLOG $\{b_1,b_2\cdots{b_3},b_4\}$ or $\{b_1,b_2,b_3\cdots\}$ where $b_4\neq{8}$. The former yields the case $b_3=5$ and $b_4=6$ by simply rotating it 2 times. The latter yields none. There are 2 axes of symmetry so
{"source": 6134, "title": "from dpo"}
electronic spectroscopy; also known as UV-Vis. For simple compounds with high symmetry, the d–d transitions can be assigned using Tanabe–Sugano diagrams. These assignments are gaining increased support with computational chemistry. === Colors of lanthanide complexes === Superficially lanthanide complexes are similar to those of the transition metals in that some are colored. However, for the common Ln3+ ions (Ln = lanthanide) the colors are all pale, and hardly influenced by the nature of the ligand. The colors are due to 4f electron transitions. As the 4f orbitals in lanthanides are "buried" in the xenon core and shielded from the ligand by the 5s and 5p orbitals they are therefore not influenced by the ligands to any great extent leading to a much smaller crystal field splitting than in the transition metals. The absorption spectra of an Ln3+ ion approximates to that of the free ion where the electronic states are described by spin-orbit coupling. This contrasts to the transition metals where the ground state is split by the crystal field. Absorptions for Ln3+ are weak as electric dipole transitions are parity forbidden (Laporte forbidden) but can gain intensity due to the effect of a low-symmetry ligand field or mixing with higher electronic states (e.g. d orbitals). f-f absorption bands are extremely sharp which contrasts with those observed for transition metals which generally have broad bands. This can lead to extremely unusual effects, such as significant color changes under different forms of lighting. === Magnetism === Metal complexes that have unpaired electrons are paramagnetic. This can be due to an odd number of electrons overall, or to incomplete electron-pairing. Thus, monomeric Ti(III) species have one "d-electron" and must be (para)magnetic, regardless of the geometry or the nature of the ligands. Ti(II), with two d-electrons, forms some complexes that have two unpaired
{"page_id": 7304, "title": "Coordination complex"}
loss of moral sense, a tendency to impulsiveness or self-doubt, an unevenness of mental qualities such as unusual memory or aesthetic abilities, a tendency to mutism or to verbosity, excessive originality, preoccupation with the self, mystical interpretations placed on simple facts or perceptions, the abuse of symbolic meanings and the magical use of words, or mantras. Lombroso, with his concept of atavistic retrogression, suggested an evolutionary reversion, complementing hereditary degeneracy, and his work in the medical examination of criminals in Turin resulted in his theory of criminal anthropology—a constitutional notion of abnormal personality that was not actually supported by his own scientific investigations. In his later life, Lombroso developed an obsession with spiritualism, engaging with the spirit of his long dead mother. In 1892, Max Nordau, an expatriate Hungarian living in Paris, published his extraordinary bestseller Degeneration, which greatly extended the concepts of Bénédict Morel and Cesare Lombroso (to whom he dedicated the book) to the entire civilization of western Europe, and transformed the medical connotations of degeneration into a generalized cultural criticism. Adopting some of Charcot's neurological vocabulary, Nordau identified a number of weaknesses in contemporary Western culture which he characterized in terms of ego-mania, i.e., narcissism and hysteria. He also emphasized the importance of fatigue, enervation and ennui. Nordau, horrified by the anti-Semitism surrounding the Dreyfus affair, devoted his later years to Zionist politics. Degeneration theory fell from favour around the time of the First World War because of an improved understanding of the mechanisms of genetics as well as the increasing vogue for psychoanalytic thinking. However, some of its preoccupations lived on in the world of eugenics and social Darwinism. It is notable that the Nazi attack on western liberal society was largely couched in terms of degenerate art with its associations of racial miscegenation and fantasies
{"page_id": 1014250, "title": "Social degeneration"}
various NLP groups of universities and companies in China, including several world-known NLP scholars, such as Wang Haifeng of Baidu, Zhou Ming of Microsoft Research, Zhang Min (张民) of Soochow University (China), and Zhao Tiejun (赵铁军) and Liu Ting (刘挺) of HIT. Owing to his contributions in Chinese language processing, Li was elected as the President of Chinese Information Processing Society of China (CIPSC) in 2011. He scaled this top level academic organization in China up to more than 3000 registered members, and promoted NLP into several national projects for research or industry development. In addition, the CIPSC is now enhancing its co-operations with world NLP organizations including ACL. == Machine Intelligence & Translation Laboratory (MI&TLAB) == Originates from Machine Translation Research Group of Computer Science Department, Harbin Institute of Technology, which was started Li in 1985. It is one of the earliest institutions engaged in MT research in China, featured by its investigations into Chinese-English machine translation. It is now running under the Research Center on Language Technology, School of Computer Science and Technology, HIT. Details for staffs and publications can be found at http://mitlab.hit.edu.cn. == MOE-MS Joint Key Lab of Natural Language Processing and Speech (HIT) == In June, 2000, the Joint HIT-Microsoft Machine Translation Lab was founded by MI&T Lab and Microsoft Research (China). It was the third joint lab established by Microsoft Research (China) with Chinese universities, and the only one focusing on Machine Translation. Based on this jointly lab, the cooperation between HIT and Microsoft gradually extended to the areas of machine translation, information retrieval, speech recognition and processing, natural language understanding. In Oct, 2004, the joint key lab was granted as one of the 10 joint key labs supported by the Microsoft Research of Asia and Ministry of Education in China. In July 2006,
{"page_id": 46850543, "title": "Li Sheng (computer scientist)"}
HD 37605 b is an extrasolar planet that is 2.84 times more massive than Jupiter. It orbits close to the star, taking 54 days to revolve around the parent star HD 37605. Its orbit is highly eccentric, around 74%. Distance from HD 37605 ranges from 0.069 to 0.453 astronomical units. It is the first planet found by Hobby-Eberly Telescope (HET) in July 2004. In a simulation, HD 37605 b's orbit "sweeps clean" most test particles within 0.5 AU; leaving only asteroids "in low-eccentricity orbits near the known planet’s apastron distance, near the 1:2 mean-motion resonance" with oscillating eccentricity up to 0.06, and also at 1:3 with oscillating eccentricity up to 0.4. == References == == External links == "HD 37605". Exoplanets. Archived from the original on 2009-11-24. Retrieved 2008-10-29.
{"page_id": 16796835, "title": "HD 37605 b"}
from a liquid reservoir and appear out of an automatic door to serve them. Another automaton he created was used for hand washing to refill a basin with water after being drained. === Italy === In the 1400s, Leonardo da Vinci conceptualized a complex mechanical robot clad in a suit of armor, capable of sitting, standing, and independently moving its arms. The entire robot was operated by a system of pulleys and cables. === Japan === From the 17th to 19th centuries, the Japanese built humanoid automata called karakuri puppets. These puppets resembled dolls and were used for entertainment in theatre, homes, and religious festivals. Karakuri puppets that were used for theater plays were called butai karakuri. Small karakuri puppets found in homes, called zashiki kurakuri, were placed on tables to dance, beat drums, or serve drinks. The puppets used in religious festivals were known as Dashi karakuri, and they served to reenact myths and legends. === France === In the 18th century, French inventor Jacques de Vaucanson created a significant humanoid automaton called The Flute Player. This wooden, human-sized robot was capable of playing various melodies with the flute. It consisted of a system of bellows, pipes, weights, and other mechanical components to simulate to the muscles necessary to play the flute. === South Korea === KAIST's research team developed Hubo, South Korea's first humanoid robot, and Professor Oh Jun-ho, who led the team, founded Rainbow Robotics in 2011. In 2025, South Korean government formed the K-Humanoid Alliance, an alliance of the companies, researchers, and talents. == Applications == Humanoid robots are now used as research tools in several scientific areas. Researchers study the human body structure and behavior (biomechanics) to build humanoid robots. On the other side, the attempt to simulate the human body leads to a better
{"page_id": 48180, "title": "Humanoid robot"}
Trass is the local name of a volcanic tuff occurring in the Eifel, where it is worked for hydraulic mortar. It is a grey or cream-coloured fragmental rock, largely composed of pumiceous dust, and may be regarded as a trachytic tuff. It much resembles the Italian pozzolana and is applied to like purposes. Mixed with lime and sand, or with Portland cement, it is extensively employed for hydraulic work, especially in the Netherlands; while the compact varieties have been used as a building material and as a fire-stone in ovens. Trass was formerly worked extensively in the Brohl valley and is now obtained from the valley of the Nette, near Andernach. == See also == Pozzolan Pozzolana Pozzolanic reaction Pumice == References ==
{"page_id": 2964733, "title": "Trass"}
transverse vibrations (or shear waves) were originally described only in elastic solids exhibiting the highly ordered crystalline state of matter. In other words, simple liquids cannot support an applied force in the form of a shearing stress, and will yield mechanically via macroscopic plastic deformation (or viscous flow). Furthermore, the fact that a solid deforms locally while retaining its rigidity – while a liquid yields to macroscopic viscous flow in response to the application of an applied shearing force – is accepted by many as the mechanical distinction between the two. The inadequacies of this conclusion, however, were pointed out by Frenkel in his revision of the kinetic theory of solids and the theory of elasticity in liquids. This revision follows directly from the continuous characteristic of the viscoelastic crossover from the liquid state into the solid one when the transition is not accompanied by crystallization—ergo the supercooled viscous liquid. Thus we see the intimate correlation between transverse acoustic phonons (or shear waves) and the onset of rigidity upon vitrification, as described by Bartenev in his mechanical description of the vitrification process. The velocities of longitudinal acoustic phonons in condensed matter are directly responsible for the thermal conductivity that levels out temperature differentials between compressed and expanded volume elements. Kittel proposed that the behavior of glasses is interpreted in terms of an approximately constant "mean free path" for lattice phonons, and that the value of the mean free path is of the order of magnitude of the scale of disorder in the molecular structure of a liquid or solid. The thermal phonon mean free paths or relaxation lengths of a number of glass formers have been plotted versus the glass transition temperature, indicating a linear relationship between the two. This has suggested a new criterion for glass formation based on
{"page_id": 22122416, "title": "Glass transition"}
field strength and hence precession frequency varies with location and time. Larmor precession frequency = magnetogyric ratio x magnetic field Proton magnetogyric ratio = 42.576 Hz/μT (also written 42.576 MHz/T or 0.042576 Hz/nT) Earth's magnetic field: 30 μT near Equator to 60 μT near Poles, around 50 μT at mid-latitudes. Thus proton (hydrogen nucleus) EFNMR frequencies are audio frequencies of about 1.3 kHz near the Equator to 2.5 kHz near the Poles, around 2 kHz being typical of mid-latitudes. In terms of the electromagnetic spectrum EFNMR frequencies are in the VLF and ULF radio frequency bands, and the audio-magnetotelluric (AMT) frequencies of geophysics. Examples of molecules containing hydrogen nuclei useful in proton EFNMR are water, hydrocarbons such as natural gas and petroleum, and carbohydrates such as occur in plants and animals. == See also == Rate of change of Earth's magnetic field Zero field NMR == References == == External links == TeachSpin EFNMR web site Magritek EFNMR web site Two dimensional EFNMR imaging Earth's field NMR/MRI practical course, SS24 October 2009. Department of Physics, University of Oxford NMR Using Earth’s Magnetic Field Open source Earth's Field NMR Spectrometer Magnetic Resonance Imaging System Based on Earth’s Magnetic Field Applications of Earth’s Field NMR to porous systems and polymer gels
{"page_id": 11434033, "title": "Earth's field NMR"}
SkyRadar is a European research and development company for radar technology and aviation security. It manufactures radar hardware and simulators for training, education and research. It was founded in 2008 as a joint venture of several European participants. The team is conducting active research embedded in several international research programs and operates a not for profit knowledge portal, providing e-Learning and academic publications. == History == SkyRadar started in 2008 with the objective to make fully operational radar systems suitable for training and education purposes, responding to technical, budgetary and pedagogical requirements. It started with the assignment to disseminate research results gained in the frame of the Strategic Value Net project in cooperation with SAP Research and Karlsruhe Institute of Technology, funded in the frame of the 6th framework by the European Union. The founders experimented on close-range and low radiation radars. The first operational radars were produced in October 2008, including A-Scope and B-Scope functionality. PPI-scope functionality was added in 2009, as well as main filter functions like STC. To gain more flexibility scope of manoeuvre for pedagogical optimization, the team decoupled radars and scopes. In the solutions from 2013 onwards, low radiation radars send Q/I signals as message to browser enabled scopes, allowing students to operate scopes, filters or ampliers independently. Since 2014, a message-queue based server allows to connect high numbers of concurrent users. In the same year, SkyRadar went live with a knowledge portal, offering e-learning and providing access to recent academic publications on the subject. Since 2014, core functionalities of the technology behind SkyRadar's primary surveillance radar have been patented through the European Patent Office and the United States Patent and Trademark Office. == Research == The SkyRadar products, namely primary surveillance and secondary surveillance radars and simulators are a result of various research programs,
{"page_id": 55512492, "title": "SkyRadar"}
The SLJ900/32 or otherwise known as the Iron Monster is a superheavy launching gantry and one of the largest and most ubiquitous in the world. At over 90 meters in length, the SLJ900/32 is also one of the world's longest terrestrial vehicles. It is built by the Beijing Wowjoint Machinery Company and designed by the Shijiazhuang Railway Design Institute. == Description == The SLJ900/32 as aforementioned, is 91 metres (299 ft) long, 7 metres (23 ft) wide, 9 m (29 ft 6 in) tall and weighs around 580 tonnes (640 short tons). Compared to conventional cranes which require clear land and are typically only able to lift a few meters of bridge material per-lift, the SLJ900/32 is capable of lifting an entire large segment of bridging material (some of which weigh around 800 to 950 tons) in a relatively compact manner without the hassle of interference such as trees or rocks. Construction process is done via the vehicle picking up a new pre-cast section of concrete underneath its "belly" and carrying it all the way from the very edge of the bridge to the installation point, where it will be connected to a predetermined pillar. Then, using a pneumatic structure, the machine is moored to the first pillar to extend to the second one, anchoring to it, and places the beam. The process would repeat itself until the bridge's foundations are completed. The vehicle moves on a large 64x wheel drive system, which in itself is divided into four sections of 16 wheels each (forming two trucks, one at each end). Each of the sections can rotate up to 90 degrees allowing the machine to drive sideways for efficiency when picking up beams. The vehicle can maintain a top speed of 8 km/h (5 mph) unloaded, and 5 km/h (3
{"page_id": 57072034, "title": "SLJ900/32"}
of the resulting states (or of state-action pairs). But it is also possible to apply reinforcement learning to structured representations rather than atomic ones; this is called relational reinforcement learning (Tadepalli et al. , 2004). > RELATIONAL REINFORCEMENT LEARNING The survey by Kaelbling et al. (1996) provides a good entry point to the literature. The text by Sutton and Barto (1998), two of the field’s pioneers, focuses on architectures and algo-rithms, showing how reinforcement learning weaves together the ideas of learning, planning, and acting. The somewhat more technical work by Bertsekas and Tsitsiklis (1996) gives a rigorous grounding in the theory of dynamic programming and stochastic convergence. Re-inforcement learning papers are published frequently in Machine Learning , in the Journal of Machine Learning Research , and in the International Conferences on Machine Learning and the Neural Information Processing Systems meetings. 858 Chapter 21. Reinforcement Learning E XERCISES 21.1 Implement a passive learning agent in a simple environment, such as the 4 × 3 world. For the case of an initially unknown environment model, compare the learning performance of the direct utility estimation, TD, and ADP algorithms. Do the comparison for the optimal policy and for several random policies. For which do the utility estimates converge faster? What happens when the size of the environment is increased? (Try environments with and without obstacles.) 21.2 Chapter 17 defined a proper policy for an MDP as one that is guaranteed to reach a terminal state. Show that it is possible for a passive ADP agent to learn a transition model for which its policy π is improper even if π is proper for the true MDP; with such models, the P OLICY -E VALUATION step may fail if γ = 1 . Show that this problem cannot arise if P OLICY -E VALUATION
{"source": 1019, "title": "from dpo"}
while lemmatization can retrieve the correct base verb form, "be" . Similarly, lemmatization can normalize "worse" to "bad" , while stemming produces "wors" .Another major difference is that stemming reduces both occurrences of "meeting" to "meet" . Using lemmatization, the first occurrence of "meeting" is recognized as a Advanced Tokenization, Stemming, and Lemmatization | 345 noun and left as is, while the second occurrence is recognized as a verb and reduced to "meet" . In general, lemmatization is a much more involved process than stem‐ ming, but it usually produces better results than stemming when used for normaliz‐ ing tokens for machine learning. While scikit-learn implements neither form of normalization, CountVectorizer allows specifying your own tokenizer to convert each document into a list of tokens using the tokenizer parameter. We can use the lemmatization from spacy to create a callable that will take a string and produce a list of lemmas: In: # Technicality: we want to use the regexp-based tokenizer # that is used by CountVectorizer and only use the lemmatization # from spacy. To this end, we replace en_nlp.tokenizer (the spacy tokenizer) # with the regexp-based tokenization. import re # regexp used in CountVectorizer regexp = re.compile('(?u) \\ b\\ w\\ w+ \\ b') # load spacy language model and save old tokenizer en_nlp = spacy.load('en') old_tokenizer = en_nlp.tokenizer # replace the tokenizer with the preceding regexp en_nlp.tokenizer = lambda string: old_tokenizer.tokens_from_list( regexp.findall(string)) # create a custom tokenizer using the spacy document processing pipeline # (now using our own tokenizer) def custom_tokenizer(document): doc_spacy = en_nlp(document, entity=False, parse=False) return [token.lemma_ for token in doc_spacy] # define a count vectorizer with the custom tokenizer lemma_vect = CountVectorizer(tokenizer=custom_tokenizer, min_df=5) Let’s transform the data and inspect the vocabulary size: In: # transform text_train using CountVectorizer with lemmatization X_train_lemma = lemma_vect.fit_transform(text_train) print ("X_train_lemma.shape: {}".format(X_train_lemma.shape))
{"source": 2778, "title": "from dpo"}
construction, to derive $\tilde{O}(D)$-round distributed algorithms for MST and Min-Cut, in planar networks. This complexity nearly matches the trivial lower bound of $\Omega(D)$. We remark that this is the first result bypassing the well-known $\tilde{\Omega}(D+\sqrt{n})$ existential lower bound of general graphs (see Peleg and Rubinovich [FOCS'99]; Elkin [STOC'04]; and Das Sarma et al. [STOC'11]) in a family of graphs of interest. Timothy M. Chan)d}n$ deterministic time. The time bound was subsequently improved to $d^{(5+o(1))d}n$ by Br\"onnimann, Chazelle, and Matou\v sek [FOCS'93]. We obtain a further improved deterministic algorithm running in $d^{(1/2+o(1))d}n$ time. Along the way, we also point out a simplified version of Chazelle and Matou\v sek's algorithm, avoiding $\eps$-approximations. Ran Gelles, Bernhard Haeupler, Gillat Kol, Noga Ron-Zewi and Avi Wigderson. Towards optimal deterministic coding for interactive communication **Abstract:**We study \emph{efficient, deterministic} interactive coding schemes that simulate any interactive protocol both under adversarial and random errors, and can approach rate 1. For channels that flip bits with probability~$\epsilon<1/2$, our coding scheme achieves a communication rate of $1 -\Theta\left(\sqrt{H({\eps})}\right)$ and a failure probability of~$\exp(-n/ \log n)$ in length $n$ protocols. Prior to our work, all deterministic schemes (either efficient or not) had a rate bounded away from~$1$. Furthermore, the best failure probability achievable by an efficient deterministic protocol was only quasi-polynomial, i.e., of the form $\exp(- \log^{O(1)} n)$ (Braverman, ITCS 2012). For channels in which an adversary controls the noise pattern our coding scheme can tolerate $O(1/\log n)$ fraction of errors with rate approaching 1. No similar deterministic and efficient schemes were known with any constant rate. Essential to both results is an explicit, efficiently encodable and
{"source": 5203, "title": "from dpo"}
k + 1 primes, and that k + 1 is the least such number. Suppose also that a = p1 × · · · × pk × pk+1 = q1 × · · · × q` Note that ` > k + 1. Since pk+1 is prime we must have pk+1 | qj for some j; by swapping qj and q` if necessary, we may take j = `, so that pk+1 | q`. As before, q` = uk+1 pk+1 for some unit uk+1 . Dividing through by pk+1 gives p1 × · · · × pk = q1 × · · · × q`−1 × uk+1 140 CHAPTER 3. NUMBER THEORY Replacing q`−1 by q`−1uk+1 , which is still prime, we can apply the induction hypothesis to obtain k = ` − 1, so k + 1 = `, and, after reordering if necessary qi = uipi for all i 6 k. Since this also holds for i = k + 1, we’re done. By induction, we’re done. Example 3.2.13. Here are some examples of numbers written as products of primes: • 12 = 2 × 2 × 3. We could also write this as 2 × 3 × 2 or ( −2) × (−3) × 2, and so on. • 53 = 53 is an expression of 53 as a product of primes. • − 1000 = 2 × 5 × (−2) × 5 × 2 × 5. Exercise 3.2.14. Express the following numbers as products of primes: 16 −240 5050 111111 −123456789 To make things slightly more concise, we introduce a standard way of expressing a number as a product of primes: Definition 3.2.15. The canonical prime factorisation of a non-zero non-unit a ∈ Z is the expression in the form a = up j1 > 1
{"source": 6389, "title": "from dpo"}
the Solar System because it was the only one known until the mid-1990s. It was not confidently assumed to be widely applicable to other planetary systems, although scientists were anxious to test the nebular model by finding protoplanetary discs or even planets around other stars. As of August 30, 2013, the discovery of 941 extrasolar planets has turned up many surprises, and the nebular model must be revised to account for these discovered planetary systems, or new models considered. Among the extrasolar planets discovered to date are planets the size of Jupiter or larger, but that possess very short orbital periods of only a few hours. Such planets would have to orbit very closely to their stars, so closely that their atmospheres would be gradually stripped away by solar radiation. There is no consensus on how to explain these so-called hot Jupiters, but one leading idea is that of planetary migration, similar to the process which is thought to have moved Uranus and Neptune to their current, distant orbit. Possible processes that cause the migration include orbital friction while the protoplanetary disk is still full of hydrogen and helium gas and exchange of angular momentum between giant planets and the particles in the protoplanetary disc. One other problem is the detailed features of the planets. The solar nebula hypothesis predicts that all planets will form exactly in the ecliptic plane. Instead, the orbits of the classical planets have various small inclinations with respect to the ecliptic. Furthermore, for the gas giants, it is predicted that their rotations and moon systems will not be inclined with respect to the ecliptic plane. However, most gas giants have substantial axial tilts with respect to the ecliptic, with Uranus having a 98° tilt. The Moon being relatively large with respect to the Earth and
{"page_id": 17052696, "title": "History of Solar System formation and evolution hypotheses"}
Omega to stop, appealing to his desire to escape. Now calmer, Omega explains further to the two that he shaped this reality within the black hole both with his willpower and the power of its singularity. However, due to this, his will is the only support keeping this reality stable. He cannot freely leave without releasing control, but releasing control would collapse the antimatter universe instantly, annihilating everything in it; and so Omega's intention is for the Doctors to take his place maintaining it. As he prepares to leave with the Doctors' help, they are stunned to find that the extremely prolonged exposure to the singularity has destroyed Omega's physical body; his willpower now also maintains his essence, and he will cease to exist if he leaves. Suffering a nervous breakdown from the shock, Omega now seeks to destroy all creation. Taking advantage of his neurosis, the two Doctors escape back to the TARDIS with all of the abductees. With the First Doctor, the two devise a way to defeat Omega. By accident, the two also discover the Second Doctor's previously-lost recorder within the TARDIS' force field generator, and integrate it into their plan. The two meet with Omega again, claiming they can give him his freedom. Omega, though, sharply retorts that he cannot be freed, and demands that they share his exile. The Doctors agree, on the condition that all of his abductees are sent safely back to Earth. Once done, the two present him with the generator. Omega knocks it over in rage at the paltry offer and the recorder falls out; having fallen into the generator during the abduction, the recorder was protected from conversion, remaining as normal matter. The resulting contact annihilates the antimatter universe, creating a new universal source of energy, and ejecting the Doctors
{"page_id": 366726, "title": "The Three Doctors (Doctor Who)"}
The vehicle completed the first official autonomous run on 13 July 2018 within the framework of the event. === Season Alpha (2019) === Season Alpha took place at various locations in Europe and North America with the aim of testing several competition formats using the new DevBot 2.0. The first event was held at the Circuito Monteblanco in Spain, and featured the first race between two fully autonomous cars. The events were not broadcast live, instead short clips on YouTube were released. Two teams were competing: Arrival and Technical University of Munich. On 7 July 2019, the Roborace DevBot 2.0 car set the first ever autonomous official timed run at Goodwood Festival of Speed, with a time of 66.96 s and a top speed of 162.8 km/h (101.2 mph). This is currently the record for autonomous vehicles. Roborace also set the Guinness World Record for having the fastest autonomous car in the world. The Robocar reached a speed of 282.42 km/h (175.49 mph). === Season Beta (2020–21) === The second testing season took place at various locations between September 2020 and October 2021, featuring 16 races and involving mixed reality elements dubbed "Roborace Metaverse", which is based on Roborace's patented technology. The program of Season Beta competitions has gradually complicating rules arranged in a progression of so-called missions. Each mission consists of two racing rounds — one round per day. A mission plan issued by Roborace for each mission defines its objectives, rules, and point-scoring system. The key objective of Season Beta is to come to the point when the majority of competing teams have developed sufficient capability for wheel-to-wheel racing in Season 1. There were 7 teams competing in Season Beta: Arrival Racing (UK/Russia), Autonomous Racing Graz (Austria), MIT Driverless (United States), Acronis SIT (Switzerland), University of Pisa (Italy),
{"page_id": 48760514, "title": "Roborace"}
Huawei played key roles enabling their governments to spy on political opponents. Several IT sources told Le Monde that inside the African Union headquarters, whose computer systems were supplied by Huawei, data transfers on its servers peaked after hours from January 2012 to January 2017, with the AU's internal data sent to unknown servers hosted in Shanghai. In May 2019, a Huawei Mediapad M5 belonging to a Canadian IT engineer living in Taiwan was found to be sending data to servers in China despite never being authorized to do so, as the apps could not be disabled and continued to send sensitive data even after appearing to be deleted. At the end of 2019, United States officials disclosed to the United Kingdom and Germany that Huawei has had the ability to covertly exploit backdoors intended for law enforcement officials since 2009, as these backdoors are found on carrier equipment like antennas and routers, and Huawei's equipment is widely used around the world due to its low cost. The United Kingdom established a lab that it ran, but which was paid for by Huawei, to evaluate Huawei equipment.: 322 After eight years of study, the lab did not identify any Huawei backdoor, but concluded that Huawei's equipment had bugs that could be exploited by hackers.: 322 ==== Timeline ==== Yale University economist Stephen Roach stated in 2022 that there was no hard evidence to support the allegations of Huawei having a backdoor for industrial espionage other than one arguable instance,: 118 which was when UK telecom Vodafone disclosed in 2011 that its Italian fixed line network contained a security vulnerability in its Huawei-installed software.: 118–119 Huawei fixed the vulnerability at Vodafone's request.: 118 There was no report of any suspicious data capture or systems control activity. Vodafone was satisfied with the
{"page_id": 290909, "title": "Huawei"}
include the selection of a day on which to do something (including taking a trip, planting, a wedding, and the exhumation of ancestral corpses), whether a newborn child's destiny is compatible with its parents and thus whether it ought to be cared for by another family, the finding of a spouse, the finding of lost objects, the identification of a thief, and the explanation for a misfortune, including illness or sterility. Raymond Decary writes that the sikidy is consulted "in all circumstances", but especially: In cases of illness, which are understood to be either punishments or warnings from supernatural powers due to the transgression of a fady (taboo), or poisonings or curses (called tolaka) from other humans. Before undertaking a journey, in order to divine an auspicious day for travel. To acquire wealth or foresee the growth of herds (gold prospecting and panning must take place on a day selected by the sikidy). For all questions relating to women, including whether a potential bride has a fate aligned with her suitor's. In order to cast a bad spell on someone. To search for or track down thieves. The kind and color of sheep to be sacrificed in a wedding procession is also divined by sikidy. Among the forest-dwelling Mikea people, sikidy is used "to direct the timing of residential movements to the forest (mihemotse)". William Ellis describes two ritual occasions for sikidy relating to infants: the declaring of the child's destiny, and the "scrambling" ceremony. As one of the "first acts" following a child's birth, the child's father or close relative consults the local mpisikidy, who works the sikidy in order to read the child's destiny. When a child's destiny is declared to be favorable, "the child is nurtured with that tenderness and affection which nature inspires, and the warmest
{"page_id": 76866914, "title": "Sikidy"}
which no one else has been able to decipher with any certainty. They posit that these are all part of a "Usko-Mediterranean" branch of the speculative Dené–Caucasian language family, which they extend to include the Berber languages of North Africa. This thesis flatly contradicts basic Egyptological, Sumerian, Semitic, Indo-European, and Mesoamerican scholarship. Phoenician, Akkadian/Babylonian, Ugaritic, and Eblaite, for example, are transparently Semitic languages, and Arnaiz-Villena excludes the rest of the Semitic languages from his family; Egyptian and Berber along with Semitic have been demonstrated to be Afro-Asiatic, and generations of linguists have been unable to find a connection between Berber and Basque or Afro-Asiatic and Basque; and Hittite is widely acclaimed as a key in the reconstruction of Proto-Indo-European, which Arnaiz-Villena acknowledges is completely unrelated to Basque. De Hoz says their work "lacks the slightest value and is contrary not just to the scientific method but to common sense", and "is an unmitigated disaster which in principle should not be reviewed", but that he does so because it was published using public funds by the respected Editorial Complutense, which might give it undeserved credibility. He calls this a "crime" against legitimate research which has gone unpublished for lack of funds. Pichler likewise describes the "decipherment" of the Canary Island inscriptions as "comic", pointing out that Arnaiz-Villena "translated" an inscription of the alphabet as if it formed words (starting with "fire deceased earth prayer" in Basque), and also found it amazing that the university would publish his books. The "Basque" words he translated into are themselves dubious, including some that are modern neologisms and some that are loanwords from Romance languages, such as bake (from Latin pace "peace"), and which therefore can say nothing about ancient Basque connections. Lakarra, taking as a sample the list of 32 items entitled "Lenguaje religioso-funerario
{"page_id": 10841641, "title": "Antonio Arnaiz-Villena"}
prevent joint separation and by that means to reduce the range of stress the bolt experiences as the tension load is repeatedly applied. This determines the durability of the bolt when subjected to repeated tension loads. Maintaining a sufficient joint preload also prevents relative slippage of the joint components that would produce fretting wear that could result in a fatigue failure of those parts. The clamp load, also called preload of a fastener, is created when a torque is applied, and so develops a tensile preload that is generally a substantial percentage of the fastener's proof strength. Fasteners are manufactured to various standards that define, among other things, their strength. Torque charts are available to specify the required torque for a given fastener based on its property class (fineness of manufacture and fit) and grade (tensile strength). When a fastener is tightened, a tension preload is develops in the bolt, while an equal compressive preload forms in the clamped parts. This system can be modeled as a spring-like assembly, where the clamped parts experience compressive strain, and the bolt tensile strain. When an external tensile load is applied, it reduces the compressive strain in the clamped parts and increases the tensile strain in the bolt. The load carried by the bolt and the clamped parts is in proportion to their stiffness because they both experience the same induced strain. As a result, the external load is shared across the joint rather than being solely carried by the bolt. In a well-designed joint, about 10-20% of the applied tensile load is carried by the bolt with the greater part transferred through the clamped parts because they are much stiffer than it. This reduction in the proportion of load transferred to the bolt is important in applications with cyclic loading, as bolts
{"page_id": 922087, "title": "Bolted joint"}
Computer addiction is a form of behavioral addiction that can be described as the excessive or compulsive use of the computer, which persists despite serious negative consequences for personal, social, or occupational function. Another clear conceptualization is made by J. J. Block, who stated in a journal entry for the American Journal of Psychiatry that "Conceptually, the diagnosis is a compulsive-impulsive spectrum disorder that involves online and/or offline computer usage and consists of at least three subtypes: excessive gaming, sexual preoccupations, and e-mail/text messaging". Computer addiction is not currently included in the Diagnostic and Statistical Manual of Mental Disorders (DSM-5) as an official disorder. The concept of computer addiction is broadly divided into two types, namely offline computer addiction, and online computer addiction. Offline computer addiction is normally used when speaking about excessive gaming behavior, which can be practiced both offline and online. Online computer addiction, also known as Internet addiction, gets more attention in general from scientific research than offline computer addiction, mainly because most cases of computer addiction are related to the excessive use of the Internet. Experts on Internet addiction have described this syndrome as an individual working intensely on the Internet, prolonged use of the Internet, uncontrollable use of the Internet, unable to use the Internet in an efficient, timely matter, not being interested in the outside world, not spending time with people from the outside world, and an increase in their loneliness and dejection. == Symptoms == Being drawn by the computer as soon as one wakes up and before one goes to bed Replacing old hobbies with excessive use of the computer and using the computer as one's primary source of entertainment and procrastination Lacking physical exercise and/or outdoor exposure because of constant use of the computer, which could contribute to many health problems
{"page_id": 174397, "title": "Computer addiction"}
) ≜ diag ⁡ ( m 1 ( x ^ ) , m 2 ( x ^ ) , … , m n ( x ^ ) ) = [ m 1 ( x ^ ) m 2 ( x ^ ) ⋱ m i ( x ^ ) ⋱ m n ( x ^ ) ] {\displaystyle M({\hat {x}})\triangleq \operatorname {diag} (m_{1}({\hat {x}}),m_{2}({\hat {x}}),\ldots ,m_{n}({\hat {x}}))={\begin{bmatrix}m_{1}({\hat {x}})&&&&&\\&m_{2}({\hat {x}})&&&&\\&&\ddots &&&\\&&&m_{i}({\hat {x}})&&\\&&&&\ddots &\\&&&&&m_{n}({\hat {x}})\end{bmatrix}}} where, for each i ∈ { 1 , 2 , … , n } {\displaystyle i\in \{1,2,\dots ,n\}} , element m i ( x ^ ) > 0 {\displaystyle m_{i}({\hat {x}})>0} and suitably large to ensure reachability of the sliding mode. The observer vector V ( t ) {\displaystyle V(t)} is such that V ( t ) ≜ [ v 1 ( t ) v 2 ( t ) v 3 ( t ) ⋮ v i ( t ) ⋮ v n ( t ) ] ≜ [ y ( t ) { m 1 ( x ^ ) sgn ⁡ ( v 1 ( t ) − h 1 ( x ^ ( t ) ) ) } eq { m 2 ( x ^ ) sgn ⁡ ( v 2 ( t ) − h 2 ( x ^ ( t ) ) ) } eq ⋮ { m i − 1 ( x ^ ) sgn ⁡ ( v i − 1 ( t ) − h i − 1 ( x ^ ( t ) ) ) } eq ⋮ { m n − 1 ( x ^ ) sgn ⁡ ( v n − 1 ( t ) − h n − 1 ( x ^ ( t ) ) ) } eq ] {\displaystyle V(t)\triangleq {\begin{bmatrix}v_{1}(t)\\v_{2}(t)\\v_{3}(t)\\\vdots \\v_{i}(t)\\\vdots \\v_{n}(t)\end{bmatrix}}\triangleq {\begin{bmatrix}y(t)\\\{m_{1}({\hat {x}})\operatorname
{"page_id": 1261170, "title": "State observer"}
from Horn Pond in Woburn. The site where the canal met the Concord River had been the site of a grist mill since the 17th century, which the proprietors purchased along with all of its water rights. From this point, the canal descended six miles to the Merrimack River in East Chelmsford (now western Lowell) and 22 miles to the Charles River in Charlestown. In late September 1794 ground was broken in North Billerica. Work on the canal was performed by a number of contractors. In some instances, local workers were contracted to dig sections, while in other areas contract labor was brought in from Massachusetts and New Hampshire for the construction work. A variety of engineering challenges were overcome, leading to innovations in construction materials and equipment. A form of hydraulic cement (made in part from volcanic materials imported at great expense from Sint Eustatius in the West Indies) was used to make the stone locks watertight. Because of its cost and the cost of working in stone, a number of the locks were made of wood instead. An innovation was made in earth-moving equipment with the development of a precursor of the dump truck, where one side of the carrier was hinged to allow the rapid dumping of material at the desired location. Water was diverted into the canal in December 1800, and by 1803 the canal was filled to Charlestown. The first boat operated on part of the canal on April 22, 1802. === Merrimack canals === A variety of enterprises by all or a few of "the proprietors of the Middlesex Canal" which were the corporations' principal stockholders and the board came together with other third parties or acted in a few cases as a combined whole to fund the development of other stretches of the
{"page_id": 1300709, "title": "Middlesex Canal"}
target MUST be a byte string.( The representation may be a full, partial, or inconsistent version of the representation served from the URI of the resource.( A link with this link relation type can occur as a top-level element in a document or as a nested element within a link. When it occurs as a top-level element, it provides an alternate representation of the document's retrieval context. When it occurs nested within a link, it provides a representation of link target of the enclosing link.( Operation Types:( Indicates that the state of the form's context can be replaced with the state described by a representation submitted to the server.( This operation type defaults to the PUT method [[RFC7231]( for both HTTP and CoAP. Typical overrides by a form field include the PATCH method [[RFC5789]( for HTTP and CoAP and the iPATCH method [[RFC8132]( for CoAP.( Indicates that the form's context can be searched by submitting a search query.( This operation type defaults to the POST method [[RFC7231]( for HTTP and the FETCH method [[RFC8132]( for CoAP. Typical overrides by a form field include the POST method [[RFC7252]( for CoAP.( ### [A.2.]( Link Relation Types:( Indicates that the link's context is a collection and that the link's target is a member of that collection, as defined in [Section 2.1]( of [[RFC6573]( Indicates that the link's target is a collection and that the link's context is a member of that collection, as defined in [Section 2.2]( of [[RFC6573]( Operation Types:( Indicates that the form's context is a collection and that a new item can be created in that collection with the state defined by a representation submitted to the server.( This operation type defaults to the POST method [[RFC7231]( for both HTTP and CoAP.( Indicates that the form's context is a member
{"source": 1470, "title": "from dpo"}
the average ranks of the different instantiations along with the results of the Friedman and the Nemenyi tests for the analysis per dataset (left) and per target (right). In both analyses, the lowest average rank is obtained by ST-bag, followed by ST-sgb and ST-ridge. In the per dataset analysis, the Nemenyi test finds that ST-bag is significantly better than ST-tree and ST-svr while in the per target analysis, ST-bag is found significantly better than all the other instantiations. Therefore, we use bag as the base regressor for all problem transformation methods in the rest of the experiments. **Fig. 3** ![Image 10: figure 3) are connected. **a** Per dataset analysis. **b** Per target analysis Full size image and ERC\(_{true}\), the versions of SST and ERC that use the same type of values for the meta-inputs as their MLC counterparts, and compare their performance to that of ST, MORF, RLC, TNR and Dirty to see where these methods stand with respect to the state-of-the-art. Figure4 and per target (right).Footnote 14_and ERC_\(_{true}\)_are competitive with state-of-the-art methods_. SST\(_{train}\) obtains the lowest average rank in both the per dataset and the per target analysis. In the per dataset analysis, it is found significantly better than TNR and Dirty and similar with MORF and RLC and in the per target analysis it is found better than TNR, Dirty and MORF and similar
{"source": 3868, "title": "from dpo"}
Abstract: transparency is a key requirement for ethical machines. verified ethical behavior is not enough to establish justified trust in autonomous intelligent agents: it needs to be supported by the ability to explain decisions. logic programming (lp) has a great potential for developing such perspective ethical systems, as in fact logic rules are easily comprehensible by humans. furthermore, lp is able to model causality, which is crucial for ethical decision making. Expand Abstract 2020-09-14 Gedi: Generative Discriminator Guided Sequence Generation Ben Krause, Akhilesh Deepak Gotmare, Bryan Mccann, Nitish Shirish Keskar, Shafiq Joty, Richard Socher, Nazneen Fatema Rajani Abstract: while large-scale language models (lms) are able to imitate the distribution of natural language well enough to generate realistic text, it is difficult to control which regions of the distribution they generate. this is especially problematic because datasets used for training large lms usually contain significant toxicity, hate, bias, and negativity. we propose gedi as an efficient method for using smaller lms as generative discriminators to guide generation from large lms to make them safer and more controllable. gedi guides generation at each step by computing classification probabilities for all possible next tokens via bayes rule by normalizing over two class-conditional distributions; one conditioned on the desired attribute, or control code, and another conditioned on the undesired attribute, or anti control code. we find that gedi gives stronger controllability than the state of the art method while also achieving generation speeds more than 30 times faster. additionally, training gedi on only four topics allows us to controllably generate new topics zero-shot from just a keyword, unlocking a new capability that previous controllable generation methods do not have. lastly, we show that gedi can make gpt-2 (1.5b parameters) significantly less toxic without sacrificing linguistic quality, making it by far the most practical existing
{"source": 5788, "title": "from dpo"}
of the psyllium husks? Reply EMMA March 14, 2017 at 3:29 pm See the Q&A above, in short ‘No’! Reply PATRICIA March 28, 2017 at 5:46 pm Hi Ray, Without the psyllium husks your bread will not hold together. Reply LAURA October 30, 2017 at 5:59 am I know you commented months ago, Ray, but I used apple pectin instead of psyllium husk. Like psyllium, it is also a gel-forming fiber. I haven’t tried psyllium yet, and it may work better, but the apple pectin was at least passable. Reply Pingback: The Life Changing Protein Loaf of Bread | Grilo Protein Pingback: ZNAKOMITY CHLEB (NIE?) ZMIENIAJĄCY ŻYCIE/A – BEZGLUTENOWY – Smakoterapia Pingback: Cu ce să înlocuieşti pâinea? 5 alternative – gândul B February 16, 2017 at 11:06 am Hi there Sarah, I just wanted to ask how much extra water should be used if using ground or milled flax seeds instead of whole flax seeds? Your notes say to add a “lot” more water. How much more water would you suggest? Reply JUDI February 16, 2017 at 9:23 am Has anyone tried freezing the raw seed/psyllium mix/dough and baking it later? Reply Pingback: Tofu Spread with Herbs and Seeds - Delicious and Healthy by Maya JEANNE February 4, 2017 at 6:08 am Healthy, satisfying,easy, versatile, love this entire website. This bread recipe is one of the best. Thankyou Sarah for sharing and helping so many. Reply Pingback: How To Bake The Bread | Information Pingback: How To Bake Pre Made Garlic Bread | Information ZULEJKA February 1, 2017 at 11:40 am Love it. I made avocado+spinach+black salt toast with it and it was so so good! (I put a link to your recipe in this post, I hope you don’t mind: Love your work! Reply Pingback: ‘Life Changing Loaf’
{"source": 7370, "title": "from dpo"}
on average about 100 g, and an increased hemoglobin concentration of on average 1.5 g/dL with half the risk of being iron deficient at three and six months, but an increased risk of jaundice requiring phototherapy. In 2012, the American College of Obstetricians and Gynecologists officially endorsed delaying clamping of the umbilical cord for 30–60 seconds with the newborn held below the level of the placenta in all cases of preterm delivery based largely on evidence that it reduces the risk of intraventricular hemorrhage in these children by 50%. In the same committee statement, ACOG also recognize several other likely benefits for preterm infants, including "improved transitional circulation, better establishment of red blood cell volume, and decreased need for blood transfusion". In January 2017, a revised Committee Opinion extended the recommendation to term infants, citing data that term infants benefit from increased hemoglobin levels in the newborn period and improved iron stores in the first months of life, which may result in improved developmental outcomes. ACOG recognized a small increase in the incidence of jaundice in term infants with delayed cord clamping, and recommended policies be in place to monitor for and treat neonatal jaundice. ACOG also noted that delayed cord clamping is not associated with increased risk of postpartum hemorrhage. Several studies have shown benefits of delayed cord clamping: A meta-analysis showed that delaying clamping of the umbilical cord in full-term neonates for a minimum of two minutes following birth is beneficial to the newborn in giving improved hematocrit, iron status as measured by ferritin concentration and stored iron, as well as a reduction in the risk of anemia (relative risk, 0.53; 95% CI, 0.40–0.70). A decrease was also found in a study from 2008. Although there is higher hemoglobin level at 2 months, this effect did not persist
{"page_id": 233253, "title": "Umbilical cord"}
and r {\displaystyle r} : k 2 = 4 9 q E I , r 2 = x 3 , x z = d w d x , n 2 = 1 3 2 {\displaystyle k^{2}={4 \over 9}{q \over EI},\quad r^{2}=x^{3},\quad {\sqrt {x}}z={\mathrm {d} w \over \mathrm {d} x},\quad n^{2}={1 \over 3^{2}}} Then, the equation transforms to the Bessel equation r 2 d 2 z d r 2 + r d z d r + ( k 2 r 2 − n 2 ) z = 0 {\displaystyle r^{2}{\mathrm {d} ^{2}z \over \mathrm {d} r^{2}}+r{\mathrm {d} z \over \mathrm {d} r}+\left(k^{2}r^{2}-n^{2}\right)z=0} The solution of the transformed equation is z = A J 1 3 ( k r ) + B J − 1 3 ( k r ) {\displaystyle z=AJ_{\frac {1}{3}}(kr)+BJ_{-{\frac {1}{3}}}(kr)} Where J n {\displaystyle J_{n}} is the Bessel function of the first kind. Then, the solution of the original equation is: d w d x = x ( A J 1 3 ( k x 3 2 ) + B J − 1 3 , 1 ( k x 3 2 ) ) {\displaystyle {\mathrm {d} w \over \mathrm {d} x}={\sqrt {x}}\left(AJ_{\frac {1}{3}}\left(kx^{\frac {3}{2}}\right)+BJ_{-{\frac {1}{3}},1}\left(kx^{\frac {3}{2}}\right)\right)} Now, we will use the boundary conditions: No moment at x = 0 → d 2 w d x 2 ( x = 0 ) = 0 → A = 0 {\displaystyle x=0\rightarrow {\mathrm {d} ^{2}w \over \mathrm {d} x^{2}}(x=0)=0\rightarrow A=0} Fixed at x = l → w ( l ) = 0 → J − 1 3 ( k l 3 2 ) = 0 {\displaystyle x=l\rightarrow w(l)=0\rightarrow J_{-{\frac {1}{3}}}\left(kl^{\frac {3}{2}}\right)=0} From the second boundary condition, we get that the critical length in which a vertical column will buckle under its own weight is: l max = ( j 1 3
{"page_id": 50642899, "title": "Self-buckling"}
these limits, any form of mathematical reasoning could be mechanized. The Church-Turing thesis implied that a mechanical device, shuffling symbols as simple as 0 and 1, could imitate any conceivable process of mathematical deduction. The key insight was the Turing machine—a simple theoretical construct that captured the essence of abstract symbol manipulation. This invention would inspire a handful of scientists to begin discussing the possibility of thinking machines. === Computer science === Calculating machines were designed or built in antiquity and throughout history by many people, including Gottfried Leibniz, Joseph Marie Jacquard, Charles Babbage, Percy Ludgate, Leonardo Torres Quevedo, Vannevar Bush, and others. Ada Lovelace speculated that Babbage's machine was "a thinking or ... reasoning machine", but warned "It is desirable to guard against the possibility of exaggerated ideas that arise as to the powers" of the machine. The first modern computers were the massive machines of the Second World War (such as Konrad Zuse's Z3, Alan Turing's Heath Robinson and Colossus, Atanasoff and Berry's ABC and ENIAC at the University of Pennsylvania). ENIAC was based on the theoretical foundation laid by Alan Turing and developed by John von Neumann, and proved to be the most influential. == Birth of artificial intelligence (1941-56) == The earliest research into thinking machines was inspired by a confluence of ideas that became prevalent in the late 1930s, 1940s, and early 1950s. Recent research in neurology had shown that the brain was an electrical network of neurons that fired in all-or-nothing pulses. Norbert Wiener's cybernetics described control and stability in electrical networks. Claude Shannon's information theory described digital signals (i.e., all-or-nothing signals). Alan Turing's theory of computation showed that any form of computation could be described digitally. The close relationship between these ideas suggested that it might be possible to construct an "electronic brain".
{"page_id": 2894560, "title": "History of artificial intelligence"}
Declarative knowledge is an awareness of facts that can be expressed using declarative sentences. It is also called theoretical knowledge, descriptive knowledge, propositional knowledge, and knowledge-that. It is not restricted to one specific use or purpose and can be stored in books or on computers. Epistemology is the main discipline studying declarative knowledge. Among other things, it studies the essential components of declarative knowledge. According to a traditionally influential view, it has three elements: it is a belief that is true and justified. As a belief, it is a subjective commitment to the accuracy of the believed claim while truth is an objective aspect. To be justified, a belief has to be rational by being based on good reasons. This means that mere guesses do not amount to knowledge even if they are true. In contemporary epistemology, additional or alternative components have been suggested. One proposal is that no contradicting evidence is present. Other suggestions are that the belief was caused by a reliable cognitive process and that the belief is infallible. Types of declarative knowledge can be distinguished based on the source of knowledge, the type of claim that is known, and how certain the knowledge is. A central contrast is between a posteriori knowledge, which arises from experience, and a priori knowledge, which is grounded in pure rational reflection. Other classifications include domain-specific knowledge and general knowledge, knowledge of facts, concepts, and principles as well as explicit and implicit knowledge. Declarative knowledge is often contrasted with practical knowledge and knowledge by acquaintance. Practical knowledge consists of skills, like knowing how to ride a horse. It is a form of non-intellectual knowledge since it does not need to involve true beliefs. Knowledge by acquaintance is a familiarity with something based on first-hand experience, like knowing the taste of chocolate.
{"page_id": 23369987, "title": "Declarative knowledge"}
another moment of victory over Spider-Man when he teamed up with the Trapster. Sent by the Friends of Humanity to assassinate Paul Stacy, the duo managed to glue Stacy and Spider-Man together and corner them, only to be called off their assignment moments before finishing them off. Shocker would later turn on the Trapster at the behest of Osborn of having Trapster eliminated as a potential witness against his recent attempt to frame Spider-Man for murder. Preparing to kill the resigned Trapster in an alley, he was attacked and rendered unconscious at the last minute by Spider-Man (then disguised under the alias Dusk). Shocker has a run-in with Iron Man on a subway car in NYC. Shocker is hired by the Golem, along with Constrictor and Jack O'Lantern, to protect a diamond delivery. He battles Hood. Shocker was terminated as a freelance agent by Hammer Industries and captured by Spider-Man while trying to rob a federal bank alongside Hydro-Man. He accidentally evaporates his water-based ally with a misplaced vibration. He allied himself with Speed Demon of the New Thunderbolts to break into a particularly well-guarded safe. Though the police arrived at his hide-out in quick pursuit, he was saved by Speed Demon dashing in and stealing both the loot and Shocker's weapons, removing all traces of evidence (and, unfortunately for Shocker, also keeping the money to fund the Thunderbolts' further activities). Shocker plays a small role in the "Secret War" crossover event where he was among the villains used by Lucia von Bardas to attack New York. A new version of the Sinister Six, with Shocker as a member, bands together during the "Civil War" but were stopped by Captain America and his Secret Avengers. Shocker was among the group of villains that the Punisher attempts to kill at the
{"page_id": 331962, "title": "Shocker (character)"}
materials), as the reaction mechanism is strongly influenced by the material selection. Cathodes are mostly made up by stainless steel plates, Platinum mesh or carbon felt electrodes. Depending on the effluent nature, an increase of the conductivity of the solution may be required: the value of 1000 mS/cm is commonly taken as a threshold. Salts like sodium chloride or sodium sulfate can be added to the solution, acting as electrolytes, thus raising the conductivity. Typical values of salts concentration are in the range of few grams per liter, but the addition has a significant impact on power consumption and can reduce it by up to 30%. As the main cost associated to electro-oxidation process is the consumption of electricity, its performance are typically assessed through two main parameters, namely current efficiency and specific energy consumption. Current efficiency is generally defined as the charge required for the oxidation of the considered species over the total charged passed during electrolysis. Although some expressions have been proposed to evaluate the instantaneous current efficiency, they have several limitations due to the presence of volatile intermediates or the need for specialized equipment. Thus, it is much easier to define a general current efficiency (GCE), defined as an average of the value of current efficiency along the entire process and formulated as follows: G C E = F V ( C O D 0 − C O D t ) 8 I t {\displaystyle GCE=FV{\frac {(COD_{0}-COD_{t})}{8It}}} Where COD0 and CODt are the chemical oxygen demand (g/dm3) at time 0 and after the treatment time t, F is the Faraday's constant (96'485 C/mol), V is the electrolyte volume (dm3), I is the current (A), t is the treatment time (h) and 8 is the oxygen equivalent mass. Current efficiency is a time dependent parameter and it decreases
{"page_id": 61186827, "title": "Electro-oxidation"}
United States Green Building Council enable access to “environmentally and socially responsible, healthy, and prosperous environment[s] that improve[s] the quality of life.” A system by the name of LEED, is “the world’s most widely used green building system with more than 100,000 buildings participating” to date. Buildings that are funded by the Green Building Initiative and LEED have been proven to be financially, environmentally, and efficiently healthier for individuals. Lower carbon emissions, healthier living spaces, and improved efficiency are all the reap of the crop of the USGBC’s remedial efforts that are “constructed and operated through LEED.” == See also == Alternative natural materials Autonomous building Ecovillage Integrated Pest Management Permaculture Principles of Intelligent Urbanism == References == == External links == Creating sustainable communities in harmony with nature. Urban Permaculture. Path to Freedom - Urban Agriculture & Sustainability Helping create sustainable habitats around the world-the SHIRE
{"page_id": 4081192, "title": "Sustainable habitat"}
had the most balls. American coot mothers have the ability to recognize their chicks by imprinting on cues from the first chick that hatches. This allows mothers to distinguish their chicks from parasitic chicks. The peregrine falcon has also been known to imprint on specific structures for their breeding grounds such as cliff sides and bridges and thus will favour that location for breeding. == Sexual imprinting == Sexual imprinting is the process by which a young animal learns the characteristics of a desirable mate. For example, male zebra finches appear to prefer mates with the appearance of the female bird that rears them. Sexual attraction to humans can develop in non-human mammals or birds as a result of sexual imprinting when reared from young by humans. One example is London Zoo female giant panda Chi Chi. When taken to Moscow Zoo for mating with the male giant panda An An, she refused his attempts to mate with her, but made a full sexual self-presentation to a zookeeper. It commonly occurs in falconry birds reared from hatching by humans. Such birds are called "imprints" in falconry. When an imprint must be bred from, the breeder lets the male bird copulate with their head while they are wearing a special hat with pockets to catch the male bird's semen. The breeder then courts a suitable imprint female bird (including offering food, if it is part of that species's normal courtship). At "copulation", the breeder puts the flat of one hand on the female bird's back to represent the weight of a male bird, and with the other hand uses a pipette, or a hypodermic syringe without a needle, to squirt the semen into the female's cloaca. A bird known for trying to mate with humans is the kākāpō Sirocco. Sexual imprinting
{"page_id": 187721, "title": "Imprinting (psychology)"}
A monotube steam generator is a type of steam generator consisting of a single tube, usually in a multi-layer spiral, that forms a once-through steam generator (OTSG). The first of these was the Herreshoff steam generator of 1873. == Principles == For the sake of efficiency, it is desirable to minimise the steam content of the generator. Heat can then be transferred efficiently into liquid water, rather than into low-density steam. Monotube steam generators may either boil gradually along their length, usually pumped circulation systems, but where this boiling does not disrupt the circulation. Otherwise they can use the Benson supercritical system, where the pressure is sufficient to prevent boiling (within the heated volume) altogether. == Examples == Examples of Monotube steam generators include: Industrial steam generators The water-tube boilers of the monotube type used in steam cars, such as: AMC Clayton Steam Generator Doble steam car Gardner-Serpollet Locomobile Company of America White Motor Company, US patent 659,837 of 1900 == Flash boilers == A flash boiler is a particular type of low-water-content monotube boiler. Modern use is confined to model steam boats but, historically, flash boilers were used in Gardner-Serpollet steam cars. == See also == List of boiler types, by manufacturer Steam generator (boiler) Steam generator (railroad) == References ==
{"page_id": 29061881, "title": "Monotube steam generator"}
combination of the two. For information on selecting the processor-clock multiplier, see the BIOS and Kernel Developer’s Guide (BKDG) or Processor Programming Reference Manual applicable to your product. 14.1.3 Processor Initialization State Table 14 -1 shows the initial processor state following either RESET or INIT. Except as indicated, processor resources generally are set to the same value after either RESET or INIT. Table 14 -1. Initial Processor State > Processor Resource Value After RESET Value After INIT > CR0 0000_0000_6000_0010h CD and NW are unchanged Bit 4 (reserved) = 1 All others = 0 CR2, CR3, CR4 0CR8 00RFLAGS 0000_0000_0000_0002h EFER 0RIP 0000_0000_0000_FFF0h CS Selector = F000h Base = 0000_0000_FFFF_0000h Limit = FFFFh Attributes = See Table 14 -2 on page 482 DS, ES, FS, GS, SS Selector = 0000h Base = 0 Limit = FFFFh Attributes = See Table 14 -2 on page 482 GDTR, IDTR Base = 0 Limit = FFFFh LDTR, TR Selector = 0000h Base = 0 Limit = FFFFh Attributes = See Table 14 -2 on page 482 RAX 0(non-zero if BIST is run and fails) 0 [AMD Public Use] 481 Processor Initialization and Long Mode Activation AMD64 Technology 24593—Rev. 3.42—March 2024 Table 14 -2 on page 482 shows the initial state of the segment-register attributes (located in the hidden portion of the segment registers) following either RESET or INIT. RDX Family/Model/Stepping, including extended family and extended model—see “Processor Implementation Information” on page 483 RBX, RCX, RBP, RSP, RDI, RSI, R8, R9, R10, R11, R12, R13, R14, R15 0x87 Floating-Point State FPR0–FPR7 = 0 Control Word = 0040h Status Word = 0000h Tag Word = 5555h Instruction CS = 0000h Instruction Offset = 0 x87 Instruction Opcode = 0 Data-Operand DS = 0000h Data-Operand Offset = 0 Not modified 64-Bit Media State MMX0–MMX7 = 0
{"source": 76, "title": "from dpo"}