text
stringlengths 5
10.5k
| source
stringlengths 33
146
|
|---|---|
she went to the supermarket, the cashier asked her if she wanted to drive out or not. So I learned that phrase from her. Last Sunday, when I went to the supermarket, I was ready to hear that again and I was so excited about it. Because most of time, I was so nervous when the cashier asked me some questions and they all spoke quickly. But not this time, finally, after the cashier packed all my stuff into the plastic bag, he asked “Do you want to drive ?” “No, thanks.” I said. But I noticed he seemed to say some word instead of “out”. The last word sounded like “off” or “up” or I was wrong. But I checked it up in the dictionary, “drive out” has a different meaning. The carryover from the metalinguistic sensitization of the interaction logs into the classroom can be seen in the following example, observed by one of the authors of this book. (11-6) T = Teacher; S = Student S: He finally success. T: What? S: He finally succeed. T: Succeeds. S: Yes. Even though the student does not appreciate the full force of the teacher’s indirect question, he understands that she is making a correc-tion of form (rather than just indicating that she does not understand, which might yield merely a repetition of the early utterance), and he modifies his original utterance accordingly. Whether his yes indicates any-thing more than closure to the exchange is, of course, unclear. This example (as well as the examples from the interaction logs themselves) shows that metalinguistic training in focusing on form can result in sensitivity to grammatical form rather than just to lexical form, as occurs in most instances. Ohta (2001) noted that students in a classroom context can assimilate S E C
|
{"source": 982, "title": "from dpo"}
|
{base type definition} and {value} is not equal to the {value} of the parent length. 4.3.2 minLength [Definition:] minLength is the minimum number of units of length, where units of length varies depending on the type that is being ·derived· from. The value of minLength ·must· be a nonNegativeInteger. For string and datatypes ·derived· from string, minLength is measured in units of characters as defined in [XML 1.0 (Second Edition)]. For hexBinary and base64Binary and datatypes ·derived· from them, minLength is measured in octets (8 bits) of binary data. For datatypes ·derived· by ·list·, minLength is measured in number of list items. Note: For string and datatypes ·derived· from string, minLength will not always coincide with "string length" as perceived by some users or with the number of storage units in some digital representation. Therefore, care should be taken when specifying a value for minLength and in attempting to infer storage requirements from a given value for minLength. ·minLength· provides for: Constraining a ·value space· to values with at least a specific number of units of length, where units of length varies depending on {base type definition}. Example The following is the definition of a ·user-derived· datatype which requires strings to have at least one character (i.e., the empty string is not in the ·value space· of this datatype). 4.3.2.1 The minLength Schema Component Schema Component: minLength {value} A nonNegativeInteger. {fixed} A boolean. {annotation} Optional. An annotation. If {fixed} is true, then types for which the current type is the {base type definition} cannot specify a value for minLength other than {value}. 4.3.2.2 XML Representation of minLength Schema Component The XML representation for a minLength schema component is a element information item. The correspondences between the properties of the information item and properties of the component are as follows: XML
|
{"source": 2668, "title": "from dpo"}
|
initial detection of the suspicious activity. The BSA officer failed to file any SARs for the time period of June through August 20XX. This information was verified through use of the FinCEN database, which showed than no SARs had been filed during that time period. In addition, SARs filed from February through May of 20XX were filed between 65 days and 82 days of the initial detection of the activity. Management must ensure that suspicious activity reports are not only identified, but also filed in a timely manner. Part 353.3(f) of the FDIC Rules and Regulations Part 353.3(f) of the FDIC Rules and Regulations states that bank management must promptly notify its board of directors, or a committee thereof, of any report filed pursuant to Part 353 (Suspicious Activity Reports). Management has not properly informed the board of directors of SARs filed to report suspicious activities. The management team has provided the board with erroneous reports showing that the bank has filed SARs, when, in fact, the management team never did file such SARs. Board and committee minutes clearly indicate a reliance on these reports as accurate. 31 C.F.R. 103.22(c)(2) This section of the Financial Recordkeeping Regulations requires the bank to treat multiple transactions totaling over $10,000 as a single transaction. Management’s large cash aggregation reports include only those cash transactions above $9,000. Because of this weakness in the reporting system’s set-up, the report failed to pick up transactions below $9,000 from multiple accounts with one owner. The following transactions were identified which should have been aggregated and a CTR filed. Management needs to alter or improve their system in order to identify such transactions. Customer Name Date Amount Account # Mini Meat Market 122222222 12/12/xx $8,000 122233333 12/12/xx $4,000 122222222 12/16/xx $6,000 122233333 12/16/xx $5,000 Claire’s Club Sandwiches a/k/a
|
{"source": 4963, "title": "from dpo"}
|
noted earlier, when manipulation of the independent variable is possible but random assignment is not, quasi-experiments can be utilized. Quasi-experiments incorporate features in their design that attempt to eliminate specific classes of extraneous factors, thereby minimizing the number of plausible alternative explanations of the results (e.g., by examining treatments within particular demographic groups). More commonly hi the behavioral sciences, control takes the form of measuring these alternatives and partialing their effects. Difficulties arise because potential causes are often many; effects may be subtle, interactive, and delayed; and strong measurement is difficult. Consequently, the process of ruling out the influence of extraneous variables is often lengthy, involving the testing of multiple statistical models in multiple studies. Drawing on precedents in other sciences (e.g., geology, astronomy) in which randomization and manipulation are not possible, two tacks may be particularly useful in limiting the range of extraneous variables that can explain observed relationships of interest (Shadish, Cook, & Campbell, 2002). The first is to make precise predictions about the direction, form, and strength of relationships, an approach that is widely used in the natural sciences such as astronomy and geology, but to date has been rarely attempted in the behavioral sciences (Meehl, 1967). One example is provided by behavioral genetic theory, which makes precise predictions about heritability differences between fraternal and identical twins for traits determined by single genes. 5 An approach that reflects the complexity of relationships typically observed in the behavioral sciences is to propose a causal network of relationships among a set of variables (Cochran, 1965). To the degree that the entire hypothesized network of relationships is supported, it may be more difficult to think of parsimonious alternative explanations for the results. > 3We ignore here any issues of problems of external validity in matching experimental manipulations of variables to
|
{"source": 6256, "title": "from dpo"}
|
moves through space along a fluctuating orbit in a fashion that resembles a random polymer chain. The immediate conclusion to be drawn is that large groups of polymers may also be described by a single fluctuating field. As it turns out, the same can be said of a single polymer as well. In analogy to the original path integral expression presented, the end to end distribution of the polymer now takes the form: Φ ( R → , N ) = ∫ 0 , 0 R → , N e − A [ η ] P η ( N , l ) D η {\displaystyle \Phi ({\vec {R}},N)=\int _{0,0}^{{\vec {R}},N}e^{-{\mathcal {A}}[\eta ]}P^{\eta }(N,l){\mathcal {D}}\eta } Our new path integrand consists of: The fluctuating field η ( R → ) {\displaystyle \eta ({\vec {R}})} The action : A [ η ] = − 1 2 ∫ d R → d R → ′ η ( R → ) V − 1 ( R → , R → ′ ) η ( R → ′ ) {\displaystyle {\mathcal {A}}[\eta ]=-{\frac {1}{2}}\int d{\vec {R}}~d{\vec {R}}'~\eta ({\vec {R}})V^{-1}({\vec {R}},{\vec {R}}')\eta ({\vec {R}}')} with V ( R → , R → ′ ) {\displaystyle V({\vec {R}},{\vec {R}}')} denoting the monomer-monomer repulsive potential. P η ( N , L ) = ∫ exp { − ∫ 0 N d ν [ M 2 R → ˙ + η ( R → ( ν ) ) ] } D R → {\displaystyle P^{\eta }(N,L)=\int \exp \left\{-\int _{0}^{N}d\nu \left[{\frac {M}{2}}{\dot {\vec {R}}}+\eta ({\vec {R}}(\nu ))\right]\right\}{\mathcal {D}}{\vec {R}}} which satisfies the Schrödinger equation: [ ∂ ∂ N − 1 2 M ∇ 2 + η ( R → ) ] P η ( N , L ) = δ ( 3 ) ( R → − R →
|
{"page_id": 43265702, "title": "Path integrals in polymer science"}
|
involved are members of the same genus. There are now a substantial number of cases in which two or more species of the same cytodeme are clearly different and even in different genera. Thus all five diploid species recognised in the genus Brassicella and both species recognised in the genus Hutera (7spp. in total) are in the same cytodeme. Perhaps the most extreme case on record concerns the wild grass, teosinte, Euchlaena mexicana and the strikingly different Maize, or Indian Corn, Zea mays, both 2n=20, fully interfertile and yielding fertile hybrids. There is perhaps a remote possibility that two such very different species could have evolved independently from distinct sources and converged in their chromosomal ideotype until they became members of the same cytodeme (including the capacity to cross-breed). It is more likely, though, that the cytodeme arose first complete with its suite of chromosomes and breeding patterns all intact and then, remaining constant in its fundamentals, it diversified into species sometimes so different as to merit generic distinction. Thus, in what may be termed the cytodeme adjunct to Darwinian theory, evolution becomes a two-stage process - first, the establishment of distinct cytodemes reproductively isolated both from one another and from all previously existing cytodemes; second, diversification within cytodeme to yield taxonomically recognisable (but not reproductively isolated) species. Whereas the second stage is Natural Selection as expounded by Darwin, the first stage is not necessarily Darwinian. In the special case of polyploidy it is known that the first stage is not Darwinian. Doubling the chromosome number of the sterile hybrid between two diploid cytodemes yields an allotetraploid which is both fertile within its bounds and essentially incompatible with all previous life forms. As yet no mechanism is known which could account for the origin of new diploid cytodemes - nor
|
{"page_id": 42700066, "title": "Cytodeme"}
|
The molecular formula C4H9N3O2 (molar mass: 131.13 g/mol) may refer to: Creatine Guanidinopropionic acid
|
{"page_id": 23524665, "title": "C4H9N3O2"}
|
Andromeda III is a dwarf spheroidal galaxy about 2.44 million light-years away in the constellation Andromeda. It is part of the Local Group and is a satellite galaxy of the Andromeda Galaxy (M31). The galaxy was discovered by Sidney van den Bergh on photographic plates taken in 1970 and 1971. Observations of the dwarf galaxy using the WFPC2 in 2002 indicate that the bulk of the galaxy is around three billion years younger than the general population of globular clusters in our own galaxy. However, there are some older stars that are comparable in age to the Milky Way galactic clusters. There is no evidence for younger stars in this dwarf galaxy, suggesting no star formation is occurring. The dwarf galaxy is located at a distance of around 75 kpc from the center of M31. A total of 56 variable stars have been discovered in And III, including 51 RR Lyrae variables. == See also == List of Andromeda's satellite galaxies == References == == External links == SEDS: Dwarf Spheroidal Galaxy Andromeda III
|
{"page_id": 3170566, "title": "Andromeda III"}
|
or more processors and one or more IO hubs or routing hubs in a network on the motherboard, allowing all of the components to access other components via the network. As with HyperTransport, the QuickPath Architecture assumes that the processors will have integrated memory controllers, and enables a non-uniform memory access (NUMA) architecture. Each QPI comprises two 20-lane point-to-point data links, one in each direction (full duplex), with a separate clock pair in each direction, for a total of 42 signals. Each signal is a differential pair, so the total number of pins is 84. The 20 data lanes are divided onto four "quadrants" of 5 lanes each. The basic unit of transfer is the 80-bit flit, which has 8 bits for error detection, 8 bits for "link-layer header", and 64 bits for data. One 80-bit flit is transferred in two clock cycles (four 20-bit transfers, two per clock tick.) QPI bandwidths are advertised by computing the transfer of 64 bits (8 bytes) of data every two clock cycles in each direction. Although the initial implementations use single four-quadrant links, the QPI specification permits other implementations. Each quadrant can be used independently. On high-reliability servers, a QPI link can operate in a degraded mode. If one or more of the 20+1 signals fails, the interface will operate using 10+1 or even 5+1 remaining signals, even reassigning the clock to a data signal if the clock fails. The initial Nehalem implementation used a full four-quadrant interface to achieve 25.6 GB/s (6.4GT/s × 1 byte × 4), which provides exactly double the theoretical bandwidth of Intel's 1600 MHz FSB used in the X48 chipset. Although some high-end Core i7 processors expose QPI, other "mainstream" Nehalem desktop and mobile processors intended for single-socket boards (e.g. LGA 1156 Core i3, Core i5, and other
|
{"page_id": 1626918, "title": "Intel QuickPath Interconnect"}
|
Illusion of validity is a cognitive bias in which a person overestimates their ability to interpret and predict accurately the outcome when analyzing a set of data, in particular when the data analyzed show a very consistent pattern—that is, when the data "tell" a coherent story. This effect persists even when the person is aware of all the factors that limit the accuracy of their predictions, that is when the data and/or methods used to judge them lead to highly fallible predictions. Daniel Kahneman, Paul Slovic, and Amos Tversky explain the illusion as follows: "people often predict by selecting the output...that is most representative of the input....The confidence they have in their prediction depends primarily on the degree of representativeness...with little or no regard for the factors that limit predictive accuracy. Thus, people express great confidence in the prediction that a person is a librarian when given a description of his personality which matches the stereotype of librarians, even if the description is scanty, unreliable, or outdated. The unwarranted confidence which is produced by a good fit between the predicted outcome and the input information may be called the illusion of validity." Consistent patterns may be observed when input variables are highly redundant or correlated, which may increase subjective confidence. However, a number of highly correlated inputs should not increase confidence much more than only one of the inputs; instead higher confidence should be merited when a number of highly independent inputs show a consistent pattern. == Discovery and description == This bias was first described by Amos Tversky and Daniel Kahneman in their 1973 paper "On the Psychology of Prediction". In a 2011 article, Kahneman recounted the story of his discovery of the illusion of validity. After completing an undergraduate psychology degree and spending a year as an infantry
|
{"page_id": 34953080, "title": "Illusion of validity"}
|
In addition to providing the necessary documentation, AMD employees contribute code to support their hardware and features. All components of the Radeon graphics device driver are developed by core contributors and interested parties worldwide. In 2011, the r300g outperformed Catalyst in some cases. ==== AMDGPU ==== At the 2014 Game Developers Conference, AMD announced that they were exploring a strategy change to re-base the user-space part of Catalyst on a free and open-source DRM kernel module instead of their proprietary kernel blob. The release of the new AMDGPU kernel module and stack was announced on the dri-devel mailing list in April 2015. Although AMDGPU only officially supports GCN 1.2 and later graphics cards, experimental support for GCN 1.0 and 1.1 graphics cards (which are only officially supported by the Radeon driver) may be enabled via a kernel parameter. A separate libdrm, libdrm-amdgpu, has been included since libdrm 2.4.63. The radeonsi 3D code mentioned in the previous Radeon paragraph is also used with amdgpu; the 3D driver has back ends for both radeon and amdgpu. === Nvidia === Nvidia's proprietary driver, Nvidia GeForce driver for GeForce, is available for Windows x86/x86-64, Linux x86/x86-64/ARM, OS X 10.5 and later, Solaris x86/x86-64 and FreeBSD x86/x86-64. A current version can be downloaded from the Internet, and some Linux distributions contain it in their repositories. The 4 October 2013 beta Nvidia GeForce driver 331.13 supports the EGL interface, enabling support for Wayland in conjunction with this driver. Nvidia's free and open-source driver is named nv. It is limited (supporting only 2D acceleration), and Matthew Garrett, Dirk Hohndel and others have called its source code confusing. Nvidia decided to deprecate nv, not adding support for Fermi or later GPUs and DisplayPort, in March 2010. In December 2009, Nvidia announced they would not support free graphics initiatives.
|
{"page_id": 2226118, "title": "Free and open-source graphics device driver"}
|
In algebra, Posner's theorem states that given a prime polynomial identity algebra A with center Z, the ring A ⊗ Z Z ( 0 ) {\displaystyle A\otimes _{Z}Z_{(0)}} is a central simple algebra over Z ( 0 ) {\displaystyle Z_{(0)}} , the field of fractions of Z. It is named after Ed Posner. == Notes == == References == Artin, Michael (1999). "Noncommutative Rings" (PDF). Chapter V. Formanek, Edward (1991). The polynomial identities and invariants of n×n matrices. Regional Conference Series in Mathematics. Vol. 78. Providence, RI: American Mathematical Society. ISBN 0-8218-0730-7. Zbl 0714.16001. Edward C. Posner, Prime rings satisfying a polynomial identity, Proc. Amer. Math. Soc. 11 (1960), pp. 180–183. doi:10.2307/2032951
|
{"page_id": 43160575, "title": "Posner's theorem"}
|
is flat and the shaping filter has a finite bandwidth, it is possible to communicate with no ISI at all. Often the channel response is not known beforehand, and an adaptive equalizer is used to compensate the frequency response. == Effects on eye patterns == One way to study ISI in a PCM or data transmission system experimentally is to apply the received wave to the vertical deflection plates of an oscilloscope and to apply a sawtooth wave at the transmitted symbol rate R (R = 1/T) to the horizontal deflection plates. The resulting display is called an eye pattern because of its resemblance to the human eye for binary waves. The interior region of the eye pattern is called the eye opening. An eye pattern provides a great deal of information about the performance of the pertinent system. The width of the eye opening defines the time interval over which the received wave can be sampled without error from ISI. It is apparent that the preferred time for sampling is the instant of time at which the eye is open widest. The sensitivity of the system to timing error is determined by the rate of closure of the eye as the sampling time is varied. The height of the eye opening, at a specified sampling time, defines the margin over noise. An eye pattern, which overlays many samples of a signal, can give a graphical representation of the signal characteristics. The first image above is the eye pattern for a binary phase-shift keying (PSK) system in which a one is represented by an amplitude of −1 and a zero by an amplitude of +1. The current sampling time is at the center of the image and the previous and next sampling times are at the edges of the image.
|
{"page_id": 41287, "title": "Intersymbol interference"}
|
codes of practice, 12 months for general-purpose AI systems, 36 months for some obligations related to "high-risk" AI systems, and 24 months for everything else. == Reactions == Experts have argued that though the jurisdiction of the law is European, it could have far-ranging implications for international companies that plan to expand to Europe. Anu Bradford at Columbia has argued that the law provides significant momentum to the world-wide movement to regulate AI technologies. Amnesty International criticized the AI Act for not completely banning real-time facial recognition, which they said could damage "human rights, civil space and rule of law" in the European Union. It also criticized the absence of ban on exporting AI technologies that can harm human rights. Some tech watchdogs have argued that there were major loopholes in the law that would allow large tech monopolies to entrench their advantage in AI, or to lobby to weaken rules. Some startups welcomed the clarification the act provides, while others argued the additional regulation would make European startups uncompetitive compared to American and Chinese startups. La Quadrature du Net (LQDN) described the AI Act as "tailor-made for the tech industry, European police forces as well as other large bureaucracies eager to automate social control". LQDN described the role of self-regulation and exemptions in the act to render it "largely incapable of standing in the way of the social, political and environmental damage linked to the proliferation of AI". Building on these critiques, scholars have raised concerns in particular about the Act's approach to regulating the secondary uses of trained AI models, which may have significant societal impacts. They argue that the Act’s narrow focus on deployment contexts and reliance on providers to self-declare intended purposes creates opportunities for misinterpretation and insufficient oversight. Additionally, the Act often exempts open-source models
|
{"page_id": 68665578, "title": "Artificial Intelligence Act"}
|
Use of the term Big Data is usually associated with such technologies. Because a prime use of storing such data is generating insights through analytics, sometimes the term Big Data is expanded as Big Data analytics. But the term is becoming content free in that it can mean different things to different people. Because our goal is to introduce you to the large data sets and their potential in generating insights, we will use the original term in this chapter. Where does Big Data come from? A simple answer is “everywhere.” The sources that were ignored because of the technical limitations are now treated as gold mines. Big Data may come from Web logs, radio-frequency identification (RFID), global positioning systems (GPS), sensor networks, social networks, Internet-based text documents, Internet search indexes, detail call records, astronomy, atmospheric science, biology, genomics, nuclear physics, biochemical experiments, medical records, scientific research, military surveillance, photography archives, video archives, and large-scale e-commerce practices. Big Data is not new. What is new is that the definition and the structure of Big Data constantly change. Companies have been storing and analyzing large volumes of data since the advent of the data warehouses in the early 1990s. Whereas terabytes used to be synonymous with Big Data warehouses, now it’s exabytes, and the rate of growth in data volume continues to escalate as organizations seek to store and analyze greater levels of transaction details, as well as Web- and machine-generated data, to gain a better under-standing of customer behavior and business drivers. Many (academics and industry analysts/leaders alike) think that “Big Data” is a misnomer. What it says and what it means are not exactly the same. That is, Big Data is not just “big.” The sheer volume of the data is only one of many characteristics that are often
|
{"source": 1196, "title": "from dpo"}
|
each six-month period as fixed by the Secretary until the amount is paid. (b) The Government _may_ issue a demand for payment to the Contractor upon finding a debt is due under the contract. (c)_Final Decisions_. The _Contracting Officer_ will issue a final decision as required by 33.211 if- (1) The _Contracting Officer_ and the Contractor are unable to reach agreement on the existence or amount of a debt in a timely manner; (2) The Contractor fails to liquidate a debt previously demanded by the _Contracting Officer_ within the timeline specified in the demand for payment unless the amounts were not repaid because the Contractor has requested an installment payment agreement; or (3) The Contractor requests a deferment of collection on a debt previously demanded by the _Contracting Officer_ (see 32.607-2). (d) If a demand for payment was previously issued for the debt, the demand for payment included in the final decision _shall_ identify the same due date as the original demand for payment. (e) Amounts _shall_ be due at the earliest of the following dates: (1) The date fixed under this contract. (2) The date of the first written demand for payment, including any demand for payment resulting from a default termination. (f) The interest charge _shall_ be computed for the actual number of calendar days involved beginning on the due date and ending on- (1) The date on which the designated office receives payment from the Contractor; (2) The date of issuance of a Government check to the Contractor from which an amount otherwise payable has been withheld as a credit against the contract debt; or (3) The date on which an amount withheld and applied to the contract debt would otherwise have become payable to the Contractor. (g) The interest charge made under this clause
|
{"source": 3644, "title": "from dpo"}
|
w_m so that \operatorname {dist}(w_{i+1},\operatorname {Span}(w_1,\dots , w_i))> \delta ^{1/k}. By Lemma 10.1 (and a simple induction), we have \operatorname {vol}(w_1,\dots , w_{i+1})> \delta ^{i/k}. If this continues as far as i = k, then we have (1). 11 Nonconcentration on subspaces In this section, we establish a key technical result, Lemma 11.1 below, which says that orbits \{ \theta d n : n \leqslant X\} with \theta diophantine cannot concentrate too much near low-dimensional subspaces (or, more accurately, the image under \pi : \mathbf {R}^D \rightarrow \mathbf {T}^D of small balls in such subspaces). Here, we use \mathcal {N}_{\delta }(U) to denote the \delta -neighbourhood (in the Euclidean metric) of a set U \subset \mathbf {R}^D, and as usual B_{\delta }(0) denotes the Euclidean ball about 0 \in \mathbf {R}^D. Write \theta d [X] = \{ \theta d n : n \leqslant X\} for short. Lemma 11.1. Let r be sufficiently large, suppose that D = C_3 r^2, and assume that N \geqslant D^{C_2 D^2}. Let \varepsilon \in (\frac {1}{r},1). Let \theta \in \mathbf {T}^D be diophantine, let d \leqslant N/X, and let V \leqslant \mathbf {R}^D be a subspace of dimension at most D(1 - \varepsilon ). Then for all d \leqslant N/X, we have (11.1) \begin{align} \# \{ \theta d [X] \cap \pi(\mathcal{N}_{D^{-C_1}}(V) \cap B_{1/10}(0))\} \leqslant 2D^{( 1-\varepsilon C_1/2)D} X. \end{align} Proof. We may assume, by adding elements if necessary, that \dim V = \lfloor (1 - \varepsilon ) D\rfloor . Set m := \lceil \varepsilon D\rceil , \delta := D^{-C_1} and S := \mathcal {N}_{D^{-C_1}}(V) \cap B_{1/10}(0). Let (v_i)_{i = 1}^D be an orthonormal basis for \mathbf {R}^D with V = \operatorname {Span}_{\mathbf {R}}(v_{m+1},\dots , v_D). Let \psi : \mathbf {R} \rightarrow \mathbf {R} be a smooth cutoff satisfying the following conditions: 1. \psi \geqslant 0
|
{"source": 5659, "title": "from dpo"}
|
BY-SA 4.0", "answer_id": 33358, "answer_last_activity_date": "2023-07-13T20:59:41.290", "answer_last_edit_date": "2023-07-13T20:59:41.290", "answer_last_editor_user_id": 119, "answer_owner_user_id": 119, "answer_score": 7, "answer_view_count": 0, "forum": "quantumcomputing_stackexchange_com", "provenance": "20241028_173636_00007_mgema_03d4319f-fb4d-4471-84f4-9574879aedac.zst:134", "question_comment_count": 0, "question_content_license": "CC BY-SA 4.0", "question_id": 33357, "question_last_activity_date": "2023-07-13T20:59:41.290", "question_last_edit_date": "", "question_last_editor_user_id": 0, "question_owner_user_id": 19679, "question_score": 9, "question_view_count": 1087 } | stackexchange | Universal Gate Set, Magic States, and costliness of the T gate The usual universal gate set is $\mathcal{C} + T$ where $\mathcal{C}$ is the Clifford group and $T = \begin{pmatrix} 1 & 0 \\ 0 & e^{i\pi/4} \end{pmatrix} $ is the $\pi/8$ rotation gate. In practice we find a code that has $\mathcal{C}$ transversal and then we use magic state distillation to "simulate" the $T$ gate. However, we could have just as easily used the universal gate set $\mathcal{C} + M$ where $M$ is any matrix outside of the Clifford group. For example, the $T$ gate above is in the 3rd level of the Clifford hierarchy, but we could instead choose $M$ in the $4$th level (or even outside of the hierarchy). There must be some reason people mainly use $T$? I have heard it said that "implementing $T$ is extremely costly" but "even higher levels of the hierarchy are worse." Can someone maybe explain this to me? Note: On the other hand consider a 1-qubit universal gate set $\langle X, Z, T\rangle + H$ and find a code with transversal $X$, $Z$, and $T$ (like the $[[15,1,3]]$ Reed Muller code) whereby we simulate $H$ using magic states. Since $H$ is in the Clifford group wouldn't this offer up even more cost reduction? The T state $Z^{1/4}|+\rangle$ has four core advantages over most other states: 1. You can physically inject T states at pretty high fidelity. 2. It has a reasonably cheap distillation circuit, as far as the sizes of these circuits typically go. 3. Four
|
{"source": 6686, "title": "from dpo"}
|
y_{1},\dots ,y_{l}\in Y} and u {\displaystyle u} unlabeled examples x l + 1 , … , x l + u ∈ X {\displaystyle x_{l+1},\dots ,x_{l+u}\in X} are processed. Semi-supervised learning combines this information to surpass the classification performance that can be obtained either by discarding the unlabeled data and doing supervised learning or by discarding the labels and doing unsupervised learning. Semi-supervised learning may refer to either transductive learning or inductive learning. The goal of transductive learning is to infer the correct labels for the given unlabeled data x l + 1 , … , x l + u {\displaystyle x_{l+1},\dots ,x_{l+u}} only. The goal of inductive learning is to infer the correct mapping from X {\displaystyle X} to Y {\displaystyle Y} . It is unnecessary (and, according to Vapnik's principle, imprudent) to perform transductive learning by way of inferring a classification rule over the entire input space; however, in practice, algorithms formally designed for transduction or induction are often used interchangeably. == Assumptions == In order to make any use of unlabeled data, some relationship to the underlying distribution of data must exist. Semi-supervised learning algorithms make use of at least one of the following assumptions: === Continuity / smoothness assumption === Points that are close to each other are more likely to share a label. This is also generally assumed in supervised learning and yields a preference for geometrically simple decision boundaries. In the case of semi-supervised learning, the smoothness assumption additionally yields a preference for decision boundaries in low-density regions, so few points are close to each other but in different classes. === Cluster assumption === The data tend to form discrete clusters, and points in the same cluster are more likely to share a label (although data that shares a label may spread across multiple clusters).
|
{"page_id": 60968880, "title": "Weak supervision"}
|
A. clarkii is the most generalised species and utilises all ten anemone species, while nine — A. frenatus, A. chagosensis, A. pacificus, A. fuscocaudatus, A. latifasciatus, A. mccullochi, A. nigripes, A. sebae, and A. biaculeatus — use just one anemone species. Desirable traits in a host include long tentacles to hide among. In addition, certain anemones like H. aurora and E. quadricolor have tentacles with knob-like structures which provide more surface area for the fish to conceal itself. R. magnifica can provide extra protection when it pulls all its tentacles inside a soft body. The potency of venom is also a desirable trait; highly toxic anemone species tend to have smaller tentacles and so provide less shelter but more protection. Their ability to avoid being stung is attributed to their mucus coating. There is evidence that clownfish mucus mimics the molecules or bacteria of anemone mucus and lacks trigger for the anemone’s nematocysts (stinging barbs). Mucus thickness may also play a role, but this is not clear. There is dispute over how much of the mucus is innate to the clownfish and how much is gained from the anemone during the acclimation period. Nguyen and colleagues (2023) write "Whereas some anemone fish species seem to produce their own protective mucous coating, others may acquire mucus (or biomolecules within) from the sea anemone during an acclimation period". Roux and colleagues (2019) found evidence that clownfish exchange microbiota with their anemone hosts. === Social structure === A group of clownfish occupying an anemone usually consists of a breeding female and male along with some non-breeding individuals. Dominance in clownfish groups is based on size, with the female being the largest and most dominant, followed by the male and then the largest non-breeder and so on. An individual is 20 percent larger than
|
{"page_id": 176332, "title": "Clownfish"}
|
to bugs or slow performance. === Bugs === Improper management of memory is a common cause of bugs and security vulnerabilities, including the following types: A memory leak occurs when a program requests memory from the operating system and never returns the memory when it is done with it. A program with this bug will gradually require more and more memory until the program fails as the operating system runs out. A segmentation fault results when a program tries to access memory that it does not have permission to access. Generally, a program doing so will be terminated by the operating system. A buffer overflow occurs when a program writes data to the end of its allocated space and then continues to write data beyond this to memory that has been allocated for other purposes. This may result in erratic program behavior, including memory access errors, incorrect results, a crash, or a breach of system security. They are thus the basis of many software vulnerabilities and can be maliciously exploited. === Virtual memory === Virtual memory is a system where physical memory is managed by the operating system typically with assistance from a memory management unit, which is part of many modern CPUs. It allows multiple types of memory to be used. For example, some data can be stored in RAM while other data is stored on a hard drive (e.g. in a swapfile), functioning as an extension of the cache hierarchy. This offers several advantages. Computer programmers no longer need to worry about where their data is physically stored or whether the user's computer will have enough memory. The operating system will place actively used data in RAM, which is much faster than hard disks. When the amount of RAM is not sufficient to run all the current programs,
|
{"page_id": 6806, "title": "Computer memory"}
|
as its co-production company. The project was described as an "audience-driven story experience", with anticipated launch in 2016. In March 2016, Sam Barlow announced he had joined Interlude and would be serving as a creative lead in the series, on the basis of his work from his video game, "Her Story", which required the player to piece together a mystery based on a series of video clips. Interlude rebranded itself as Eko in December 2016, and the six-episode series was released in March 2018. == Soundtrack == The film's music was composed and conducted by Arthur B. Rubinstein and performed by the Hollywood Studio Symphony. A soundtrack album including songs and dialogue excerpts was released by Polydor Records. Intrada Records issued an expanded release in 2008 with the complete score, with expanded horn sections and without the film dialogue. In 2018, Quartet Records issued a 35th anniversary expanded 2-CD edition containing the score as presented in the film, and the 1983 Polydor album on disc 2. == Legacy == Critics have cited the film as an influence on Mamoru Hosoda's 2000 short film Digimon Adventure: Our War Game!, with critic Geoffrey G. Thew, writing in Anime Impact: The Movies and Shows that Changed the World of Japanese Animation, noting that both films share a title and a plot of "a rogue AI hijacking the Internet to spread chaos and potentially destroy the world, only to be stopped by some kids on their computers." Hosoda later stated that Our War Game "kind of started my idea for [his 2009 film] Summer Wars," noting that Summer Wars "became the feature-length version of that idea" and allowed him to explore material he was unable to in Our War Game's 40 minute runtime. == See also == 1983 Soviet nuclear false alarm incident, which
|
{"page_id": 34130, "title": "WarGames"}
|
with his research on annihilation radiation, made Blackett a leading expert in the new theory of antimatter. That same year, he moved to Birkbeck, University of London, as professor of Physics, and stayed for four years. In 1937, he went to the Victoria University of Manchester where he was elected to the Langworthy Professorship and created a major international research laboratory. The Blackett Memorial Hall and Blackett Lecture Theatre at the University of Manchester were subsequently named after him. He was elected to membership of the Manchester Literary and Philosophical Society on 14.12.1937. In 1947, Blackett introduced a theory to account for the Earth's magnetic field as a function of its rotation, with the hope that it would unify both the electromagnetic force and the force of gravity. He spent a number of years developing high-quality magnetometers to test his theory, but eventually found it to be without merit. However, his work on the subject led him into geophysics, where he later helped process data relating to paleomagnetism, and also provided strong evidence for continental drift. He was awarded the 1948 Nobel Prize in Physics for his investigation of cosmic rays using his invention of the counter-controlled cloud chamber. In 1953, he was appointed head of the Physics Department at Imperial College London, and retired from there in July 1963. The Physics department building of Imperial College, the Blackett Laboratory, is named in his honour. In 1957, Blackett gave the presidential address ("Technology and World Advancement") to the British Association meeting in Dublin In 1965, he delivered the MacMillan Memorial Lecture to the Institution of Engineers and Shipbuilders in Scotland. He chose the subject "Continental Drift". === World War II and operational research === In 1935, Blackett was invited to join the Aeronautical Research Committee chaired by Sir Henry Tizard.
|
{"page_id": 326834, "title": "Patrick Blackett"}
|
The Terrestrial Planet Finder (TPF) was a proposed project by NASA to construct a system of space telescopes for detecting extrasolar terrestrial planets. TPF was postponed several times and finally cancelled in 2011. There were two telescope systems under consideration, the TPF-I, which had several small telescopes, and TPF-C, which used one large telescope. == History == In May 2002, NASA chose two TPF mission architecture concepts for further study and technology development. Each would use a different means to achieve the same goal—to block the light from a parent star in order to see its much smaller, dimmer planets. The technological challenge of imaging planets near their much brighter star has been likened to finding a firefly near the beam of a distant searchlight. Additional goals of the mission would include the characterization of the surfaces and atmospheres of newfound planets, and looking for the chemical signatures of life. The two planned architectures were: Infrared astronomical interferometer (TPF-I): Multiple small telescopes on a fixed structure or on separated spacecraft floating in precision formation would simulate a much larger, very powerful telescope. The interferometer would use a technique called nulling to reduce the starlight by a factor of one million, thus enabling the detection of the very dim infrared emission from the planets. Visible Light Coronagraph (TPF-C): A large optical telescope, with a mirror three to four times bigger and at least 100 times more precise than the Hubble Space Telescope, would collect starlight and the very dim reflected light from the planets. The telescope would have special optics to reduce the starlight by a factor of one billion, thus enabling astronomers to detect faint planets. NASA and Jet Propulsion Laboratory (JPL) were to issue calls for proposals seeking input on the development and demonstration of technologies to implement the
|
{"page_id": 101232, "title": "Terrestrial Planet Finder"}
|
Jeffrey C. Wynn is a research geophysicist with the United States Geological Survey (USGS). He is currently based in the Cascades Volcano Observatory in Vancouver, WA, one of the five USGS volcano observatories in the United States [1]. == Professional career == Wynn has served as vice president for R&D of Zonge International, and in several rotational management positions in the USGS. These include Chief Scientist for Volcano Hazards, Chief of the Office of Geochemistry & Geophysics, Chief of the Venezuelan Guayana & Amazonas Exploration Mission ("Jefe del Grupo Asesor"), where he was first author of the first complete geologic map of southern Venezuela, and also published a full assessment of discovered and undiscovered mineral resources for the roadless southern half of Venezuela. Wynn also served for four years as the Deputy Chief for Science and Chief of the USGS Saudi Arabian Mission before assuming responsibility for volcano research and monitoring as Chief Scientist for Volcano Hazards in the USGS. Wynn has studied and published on the historical era Wabar craters asteroid impact event in the Empty Quarter of Saudi Arabia. He has also developed a new technology for mapping sub-seafloor minerals, buried wrecks, buried oil and gas infrastructure, and migrating hydrocarbon plumes in the open ocean using a physical property called induced polarization ("IP"). He has also done extensive geophysical mapping in SE Alaska. Wynn co-developed an airborne electromagnetic technology to rapidly map groundwater deep beneath arid basins in 3D. Using this technology, he successfully mapped the groundwater of the San Pedro Basin in southern Arizona and northern (Sonora), Mexico in three dimensions. == Publications == He has published over 300 articles, books, patents, and maps in fields as diverse as geology, oceanography, hydrology, geophysics, archeology, and astrophysics. Wynn is a past president of the Environmental & Engineering Geophysical
|
{"page_id": 667613, "title": "Jeffrey C. Wynn"}
|
Natural resource valuation is a process of providing of benefits, costs, damage of or to natural and environmental resources. It has a fundamental role in the practice of cost-benefit analysis of health, safety, and environmental issues. Natural resource valuation is performed in the conduct of natural resource damage assessments (NRDA) done under the Comprehensive Environmental Response, Compensation, and Liability Act of 1980 (CERCLA, or Superfund), the Oil Pollution Act (OPA), and state regulations. It is also performed in cost-benefit analysis of environmental restoration (ER) and waste management. It is a key exercise in economic analysis and its results provide important information about values of environmental goods and services. Natural resource valuation studies are often aimed at assessing economic values that represent the public good characteristics of natural systems. Willingness to pay measures are typically used to estimate ecosystem goods and services that benefit not only a select few but wider society. There are a two types of valuation including market valuation and non market valuation. Market valuation estimates the total willingness to pay based on price (demand) whilst non market valuation estimates willingness to pay either through examining behavior of respondents or demand for related goods. Most of the environmental resources are valued using the constructive approach. There are number of methods involved in natural resource valuation including revealed preference and stated preference method. It is important to value natural resources because they contribute towards fiscal revenue, income, and poverty reduction. Sectors related to natural resources use provide jobs and are often the basis of livelihoods in poorer communities. Owing to this fundamental importance of natural resources, they must be managed sustainably. == Importance == Resource valuation is used as an input to generate better policy recommendations to support protected areas management and its linkages to spatial planning process. Resource
|
{"page_id": 71484847, "title": "Natural resource valuation"}
|
MIR22HG (MIR22 host gene), also known as C17orf91, MGC14376, MIRN22, hsa-mir-22, and miR-22 is a human gene that encodes a noncoding RNA (ncRNA).This RNA molecule is not translated into a protein but nonetheless may have important functions. MIR22HG ncRNA is greater than 200 nucleotides in length and therefore informally classified as a long non-coding RNA, i.e. lncRNA. The MIR22HG gene is located at band 13.3 on the short (or "p") arm of chromosome 17. It is expressed in each of the 27 human tissues tested. == Function == Many lncRNAs regulate diverse processes including cellular metabolism, proliferation, movement, differentiation (i.e. change of a cell from one type to another, usually more mature, cell type), apoptosis (i.e. programmed cell death), and the expression of various genes through chromatin remodeling, genomic imprinting, modulating the actions of other RNAs, and various other ways. The normal actions and functions of the MIR22HG gene are complex and have not been fully elucidated, but its primary function may be as a tumor suppressor gene. It is involved in the regulation of several signaling pathways including Wnt/β-catenin, epithelial-mesenchymal transition (EMT), notch, and STAT3. == Clinical significance == When overexpressed, it acts as a tumor suppressor gene in many cancer types but in a few cancer types it acts as an oncogene, i.e. a tumor promotor gene: MIR22HG gene's impact on various cancers is strictly dependent on the type of cancer in which it operates. The roles or the MIR22HG gene in cancers have generally been evaluated by: a) next-generation sequencing to quantify the levels of MIR22HG RNA in cancer samples, samples of nearby normal tissue of the same type as the cancer, and cultured cancer cells of the same type as the cancers; b) comparing MIR22HG lncRNA levels in patients with the same cancer to the severity
|
{"page_id": 68639621, "title": "MIR22HG"}
|
In mathematics, a flow formalizes the idea of the motion of particles in a fluid. Flows are ubiquitous in science, including engineering and physics. The notion of flow is basic to the study of ordinary differential equations. Informally, a flow may be viewed as a continuous motion of points over time. More formally, a flow is a group action of the real numbers on a set. The idea of a vector flow, that is, the flow determined by a vector field, occurs in the areas of differential topology, Riemannian geometry and Lie groups. Specific examples of vector flows include the geodesic flow, the Hamiltonian flow, the Ricci flow, the mean curvature flow, and Anosov flows. Flows may also be defined for systems of random variables and stochastic processes, and occur in the study of ergodic dynamical systems. The most celebrated of these is perhaps the Bernoulli flow. == Formal definition == A flow on a set X is a group action of the additive group of real numbers on X. More explicitly, a flow is a mapping φ : X × R → X {\displaystyle \varphi :X\times \mathbb {R} \to X} such that, for all x ∈ X and all real numbers s and t, φ ( x , 0 ) = x ; φ ( φ ( x , t ) , s ) = φ ( x , s + t ) . {\displaystyle {\begin{aligned}&\varphi (x,0)=x;\\&\varphi (\varphi (x,t),s)=\varphi (x,s+t).\end{aligned}}} It is customary to write φt(x) instead of φ(x, t), so that the equations above can be expressed as φ 0 = Id {\displaystyle \varphi ^{0}={\text{Id}}} (the identity function) and φ s ∘ φ t = φ s + t {\displaystyle \varphi ^{s}\circ \varphi ^{t}=\varphi ^{s+t}} (group law). Then, for all t ∈ R , {\displaystyle t\in \mathbb
|
{"page_id": 2619735, "title": "Flow (mathematics)"}
|
it turns out, this was the wrong problem to solve because you don't always need to generate lots of options. (“If your software was any good, it’d produce fewer designs, not more.”) So while Autodesk was messing around with ways to generate and sort hundreds of design options, the other companies were inventing these powerful, general-purpose models (like GPT). And ultimately these versatile algorithms that could be applied to any project proved more productive than developing an interface to create a bespoke algorithm for every project. At the moment, generative AI is at a really interesting stage. While the underlying technology continues to improve, much of the innovation from companies like Google and Microsoft centers around the interface. It's a critical detail (“the real challenge isn’t the technology, it’s the interface, it’s how the algorithms fit the designer and their process.") Even for something like ChatGPT, the underlying technology has been around for a while. When I wrote the article, I used GPT2 for the email tool, which is a slightly earlier version of the language model that ChatGPT uses. So the technology was out there. People were building things with it. But it wasn’t until OpenAI hit upon the chat interface that it really took off. I’m still skeptical about whether we can create a GPT for architecture. The main problem is the data. Tools like Dall-E or GPT require huge data sets to train. But in architecture, there isn’t a large corpus of models to train an algorithm on. Autodesk certainly has the potential to gather this data (and in some alternate reality, they spent time getting this data rather than going down the rabbit hole of creating interfaces to sort design options). But even if a GPT for architecture is never created, these other tools will still
|
{"source": 1757, "title": "from dpo"}
|
to use exceptions for input validation and parsing, which can be thousands of times slower than the equivalent functions written with \/ or Validation. Some people claim that predictable exceptions for input validation are referentially transparent because they will occur every time. However, the stacktrace inside the exception depends on the call chain, giving a different value depending on who calls it, thus breaking referential transparency. Regardless, throwing an exception is not pure because it means the function is not Total. 6.7.4 These We encountered These, a data encoding of inclusive logical OR, when we learnt about Align. sealed abstract class \&/[+A, +B] { ... } object \&/ { type These[A, B] = A \&/ B final case class ThisA extends (A \&/ Nothing) final case class ThatB extends (Nothing \&/ B) final case class BothA, B extends (A \&/ B) def applyA, B: These[A, B] = Both(a, b) } with convenient construction syntax implicit class TheseOpsA { final def wrapThis[B]: A \&/ B = \&/.This(self) final def wrapThat[B]: B \&/ A = \&/.That(self) } implicit class ThesePairOpsA, B) { final def both: A \&/ B = \&/.Both(self._1, self._2) } These has typeclass instances for Monad Bitraverse Traverse Cobind and depending on contents Semigroup / Monoid / Band Equal / Order Show These (\&/) has many of the methods we have come to expect of Disjunction (\/) and Validation (\?/) sealed abstract class \&/[+A, +B] { def foldX => X): X = ... def swap: (B \&/ A) = ... def appendX >: A: Semigroup, Y >: B: Semigroup): X \&/ Y = ... def &&&X >: A: Semigroup, C: X \&/ (B,
|
{"source": 4170, "title": "from dpo"}
|
shooting with both eyes open. A few years ago i had an oh shit bear hunting situation where I made the shot with my front lens cover (accidently) on. Reply reply } Share Share [More replies]( []( [11b68w]( •[5y ago]( Either/or. An optic 1.93 or higher makes passive aiming pretty easy. Its doable with lower third, but less easy. Also helps to have something like a T2 or EoTech mounted all the way forward on the receiver. Dual illuminated RMR on the pistol works super easy for me. Easier than I expected, actually. If you have no concern about an enemy with NVG’s, a rifle laser is super nice, due to the illuminator helping you see better, in general. Especially in low ambient illum conditions or with mediocre tubes. The lasers are also useful as designators in a LE/Mil context. DBAL D2 and MAWL are what the cool kids are recommending for civvie lasers. Reply reply } Share Share []( [IHTFP08]( •[5y ago]( Yes. Higher mount helps, NV settings are a must. Also rmr on pistol is standard NV load out more than IR lasers. Reply reply } Share Share []( [dirttrack6531]( •[5y ago]( Frickin IR laser beams!!! Reply reply } Share Share []( [Chad3205]( •[5y ago]( Both, although majority is Ir laser Reply reply } Share Share []( [djsolly]( •[5y ago]( I have a vortex huey gen II mounted as far forward on the upper with no riser as I can get it and I can get behind the optic comfortably. Reply reply } Share Share New to Reddit? Create your account and connect with a world of communities. المواصلة باستخدام Google المواصلة باستخدام Google Continue with Email Continue With Phone Number By continuing, you agree to our[User Agreement]( acknowledge that you understand the[Privacy Policy]( More posts you may
|
{"source": 5985, "title": "from dpo"}
|
in Physics Reports that has accrued more than 4500 citations and is the most cited and downloaded article in the history of Physics Reports. His work has been featured in many venues, including being on the cover of the Physical Review Letters issue of 19 March 2010 He is editor of the Journal of Complex Networks since 2013. He is also member of the advisory board of Chaos, a scientific journal of the American Institute of Physics. == Awards == In 2008, Latora has been awarded with the "Giovan Pietro Grimaldi Prize". The motivation of the prize was: "For his works on the structure and dynamics of complex networks". In 2021, Latora has been nominated fellow of the Network Science Society with the following citation: "For seminal work on the structure and dynamics of temporal, multilayer, and higher-order networks, and innovative applications of network science to other disciplines, including neuroscience and urban design." == References == == External links == Personal webpage
|
{"page_id": 42288791, "title": "Vito Latora"}
|
In addition to being an extraordinary artist, he created new techniques, such as carbon dust, that were especially suitable to his subject matter and then-current printing technologies. In 1911 he presided over the creation of the first academic department of medical illustration, which continues to this day. His graduates spread out across the world, and founded a number of the academic programs listed below under "Education". Notable medical illustrators include Max Brödel and Dr. Frank H. Netter. For an online inventory of scientific illustrators including currently already more than 1000 medical illustrators active 1450-1950 and specializing in anatomy, dermatology and embryology, see the Stuttgart Database of Scientific Illustrators 1450–1950 Medical illustration is used in the history of medicine. == Profession == Medical illustrators not only produce such material but can also function as consultants and administrators within the field of biocommunication. A certified medical illustrator continues to obtain extensive training in medicine, science, and art techniques throughout his or her career. The Association of Medical Illustrators is an international organization founded in 1945, and incorporated in Illinois. Its members are primarily artists who create material designed to facilitate the recording and dissemination of medical and bioscientific knowledge through visual communication media. Members are involved not only in the creation of such material, but also serve in consultant, advisory, educational and administrative capacities in all aspects of bioscientific communications and related areas of visual education. The professional objectives of the AMI are to promote the study and advancement of medical illustration and allied fields of visual communication, and to promote understanding and cooperation with the medical profession and related health science professions. The AMI publishes an annual Medical Illustration Source Book which is distributed to creative and marketing professionals that regularly hire medical/scientific image makers for editorial, publishing, educational and advertising
|
{"page_id": 3229777, "title": "Medical illustration"}
|
the amount of hot PV-absorber area and therefore reduces heat losses to the ambient, which improves significantly the efficiency for higher application temperatures. Concentrator systems also often require reliable control systems to accurately track the Sun and to protect the PV cells from damaging over-temperature conditions. However, there are also stationery PVT collector types that use nonimaging reflectors, such as the Compound Parabolic Concentrator (CPC), and do not have to track the Sun. Under ideal conditions, about 75% of the Sun's power directly incident upon such systems can be gathered as electricity and heat at temperatures up to 160 °C. CPVT units that are coupled with thermal energy storage and organic Rankine cycle generators can provide on-demand recovery of up to 70% of their instantaneous electricity generation, and may thus be a fairly efficient alternative to the types of electrical storage which are joined with traditional PV systems. A limitation of high-concentrator (i.e. HCPV and HCPVT) systems is that they maintain their long-term advantages over conventional c-Si/mc-Si collectors only in regions that remain consistently free of atmospheric aerosol contaminants (e.g. light clouds, smog, etc.). Power production is rapidly degraded because 1) radiation is reflected and scattered outside of the small (often less than 1–2 °) acceptance angle of the collection optics, and 2) absorption of specific components of the solar spectrum causes one or more series junctions within the multi-junction cells to under-perform. The short-term impacts of such power generation irregularities can be reduced to some degree with inclusion of electrical and thermal storage in the system. == PVT applications == The range of applications of PVT collectors, and in general solar thermal collectors, can be divided according to their temperature levels: low temperature applications up to 50 °C medium temperature applications up to 80 °C high temperature applications above
|
{"page_id": 21595073, "title": "Photovoltaic thermal hybrid solar collector"}
|
Timaeus (; Ancient Greek: Τίμαιος, romanized: Timaios, pronounced [tǐːmai̯os]) is one of Plato's dialogues, mostly in the form of long monologues given by Critias and Timaeus, written c. 360 BC. The work puts forward reasoning on the possible nature of the physical world and human beings and is followed by the dialogue Critias. Participants in the dialogue include Socrates, Timaeus, Hermocrates, and Critias. Some scholars believe that it is not the Critias of the Thirty Tyrants who appears in this dialogue, but his grandfather, also named Critias. At the beginning of the dialogue, the absence of another, unknown dialogue participant, present on the day before, is bemoaned. It has been suggested from some traditions—Diogenes Laertius (VIII 85) from Hermippus of Smyrna (3rd century BC) and Timon of Phlius (c. 320 – c. 235 BC)—that Timaeus was influenced by a book about Pythagoras, written by Philolaus, although this assertion is generally considered false. == Introduction == The dialogue takes place the day after Socrates described his ideal state. In Plato's works, such a discussion occurs in the Republic. Socrates feels that his description of the ideal state was not sufficient for the purposes of entertainment and that "I would be glad to hear some account of it engaging in transactions with other states" (19b). Hermocrates wishes to oblige Socrates and mentions that Critias knows just the account (20b) to do so. Critias proceeds to tell the story of Solon's journey to Egypt where he hears the story of Atlantis, and how Athens used to be an ideal state that subsequently waged war against Atlantis (25a). Critias believes that he is getting ahead of himself, and mentions that Timaeus will tell part of the account from the origin of the universe to man. Critias also cites the Egyptian priest in Sais about
|
{"page_id": 540497, "title": "Timaeus (dialogue)"}
|
it as their standard. In the 19th century and early 20th century, the pedal division also underwent changes. The pedal divisions of the Baroque era often included a small number of higher-pitched stops, which allowed performers to perform higher melodies on the pedalboard. In the 19th century and early 20th century, organ designers omitted most of these higher-pitched stops, and used pedal divisions which were dominated by 8′ and 16′ stops. This design change, which coincided with the musical trend for music with a deep, rich bass part, meant that players used the pedalboard mainly for bass parts. By the mid-19th century, the pedal part of organ music was increasingly given its own staff, which meant that composers and transcribers began writing organ music in three-stave systems (right hand, left hand, and pedal keyboard). Whereas early organ composers left the way that pedal keyboard lines were played to the player's discretion, in the later 19th century, composers began to indicate specific foot actions. In addition to telling the organist whether to use the left or right foot, symbols indicate whether they should use the toe or heel. A "^" symbol indicates the toe, and a "u" or "o" indicates the heel. Symbols below notes indicate the left foot, and above notes indicates the right foot. Swedish organist L. Nilson published a method for the pedal keyboard, the English translation of which was titled A System of Technical Studies in Pedal Playing for the Organ (Schirmer, 1904). Nilson lamented that it "...is a melancholy fact that only very few eminent organists since Bach's time have made it their business to lift pedal-playing out of its primitive confusion..." (page 1 of Preface). He argued that the great organ pedagogues such as Kittel and Abbe Vogler did not make any efforts to improve
|
{"page_id": 1093053, "title": "Pedal keyboard"}
|
Shape context is a feature descriptor used in object recognition. Serge Belongie and Jitendra Malik proposed the term in their paper "Matching with Shape Contexts" in 2000. == Theory == The shape context is intended to be a way of describing shapes that allows for measuring shape similarity and the recovering of point correspondences. The basic idea is to pick n points on the contours of a shape. For each point pi on the shape, consider the n − 1 vectors obtained by connecting pi to all other points. The set of all these vectors is a rich description of the shape localized at that point but is far too detailed. The key idea is that the distribution over relative positions is a robust, compact, and highly discriminative descriptor. So, for the point pi, the coarse histogram of the relative coordinates of the remaining n − 1 points, h i ( k ) = # { q ≠ p i : ( q − p i ) ∈ bin ( k ) } {\displaystyle h_{i}(k)=\#\{q\neq p_{i}:(q-p_{i})\in {\mbox{bin}}(k)\}} is defined to be the shape context of p i {\displaystyle p_{i}} . The bins are normally taken to be uniform in log-polar space. The fact that the shape context is a rich and discriminative descriptor can be seen in the figure below, in which the shape contexts of two different versions of the letter "A" are shown. (a) and (b) are the sampled edge points of the two shapes. (c) is the diagram of the log-polar bins used to compute the shape context. (d) is the shape context for the point marked with a circle in (a), (e) is that for the point marked as a diamond in (b), and (f) is that for the triangle. As can be seen, since (d)
|
{"page_id": 15281107, "title": "Shape context"}
|
exploding as a type Ib or Ic supernova. For some intermediate range of masses, stars are thought to undergo core collapse at the yellow hypergiant or LBV stage of their lives, resulting in a type IIb or perhaps IIn supernova. HR 8752 may be such a star, and may never make it beyond its current evolutionary state before exploding. == Possible binary == HR 8752 may have a companion. Measurements of the ultraviolet spectral distribution show an excess that corresponds to the output of a B1 main-sequence star. The absolute magnitude was estimated at -4.5, approximately 40 times fainter than the primary at visual wavelengths. Although the stars must be fairly close (< 1400AU), no radial velocity variations have been detected in the spectral lines of the primary, and no lines are observed which can be attributed directly to the secondary. The observed spectrum may be mostly from a shell surrounding both stars. It has been suggested that some variations in spectral line profiles are caused by variations in colliding winds or disturbances of previously ejected material, caused during a periastron passage of the companion. == Notes == == References == == External links == James Kaler's page on HR 8752 Juergen Kummer's page on V509 Cassiopeiae
|
{"page_id": 5096042, "title": "V509 Cassiopeiae"}
|
In population genetics an idealised population is one that can be described using a number of simplifying assumptions. Models of idealised populations are either used to make a general point, or they are fit to data on real populations for which the assumptions may not hold true. For example, coalescent theory is used to fit data to models of idealised populations. The most common idealized population in population genetics is described in the Wright-Fisher model after Sewall Wright and Ronald Fisher (1922, 1930) and (1931). Wright-Fisher populations have constant size, and their members can mate and reproduce with any other member. Another example is a Moran model, which has overlapping generations, rather than the non-overlapping generations of the Fisher-Wright model. The complexities of real populations can cause their behavior to match an idealised population with an effective population size that is very different from the census population size of the real population. For sexual diploids, idealized populations will have genotype frequencies related to the allele frequencies according to Hardy-Weinberg equilibrium. == Hardy-Weinberg == In 1908, G. H. Hardy and Wilhelm Weinberg modeled an idealised population to demonstrate that in the absence of selection, migration, random genetic drift, allele frequencies stay constant over time, and that in the presence of random mating, genotype frequencies are related to allele frequencies according to a binomial square principle called the Hardy-Weinberg law. == Usage in population dynamics == A good example of usage idealised population model, in tracking natural population conditions, could be found in a research of Joe Roman and Stephen R. Palumbi (2003). Using genetic diversity data, they questioned: have populations of North Atlantic great whales recovered enough for commercial whaling? To calculate genetic diversity the authors multiply long term effective population size of the females by two, assuming sex ratio 1:1,
|
{"page_id": 2279743, "title": "Idealised population"}
|
beach in Yabucoa which had become a dumping ground for unwanted pets and stray dogs for years. By late 2017, with the help of the American Humane Society, and Wings of Rescue, the Sato Project had flown 2,000 dogs from the beach in Yabucoa to locations in the U.S. Before and after Hurricane Maria struck the island of Puerto Rico in September 2017, Beckles evacuated many dogs saving them from certain death. Beckles worked to reunite pets with their owners, residents who had left Puerto Rico after the hurricane. Volunteer pilots and "Wings of Rescue" helped the organization with the efforts to evacuate the animals. In the spring of 2018, around the same time that the University of Puerto Rico had taken its first estimated census of stray animals in Puerto Rico (finding there were 300,000 satos and a million stray cats), Sato Project was part of a coalition that launched Puerto Rico's first "spayathon", a free-of-charge, spaying and neutering event to help curb the island's sato (dog and cat) overpopulation. The spayathon also microchipped the animals. Sato Project pushed to have a gate put up to curtail the dumping of dogs on "Dead Dog Beach". More than 100,000 dogs have been spayed or neutered in subsequent "spayathons". In 2020, the organization continued helping by evacuating dogs that had been satos affected by the 2019–2020 Puerto Rico earthquakes. The rescue animals were flown and taken to shelters in Philadelphia, New Jersey and Florida. The project's founder would like to see the sato become Puerto Rico's national dog and advocates for their adoption by people living in the United States and some American celebrities have adopted a sato. Puppies rescued by the Sato Project have been featured in Puppy Bowl, a puppy "parody of the U.S. Super Bowl". == Gallery ==
|
{"page_id": 63371794, "title": "Sato Project"}
|
to the capacity of inferring the state of the system by measuring the trajectories of the system output. It is also closely related to data informativity, which refers to the proper selection of inputs that enables the inference of the unknown parameters. The (lack of) structural identifiability is also important in the context of dynamical compensation of physiological control systems. These systems should ensure a precise dynamical response despite variations in certain parameters. In other words, while in the field of systems identification, unidentifiability is considered a negative property, in the context of dynamical compensation, unidentifiability becomes a desirable property. Identifiability also appears in the context of inverse optimal control. Here, one assumes that the data comes from a solution of an optimal control problem with unknown parameters in the objective function. Here, identifiability refers to the possibility of infering the parameters present in the objective function by using the measured data. == Software == There exist many software that can be used for analyzing the identifiability of a system, including non-linear systems: PottersWheel: MATLAB toolbox that uses profile likelihood for structural and practical identifiability analysis. STRIKE-GOLDD: MATLAB toolbox for structural identifiability analysis. StructuralIdentifiability.jl: Julia library for assessing structural parameter identifiability. LikelihoodProfiler.jl: Julia library for practical identifiability analysis. == See also == System identification Observability Model order reduction Adaptive control == References ==
|
{"page_id": 72323059, "title": "Structural identifiability"}
|
in which research that focusses on specific actions, processes or platforms can rapidly become irrelevant. ‘General’ typologies also provide a more holistic view of the field. ##### 3.7.1.3. The wide range of theoretical perspectives and methods The final challenge arises from the diversity of the theoretical perspectives and research methodologies that are used in social media research. Ngai, Tao & Moon (2015), technology acceptance model, and the theory of reasoned action), _social behavior_ _theories_ (e.g. social aspects theory, social loafing theory, and social power) and _mass communications_ _theories_ (e.g. para-social interaction, and uses and gratifications theory). This contrasts with Leung, Sun & Bai (2017) the main theoretical foundation is word-of mouth. This diverse range of theoretical perspectives is, at least partially, responsible for the wide range of methodological approaches that have been employed in social media marketing research. These include a range of traditional qualitative and quantitative approaches, including questionnaires, interviews, focus groups, diaries, observation, secondary analysis and official statistics, ethnography, participant observation, and mixed methods, as well as more social media-specific approaches such as content analysis applied to social media platforms, social media-based focus groups, social media netnography and virtual ethnography, and social media analytics. #### 3.7.2. Proposals and propositions for future research In developing
|
{"source": 984, "title": "from dpo"}
|
in which each biography is available. Source: Beytía and Schobin (2020) Chapter 3 - Pantheon dataset application 81 > Figure 3.7: Geographical distribution of biographical reference flows. Source: Beytía and Schobin (2020) a different network with different shape, size, and density. From here, it is also possible to see how, in accordance with the intuitions of Figure 3.4, the importance and the presence of the different domains shift over time. Looking at the geographical distribution, Figure 3.7 shows that there is a heavy con-centration of links between Western Europe and the United States, reflecting the strong Western bias already evident in the dataset when dealing with the node’s attributes. In particular, Figure 3.8 (C) shows that individuals from North America tend to cluster together, with relatively few outward connections, mostly directed toward European figures. Meanwhile, European figures exhibit more extensive interactions with other continents while maintaining a significant level of internal connectivity. Figure 3.8 (A) highlights the strong influence of an individual’s occupational domain on the network’s topology. Indeed, clear clusters form around domain affiliations. On the same note, (A) Domains (B) Gender (C) Continents > Figure 3.8: Pantheon network graph using the Yifan Hu algorithm (Hu, 2011). Node colours represent domains (A), gender (B) or continents (C). Edges coloured according to the attribute value of node they originate from. 82 Section 3.3 - Clustering of the Pantheon dataset the distribution of nodes by gender, represented in Figure 3.8 (B), aligns with earlier findings: the majority of the network is predominantly male, with only the artistic and sport domain being more balanced, with female figures forming a denser, more intercon-nected subnetwork, suggesting that women tend share more biographical links in their Wikipedia pages. # 3.3 Clustering of the Pantheon dataset 3.3.1 Non-parametric clustering of Pantheon dataset To apply the non-parametric
|
{"source": 2689, "title": "from dpo"}
|
a large portion of the best thought and consideration of the present generation. The conservation of individual resources and energy will follow as a natural consequence, and that is the keynote of this volume. "The reckless destruction of our forests by the early settlers, supplemented by the wasteful practices of the large lumber companies, and finally, the enormous destruction by forest fires, have brought about greatly changed conditions. Our timber supply has been depleted to such an extent that according to statistics prepared by the United States Government, we take from our forests each year, not counting loss and waste, three and one half times their yearly growth; in other words, we use forty cubic feet per acre for each twelve cubic feet grown. We use two hundred and sixty cubic feet per capita, while Germany uses thirty-seven and France thirty-five. The large per capita consumption in the United States is attributable partly to our great fire losses, but principally to the ephemeral and flimsy character of our buildings. If the present rate of demand and supply is not changed, our timber resources will be entirely exhausted in a few years. Such a result would have far reaching consequences, and would be a national calamity. This can be averted, (1) by substituting incombustible and more durable materials for wood, and (2) by employing proper safeguards and appliances, thus preventing a large part of the annual fire loss. "Twenty to fifty years ago when timber was abundant and the best only was selected for building purposes, much more durable houses were built than at present. Large sized and heavy pieces were used for the framing and the beams. These were frequently hewn by hand, so as to utilize the heart which is the best and most durable part of the timber.
|
{"page_id": 18721214, "title": "A. L. A. Himmelwright"}
|
prove results about skew Young tableaux. We finish with a note on Knuth equivalence and its analog for cylindric tableaux. 56) Yilun Du , On the Algorithmic and Theoretical Exploration of Tiling-Harmonic Functions (6 Jan 2015) In this paper, we explore a new class of harmonic functions defined on a tiling T , a square tiling of a region D , in C . We define these functions as tiling harmonic functions. We develop an efficient algorithm for computing interior values of tiling harmonic functions and graph harmonic functions in a tiling. Using our algorithm, we find that in general tiling harmonic functions are not generally equivalent to graph harmonic functions. In addition, we prove some theoretical results on the structure of tiling harmonic functions and classify one type of tiling harmonic function. 55) Jessica Li , On the Modeling of Snowflake Growth Using Hexagonal Automata (2 Jan 2015; arXiv.org , 8 May 2015; pubished (with Laura P. Schaposnik) in Physical Review E 93:2 (Feb. 2016) ) Snowflake growth is an example of crystallization, a basic phase transition in physics. Studying snowflake growth helps gain fundamental understanding of this basic process and may help produce better crystalline materials and benefit several major industries. The basic theoretical physical mechanisms governing the growth of snowflake are not well understood: whilst current computer modeling methods can generate snowflake images that successfully capture some basic features of actual snowflakes, so far there has been no analysis of these computer models in the literature, and more importantly, certain fundamental features of snowflakes are not well understood. A key challenge of analysis is that the snowflake growth models consist of a large set of partial difference equations, and as in many chaos theory problems, rigorous study is difficult. In this paper we analyze a popular model
|
{"source": 6264, "title": "from dpo"}
|
greenhouse gas. Rather than thinking of longwave radiation headed to space as coming from the surface itself, it is more realistic to think of this outgoing radiation as being emitted by a layer in the mid-troposphere, which is effectively coupled to the surface by a lapse rate. The difference in temperature between these two locations explains the difference between surface emissions and emissions to space, i.e., it explains the greenhouse effect. == Infrared absorbing constituents in the atmosphere == === Greenhouse gases === A greenhouse gas (GHG) is a gas which contributes to the trapping of heat by impeding the flow of longwave radiation out of a planet's atmosphere. Greenhouse gases contribute most of the greenhouse effect in Earth's energy budget. ==== Infrared active gases ==== Gases which can absorb and emit longwave radiation are said to be infrared active and act as greenhouse gases. Most gases whose molecules have two different atoms (such as carbon monoxide, CO), and all gases with three or more atoms (including H2O and CO2), are infrared active and act as greenhouse gases. (Technically, this is because when these molecules vibrate, those vibrations modify the molecular dipole moment, or asymmetry in the distribution of electrical charge. See Infrared spectroscopy.) Gases with only one atom (such as argon, Ar) or with two identical atoms (such as nitrogen, N2, and oxygen, O2) are not infrared active. They are transparent to longwave radiation, and, for practical purposes, do not absorb or emit longwave radiation. (This is because their molecules are symmetrical and so do not have a dipole moment.) Such gases make up more than 99% of the dry atmosphere. ==== Absorption and emission ==== Greenhouse gases absorb and emit longwave radiation within specific ranges of wavelengths (organized as spectral lines or bands). When greenhouse gases absorb radiation,
|
{"page_id": 12395, "title": "Greenhouse effect"}
|
inference === The extant phylogenetic bracket approach allows researchers to infer traits in extinct animals with varying levels of confidence. This is referred to as the levels of inference. There are three levels of inference, with each higher level indicating less confidence for the inference. ==== Inferences based on osteological correlates ==== Level 1 — The inference of a character that leaves a bony signature on the skeleton in both members of the extant sister groups. Example: Saying that Tyrannosaurus rex had an eyeball is a level 1 inference because both extant members of the groups encompassing Tyrannosaurus rex have eyeballs, and eyeball sockets (orbital excavations) in the skull, the homology of which is well established, and the fossils of Tyrannosaurus rex skulls have similar morphology. Level 2 — The inference of a character that leaves a signature on the skeleton of only one of the extant sister groups. For example, saying that Tyrannosaurus rex had air sacs running through its skeleton is a level 2 inference as birds are the only extant sister group to Tyrannosaurus rex to show such air sacs. However the underlying pneumatic fossae, air sacs, in the bones of extant birds are remarkably similar to the cavities seen in the fossil vertebrae of Tyrannosaurus rex. The high degree of similarity between the pneumatic fossae in Tyrannosaurus rex and extant birds makes this a fairly strong inference, yet not as strong as a level 1 inference. Level 3 — The inference of a character that leaves a bony signature on the skeleton but is not present in either extant sister group to the taxon in question. For example, saying that ceratopsian dinosaurs such as Triceratops horridus had horns in life would be a level 3 inference. Neither extant crocodylians, nor extant birds have horns today, but
|
{"page_id": 11000831, "title": "Phylogenetic bracketing"}
|
John H. Safer (September 6, 1922 – December 7, 2018) was an American sculptor. Safer's varied career spanned work in theater lighting, television, real estate, politics and banking. Safer was best known for his monumental sculptures, but he has also created many smaller works. These include award sculptures for organizations such as the National Air and Space Museum, the PGA Tour, the Georgetown University Lombardi Cancer Center, the World Peace Foundation, and the Shakespeare Guild. Safer's works stand in museums, galleries and embassies throughout the world. In 1972 and in 1989 the U.S. Department of State sent a group of Safer sculptures abroad to be exhibited as examples of America's finest art. He died in December 2018 at the age of 96. == Sculpture == Safer's earliest sculptures in the 1950s and 1960s were small works of Lucite. Over time he also began to work in bronze and stainless steel. The pieces became larger and in 1979 his first public commission, Judgment, a multi-ton patinated bronze, was installed at Harvard Law School in Cambridge, Massachusetts. This was the first in a long string of public installations. As the commissions grew in number they grew in size as well. Interplay, created in 1987, is 18 feet (5.5 m) high. Leading Edge, created in 1989, is 20 feet (6.1 m) high. His hallmark work, Ascent, which stands at the entrance of the Smithsonian Institution's Udvar-Hazy Center at Dulles Airport in Virginia, is 75 feet (23 m) high. "Through his work, John has tried to capture the essence and reduce the subject to the pure line in space that Aristotle believed to be the basis of sculpture." == Early life == John Safer was born and raised in Washington, D.C., the only child of John and Rebecca Herzmark Safer. His father, who operated
|
{"page_id": 18937712, "title": "John Safer"}
|
, photon 3 carries this desired state. If the Bell state detected is | ϕ − ⟩ 12 = 1 2 ( | H ⟩ 1 | H ⟩ 2 − | V ⟩ 1 | V ⟩ 2 ) {\displaystyle |\phi ^{-}\rangle _{12}={\frac {1}{\sqrt {2}}}(|H\rangle _{1}|H\rangle _{2}-|V\rangle _{1}|V\rangle _{2})} , then a phase shift of π {\displaystyle \pi } is applied to the state to get the desired quantum state. The distance between the ground station and the satellite changes from as little as 500 km to as large as 1,400 km. Because of the changing distance, the channel loss of the uplink varies between 41 dB and 52 dB. The average fidelity obtained from this experiment was 0.80 with a standard deviation of 0.01. Therefore, this experiment successfully established a ground-to-satellite uplink over a distance of 500–1,400 km using quantum teleportation. This is an essential step towards creating a global-scale quantum internet. === Quantum teleportation over internet cables === Quantum teleportation has been demonstrated over fiber optic cables simultaneously carrying regular telecommunications traffic. This eliminates the need for separate, dedicated infrastructure for quantum networking and shows that quantum teleportation and classical communications can coexist on the same fiber optic cables. A less crowded wavelength of light was used for the quantum signal and special filters were required to reduce noise from other traffic. === Quantum teleportation with nonlinear sum frequency generation === In April 2025, researchers at the University of Illinois Urbana-Champaign achieved quantum teleportation with 94% fidelity using a nanophotonic indium-gallium-phosphide platform to perform nonlinear sum frequency generation (SFG). This method mitigated multiphoton noise and boosted teleportation efficiency by a factor of 10,000 compared to prior SFG-based systems. == Formal presentation == There are a variety of ways in which the teleportation protocol can be written mathematically.
|
{"page_id": 25280, "title": "Quantum teleportation"}
|
6 Plus are its displays; both branded as "Retina HD displays" and "ion-strengthened", the iPhone 6 display is 4.7 inches in size with a 16:9 resolution of 1334x750 (326 ppi, minus one row of pixels), while the iPhone 6 Plus includes a 5.5-inch 1920x1080 (1080p) display (401 PPI). The displays use a multiple-domain LCD panel, dubbed "dual-domain pixels"; the RGB pixels themselves are skewed in a pattern so every pixel is seen from a different angle. This technique helps improve the viewing angles of the display. To accommodate the larger physical size of the iPhone 6 and iPhone 6 Plus, the top button was renamed to "side button" and moved to the side of the phone instead of the top to improve its accessibility. ==== Batteries ==== The iPhone 6 features a 6.91 Wh (1,810 mAh) battery, while the iPhone 6 Plus features an 11.1 Wh (2,915 mAh) battery, neither of which are user-replaceable. ==== Chipsets ==== Both models include an Apple A8 system-on-chip, and an M8 motion co-processor—an update of the M7 chip from the iPhone 5s. The primary difference between the M8 and the original M7 coprocessor is the M8 also includes a barometer to measure altitude changes. Phil Schiller said the A8 chip would provide, in comparison to the 5s, a 25% increase in CPU performance, a 50% increase in graphics performance, and less heat output. Early hands-on reports said the A8's GPU performance might indeed break away from previous generations doubling of performance at each yearly release, scoring 21204.26 in Base mark X compared to 20253.80, 10973.36 and 5034.75 on respectively the 5s, 5 and 4s. ==== Connectivity ==== The expanded LTE connectivity on the iPhone 6 and iPhone 6 Plus is improved to LTE Advanced, with support for over 20 LTE bands (seven more than
|
{"page_id": 42259842, "title": "IPhone 6"}
|
did not engage in juggling. The extent of these changes in the jugglers reduced during a subsequent three-month period in which they did not practice juggling. To further resolve the time course of these changes, the experiment was repeated with another young cohort scanned in shorter intervals, and the by then typical changes in V5 could already be found after just seven days of juggling practice. The observed changes were larger in the initial learning phase than during continued training. Whereas the former two studies involved students in their early twenties, the experiments were recently repeated with an elderly cohort, revealing the same kind of structural changes, although attenuated by lower juggling performance of this group. Using a completely different kind of intervention—application of Transcranial Magnetic Stimulation in daily sessions over five days—changes were observed in and near the TMS target areas as well as in the basal ganglia of volunteers in their mid-twenties, compared to a control group that had received placebo. It is possible, though, that these changes simply reflect vascularization effects. Taken together, these morphometric studies strongly support the notion that brain plasticity—changes of brain structure—remains possible throughout life and may well be an adaptation to changes in brain function which has also been shown to change with experience. The title of this section was meant to emphasize this, namely that plasticity and learning provide two perspectives—functional and structural—at the same phenomenon, a brain that changes over time. === Brain disease === Brain diseases are the field to which brain morphometry is most often applied, and the volume of the literature on this is vast. === Brain evolution === Brain changes also accumulate over periods longer than an individual life but even though twin studies have established that human brain structure is highly heritable, brain morphometric studies
|
{"page_id": 26827372, "title": "Brain morphometry"}
|
Operations engineering is a branch of engineering that is mainly concerned with the analysis and optimization of operational problems using scientific and mathematical methods. More frequently it has applications in the areas of Broadcasting/Industrial Engineering and also in the Creative and Technology Industries. Operations engineering is considered to be a subdiscipline of Operations Research and Operations Management. == Associations == INFORMS Society of Operations Engineers industrial operation == References == == See also == Operations research Systems engineering Enterprise engineering Engineering management Business engineering
|
{"page_id": 49937269, "title": "Operations engineering"}
|
The Mummy! A Tale of the Twenty-Second Century is an 1827 three-volume novel written by Jane Webb (later Jane C. Loudon). It concerns the Egyptian mummy of Cheops, who is brought back to life in the year 2126. The novel describes a future filled with advanced technology, and was the first English-language story to feature a reanimated mummy. After her father's death, making her an orphan at the age of 17, Webb found that: on the winding up of his affairs that it would be necessary to do something for my support. I had written a strange, wild novel, called the Mummy, in which I had laid the scene in the twenty-second century, and attempted to predict the state of improvement to which this country might possibly arrive. She may have drawn inspiration from the general fashion for anything pharaonic, inspired by the French researches during the Napoleonic invasion of Egypt; the 1821 public unwrappings of Egyptian mummies in a theatre near Piccadilly, which she may have attended as a girl; and, very likely, the 1818 novel by Mary Shelley, Frankenstein; or, The Modern Prometheus. As Shelley had written of Frankenstein's creation, "A mummy again endued with animation could not be so hideous as that wretch," which may have triggered her later concept. In any case, at many points she deals in greater clarity with elements from the earlier book such as the loathing for the much-desired object and the immediate arrest for crime and attempt to lie one's way out of it. However, unlike the Frankenstein monster, the hideous revived Cheops is not shuffling around dealing out horror and death, but giving canny advice on politics and life to those who befriend him. In some ways The Mummy! may be seen as her reaction to themes in Frankenstein: her
|
{"page_id": 28219057, "title": "The Mummy! A Tale of the Twenty-Second Century"}
|
HD 18970 is a class G9.5III (yellow giant) star in the constellation Perseus. Its apparent magnitude is 4.77 and it is approximately 211 light years away based on parallax. == References ==
|
{"page_id": 38218575, "title": "HD 18970"}
|
the Queen Square hammer, except that it has a metallic handle that is often detachable. Babinski hammers can also be telescoping, allowing for compact storage. Babinski's hammer was popularized in clinical use in America by the neurologist Abraham Rabiner, who was given the instrument as a peace offering by Babinski after the two brawled at a black tie affair in Vienna. The Trömner reflex hammer was designed by Ernst Trömner. This model is shaped like a two-headed mallet. The larger mallet is used to elicit tendon stretch reflexes, and the smaller mallet is used to elicit percussion myotonia. Other reflex hammer types include the Buck, Berliner and Stookey reflex hammers. There are numerous models available from various commercial sources. == Method of use == The strength of a reflex is used to gauge central and peripheral nervous system disorders, with the former resulting in hyperreflexia, or exaggerated reflexes, and the latter resulting in hyporeflexia or diminished reflexes. However, the strength of the stimulus used to extract the reflex also affects the magnitude of the reflex. Attempts have been made to determine the force required to elicit a reflex, but vary depending on the hammer used, and are difficult to quantify. The Taylor hammer is usually held at the end by the physician, and the entire device is swung in an arc-like motion onto the tendon in question. The Queen Square and Babinski hammers are usually held perpendicular to the tendon in question, and are passively swung with gravity assistance onto the tendon. The Jendrassik maneuver, which entails interlocking of flexed fingers to distract a patient and prime the reflex response, can also be used to accentuate reflexes. In cases of hyperreflexia, the physician may place his finger on top of the tendon, and tap the finger with the hammer. Sometimes
|
{"page_id": 5489731, "title": "Reflex hammer"}
|
Nios Tutorial (Version 1.1) February 2004 Microtronix Embedded Linux Development - Getting Started Guide: Document Revision 1.2 getting_started_guide.pdf Stratix EP1S10 Device: Pin Information February 2004 TMS320C6000 Assembly Language Tools User's Guide toolspdf6000/spru186i.pdf Dynamic Spectrum Allocation In Composite Reconfigurable Wireless Networks IEEE Communications Magazine, May 2004. 1299346&isnumber=28868 TOA - VX-2000 (Digital Matrix System) Klotz Digital - Vadis (Audio Matrix), VariZone (Complex Digital PA System For Emergency Evacuation Applications) Peavey - MediaMatrix System Optimus - Optimus (Audio & Communication), Improve (Distributed Audio) Simplex - TrueAlarm (Fire Alarm Systems) Tyco - Fire Detection and Alarm, Integrated Security Systems, Health Care Communication Systems 10Base-T FPGA Interface - Ethernet Packets: Sending and Receiving Ethernet Receiver The OPC Foundation www.ubicom.com (IP2022) (eZ80) |=[ EOF ]=---------------------------------------------------------------=| -------------------------------------------------------------------------------- ==Phrack Inc.== Volume 0x0b, Issue 0x3f, Phile #0x12 of 0x14 |=------=[ hiding processes ( understanding the linux scheduler ) ]=----=| |=-----------------------------------------------------------------------=| |=------------=[ by ubra from PHI Group -- 17 October 2004 ]=-----------=| |=-----=[ mail://ubra_phi.group.za.org ]=----=| --[ Table of contents 1 - looking back 2 - the schedule(r) inside 3 - abusing the silence ( attacking ) 4 - can you scream ? ( countering ) 5 - references 6 - and the game dont stop.. 7 - sources --[ 1 - looking back We begin our journey in the old days, when simply giving your process a weird name was enough to hide inside the tree. Sadly this is also quite effective these days due to lack of skill from stock admins. In the last millenium ..well actualy just before 1999, backdooring binaries was very popular (ps, top, pstree and others ) but this
|
{"source": 1201, "title": "from dpo"}
|
is a normalization coefficient that should approximate the maximum expected value (the default value set in NLTK is 15). The usage of this class is immediate, as the following snippet can confirm: > from nltk.sentiment.vader import SentimentIntensityAnalyzer >>> text ='This is avery interesting and quite powerful sentiment analyzer' >>> vader = SentimentIntensityAnalyzer() >>> print(vader.polarity_scores(text)) {'neg': 0.0, 'neu': 0.535, 'pos': 0.465, 'compound': 0.7258} The NLTK Vader implementation uses the library Twython for some functionalities. Even though it's not necessary, in order to avoid a warning, it's possible to install it using pip ( pip install twython ). Topic Modeling and Sentiment Analysis in NLP [ 287 ] # References Hofmann T., Unsupervised Learning by Probabilistic Latent Semantic Analysis , Machine Learning 42, 177-196, 2001, Kluwer Academic Publishers. Blei D., Ng A., Jordan M., Latent Dirichlet Allocation, Journal of Machine Learning Research , 3, (2003) 993-1022. Hutto C.J., Gilbert E., VADER: A Parsimonious Rule-based Model for Sentiment Analysis of Social Media Text , AAAI, 2014. # Summary In this chapter, we introduced topic modeling. We discussed latent semantic analysis based on truncated SVD, probabilistic latent semantic analysis (which aims to build a model without assumptions about latent factor prior probabilities), and latent Dirichlet allocation, which outperformed the previous method and is based on the assumption that the latent factor has a sparse prior Dirichlet distribution. This means that a document normally covers only a limited number of topics and a topic is characterized only by a few important words. In the last section, we discussed sentiment analysis of documents, which is aimed at determining whether a piece of text expresses a positive or negative feeling. In order to show a feasible solution, we built a classifier based on an NLP pipeline and a random forest with average performances that can be used
|
{"source": 3662, "title": "from dpo"}
|
a correct answer can be rejected after only two tokens, underscoring the potential for relaxed verification schemes.1 1 1 Additional examples of this phenomenon are provided in Appendix C.1 for further examination. Intuitively, one would expect from a well-calibrated verification scheme to allow for accepting candidate tokens whenever they are contextually correct. However, as we show in the following two paragraphs, this not the case for standard logits-based verification. #### High-quality draft model. To further demonstrate how valid responses incur high rejection rates, we perform the following experiment: We take a very powerful LLM as the draft model and evaluate whether the target model accepts more candidate tokens, which are now guaranteed to be of high quality. While such a setup does not make sense for SD from an efficiency point of view (a powerful drafter is of course too slow), it further investigates if acceptance rates improve with the quality of responses. To that end we use GPT-4o as draft model for the target Llama-405B. We generate full answers with GPT-4o on MT-Bench,GSM8K and HumanEval and simply check how many tokens the target accepts under greedy decoding before the first rejection, as there is no way to properly perform SD with closed-source models. In order to ensure that the target model is able to “recognize” the high-quality tokens, we use the performant Llama-405B. We display the average acceptance length and an example prompt in Fig.3. Counter-intuitively, we find that the target model does not reward the higher quality of
|
{"source": 5688, "title": "from dpo"}
|
Like the Preferred AEAD Ciphersuites, the list is ordered. Algorithm IDs are defined in Section 9.4¶¶, any local certification signatures are deleted from the key.( The receiver of a transported key "imports" it and likewise trims any local certifications. In normal operation, there won't be any local certifications, assuming the import is performed on an exported key. However, there are instances where this can reasonably happen. For example, if an implementation allows keys to be imported from a key database in addition to an exported key, then this situation can arise.( Some implementations do not represent the interest of a single user (for example, a
|
{"source": 6720, "title": "from dpo"}
|
classes of dienes are oxo- and aza- butadienes. The key quality of a good DAINV diene is a significantly lowered HOMO and LUMO, as compared to standard DA dienes. Below is a table showing a few commonly used DAINV dienes, their HOMO and LUMO energies, and some standard DA dienes, along with their respective MO energies. == Common dienophiles == The dienophiles used in inverse electron demand Diels-Alder reactions are, unlike in the standard DA, very electron rich, containing one or more electron donating groups. This results in higher orbital energies, and thus more orbital overlap with the LUMO of the diene. Common classes of dieneophiles for DAINV reaction include vinyl ethers and vinyl acetals, imine, enamines, alkynes and highly strained olefins. The most important consideration in choice of dienophile is its relative orbital energies. Both HOMO and LUMO impact the rate and selectivity of the reaction. A table of common DAINV dienophiles, standard DA dienophiles, and their respective MO energies can be seen below. A second table shows how electron richness in the dienophiles affects the rate of reaction with a very electron poor diene, namely hexachlorocyclopentadiene. The more electron rich the dienophile is, the higher the rate of the reaction will be. This is very clear when comparing the relative rates of reaction for styrene and the less electron rich p-nitrostyrene; the more electron rich styrene reactions roughly 40% faster than p-nitrostyrene. == Scope and applications == DAINV reactions provide a pathway to a rich library of synthetic targets, and have been utilized to form many highly functionalized systems, including selectively protected sugars, an important contribution to the field of sugar chemistry. In addition, DAINV reactions can produce an array of different products from a single starting material, such as tetrazine. DAINV reactions have been utilized for the
|
{"page_id": 33509814, "title": "Inverse electron-demand Diels–Alder reaction"}
|
intermediate areas, and steep at the largest areas. The formula for species composition may be used to calculate the expected number of species present in a community under the assumptions of the Unified Theory. In symbols E { S | θ , J } = θ θ + θ θ + 1 + θ θ + 2 + ⋯ + θ θ + J − 1 {\displaystyle E\left\{S|\theta ,J\right\}={\frac {\theta }{\theta }}+{\frac {\theta }{\theta +1}}+{\frac {\theta }{\theta +2}}+\cdots +{\frac {\theta }{\theta +J-1}}} where θ is the fundamental biodiversity number. This formula specifies the expected number of species sampled in a community of size J. The last term, θ / ( θ + J − 1 ) {\displaystyle \theta /(\theta +J-1)} , is the expected number of new species encountered when adding one new individual to the community. This is an increasing function of θ and a decreasing function of J, as expected. By making the substitution J = ρ A {\displaystyle J=\rho A} (see section on saturation above), then the expected number of species becomes Σ θ / ( θ + ρ A − 1 ) {\displaystyle \Sigma \theta /(\theta +\rho A-1)} . The formula above may be approximated to an integral giving S ( θ ) = 1 + θ ln ( 1 + J − 1 θ ) . {\displaystyle S(\theta )=1+\theta \ln \left(1+{\frac {J-1}{\theta }}\right).} This formulation is predicated on a random placement of individuals. === Example === Consider the following (synthetic) dataset of 27 individuals: a,a,a,a,a,a,a,a,a,a,b,b,b,b,c,c,c,c,d,d,d,d,e,f,g,h,i There are thus 27 individuals of 9 species ("a" to "i") in the sample. Tabulating this would give: a b c d e f g h i 10 4 4 4 1 1 1 1 1 indicating that species "a" is the most abundant with 10 individuals and species
|
{"page_id": 635546, "title": "Unified neutral theory of biodiversity"}
|
Titan Commerce for distribution and support operations of the NeuroSky MindSet and ThinkGear technology components to the EU. Mind Games: MindGames is an Iceland-based developer of games that utilize EEG technology. MindGames is known for Tug of Mind, a game set to be released for future NeuroSky products. MindGames' long-term goal is to develop games that teach people how to control negative emotions. However, it is not clear whether or not MindGames simply produces games for NeuroSky products or whether they have an official partnership. Universities and Research Institutions: Because NeuroSky also produces research grade EEG technology it has partnerships with a number of universities and research intuitions including: Johns Hopkins, Brown University, Duke University, University of California San Diego, San Jose State University, Dongguk University, University of Glasgow, The Hong Kong PolyTechnic University, and Trinity College, Dublin. == NeuroSky products == MindWave: The MindWave is a NeuroSky product released in 2010 in China and 2011 in the US and the EU. It costs $99 US making it one of the least expensive EEG device to ever be produced. The MindWave has been marketed as both an education and entertainment device. The MindWave won the Guinness Book of World Records award for “Heaviest machine moved using a brain control interface”. Research Products: Neurosky has produced a number of research products (generally very similar to their direct to consumer products ). These products have been widely adopted due to positive reviews by early adopters with Richard Reilly Ph.D., professor at Trinity College in Dublin stating that, “We have been impressed with the quality of data from the MindKit Pro, compared to our gold-standard EEG acquisitions system. It has opened new possibilities for the remote monitoring patients’ activity a great benefit in patient care.” Modules: TGAM, ThinkGear Asic Module, hardware module for
|
{"page_id": 10231053, "title": "NeuroSky"}
|
it. Each person has two choices – yes or no – and these choices are independent of those of the other people. Therefore there are 2 × 2 × ⋯ 2 = 2 n {\displaystyle 2\times 2\times \cdots 2=2^{n}} possibilities. Alternatively, one may observe that the size of the committee must be some number between 0 and n {\displaystyle n} . For each possible size k {\displaystyle k} , the number of ways in which a committee of k {\displaystyle k} people can be formed from n {\displaystyle n} people is the binomial coefficient ( n k ) . {\displaystyle {n \choose k}.} Therefore the total number of possible committees is the sum of binomial coefficients over k = 0 , 1 , 2 , … , n {\displaystyle k=0,1,2,\dots ,n} . Equating the two expressions gives the identity ∑ k = 0 n ( n k ) = 2 n , {\displaystyle \sum _{k=0}^{n}{n \choose k}=2^{n},} a special case of the binomial theorem. A similar double counting method can be used to prove the more general identity ∑ k = d n ( n k ) ( k d ) = 2 n − d ( n d ) {\displaystyle \sum _{k=d}^{n}{n \choose k}{k \choose d}=2^{n-d}{n \choose d}} === Handshaking lemma === Another theorem that is commonly proven with a double counting argument states that every undirected graph contains an even number of vertices of odd degree. That is, the number of vertices that have an odd number of incident edges must be even. In more colloquial terms, in a party of people some of whom shake hands, an even number of people must have shaken an odd number of other people's hands; for this reason, the result is known as the handshaking lemma. To prove this by double
|
{"page_id": 296942, "title": "Double counting (proof technique)"}
|
a process perspective on resilience: IRP can be interpreted either as an antecedent of how a setback affects an innovator, or as an outcome of the process that is influenced by the setback situation. A measurement scale of IRP was developed and validated in 2018. == Cultural differences == There is controversy about the indicators of good psychological and social development when resilience is studied across different cultures and contexts. The American Psychological Association's Task Force on Resilience and Strength in Black Children and Adolescents, for example, notes that there may be special skills that these young people and families have that help them cope, including the ability to resist racial prejudice. Researchers of indigenous health have shown the impact of culture, history, community values, and geographical settings on resilience in indigenous communities. People who cope may also show "hidden resilience" when they do not conform with society's expectations for how someone is supposed to behave (for example, in some contexts aggression may aid resilience, or less emotional engagement may be protective in situations of abuse). === Resilience in individualist and collectivist communities === Individualist cultures, such as those of the U.S., Austria, Spain, and Canada, emphasize personal goals, initiatives, and achievements. Independence, self-reliance, and individual rights are highly valued by members of individualistic cultures. The ideal person in individualist societies is assertive, strong, and innovative. People in this culture tend to describe themselves in terms of their unique traits—"I am analytical and curious". Economic, political, and social policies reflect the culture's interest in individualism. Collectivist cultures, such as those of Japan, Sweden, Turkey, and Guatemala, emphasize family and group work goals. The rules of these societies promote unity, brotherhood, and selflessness. Families and communities practice cohesion and cooperation. The ideal person in collectivist societies is trustworthy, honest, sensitive, and
|
{"page_id": 3735409, "title": "Psychological resilience"}
|
tyre. For The Sixth Wars the tyre was replaced by a metal bumper. The tyre returned in Extreme 2 and remained for the remainder of the shows' run. The pit was enlarged and given an LED decoration on the inside for the eighth series, with the pyrotechnic no longer used. Since Series 9, the pyrotechnics have returned to the Pit. Flame Pit - The Second Wars onwards - A large square grill located close to the right bottom corner of the arena resembling a kitchen stove. Used to set fire to flammable robots and damage electronics. For the eighth series, this was moved to the upper left corner of the arena. (similar to the floor grills which only appeared in the first wars) The Corner Patrol Zones (commonly known as CPZs) - The Third Wars onwards - the House Robots were now confined to the four corners of the arena much like in The First Wars. Which house robot was used in which battle was defined by a rota system but would also sometimes depend on if a house robot was being repaired. In The Fourth Wars, the house robots switched to a rota system of Shunt, Matilda, Sgt. Bash and Dead Metal. Sir Killalot consistently appeared every round excepting battles 3 & 4 of the War of Independence when the original 4 house robots occupy all the CPZs. From The Fifth Wars and Extreme 1 onwards only two house robots were allowed in the arena at a time and this moved in a rota through all of the machines. However some House Robots did not appear in a show or on some occasions missed several shows before making an appearance. For Series 8, the CPZs are larger and L-shaped. The house robots no longer attack competitors who enter any
|
{"page_id": 364093, "title": "Robot Wars (TV series)"}
|
Moss is considered a raised bog, because its general situation is on higher ground, at the edge of the Mounth, a coastal mountain spur of the Grampian Mountains overlooking the North Sea. There are rock outcrops and strewn boulders relict from the glacial age at this site. Elevations within the Portlethen Moss range from approximately 35 to 60 metres above sea level. Formation of this moss has occurred due to extensive sets of depressions in the underlying rock formations of Old Red Sandstone. The entire water composition of the bog thus has been provided by precipitation with no source of surface runoff, since the topography reduces to lower elevations in every direction. Due to the high winds, moderate precipitation and cool temperatures that generally prevail, conditions are favourable for the formation of an acid bog, since water stagnates, but eventually evaporates with ensuing acidity enhancement of decaying organic matter. There was virtually no drainage outlet in prehistoric times, and little drainage even in modern times. == Evolution of Portlethen Moss == Many coastal mosses were initiated by the process of glaciation, which sheared rock formations to a generally level terrain, while also gouging moderate-sized craters that would pond. This description fits the fundamental situation of Portlethen Moss, where sphagnum would have flourished over millennia of evaporation, further intensifying the soil acidity, fueled by organic matter decaying, with little drainage outlet. A layer of sphagnum moss would have developed at the benthic level of the bog, and additional sphagnum layers floated in mats atop the bog. At an intermediate level of evolution, thick peat layers formed from the decay and carbonisation of the rotting sphagnum. Generations of Carex and Juncus flourished, leading to further decay of these materials and eventual heightening of the organic mass. Finally, secondary vegetation took root in
|
{"page_id": 6622971, "title": "Portlethen Moss"}
|
The OpenHPSDR (High Performance Software Defined Radio) project dates from 2005 when Phil Covington, Phil Harman, and Bill Tracey combined their separate projects to form the HPSDR group. It is built around a modular concept which encourages experimentation with new techniques and devices (e.g. SDR, Envelope Elimination and Restoration) without the need to replace the entire set of boards. The project has expanded from the original group, and several additional people have been involved in recent HPSDR module designs. The core modules of the project are the Atlas passive backplane, the Ozy interface which provides a USB 2.0 data channel between the HPSDR system and the host PC, and the Mercury and Penelope receiver and exciter boards, which use high speed ADCs and DACs for direct conversion of received or transmitted signals in the DC to 55 MHz frequency range. Mercury has attracted wide interest within the HPSDR community as a general-coverage, high performance, HF receiver. It uses a 16-bit 135MSPS analog-to-digital converter that provides performance over the range 0 to 55 MHz comparable to that of a conventional analog HF radio. The receiver will also operate in the VHF and UHF range using either mixer image or alias responses. The host computer uses DSP techniques to process the digital bitstream it receives from the HPSDR system. Currently, the HPSDR hardware has been interfaced with the Flex-Radio PowerSDR Windows-based software, which is licensed under the GPL. As of February, 2011, the following HPSDR modules have been released: Atlas backplane Magister control board Janus I/Q interface board Mercury Direct Down-conversion receiver LPU linear power supply Pandora enclosure Excalibur frequency reference board PennyWhistle 20 watt RF power amplifier Hercules 100 watt RF power amplifier Metis 1Gigbit (1000T) PC interface board Other modules nearing release include: Alex RX/TX filter bank Replaced by newer
|
{"page_id": 22202617, "title": "OpenHPSDR"}
|
architecture characterised by columns which stood on the flat pavement of a temple without a base, their vertical shafts fluted with parallel concave grooves topped by a smooth capital that flared from the column to meet a square abacus at the intersection with the horizontal beam that they carried. Dormer A structural element of a building that protrudes from the plane of a sloping roof surface. Dormers are used, either in original construction or as later additions, to create usable space in the roof of a building by adding headroom and usually also by enabling addition of windows. Dosseret, or impost block A cubical block of stone above the capitals in a Byzantine church, used to carry the arches and vault, the springing of which had a superficial area greatly in excess of the column which carried them. Double-depth plan A plan for a structure that is two rooms deep but lacking a central corridor. Dromos An entrance passage or avenue leading to a building, tomb or passageway. Those leading to beehive tombs are enclosed between stone walls and sometimes in-filled between successive uses of the tomb. In ancient Egypt the dromos was a straight, paved avenue flanked by sphinxes. Dutch gable A gable whose sides have a shape made up of one or more curves and has a pediment at the top. == E == Eave return An element of Classical Revival architecture in American domestic architecture. Egg-and-dart An ornamental moulding in which an ovolo is inscribed with alternating oval and V-shaped motifs. Enfilade A row of rooms with aligned doorways, creating a linear processional route. Enfilades were common in upper-class Baroque architecture and are used in museum layouts to manage flow. Engaged column A column built into and partially projecting from a wall, particularly notable in Roman architecture.
|
{"page_id": 3104368, "title": "Glossary of architecture"}
|
hypertensive rats that were subjected to different combinations of folic acid treatments. The results showed that folic acid attributed to the increase of nitric oxide synthesis of endothelial cell and decrease of plasma. The effects of folic acid consumption caused the offset of hypertension.Curcumin is a nutritional agent that protects against damage to organs, genotoxicty, and oxidative stress, induced by tartrazine, a synthetic food coloring agent. Tartrazine is a known toxin if exceeded upon the recommended dosage amount. The research was done on rats that were subjected to different dosages of tartrazine and curcumin. In the findings it showed curcumin was able to counteract the adverse effects and protect from further infliction. Restricting caloric and protein intake are also two most common means of observing impact of nutritional agents. A pregnant rodent may have their caloric intake reduced up to 30-50% of normal intake. Protein restricted rodents are given 8-9% casein, as opposed to control rats that are fed 20% casein. Micronutrients, such as zinc and iron, may also be restricted to investigate the effects on offspring. Additionally, rats fed diets lacking or including methyl donors are often used to study the effects of diet on epigenomics, as variations within the methylation of DNA are common means of silencing or expressing genes. Supplementing maternal mice with folic acid, vitamin B12, choline and betaine leads to increased levels of DNA methylation at CpG sites and causes a coat color change. This is an example of epigenetically modifiable loci called a “metastable epiallele”, of which only a few have been identified. == Future directions == The possibilities of utilizing nutriepigenomics for intervention are quite expansive. This can include preventative therapies, such as providing an optimal regime for nutrition during pregnancy and lactation. It is already common place for pregnant mothers to supplement
|
{"page_id": 30932051, "title": "Nutriepigenomics"}
|
in OpenSSL. In this paper, we focus on its impact on traditional cache side channel attacks, specifically on Flush +Reload. In Sandy Bridge and successive processors, each core is equipped with four types of hardware prefetchers: Streaming prefetcher, Spatial prefetcher, Data Cache Unit (DCU) prefetcher, and Instruction pointer (IP)-based stride prefetcher . Unfortunately, details about the behavior of those prefetchers are not publicly known, except for a brief explanation of each prefetching mechanism in Intel docu-ments , . Those prefetchers can be quite aggressive. For example, according to the documents, the streaming prefetcher could prefetch up to 20 cache lines ahead. > III. DESIGN AND IMPLEMENTATION A revised Flush +Reload attack that overcomes the nega-tive impacts caused by hardware prefetching is proposed here. Figure 2 demonstrates a high-level workflow of our > 21252 VOLUME 9, 2021 Z. Wang et al. : Defeating Hardware Prefetchers in Flush +Reload Side-Channel Attack > FIGURE 2. Workflow of Prefetcher Aware Flush +Reload Attack. attack. We first conduct reverse engineering towards the prefetcher, and then use the obtained statistical description of the prefetcher to build up an evaluation model of the prefetch-ing impact on Flush +Reload attack, which is regarded as the loss in the attack scenario. Using the obtained loss function, we then fine tune the placement of probes to minimize the loss value. We then take the new probe set into a Flush +Reload attack that circumvents the interference of the prefetcher. The whole process is done automatically, which potentially enables the attack to be migrated to other CPUs with little effort. To simplify explanations, and without loss of generality, we assume that the attacker targets a single memory page. For the cases that target multiple memory pages, the attacker can do the same for each page. In the case of cache miss,
|
{"source": 2289, "title": "from dpo"}
|
Writer is a type alias for WriterT , so we can read types like WriterT[Id, W, A] as Writer[W, A] :110 CHAPTER 4. MONADS type Writer[W, A] = WriterT[Id, W, A] For convenience, Cats provides a way of creaঞng Writers specifying only the log or the result. If we only have a result we can use the standard pure syntax. To do this we must have a Monoid[W] in scope so Cats knows how to produce an empty log: import cats.instances.vector._ // for Monoid import cats.syntax.applicative._ // for pure type Logged[A] = Writer[Vector[String], A] 123.pure[Logged] // res2: Logged[Int] = WriterT((Vector(),123)) If we have a log and no result we can create a Writer[Unit] using the tell syntax from cats.syntax.writer : import cats.syntax.writer._ // for tell Vector("msg1", "msg2", "msg3").tell // res3: cats.data.Writer[scala.collection.immutable.Vector[String], Unit] = WriterT((Vector(msg1, msg2, msg3),())) If we have both a result and a log, we can either use Writer.apply or we can use the writer syntax from cats.syntax.writer : import cats.syntax.writer._ // for writer val a = Writer(Vector("msg1", "msg2", "msg3"), 123) // a: cats.data.WriterT[cats.Id,scala.collection.immutable.Vector[ String],Int] = WriterT((Vector(msg1, msg2, msg3),123)) val b = 123.writer(Vector("msg1", "msg2", "msg3")) // b: cats.data.Writer[scala.collection.immutable.Vector[String],Int] = WriterT((Vector(msg1, msg2, msg3),123)) We can extract the result and log from a Writer using the value and written methods respecঞvely: 4.7. THE WRITER MONAD 111 val aResult: Int = a.value // aResult: Int = 123 val aLog: Vector[String] = a.written // aLog: Vector[String] = Vector(msg1, msg2, msg3) We can extract both values at the same ঞme using the run method: val (log, result) = b.run // log: scala.collection.immutable.Vector[String] = Vector(msg1, msg2, msg3) // result: Int = 123 4.7.2 Composing and Transforming Writers The log in a Writer is preserved when we map or flatMap over it. flatMap appends the logs from the source Writer and the result of the user’s
|
{"source": 4183, "title": "from dpo"}
|
within the array will be set to the required length. If the pValue of elements within the array is not NULL_PTR, then the ulValueLen element of attributes within the array MUST reflect the space that the corresponding pValue points to, and pValue is filled in if there is sufficient room. Therefore it is important to initialize the contents of a buffer before calling C_GetAttributeValue to get such an array value. If any ulValueLen within the array isn't large enough, it will be set to CK_UNAVAILABLE_INFORMATION and the function will return CKR_BUFFER_TOO_SMALL, as it does if an attribute in the pTemplate argument has ulValueLen too small. Note that any attribute whose value is an array of attributes is identifiable by virtue of the attribute type having the CKF_ARRAY_ATTRIBUTE bit set. Note that the error codes CKR_ATTRIBUTE_SENSITIVE, CKR_ATTRIBUTE_TYPE_INVALID, and CKR_BUFFER_TOO_SMALL do not denote true errors for C_GetAttributeValue. If a call to C_GetAttributeValue returns any of these three values, then the call MUST nonetheless have processed every attribute in the template supplied to C_GetAttributeValue. Each attribute in the template whose value can be returned by the call to C_GetAttributeValue will be returned by the call to C_GetAttributeValue. Return values: CKR_ARGUMENTS_BAD, CKR_ATTRIBUTE_SENSITIVE, CKR_ATTRIBUTE_TYPE_INVALID, CKR_BUFFER_TOO_SMALL, CKR_CRYPTOKI_NOT_INITIALIZED, CKR_DEVICE_ERROR, CKR_DEVICE_MEMORY, CKR_DEVICE_REMOVED, CKR_FUNCTION_FAILED, CKR_GENERAL_ERROR, CKR_HOST_MEMORY, CKR_OBJECT_HANDLE_INVALID, CKR_OK, CKR_SESSION_CLOSED, CKR_SESSION_HANDLE_INVALID. Example: CK_SESSION_HANDLE hSession; CK_OBJECT_HANDLE hObject; CK_BYTE_PTR pModulus, pExponent; CK_ATTRIBUTE template[] = { {CKA_MODULUS, NULL_PTR, 0}, {CKA_PUBLIC_EXPONENT, NULL_PTR, 0} }; CK_RV rv; . . rv = C_GetAttributeValue(hSession, hObject, &template, 2); if (rv == CKR_OK) { pModulus = (CK_BYTE_PTR) malloc(template.ulValueLen); template.pValue = pModulus; /* template.ulValueLen was set by C_GetAttributeValue */ pExponent = (CK_BYTE_PTR) malloc(template.ulValueLen); template.pValue = pExponent; /* template.ulValueLen was set by C_GetAttributeValue */ rv = C_GetAttributeValue(hSession, hObject, &template, 2); if (rv == CKR_OK) { . . } free(pModulus); free(pExponent); } ¨ C_SetAttributeValue CK_DEFINE_FUNCTION(CK_RV, C_SetAttributeValue)( CK_SESSION_HANDLE hSession, CK_OBJECT_HANDLE hObject, CK_ATTRIBUTE_PTR pTemplate,
|
{"source": 6021, "title": "from dpo"}
|
binding strength to passively release the growth factors, along with active release via mechanisms such as hybridization of the aptamer with complementary oligonucleotides or unfolding of the aptamer due to cellular traction forces. === Tissue Engineering Application === Aptamer, known for their ability to bind specific molecules reversibly, have been used in 3D bioprinting tissues to precisely deliver growth factors to promote vascularization. This controlled delivery allows growth factors to be released at the right place and time, encouraging the formation of localized and complex vascular networks. Additionally, the properties of these networks can be fine-tuned by adjusting how growth factors are released over time, making this approach a powerful tool for creating vascularized engineered tissues. === AptaBiD === AptaBiD (Aptamer-Facilitated Biomarker Discovery) is an aptamer-based method for biomarker discovery. == Peptide Aptamers == While most aptamers are based on DNA, RNA, or XNA, peptide aptamers are artificial proteins selected or engineered to bind specific target molecules. === Structure === Peptide aptamers consist of one or more peptide loops of variable sequence displayed by a protein scaffold. Derivatives known as tadpoles, in which peptide aptamer "heads" are covalently linked to unique sequence double-stranded DNA "tails", allow quantification of scarce target molecules in mixtures by PCR (using, for example, the quantitative real-time polymerase chain reaction) of their DNA tails. The peptides that form the aptamer variable regions are synthesized as part of the same polypeptide chain as the scaffold and are constrained at their N and C termini by linkage to it. This double structural constraint decreases the diversity of the 3D structures that the variable regions can adopt, and this reduction in structural diversity lowers the entropic cost of molecular binding when interaction with the target causes the variable regions to adopt a uniform structure. === Selection === The most
|
{"page_id": 1970691, "title": "Aptamer"}
|
remaining after the evaporation of water from a sample. Upper critical solution temperature – Critical temperature of miscibility in a mixture Lower critical solution temperature – Critical temperature below which components of a mixture are miscible for all compositions Coil–globule transition – Collapse of a macromolecule from an expanded coil state to a collapsed globule state == References == IUPAC, Compendium of Chemical Terminology, 2nd ed. (the "Gold Book") (1997). Online corrected version: (2006–) "solution". doi:10.1351/goldbook.S05746 == External links == Media related to Solutions at Wikimedia Commons
|
{"page_id": 28729, "title": "Solution (chemistry)"}
|
is the central element of the traditional "French polish" method of finishing furniture, fine string instruments, and pianos. Shellac, being edible, is used as a glazing agent on pills (see excipient) and sweets, in the form of pharmaceutical glaze (or, "confectioner's glaze"). Because of its acidic properties (resisting stomach acids), shellac-coated pills may be used for a timed enteric or colonic release. Shellac is used as a 'wax' coating on citrus fruit to prolong its shelf/storage life. It is also used to replace the natural wax of the apple, which is removed during the cleaning process. When used for this purpose, it has the food additive E number E904. Shellac is an odour and stain blocker and so is often used as the base of "all-purpose" primers. Although its durability against abrasives and many common solvents is not very good, shellac provides an excellent barrier against water vapour penetration. Shellac-based primers are an effective sealant to control odours associated with fire damage. Shellac has traditionally been used as a dye for cotton and, especially, silk cloth in Thailand, particularly in the north-eastern region. It yields a range of warm colours from pale yellow through to dark orange-reds and dark ochre. Naturally dyed silk cloth, including that using shellac, is widely available in the rural northeast, especially in Ban Khwao District, Chaiyaphum province. The Thai name for the insect and the substance is "khrang" (Thai: ครั่ง). ==== Wood finish ==== Wood finishing is one of the most traditional and still popular uses of shellac mixed with solvents or alcohol. This dissolved shellac liquid, applied to a piece of wood, is an evaporative finish: the alcohol of the shellac mixture evaporates, leaving behind a protective film. Shellac as wood finish is natural and non-toxic in its pure form. A finish made of
|
{"page_id": 54813, "title": "Shellac"}
|
most of his time telling stories to the child and to protect the dangers of the forest. The four passengers take turns at the microphone to occupy the little Claude. Impatient at the side trip, Martin tried to encourage Claude to enter a dangerous cave, and arrive sooner at Sidoine. Silbad surprises and violently beats him, then Max locks Martin in a cabin of his spaceship. Belle is shocked by the attitude of her husband. When they make a stopover on the planet Gamma 10, Martin takes the opportunity to escape. Max starts to pursue him, but found Martin's dead body, inert in the sand. He is soon surrounded by a bunch of ex-prisoners of Sidoine stranded on the planet and is taken to a troglodyte village. Silbad eventually comes to search for Max, and is also captured. The two men are presented to the Master who intends to feed them to his hungry monster. But Max and Silbad manage to escape, then provide food to the Master and his monster, and promises to bring back help for the prisoners stranded on the planet. In his approach to the planet Perdide, The Big Max is caught in unexpectedly intense interstellar and interplanetary traffic and are detained by the police. Max and Silbad question the authorities and learn to their amazement that Perdide is a developed planet, urbanized and largely colonized since its "Enhancement", nearly sixty years earlier. Max and Silbad realise that their space journey at 99% of the speed of light created a time lag of a century between them and the planet Perdide. Hearing the news, Silbad has a heart attack. The Big Max lands on the planet Perdide and Max goes in search of a local historian to understand how the tragic story of the little Claude
|
{"page_id": 52183575, "title": "L'Orphelin de Perdide"}
|
Japan, Taiwan or China (Mainland China and Hong Kong)). KOO: South Korea XAR: North America (USA) XAC: North America (Canada) MXO: Latin America (Mexico) ZBR: Latin America (Other) EUX: Europe (1) EUY: Europe (2) SER: CIS (Russia) SEK: CIS (Ukraine) CAU: CIS (Caucasus) ILO: Middle East (Israel) TUR: Middle East (Turkey) XSG: Middle East (Other, formerly UAE) INU: India XSP: Asia Pacific (formerly Singapore/Southeast Asia) BRI: Taiwan ===== Cellular devices ===== OKR: South Korea (cellular devices only) OYM/OYN: United States (cellular devices only) OYA/OYV: Canada (cellular devices only) OWB/OWO/OWE: Latin America and the Caribbean (some devices) OJM: Middle East (including Israel) and Africa (some devices) ODM/ODI: India/South Asia (some devices) OLM/OLN/OLO: Southeast Asia (excluding Indonesia) and Oceania (some devices) OLE/OLP: Indonesia (all devices, except smartwatches) OXE: Post-Soviet states (some devices) CHN/CHC: Mainland China (all devices) OZS: Taiwan and Hong Kong (smartphones only) XJP/SJP: Japan (some devices) OXM: Europe/All other countries (including US/Canada/South Korea/Hong Kong/Taiwan Wi-Fi only devices, as well as Hong Kong/Taiwan cellular tablets) As of the Galaxy Tab S6, Wi-Fi only devices are no longer region locked, and the Call and Text on Other Devices features on those devices can work on SIM cards from most regions (excluding mainland China and Indonesia). All regions (except for mainland China and Indonesia) share the same CSC in their firmware version. Starting with Windows Phone 7 for mobile devices and Windows 8 for computers, not all display languages are preinstalled/available for download on all devices with an OEM license of Windows. Users may not see all the display languages listed as options on the device or as options available for download as separate language packs. These exact options depend on the device manufacturer and country/region of purchase (and the mobile operator, if the device features cellular connectivity). Microsoft believes region locking is
|
{"page_id": 307229, "title": "Regional lockout"}
|
The Mobile Servicing System (MSS) is a robotic system on board the International Space Station (ISS). Launched to the ISS in 2001, it plays a key role in station assembly and maintenance; it moves equipment and supplies around the station, supports astronauts working in space, services instruments and other payloads attached to the ISS, and is used for external maintenance. Astronauts receive specialized training to perform these functions with the various systems of the MSS. The MSS is composed of three components: the Space Station Remote Manipulator System (SSRMS), known as Canadarm2. the Mobile Remote Servicer Base System (MBS). the Special Purpose Dexterous Manipulator (SPDM, also known as "Dextre" or "Canada hand"). The system can move along rails on the Integrated Truss Structure on top of the US-provided Mobile Transporter cart, which hosts the MRS Base System. The system's control software was written in the Ada 95 programming language. The MSS was designed and manufactured by MDA (previously divisions of MacDonald Dettwiler Associates called MDA Space Missions, MD Robotics, and previously called SPAR Aerospace) for the Canadian Space Agency's contribution to the International Space Station. == Canadarm2 == Officially known as the Space Station Remote Manipulator System (SSRMS), Canadarm2 was launched on STS-100 in April 2001. This second generation arm is a larger, more advanced version of the Space Shuttle's original Canadarm. Canadarm2 is 17.6 m (58 ft) when fully extended and has seven motorized joints (an 'elbow' hinge in the middle, and three rotary joints at each of the 'wrist/shoulder' ends). It has a mass of 1,800 kg (4,000 lb), a diameter of 35 cm (14 in), and is made from titanium. The arm can handle large payloads of up to 116,000 kg (256,000 lb) and could assist with docking the space shuttle. It is self-relocatable and can move
|
{"page_id": 339639, "title": "Mobile Servicing System"}
|
copyrights because Accolade's games contained Sega's material. Accolade insisted that their use of Sega's material constituted fair use. However, Judge Caulfield did not accept this explanation since Accolade was a game manufacturer, their works were for financial gain, and because their works competed directly with Sega's licensed games, likely resulting in a sales decrease for Sega's games.: 384 Accolade's case was further hurt by a presentation by a Sega engineer named Takeshi Nagashima, who showed two Sega game cartridges that were able to run on the Genesis III without the trademark-displaying TMSS, and offered them to Accolade's defense team but would not reveal how that was possible.: 385 Ultimately, this would result in Accolade's defeat on April 3, 1992, when Judge Caulfield ruled in favor of Sega and issued an injunction prohibiting future sales by Accolade of Genesis-compatible games incorporating the Sega message or using the results of the reverse engineering. Almost a week later, Accolade was also required by the court to recall all of their Genesis-compatible games.: 386 == Appeal == The decision in the district court ruling had been very costly to Accolade. According to Accolade co-founder Alan Miller, "Just to fight the injunction, we had to pay at least half a million dollars in legal fees.": 386 On April 14, 1992, Accolade asked the district court to stay the preliminary injunction pending appeal, but when the court did not rule by April 21, Accolade appealed the injunction to the Ninth Circuit of the U.S. Court of Appeals. A stay was granted on the mandate to recall all of Accolade's Genesis games, but the injunction preventing further reverse engineering and development of Genesis software was maintained until August 28, when the Ninth Circuit ordered it dissolved pending the appeal review. In support of the appeal, the Computer
|
{"page_id": 24495502, "title": "Sega v. Accolade"}
|
our limited knowledge about chance, the wide space of the possible. Over time, the question 'what is chance' has become more and more important to me." (In German: "Zufall bedeutet für mich eine Befreiung subjektiver und emotionaler Elemente zugunsten objektiver Prozesse. Zufall ist nicht Willkür. Er unterliegt seinen Bedingungen und dazu gehört, dass ich sein Resultat anerkenne. Ähnlich einem wissenschaftlichen Verfahren versuche ich Antworten auf unser begrenztes Wissen über den Zufall zu finden, den weiten Raum des Möglichen. Im Laufe der Zeit wurde die Frage ‚Was ist Zufall‘ immer wichtiger für mich.") Since 2002, Hornef has devoted herself exclusively to researching this phenomenon. An expression of this is her series of works Alea iacta est, which she constantly expands, using an ordinary dice. Not only paintings, but also reliefs and three-dimensional objects are created. For the spatial installation Mikado at Skulpturen im Park in Mörfelden, she also used the dice as a random number generator: the 20 round logs, each three meters long, were divided into six segments, and Hornef let the cube decide which of them would be painted black. In her other works, she begins with an MDF board, which she paints black or white, and uses drilled holes, toothpicks, and Wooden dowels to structure the works. In a kind of controlled chance Hornef determines the basic rules of the work, such as the dimensions of the images, the lines (horizontal, vertical, or diagonal), and also the criteria for the elements to be thrown. The artist mostly works with square panels on which she draws a square grid with a pencil that remains visible. It is noteworthy that the artist herself indirectly provides the technique for decoding the works by addressing chance as the medium of her work. Albrecht Beutelspacher explains the method of working of Ingrid Hornef
|
{"page_id": 73295838, "title": "Ingrid Hornef"}
|
Greibach's theorem is independent of a particular approach to describe a formal language. It just considers a set C of formal languages over an alphabet Σ∪{#} such that each language in C has a finite description, each regular language over Σ∪{#} is in C, given descriptions of languages L1, L2 ∈ C and of a regular language R ∈ C, a description of the products L1R and RL1, and of the union L1∪L2 can be effectively computed, and it is undecidable for any member language L ∈ C with L ⊆ Σ* whether L = Σ*. Let P be any nontrivial subset of C that contains all regular sets over Σ∪{#} and is closed under quotient by each single symbol in Σ∪{#}. Then the question whether L ∈ P for a given description of a language L ∈ C is undecidable. === Proof === Let M ⊆ Σ*, such that M ∈ C, but M ∉ P. For any L ∈ C with L ⊆ Σ*, define φ(L) = (M#Σ*) ∪ (Σ*#L). From a description of L, a description of φ(L) can be effectively computed. Then L = Σ* if and only if φ(L) ∈ P: If L = Σ*, then φ(L) = Σ*#Σ* is a regular language, and hence in P. Else, some w ∈ Σ* \ L exists, and the quotient φ(L)/(#w) equals M. Therefore, by repeated application of the quotient-closure property, φ(L) ∈ P would imply M = φ(L)/(#w) ∈ P, contradicting the definition of M. Hence, if membership in P would be decidable for φ(L) from its description, so would be L’s equality to Σ* from its description, which contradicts the definition of C. == Applications == Using Greibach's theorem, it can be shown that the following problems are undecidable: Given a context-free grammar, does it
|
{"page_id": 42138836, "title": "Greibach's theorem"}
|
Webb's First Deep Field is the first operational image taken by the James Webb Space Telescope (JWST). The deep-field photograph, which covers a tiny area of sky visible from the Southern Hemisphere, is centered on SMACS 0723, a galaxy cluster in the constellation of Volans. Thousands of galaxies are visible in the image, some as old as 13 billion years. It is the highest-resolution image of the early universe ever taken. Captured by the telescope's Near-Infrared Camera (NIRCam), the image was revealed to the public by NASA on 11 July 2022. == Background == The James Webb Space Telescope is a space telescope operated by NASA and designed primarily to conduct infrared astronomy. Launched in December 2021, the spacecraft has been in a halo orbit around the second Sun–Earth Lagrange point (L2), about 1.5 million kilometers (900,000 mi) from Earth, since January 2022. At L2, the gravitational pull of the Sun combines with the gravitational pull of the Earth to produce an orbital period that matches Earth's, and the Earth and Sun remain co-aligned (as seen from that point) as the Earth and the spacecraft orbit the Sun together. Webb's First Deep Field was taken by the telescope's Near-Infrared Camera (NIRCam) and is a composite produced from images at different wavelengths, totalling 12.5 hours of exposure time. SMACS 0723 is a galaxy cluster visible from Earth's Southern Hemisphere, and has often been examined by Hubble and other telescopes in search of the deep past. == Scientific results == The image shows the galaxy cluster SMACS 0723 as it appeared 4.6 billion years ago, covering an area of sky with an angular size of 2.4 arcminutes, approximately equal to a grain of sand held at arm's length. Many of the objects in the image have undergone notable redshift due to the
|
{"page_id": 71275203, "title": "Webb's First Deep Field"}
|
to perform an unauthorized process that will have adverse impacts on the confidentiality, integrity, or availability of a system. A virus, worm, Trojan horse, or other code-based entity that infects a host. Spyware and some forms of adware are also examples of malicious code. managed interface An interface within a system that provides boundary protection capabilities using automated mechanisms or devices. mandatory access control An access control policy that is uniformly enforced across all subjects and objects within a system. A subject that has been granted access to information is constrained from: passing the information to unauthorized subjects or objects; granting its privileges to other subjects; changing one or more security attributes on subjects, objects, the system, or system components; choosing the security attributes to be associated with newly created or modified objects; or changing the rules for governing access control. Organization-defined subjects may explicitly be granted organization-defined privileges (i.e., they are trusted subjects) such that they are not limited by some or all of the above constraints. Mandatory access control is considered a type of nondiscretionary access control. marking See security marking . matching agreement [OMB A-108 ] A written agreement between a recipient agency and a source agency (or a non-Federal agency) that is required by the Privacy Act for parties engaging in a matching program. media [FIPS 200 ] Physical devices or writing surfaces including magnetic tapes, optical disks, magnetic disks, Large-Scale Integration memory chips, and printouts (but excluding display media) onto which information is recorded, stored, or printed within a system. metadata Information that describes the characteristics of data, including structural metadata that describes data structures (i.e., data format, syntax, semantics) and descriptive metadata that describes data contents (i.e., security labels). NIST SP 800-53, REV . 5 SECURITY AND P RIVACY C ONTROLS FOR I NFORMATION
|
{"source": 985, "title": "from dpo"}
|
reasons for this include that, in some cases, the current RE workflow outputs a number of incomplete statements leading to a larger number of extracted statements. Trained BELIEF curators use those incomplete statements to assemble correct statements and mark them for export. Unmarked statements are ignored by the system in the later export phase. In the BioCreative V IAT, some participants did not follow the guidelines correctly and did not only select statements for export, but removed all false BEL statements from the system. This increased the curation time and reduced the time-on-task performance for assisted curation. As mentioned above, we plan to improve the RE in order to reduce the overall number of statements. The evaluation environment developed in the course of the BEL track will help us in further optimizing the text mining output. The integrated BEL validator component was used to validate the added or modified BEL statements with the context annotations in the curation interface. This component generates syntax and semantic error messages for invalid BEL statements. As a result, BEL documents exported by the system are always syntactically correct. In the feedback from the survey, the biocurators criticized the complexity of these messages. It is most likely that the performance of curation could improve if these messages were more instructive. In future versions, an additional help page introducing various categories of error and instructions on fixing the invalid statements will hopefully lead to a better understanding of the messages. Nevertheless, to correct invalid BEL statements, use of the BEL framework is even more time-consuming than using BELIEF. In the presented evaluation, we only measured the time OpenBEL experts needed to correct the syntactically invalid statements. Without experience, the correction time would be much longer. Therefore, from the perspective of usability, it makes perfect sense to
|
{"source": 2716, "title": "from dpo"}
|
In TDS Archive]( , can we actually determine a solution, and if so, how? Note that if we find a solution of the given equation, we have at the same time answered the question of the existence of a solution. However, without knowledge of existence theory we might, for example, use a computer to find a numerical approximation to a “solution” that does not exist. On the other hand, even though we may know that a solution exists, it may be that the solution is not expressible in terms of the usual elementary functions—polynomial, trigonometric, exponential, logarithmic, and hyperbolic functions. Unfortunately, this is the situation for most differential equations. Thus, we discuss both elemen-tary methods that can be used to obtain exact solutions of certain relatively simple problems, and also methods of a more general nature that can be used to find ap-proximations to solutions of more difficult problems. Computer Use in Differential Equations. A computer can be an extremely valuable tool in the study of differential equations. For many years computers have been used to execute numerical algorithms, such as those described in Chapter 8, to construct numerical approximations to solutions of differential equations. These algorithms have been refined to an extremely high level of generality and efficiency. A few lines of computer code, written in a high-level programming language and executed (often within a few seconds) on a relatively inexpensive computer,
|
{"source": 6286, "title": "from dpo"}
|
on sales of preparative ultracentrifuge models, which were becoming popular as workhorses in biomedical laboratories. In 1949, Spinco introduced the Model L, the first preparative ultracentrifuge to reach a maximum speed of 40,000 rpm. In 1954, Beckman Instruments (later Beckman Coulter) purchased the company, forming the basis of its Spinco centrifuge division. == Instrumentation == Ultracentrifuges are available with a wide variety of rotors suitable for a great range of experiments. Most rotors are designed to hold tubes that contain the samples. Swinging bucket rotors allow the tubes to hang on hinges so the tubes reorient to the horizontal as the rotor initially accelerate. Fixed angle rotors are made of a single block of material and hold the tubes in cavities bored at a predetermined angle. Zonal rotors are designed to contain a large volume of sample in a single central cavity rather than in tubes. Some zonal rotors are capable of dynamic loading and unloading of samples while the rotor is spinning at high speed. Preparative rotors are used in biology for pelleting of fine particulate fractions, such as cellular organelles (mitochondria, microsomes, ribosomes) and viruses. They can also be used for gradient separations, in which the tubes are filled from top to bottom with an increasing concentration of a dense substance in solution. Sucrose gradients are typically used for separation of cellular organelles. Gradients of caesium salts are used for separation of nucleic acids. After the sample has spun at high speed for sufficient time to produce the separation, the rotor is allowed to come to a smooth stop and the gradient is gently pumped out of each tube to isolate the separated components. == Hazards == The tremendous rotational kinetic energy of the rotor in an operating ultracentrifuge makes the catastrophic failure of a spinning rotor a
|
{"page_id": 106291, "title": "Ultracentrifuge"}
|
the bulk phase theories and proposed his association-induction hypothesis of living cells. == See also == Cell adhesion Cytoskeleton Cell biology Cellular differentiation Germ theory of disease Membrane models == References == == Bibliography == Tavassoli, M. (1980). "The cell theory: a foundation to the edifice of biology". American Journal of Pathology. 98 (1): 44. PMC 1903404. PMID 6985772. Turner, W. (January 1890). "The Cell Theory Past and Present". Journal of Anatomy and Physiology. 24 (Pt 2): 253–87. PMC 1328050. PMID 17231856. Wolfe, Stephen L. (1972). Biology of the cell. Wadsworth Pub. Co. ISBN 978-0-534-00106-3. == External links == Mallery, C. (2008-02-11). "Cell Theory". Archived from the original on 2018-12-25. Retrieved 2008-11-25. "Studying Cells Tutorial". 2004. Retrieved 2008-11-25.
|
{"page_id": 102858, "title": "Cell theory"}
|
greater variation in length, but the denaturant gradient uses a second element (of melting point) to further distinguish between the samples. The DGGE gel will separate genes of the same size based on base sequence. This technique shows to what extent microbial communities are the same or different in taxonomic composition. Each band in a different location on the gel represents a different phylotype (one unique sequence of a phylogenetic marker gene). For microbial communities this method profiles many individual 16S rRNA sequences. The number of bands at differing horizontal positions can be used to estimate the level of biodiversity in that sample and infer phylogenetic affiliation. In order to know more about phylogenetic affiliation, one could excise those bands from the gel and then sequence them. ==== Advantages and disadvantages ==== The use of denaturing profile serves as a way to separate DNA fragments of similar sizes. This is beneficial in assessing microbial diversity due to the fact that the 16S rRNA gene does not vary much in size across bacterial phyla. The DGGE gel provides a quick way of looking at biodiversity in a microbial sample and does not preclude the option of sequencing the bands of interest. This method does not require that the microbes be cultured in the lab and does not require any sequence data needed to design probes for hybridization methods. The main disadvantage is that this is a qualitative assessment of biodiversity and one must sequence the genes in order to make inferences about the phylogenetic relatedness. Another disadvantage is that the GC clamp can be variable each time it is synthesized. This leads to the potential for different DGGE profiles for the same 16S rRNA sequence. ==== Applications ==== Stephen et al. utilized DGGE for a rapid analysis of Proteobacteria in soil.
|
{"page_id": 35256679, "title": "Community fingerprinting"}
|
A more concise statement can be made in terms of index sets: The only decidable index sets are ∅ and N {\displaystyle \mathbb {N} } . == Examples == Given a program P which takes a natural number n and returns a natural number P(n), the following questions are undecidable: Does P terminate on a given n? (This is the halting problem.) Does P terminate on 0? Does P terminate on all n (i.e., is P total)? Does P terminate and return 0 on every input? Does P terminate and return 0 on some input? Does P terminate and return the same value for all inputs? Is P equivalent to a given program Q? == Proof by Kleene's recursion theorem == Assume for contradiction that P {\displaystyle P} is a non-trivial, extensional and computable set of natural numbers. There is a natural number a ∈ P {\displaystyle a\in P} and a natural number b ∉ P {\displaystyle b\notin P} . Define a function Q {\displaystyle Q} by Q e ( x ) = φ b ( x ) {\displaystyle Q_{e}(x)=\varphi _{b}(x)} when e ∈ P {\displaystyle e\in P} and Q e ( x ) = φ a ( x ) {\displaystyle Q_{e}(x)=\varphi _{a}(x)} when e ∉ P {\displaystyle e\notin P} . By Kleene's recursion theorem, there exists e {\displaystyle e} such that φ e = Q e {\displaystyle \varphi _{e}=Q_{e}} . Then, if e ∈ P {\displaystyle e\in P} , we have φ e = φ b {\displaystyle \varphi _{e}=\varphi _{b}} , contradicting the extensionality of P {\displaystyle P} since b ∉ P {\displaystyle b\notin P} , and conversely, if e ∉ P {\displaystyle e\notin P} , we have φ e = φ a {\displaystyle \varphi _{e}=\varphi _{a}} , which again contradicts extensionality since a ∈ P {\displaystyle a\in P}
|
{"page_id": 25852, "title": "Rice's theorem"}
|
connectivity. TAPI 3.1 exposes COM interfaces. H.323v2 based IP telephony and IP multicast AV conferencing Telephony Service Providers are included. TAPI 3.1 also includes File Terminals (record streaming data), Pluggable Terminals (add external terminal object), USB/HID Phone TSP (control a USB phone and use it as a streaming endpoint) and support for Auto Discovery of TAPI Servers. Several H.323 supplementary services have been implemented for richer call control features (Call Transfer, Call Hold, Call Diversion, Call Park and Pickup). Windows Messenger and RTC (Real-Time Communication) Client API to provide IM, presence, AV communications, whiteboarding, application sharing, Acoustic Echo Cancellation, media encryption, PC to phone and phone to PC services to applications. For computers in a workgroup, the Windows Time Service in Windows XP supports a new Internet Time feature (NTP client), which updates the clock on the user's computer by synchronizing with an NTP time server on the Internet. This feature is useful for computers whose real-time clock does not maintain the correct time. Microsoft Message Queuing 3.0 supports: Internet Messaging (referencing queues via HTTP, SOAP-formatted messages, MSMQ support for Internet Information Services), queue aliases, multicasting of messages, and additional support for programmatic maintenance and administration of queues and MSMQ itself. MSMQ 3 clients directly communicate with Active Directory using LDAP. == Other features == Internet Information Services 5.1 COM+ 1.5 Speech Application Programming Interface 5.1 SAPI 5 support in Microsoft Narrator Paint is based on GDI+ and therefore, images can be natively saved as JPEG, GIF, TIFF and PNG without requiring additional graphics filters (in addition to BMP). However, alpha channel transparency is still not supported because the GDI+ version of Paint can only handle up to 24-bit depth images. Support for acquiring images from a scanner or a digital camera was also added to Paint. WordPad has full
|
{"page_id": 3463130, "title": "Features new to Windows XP"}
|
V. S. Sobolev Institute of Geology and Mineralogy of the Siberian Branch of the RAS, IGM SB RAS (Russian: Институт геологии и минералогии имени В. С. Соболева СО РАН, ИГМ СО РАН) is a research institute in Novosibirsk, Russia. It was founded in 2006. == Activity == Since 2008, the IGM SB RAS, togetger with the Institute of Cytology and Genetics, was engaged in the study of the ecosystem of Lake Solyonoye located in Bagansky District of Novosibirsk Oblast. In 2019, a number of media sources reported that the Sobolev Institute of Geology and Mineralogy, together with the University of Arizona, determined the age of the Aral Sea. == References ==
|
{"page_id": 71643718, "title": "V. S. Sobolev Institute of Geology and Mineralogy"}
|
chooses to use a simpler version of the object (e.g. with fewer polygons or textures).: 852 Light probe Object used to capture light parameters at a specific point in space in order to help compute scene lighting. Low level rendering API A library providing a minimal abstraction layer over a graphics processing unit's raw command lists, such as Vulkan, LibGCM, or Metal (API). The user typically has more control over (and responsibility for) resource management, command buffers, synchronization issues. Lighting Computations simulating the behavior of light. Light vector In shading calculations, a 3d unit vector representing the direction of incident light onto a model's surface. Light field A data structure approximating the 4D flux of light rays through a space (or in the general case, 5D); it may be captured using multiple cameras (e.g. light stage), or rendered from a 3D model by ray tracing. Line primitive A rendering primitive or modelling primitive representing a line segment, used for wireframes. Lumels A term for texels in the texture map representing a lightmap. == M == Manhattan distance Measure of distance between two points, different from Euclidean distance, that sums the distances along principal axes. Marching cubes A method for triangulating implicit surfaces. MegaTexturing Texturing technique that works with extremely large textures which are not loaded into memory all at once, but rather streamed from the hard disk depending on the camera view.: 176 Microtexture An alternative term sometimes used for Detail textures. Mipmap Method of preventing aliasing by storing differently scaled versions of the same image and using the correct one during rendering. Modelling primitive Basic elements from which 3D models and 3D scenes are composed. Also known as a Geometric primitive. Model space Coordinate system in which a 3D model is created and stored. Model transformation matrix A transformation
|
{"page_id": 50724448, "title": "Glossary of computer graphics"}
|
a manifold. Note that both super Minkowski spaces and super vector spaces can be taken as special cases of supermanifolds. A fourth, and completely unrelated meaning saw a brief usage in general relativity; this is discussed in greater detail at the bottom. == Examples == Several examples are given below. The first few assume a definition of superspace as a super vector space. This is denoted as Rm|n, the Z2-graded vector space with Rm as the even subspace and Rn as the odd subspace. The same definition applies to Cm|n. The four-dimensional examples take superspace to be super Minkowski space. Although similar to a vector space, this has many important differences: First of all, it is an affine space, having no special point denoting the origin. Next, the fermionic coordinates are taken to be anti-commuting Weyl spinors from the Clifford algebra, rather than being Grassmann numbers. The difference here is that the Clifford algebra has a considerably richer and more subtle structure than the Grassmann numbers. So, the Grassmann numbers are elements of the exterior algebra, and the Clifford algebra has an isomorphism to the exterior algebra, but its relation to the orthogonal group and the spin group, used to construct the spin representations, give it a deep geometric significance. (For example, the spin groups form a normal part of the study of Riemannian geometry, quite outside the ordinary bounds and concerns of physics.) === Trivial examples === The smallest superspace is a point which contains neither bosonic nor fermionic directions. Other trivial examples include the n-dimensional real plane Rn, which is a vector space extending in n real, bosonic directions and no fermionic directions. The vector space R0|n, which is the n-dimensional real Grassmann algebra. The space R1|1 of one even and one odd direction is known as the
|
{"page_id": 432630, "title": "Superspace"}
|
white dwarf (175 Myr), so they must have been deposited fairly recently, as much as probably only 1–2 million years ago. It was suggested that this was evidence for a disintegrating rocky minor planet orbiting around WD 1145+047 with a low mass of one-tenth that of Ceres, comparable to the mass of some of the large asteroids in the Solar System. The discovery was then published in the online journal Nature on 22 October 2015, describing the nature of the system. == See also == Disrupted planet GD 66 – white dwarf assumed to have an orbiting giant planet, but later ruled out Pulsar planet – type of planet that orbits another stellar remnant, a pulsar == References == == External links == NASA – Kepler Mission. NASA – Kepler Discoveries – Summary Table. NASA – WD 1145+017 b at The Extrasolar Planets Encyclopaedia.
|
{"page_id": 51326627, "title": "WD 1145+017 b"}
|
ILIOS is an acronym of InterLink Internet Operating System. It is an attempt to create a router-only operating system; one specifically oriented towards computer networking purposes, especially routing. It supports IPv4 routing and is a good educational OS, though it is single tasking and does everything via interrupts. It is released under the BSD License. The author of this research OS is Rink Springer, who is also responsible for porting FreeBSD to the Xbox. == External links == ILIOS - Trac Rink Springer's website
|
{"page_id": 2869030, "title": "ILIOS"}
|
representation learning and generative modeling. In this paper, we look at geometric data represented as point clouds. We introduce a deep AutoEncoder (AE) network with state-of-the-art reconstruction quality and generalization ability. The learned representations outperform existing methods on 3D recognition tasks and enable shape editing via simple algebraic manipulations, such as semantic part editing, shape analogies and shape interpolation, as well as shape completion. We perform a thorough study of different generative models including GANs operating on the raw point clouds, significantly improved GANs trained in the fixed latent space of our AEs, and Gaussian Mixture Models (GMMs). To quantitatively evaluate generative models we introduce measures of sample fidelity and diversity based on matchings between sets of point clouds. Interestingly, our evaluation of generalization, fidelity and diversity reveals that GMMs trained in the latent space of our AEs yield the best results overall.} } @InProceedings{adel2018a, title = {Discovering Interpretable Representations for Both Deep Generative and Discriminative Models}, author = {Tameem Adel and Zoubin Ghahramani and Adrian Weller}, booktitle = {Proceedings of the 35th International Conference on Machine Learning}, pages = {50--59}, year = {2018}, editor = {Jennifer Dy and Andreas Krause}, volume = {80}, series = {Proceedings of Machine Learning Research}, address = {Stockholm, Sweden}, month = {10--15 July}, publisher = {PMLR}, pdf = { url = { abstract = {Interpretability of representations in both deep generative and discriminative models is highly desirable. Current methods jointly optimize an objective combining accuracy and interpretability. However, this may reduce accuracy, and is not applicable to already trained models. We propose two interpretability frameworks. First, we provide an interpretable lens for an existing model. We use a generative model which takes as input the representation in an existing (generative or discriminative) model, weakly supervised by limited side information. Applying a flexible and invertible
|
{"source": 1216, "title": "from dpo"}
|
to which some function has been assigned? But let’s go back again to the same examples; is there some function of eyes, according to us? There is. And in that case, is there also an excellence of eyes? An excellence too. What about ears? Do they have a function? Yes. Haven’t they an excellence too? An excellence too. And in all other cases, doesn’t the same apply? It does. In that case could the eyes ever perform **353C**their own function properly, without possessing their own particular excellence, but possessing a defect instead of the excellence? Well, how could they, he replied, since, presumably you mean that they possess blindness instead of sight? I am not yet asking what their excellence might be, but whether anything that exercises a function carries out its own function well, by means of its own excellence, and badly, by means of its defect. Well what you are saying is true, in this case, said he. Won’t the ears carry out their own function badly when deprived of their own excellence? Entirely so. **353D** So, do we apply the same argument to all the other instances? Well, I think so. Come on then, let’s proceed to consider the following; is there some function of the soul that you could not perform by means of anything else there is? For instance, caring, ruling and deliberating, and everything of that sort; is there anything else besides soul to which we might properly attribute these? Could we maintain that they are not particular to soul? They are particular to nothing else. And again, what about living? Shall we say that it is a function of soul? More than anything, said he. Don’t we also say that there is some excellence of a soul? We say so. **353E** In that
|
{"source": 3689, "title": "from dpo"}
|
onto, show that g is one-to-one. (e) If g ◦ f is onto and g is one-to-one, show that f is onto. 23. Define a function on the real numbers by f (x) = x + 1 x − 1 . What are the domain and range of f ? What is the inverse of f ? Compute f ◦ f −1 and f −1 ◦ f .24. Let f : X → Y be a map with A1, A 2 ⊂ X and B1, B 2 ⊂ Y .(a) Prove f (A1 ∪ A2) = f (A1) ∪ f (A2). (b) Prove f (A1 ∩ A2) ⊂ f (A1) ∩ f (A2). Give an example in which equality fails. (c) Prove f −1(B1 ∪ B2) = f −1(B1) ∪ f −1(B2), where f −1(B) = {x ∈ X : f (x) ∈ B}. (d) Prove f −1(B1 ∩ B2) = f −1(B1) ∩ f −1(B2). (e) Prove f −1(Y \ B1) = X \ f −1(B1). 25. Determine whether or not the following relations are equivalence relations on the given set. If the relation is an equivalence relation, describe the partition given by it. If the relation is not an equivalence relation, state why it fails to be one. (a) x ∼ y in R if x ≥ y (b) m ∼ n in Z if mn > 0(c) x ∼ y in R if |x − y| ≤ 4(d) m ∼ n in Z if m ≡ n (mod 6) 26. Define a relation ∼ on R2 by stating that ( a, b ) ∼ (c, d ) if and only if a2 +b2 ≤ c2 + d2. Show that ∼ is reflexive and transitive but not symmetric. 27. Show that an m × n matrix gives rise
|
{"source": 5734, "title": "from dpo"}
|
does not write on Aristotle and friendship, then Sorabji works on Aristotle’s commentators, and Sim compares Aristotle and Confucius if Broadie does not look at Aristotle’s ethics in contemporary work. It is not the case that Annas wrote on Aristotle and friendship and Broadie looks at Aristotle’s ethics in contemporary work. Sorabji does not work on Aristotle’s commentators. Sim compares Aristotle and Confucius just in case either Annas wrote on Aristotle and friendship or Sorabji works on Aristotle’s commentators. If Sorabji works on Aristotle’s commentators, then Broadie looks at Aristotle’s ethics in contemporary work, if, and only if, Sim compares Aristotle and Confucius. Hence, Broadie not looking at Aristotle’s ethics in contemporary work is necessary and sufficient for Sim comparing Aristotle and Confucius. F: Foot developed the trolley problem. T: Thomson introduced the fat man scenario. K: Kamm presents the looping trolley case. 11. Foot developed the trolley problem. Thomson introduced the fat man scenario. Kamm presents the looping trolley case, if Foot developed the trolley problem and Thomson introduced the fat man scenario. Therefore, Kamm presents the looping trolley case; however, Thomson introduced the fat man scenario. 12. Foot developing the trolley problem is sufficient for Thomson introducing the fat man scenario. Kamm presents the looping trolley case if Thomson introduced the fat man scenario. Foot developed the trolley problem. So, Kamm presents the looping trolley case. 13. Either Foot developed the trolley problem or Thomson introduced the fat man scenario. Thomson introduces the fat man scenario unless Kamm does not present the looping trolley case. Foot developing the trolley problem is necessary and sufficient for Kamm presenting the looping trolley case. Either Kamm presents the looping trolley case or Foot developed the trolley problem, given that Thomson introduced the fat man scenario. So, Thomson introduced the fat man
|
{"source": 6816, "title": "from dpo"}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.