text
stringlengths
5
1.02k
source
stringclasses
1 value
"Software transactional memory-Composable operations-\u2014Tim Harris et al., \"Composable Memory Transactions\", Section 2: Background, pg.2 With STM, this problem is simple to solve: simply wrapping two operations in a transaction makes the combined operation atomic. The only sticking point is that it is unclear to the caller, who is unaware of the implementation details of the component methods, when it should attempt to re-execute the transaction if it fails. In response, the authors proposed a retry command which uses the transaction log generated by the failed transaction to determine which memory cells it read, and automatically retries the transaction when one of these cells is modified, based on the logic that the transaction will not behave differently until at least one such value is changed."
large_corpus_distillation
"Units of information-Units derived from bit-In the past, uppercase K has been used instead of lowercase k to indicate 1024 instead of 1000. However, this usage was never consistently applied."
large_corpus_distillation
"Support vector machine-Motivation-Suppose some given data points each belong to one of two classes, and the goal is to decide which class a new data point will be in. In the case of support vector machines, a data point is viewed as a p -dimensional vector (a list of p numbers), and we want to know whether we can separate such points with a (p\u22121) -dimensional hyperplane. This is called a linear classifier. There are many hyperplanes that might classify the data. One reasonable choice as the best hyperplane is the one that represents the largest separation, or margin, between the two classes. So we choose the hyperplane so that the distance from it to the nearest data point on each side is maximized. If such a hyperplane exists, it is known as the maximum-margin hyperplane and the linear classifier it defines is known as a maximum-margin classifier; or equivalently, the perceptron of optimal stability.More formally, a support vector machine constructs a hyperplane or set of hyperplanes in a high or
large_corpus_distillation
"Comparison of programming languages (strings)-String literals-Multiline string Many languages have a syntax specifically intended for strings with multiple lines. In some of these languages, this syntax is a here document or \"heredoc\": A token representing the string is put in the middle of a line of code, but the code continues after the starting token and the string's content doesn't appear until the next line. In other languages, the string's content starts immediately after the starting token and the code continues after the string literal's terminator."
large_corpus_distillation
"Molecular logic gate-Working principle-For logic gates with a single input, there are four possible output patterns. When the input is 0, the output can be either a 0 or 1. When the input is 1, the output can again be 0 or 1. The four output bit patterns that can arise corresponds to a specific logic type: PASS 0, YES, NOT, and PASS 1. PASS 0 always outputs 0, whatever the input. PASS 1 always outputs 1, whatever the input. YES outputs a 1 when the input is 1, and NOT is the inverse of YES \u2013 it outputs a 0 when the input is 1.AND, OR, XOR, NAND, NOR, XNOR, and INH are two-input logic gates. The AND, OR, and XOR gates are fundamental logic gates, and the NAND, NOR, and XNOR gates are complementary to AND, OR, and XOR gates, respectively. An INHIBIT (INH) gate is a special conditional logic gate that includes a prohibitory input. When the prohibitory input is absent, the output produced depends solely on the other input."
large_corpus_distillation
"2. From (a) and the conditioning rule, E(etXN)=E[E(etXN \u2223N)]=E[M N (t)]=\u2211\u221e n=1 p n M n (t)\n1\nIn the coin-die experiment, a biased coin is tossed with probability of heads 3. If the coin lands tails, a fair die is rolled; if the coin lands heads, an ace-six flat die is\n1 1\nrolled (faces 1 and 6 have probability 4 each, and faces 2, 3, 4, 5 have probability 8 each). Find the mean and standard deviation of the die score.\nAnswer\n7\n1. 2\n2. 1.8634\nRun the coin-die experiment 1000 times and note the apparent convergence of the empirical mean and standard deviation to the distribution mean and standard\ndeviation.\nThis page titled 4.7: Conditional Expected Value is shared under a CC BY 2.0 license and was authored, remixed, and/or curated by Kyle Siegrist (Random Services) via source content that\nwas edited to the style and standards of the LibreTexts platform.\n4.7.12 https://stats.libretexts.org/@go/page/10163"
large_corpus_distillation
"Graph partition-Other graph partition methods-Spin models have been used for clustering of multivariate data wherein similarities are translated into coupling strengths. The properties of ground state spin configuration can be directly interpreted as communities. Thus, a graph is partitioned to minimize the Hamiltonian of the partitioned graph. The Hamiltonian (H) is derived by assigning the following partition rewards and penalties."
large_corpus_distillation
"Modern portfolio theory-Mathematical model-Matrices are preferred for calculations of the efficient frontier.\nIn matrix form, for a given \"risk tolerance\" q\u2208[0,\u221e) , the efficient frontier is found by minimizing the following expression: wT\u03a3w\u2212q\u00d7RTw where w is a vector of portfolio weights and 1."
large_corpus_distillation
"Michal Rosen-Zvi-Research-This is an extension of latent Dirichlet allocation, a method for understanding the word frequencies of documents by fitting them to a model in which a document is associated with a mixture of topics, each of which is a probability distribution over words. The author-topic model was introduced in a paper by Rosen-Zvi and three other authors in 2004. It adds the identity of document authors and their preferences for topics to the model, and became \"one of the earliest attempts at modeling the interests of authors\" used in this area of machine learning.More recent research of Rosen-Zvi has focused on the application of machine learning techniques to problems in healthcare including medical diagnosis, choice of treatment, and causal inference in the observation of medical treatments and their outcomes."
large_corpus_distillation
"Inheritance (object-oriented programming)-Applications-In most quarters, class inheritance for the sole purpose of code reuse has fallen out of favor. The primary concern is that implementation inheritance does not provide any assurance of polymorphic substitutability\u2014an instance of the reusing class cannot necessarily be substituted for an instance of the inherited class. An alternative technique, explicit delegation, requires more programming effort, but avoids the substitutability issue. In C++ private inheritance can be used as a form of implementation inheritance without substitutability. Whereas public inheritance represents an \"is-a\" relationship and delegation represents a \"has-a\" relationship, private (and protected) inheritance can be thought of as an \"is implemented in terms of\" relationship.Another frequent use of inheritance is to guarantee that classes maintain a certain common interface; that is, they implement the same methods. The parent class can be a combination of implemented
large_corpus_distillation
"Idempotent (ring theory)-Examples-Quotient of polynomial ring Given a ring R and an element f\u2208R such that f2\u22600 , then the quotient ring R/(f2\u2212f) has the idempotent f . For example, this could be applied to x\u2208Z[x] , or any polynomial f\u2208k[x1,\u2026,xn] Idempotents in split-quaternion rings There is a catenoid of idempotents in the split-quaternion ring."
large_corpus_distillation
"User story-Limitations-Limitations of user stories include: Scale-up problem: User stories written on small physical cards are hard to maintain, difficult to scale to large projects and troublesome for geographically distributed teams.\nVague, informal and incomplete: User story cards are regarded as conversation starters. Being informal, they are open to many interpretations. Being brief, they do not state all of the details necessary to implement a feature. Stories are therefore inappropriate for reaching formal agreements or writing legal contracts.\nLack of non-functional requirements: User stories rarely include performance or non-functional requirement details, so non-functional tests (e.g. response time) may be overlooked."
large_corpus_distillation
"Determining the number of clusters in a data set-The gap statistics-Robert Tibshirani, Guenther Walther, and Trevor Hastie proposed estimating the number of clusters in a data set via the gap statistic.\nThe gap statistics, based on theoretical grounds, measures how far is the pooled within-cluster sum of squares around the cluster centers from the sum of squares expected under the null reference distribution of data. The expected value is estimated by simulating null reference data of characteristics of the original data, but lacking any clusters in it.\nThe optimal number of clusters is then estimated as the value of k for which the observed sum of squares falls farthest below the null reference.\nUnlike many previous methods, the gap statistics can tell us that there is no value of k for which there is a good clustering.\nThe gap statistics is implemented as the clusGap function in the cluster package in R."
large_corpus_distillation
nonresidue, or zero.\nFor example, from the table for modulus 15 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14 (residues in bold).\nThe product of the nonresidues 2 and 8 is the residue 1, whereas the product of the nonresidues 2 and 7 is the nonresidue 14."
large_corpus_distillation
"Examples of groups-Permutations of a set of three elements-By inspection, we can determine associativity and closure; note in particular that (ba)b = bab = b(ab).\nSince it is built up from the basic operations a and b, we say that the set {a,\u2009b} generates this group. The group, called the symmetric group S3, has order 6, and is non-abelian (since, for example, ab \u2260 ba)."
large_corpus_distillation
"Message Authenticator Algorithm-Formal specifications of the MAA-The MAA has been used as a prominent case study for assessing various formal methods."
large_corpus_distillation
"Classic monolingual word-sense disambiguation-Classic monolingual word-sense disambiguation-Classic monolingual Word Sense Disambiguation evaluation tasks uses WordNet as its sense inventory and is largely based on supervised / semi-supervised classification with the manually sense annotated corpora: Classic English WSD uses the Princeton WordNet as it sense inventory and the primary classification input is normally based on the SemCor corpus.\nClassical WSD for other languages uses their respective WordNet as sense inventories and sense annotated corpora tagged in their respective languages. Often researchers will also tapped on the SemCor corpus and aligned bitexts with English as its source language "
large_corpus_distillation
"Twincharger-Disadvantages-The main disadvantage of twincharging is the complexity and expense of components. Usually, to provide acceptable response, smoothness of power delivery, and adequate power gain over a single-compressor system, expensive electronic and/or mechanical controls must be used. In a spark-ignition engine, a low compression ratio must also be used if the supercharger produces high boost levels, negating some of the efficiency benefits of a lower-displacement engine."
large_corpus_distillation
check shows that, indeed, 49 mod 11 {\\textstyle 7^{2}\\equiv 49\\equiv 5{\\pmod {11}}} and 16 mod 11 {\\textstyle 4^{2}\\equiv 16\\equiv 5{\\pmod {11}}} "
large_corpus_distillation
"Critical graph-Critical graph-In graph theory, a critical graph is an undirected graph all of whose subgraphs have smaller chromatic number. In such a graph, every vertex or edge is a critical element, in the sense that its deletion would decrease the number of colors needed in a graph coloring of the given graph. The decrease in the number of colors cannot be by more than one."
large_corpus_distillation
"Managed private cloud-Advantages and challenges of managed private cloud-A managed private cloud cuts down on upkeep costs by outsourcing infrastructure management and maintenance to the managed cloud provider. It is easier to integrate an organization's existing software, services, and applications into a dedicated cloud hosting infrastructure which can be customized to the client's needs, instead of a public cloud platform, whose hardware or infrastructure/software platform cannot be individualized to each client.Customers who choose a managed private cloud deployment usually choose them because of their desire for an efficient cloud deployment, but also have the need for service customization or integration only available in a single-tenant environment."
large_corpus_distillation
"CIPHERUNICORN-E-CIPHERUNICORN-E-In cryptography, CIPHERUNICORN-E is a block cipher created by NEC in 1998. It was among the cryptographic techniques recommended for Japanese government use by CRYPTREC in 2003. However, it has been dropped to \"candidate\" level by the CRYPTREC revision of 2013.\nThe algorithm has a 16-round modified Feistel network structure, with an additional key-dependent mixing function after every 2 rounds. The block size is 64 bits, and the key size 128 bits. The round function is fairly complicated, split into two nearly parallel computations."
large_corpus_distillation
"Information gain ratio-Information gain ratio-In decision tree learning, Information gain ratio is a ratio of information gain to the intrinsic information. It was proposed by Ross Quinlan, to reduce a bias towards multi-valued attributes by taking the number and size of branches into account when choosing an attribute.Information Gain is also known as Mutual Information."
large_corpus_distillation
"PL/C-Design and implementation-The implementation of the PL/C compiler itself was done in IBM 360 Assembly Language, with a heavy use of assembly language macros. It was done as a three-pass compiler, with the first pass doing syntax analysis and error correction of syntactical mistakes, the second pass doing semantic analysis and further error correction, and the third pass doing code generation. PL/C needed to avoid the longstanding problem of \"cascading diagnostics\", wherein one error is internally but often mistakenly \"fixed\" by the compiler, resulting in a secondary, tertiary and additional series of unhelpful error messages. The design of the intermediate language used between the passes and of the accompanying symbol table was tailored towards, and key to the achievement of, the error-repair and diagnostic reporting capabilities of PL/C. While a number of these techniques dated from the Cornell CORC and CUPL efforts, here they were being used for the first time on large programming language with a
large_corpus_distillation
"Singular value decomposition-Calculating the SVD-The approaches that use eigenvalue decompositions are based on the QR algorithm, which is well-developed to be stable and fast. Note that the singular values are real and right- and left- singular vectors are not required to form similarity transformations. One can iteratively alternate between the QR decomposition and the LQ decomposition to find the real diagonal Hermitian matrices. The QR decomposition gives M \u21d2 Q R and the LQ decomposition of R gives R \u21d2 L P\u204e. Thus, at every iteration, we have M \u21d2 Q L P\u204e, update M \u21d0 L and repeat the orthogonalizations. Eventually, this iteration between QR decomposition and LQ decomposition produces left- and right- unitary singular matrices. This approach cannot readily be accelerated, as the QR algorithm can with spectral shifts or deflation. This is because the shift method is not easily defined without using similarity transformations. However, this iterative approach is very simple to
large_corpus_distillation
"Run to completion scheduling-Run to completion scheduling-Run-to-completion scheduling or nonpreemptive scheduling is a scheduling model in which each task runs until it either finishes, or explicitly yields control back to the scheduler. Run to completion systems typically have an event queue which is serviced either in strict order of admission by an event loop, or by an admission scheduler which is capable of scheduling events out of order, based on other constraints such as deadlines."
large_corpus_distillation
"Schuette\u2013Nesbitt formula-Remarks-The probabilistic versions (4'), (5') and (6') can be generalized to every finite measure space.For textbook presentations of the probabilistic Schuette\u2013Nesbitt formula (6') and their applications to actuarial science, cf. Gerber (1997). Chapter 8, or Bowers et al. (1997), Chapter 18 and the Appendix, pp. 577\u2013578."
large_corpus_distillation
for the platform.\nThus, a user-agent may restrict file selection, as, for example, in the following, restricted to GIF and PNG images or any images: On a Windows platform, this might mean that the user agent would show files only of the types specified in the browse-file dialog."
large_corpus_distillation
"Region-based memory management-History and concepts-Region inference In 1988, researchers began investigating how to use regions for safe memory allocation by introducing the concept of region inference, where the creation and deallocation of regions, as well as the assignment of individual static allocation expressions to particular regions, is inserted by the compiler at compile-time. The compiler is able to do this in such a way that it can guarantee dangling pointers and leaks do not occur."
large_corpus_distillation
"IDEA NXT-IDEA NXT-In cryptography, the IDEA NXT algorithm (previously known as FOX) is a block cipher designed by Pascal Junod and Serge Vaudenay of EPFL (Lausanne, Switzerland). It was conceived between 2001 and 2003. The project was originally named FOX and was published in 2003. In May 2005, it was announced by MediaCrypt under the name IDEA NXT. IDEA NXT is the successor to the International Data Encryption Algorithm (IDEA) and also uses the Lai\u2013Massey scheme. MediaCrypt AG holds patents on elements of IDEA and IDEA NXT. The cipher is specified in two configurations: NXT64 (with block of 64 bits, key of 128 bits, 16 rounds) and NXT128 (with block of 128 bits, key of 256 bits, 16 rounds)."
large_corpus_distillation
"Cc x\u2208Cc y\u2208C x y\nWe will show that is open, so fix . For each , the points and can be separated, so there exist disjoint open\nU V x\u2208U y\u2208V {V :y\u2208C} C\nsets y and y such that y and y. Trivially, the collection y is an open cover of , and hence there\nB\u2286C {V :y\u2208B} C U =\u22c2 U \u22c3 V\nexist a finite subset such that y covers . But then y\u2208B y is open and is disjoint from y\u2208B y.\nU C U x\u2208U \u2286Cc\nHence also is disjoint from . So to summarize, is open and .\nAlso in a Hausdorff space, a point can be separated from a compact set that does not contain the point.\n(S,S) x\u2208S C \u2286S x\u2209C U\nSuppose that is a Hausdorff space. If , is compact, and , then there exist disjoint open sets and\nV x\u2208U C \u2286V\nwith and\nProof\ny\u2208C U V x\u2208U y\u2208V\nSince the space is Hausdorff, for each there exist disjoint open sets y and y with y and y. The collection\n{V :y\u2208C} C B\u2282C {V :y\u2208B} C\ny is an open cover of , and hence there exists
large_corpus_distillation
"Lambda lifting-Lambda lifting in lambda calculus-lambda-named\u2061[X]\u2228lambda-anon\u2061[X]\u2192lift-choice\u2061[X]=X lift-choice\u2061[\u03bbF.X]=lift-choice\u2061[X] none \u2192lift-choice\u2061[MN]=lift-choice\u2061[M] lift-choice\u2061[MN]=lift-choice\u2061[N] none For example, Examples For example, the Y combinator, \u03bbf.(\u03bbx.f(xx))(\u03bbx.f(xx)) is lifted as, let in \u2061qx and after Parameter dropping, let in \u2061q As a lambda expression (see Conversion from let to lambda expressions), (\u03bbx.(\u03bbq.q)\u03bbf.f(xf)(xf))\u03bbf.\u03bby.f(yy) If lifting anonymous functions only, the Y combinator is, let in \u2061qp and after Parameter dropping, let in \u2061q As a lambda expression, (\u03bbp.(\u03bbq.q)\u03bbf.(pf)(pf))\u03bbf.\u03bbx.f(xx) The first sub expression to be chosen for lifting is \u03bbx.f(xx) . This transforms the lambda expression into \u03bbf.(pf)(pf) and creates the equation pfx=f(xx) The second sub expression to be chosen for lifting is \u03bbf.(pf)(pf) . This
large_corpus_distillation
"Predeclared-Pre-declared entity-A pre-declared entity is a built-in notation convention for a character or a string. For example, in the HTML markup language, a large number of character and numeric entities are available to represent characters. In HTML, '&lt;' is a possible pre-declared entity to represent '<'. The programmer must not declare this entity by himself before he can use it, since it is already pre-declared by the specifications of the HTML language. Pre-declared entities are often used as escape sequences to represent information that would otherwise cause possible conflicts in its non-encoded form."
large_corpus_distillation
"Quickhull-Quickhull-Quickhull is a method of computing the convex hull of a finite set of points in n-dimensional space. It uses a divide and conquer approach similar to that of quicksort, from which its name derives. Its worst case time complexity for 2-dimensional and 3-dimensional space is O(n2) , but when the input precision is restricted to log \u2061n) bits, its worst case time complexity is conjectured to be log \u2061r) , where n is the number of input points and r is the number of processed points (up to n ).N-dimensional Quickhull was invented in 1996 by C. Bradford Barber, David P. Dobkin, and Hannu Huhdanpaa. It was an extension of Jonathan Scott Greenfield's 1990 planar Quickhull algorithm, although the 1996 authors did not know of his methods. Instead, Barber et al. describe it as a deterministic variant of Clarkson and Shor's 1989 algorithm."
large_corpus_distillation
"BitFunnel-Algorithm-Initial problem and solution overview The BitFunnel paper describes the \"matching problem\", which occurs when an algorithm must identify documents through the usage of keywords. The goal of the problem is to identify a set of matches given a corpus to search and a query of keyword terms to match against. This problem is commonly solved through inverted indexes, where each searchable item is maintained with a map of keywords.In contrast, BitFunnel represents each searchable item through a signature. A signature is a sequence of bits which describe a Bloom filter of the searchable terms in a given searchable item. The bloom filter is constructed through hashing through several bit positions."
large_corpus_distillation
of estimators) distinct reference values \u03b2. The true distribution is then approximated by a linear regression, and the best estimators are obtained in closed form as where Y denotes the template matrix with the values of the known or previously determined model for any of the reference values \u03b2, d are the random variables (e.g. a measurement), and the matrix M~ and the vector m~ are calculated from the values of \u03b2. The LTF can also be expressed for Log-normal distribution distributed random variables. A generalization of the LTF is the Quadratic Template Fit, which assumes a second order regression of the model, requires predictions for at least n2+2n distinct values \u03b2, and it finds the best estimator using Newton's method.Percentage least squares focuses on reducing percentage errors, which is useful in the field of forecasting or time series analysis. It is also useful in situations where the dependent variable has a wide range without constant variance, as here the larger
large_corpus_distillation
method, and as its difficulty does not depend on the cipher but only on the key length - it's not considered a real cryptanalysis of the cipher. If the key has N bits, there are 2N possible keys to try, so a brute-force attack can recover the cipher in a worst-case time proportional to 2N and an average time of 2N-1. This is often used as a standard of comparison for other attacks. Brute-force can be applied in ciphertext-only settings, but the cryptanalyst must have enough information about the plaintext (at least N bits) to allow the identification of the correct key once it is tried."
large_corpus_distillation
"Object lifetime-Overview-In garbage-collected languages, objects are generally dynamically allocated (on the heap) even if they are initially bound to an automatic variable, unlike automatic variables with primitive values, which are typically automatically allocated (on the stack or in a register). This allows the object to be returned from a function (\"escape\") without being destroyed. However, in some cases a compiler optimization is possible, namely performing escape analysis and proving that escape is not possible, and thus the object can be allocated on the stack; this is significant in Java. In this case object destruction will occur promptly \u2013 possibly even during the variable's lifetime (before the end of its scope), if it is unreachable."
large_corpus_distillation
"One-way function-One-way function-The existence of such one-way functions is still an open conjecture. Their existence would prove that the complexity classes P and NP are not equal, thus resolving the foremost unsolved question of theoretical computer science.:\u200aex. 2.2,\u200apage 70\u200a The converse is not known to be true, i.e. the existence of a proof that P\u2260NP would not directly imply the existence of one-way functions.In applied contexts, the terms \"easy\" and \"hard\" are usually interpreted relative to some specific computing entity; typically \"cheap enough for the legitimate users\" and \"prohibitively expensive for any malicious agents\". One-way functions, in this sense, are fundamental tools for cryptography, personal identification, authentication, and other data security applications. While the existence of one-way functions in this sense is also an open conjecture, there are several candidates that have withstood decades of intense scrutiny. Some of them are essential ingredients
large_corpus_distillation
"Shell (computing)-Command-line shells-A feature of many command-line shells is the ability to save sequences of commands for re-use. A data file can contain sequences of commands which the CLI can be made to follow as if typed in by a user. Special features in the CLI may apply when it is carrying out these stored instructions. Such batch files (script files) can be used repeatedly to automate routine operations such as initializing a set of programs when a system is restarted. Batch mode use of shells usually involves structures, conditionals, variables, and other elements of programming languages; some have the bare essentials needed for such a purpose, others are very sophisticated programming languages in and of themselves. Conversely, some programming languages can be used interactively from an operating system shell or in a purpose-built program."
large_corpus_distillation
"65,536-In computing-65536 (216) is the number of different values representable in a number of 16 binary digits (or bits), also known as an unsigned short integer in many computer programming systems. A 65,536-bit integer can represent up to 265536 (2.00352993...\u00d71019,728) values."
large_corpus_distillation
"HTTP Public Key Pinning-Mechanism-At least one backup key must be pinned, in case the current pinned key needs to be replaced. The HPKP is not valid without this backup key (a backup key is defined as a public key not present in the current certificate chain).HPKP is standardized in RFC 7469. It expands on static certificate pinning, which hardcodes public key hashes of well-known websites or services within web browsers and applications.Most browsers disable pinning for certificate chains with private root certificates to enable various corporate content inspection scanners and web debugging tools (such as mitmproxy or Fiddler). The RFC 7469 standard recommends disabling pinning violation reports for \"user-defined\" root certificates, where it is \"acceptable\" for the browser to disable pin validation."
large_corpus_distillation
"Note-taking-History-Note-taking has been an important part of human history and scientific development. The Ancient Greeks developed hypomnema, personal records on important subjects. In the Renaissance and early modern period, students learned to take notes in schools, academies and universities, often producing beautiful volumes that served as reference works after they finished their studies. In pre-digital times, people used many kinds of notebooks, including commonplace books, accounting waste books, and marginalia. Philosopher John Locke developed and published a popular indexing system which served as a model for commonplace books and inspired at least ten different published editions of commonplace book templates in Europe and the Americas as well as Bell's Common-Place Book, Form'd Generally upon the Principles Recommended and Practised by Mr Locke (London, 1770)."
large_corpus_distillation
"Latin tenses-Perfect indicative-Irregular verbs: 'I give': ded\u012b 'I bear': tul\u012b 'I go': i\u012b (rarely \u012bv\u012b) 'I want': volu\u012b 'I am': fu\u012b 'I am able': potu\u012bPassive and deponent verbs 1st conjugation: am\u0101tus sum (rarely am\u0101tus fu\u012b) 2nd conjugation: v\u012bsus sum (v\u012bsus fu\u012b) 3rd conjugation (-\u014d): ductus sum (ductus fu\u012b) 3rd conjugation (-i\u014d): captus sum (captus fu\u012b) 4th conjugation: aud\u012btus sum (auditus fu\u012b)The forms with fu\u012b are much less common. These forms are discussed in a separate section below."
large_corpus_distillation
"RedSleeve-Releases-Versioning scheme RedSleeve aims to maintain equal versioning with the upstream distribution, both in terms of distribution release numbering and the individual package release numbering. The only exception is with the packages that had to be modified from the upstream release. This is only done to either remove upstream branding as required by the upstream distributions terms and conditions, or to apply additional patches required to make the package build and work on the ARM architecture which at the time of writing the upstream distribution does not support. In such cases, .0 is appended to the package version after the distribution tag."
large_corpus_distillation
"Capacitated arc routing problem-Capacitated arc routing problem-In mathematics, the capacitated arc routing problem (CARP) is that of finding the shortest tour with a minimum graph/travel distance of a mixed graph with undirected edges and directed arcs given capacity constraints for objects that move along the graph that represent snow-plowers, street sweeping machines, or winter gritters, or other real-world objects with capacity constraints. The constraint can be imposed for the length of time the vehicle is away from the central depot, or a total distance traveled, or a combination of the two with different weighting factors."
large_corpus_distillation
"Supersingular isogeny key exchange-Similar systems, signatures, and uses-A predecessor to the SIDH was published in 2006 by Rostovtsev and Stolbunov. They created the first Diffie-Hellman replacement based on elliptic curve isogenies. Unlike the method of De Feo, Jao, and Plut, the method of Rostovtsev and Stolbunov used ordinary elliptic curves and was found to have a subexponential quantum attack.In March 2014, researchers at the Chinese State Key Lab for Integrated Service Networks and Xidian University extended the security of the SIDH to a form of digital signature with strong designated verifier. In October 2014, Jao and Soukharev from the University of Waterloo presented an alternative method of creating undeniable signatures with designated verifier using elliptic curve isogenies."
large_corpus_distillation
enumeration of all the conditions on a rule. It often happens that a plausible commonsense rule, when examined closely, has an almost unlimited number of possible types of exceptions. The problem of dealing with all these potential exceptions is known as the qualification problem.\n5.\nInference from the absence of information.\nIt is often reasonable to infer that a statement A is false from the fact that one does not know A to be true, or from the fact that it is not stated to be true in a problem statement.\n6.\nLimiting the extent of inference.\nMany intuitively appealing sets of axioms have the property that the first few inferences all seem to be reasonable and to have reasonable conclusions, but that, as the inferences get further and further from the starting axioms, the conclusions seem less and less sensible, and they eventually end up in pure nonsense.\n7.\nInference using vague concepts. Inferences that involve reasoning near the boundaries of a vague concept are often uncertain.\n8.\nFinding
large_corpus_distillation
verbs and adjectives. There is a greater use of one word utterances and the pragmatics uses supportive language like expansions and re-casting."
large_corpus_distillation
differences are a more intensive usage of the feedback channel, and a stricter logic at both the compressor and the decompressor that prevents loss of context synchronization between compressor and decompressor, except for very high residual bit error rates."
large_corpus_distillation
"Cross-validation (statistics)-Limitations and misuse-In many applications, models also may be incorrectly specified and vary as a function of modeler biases and/or arbitrary choices. When this occurs, there may be an illusion that the system changes in external samples, whereas the reason is that the model has missed a critical predictor and/or included a confounded predictor. New evidence is that cross-validation by itself is not very predictive of external validity, whereas a form of experimental validation known as swap sampling that does control for human bias can be much more predictive of external validity. As defined by this large MAQC-II study across 30,000 models, swap sampling incorporates cross-validation in the sense that predictions are tested across independent training and validation samples. Yet, models are also developed across these independent samples and by modelers who are blinded to one another. When there is a mismatch in these models developed across these swapped training and
large_corpus_distillation
"STM-4-STM-1 frame structure-The first 36 bytes of each row of a frame (with the exception of row 16) contain a Section OverHead. The first 36 bytes of row 16 contain the AU pointer (payload address). The field consisting of 261X36 bytes is for payload transmission (The STM payload is known as the Virtual Container (VC) and consists of the actual payload, known as the Container (C) and additional overhead known as Path Over Head (POH), for path related information)."
large_corpus_distillation
"Diceball-Gameplay and rules-To start the game, the visiting team puts a pawn (as a batter and eventually runner) in the batter's box, to get the pitcher's throws. The die replaces the ball. The pitcher rolls the die until either the batter is struck out, the batter is walked, or the ball is hit. If the ball is hit, the offensive team rolls a die to determine the number of dice to be used to hit the ball."
large_corpus_distillation
"Dead Sea salt-Therapeutic benefits-Dead Sea salts have been claimed to treat the following conditions: Rheumatologic conditions Rheumatologic conditions can be treated in the balneotherapy of rheumatoid arthritis, psoriatic arthritis, and osteoarthritis. The minerals are absorbed while soaking, stimulating blood circulation."
large_corpus_distillation
"Kruskal's tree theorem-Statement-Take X to be a partially ordered set. If T1, T2 are rooted trees with vertices labeled in X, we say that T1 is inf-embeddable in T2 and write T1 \u2264 T2 if there is an injective map F from the vertices of T1 to the vertices of T2 such that For all vertices v of T1, the label of v precedes the label of F(v), If w is any successor of v in T1, then F(w) is a successor of F(v), and If w1, w2 are any two distinct immediate successors of v, then the path from F(w1) to F(w2) in T2 contains F(v)."
large_corpus_distillation
"Situation semantics-HPSG-Situation semantics is the first semantic theory that was used in head-driven phrase structure grammar (HPSG)."
large_corpus_distillation
following are the sufficiency determiners: enough sufficientThese determiners convey inexact quantification that is framed in terms of some minimum quantity needed. For instance, enough money for a taxi implies that a minimum amount of money is necessary to pay for a taxi and that the amount of money in question is sufficient for the purpose. When functioning as determinatives in a noun phrase, sufficiency determiners select plural count nouns (e.g., sufficient reasons) or non-count nouns (e.g., enough money).:\u200a396 Interrogative determiners The following are the interrogative determiners: what whichThese determiners can also be followed by -ever and -soever. Interrogative determiners are typically used in the formation of questions, as in what/which conductor do you like best? Using what marks a noun phrase as indefinite while using which marks the noun phrase as definite, being used when the context implies a limited number of choices.:\u200a369 Relative determiners The following are the relative
large_corpus_distillation
"Low-rank matrix approximations-Randomized feature maps approximation-Random binning features A random binning features map partitions the input space using randomly shifted grids at randomly chosen resolutions and assigns to an input point a binary bit string that corresponds to the bins in which it falls. The grids are constructed so that the probability that two points x,x\u2032\u2208Rd are assigned to the same bin is proportional to K(x,x\u2032) . The inner product between a pair of transformed points is proportional to the number of times the two points are binned together, and is therefore an unbiased estimate of K(x,x\u2032) . Since this mapping is not smooth and uses the proximity between input points, Random Binning Features works well for approximating kernels that depend only on the L1 distance between datapoints."
large_corpus_distillation
"Helium mass spectrometer-Test methods-Bombing test This method applies to objects that are supposedly sealed.\nFirst the device under test will be exposed for an extended length of time to a high helium pressure in a \"bombing\" chamber.\nIf the part is leaky, helium will be able to penetrate the device.\nLater the device will be placed in a vacuum chamber, connected to a vacuum pump and a mass spectrometer. The tiny amount of gas that entered the device under pressure will be released in the vacuum chamber and sent to the mass spectrometer where the leak rate will be measured.\nThis test method applies to implantable medical devices, crystal oscillator, saw filter devices.\nThis method is not able to detect a massive leak as the tracer gas will be quickly pumped out when test chamber is pumped down.\nHelium charged sniffer test In this last case the part is pressurized with helium. The mass spectrometer is fitted with a special device, a sniffer probe, that allows it to sample air (and tracer gas when
large_corpus_distillation
"Layered permutation-Characterization by forbidden patterns-The layered permutations can also be equivalently described as the permutations that do not contain the permutation patterns 231 or 312. That is, no three elements in the permutation (regardless of whether they are consecutive) have the same ordering as either of these forbidden triples."
large_corpus_distillation
The more direct consequence of this situation is that it is also not possible for Teams A and B to finish in a tie with each other, and Team C cannot win the division."
large_corpus_distillation
"a. Create a table displaying this data.\nb. What is the expected frequency in each cell?\nc. Conduct a significance test to see if there is a relationship between smoking and incontinence. What Chi Square value do you\np\nget? What value do you get?\nd. What do you conclude? (relevant section)\nQ5\nAt a school pep rally, a group of sophomore students organized a free raffle for prizes. They claim that they put the names of all of\n36 6\nthe students in the school in the basket and that they randomly drew names out of this basket. Of the prize winners, were\n14 9 7\nfreshmen, were sophomores, were juniors, and were seniors. The results do not seem that random to you. You think it is a\n30% 25%\nlittle fishy that sophomores organized the raffle and also won the most prizes. Your school is composed of freshmen,\n25% 20%\nsophomores, juniors, and seniors.\na. What are the expected frequencies of winners from each class?\nb. Conduct a significance test to determine whether the winners of the prizes were
large_corpus_distillation
now supported performance improvements New camera support New lens correction support Lightroom Cloud ecosystem has also been updated Bug fixes Version 11.3 (April 12, 2022 (2022-04-12))Masking Update Reset Preferences Performance New camera support New lens correction support Bug fixes Version 11.3.1 (April 21, 2022 (2022-04-21))Bug fixed: Open Catalog causing substrate.dll issue and crashing (some Windows systems)."
large_corpus_distillation
"ARM Cortex-A9-Features-Key features of the Cortex-A9 core are: Out-of-order speculative issue superscalar execution 8-stage pipeline giving 2.50 DMIPS/MHz/core.\nNEON SIMD instruction set extension performing up to 16 operations per instruction (optional).\nHigh performance VFPv3 floating point unit doubling the performance of previous ARM FPUs (optional).\nThumb-2 instruction set encoding reduces the size of programs with little impact on performance.\nTrustZone security extensions.\nJazelle DBX support for Java execution.\nJazelle RCT for JIT compilation.\nProgram Trace Macrocell and CoreSight Design Kit for non-intrusive tracing of instruction execution.\nL2 cache controller (0\u20134 MB).\nMulti-core processing.ARM states that the TSMC 40G hard macro implementation typically operates at 2 GHz; a single core (excluding caches) occupies less than 1.5 mm2 when designed in a TSMC 65 nanometer (nm) generic process and can be clocked at speeds over 1 GHz, consuming less than 250 mW per core."
large_corpus_distillation
"Repetition code-Code parameters-In the case of a binary repetition code, there exist two code words - all ones and all zeros - which have a length of n . Therefore, the minimum Hamming distance of the code equals its length n . This gives the repetition code an error correcting capacity of n\u221212 (i.e. it will correct up to n\u221212 errors in any code word)."
large_corpus_distillation
"Reblogging-History and comparison-Twitter On Twitter, another micro-blogging service, the \"retweet\" phenomenon began in March 2007, whereby an informal protocol slowly developed among Twitter users. Initially called an \"echo\" the first instance was by Narendra Rocherolle. A month later the term \"retweet\" was first referenced. In this protocol, a post by User A which was considered notable by another Twitter user, User B, was copied by User B and reposted under User B's account and prefixed with \"RT @username\", with \"RT\" meaning \"retweet\". The at sign had previously been developed as a means of replying to Twitter posts, while the hashtag protocol would later be developed by users within the same year to highlight important keywords which they wished to be indexed in a search engine."
large_corpus_distillation
symmetry, we need only prove (a). The conditional PDF of Y given X=x is y\u21a6f(x,y)/g(x) where f is the joint PDF, and where g is\nthe PDF of X, namely the normal PDF with mean \u03bc and standard deviation \u03c3. The result then follows after some algebra.\nProof from random variables\nAgain, we only need to prove (a). We can assume that (X,Y) is defined in terms of a standard normal pair (Z,W) as in the definition. Hence\nX\u2212\u03bc \u2212\u2212\u2212\u2212\u2212\nY =\u03bd+\u03c1\u03c4 +\u03c4\u221a1\u2212\u03c12W (5.7.14)\n\u03c3\nSince that X and W are independent, the conditional distribution of Y given X=x is the distribution of \u03bd+\u03c1\u03c4x\u2212 \u03c3 \u03bc +\u03c4\u221a \u2212 1 \u2212 \u2212 \u2212\u2212 \u03c1 \u2212 2W . The\nlatter distribution is normal, with mean and variance specified in the theorem.\nNote that the conditional variances do not depend on the value of the given variable.\nIn the bivariate normal experiment, set the standard deviation of X to 1.5, the standard
large_corpus_distillation
"Error correction code-Code-rate and the tradeoff between reliability and data rate-The most popular ECCs have a trade-off between performance and computational complexity. Usually, their parameters give a range of possible code rates, which can be optimized depending on the scenario. Usually, this optimization is done in order to achieve a low decoding error probability while minimizing the impact to the data rate. Another criterion for optimizing the code rate is to balance low error rate and retransmissions number in order to the energy cost of the communication."
large_corpus_distillation
"Bergman's diamond lemma-Examples-Since p,q,r are commutative the above are equal. Thus the ambiguity resolves and the Lemma implies that {yx\u2212pxy,zx\u2212qxz,zy\u2212ryz} is a Gr\u00f6bner basis of I Non-resolving ambiguities Let A=k\u27e8x,y,z\u27e9/(z2\u2212xy\u2212yx,zx\u2212xz,zy\u2212yz) . Under the same ordering as in the previous example, the leading words of the generators of the ideal are z2 , zx and zy . There are two overlap ambiguities, namely z2x and z2y . Let us consider z2x . If we resolve z2 first we get, z2x=(xy+yx)x=xyx+yx2, which contains no leading words and is therefore reduced. Resolving zx first we obtain, z2x=zxz=xz2=x(xy+yx)=x2y+xyx."
large_corpus_distillation
"Arithmetic circuit complexity-Overview-Note that we are interested in the formal computation of polynomials, rather than the functions that the polynomials define. For example, consider the polynomial x2+x; over the field of two elements this polynomial represents the zero function, but it is not the zero polynomial. This is one of the differences between the study of arithmetic circuits and the study of Boolean circuits. In Boolean complexity, one is mostly interested in computing a function, rather than some representation of it (in our case, a representation by a polynomial). This is one of the reasons that make Boolean complexity harder than arithmetic complexity. The study of arithmetic circuits may also be considered as one of the intermediate steps towards the study of the Boolean case, which we hardly understand."
large_corpus_distillation
"Piling-up lemma-Expected value formulation-For dependent variables the above formulation gains a (positive or negative) covariance term, thus the lemma does not hold. In fact, since two Bernoulli variables are independent if and only if they are uncorrelated (i.e. have zero covariance; see uncorrelatedness), we have the converse of the piling up lemma: if it does not hold, the variables are not independent (uncorrelated)."
large_corpus_distillation
"MMB (cipher)-MMB (cipher)-MMB has a key size and block size of 128 bits."
large_corpus_distillation
"Perceptual computing-Perceptual computer-A vocabulary is application (context) dependent, and must be large enough so that it lets the end-user interact with the Per-C in a user-friendly manner. The encoder transforms words into fuzzy sets (FSs) and leads to a codebook \u2013 words with their associated FS models. The outputs of the encoder activate a Computing With Words (CWW) engine, whose output is one or more other FSs, which are then mapped by the decoder into a recommendation (subjective judgment) with supporting data. The recommendation may be in the form of a word, group of similar words, rank or class."
large_corpus_distillation
"6.867 Machine learning, lecture 6 (Jaakkola) 1\nLecture topics:\n\u2022 Active learning\n\u2022 Non-linear predictions, kernels\nActive learning\nWe can use the expressions for the mean squared error to actively select input points\nx ,..., x , when possible, so as to reduce the resulting estimation error. This is an active\n1 n\nlearning (experiment design) problem. By letting the method guide the selection of the\ntraining examples (inputs), we will generally need far fewer examples in comparison to\nselecting them at random from some underlying distribution, database, or trying available\nexperiments at random.\nTo develop this further, recall that we continue to assume that the responses y come from\nsome linear model y = \u03b8\u2217T x + \u03b8\u2217 + \ufffd where \ufffd \u223c N(0,\u03c3\u22172). Nothing is assumed about the\n0\ndistribution of x as the choice of the inputs is in our control. For any given set of inputs,\nx ,..., x , we derived last time an expression for the mean squared error of
large_corpus_distillation
"Square-difference-free set-Upper bounds-According to the Furstenberg\u2013S\u00e1rk\u00f6zy theorem, if S is a square-difference-free set, then the natural density of S is zero. That is, for every \u03b5>0 , and for all sufficiently large n , the fraction of the numbers up to n that are in S is less than \u03b5 . Equivalently, every set of natural numbers with positive upper density contains two numbers whose difference is a square, and more strongly contains infinitely many such pairs. The Furstenberg\u2013S\u00e1rk\u00f6zy theorem was conjectured by L\u00e1szl\u00f3 Lov\u00e1sz, and proved independently in the late 1970s by Hillel Furstenberg and Andr\u00e1s S\u00e1rk\u00f6zy, after whom it is named. Since their work, several other proofs of the same result have been published, generally either simplifying the previous proofs or strengthening the bounds on how sparse a square-difference-free set must be. The best upper bound currently known is due to Thomas Bloom and James Maynard, who show that a
large_corpus_distillation
"Algebra over a field-Kinds of algebras and examples-An example of unital zero algebra is the algebra of dual numbers, the unital zero R-algebra built from a one dimensional real vector space."
large_corpus_distillation
"900 9 \u2022 Sequences, Probability, and Counting Theory\n\u2022 In the line headed\u201cVariable:\u201dtype in the variable used on the previous step.\n\u2022 In the line headed\u201cstart:\u201dkey in the value of that begins the sequence.\n\u2022 In the line headed\u201cend:\u201dkey in the value of that ends the sequence.\n\u2022 Press[ENTER]3 times to return to the home screen. You will see the sequence syntax on the screen. Press[ENTER]\nto see the list of terms for the finite sequence defined. Use the right arrow key to scroll through the list of terms.\nUsing a TI-83, do the following.\n\u2022 In the home screen, press[2ND] LIST.\n\u2022 Scroll over toOPSand choose\u201cseq(\u201dfrom the dropdown list. Press[ENTER].\n\u2022 Enter the items in the order\u201cExpr\u201d,\u201cVariable\u201d,\u201cstart\u201d,\u201cend\u201dseparated by commas. See the instructions above\nfor the description of each item.\n\u2022 Press[ENTER]to see the list of terms for the finite sequence defined. Use the right arrow
large_corpus_distillation
"Newick format-Examples-The following tree: could be represented in Newick format in several ways (,,(,)); no nodes are named (A,B,(C,D)); leaf nodes are named (A,B,(C,D)E)F; all nodes are named (:0.1,:0.2,(:0.3,:0.4):0.5); all but root node have a distance to parent (:0.1,:0.2,(:0.3,:0.4):0.5):0.0; all have a distance to parent (A:0.1,B:0.2,(C:0.3,D:0.4):0.5); distances and leaf names (popular) (A:0.1,B:0.2,(C:0.3,D:0.4)E:0.5)F; distances and all names ((B:0.2,(C:0.3,D:0.4)E:0.5)F:0.1)A; a tree rooted on a leaf node (rare) Newick format is typically used for tools like PHYLIP and is a minimal definition for a phylogenetic tree."
large_corpus_distillation
"Wireless Application Protocol-Technical specifications-WAP stack The WAP standard described a protocol suite or stack allowing the interoperability of WAP equipment and software with different network technologies, such as GSM and IS-95 (also known as CDMA)."
large_corpus_distillation
"Inversion list-Inversion list-For example, for ranges 10-14, 25-37, the inversion list would be: 10 15 25 38 To search whether an item belongs to any of the ranges, a binary search is made. If the search ends in a \"first\" element, the searched item is in the set. If the search ends in an \"after\" element, or outside the array, the searched item is not in the set."
large_corpus_distillation
to, yet distinct from, the uses of those terms in RDF. This example demonstrates: In the English language statement 'New York has the postal abbreviation NY' , 'New York' would be the subject, 'has the postal abbreviation' the predicate and 'NY' the object."
large_corpus_distillation
"Generalized suffix array-Construction Algorithms and Implementations-Algorithms and tools for constructing a generalized suffix array include: Fei Shi's (1996) algorithm which runs in log \u2061n) worst case time and O(N) space, where N is the sum of the lengths of all strings in S and n the length of the longest string in S . This includes sorting, searching and finding the longest common prefixes."
large_corpus_distillation
"Coefficient of determination-Comparison with norm of residuals-Occasionally, the norm of residuals is used for indicating goodness of fit. This term is calculated as the square-root of the sum of squares of residuals: norm of residuals res =\u2016e\u2016."
large_corpus_distillation
"Computer performance-Aspects of performance-Availability Availability of a system is typically measured as a factor of its reliability - as reliability increases, so does availability (that is, less downtime). Availability of a system may also be increased by the strategy of focusing on increasing testability and maintainability and not on reliability. Improving maintainability is generally easier than reliability. Maintainability estimates (Repair rates) are also generally more accurate. However, because the uncertainties in the reliability estimates are in most cases very large, it is likely to dominate the availability (prediction uncertainty) problem, even while maintainability levels are very high."
large_corpus_distillation
"Variational autoencoder-Evidence lower bound (ELBO)-As distance loss between the two distributions the Kullback\u2013Leibler divergence DKL(q\u03d5(z|x)\u2225p\u03b8(z|x)) is a good choice to squeeze q\u03d5(z|x) under p\u03b8(z|x) .The distance loss just defined is expanded as ln ln ln ln \u2061q\u03d5(z|x)p\u03b8(x,z)] Now define the evidence lower bound (ELBO):Maximizing the ELBOis equivalent to simultaneously maximizing ln \u2061p\u03b8(x) and minimizing DKL(q\u03d5(z|x)\u2225p\u03b8(z|x)) . That is, maximizing the log-likelihood of the observed data, and minimizing the divergence of the approximate posterior q\u03d5(\u22c5|x) from the exact posterior p\u03b8(\u22c5|x) The form given is not very convenient for maximization, but the following, equivalent form, is:where ln \u2061p\u03b8(x|z) is implemented as \u221212\u2016x\u2212D\u03b8(z)\u201622 , since that is, up to an additive constant, what x\u223cN(D\u03b8(z),I) yields. That is, we model the distribution of x conditional on z to be a
large_corpus_distillation
"Document Object Model-DOM tree structure-A Document Object Model (DOM) tree is a hierarchical representation of an HTML or XML document. It consists of a root node, which is the document itself, and a series of child nodes that represent the elements, attributes, and text content of the document. Each node in the tree has a parent node, except for the root node, and can have multiple child nodes."
large_corpus_distillation
"Author profiling-Author profiling and the Internet-Blogs Author profiling can be used to identify characteristics of blog writers, such as their age, gender and geographical location, based on their different writing styles, This is especially useful when it comes to anonymous blogs. The choice of content words, style-based features and topic-based features are analyzed to discover characteristics of the author.In general, features that are frequently occur in blogs include a high distribution of verbs per writing and a relatively high use of pronouns. The frequency of verbs, pronouns and other word classes are used to profile and classify emotions in the writings of authors, as well as their gender and age. Author profiling using classification models that were used on physical documents in the past, such as Support Vector Machines, have also been tested on blogs. However, it has been proven to be unsuitable for the latter due to its low performance.The machine learning algorithms that work well for author
large_corpus_distillation
is divisible by 14, then the given number is divisible by 14.\nExample: 1468, 7394 rule => since 68(14*5) divisible by 14, then 1468 is divisible by 14.\nrule => since 94(14*7) divisible by 14, then 7394 is divisible by 14."
large_corpus_distillation
n denotes the number of vertices.\nAfter m times executions of the outer loop, we output the minimum cut among all the results. The figure 2 gives an example of one execution of the algorithm. After execution, we get a cut of size 3."
large_corpus_distillation
fast and efficient computation. The main idea of this algorithm is to use the Polynomial Transform to convert the multidimensional DCT into a series of 1-D DCTs directly. MD DCT-IV also has several applications in various fields."
large_corpus_distillation
"Image server-Uses-In e-commerce, image servers are qualified by their abilities to scale to hundreds of thousands of images, to multiple CPUs or load-balanced server machines, and to the quantity and quality of their image processing functionalities, such as resizing, compositing, zoom and 3D viewers, and the addition of dynamic data to the images in the form of overlaid text or graphics."
large_corpus_distillation
"DriveSpace-Bugs and data loss-The fact that the compressed contents of a compressed drive was stored in a single file implied the possibility of a user accidentally deleting all of their data by deleting just that file. This could happen if the user inadvertently got access to the host drive, containing this file. The host drive was usually mapped to the letter H: by the compression driver. However, if the compression driver had failed to load the user might see it as drive C:."
large_corpus_distillation
"Transport Layer Security-Algorithms-Key exchange or key agreement Before a client and server can begin to exchange information protected by TLS, they must securely exchange or agree upon an encryption key and a cipher to use when encrypting data (see \u00a7 Cipher). Among the methods used for key exchange/agreement are: public and private keys generated with RSA (denoted TLS_RSA in the TLS handshake protocol), Diffie\u2013Hellman (TLS_DH), ephemeral Diffie\u2013Hellman (TLS_DHE), elliptic-curve Diffie\u2013Hellman (TLS_ECDH), ephemeral elliptic-curve Diffie\u2013Hellman (TLS_ECDHE), anonymous Diffie\u2013Hellman (TLS_DH_anon), pre-shared key (TLS_PSK) and Secure Remote Password (TLS_SRP).The TLS_DH_anon and TLS_ECDH_anon key agreement methods do not authenticate the server or the user and hence are rarely used because those are vulnerable to man-in-the-middle attacks. Only TLS_DHE and TLS_ECDHE provide forward secrecy."
large_corpus_distillation
"Cache prefetching-Methods of hardware prefetching-Stream buffers are one of the most common hardware based prefetching techniques in use. This technique was originally proposed by Norman Jouppi in 1990 and many variations of this method have been developed since. The basic idea is that the cache miss address (and k subsequent addresses) are fetched into a separate buffer of depth k . This buffer is called a stream buffer and is separate from cache. The processor then consumes data/instructions from the stream buffer if the address associated with the prefetched blocks match the requested address generated by the program executing on the processor. The figure below illustrates this setup: Whenever the prefetch mechanism detects a miss on a memory block, say A, it allocates a stream to begin prefetching successive blocks from the missed block onward. If the stream buffer can hold 4 blocks, then we would prefetch A+1, A+2, A+3, A+4 and hold those in the allocated stream buffer. If the processor consumes A+1
large_corpus_distillation
"Square number-Properties-Squarity testing can be used as alternative way in factorization of large numbers. Instead of testing for divisibility, test for squarity: for given m and some number k, if k2 \u2212 m is the square of an integer n then k \u2212 n divides m. (This is an application of the factorization of a difference of two squares.) For example, 1002 \u2212 9991 is the square of 3, so consequently 100 \u2212 3 divides 9991. This test is deterministic for odd divisors in the range from k \u2212 n to k + n where k covers some range of natural numbers k\u2265m."
large_corpus_distillation
"API testing-API testing overview-API testing involves testing APIs directly (in isolation) and as part of the end-to-end transactions exercised during integration testing. Beyond RESTful APIs, these transactions include multiple types of endpoints such as web services, ESBs, databases, mainframes, web UIs, and ERPs. API testing is performed on APIs that the development team produces as well as APIs that the team consumes within their application (including third-party APIs).API testing is used to determine whether APIs return the correct response (in the expected format) for a broad range of feasible requests, react properly to edge cases such as failures and unexpected/extreme inputs, deliver responses in an acceptable amount of time, and respond securely to potential security attacks. Service virtualization is used in conjunction with API testing to isolate the services under test as well as expand test environment access by simulating APIs/services that are not accessible for testing.API testing commonly
large_corpus_distillation
of these updates, releasing updates for over half of the top 50 devices worldwide in the last quarter of 2016\". They also wrote that \"About half of devices in use at the end of 2016 had not received a platform security update in the previous year\", stating that their work would continue to focus on streamlining the security updates program for easier deployment by manufacturers. Furthermore, in a comment to TechCrunch, Ludwig stated that the wait time for security updates had been reduced from \"six to nine weeks down to just a few days\", with 78% of flagship devices in North America being up-to-date on security at the end of 2016.Patches to bugs found in the core operating system often do not reach users of older and lower-priced devices. However, the open-source nature of Android allows security contractors to take existing devices and adapt them for highly secure uses. For example, Samsung has worked with General Dynamics through their Open Kernel Labs acquisition to rebuild Jelly Bean on top of their
large_corpus_distillation
the product manager represents the needs of end-users, evaluates market trends and competition, and uses this information to determine what features to build. For example, a product manager may decide a feature is needed because users are asking for it, or because the feature is needed to stay competitive. In order to facilitate this decision-making process the product manager may set out a vision for the product or a general framework for making product decisions. The product manager also ensures an atmosphere of cohesiveness and focused collaboration between all the members of the team, all in the interest of driving the product forward. Product managers are often thought of as sitting at the intersection of business, design, and technology."
large_corpus_distillation
of linguistic features, such as part of speech (POS). POS is introduced as a certain functionality of both syntax and semantic point of view, e.g. if a token of output sentence is a verb while it is expected to be a noun, then there shall be a penalty; also, if the POS is the same but the exact word is not the same, e.g. good vs nice, then this candidate shall gain certain credit. The overall score of hLEPOR then is calculated as the combination of word level score and POS level score with a weighting set. Language modelling inspired n-gram knowledge is also extensively explored in nLEPOR. In addition to the n-gram knowledge for n-gram position difference penalty calculation, n-gram is also applied to n-gram precision and n-gram recall in nLEPOR, and the parameter n is an adjustable factor. In addition to POS knowledge in hLEPOR, phrase structure from parsing information is included in a new variant HPPR. In HPPR evaluation modeling, the phrase structure set, such as noun phrase, verb phrase, prepositional
large_corpus_distillation
"Mean squared error-Applications-Minimizing MSE is a key criterion in selecting estimators: see minimum mean-square error. Among unbiased estimators, minimizing the MSE is equivalent to minimizing the variance, and the estimator that does this is the minimum variance unbiased estimator. However, a biased estimator may have lower MSE; see estimator bias.\nIn statistical modelling the MSE can represent the difference between the actual observations and the observation values predicted by the model. In this context, it is used to determine the extent to which the model fits the data as well as whether removing some explanatory variables is possible without significantly harming the model's predictive ability.\nIn forecasting and prediction, the Brier score is a measure of forecast skill based on MSE."
large_corpus_distillation