source stringclasses 6
values | text stringlengths 2 99.5k |
|---|---|
AlaaElhilo/Wikipedia_ComputerScience | Operations and structures for "sequence control" allow controlling the execution flow of program instructions. When certain conditions are met, it is necessary to change the typical sequential execution of a program. Therefore, the interpreter employs data structures that are modified by operations distinct from those... |
AlaaElhilo/Wikipedia_ComputerScience | Data transfer operations are used to control how operands and data are transported from memory to the interpreter and vice versa. These operations deal with the store and the retrieval order of operands from the store. |
AlaaElhilo/Wikipedia_ComputerScience | Memory management is concerned with the operations performed in memory to allocate data and applications. In the abstract machine, data and programmes can be held indefinitely, or in the case of programming languages, memory can be allocated or deallocated using a more complex mechanism. |
AlaaElhilo/Wikipedia_ComputerScience | Abstract machine hierarchies are often employed, in which each machine uses the functionality of the level immediately below and adds additional functionality of its own to meet the level immediately above. A hardware computer, constructed with physical electronic devices, can be added at the most basic level. Above th... |
AlaaElhilo/Wikipedia_ComputerScience | Cell, a cellular architecture containing 9 cores, is the processor used in the PlayStation 3. Another prominent cellular architecture is Cyclops64, a massively parallel architecture currently under development by IBM. |
AlaaElhilo/Wikipedia_ComputerScience | Cellular architectures follow the low-level programming paradigm, which exposes the programmer to much of the underlying hardware. This allows the programmer to greatly optimize their code for the platform, but at the same time makes it more difficult to develop software. |
AlaaElhilo/Wikipedia_ComputerScience | In mathematical logic and theoretical computer science, a register machine is a generic class of abstract machines used in a manner similar to a Turing machine. All models of register machines are Turing equivalent. |
AlaaElhilo/Wikipedia_ComputerScience | The register machine gets its name from its use of one or more "registers". In contrast to the tape and head used by a Turing machine, the model uses multiple, uniquely addressed registers, each of which holds a single positive integer. |
AlaaElhilo/Wikipedia_ComputerScience | There are at least four sub-classes found in literature, here listed from most primitive to the most like a computer: |
AlaaElhilo/Wikipedia_ComputerScience | Any properly defined register machine model is Turing equivalent. Computational speed is very dependent on the model specifics. |
AlaaElhilo/Wikipedia_ComputerScience | In practical computer science, a related concept known as a virtual machine is occasionally employed to reduce reliance on underlying machine architectures. These virtual machines are also utilized in educational settings. In textbooks, the term "register machine" is sometimes used interchangeably to describe a virtual... |
AlaaElhilo/Wikipedia_ComputerScience | A register machine consists of: |
AlaaElhilo/Wikipedia_ComputerScience | An unbounded number of labelled, discrete, unbounded registers unbounded in extent : a finite set of registers
r
0
…
r
n
{\displaystyle r_{0}\ldots r_{n}}
... |
AlaaElhilo/Wikipedia_ComputerScience | Tally counters or marks: discrete, indistinguishable objects or marks of only one sort suitable for the model. In the most-reduced counter machine model, per each arithmetic operation only one object/mark is either added to or removed from its location/tape. In some counter machine models and most RAM and RASP models ... |
AlaaElhilo/Wikipedia_ComputerScience | A limited set of instructions: the instructions tend to divide into two classes: arithmetic and control. The instructions are drawn from the two classes to form "instruction-sets", such that an instruction set must allow the model to be Turing equivalent .
Arithmetic: Arithmetic instructions may operate on all register... |
AlaaElhilo/Wikipedia_ComputerScience | Arithmetic: Arithmetic instructions may operate on all registers or on a specific register, such as an accumulator. Typically, they are selected from the following sets, though exceptions exist: Counter machine: { Increment , Decrement , Clear-to-zero } Reduced RAM, RASP: { Increment , Decrement , Clear-to-zero , Loa... |
AlaaElhilo/Wikipedia_ComputerScience | Control: Counter machine models: Optionally include { Copy }. RAM and RASP models: Most include { Copy }, or { Load Accumulator from r, Store accumulator into r, Load Accumulator with an immediate constant }. All models: Include at least one conditional "jump" following the test of a register, such as { Jump-if-ze... |
AlaaElhilo/Wikipedia_ComputerScience | Counter machine: no indirect addressing, immediate operands possible in highly atomized models |
AlaaElhilo/Wikipedia_ComputerScience | RAM and RASP: indirect addressing available, immediate operands typical |
AlaaElhilo/Wikipedia_ComputerScience | Input-output: optional in all models |
AlaaElhilo/Wikipedia_ComputerScience | State register: A special Instruction Register , distinct from the registers mentioned earlier, stores the current instruction to be executed along with its address in the instruction table. This register, along with its associated table, is located within the finite state machine. The IR is inaccessible in all models.... |
AlaaElhilo/Wikipedia_ComputerScience | Two trends appeared in the early 1950s—the first to characterize the computer as a Turing machine, the second to define computer-like models—models with sequential instruction sequences and conditional jumps—with the power of a Turing machine, i.e. a so-called Turing equivalence. Need for this work was carried out in c... |
AlaaElhilo/Wikipedia_ComputerScience | The first trend toward characterizing computers has originated with Hans Hermes , Rózsa Péter , and Heinz Kaphengst , the second trend with Hao Wang and, as noted above, furthered along by Zdzislaw Alexander Melzak , Joachim Lambek and Marvin Minsky . |
AlaaElhilo/Wikipedia_ComputerScience | The last five names are listed explicitly in that order by Yuri Matiyasevich. He follows up with: |
AlaaElhilo/Wikipedia_ComputerScience | Lambek, Melzak, Minsky and Shepherdson and Sturgis independently discovered the same idea at the same time. See note on precedence below. |
AlaaElhilo/Wikipedia_ComputerScience | The history begins with Wang's model. |
AlaaElhilo/Wikipedia_ComputerScience | Wang's work followed from Emil Post's paper and led Wang to his definition of his Wang B-machine—a two-symbol Post–Turing machine computation model with only four atomic instructions: |
AlaaElhilo/Wikipedia_ComputerScience | To these four both Wang and then C. Y. Lee added another instruction from the Post set { ERASE }, and then a Post's unconditional jump { JUMP_to_ instruction_z } (or to make things easier, the conditional jump JUMP_IF_blank_to_instruction_z, or both. Lee named this a "W-machine" model: |
AlaaElhilo/Wikipedia_ComputerScience | Wang expressed hope that his model would be "a rapprochement": 63 between the theory of Turing machines and the practical world of the computer. |
AlaaElhilo/Wikipedia_ComputerScience | Wang's work was highly influential. We find him referenced by Minsky and , Melzak , Shepherdson and Sturgis . Indeed, Shepherdson and Sturgis remark that: |
AlaaElhilo/Wikipedia_ComputerScience | Martin Davis eventually evolved this model into the Post–Turing machine. |
AlaaElhilo/Wikipedia_ComputerScience | Difficulties with the Wang/Post–Turing model: |
AlaaElhilo/Wikipedia_ComputerScience | Except there was a problem: the Wang model was still a single-tape Turing-like device, however nice its sequential program instruction-flow might be. Both Melzak and Shepherdson and Sturgis observed this : |
AlaaElhilo/Wikipedia_ComputerScience | Indeed, as examples at Turing machine examples, Post–Turing machine and partial function show, the work can be "complicated". |
AlaaElhilo/Wikipedia_ComputerScience | So why not 'cut the tape' so each is infinitely long but left-ended, and call these three tapes "Post–Turing tapes"? The individual heads will move left and right . In one sense the heads indicate "the tops of the stack" of concatenated marks. Or in Minsky and Hopcroft and Ullman : 171ff the tape is always blank e... |
AlaaElhilo/Wikipedia_ComputerScience | Care must be taken to write the instructions so that a test-for-zero and jump occurs before decrementing, otherwise the machine will "fall off the end" or "bump against the end"—creating an instance of a partial function. |
AlaaElhilo/Wikipedia_ComputerScience | Minsky and Shepherdson–Sturgis prove that only a few tapes—as few as one—still allow the machine to be Turing equivalent if the data on the tape is represented as a Gödel number ; this number will evolve as the computation proceeds. In the one tape version with Gödel number encoding the counter machine must be able t... |
AlaaElhilo/Wikipedia_ComputerScience | Melzak's model is significantly different. He took his own model, flipped the tapes vertically, called them "holes in the ground" to be filled with "pebble counters". Unlike Minsky's "increment" and "decrement", Melzak allowed for proper subtraction of any count of pebbles and "adds" of any count of pebbles. |
AlaaElhilo/Wikipedia_ComputerScience | He defines indirect addressing for his model: 288 and provides two examples of its use;: 89 his "proof": 290–292 that his model is Turing equivalent is so sketchy that the reader cannot tell whether or not he intended the indirect addressing to be a requirement for the proof. |
AlaaElhilo/Wikipedia_ComputerScience | Legacy of Melzak's model is Lambek's simplification and the reappearance of his mnemonic conventions in Cook and Reckhow 1973. |
AlaaElhilo/Wikipedia_ComputerScience | Lambek took Melzak's ternary model and atomized it down to the two unary instructions—X+, X− if possible else jump—exactly the same two that Minsky had come up with. |
AlaaElhilo/Wikipedia_ComputerScience | However, like the Minsky model, the Lambek model does execute its instructions in a default-sequential manner—both X+ and X− carry the identifier of the next instruction, and X− also carries the jump-to instruction if the zero-test is successful. |
AlaaElhilo/Wikipedia_ComputerScience | A RASP or random-access stored-program machine begins as a counter machine with its "program of instruction" placed in its "registers". Analogous to, but independent of, the finite state machine's "Instruction Register", at least one of the registers ) and one or more "temporary" registers maintain a record of, and ope... |
AlaaElhilo/Wikipedia_ComputerScience | Except there is a problem: If based on the counter machine chassis this computer-like, von Neumann machine will not be Turing equivalent. It cannot compute everything that is computable. Intrinsically the model is bounded by the size of its finite state machine's instructions. The counter machine based RASP can comput... |
AlaaElhilo/Wikipedia_ComputerScience | Elgot–Robinson investigate the possibility of allowing their RASP model to "self modify" its program instructions. The idea was an old one, proposed by Burks–Goldstine–von Neumann , and sometimes called "the computed goto." Melzak specifically mentions the "computed goto" by name but instead provides his model with in... |
AlaaElhilo/Wikipedia_ComputerScience | Computed goto: A RASP program of instructions that modifies the "goto address" in a conditional- or unconditional-jump program instruction. |
AlaaElhilo/Wikipedia_ComputerScience | But this does not solve the problem . What is necessary is a method to fetch the address of a program instruction that lies "beyond/above" the upper bound of the finite state machine instruction register and TABLE. |
AlaaElhilo/Wikipedia_ComputerScience | Minsky hints at the issue in his investigation of a counter machine equipped with the instructions { CLR , INC , and RPT }. He doesn't tell us how to fix the problem, but he does observe that: |
AlaaElhilo/Wikipedia_ComputerScience | But Elgot and Robinson solve the problem: They augment their P0 RASP with an indexed set of instructions—a somewhat more complicated form of indirect addressing. Their P'0 model addresses the registers by adding the contents of the "base" register to the "index" specified explicitly in the instruction . Thus the inde... |
AlaaElhilo/Wikipedia_ComputerScience | By 1971 Hartmanis has simplified the indexing to indirection for use in his RASP model. |
AlaaElhilo/Wikipedia_ComputerScience | Indirect addressing: A pointer-register supplies the finite state machine with the address of the target register required for the instruction. Said another way: The contents of the pointer-register is the address of the "target" register to be used by the instruction. If the pointer-register is unbounded, the RAM, and... |
AlaaElhilo/Wikipedia_ComputerScience | Note that the finite state machine does not have to explicitly specify this target register's address. It just says to the rest of the machine: Get me the contents of the register pointed to by my pointer-register and then do xyz with it. It must specify explicitly by name, via its instruction, this pointer-register b... |
AlaaElhilo/Wikipedia_ComputerScience | Cook and Reckhow cite Hartmanis and simplify his model to what they call a random-access machine . In a sense we are back to Melzak but with a much simpler model than Melzak's. |
AlaaElhilo/Wikipedia_ComputerScience | Minsky was working at the MIT Lincoln Laboratory and published his work there; his paper was received for publishing in the Annals of Mathematics on 15 August 1960, but not published until November 1961. While receipt occurred a full year before the work of Melzak and Lambek was received and published . That both were... |
AlaaElhilo/Wikipedia_ComputerScience | Almost exactly the same thing happened to Shepherdson and Sturgis. Their paper was received in December 1961—just a few months after Melzak and Lambek's work was received. Again, they had little or no benefit of reviewing the work of Minsky. They were careful to observe in footnotes that papers by Ershov, Kaphengst an... |
AlaaElhilo/Wikipedia_ComputerScience | The final paper of Shepherdson and Sturgis did not appear in a peer-reviewed journal until 1963. And as they note in their Appendix A, the 'systems' of Kaphengst , Ershov , Péter are all so similar to what results were obtained later as to be indistinguishable to a set of the following: |
AlaaElhilo/Wikipedia_ComputerScience | Indeed, Shepherson and Sturgis conclude |
AlaaElhilo/Wikipedia_ComputerScience | By order of publishing date the work of Kaphengst , Ershov , Péter were first. |
AlaaElhilo/Wikipedia_ComputerScience | Background texts: The following bibliography of source papers includes a number of texts to be used as background. The mathematics that led to the flurry of papers about abstract machines in the 1950s and 1960s can be found in van Heijenoort —an assemblage of original papers spanning the 50 years from Frege to Gödel .... |
AlaaElhilo/Wikipedia_ComputerScience | For a good treatment of the counter machine see Minsky Chapter 11 "Models similar to Digital Computers"—he calls the counter machine a "program computer". A recent overview is found at van Emde Boas . A recent treatment of the Minsky /Lambek model can be found Boolos–Burgess–Jeffrey ; they reincarnate Lambek's "abacu... |
AlaaElhilo/Wikipedia_ComputerScience | The papers: The papers begin with Wang and his dramatic simplification of the Turing machine. Turing , Kleene , Davis and in particular Post are cited in Wang ; in turn, Wang is referenced by Melzak , Minsky and Shepherdson–Sturgis as they independently reduce the Turing tapes to "counters". Melzak provides his p... |
AlaaElhilo/Wikipedia_ComputerScience | For the most part the papers contain mathematics beyond the undergraduate level—in particular the primitive recursive functions and mu recursive functions presented elegantly in Kleene and less in depth, but useful nonetheless, in Boolos–Burgess–Jeffrey . |
AlaaElhilo/Wikipedia_ComputerScience | All texts and papers excepting the four starred have been witnessed. These four are written in German and appear as references in Shepherdson–Sturgis and Elgot–Robinson ; Shepherdson–Sturgis offer a brief discussion of their results in Shepherdson–Sturgis' Appendix A. The terminology of at least one paper seems to h... |
AlaaElhilo/Wikipedia_ComputerScience | NUMA architectures logically follow in scaling from symmetric multiprocessing architectures. They were developed commercially during the 1990s by Unisys, Convex Computer , Honeywell Information Systems Italy , Silicon Graphics , Sequent Computer Systems , Data General , Digital and ICL. Techniques developed by these... |
AlaaElhilo/Wikipedia_ComputerScience | The first commercial implementation of a NUMA-based Unix system was the Symmetrical Multi Processing XPS-100 family of servers, designed by Dan Gielan of VAST Corporation for Honeywell Information Systems Italy. |
AlaaElhilo/Wikipedia_ComputerScience | Modern CPUs operate considerably faster than the main memory they use. In the early days of computing and data processing, the CPU generally ran slower than its own memory. The performance lines of processors and memory crossed in the 1960s with the advent of the first supercomputers. Since then, CPUs increasingly have... |
AlaaElhilo/Wikipedia_ComputerScience | Limiting the number of memory accesses provided the key to extracting high performance from a modern computer. For commodity processors, this meant installing an ever-increasing amount of high-speed cache memory and using increasingly sophisticated algorithms to avoid cache misses. But the dramatic increase in size of ... |
AlaaElhilo/Wikipedia_ComputerScience | NUMA attempts to address this problem by providing separate memory for each processor, avoiding the performance hit when several processors attempt to address the same memory. For problems involving spread data , NUMA can improve the performance over a single shared memory by a factor of roughly the number of processor... |
AlaaElhilo/Wikipedia_ComputerScience | Of course, not all data ends up confined to a single task, which means that more than one processor may require the same data. To handle these cases, NUMA systems include additional hardware or software to move data between memory banks. This operation slows the processors attached to those banks, so the overall speed ... |
AlaaElhilo/Wikipedia_ComputerScience | AMD implemented NUMA with its Opteron processor , using HyperTransport. Intel announced NUMA compatibility for its x86 and Itanium servers in late 2007 with its Nehalem and Tukwila CPUs. Both Intel CPU families share a common chipset; the interconnection is called Intel QuickPath Interconnect , which provides extremely... |
AlaaElhilo/Wikipedia_ComputerScience | Nearly all CPU architectures use a small amount of very fast non-shared memory known as cache to exploit locality of reference in memory accesses. With NUMA, maintaining cache coherence across shared memory has a significant overhead. Although simpler to design and build, non-cache-coherent NUMA systems become prohibit... |
AlaaElhilo/Wikipedia_ComputerScience | Typically, ccNUMA uses inter-processor communication between cache controllers to keep a consistent memory image when more than one cache stores the same memory location. For this reason, ccNUMA may perform poorly when multiple processors attempt to access the same memory area in rapid succession. Support for NUMA in ... |
AlaaElhilo/Wikipedia_ComputerScience | Alternatively, cache coherency protocols such as the MESIF protocol attempt to reduce the communication required to maintain cache coherency. Scalable Coherent Interface is an IEEE standard defining a directory-based cache coherency protocol to avoid scalability limitations found in earlier multiprocessor systems. For... |
AlaaElhilo/Wikipedia_ComputerScience | One can view NUMA as a tightly coupled form of cluster computing. The addition of virtual memory paging to a cluster architecture can allow the implementation of NUMA entirely in software. However, the inter-node latency of software-based NUMA remains several orders of magnitude greater than that of hardware-based NU... |
AlaaElhilo/Wikipedia_ComputerScience | Since NUMA largely influences memory access performance, certain software optimizations are needed to allow scheduling threads and processes close to their in-memory data. |
AlaaElhilo/Wikipedia_ComputerScience | As of 2011, ccNUMA systems are multiprocessor systems based on the AMD Opteron processor, which can be implemented without external logic, and the Intel Itanium processor, which requires the chipset to support NUMA. Examples of ccNUMA-enabled chipsets are the SGI Shub , the Intel E8870, the HP sx2000 , and those found ... |
AlaaElhilo/Wikipedia_ComputerScience | A scalar processor is classified as a single instruction, single data processor in Flynn's taxonomy. The Intel 486 is an example of a scalar processor. It is to be contrasted with a vector processor where a single instruction operates simultaneously on multiple data items processor). The difference is analogous to t... |
AlaaElhilo/Wikipedia_ComputerScience | The term scalar in computing dates to the 1970 and 1980s when vector processors were first introduced. It was originally used to distinguish the older designs from the new vector processors. |
AlaaElhilo/Wikipedia_ComputerScience | A superscalar processor may execute more than one instruction during a clock cycle by simultaneously dispatching multiple instructions to redundant functional units on the processor. Each functional unit is not a separate CPU core but an execution resource within a single CPU such as an arithmetic logic unit, a bit sh... |
AlaaElhilo/Wikipedia_ComputerScience | A scalar data type, or just scalar, is any non-composite value. |
AlaaElhilo/Wikipedia_ComputerScience | Generally, all basic primitive data types are considered scalar: |
AlaaElhilo/Wikipedia_ComputerScience | Some programming languages also treat strings as scalar types, while other languages treat strings as arrays or objects. |
AlaaElhilo/Wikipedia_ComputerScience | A quantum computer is a computer that takes advantage of quantum mechanical phenomena. |
AlaaElhilo/Wikipedia_ComputerScience | On small scales, physical matter exhibits properties of both particles and waves, and quantum computing leverages this behavior, specifically quantum superposition and entanglement, using specialized hardware that supports the preparation and manipulation of quantum states. |
AlaaElhilo/Wikipedia_ComputerScience | Classical physics cannot explain the operation of these quantum devices, and a scalable quantum computer could perform some calculations exponentially faster than any modern "classical" computer. In particular, a large-scale quantum computer could break widely used encryption schemes and aid physicists in performing p... |
AlaaElhilo/Wikipedia_ComputerScience | The basic unit of information in quantum computing is the qubit, similar to the bit in traditional digital electronics. Unlike a classical bit, a qubit can exist in a superposition of its two "basis" states. When measuring a qubit, the result is a probabilistic output of a classical bit, therefore making quantum comput... |
AlaaElhilo/Wikipedia_ComputerScience | Physically engineering high-quality qubits has proven challenging. If a physical qubit is not sufficiently isolated from its environment, it suffers from quantum decoherence, introducing noise into calculations. Paradoxically, perfectly isolating qubits is also undesirable because quantum computations typically need to... |
AlaaElhilo/Wikipedia_ComputerScience | In principle, a non-quantum computer can solve the same computational problems as a quantum computer, given enough time. Quantum advantage comes in the form of time complexity rather than computability, and quantum complexity theory shows that some quantum algorithms for carefully selected tasks require exponentially ... |
AlaaElhilo/Wikipedia_ComputerScience | For many years, the fields of quantum mechanics and computer science formed distinct academic communities. Modern quantum theory developed in the 1920s to explain the wave–particle duality observed at atomic scales, and digital computers emerged in the following decades to replace human computers for tedious calculatio... |
AlaaElhilo/Wikipedia_ComputerScience | A measurement-based quantum computer decomposes computation into a sequence of Bell state measurements and single-qubit quantum gates applied to a highly entangled initial state , using a technique called quantum gate teleportation. |
AlaaElhilo/Wikipedia_ComputerScience | An adiabatic quantum computer, based on quantum annealing, decomposes computation into a slow continuous transformation of an initial Hamiltonian into a final Hamiltonian, whose ground states contain the solution. |
AlaaElhilo/Wikipedia_ComputerScience | Neuromorphic quantum computing is an unconventional computing type of computing that uses neuromorphic computing to perform quantum operations. It was suggested that quantum algorithms, which are algorithms that run on a realistic model of quantum computation, can be computed equally efficiently with neuromorphic quan... |
AlaaElhilo/Wikipedia_ComputerScience | A topological quantum computer decomposes computation into the braiding of anyons in a 2D lattice. |
AlaaElhilo/Wikipedia_ComputerScience | A quantum Turing machine is the quantum analog of a Turing machine. All of these models of computation—quantum circuits, one-way quantum computation, adiabatic quantum computation, and topological quantum computation—have been shown to be equivalent to the quantum Turing machine; given a perfect implementation of one s... |
AlaaElhilo/Wikipedia_ComputerScience | Quantum computing has significant potential applications in the fields of cryptography and cybersecurity. Quantum cryptography, which relies on the principles of quantum mechanics, offers the possibility of secure communication channels that are resistant to eavesdropping. Quantum key distribution protocols, such as B... |
AlaaElhilo/Wikipedia_ComputerScience | However, quantum computing also poses challenges to traditional cryptographic systems. Shor's algorithm, a quantum algorithm for integer factorization, could potentially break widely used public-key cryptography schemes like RSA, which rely on the difficulty of factoring large numbers. Post-quantum cryptography, which ... |
AlaaElhilo/Wikipedia_ComputerScience | Ongoing research in quantum cryptography and post-quantum cryptography is crucial for ensuring the security of communication and data in the face of evolving quantum computing capabilities. Advances in these fields, such as the development of new QKD protocols, the improvement of QRNGs, and the standardization of post-... |
AlaaElhilo/Wikipedia_ComputerScience | Quantum cryptography enables new ways to transmit data securely; for example, quantum key distribution uses entangled quantum states to establish secure cryptographic keys. When a sender and receiver exchange quantum states, they can guarantee that an adversary does not intercept the message, as any unauthorized eavesd... |
AlaaElhilo/Wikipedia_ComputerScience | Modern fiber-optic cables can transmit quantum information over relatively short distances. Ongoing experimental research aims to develop more reliable hardware , hoping to scale this technology to long-distance quantum networks with end-to-end entanglement. Theoretically, this could enable novel technological applicat... |
AlaaElhilo/Wikipedia_ComputerScience | Progress in finding quantum algorithms typically focuses on this quantum circuit model, though exceptions like the quantum adiabatic algorithm exist. Quantum algorithms can be roughly categorized by the type of speedup achieved over corresponding classical algorithms. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.