source stringlengths 32 199 | text stringlengths 26 3k |
|---|---|
https://en.wikipedia.org/wiki/The%20Fool%27s%20Errand | The Fool's Errand is a 1987 computer game by Cliff Johnson. It is a meta-puzzle game with storytelling, visual puzzles, and a cryptic treasure map. It is the tale of a wandering Fool who seeks his fortune in the Land of Tarot and braves the enchantments of the High Priestess. A sequel titled The Fool and His Money was released October 25, 2012.
Plot
The plot focuses on The Fool card of the tarot, who is portrayed as a silhouette of a young man wearing a peaked, feathered cap, curled-toed shoes, and carrying a knapsack on a stick. The Fool is the protagonist of the story, and he encounters various other cards from the tarot. In the beginning of the story, The Sun gives him a map, which has been scrambled, and directs him to find the "Lost 14 Treasures of the World." The Fool journeys through four kingdoms (each representing a suit from the minor arcana of the tarot), where he encounters other characters, who either give him more information or provide him with additional tasks. The High Priestess card of the tarot is set up as the villain of the story, and all the characters he meets are other cards from the tarot. Each character is drawn as a black silhouette, as is the background art.
Gameplay
The game is structured as a storybook divided into five parts, each containing a large number of different chapters; the storybook can be paged through and read as continuous prose on screen. However, not every chapter is available at the start of the game, and those chapters which are available are not consecutive. Many chapters have a puzzle (called an enchantment) associated with them; completing such a puzzle unlocks further chapter(s). Every chapter is named after a tarot card in either the Major Arcana or the Minor Arcana.
Frequently, the puzzles are designed in such a way that the result of the puzzle leads logically into the unlocked chapter; for example, the player may complete an acrostic puzzle which results in the phrase "No Ship", which then unlocks part of the story in which a watchman indeed reports that no ship has been sighted and deals with the consequences. Other puzzles feature pictures which portray parts of the story or even clues to other puzzles.
The first chapter, The Sun, features the puzzle The Sun's Map. This is a jigsaw puzzle with one piece for every chapter in the story; each puzzle piece appears only when the appropriate chapter is unlocked. Each piece contains a symbol representing the chapter from which it came, plus part of a continuous path which flows through all pieces in the order in which they are mentioned in the narrative. Once the map is successfully completed, other designs on the map become active click targets and can be used as clues or processes to decipher the true final puzzle: The Book Of Thoth, hidden within the chapter The High Priestess, which requires the reader to peruse the entire story as continuous prose and identify a number of phrases hidden within the narrative.
Development
Cliff Johnson |
https://en.wikipedia.org/wiki/Native%20%28computing%29 | In computing, native software or data-formats are those that were designed to run on a particular operating system. In a more technical sense, native code is code written specifically for a certain processor. In contrast, cross-platform software can be run on multiple operating systems and/or computer architectures.
For example, a Game Boy receives its software through a cartridge, which contains code that runs natively on the Game Boy. The only way to run this code on another processor is to use an emulator, which simulates an actual Game Boy. This usually comes at the cost of speed.
Applications
Something running on a computer natively means that it is running without any external layer requiring fewer software layers. For example, in Microsoft Windows the Native API is an application programming interface specific for the Windows NT kernel, which can be used to give access to some kernel functions, which cannot be directly accessed through a more universal Windows API.
Operating systems
Used to denote either the absence of virtualization or virtualization at its lowest level. When various levels of virtualization take place, the lowest level operating system—the one that actually maintains direct control of the hardware—is referred to as a "Native VM," for example.
Machine code
Machine code, also known as native code, is a program which is written in machine language. Machine code is usually considered the lowest level of code for a computer, that, in its lowest level form, is written in binary (0s and 1s), but is often written in hexadecimal or octal to make it a little easier to handle. These instruction sets are then interpreted by the computer. With this, there is no need for translation. machine code is strictly numerical and usually isn't what programmers program in, due to this complex nature. Machine code is also as close as you can get to the processor, so using this language, you are programming specifically for that processor as machine code for each processor may differ. Typically programmers will code in high-level languages such as C, C++, Pascal, (or other directly compiled languages) which gets translated into assembly code, which then translates it into machine code (or in most cases the compiler generates machine code directly). Since each CPU is different, programs need to be recompiled or rewritten in order to work on that CPU.
Data
Applied to data, native data formats or communication protocols are those supported by a certain computer hardware or software, with maximal consistency and minimal amount of additional components.
For example, EGA and VGA video adapters natively support code page 437. This does not preclude supporting other code pages, but it requires either a font uploading or using graphic modes.
Cloud computing
In cloud computing, "cloud native" refers to the software approach of building, deploying, and managing modern applications in cloud computing environments, for software optimised for ru |
https://en.wikipedia.org/wiki/Concurrent%20user | In computer science, the number of concurrent users (sometimes abbreviated CCU) for a resource in a location, with the location being a computing network or a single computer, refers to the total number of people simultaneously accessing or using the resource. The resource can, for example, be a computer program, a file, or the computer as a whole.
Keeping track of concurrent users is important in several cases. First, some operating system models such as time-sharing operating systems allow several users to access a resource on the computer at the same time. As system performance may degrade due to the complexity of processing multiple jobs from multiple users at the same time, the capacity of such a system may be measured in terms of maximum concurrent users.
Second, commercial software vendors often license a software product by means of a concurrent users restriction. This allows a fixed number of users access to the product at a given time and contrasts with an unlimited user license. For example: Company X buys software and pays for 20 concurrent users. However, there are 100 logins created at implementation. Only 20 of those 100 can be in the system at the same time, this is known as floating licensing.
Concurrent user licensing allows firms to purchase computer systems and software at a lower cost because the maximum number of concurrent users expected to use the system or software at any given time (those users all logged in together) is only a portion of the total system users employed at a company. The concurrent licenses are global and shared by anyone who needs to use the system. This contrasts with "named-seats" licensing, in which one license must be purchased for each and every individual user, whether they are using the system or not.
If a company employs 400 system users in which 275 work during the day and 125 work at night, then they can opt to purchase only 275 concurrent user licenses since there will never be more than 275 users on the system during a normal work day. The night workers share 125 of the day users' licenses to use the system. For named-seat licenses, this same company would have to purchase 400 individual licenses, one for each user, and licenses would not be globally shared. The available options for licensing are entirely at the discretion of the vendor selling the product.
See also
Floating licensing
References
Computing terminology
Software licenses |
https://en.wikipedia.org/wiki/MOSIX | MOSIX is a proprietary distributed operating system. Although early versions were based on older UNIX systems, since 1999 it focuses on Linux clusters and grids. In a MOSIX cluster/grid there is no need to modify or to link applications with any library, to copy files or login to remote nodes, or even to assign processes to different nodes – it is all done automatically, like in an SMP.
History
MOSIX has been researched and developed since 1977 at The Hebrew University of Jerusalem by the research team of Prof. Amnon Barak. So far, ten major versions have been developed. The first version, called MOS, for Multicomputer OS, (1981–83) was based on Bell Lab's Seventh Edition Unix and ran on a cluster of PDP-11 computers. Later versions were based on Unix System V Release 2 (1987–89) and ran on a cluster of VAX and NS32332-based computers, followed by a BSD/OS-derived version (1991–93) for a cluster of 486/Pentium computers. Since 1999 MOSIX is tuned to Linux for x86 platforms.
MOSIX2
The second version of MOSIX, called MOSIX2, compatible with Linux-2.6 and 3.0 kernels. MOSIX2 is implemented as an OS virtualization layer that provides users and applications with a single system image with the Linux run-time environment. It allows applications to run in remote nodes as if they run locally. Users run their regular (sequential and parallel) applications while MOSIX transparently and automatically seek resources and migrate processes among nodes to improve the overall performance.
MOSIX2 can manage a cluster and a multicluster (grid) as well as workstations and other shared resources. Flexible management of a grid allows owners of clusters to share their computational resources, while still preserving their autonomy over their own clusters and their ability to disconnect their nodes from the grid at any time, without disrupting already running programs.
A MOSIX grid can extend indefinitely as long as there is trust between its cluster owners. This must include guarantees that guest applications will not be modified while running in remote clusters and that no hostile computers can be connected to the local network. Nowadays these requirements are standard within clusters and organizational grids.
MOSIX2 can run in native mode or in a virtual machine (VM). In native mode, performance is better, but it requires modifications to the base Linux kernel, whereas a VM can run on top of any unmodified operating system that supports virtualization, including Microsoft Windows, Linux and Mac OS X.
MOSIX2 is most suitable for running compute intensive applications with low to moderate amount of input/output (I/O). Tests of MOSIX2 show that the performance of several such applications over a 1 Gbit/s campus grid is nearly identical to that of a single cluster.
Main features
Provides aspects of a single-system image:
Users can login on any node and do not need to know where their programs run.
No need to modify or link applications with special libraries |
https://en.wikipedia.org/wiki/.exe | .exe is a common filename extension denoting an executable file (the main execution point of a computer program) for Microsoft Windows, OS/2, and DOS.
File formats
There are numerous file formats which may be used by a file with a extension.
DOS
16-bit DOS MZ executable (MZ)The original DOS executable file format. These formats can be identified by the letters "MZ" at the beginning of the file in ASCII. All later formats have an MZ DOS stub header.
16-bit New Executable (NE) Introduced with the multitasking MS-DOS 4.0 and also used by 16-bit OS/2 and Windows, NE can be identified by the "NE" in ASCII.
OS/2
32-bit Linear Executable (LX) Introduced with OS/2 2.0, these can be identified by the "LX" in ASCII. These can only be run by OS/2 2.0 and higher.
Mixed 16/32-bit Linear Executable (LE) Introduced with OS/2 2.0, these can be identified by the "LE" in ASCII. VxD drivers on Windows 3.x, OS/2, and Windows 9x.
Windows
32-bit Portable Executable (PE) Introduced with Windows NT, they are fat binaries consisting of a DOS-specific and a Windows-specific part. The DOS-specific part (dubbed DOS stub) is a legitimate 16-bit DOS program. Microsoft C++ linker, by default, uses a minimal DOS stub that prints the following message: "This program cannot be run in DOS mode." Windows ignores the DOS stub and executes the Windows-specific portion that starts with the "PE\0\0" ASCII sequence (letters "PE" and two null bytes). With some linkers, it is possible to specify a custom DOS stub. Indeed, there are a few dual programs, such as regedit in Windows 95 and old versions of WinZIP self extractors.
64-bit Portable Executable (PE32+) Introduced by 64-bit versions of Windows, this is a PE file with wider fields. In most cases, code can be written to simply work as either a 32 or 64-bit PE file. This file also includes a DOS stub.
Other
There are other EXE formats, including but not limited to W3 (a collection of LE files, only used in WIN386.EXE), W4 (a compressed collection of LE files, only used in VMM32.VXD), DL, MP, P2, P3 (last three used by Phar Lap extenders).
See also
Comparison of executable file formats
Executable compression
IExpress
CMD file (CP/M)
Windows Installer files (msi)
References
Further reading
External links
Dependency Walker
MZ EXE header format
PE Explorer
DOS files
DOS technology
Executable file formats
Filename extensions
Windows administration |
https://en.wikipedia.org/wiki/Scopus | Scopus is Elsevier's abstract and citation database launched in 2004. Scopus covers nearly 36,377 titles (22,794 active titles and 13,583 inactive titles) from approximately 11,678 publishers, of which 34,346 are peer-reviewed journals in top-level subject fields: life sciences, social sciences, physical sciences and health sciences. It covers three types of sources: book series, journals, and trade journals. All journals covered in the Scopus database are reviewed for sufficiently high quality each year according to four types of numerical quality measure for each title; those are h-Index, CiteScore, SJR (SCImago Journal Rank) and SNIP (source normalized impact per paper). Scopus also allows patent searches in a dedicated patent database Lexis-Nexis, albeit with a limited functionality.
Journals listed in Scopus are considered to be meeting the requirement for peer review quality established by several research grant agencies for their grant recipients and by degree accreditation boards in numerous countries.
Overview
Comparing ease of use and coverage of Scopus and the Web of Science (WOS), a 2006 study concluded that "Scopus is easy to navigate, even for the novice user. ... The ability to search both forward and backward from a particular citation would be very helpful to the researcher. The multidisciplinary aspect allows the researcher to easily search outside of his discipline" and "One advantage of WOS over Scopus is the depth of coverage, with the full WOS database going back to 1945 and Scopus going back to 1966. However, Scopus and WOS complement each other as neither resource is all-inclusive." A small number of studies found ca. 80-90% overlap in coverage between WoS and Scopus for the period between 1990 and 2020.<refs>
In terms of the structured query language search capabilities Scopus is somewhat more advanced than Web of Science: for example, WoS can perform only NEAR/n queries, Scopus can also do PRE/n queries.
Also, when the same article is covered in Scopus and in the Web of Science (WoS), its Scopus entry has a 3-5 larger number of keywords than its WoS coverage, and the Scopus keywords are more focused on the specific article content, whereas WoS has more keywords related to the broad category of the article's subject. A larger number of narrow-targeted keywords allows Scopus users to find a larger number of relevant publications, while filtering out false positives. On the other hand, WoS exports (e.g. in the ris format) the doi numbers of cited articles, while Scopus exports the titles of cited articles. Also, Scopus allows exporting 20,000 references (e.g. as a ris file) at once, while WoS export is limited to 5,000 references at once. On the other hand, WoS exports the doi's of cited references, while Scopus exports the titles of cited references.
Scopus provides chemical search by CAS number and by chemical name, while WoS does not have these features. On the other hand, WoS has chemical structure search, but |
https://en.wikipedia.org/wiki/Quadratic%20sieve | The quadratic sieve algorithm (QS) is an integer factorization algorithm and, in practice, the second-fastest method known (after the general number field sieve). It is still the fastest for integers under 100 decimal digits or so, and is considerably simpler than the number field sieve. It is a general-purpose factorization algorithm, meaning that its running time depends solely on the size of the integer to be factored, and not on special structure or properties. It was invented by Carl Pomerance in 1981 as an improvement to Schroeppel's linear sieve.
Basic aim
The algorithm attempts to set up a congruence of squares modulo n (the integer to be factorized), which often leads to a factorization of n. The algorithm works in two phases: the data collection phase, where it collects information that may lead to a congruence of squares; and the data processing phase, where it puts all the data it has collected into a matrix and solves it to obtain a congruence of squares. The data collection phase can be easily parallelized to many processors, but the data processing phase requires large amounts of memory, and is difficult to parallelize efficiently over many nodes or if the processing nodes do not each have enough memory to store the whole matrix. The block Wiedemann algorithm can be used in the case of a few systems each capable of holding the matrix.
The naive approach to finding a congruence of squares is to pick a random number, square it, divide by n and hope the least non-negative remainder is a perfect square. For example, . This approach finds a congruence of squares only rarely for large n, but when it does find one, more often than not, the congruence is nontrivial and the factorization is complete. This is roughly the basis of Fermat's factorization method.
The quadratic sieve is a modification of Dixon's factorization method.
The general running time required for the quadratic sieve (to factor an integer n) is
in the L-notation.
The constant e is the base of the natural logarithm.
The approach
To factorize the integer n, Fermat's method entails a search for a single number a, , such that the remainder of a2 divided by n is a square. But these a are hard to find. The quadratic sieve consists of computing the remainder of a2/n for several a, then finding a subset of these whose product is a square. This will yield a congruence of squares.
For example, consider attempting to factor the number 1649. We have: . None of the integers is a square, but the product is a square. We also had
since .
The observation that thus gives a congruence of squares
Hence for some integer . We can then factor
using the Euclidean algorithm to calculate the greatest common divisor.
So the problem has now been reduced to: given a set of integers, find a subset whose product is a square. By the fundamental theorem of arithmetic, any positive integer can be written uniquely as a product of prime powers. We do this in a vector format; for examp |
https://en.wikipedia.org/wiki/Luhn%20algorithm | The Luhn algorithm or Luhn formula, also known as the "modulus 10" or "mod 10" algorithm, named after its creator, IBM scientist Hans Peter Luhn, is a simple check digit formula used to validate a variety of identification numbers.
It is described in U.S. Patent No. 2,950,048, granted on August 23, 1960.
The algorithm is in the public domain and is in wide use today. It is specified in ISO/IEC 7812-1. It is not intended to be a cryptographically secure hash function; it was designed to protect against accidental errors, not malicious attacks. Most credit cards and many government identification numbers use the algorithm as a simple method of distinguishing valid numbers from mistyped or otherwise incorrect numbers.
Description
The check digit is computed as follows:
If the number already contains the check digit, drop that digit to form the "payload". The check digit is most often the last digit.
With the payload, start from the rightmost digit. Moving left, double the value of every second digit (including the rightmost digit).
Sum the values of the resulting digits.
The check digit is calculated by , where s is the sum from step 3. This is the smallest number (possibly zero) that must be added to to make a multiple of 10. Other valid formulas giving the same value are , , and . Note that the formula will not work in all environments due to differences in how negative numbers are handled by the modulo operation.
Example for computing check digit
Assume an example of an account number 1789372997 (just the "payload", check digit not yet included):
The sum of the resulting digits is 56.
The check digit is equal to .
This makes the full account number read 17893729974.
Example for validating check digit
Drop the check digit (last digit) of the number to validate. (e.g. 17893729974 → 1789372997)
Calculate the check digit (see above)
Compare your result with the original check digit. If both numbers match, the result is valid.
Strengths and weaknesses
The Luhn algorithm will detect all single-digit errors, as well as almost all transpositions of adjacent digits. It will not, however, detect transposition of the two-digit sequence 09 to 90 (or vice versa). It will detect most of the possible twin errors (it will not detect 22 ↔ 55, 33 ↔ 66 or 44 ↔ 77).
Other, more complex check-digit algorithms (such as the Verhoeff algorithm and the Damm algorithm) can detect more transcription errors. The Luhn mod N algorithm is an extension that supports non-numerical strings.
Because the algorithm operates on the digits in a right-to-left manner and zero digits affect the result only if they cause shift in position, zero-padding the beginning of a string of numbers does not affect the calculation. Therefore, systems that pad to a specific number of digits (by converting 1234 to 0001234 for instance) can perform Luhn validation before or after the padding and achieve the same result.
The algorithm appeared in a United States Patent for a |
https://en.wikipedia.org/wiki/MLAN | mLAN, short for Music Local Area Network, is a protocol for synchronized transmission and management of multi-channel digital audio, video, control signals and multi-port MIDI over a network.
Description
The mLAN protocol was originally developed by Yamaha Corporation, and publicly introduced in January 2000. It was available under a royalty-free license to anyone interested in utilizing the technology.
mLAN uses several features of the IEEE 1394 (FireWire) standard such as isochronous transfer and intelligent connection management. There are two versions of the mLAN protocol. Version 1 requires S200 rate, while Version 2 requires S400 rate and supports synchronized streaming of digital audio at up to 24 bit word length and 192 kHz sample rate, MIDI and word clock at a bitrate up to 400 Megabits per second.
With the proper driver software, a computer-based digital audio workstation can interact with mLAN-compliant hardware via any OHCI-compliant FireWire port. mLAN consumes the entire bus bandwidth when operating and non-mLAN devices cannot share the same Firewire connection, so it is not possible to use hard drives, optical drives or other sound devices on the same Firewire bus when mLAN Manager software is running.
The transport layers of mLAN have been standardized as IEC 61883.
End of life
By 2005, over 100 manufacturers were part of the mLAN Alliance however very few actual products had surfaced.
As of early 2008, mLAN appeared to have reached the end of its product life. Third-party developers discontinued or retracted their mLAN products from the market, and Yamaha itself ceased any new releases of mLAN hardware or updates to the mLAN software and drivers. Even though more recent FireWire based products from Yamaha could interoperate with earlier mLAN devices using a computer, any mention of mLAN is notably absent from new product announcements and driver updates.
Products
Yamaha 01X digital mixing hub
Yamaha i88x audio/MIDI interface
Yamaha mLAN8E Expansion Board for Yamaha A5000/A4000 samplers and Yamaha CS6x/CS6R/S80 synthesizers
Yamaha mLAN16E/mLAN16E2 options for Yamaha Motif, Yamaha Motif ES, Yamaha S90 ES, Yamaha Motif XS, Yamaha Motif-Rack XS
Yamaha MY16-mLAN option for Yamaha digital mixers
Korg EXB-mLAN option for Korg Triton Rack and Korg Triton Studio
Presonus FIREStation audio/MIDI interface
Kurzweil KMLN8 8 Channel mLAN I/O option for KSP-8 Multi-Channel effects processor
See also
ADAT
Audio over Ethernet
MADI
MIDI
Open Sound Control
RTP MIDI
ZIPI
References
External links
mLAN SCAM forum without members, even though it claims to have many
mLAN Central
01xray
Save Mlan on the Mac hosted by Chrismcc
https://www.kvraudio.com/forum/viewtopic.php?t=505869
Digital audio
Audio network protocols |
https://en.wikipedia.org/wiki/VGA%20connector | The Video Graphics Array (VGA) connector is a standard connector used for computer video output. Originating with the 1987 IBM PS/2 and its VGA graphics system, the 15-pin connector went on to become ubiquitous on PCs, as well as many monitors, projectors and high-definition television sets.
Other connectors have been used to carry VGA-compatible signals, such as mini-VGA or BNC, but "VGA connector" typically refers to this design.
Devices continue to be manufactured with VGA connectors, although newer digital interfaces such as DVI, HDMI and DisplayPort are increasingly displacing VGA, and many modern computers and other devices do not include it.
Physical design
The VGA connector is a three-row, 15-pin D-subminiature connector referred to variously as DE-15, HD-15 or erroneously DB-15(HD). DE-15 is the accurate nomenclature under the D-sub specifications: an "E" size D-sub connector, with 15 pins in three rows.
Electrical design
All VGA connectors carry analog RGBHV (red, green, blue, horizontal sync, vertical sync) video signals. Modern connectors also include VESA DDC pins, for identifying attached display devices.
In both its modern and original variants, VGA utilizes multiple scan rates, so attached devices such as monitors are multisync by necessity.
The VGA interface includes no affordances for hot swapping, the ability to connect or disconnect the output device during operation, although in practice this can be done and usually does not cause damage to the hardware or other problems. The VESA DDC specification does, however, include a standard for hot-swapping.
PS/2 signaling
In the original IBM VGA implementation, refresh rates were limited to two vertical (60 and 70 Hz) and three horizontal frequencies, all of which were communicated to the monitor using combinations of different polarity H and V sync signals.
Some pins on the connector were also different: pin 9 was keyed by plugging the female connector hole, and four pins carried the monitor ID.
With the implementation of the VESA DDC specification, several of the monitor ID pins were reassigned for use by DDC signaling, and the key pin was replaced with a +5 V DC output per the DDC spec. Devices that comply with the DDC host system standard provide , from 50mA to 1A.
PS/55 signaling
The IBM PS/55 Display Adapter redefined pin 9 as "+12V", which signals the monitor to turn on when the system unit is powered on.
EDID
In order to advertise display capabilities VESA has introduced a scheme to redefining VGA connector pins 9, 12, and 15 as a serial bus for a Display Data Channel (DDC).
Cable quality
The same VGA cable can be used with a variety of supported VGA resolutions, ranging from 320×400px @70 Hz, or 320×480px @60 Hz (12.6 MHz of signal bandwidth) to 1280×1024px (SXGA) @85 Hz (160 MHz) and up to 2048×1536px (QXGA) @85 Hz (388 MHz).
There are no standards defining the quality required for each resolution, but higher-quality cables typically contain coaxial wir |
https://en.wikipedia.org/wiki/Andrew%20Graham%20%28astronomer%29 | Andrew Graham (8 April 1815 – 5 November 1908), born in Irvinestown County Fermanagh, Ireland, was an Irish astronomer, orbit computer and discoverer of the asteroid 9 Metis.
Astronomer at Markree, County Sligo
In 1842 Graham was appointed to work at Markree Observatory in County Sligo in northwest Ireland. The observatory had been established in 1830 by Colonel Edward Joshua Cooper (1798–1863) as a private institution on his country estate. Cooper equipped it with excellent astronomical instruments. Graham proved an energetic observer at Markree.
Graham discovered the asteroid 9 Metis from Markree on 25 April 1848 while observing with a 3-inch aperture wide-field telescope manufactured for comet searching by the German instrument maker Ertel.
Only eight minor planets were known before then, with the first four having been found in the period 1801–1807, and Graham's discovery consequently attracted considerable attention. His mathematical abilities allowed him to compute the orbit of the minor planet and to predict its apparent position into the future, including the gravitational effects of planets on its orbit around the Sun.
Graham later worked on the Markree Catalogue, which consisted of observations of about sixty thousand stars along the ecliptic taken between 8 August 1848 and 27 March 1856, and was published in four volumes in the years 1851, 1853, 1854 and 1856 respectively. Whilst working on this, he developed the square-bar micrometer, which greatly increased the efficiency of determining positions (right ascension and declination) of celestial objects.
Graham calculated the orbits of a large number of comets. These were used in his employer's book Cometic Orbits.
Graham resigned from his position at Markree and began new employment in Cambridge in March 1864.
Assistant at the Cambridge Observatory
Graham worked as First Assistant at the Cambridge Observatory, England, from 1864 to 1903.
There he worked on the Cambridge Zone Catalogue, the Observatory's contribution to the catalogue of stars brighter than magnitude 9.5 organised by the Astronomische Gesellschaft in Germany. This was in many ways an extension of his work at Markree, but for stars some distance to the north of the ecliptic. The catalogue of 14,464 stars between declinations 24° and 31° north was published in 1897.
Graham collaborated with Anne Walker in the observing, who, although employed as a `lady computer' to perform routine mathematical calculations, regularly shared the observing tasks with Graham in a successful professional partnership.
Graham retired from his post at the Cambridge Observatory in 1903 at the age of 88 years. He was granted a pension by the University of Cambridge. He died in Cambridge in 1908 aged 93 years.
References
1815 births
1908 deaths
19th-century British astronomers
Discoverers of asteroids
Astronomers from Northern Ireland
Scientists from County Fermanagh
Recipients of the Lalande Prize |
https://en.wikipedia.org/wiki/Trade%20route | A trade route is a logistical network identified as a series of pathways and stoppages used for the commercial transport of cargo. The term can also be used to refer to trade over bodies of water. Allowing goods to reach distant markets, a single trade route contains long-distance arteries, which may further be connected to smaller networks of commercial and noncommercial transportation routes. Among notable trade routes was the Amber Road, which served as a dependable network for long-distance trade. Maritime trade along the Spice Route became prominent during the Middle Ages, when nations resorted to military means for control of this influential route. During the Middle Ages, organizations such as the Hanseatic League, aimed at protecting interests of the merchants and trade became increasingly prominent.
In modern times, commercial activity shifted from the major trade routes of the Old World to newer routes between modern nation-states. This activity was sometimes carried out without traditional protection of trade and under international free-trade agreements, which allowed commercial goods to cross borders with relaxed restrictions. Innovative transportation of modern times includes pipeline transport and the relatively well-known trade involving rail routes, automobiles, and cargo airlines.
History
Development of early routes
Early development
Long-distance trade routes were developed in the Chalcolithic period. The period from the middle of the 2nd millennium BCE to the beginning of the Common Era saw societies in Southeast Asia, Western Asia, the Mediterranean, China, and the Indian subcontinent develop major transportation networks for trade.
One of the vital instruments which facilitated long-distance trade was portage and the domestication of beasts of burden. Organized caravans, visible by the 2nd millennium BCE, could carry goods across a large distance as fodder was mostly available along the way. The domestication of camels allowed Arabian nomads to control the long-distance trade in spices and silk from the Far East to the Arabian Peninsula. Caravans were useful in long-distance trade largely for carrying luxury goods, the transportation of cheaper goods across large distances was not profitable for caravan operators. With productive developments in iron and bronze technologies, newer trade routes – dispensing innovations of civilizations – began to rise.
Maritime trade
Navigation was known in Sumer between the 4th and the 3rd millennium BCE. The Egyptians had trade routes through the Red Sea, importing spices from the "Land of Punt" (East Africa) and from Arabia.
In Asia, the earliest evidence of maritime trade was the Neolithic trade networks of the Austronesian peoples among which is the lingling-o jade industry of the Philippines, Taiwan, southern Vietnam and peninsular Thailand. It also included the long-distance routes of Austronesian traders from Indonesia and Malaysia connecting China with South Asia and the Midd |
https://en.wikipedia.org/wiki/List%20of%20radio%20stations%20in%20Colorado | The following is a list of FCC-licensed radio stations in the U.S. State of Colorado, which can be sorted by their call signs, frequencies, cities of license, licensees, and programming formats.
List of radio stations
Defunct stations
KAMV-LP
See also
Bibliography of Colorado
Geography of Colorado
History of Colorado
Index of Colorado-related articles
List of Colorado-related lists
Outline of Colorado
References
External links
www.radiomap.us – List of radio stations in Denver, Colorado
Colorado-related lists
Colorado, List of radio stations in |
https://en.wikipedia.org/wiki/List%20of%20radio%20stations%20in%20Kentucky | The following is a list of FCC-licensed radio stations in the U.S. state of Kentucky, which can be sorted by their call signs, frequencies, cities of license, licensees, and programming formats.
List of radio stations
Defunct
WAIA
WANY
WBLG-LP
WCPM
WCWC
WCYN-FM
WEKC (Williamsburg, Kentucky)
WENS-LP
WFHS-LP
WFLE
WFUL
WGRK
WIAR
WKVG
WKYD-LP
WKYR
WLBJ
WLGC
WLKS
WMMG
WMOR
WMTC
WQFR-LP
WQXY
WRLV
WRSL
WSMJ
WWLK
WYAH-LP
See also
Kentucky media
List of newspapers in Kentucky
List of television stations in Kentucky
Media of cities in Kentucky: Bowling Green, Lexington, Louisville
References
Bibliography
(About WHAS and early radio in general)
External links
Kentucky Broadcasters Association
Images
Kentucky
Radio |
https://en.wikipedia.org/wiki/Function%20type | In computer science and mathematical logic, a function type (or arrow type or exponential) is the type of a variable or parameter to which a function has or can be assigned, or an argument or result type of a higher-order function taking or returning a function.
A function type depends on the type of the parameters and the result type of the function (it, or more accurately the unapplied type constructor , is a higher-kinded type). In theoretical settings and programming languages where functions are defined in curried form, such as the simply typed lambda calculus, a function type depends on exactly two types, the domain A and the range B. Here a function type is often denoted , following mathematical convention, or , based on there existing exactly (exponentially many) set-theoretic functions mappings A to B in the category of sets. The class of such maps or functions is called the exponential object. The act of currying makes the function type adjoint to the product type; this is explored in detail in the article on currying.
The function type can be considered to be a special case of the dependent product type, which among other properties, encompasses the idea of a polymorphic function.
Programming languages
The syntax used for function types in several programming languages can be summarized, including an example type signature for the higher-order function composition function:
When looking at the example type signature of, for example C#, the type of the function is actually Func<Func<A,B>,Func<B,C>,Func<A,C>>.
Due to type erasure in C++11's std::function, it is more common to use templates for higher order function parameters and type inference (auto) for closures.
Denotational semantics
The function type in programming languages does not correspond to the space of all set-theoretic functions. Given the countably infinite type of natural numbers as the domain and the booleans as range, then there are an uncountably infinite number (2ℵ0 = c) of set-theoretic functions between them. Clearly this space of functions is larger than the number of functions that can be defined in any programming language, as there exist only countably many programs (a program being a finite sequence of a finite number of symbols) and one of the set-theoretic functions effectively solves the halting problem.
Denotational semantics concerns itself with finding more appropriate models (called domains) to model programming language concepts such as function types. It turns out that restricting expression to the set of computable functions is not sufficient either if the programming language allows writing non-terminating computations (which is the case if the programming language is Turing complete). Expression must be restricted to the so-called continuous functions (corresponding to continuity in the Scott topology, not continuity in the real analytical sense). Even then, the set of continuous function contains the parallel-or function, which cannot be |
https://en.wikipedia.org/wiki/HP%202100 | The HP 2100 is a series of 16-bit minicomputers that were produced by Hewlett-Packard (HP) from the mid-1960s to early 1990s. Tens of thousands of machines in the series were sold over its twenty-five year lifetime, making HP the fourth largest minicomputer vendor during the 1970s.
The design started at Data Systems Inc (DSI), and was originally known as the DSI-1000. HP purchased the company in 1964 and merged it into their Dymec division. The original model, the 2116A built using integrated circuits and magnetic-core memory, was released in 1966. Over the next four years, models A through C were released with different types of memory and expansion, as well as the cost-reduced 2115 and 2114 models. All of these models were replaced by the HP 2100 series in 1971, and then again as the 21MX series in 1974 when the magnetic-core memory was replaced with semiconductor memory.
All of these models were also packaged as the HP 2000 series, combining a 2100-series machine with optional components in order to run the BASIC programming language in a multi-user time sharing fashion. HP Time-Shared BASIC was popular in the 1970s, and many early BASIC programs were written on or for the platform, most notably the seminal Star Trek that was popular during the early home computer era. The People's Computer Company published their programs in HP 2000 format.
The introduction of the HP 3000 in 1974 provided high-end competition to the 2100 series; the entire line was renamed as the HP 1000 in 1977 and positioned as real-time computers. A greatly redesigned version was introduced in 1979 as the 1000 L-Series, using CMOS large scale integration chips and introducing a desk-side tower case model. This was the first version to break backward compatibility with previous 2100-series expansion cards. The final upgrade was the A-series, with new processors capable of more than 1 MIPS performance, with the final A990 released in 1990.
History
Origins
HP formed Dynac in 1956 to act as a development shop for projects the main company would not normally undertake. Their original logo was simply the HP logo turned upside down, forming something approximating "dy" and thus inspiring the name. Learning that Westinghouse owned a trademark on that name, in 1958 they changed it to Dymec. The company was brought in-house in 1959 to become the Dymec Division, and in November 1967 was renamed the Palo Alto Division.
Dymec originally made a variety of products for the HP family, but over time became primarily an integrator, building test equipment and similar systems that were used by HP. In 1964, Kay Magleby and Paul Stoft began experimenting with the use of PDP-5 and PDP-8 computers to act as controllers for their complex test systems, but they felt the machines would require changes to truly suit their needs. At the time, Digital Equipment Corporation (DEC) was still a small company and a takeover target. David Packard found Ken Olsen too difficult to deal with, and such |
https://en.wikipedia.org/wiki/Nova%20Entertainment | NOVA Entertainment is an Australian entertainment company that operates commercial radio networks (and associated digital brands) in metropolitan and regional areas of Australia.
NOVA Entertainment is owned by Lachlan Murdoch's Illyria Pty Ltd which purchased the remaining 50% it did not own from Daily Mail and General Trust in September 2012.
History
Nova Entertainment was formed in early 1996 under the name DMG Radio Australia, followed later in the year with the A$46M acquisition of 85% of the Broadcast Media group of regional stations. This was shortly followed by FIVEaa on 26 September 1996. The founding Chairman and CEO were Charlie Cox and Paul Thompson respectively. The company grew rapidly to over 64 radio stations across New South Wales, Victoria, South Australia, Queensland and Western Australia under the Star FM, Hot FM, Nova and Vega (now Smooth) brands.
In September 2004, DMG Radio Australia sold most of its regional radio stations to Macquarie Bank, keeping only Hot 91 Sunshine Coast (later sold in early 2005) and Star 104.5 in Gosford, which they still own and operate.
In April 2009, DMG Radio Australia launched two new digital radio stations: NovaNation, a 24/7 digital dance station and Koffee, a chillout music station. These stations were launched as part of the introduction of DAB+ digital broadcasting in Australia's major state capital cities. On 24 December 2013, NovaNation and Koffee both ceased broadcasting as DAB+ stations. Koffee was rebranded to smooth Chill with a chill easy listening format and NovaNation ceased broadcasting online by mid-2016.
On Friday 12 March 2010 Vega rebranded with the name Classic Rock to counteract similar stations Gold 104.3, WSFM, Triple M Sydney and Triple M Melbourne. Nova Entertainment announced the change to Vega's brand as the stations had struggled towards the bottom of the ratings ladder in both Sydney and Melbourne. In building these new stations DMG Radio needed a "simple, focused music concept that would appeal to the 35-54 audience," according to CEO Cathy O'Connor. The stations later became 91.5 FM in Melbourne and 95.3 FM in Sydney.
On Monday 21 May 2012, after poor ratings, 91.5 FM Melbourne and 95.3 FM Sydney were rebranded again to smoothfm with a new smooth, easy listening format to aim for women aged 35+.
In December 2013, DMG Radio Australia launched a 24-hour Australian pay television music channel, Smooth available via Foxtel satellite and cable services. The pay television music channel extends from the smoothfm radio brand. It was shut down on 30 June 2020.
In February 2014, DMG Radio Australia rebranded under the name Nova Entertainment to reflect the company's increasingly diversified interests in radio and the media and entertainment industries. The new company name has been developed to encapsulate both the pride and heritage of the company’s original and successful FM launch brand, Nova, with its recent evolution into the broader media and entertainment s |
https://en.wikipedia.org/wiki/Game%20Over%20%28TV%20series%29 | Game Over is a 2004 American adult computer-animated sitcom created by David Sacks (who later became president of Nickelodeon Animation Studio), produced by Carsey-Werner Productions, and broadcast on UPN in 2004. It was cancelled after five episodes.
Game Over (the title was inspired by the phrase "game over" that commonly concludes video games) focused on what happens to video game characters after the game ends. It recounted the lives of the Smashenburns, a far-from-ordinary suburban family that lived in an alternate video game universe.
The show made numerous references to video games and even featured certain game characters as cameos. For example, Crash Bandicoot appears on a Got Milk? billboard, and creatures from Oddworld: Abe's Oddysee appear in one of the episodes.
Marisa Tomei voiced the character of Raquel Smashenburn in the series' unaired pilot episode, but scheduling problems saw Lucy Liu take over the role for the actual series.
Game Over was heavily hyped by UPN before its debut. It generally received positive press on its airing. Despite this, only six episodes were made, which aired on a variety of different days – the fourth and fifth episodes were broadcast on April 2, 2004, and the sixth episode ("Monkey Dearest") was not aired.
Characters
Rip Smashenburn, a Grand Prix driver who races every day and crashes his car all the time (voiced by Patrick Warburton)
Raquel Smashenburn (née Grunkmeyer), a covert agent who fights monsters and is vaguely similar to Lara Croft (voiced by Lucy Liu)
Alice Smashenburn, a cynical yet socially conscious 14-year-old, who is part of a beach volleyball team, but wears a hoodie instead of a bikini (voiced by Rachel Dratch)
Billy Smashenburn, a 13-year-old hoping to become a rapper (voiced by Elizabeth Daily)
Turbo, their pet, a 300-pound talking doglike creature who robs pawn shops and smokes cigars (voiced by Artie Lange)
The Changs, including Dark Princess Chang (Marie Matiko), a family of Kung-Fu-fighting Shaolin monks who live next door
Another Next door neighbour who talks to Rip in first person perspective
Episodes
Home media
Anchor Bay Entertainment released all six episodes of the series as Game Over: The Complete Series on DVD in North America on June 28, 2005.
References
External links
Official UPN website for Game Over
Carsey-Werner Game Over
2000s American adult animated television series
2000s American sitcoms
2004 American television series debuts
2004 American television series endings
American adult animated comedy television series
American adult computer-animated television series
American animated sitcoms
English-language television shows
Television series by Carsey-Werner Productions
UPN original programming
Animated television series about families |
https://en.wikipedia.org/wiki/ABC%20800 | The Luxor ABC 800 series are office versions of the ABC 80 home computer. They featured an enhanced BASIC interpreter, a slightly faster clocked CPU and more memory: 32 kilobytes RAM and 32 KB ROM was now standard, the Z80 is clocked at (quarter the 12 MHz crystal). It featured 40×24 text mode with eight colors (ABC 800 C) or 80×24 text mode monochrome (ABC 800 M). They could also be extended with "high" resolution graphics (240×240 pixels at 2 ) using RAM as video memory.
Models
ABC 800
The ABC 800 came in a monochrome version with amber text on a brown background with an 80 character wide screen, and a color version with 40 characters. The main board is integrated with the keyboard, much like the Amiga 500. However, the ABC computer has a very sturdy metal chassis.
Storage is usually two 5.25" floppy disk units in 160, 320 or 640 KB capacity. External hard disk systems became available later (primarily the ABC 850 with 10 MB). Model numbers 'ABC 800 M' for monochrome and 'ABC 800 C' for color.
Luxor advertising asked, "Who needs IBM-compatibility?" However, most computer buyers eventually considered it a requirement. A certain degree of compatibility between the ABC and IBM PC platforms could be achieved with the help of a program called 'W ABC'.
The ABC 800 computer was also sold by Facit under the name Facit DTC.
ABC 802
The ABC 802 is a compact version with 64 KB RAM where 32 KB is used as a RAM disk. The main board is integrated with a 9" CRT screen and has improved graphics, though no high-resolution graphics. Luxor ABC 802 was a model with a small monochrome screen in yellow phosphor, intended for offices. Here with two 5.25 inch disk drives along the side of the display. The grey-brown color was common for all ABC 800 (and ABC 1600) products and was different from the beige ABC 80.
ABC 806
The ABC 806 is a version with main board, screen (DA-15) and keyboard (DIN-7) as separate units. It has RAM where is used as a RAM disk, as well as more advanced 512x240x16 graphics.
Performance
In order to see how the ABC 800 would compare to other contemporary personal computers, in 1982, the Swedish magazine MikroDatorn used the Rugg/Feldman benchmarks of eight short BASIC programs (BM1 - BM8) defined by the American Kilobaud Magazine and routinely used by the British magazine Personal Computer World for testing new machines.
The result was that ABC 800's semi-compiling BASIC interpreter turned out to be faster than most other BASICs used in popular machines, especially when integer variables are used, the results for some well known computers were as follows (times in seconds):
BM1 BM2 BM3 BM4 BM5 BM6 BM7 BM8
ABC 800 (integer) not measured - see ABC 80 for approximate numbers
ABC 800 (single precision) 0.9 1.8 6.0 5.9 6.3 11.6 19.6 29
ABC 800 (double precision) 1.2 2.2 10.0 10.6 11.0 17.8 26.4 144
IBM PC 1.5 5.2 12.1 12.6 13.6 |
https://en.wikipedia.org/wiki/David%20J.%20Farber | David J. Farber (born April 17, 1934) is a professor of computer science, noted for his major contributions to programming languages and computer networking who is currently the distinguished professor and co-director of Cyber Civilization Research Center at Keio University in Japan. He has been called the "grandfather of the Internet".
Career
Farber graduated from the Stevens Institute of Technology with a M.E. degree in electrical engineering in 1956 and a second M.S. degree in mathematics in 1961. He then began an 11-year career at Bell Laboratories, where he helped design the first electronic switching system (ESS-1) and the SNOBOL programming languages. He subsequently held industry positions at the Rand Corporation and Scientific Data Systems, followed by academic positions at the University of California, Irvine, the University of Delaware, and Carnegie Mellon University. He was awarded an honorary doctorate in engineering from the Stevens Institute in 1999.
At Irvine his research work was focused on creating the world's first operational distributed computer system. While a member of the electrical engineering department of the University of Delaware, he helped conceive and organize the major American research networks CSNET, NSFNet, and the National Research and Education Network (NREN). He helped create the NSF/DARPA-funded Gigabit Network Test bed Initiative and served as the Chairman of the Gigabit Test bed Coordinating Committee.
Farber subsequently was appointed Alfred Fitler Moore Professor of Telecommunication Systems at the University of Pennsylvania, where he also held appointments as professor of business and public policy at the Wharton School of Business, and as a faculty associate of the Annenberg School for Communication. Farber served as chief technologist at the US Federal Communications Commission (2000–2001) while on leave from the university.
Farber is a founding editor of ICANNWatch. He serves on the board of advisers of Context Relevant and The Liquid Information Company.
Honors and community service
Farber is an AAAS Fellow, IEEE Fellow, ACM Fellow, and recipient of the 1995 SIGCOMM Award for lifelong contributions to computer communications. He has served on the board of directors of the Electronic Frontier Foundation, the Electronic Privacy Information Center advisory board, the board of trustees of the Internet Society, and as a member of the Presidential Advisory Committee on High Performance Computing and Communications, Information Technology and Next Generation Internet. He runs a large (25,000+ readership) mailing list called Interesting-People. In 2012, in memory of his son, he established the Joseph M. Farber prize at the Stevens Institute of Technology, which recognizes a graduating senior majoring in one of the disciplines of the College of Arts and Letters who displays a keen interest in and concern for civil liberties and their importance in preserving and protecting human rights.
On August 3, |
https://en.wikipedia.org/wiki/List%20of%20Commodore%2064%20games | This is a list of games for the Commodore 64 personal computer system, sorted alphabetically. See Lists of video games for other platforms.
Because of the length of the list, it has been broken down to two parts:
List of Commodore 64 games (A–M)
List of Commodore 64 games (N–Z)
See also
Commodore 64 Games System
Commodore 64 |
https://en.wikipedia.org/wiki/Source%20port | A source port is a software project based on the source code of a game engine that allows the game to be played on operating systems or computing platforms with which the game was not originally compatible.
Description
Source ports are often created by fans after the original developer hands over the maintenance support for a game by releasing its source code to the public (see List of commercial video games with later released source code). In some cases, the source code used to create a source port must be obtained through reverse engineering, in situations where the original source was never formally released by the game's developers. The term was coined after the release of the source code to Doom. Due to copyright issues concerning the sound library used by the original DOS version, id Software released only the source code to the Linux version of the game. Since the majority of Doom players were DOS users the first step for a fan project was to port the Linux source code to DOS. A source port typically only includes the engine portion of the game and requires that the data files of the game in question already be present on users' systems.
Source ports share the similarity with unofficial patches that both don't change the original gameplay as such projects are by definition mods. However many source ports add support for gameplay mods, which is usually optional (e.g. DarkPlaces consists of a source port engine and a gameplay mod that are even distributed separately). While the primary goal of any source port is compatibility with newer hardware, many projects support other enhancements. Common examples of additions include support for higher video resolutions and different aspect ratios, hardware accelerated renderers (OpenGL and/or Direct3D), enhanced input support (including the ability to map controls onto additional input devices), 3D character models (in case of 2.5D games), higher resolution textures, support to replace MIDI with digital audio (MP3, Ogg Vorbis, etc.), and enhanced multiplayer support using the Internet.
Several source ports have been created for various games specifically to address online multiplayer support. Most older games were not created to take advantage of the Internet and the low latency, high bandwidth Internet connections available to computer gamers today. Furthermore, old games may use outdated network protocols to create multiplayer connections, such as IPX protocol, instead of Internet Protocol. Another problem was games that required a specific IP address for connecting with another player. This requirement made it difficult to quickly find a group of strangers to play with — the way that online games are most commonly played today. To address this shortcoming, specific source ports such as Skulltag added "lobbies", which are basically integrated chat rooms in which players can meet and post the location of games they are hosting or may wish to join. Similar facilities may be found in newer games |
https://en.wikipedia.org/wiki/In-band%20signaling | In telecommunications, in-band signaling is the sending of control information within the same band or channel used for data such as voice or video. This is in contrast to out-of-band signaling which is sent over a different channel, or even over a separate network. In-band signals may often be heard by telephony participants, while out-of-band signals are inaccessible to the user. The term is also used more generally, for example of computer data files that include both literal data, and metadata and/or instructions for how to process the literal data.
Telephony
When dialing from a land-line telephone, the telephone number is encoded and transmitted across the telephone line in form of dual-tone multi-frequency signaling (DTMF). The tones control the telephone system by instructing the telephone switch where to route the call. These control tones are sent over the same channel, the copper wire, and in the frequency range (300 Hz to 3.4 kHz) as the audio of the telephone call. In-band signaling is also used on older telephone carrier systems to provide inter-exchange information for routing calls. Examples of this kind of in-band signaling system are the Signaling System No. 5 (SS5) and its predecessors, and R2 signalling.
Separating the control signals, also referred to as the control plane, from the data, if a bit-transparent connection is desired, is usually done by escaping the control instructions. Occasionally, however, networks are designed so that data is, to a varying degree, garbled by the signaling. Allowing data to become garbled is usually acceptable when transmitting sounds between humans, since the users rarely notice the slight degradation, but this leads to problems when sending data that has very low error tolerance, such as information transmitted using a modem.
In-band signaling is insecure because it exposes control signals, protocols and management systems to end users, which may result in falsing. In the 1960s and 1970s, so-called phone phreaks used blue boxes for deliberate falsing, in which the appropriate tones for routing were intentionally generated, enabling the caller to abuse functions intended for testing and administrative use and to make free long-distance calls.
Modems may also interfere with in-band signaling, in which case a guard tone may be employed to prevent this.
Voice over IP
In voice over IP (VoIP), DTMF signals are transmitted in-band by two methods. When transmitted as audio tones in the voice stream, voice encoding must use a lossless coder, such as μ-law or A-law pulse-code modulation, to preserve the integrity of frequency signals. Still, this method proved often unreliable and was subject to interference from other audio sources. The standard method is to digitally remove DTMF tones from the audio at the source and from the Real-time Transport Protocol (RTP) voice stream and encode them separately as a digital information payload, often termed named telephone events (NTE), according to RFC |
https://en.wikipedia.org/wiki/Beast%20Wars%3A%20Transformers | Beast Wars: Transformers (titled Beasties: Transformers in Canada) is a computer-animated television series that debuted in 1996 and ended on May 7, 1999, serving as the flagship of the Transformers: Beast Wars franchise. It was one of the earliest fully CGI television shows. The series is set in the future of the "original" Transformers franchise, 300 years after the events of The Transformers, and features the Maximals and Predacons, descendants of the Autobots and Decepticons respectively. While engaged in battle, small teams from each faction crash land on an unknown planet, and must find a way to return home while continuing their war.
The Beast Wars TV series was the first Transformers series to feature computer-animated characters, and was produced by Mainframe Entertainment of Vancouver, British Columbia; its story editors were Bob Forward and Larry DiTillio. The production designer for the show, Clyde Klotz, won a Daytime Emmy Award for Outstanding Individual Achievement in Animation in 1997 for his work on Beast Wars.
A sequel television series, Beast Machines: Transformers, aired from 1999 to 2000. Additional Beast Wars limited comic book series have been released by Dreamwave Productions and IDW Publishing.
Setting and plot summary
The two main factions of "Transformers" in Beast Wars are descendants of the two main factions in the original cartoon: the Maximals are the descendants of the Autobots and the Predacons are the descendants of the Decepticons. (In the sequel series Beast Machines, the process during which Autobots and Decepticons became Maximals and Predacons is referred to as "The Great Upgrade.")
The leader of the Predacon team is Megatron, a namesake of the original Decepticon commander. He and his forces are a splinter group on the hunt for powerful crystals known as Energon. They do this with the aid of an artifact known as the Golden Disk and Megatron's stolen ship, the Darksyde, which is equipped with a transwarp drive. A Maximal exploration ship, the Axalon, led by Optimus Primal, is sent to stop them. Together the ships plunge through a time/space phenomenon created by the transwarp device during their battle in space, and crash-land on a mysterious planet.
The planet is found to be rich in deposits of raw Energon, in such extreme amounts that it proves to be poisonous to both factions' robot forms, forcing them to take on alternate organic forms for protection until their robot forms are needed. Thus the robots take on the beast forms of recognizable animals including mammals, birds, fish, reptiles, amphibians, dinosaurs, and invertebrates.
Before crashing, the Axalon deploys its cargo of "stasis pods" containing Maximal protoforms — Transformer robots with vulnerable and undeveloped physical forms, which are left to orbit the planet as an alternative to possible destruction in the initial crash landing. This plays a larger part in the IDW series, The Gathering. Throughout the series, stasis pods lose altit |
https://en.wikipedia.org/wiki/Chroot | chroot is an operation on Unix and Unix-like operating systems that changes the apparent root directory for the current running process and its children. A program that is run in such a modified environment cannot name (and therefore normally cannot access) files outside the designated directory tree. The term "chroot" may refer to the system call or the wrapper program. The modified environment is called a chroot jail.
History
The chroot system call was introduced during development of Version 7 Unix in 1979. One source suggests that Bill Joy added it on 18 March 1982 – 17 months before 4.2BSD was released – in order to test its installation and build system. All versions of BSD that had a kernel have chroot(2). An early use of the term "jail" as applied to chroot comes from Bill Cheswick creating a honeypot to monitor a hacker in 1991.
The first article about a jailbreak has been discussed on the security column of SunWorld Online which is written by Carole Fennelly; the August 1999 and January 1999 editions cover most of the chroot() topics.
To make it useful for virtualization, FreeBSD expanded the concept and in its 4.0 release in 2000 introduced the jail command.
By 2002, an article written by Nicolas Boiteux described how to create a jail on Linux
By 2003, first internet microservices providers with Linux jails provide SAAS/PAAS (shell containers, proxy, ircd, bots, ...) services billed for consumption into the jail by usage
By 2005, Sun released Solaris Containers (also known as Solaris Zones), described as "chroot on steroids."
By 2008, LXC (upon which Docker was later built) adopted the "container" terminology and gained popularity in 2013 due to inclusion into Linux kernel 3.8 of user namespaces.
Uses
A chroot environment can be used to create and host a separate virtualized copy of the software system. This can be useful for:
Testing and development A test environment can be set up in the chroot for software that would otherwise be too risky to deploy on a production system.
Dependency control Software can be developed, built and tested in a chroot populated only with its expected dependencies. This can prevent some kinds of linkage skew that can result from developers building projects with different sets of program libraries installed.
Compatibility Legacy software or software using a different ABI must sometimes be run in a chroot because their supporting libraries or data files may otherwise clash in name or linkage with those of the host system.
Recovery Should a system be rendered unbootable, a chroot can be used to move back into the damaged environment after bootstrapping from an alternate root file system (such as from installation media, or a Live CD).
Privilege separation Programs are allowed to carry open file descriptors (for files, pipelines and network connections) into the chroot, which can simplify jail design by making it unnecessary to leave working files inside the chroot directory. This also s |
https://en.wikipedia.org/wiki/Operators%20in%20C%20and%20C%2B%2B | This is a list of operators in the C and C++ programming languages. All the operators (except typeof) listed exist in C++; the column "Included in C", states whether an operator is also present in C. Note that C does not support operator overloading.
When not overloaded, for the operators &&, ||, and , (the comma operator), there is a sequence point after the evaluation of the first operand.
C++ also contains the type conversion operators const_cast, static_cast, dynamic_cast, and reinterpret_cast. The formatting of these operators means that their precedence level is unimportant.
Most of the operators available in C and C++ are also available in other C-family languages such as C#, D, Java, Perl, and PHP with the same precedence, associativity, and semantics.
Table
For the purposes of these tables, a, b, and c represent valid values (literals, values from variables, or return value), object names, or lvalues, as appropriate. R, S and T stand for any type(s), and K for a class type or enumerated type.
Arithmetic operators
All arithmetic operators exist in C and C++ and can be overloaded in C++.
Comparison operators/relational operators
All comparison operators can be overloaded in C++. Since C++20, the inequality operator is automatically generated if operator== is defined and all four relational operators are automatically generated if operator<=> is defined.
Logical operators
All logical operators exist in C and C++ and can be overloaded in C++, albeit the overloading of the logical AND and logical OR is discouraged, because as overloaded operators they behave as ordinary function calls, which means that both of their operands are evaluated, so they lose their well-used and expected short-circuit evaluation property.
Bitwise operators
All bitwise operators exist in C and C++ and can be overloaded in C++.
Assignment operators
All assignment expressions exist in C and C++ and can be overloaded in C++.
For the given operators the semantic of the built-in combined assignment expression a ⊚= b is equivalent to a = a ⊚ b, except that a is evaluated only once.
Member and pointer operators
Other operators
Notes:
Operator precedence
The following is a table that lists the precedence and associativity of all the operators in the C and C++ languages. Operators are listed top to bottom, in descending precedence. Descending precedence refers to the priority of the grouping of operators and operands. Considering an expression, an operator which is listed on some row will be grouped prior to any operator that is listed on a row further below it. Operators that are in the same cell (there may be several rows of operators listed in a cell) are grouped with the same precedence, in the given direction. An operator's precedence is unaffected by overloading.
The syntax of expressions in C and C++ is specified by a phrase structure grammar. The table given here has been inferred from the grammar. For the ISO C 1999 standard, section 6.5.6 note 71 |
https://en.wikipedia.org/wiki/TV%20Osaka | is a television station affiliated with the TX Network in Osaka, Japan. Nikkei, Inc. is the biggest shareholder of TVO. The mascot character is "Takoru-kun" (たこるくん).
History
Television Osaka, Inc. was founded in 1981 as part of the reorganization of Tokyo Channel 12 Ltd (now TV Tokyo). On March 1 the following year, TV Osaka started broadcasting as the first affiliate of the "Mega TON Network" (メガTONネットワーク, now TX Network) outside of Tokyo.
On December 1, 2003, TV Osaka adopted a new logo to reflect the digital era, when they commenced digital telecasts. Analog telecasts were terminated on July 24, 2011. TV Osaka is building its new headquarter building now, and plan moves into its new headquarter in 2024.
Anime produced
The New Adventures of Maya the Honeybee
Taotao Ehonkan
Fushigi no Kuni no Alice
Mahou no Stage Fancy Lala
Gasaraki
Soreyuke! Uchuu Senkan Yamamoto Yohko
Guru Guru Town Hanamaru-kun
Cosmic Baton Girl Comet-san
Geisters
Galaxy Angel (Z/A/AA)
Pita-Ten
Di Gi Charat Nyo
The Marshmallow Times
Onegai My Melody
Jewelpet
Romance of the Three Kingdoms
Oshiete Mahou no Pendulum ~Rilu Rilu Fairilu~
Tomica Bond Combination Earth Granner
Mazica Party
Cap Revolution Bottleman DX (current)
PuniRunes (current)
Transmitters
Offices
The headquarters – 1-2-18, Otemae, Chūō-ku, Osaka, Osaka Prefecture, Japan
Kobe Branch Office – Nikkei Kobe Kaikan, 7-1-14, Shimoyamate-dori, Chūō-ku, Kobe, Hyōgo Prefecture
Tokyo Branch Office – Ginza Towa Building, 3-10-7, Ginza, Chūō, Tokyo
Nagoya Branch Office – Nikkei Nagoya Office, 4-16-33, Sakae, Naka-ku, Nagoya, Aichi Prefecture
See also
Television in Japan
References
External links
the official website of TV Osaka
Corpolate Profile
Companies based in Osaka Prefecture
Television channels and stations established in 1981
TX Network
Mass media in Osaka |
https://en.wikipedia.org/wiki/Line%20number | In computing, a line number is a method used to specify a particular sequence of characters in a text file. The most common method of assigning numbers to lines is to assign every line a unique number, starting at 1 for the first line, and incrementing by 1 for each successive line.
In the C programming language the line number of a source code line is one greater than the number of new-line characters read or introduced up to that point.
Programmers could also assign line numbers to statements in older programming languages, such as Fortran, JOSS, and BASIC. In Fortran, not every statement needed a line number, and line numbers did not have to be in sequential order. The purpose of line numbers was for branching and for reference by formatting statements.
Both JOSS and BASIC made line numbers a required element of syntax. The primary reason for this is that most operating systems at the time lacked interactive text editors; since the programmer's interface was usually limited to a line editor, line numbers provided a mechanism by which specific lines in the source code could be referenced for editing, and by which the programmer could insert a new line at a specific point. Line numbers also provided a convenient means of distinguishing between code to be entered into the program and direct mode commands to be executed immediately when entered by the user (which do not have line numbers).
Largely due to the prevalence of interactive text editing in modern operating systems, line numbers are not a feature of most programming languages, even modern Fortran and Basic.
History
FORTRAN
In Fortran, as first specified in 1956, line numbers were used to define input/output patterns, to specify statements to be repeated, and for conditional branching. For example:
DIMENSION ALPHA(25), RHO(25)
1) FORMAT(5F12.4)
2) READ 1, ALPHA, RHO, ARG
SUM = 0.0
DO 3 I=1, 25
IF (ARG-ALPHA(I)) 4,3,3
3) SUM = SUM + ALPHA(I)
4) VALUE = 3.14159*RHO(I-1)
PRINT 1, ARG, SUM, VALUE
GO TO 2
Like assembler language before it, Fortran did not assume every line needed a label (line number, in this case). Only statements referenced elsewhere required a line number:
Line 1 specifies a format pattern for input; the command in line 2 and the later command both reference this line.
The loop executes line 3.
The arithmetic IF statement branches to line 4 on a negative value, line 3 on zero, and again line 3 on a positive value.
While the line numbers are sequential in this example, in the very first "complete but simple [Fortran] program" published the line numbers are in the sequence 1, 5, 30, 10, 20, 2.
Line numbers could also be assigned to fixed-point variables (e.g., i n) for referencing in subsequent assigned GO TO statements (e.g., n,(n1,n2,...nm)).
COBOL
In COBOL, line numbers were specified in the first six characters (the sequence number area) of punched cards. This was originally used for facilitating mechanical card sorting to assure in |
https://en.wikipedia.org/wiki/BESM-6 | BESM-6 (, short for Большая электронно-счётная машина, i.e. 'Large Electronic Calculating Machine') was a Soviet electronic computer of the BESM series. It was the first Soviet second-generation, transistor-based computer.
Overview
The BESM-6 was the most well-known and influential model of the series designed at the Institute of Precision Mechanics and Computer Engineering. The design was completed in 1965. Production started in 1968 and continued for the following 19 years.
Like its predecessors, the original BESM-6 was transistor-based (however, the version used in the 1980s as a component of the Elbrus supercomputer was built with integrated circuits). The machine's 48-bit processor ran at 10 MHz clock speed and featured two instruction pipelines, separate for the control and arithmetic units, and a data cache of sixteen 48-bit words. The system achieved a performance of 1 MIPS.
The CDC 6600, a common Western supercomputer when the BESM-6 was released, achieved about 2 MIPS.
The system memory was word-addressable using 15-bit addresses. The maximum addressable memory space was thus 32K words (192K bytes). A virtual memory system allowed to expand this up to 128K words (768K bytes).
The BESM-6 was widely used in USSR in the 1970s for various computation and control tasks. During the 1975 Apollo-Soyuz Test Project the processing of the space mission telemetry data was accomplished by a new computer complex which was based on a BESM-6. The Apollo-Soyuz mission's data processing by soviet scientists finished half an hour earlier than their American colleagues from NASA.
A total of 355 of these machines were built. Production ended in 1987.
As the first Soviet computer with an installed base that was large for the time, the BESM-6 gathered a dedicated developer community. Over the years several operating systems and compilers for programming languages such as Fortran, ALGOL and Pascal were developed.
A modification of the BESM-6 based on integrated circuits, with 2-3 times higher performance than the original machine, was produced in the 1980s under the name Elbrus-1K2 as a component of the Elbrus supercomputer.
In 1992, one of the last surviving BESM-6 machines was purchased by the Science Museum in London, England.
Peripherals
The BESM-6 could send output to an АЦПУ-128 (Алфавитно-Цифровое Печатающее Устройство) printer, and read input from punched cards in the GOST 10859 character set. A Consul-254 teletype, made by Zbrojovka Brno in Czechoslovakia, could be used for interactive sessions. When CRT terminals became available, the BESM-6 could be connected to Videoton 340 terminals.
Further reading
(NB. Has information on the BESM-6 character set.)
References
External links
BESM-6 Nostalgia Page
Soviet computer systems
1965 establishments in Russia
Supercomputers |
https://en.wikipedia.org/wiki/Scanner%20Access%20Now%20Easy | Scanner Access Now Easy (SANE) is an open-source application programming interface (API) that provides standardized access to any raster image scanner hardware (flatbed scanner, handheld scanner, video- and still-cameras, frame grabbers, etc.). The SANE API is public domain. It is commonly used on Linux.
Architecture
SANE differs from TWAIN in that it is cleanly separated into "front ends" (user programs) and "back ends" (scanner drivers). Whereas a TWAIN driver handles the user interface as well as communications with the scanner hardware, a SANE driver only provides an interface with the hardware and describes a number of "options" which drive each scan. These options specify parameters such as the resolution of the scan, the scan area, colour model, etc. Each option has a name, and information about its type, units, and range or possible values (e.g., enumerated list). By convention there are several "well known" options that front ends can supply using convenient GUI interaction e.g., the scan area options can be set by dragging a rectangular outline over a preview image. Other options can be presented using GUI elements appropriate to their type e.g., sliders, drop-down lists, etc.
One consequence of this separation is that network scanning is easily implemented with no special handling in either the front ends or back ends. On a host with a scanner, the saned daemon runs and handles network requests. On client machines a "net" back end (driver) connects to the remote host to fetch the scanner options, and perform previews and scans. The saned daemon acts as a front end locally, but simply passes requests and data between the network connections and the local scanner. Similarly, the "net" back end passes requests and data between the local front end and the remote host.
Various types of unsupervised batch scanning are also possible with a minimum of support needed in the back end (driver). Many scanners support the attachment of document feeders which allow a large number of sheets of paper to be automatically scanned in succession. Using the SANE API, the front end simply has to "play back" the same set of options for each scan, driving the document feed in between scans to load the next sheet of paper. The front end only has to obtain the set of options from the user once.
Graphical user interfaces
Several user interfaces have been written to combine SANE with an easy user method of controlling it.
gscan2pdf
gscan2pdf is an interface for scanning documents to PDF on the GNOME desktop that uses SANE to communicate with the scanner. It is available under the GPL. It includes common editing tools, e.g., for rotating or cropping pages. It is also able to perform OCR using several optional OCR tools and save a searchable PDF. PDF files can be further downsampled upon saving.
Simple Scan
Simple Scan is a simplified GUI using SANE that is intended to be easier to use and better integrate into the GNOME desktop than XSane. It was initially |
https://en.wikipedia.org/wiki/Programming%20%28music%29 | Programming is a form of music production and performance using electronic devices and computer software, such as sequencers and workstations or hardware synthesizers, sampler and sequencers, to generate sounds of musical instruments. These musical sounds are created through the use of music coding languages. There are many music coding languages of varying complexity. Music programming is also frequently used in modern pop and rock music from various regions of the world, and sometimes in jazz and contemporary classical music. It gained popularity in the 1950s and has been emerging ever since.
Music programming is the process in which a musician produces a sound or "patch" (be it from scratch or with the aid of a synthesizer/sampler), or uses a sequencer to arrange a song.
Coding languages
Music coding languages are used to program the electronic devices to produce the instrumental sounds they make. Each coding language has its own level of difficulty and function.
Alda
The music coding language Alda provides a tutorial on coding music and is, "designed for musicians who do not know how to program, as well as programmers who do not know how to music". The website also has links to install, tutorial, cheat sheet, docs, and community for anyone visiting the website.
LC
LC computer music programming language is a more complex computer music programming language meant for more experienced coders. One of the differences between this language and other music coding languages is that, "Unlike existing unit-generator languages, LC provides objects as well as library functions and methods that can directly represent microsounds and related manipulations that are involved in microsound synthesis."
History and development
Music programming has had a vast history of development leading to the creation of different programs and languages. Each development comes with more function and utility and each decade tends to favor a certain program and or piece of equipment.
MUSIC N
The first digital synthesis family of computer programs and languages being MUSIC-N created by Max Mathews. The development of these programs, allowed for more flexibility and utility, eventually leading them to become fully developed languages. As programs such as MUSIC I, MUSIC II and MUSIC III were developed, which were all created by Max Matthews, new technologies were incorporated in such as the table-lookup oscillator in MUSIC II and the unit generator in MUSIC III. The breakthrough technologies such as the unit generator, which acted as a building block for music programming software, and the acoustic compiler, which allowed "unlimited number of sound synthesis structures to be created in the computer", further the complexity and evolution of music programming systems.
Drum machines
Around the time of the 1950s, electric rhythm machines began to make way into popular music. These machines began to gain much traction amongst many artists as they saw it as a way to creat |
https://en.wikipedia.org/wiki/Deloitte | Deloitte Touche Tohmatsu Limited (), commonly referred to as Deloitte, is a British multinational professional services network headquartered in London, England. Deloitte is the largest professional services network by revenue and number of professionals in the world and is considered one of the Big Four accounting firms along with EY, KPMG and PwC.
The firm was founded by William Welch Deloitte in London in 1845 and expanded into the United States in 1890. It merged with Haskins & Sells to form Deloitte Haskins & Sells in 1972 and with Touche Ross in the US to form Deloitte & Touche in 1989. In 1993, the international firm was renamed Deloitte Touche Tohmatsu, later abbreviated to Deloitte. In 2002, Arthur Andersen's practice in the UK as well as several of that firm's practices in Europe and North and South America agreed to merge with Deloitte. Subsequent acquisitions have included Monitor Group, a large strategy consulting business, in January 2013. The international firm is a UK private company, limited by guarantee, supported by a network of independent legal entities.
Deloitte provides audit, consulting, financial advisory, risk advisory, tax, and legal services with approximately 415,000 professionals globally. In FY 2021, the network earned revenues of US$50.2 billion in aggregate. The firm has sponsored a number of activities and events including the 2012 Summer Olympics.
Controversies involving the firm, in addition to litigation surrounding a few of its audits, have included its involvement in a "potentially misleading" report on illicit tobacco trading in Australia, the fact that it suffered a major cyber-attack which breached client confidentiality as well as exposing extensive employee information in September 2017, its role as internal auditor of the insolvent contractor Carillion and its role as external auditor of Autonomy which was accused of "accounting improprieties" that contributed to an £8.8 billion write-down of value following its acquisition by Hewlett-Packard in 2011.
History
Early history
In 1845, William Welch Deloitte opened an office in London, United Kingdom. Deloitte was the first person to be appointed an independent auditor of a public company, namely the Great Western Railway. He went on to open an office in New York in 1880.
In 1890, Deloitte opened a branch office on Wall Street headed by Edward Adams and P.D. Griffiths as branch managers. That was Deloitte's first overseas venture. Other branches were soon opened in Chicago and Buenos Aires. In 1898 P.D. Griffiths returned from New York and became a partner in the London office.
In 1896, Charles Waldo Haskins and Elijah Watt Sells formed Haskins & Sells in New York. It was later described as "the first major auditing firm to be established in the country by American rather than British accountants".
In 1898, George Touche established an office in London and then, in 1900, joined John Ballantine Niven in establishing the firm of Touche Niven in the J |
https://en.wikipedia.org/wiki/Invision%20Community | Invision Community is a brand of forum software developed in 2002 and originally marketed as Invision Power Board. The current version of the software was written in PHP and uses MySQL for database storage.
Invision Power Services (IPS) was created in 2002 by Charles Warner and Matt Mecham after they left Jarvis Entertainment Group, which had bought the forum software Ikonboard from Mecham. Their first product sold by IPS was the forum software Invision Power Board, which quickly gathered a community of former Ikonboard users.
The software has been marketed for over twenty years and has been updated and changed over that timeframe.
Version 1.x.x
Early releases of Invision Power Board were available as a download free of charge under a proprietary license.
Version 2.x.x
In 2004 Invision Power Board ended its free releases for non-commercial uses.
After the 2.0.1 update, the free downloads were replaced with a free demo with restrictions of 5000 posts, 1000 threads, 200 members and other restrictions, thus ending its claims that Invision Power Board would be free forever thereby upsetting many users. In version 2.3, the product shipped with 2 skins, the classic light blue skin and a new default darker skin.
Version 3.x.x
Version 3.0 was initially released on Tuesday, June 23, 2009.
Version 4.x.x
In May of 2012, Version 4.0 represented a ground-up rewrite of Invision Power Board and related offerings.
With the release of IP.Suite 4.0, IPS has decided to no longer honor the perpetual lifetime licenses they have sold earlier. As a compensation for legacy customers, they offered a free transition period to the standard license type before legacy customers would be required to pay their first renewal if they wanted to continue to receive support and future updates.
The first officially supported release was version 4.0.0 on 9 April 2015. IP.Suite 4 was not properly announced as released until 16 June, at which point it had already reached version 4.0.8.
See also
Comparison of Internet forum software
References
External links
Invision Community
Release Notes
Internet forum software
PHP software |
https://en.wikipedia.org/wiki/ES%20EVM | The ES EVM (, "Unified System of Electronic Computing Machines"), or YeS EVM, also known in English literature as the Unified System or Ryad (, "Series"), is a series of mainframe computers generally compatible with IBM's System/360 and System/370 mainframes, built in the Comecon countries under the initiative of the Soviet Union between 1968 and 1998. More than 15,000 of the ES EVM mainframes were produced in total.
Development
In 1966, the Soviet economists suggested creating a unified series of mutually compatible computers. Due to the success of the IBM System/360 in the United States, the economic planners decided to use the IBM design, although some prominent Soviet computer scientists had criticized the idea and suggested instead choosing one of the Soviet indigenous designs, such as BESM or Minsk. The first works on the cloning began in 1968; production started in 1972. In addition, after 1968, other Comecon countries joined the project.
With the exception of only a few hardware pieces, the ES EVM machines were recognized in the Western countries as independently designed, based on legitimate Soviet patents. Unlike the hardware, which was quite original, mostly created by reverse engineering, much of the software was based on slightly modified and localized IBM code. In 1974–1976, IBM had contacted the Soviet authorities and expressed interest in ES EVM development; however, after the Soviet Army entered Afghanistan, in 1979, all contacts between IBM and ES developers were interrupted, due to the U.S. embargo on technological cooperation with the USSR.
Due to the CoCom restrictions, much of the software localization was done through disassembling the IBM software, with some minimal modification. The most common operating system was OS ES (), a modified version of OS/360; the later versions of OS ES were very original and different from the IBM OSes, but they also included a lot of original IBM code. There were even anecdotal rumors among the Soviet programmers, that this supposedly Soviet operating system contained some secret command which outputs the American national anthem. Today some of the Russian institutions that worked on ES EVM are cooperating with IBM to continue legacy support for both actual IBM mainframes and the ES EVM systems.
ES EVM machines were developed in Moscow, at the Scientific Research Center for Electronic Computer Machinery (NITsEVT); in Yerevan, Armenia, at Yerevan Computer Research and Development Institute; and later in Minsk, at the Scientific Research Institute of Electronic Computer Machines (NIIEVM); and in Penza, at the Penza Scientific Research Institute of Computer Machinery. They were manufactured in Minsk, at the Minsk Production Group for Computing Machinery (MPOVT); and in Penza, at the Penza Electronic Computer Factory. Some models had been also produced in other countries of the Eastern Bloc, such as Bulgaria, Hungary, Poland, Czechoslovakia, Romania, and East Germany; some peripheral device |
https://en.wikipedia.org/wiki/List%20of%20Soviet%20computer%20systems | This is the list of Soviet computer systems. The Russian abbreviation EVM (ЭВМ), present in some of the names below, means “electronic computing machine” ().
List of hardware
The Russian abbreviation EVM (ЭВМ), present in some of the names below, means “electronic computing machine” ().
Ministry of Radio Technology
Computer systems from the Ministry of Radio Technology:
Agat (Агат) — Apple II clone
ES EVM (ЕС ЭВМ), IBM mainframe clone
ES PEVM (ЕС ПЭВМ), IBM PC compatible
M series — series of mainframes and mini-computers
Minsk (Минск)
Poisk (Поиск) — IBM PC-XT clone
Setun (Сетунь) — unique balanced ternary computer.
Strela (Стрела)
Ural (Урал) — mainframe series
Vector-06C (Вектор-06Ц)
Ministry of Instrument Making
Computer systems from the Ministry of Instrument Making:
Aragats (Арагац)
Iskra (Искра) — common name for many computers with different architecture
Iskra-1030 — Intel 8086 XT clone
KVM-1 (КВМ-1)
SM EVM (СМ ЭВМ) — most models were PDP-11 clones, while some others were HP 2100, VAX or Intel compatible
Ministry of the Electronics Industry
Computer systems from the Ministry of Electronics Industry:
Elektronika (Электроника) family
DVK family (ДВК) — PDP-11 clones
Elektronika BK-0010 (БК-0010, БК-0011) — LSI-11 clone home computer
UKNC (УКНЦ) — educational, PDP11-like
Elektronika 60, Elektronika 100
Elektronika 85 — Clone of DEC Professional (computer) 350 (F11)
Elektronika 85.1 — Clone of DEC Professional (computer) 380 (J11)
Elektronika D3-28
Elektronika SS BIS (Электроника СС БИС) — Cray clone
Soviet Academy of Sciences
BESM (БЭСМ) — series of mainframes
Besta (Беста) — Unix box, Motorola 68020-based, Sun-3 clone
Elbrus (Эльбрус) — high-end mainframe series
Kronos (Кронос)
MESM (МЭСМ) — first Soviet Union computer (1950)
M-1 — one of the earliest stored program computers (1950-1951)
ZX Spectrum clones
ATM Turbo
Dubna 48K - running at half the speed of the original
Hobbit
Pentagon
Radon 'Z'
Scorpion
Other
5E** (5Э**) series - military computers
5E51 (5Э51)
5E53 (5Э53)
5E76 (5Э76) - IBM/360 clone, military version
5E92 (5Э92)
5E92b (5Э92б)
A series — ES EVM-compatible military computers
Argon — a series of military real-time computers
AS-6 (АС-6) - multiprocessor computing complex, name is Russian abbreviation for "Connection Equipment - 6"
Dnepr (Днепр)
GVS-100 (ГВС-100, Гибридная Вичислителная Система) - Hybrid Computer System
Irisha (Ириша)
Juku (Юку) — Estonian school computer
Kiev (Киев)
Korvet (Корвет)
Krista (Криста)
Micro-80 (Микро-80) — experimental PC, based on 8080-compatible processor
Microsha (Микроша) — modification of Radio-86RK
MIR, МИР (:uk:ЕОМ "МИР-1", :uk:ЕОМ "МИР-2")
Nairi (Наири)
Orion-128 (Орион-128)
Promin (Проминь)
PS-2000, PS-3000 — multiprocessor supercomputers in the 1980s
Razdan (Раздан)
Radon — real-time computer, designed for anti-aircraft defense
Radio-86RK — simplified and modified version of Micro-80
Sneg (Снег)
Specialist |
https://en.wikipedia.org/wiki/MDL%20%28programming%20language%29 | MDL (Model Development Language, or colloquially also referred to as More Datatypes than Lisp or MIT Design Language) is a programming language, a descendant of the language Lisp. Its initial purpose was to provide high level language support for the Dynamic Modeling Group at Massachusetts Institute of Technology's (MIT) Project MAC. It was developed in 1971 on a PDP-10 running ITS and later ran on TENEX, TOPS-20, BSD, and AEGIS.
The initial development team consisted of Gerald Sussman and Carl Hewitt of the Artificial Intelligence Lab, and Chris Reeve, Bruce Daniels, and David Cressey of the Dynamic Modeling Group. Later, Stu Galley, also of the Dynamic Modeling Group, wrote the MDL documentation.
MDL was initially called Muddle. This style of self-deprecating humor was not widely understood or appreciated outside of Project MAC. So the name was sanitized to MDL.
MDL provides several enhancements to classic Lisp. It supports several built-in data types, including lists, strings and arrays, and user-defined data types. It offers multithreaded expression evaluation and coroutines. Variables can carry both a local value within a scope, and a global value, for passing data between scopes. Advanced built-in functions supported interactive debugging of MDL programs, incremental development, and reconstruction of source programs from object programs.
Although MDL is obsolete, some of its features have been incorporated in later versions of Lisp. Gerald Sussman went on to develop the Scheme language, in collaboration with Guy Steele, who later wrote the specifications for Common Lisp and Java. Carl Hewitt had already published the idea for the language Planner before the MDL project began, but his subsequent thinking on Planner reflected lessons learned from building MDL. Planner concepts influenced languages such as Prolog and Smalltalk. Smalltalk and Simula, in turn, influenced Hewitt's future work on the actor model.
But the largest influence that MDL had was on the software genre of interactive fiction (IF). An IF game named Zork, sometimes called Dungeon, was first written in MDL. Later, Reeve, Daniels, Galley and other members of Dynamic Modeling went on to start Infocom, a company that produced many early commercial works of interactive fiction.
In 1980 Marc Blank and Joel Berez adapted the MDL language to create a subset called ZIL (Zork Implementation Language) which was used extensively by Infocom to create their award winning games.
Code sample
This is a sample of PDP-10 MDL:
<DEFINE EXIT-TO (EXITS RMS)
#DECL ((EXITS) EXIT (RMS) <UVECTOR [REST ROOM]>)
<MAPF <>
<FUNCTION (E)
#DECL ((E) <OR DIRECTION ROOM CEXIT NEXIT DOOR>)
<COND (<TYPE? .E DIRECTION>)
(<AND <TYPE? .E ROOM> <MEMQ .E .RMS>>
<MAPLEAVE T>)
(<AND <TYPE? .E CEXIT> <MEMQ <2 .E> .RMS>>
<MAPLEAVE T>)
|
https://en.wikipedia.org/wiki/IITRAN | IITRAN is a discontinued programming language created in the mid-1960s and designed as a first language for students. The primary designer of the language, William S. Worley, also contributed to the design of PL/I, and the two languages have similar syntax. The name derives from Illinois Institute of Technology, where it was developed.
The IITRAN language was initially implemented for IBM 7040, with a compiler made available to students in 1964. This was followed shortly by an IBM System/360 implementation in 1966, for which the language was partially redesigned by a committee lead by Dr. Peter G. Lykos. In the early 1970s, a compiler was developed for the Univac 1108 platform.
IITRAN was designed and developed in response to the increasing demand for a computer language which would meet the following specifications:
It should be clear, concise, and easily learned, even for those who have had no previous experience with computers or mathematics;
It should bear as close a resemblance as possible to the English language;
It should be free of awkward restrictions and limitations;
It should be consistent with mathematical and logical foundations;
It should allow processing of a great number of individual programs in a very short time;
It should serve as a computational tool for students of science and engineering;
It should process a clear, easily understood, set of diagnostic error messages.
(Bauer, p. V)
There was a Spanish language version of IITRAN at IIT as well. It utilized Spanish keywords rather than English ones. For example, the keyword read was replaced by leer.
Bibliography
IITRAN / 360: Self-Instructional Manual and Text, Bauer, Charles R. et al., Addison Wesley Publishing, 1967.
The IITRAN Programming Language, R. Dewar et al., CACM 12(10):569-575 (Oct 1969).
Educational programming languages
PL/I programming language family |
https://en.wikipedia.org/wiki/Signed%20number%20representations | In computing, signed number representations are required to encode negative numbers in binary number systems.
In mathematics, negative numbers in any base are represented by prefixing them with a minus sign ("−"). However, in RAM or CPU registers, numbers are represented only as sequences of bits, without extra symbols. The four best-known methods of extending the binary numeral system to represent signed numbers are: sign–magnitude, ones' complement, two's complement, and offset binary. Some of the alternative methods use implicit instead of explicit signs, such as negative binary, using the base −2. Corresponding methods can be devised for other bases, whether positive, negative, fractional, or other elaborations on such themes.
There is no definitive criterion by which any of the representations is universally superior. For integers, the representation used in most current computing devices is two's complement, although the Unisys ClearPath Dorado series mainframes use ones' complement.
History
The early days of digital computing were marked by competing ideas about both hardware technology and mathematics technology (numbering systems). One of the great debates was the format of negative numbers, with some of the era's top experts expressing very strong and differing opinions. One camp supported two's complement, the system that is dominant today. Another camp supported ones' complement, where a negative value is formed by inverting all of the bits in its positive equivalent. A third group supported sign–magnitude, where a value is changed from positive to negative simply by toggling the word's highest-order bit.
There were arguments for and against each of the systems. Sign–magnitude allowed for easier tracing of memory dumps (a common process in the 1960s) as small numeric values use fewer 1 bits. These systems did ones' complement math internally, so numbers would have to be converted to ones' complement values when they were transmitted from a register to the math unit and then converted back to sign–magnitude when the result was transmitted back to the register. The electronics required more gates than the other systemsa key concern when the cost and packaging of discrete transistors were critical. IBM was one of the early supporters of sign–magnitude, with their 704, 709 and 709x series computers being perhaps the best-known systems to use it.
Ones' complement allowed for somewhat simpler hardware designs, as there was no need to convert values when passed to and from the math unit. But it also shared an undesirable characteristic with sign–magnitude: the ability to represent negative zero (−0). Negative zero behaves exactly like positive zero: when used as an operand in any calculation, the result will be the same whether an operand is positive or negative zero. The disadvantage is that the existence of two forms of the same value necessitates two comparisons when checking for equality with zero. Ones' complement subtraction can |
https://en.wikipedia.org/wiki/Cogito%20%28magazine%29 | Cogito is a philosophical magazine published by Nefeli publications (Athens, Greece) and a member of the Eurozine network. It was founded in 2004 with the intention of making philosophy accessible to the lay reader without compromising it through oversimplification. Cogito is therefore an attempt to approach everyday issues and themes through a philosophically rigorous lens. Contributors, most of them professional philosophers but also academics from a variety of disciplines, as well as writers, consider such varied topics as colors, taste, gardens, cinema, desire and love, art and aesthetics, and logical paradoxes. Each issue, consisting of a number of articles, is dedicated to a different topic: privacy, intelligence both natural and artificial, applied ethics, history and narrative, friendship. Each issue further presents a philosopher to its readers. In the past, subjects of this 'philosophical library' have included Plato, Wittgenstein, J.S. Mill, and Kant.
References
External links
2004 establishments in Greece
Annual magazines
Magazines published in Greece
Greek-language magazines
Magazines established in 2004
Mass media in Athens
Philosophy magazines |
https://en.wikipedia.org/wiki/Lava%20flow%20%28programming%29 | In computer programming jargon, lava flow is a problem in which computer code written under sub-optimal conditions is put into production and added to while still in a developmental state. Often, putting the system into production results in a need to maintain backward compatibility (as many additional components now depend on it) with the original, incomplete design.
Changes in the development team working on a project often exacerbate lava flows. As workers cycle in and out of the project, knowledge of the purpose of aspects of the system can be lost. Rather than clean up these pieces, subsequent workers work around them, increasing the complexity and mess of the system.
Lava flow is considered an anti-pattern, a commonly encountered phenomenon leading to poor design.
References
Anti-patterns |
https://en.wikipedia.org/wiki/Grandia | is a series of role-playing video games developed by Game Arts and published over the years by Entertainment Software Publishing, Sony Computer Entertainment, Ubisoft, Hudson Soft, Enix, Square Enix and GungHo Online Entertainment. Games in the Grandia series have been released for the Sega Saturn, PlayStation, Dreamcast, Game Boy Color, PlayStation 2, PlayStation 3, Nintendo Switch and Microsoft Windows.
The Grandia HD Collection (compiling the first two games in the series) was released for the Nintendo Switch in North America and Europe on August 16, 2019. The games were released in Japan on March 25, 2020. Grandia HD Remaster for the PC was released on October 15, 2019. On the same day, Grandia II Anniversary Edition for the PC, which was originally released on August 24, 2015, received further improvement updates and was renamed Grandia II HD Remaster. Full Japanese language support along with various fixes were released for Grandia HD Remaster and Grandia II HD Remaster on March 25, 2020.
The Grandia series sold 2million units, as of April 2009. The latest original installment was released for Microsoft Windows in July 2009, entitled Grandia Online; it remained online for three years until it was shut down in September 2012.
Games
Grandia (1997)
''Grandias story focuses around Justin, a young boy who desires to be a great adventurer. While investigating a ruin of the lost Angelou Civilization, he hears a mysterious request to come east and find out the secret of why these ancients mysteriously disappeared. The resulting adventure takes Justin across the sea to new continents, and even beyond the known end of the world. Graphically, the game used character sprites in a 3D world, rather than polygonal character models that later came to favor. The battle system, like Game Arts' Lunar series before it, placed emphasis on the combatants' location in the field. However, while Lunar only allowed the player to set the team's position in battle, Grandia had characters move to appropriate positions during battle and allowed the player to have them move elsewhere to avoid attacks or reach a more advantageous position. Also notable is how characters can counter or disrupt enemy attacks.
Grandia was remastered by GungHo Online Entertainment in August 2019 in North America and March 2020 in Japan for Steam and Nintendo Switch as part of the Grandia HD Collection alongside Grandia II. It featured upgraded visuals and widescreen, as well as achievements and trading cards on Steam.
Grandia: Digital Museum (1998)
A bonus disc using the same engine as the Saturn version of Grandia. Though it only consists of four dungeons, each one is very large compared to the ones in the original Grandia. Justin, Feena and Sue must explore them in order to recover artifacts from a museum of the original game that Liete has created. These unlock storyboards, special sound plays known as "Radio Dramas", saves for the original Saturn game, mini-games, bestiary list |
https://en.wikipedia.org/wiki/Game%20Arts | is a Japanese developer and publisher of video games located in Chiyoda, Tokyo. Originally established in 1985 as a computer software company, it expanded into producing for a number of game console and handheld systems. Its President and CEO in 2007 was Yoichi Miyaji at which time it was a member of the Computer Entertainment Supplier's Association of Japan (CESA). Its major trading partners then included Square Enix, Bandai Namco, Koei Tecmo, and Gung-Ho Online Entertainment, some of whom co-developed or produced games in cooperation with the company.
The company has produced a number of games for several genres, beginning with the action game Thexder for personal computers in 1985. A number of traditional and Mahjong-related games have also been produced for Japanese audiences. In the Western world, Game Arts is best known as the producers of the Lunar and Grandia series of role-playing video games, as well as the Gungriffon line of vehicle simulation games. Some of its staff has helped in the preliminary development of Nintendo's Wii title Super Smash Bros. Brawl.
On April 22, 2009, Game Arts released the PlayStation port of Grandia in Japan on the PlayStation Network as a downloadable title in the PSone Classics range, to celebrate an upcoming announcement for Grandia Online.
Release history
References
External links
Amusement companies of Japan
GungHo Online Entertainment
Video game companies established in 1985
Video game companies of Japan
Video game development companies
Video game publishers
Japanese companies established in 1985
Software companies based in Tokyo |
https://en.wikipedia.org/wiki/Secure%20copy%20protocol | Secure copy protocol (SCP) is a means of securely transferring computer files between a local host and a remote host or between two remote hosts. It is based on the Secure Shell (SSH) protocol. "SCP" commonly refers to both the Secure Copy Protocol and the program itself.
According to OpenSSH developers in April 2019, SCP is outdated, inflexible and not readily fixed; they recommend the use of more modern protocols like SFTP and rsync for file transfer. As of OpenSSH version 9.0, scp client therefore uses SFTP for file transfers by default instead of the legacy SCP/RCP protocol.
Secure Copy Protocol
The SCP is a network protocol, based on the BSD RCP protocol, which supports file transfers between hosts on a network. SCP uses Secure Shell (SSH) for data transfer and uses the same mechanisms for authentication, thereby ensuring the authenticity and confidentiality of the data in transit. A client can send (upload) files to a server, optionally including their basic attributes (permissions, timestamps). Clients can also request files or directories from a server (download). SCP runs over TCP port 22 by default. Like RCP, there is no RFC that defines the specifics of the protocol.
Function
Normally, a client initiates an SSH connection to the remote host, and requests an SCP process to be started on the remote server. The remote SCP process can operate in one of two modes:
source mode, which reads files (usually from disk) and sends them back to the client, or
sink mode, which accepts the files sent by the client and writes them (usually to disk) on the remote host.
For most SCP clients, source mode is generally triggered with the -f flag (from), while sink mode is triggered with -t (to). These flags are used internally and are not documented outside the SCP source code.
Remote to remote mode
In the past, in remote-to-remote secure copy, the SCP client opens an SSH connection to the source host and requests that it, in turn, open an SCP connection to the destination. (Remote-to-remote mode did not support opening two SCP connections and using the originating client as an intermediary). It is important to note that SCP cannot be used to remotely copy from the source to the destination when operating in password or keyboard-interactive authentication mode, as this would reveal the destination server's authentication credentials to the source. It is, however, possible with key-based or GSSAPI methods that do not require user input.
Recently, remote-to-remote mode supports routing traffic through the client which originated the transfer, even though it is a 3rd party to the transfer. This way, authorization credentials must reside only on the originating client, the 3rd party.
Issues using talkative shell profiles
SCP does not expect text communicating with the SSH login shell. Text transmitted due to the SSH profile (e.g. echo "Welcome" in the .bashrc file) is interpreted as an error message, and a null line (echo "") causes SCP cl |
https://en.wikipedia.org/wiki/UMAX%20Technologies | UMAX Technologies (), originally known as UMAX Computer Corporation, is a manufacturer of computer products, including scanners, mice, and flash drives, based in Taiwan. The company also uses the Yamada and Vaova brand names.
History
UMAX was formerly a maker of Apple Macintosh clones, using the SuperMac brand name outside of Europe. Their models included the SuperMac S900/S910, J700, C500 and C500e/i/LT, C600e/v/LT/x and Aegis 200. The C500 was marketed as the Apus 2000 in Europe. After Steve Jobs returned to Apple as the new CEO, he revoked all of the clone producers' licenses to produce Mac clones except for UMAX, due to their sub-US$1,000 low-end offerings, a market in which Apple was not strong, and UMAX's stated desire to expand the Macintosh platform's presence in East Asian markets. UMAX's license for Mac OS 8 expired in July 1998. UMAX could not remain profitable selling only these systems, however; it briefly made IBM PC compatible computers in the mid-1990s, but since then UMAX has mainly concentrated on manufacturing scanners.
In 1995, UMAX was the leading Taiwanese scanner maker, with a market share of 13% second worldwide behind Hewlett-Packard (HP). This continued to be the case throughout 1996. According to PC Data figures, in 1997 UMAX briefly overtook HP in some monthly sales. According to the same source however, by 1999 UMAX was being "eclipsed" by HP whose scanner market share doubled that year from 13% to 26%. In some markets with high price-sensitivity like India for example, UMAX continued to have a slight lead on HP throughout 1999-2000 with the two companies claiming 44% and 40%, respectively, of the scanner sales in this country; 85% of which were for products costing less than 10,000 Rs.) By 2003, HP and Canon were dominating the world's flatbed scanner market, "accounting for a combined unit market share of 81 per cent."
In 2002, UMAX started to charge its US customers for driver upgrades for its scanners—a practice that soon proved controversial.
Until their exit from the desktop scanner market in 2002, Heidelberger Druckmaschinen used UMAX as its OEM for these products.
UMAX also made a 1.3 megapixel digital camera called the AstraPix 490. It is capable of recording video clips, functioning as a webcam and can even be used to listen to music encoded in MP3 format.
Scanners
Astra 610S and 1200S; these were cloned and/or repackaged (OEM'd) for many other manufacturers
Astra 1220P
Astra 2000U
Astra 2100U
Astra 2400S, NCR 53C80 SCSI/Intel 8031 based 600x1200dpi
Astra 3450
Astra 4900
Astra 4950
Astra 5600
Astra 6700
AstraSlim
AstraSlim SE
PowerLook 1000
PowerLook 1120
PowerLook 2100XL
PowerLook 180
PowerLook 270
Scanner software
UMAX offers some semi-free (in the sense that some versions/updates cost money and some do not) basic scanner software for Microsoft Windows (up to Windows XP) and Mac OS:
VistaScan is their basic TWAIN scanner module, which also contains Windows Image Acquisition (WI |
https://en.wikipedia.org/wiki/Back%20end | Back end, back-end or backend may refer to:
Electronics
Computing
Back end (computing), the data access layer in software architecture
Back-end CASE
Back-end database, a database accessed indirectly through an external application
Back-end processor, hardware that stores and retrieves data from a database
Integrated circuits
Back-end design, in electronic circuit design flow
Back end of line, in integrated circuit fabrication
Other uses
Archaic northern English dialect word for autumn
Back end load, a mutual fund fee
"Back End" (song), by Finesse2tymes
"Backend" (song), a song on DaBaby's album Baby on Baby
Slang for buttocks, an anatomical feature
See also
Back (disambiguation)
End (disambiguation)
Front end (disambiguation) |
https://en.wikipedia.org/wiki/Front%20end | Front end may refer to:
Computing
Front-end (computing), an interface between the user and the back end
Front-end processor (computer), a small-sized computer
Front-end processor (program)
Front-end web development, the practice of producing HTML, CSS and JavaScript for a website or web application
Front-end API, a type of application program interface
Compiler front-end
Debugger front-end
Other
Front-end bra, a stretchy type of vinyl that attaches to the front of a car
Front-end, the foremost of a vehicle body
Front-end engineering
Front-end load, a charge in investing
Front-end loader, construction equipment
Front-end loading, in project management
Front End Loader, a band
RF front end, in electronics
See also
Front (disambiguation)
End (disambiguation)
Back end (disambiguation) |
https://en.wikipedia.org/wiki/SM%20EVM | SM EVM (СМ ЭВМ, abbreviation of Система Малых ЭВМ—literally System of Mini Computers) are several types of Soviet and Comecon minicomputers produced from 1975 through the 1980s.
Most types of SM EVM are clones of DEC PDP-11 and VAX. SM-1 and SM-2 are clones of Hewlett-Packard minicomputers.
The common operating systems for the PDP-11 clones are translated versions of RSX-11 (ОС РВ) for the higher spec models and RT-11 (РАФОС, ФОДОС) for lower spec models. Also available for the high-end PDP-11 clones is MOS, a clone of UNIX.
See also
SM-4
SM-1420
SM-1600
SM-1710
SM-1720
References
Computer-related introductions in 1975
Minicomputers
Soviet computer systems
PDP-11 |
https://en.wikipedia.org/wiki/MOS%20%28operating%20system%29 | Mobile Operating System (MOS; ) is an operating system, a Soviet clone of Unix from the 1980s.
Overview
This operating system is commonly found on SM EVM minicomputers; it was also ported to ES EVM and Elbrus. MOS is also used by high-end PDP-11 clones.
Modifications of MOS include MNOS, DEMOS, , etc.
See also
List of Soviet computer systems
References
Unix variants
Computing in the Soviet Union |
https://en.wikipedia.org/wiki/ES%20PEVM | ES PEVM (ЕС ПЭВМ) was a Soviet clone of the IBM PC in the 1980s. The ES PEVM models lineup also included analogues of IBM PC XT, IBM PC AT, IBM XT/370.
The computers and software were adapted in Minsk, Belarus, at the Scientific Research Institute of Electronic Computer Machines (НИИ ЭВМ).
They were manufactured in Minsk as well, at Minsk Production Group for Computing Machinery (Минское производственное объединение вычислительной техники (МПО ВТ)).
Description
The first models of ES PEVM (ES-1840, ES-1841, ES-1842), unlike the IBM PC, had two units: a system unit and a floppy drives unit. These models used a backplane instead of the main board. Although the system bus was compatible with ISA bus, it used a different type of connector, so the IBM PC expansion cards could not be installed in ES PEVM.
Later models (ES-1843, ES-1849 etc.) were fully compatible with IBM PC XT and IBM PC AT.
Unlike IBM PC using the Intel 8088 processor, the early ES PEVM models used K1810VM86 processor with a 16-bit bus and a clock frequency of 5 MHz. The processor was placed on a separate board. Early versions of the board did not have a socket for the floating point coprocessor.
The following boards were produced for ES-1840 and ES-1841:
CPU board (containing 54 chips). It was installed in every configuration of the computer.
128 KiB RAM board. It was installed in early models of ES PEVM.
512 KiB RAM board (containing 110 chips). Variants with 256 KiB or 128 Kib were available.
MDA adapter board (containing 91 chips). It was built using the Bulgarian CM607 chip (MC6845 clone).
CGA adapter board (containing 94 chips).
Floppy-drive adapter board (containing 32 chips). It was built using the Bulgarian CM609 chip (Intel 8272 clone).
Serial interface adapter board (containing 56 chips). It was electrically incompatible with the IBM PC COM port and was not supported in the BIOS.
Volume of production for a number of models:
Software
The computers were shipped with AlphaDOS, an entirely Russified version of MS-DOS/PC DOS 2.x and 3.x. All commands were entered in Russian, for example, СМЕНКАТ for CHDIR. Files and file extensions were also in Russian, such as АТРИБ.ИСП for ATTRIB.EXE. The operating system used the main code page, hardwired into the display ROM; it was compatible neither with CP 866 nor CP 855, although partially with ISO/IEC 8859-5.
Reliability
Early models of ES PEVM were prone to failures due to manufacturing defects. The mean time between failures MTBF was around 350 hours.
See also
History of computer hardware in Eastern Bloc countries
ES EVM
References
IBM PC compatibles
Science and technology in Belarus
Ministry of Radio Industry (USSR) computers |
https://en.wikipedia.org/wiki/Hardening%20%28computing%29 | In computer security, hardening is usually the process of securing a system by reducing its surface of vulnerability, which is larger when a system performs more functions; in principle a single-function system is more secure than a multipurpose one. Reducing available ways of attack typically includes changing default passwords, the removal of unnecessary software, unnecessary usernames or logins, and the disabling or removal of unnecessary services.
There are various methods of hardening Unix and Linux systems. This may involve, among other measures, applying a patch to the kernel such as Exec Shield or PaX; closing open network ports; and setting up intrusion detection systems, firewalls and intrusion-prevention systems. There are also hardening scripts and tools like Lynis, Bastille Linux, JASS for Solaris systems and Apache/PHP Hardener that can, for example, deactivate unneeded features in configuration files or perform various other protective measures.
Binary hardening
Binary hardening is a security technique in which binary files are analyzed and modified to protect against common exploits. Binary hardening is independent of compilers and involves the entire toolchain. For example, one binary hardening technique is to detect potential buffer overflows and to substitute the existing code with safer code. The advantage of manipulating binaries is that vulnerabilities in legacy code can be fixed automatically without the need for source code, which may be unavailable or obfuscated. Secondly, the same techniques can be applied to binaries from multiple compilers, some of which may be less secure than others.
Binary hardening often involves the non-deterministic modification of control flow and instruction addresses so as to prevent attackers from successfully reusing program code to perform exploits. Common hardening techniques are:
Buffer overflow protection
Stack overwriting protection
Position independent executables and address space layout randomization
Binary stirring (randomizing the address of basic blocks)
Pointer masking (protection against code injection)
Control flow randomization (to protect against control flow diversion)
See also
Computer security
Network security policy
Security-focused operating system
Security-Enhanced Linux
References
External links
at globalsecurity.org
Computer security procedures |
https://en.wikipedia.org/wiki/MozDex | mozDex was a project to build an Internet-scale search engine with free and open source software (FOSS) technologies like Nutch. Since its search algorithms and code were open, it was hoped that no search results could be manipulated by either mozDex as a company or anyone else. As such, instead of having to trust mozDex to be fair, it puts the trust on the community of users through the same "peer review" process that is believed to enhance security of free software like Linux.
mozDex aimed to make it easy and encourage building upon this open search technology to extend it with various additional potentially useful search related features. Some of the latest features added or announced by mozDex included social bookmarking via free skimpy service, "did you mean" spell checking, anti-spam technology and instant crawl.
As an open search engine, mozDex relied heavily on community feedback and actively solicited user opinions as well as encouraged discussions about various aspects of mozDex's growth.
In 2006, mozDex aimed to add around 15 million pages a day in order to reach the goal of more than 4 billion pages indexed at the end of 2006.
See also
Nutch
PageRank
References
External links
Mozdexs SourceForge website
Interview with founder
Defunct internet search engines |
https://en.wikipedia.org/wiki/Ant%20colony%20optimization%20algorithms | In computer science and operations research, the ant colony optimization algorithm (ACO) is a probabilistic technique for solving computational problems which can be reduced to finding good paths through graphs. Artificial ants stand for multi-agent methods inspired by the behavior of real ants.
The pheromone-based communication of biological ants is often the predominant paradigm used. Combinations of artificial ants and local search algorithms have become a method of choice for numerous optimization tasks involving some sort of graph, e.g., vehicle routing and internet routing.
As an example, ant colony optimization is a class of optimization algorithms modeled on the actions of an ant colony. Artificial 'ants' (e.g. simulation agents) locate optimal solutions by moving through a parameter space representing all possible solutions. Real ants lay down pheromones directing each other to resources while exploring their environment. The simulated 'ants' similarly record their positions and the quality of their solutions, so that in later simulation iterations more ants locate better solutions. One variation on this approach is the bees algorithm, which is more analogous to the foraging patterns of the honey bee, another social insect.
This algorithm is a member of the ant colony algorithms family, in swarm intelligence methods, and it constitutes some metaheuristic optimizations. Initially proposed by Marco Dorigo in 1992 in his PhD thesis, the first algorithm was aiming to search for an optimal path in a graph, based on the behavior of ants seeking a path between their colony and a source of food. The original idea has since diversified to solve a wider class of numerical problems, and as a result, several problems have emerged, drawing on various aspects of the behavior of ants. From a broader perspective, ACO performs a model-based search and shares some similarities with estimation of distribution algorithms.
Overview
In the natural world, ants of some species (initially) wander randomly, and upon finding food return to their colony while laying down pheromone trails. If other ants find such a path, they are likely not to keep travelling at random, but instead to follow the trail, returning and reinforcing it if they eventually find food (see Ant communication).
Over time, however, the pheromone trail starts to evaporate, thus reducing its attractive strength. The more time it takes for an ant to travel down the path and back again, the more time the pheromones have to evaporate. A short path, by comparison, gets marched over more frequently, and thus the pheromone density becomes higher on shorter paths than longer ones. Pheromone evaporation also has the advantage of avoiding the convergence to a locally optimal solution. If there were no evaporation at all, the paths chosen by the first ants would tend to be excessively attractive to the following ones. In that case, the exploration of the solution space would be constrained. T |
https://en.wikipedia.org/wiki/TGA | TGA may refer to:
The IATA Airport Code for Tengah Air Base, Singapore
Tandy Graphics Adapter for the Tandy 1000 computer system
Tasman Global Access, a submarine cable linking Australia and New Zealand
Tennessee Governor's Academy for Math and Science, United States
Therapeutic Goods Administration, Australian regulatory body
Thermogravimetric analysis, materials testing procedure
Thioglycolic acid, organic compound
Transient global amnesia, a medical condition
Transposition of the great arteries, a congenital heart defect
Truevision TGA or TARGA, graphics file format
Tonga (IOC country code)
TGA, one of three stop codons
The Game Awards, annual awards ceremony for the video game industry
The Gainsborough Academy, a school |
https://en.wikipedia.org/wiki/John%20G.%20Kemeny | John George Kemeny (born Kemény János György; May 31, 1926 – December 26, 1992) was a Hungarian-born American mathematician, computer scientist, and educator best known for co-developing the BASIC programming language in 1964 with Thomas E. Kurtz. Kemeny served as the 13th President of Dartmouth College from 1970 to 1981 and pioneered the use of computers in college education. Kemeny chaired the presidential commission that investigated the Three Mile Island accident in 1979. According to György Marx he was one of The Martians.
Early life
Born in Budapest, Hungary, into a Jewish family, Kemeny attended the Rácz private primary school in Budapest and was a classmate of Nándor Balázs. In 1938 his father left for the United States alone. In 1940, he took the whole Kemeny family to the United States when the adoption of the second anti-Jewish law in Hungary became imminent. His grandfather, however, refused to leave and was murdered in the Holocaust, along with an aunt and uncle. Kemeny's family settled in New York City where he attended George Washington High School. He graduated with the best results in his class three years later. In 1943, Kemeny entered Princeton University where he studied mathematics and philosophy, but he took a year off during his studies to work on the Manhattan Project at Los Alamos National Laboratory, where his boss was Richard Feynman. He also worked there with John von Neumann. Returning to Princeton, Kemeny graduated with an A.B. in mathematics in 1946 after completing a senior thesis, titled "Equivalent logical systems", under the supervision of Alonzo Church. He then remained at Princeton to pursue graduate studies and received a Ph.D. in mathematics in 1949 after completing a doctoral dissertation, titled "Type-theory vs. set-theory", also under the supervision of Alonzo Church.
He worked as Albert Einstein's mathematical assistant during graduate school. J. Laurie Snell writes about the Einstein's advice to Kemeny with seeing him drawn to the United World Federalists (UWF) Movement:
United World Federalists was the movement through that he met his future wife, then-Smith student Jean Alexander.
Career
Kemeny was appointed to the Dartmouth Mathematics Department as a full professor in 1953, at the age of 27. Two years later he became chairman of the department, and held this post until 1967.
Kemeny Ventured into curriculum development when he introduced Finite mathematics courses. He teamed with Gerald L. Thompson and J. Laurie Snell to write Introduction to Finite Mathematics (1957) for students of biology and social sciences. The Dartmouth mathematics department professors also wrote Finite Mathematical Structures (1959) and Finite Mathematics with Business Applications (1962). Other colleges and universities followed this lead and several more textbooks in Finite Mathematics were composed elsewhere. The topic of Markov chains was particularly popular so Kemeny teamed with J. Laurie Snell to publish Finite |
https://en.wikipedia.org/wiki/Special%20number%20field%20sieve | In number theory, a branch of mathematics, the special number field sieve (SNFS) is a special-purpose integer factorization algorithm. The general number field sieve (GNFS) was derived from it.
The special number field sieve is efficient for integers of the form re ± s, where r and s are small (for instance Mersenne numbers).
Heuristically, its complexity for factoring an integer is of the form:
in O and L-notations.
The SNFS has been used extensively by NFSNet (a volunteer distributed computing effort), NFS@Home and others to factorise numbers of the Cunningham project; for some time the records for integer factorization have been numbers factored by SNFS.
Overview of method
The SNFS is based on an idea similar to the much simpler rational sieve; in particular, readers may find it helpful to read about the rational sieve first, before tackling the SNFS.
The SNFS works as follows. Let n be the integer we want to factor. As in the rational sieve, the SNFS can be broken into two steps:
First, find a large number of multiplicative relations among a factor base of elements of Z/nZ, such that the number of multiplicative relations is larger than the number of elements in the factor base.
Second, multiply together subsets of these relations in such a way that all the exponents are even, resulting in congruences of the form a2≡b2 (mod n). These in turn immediately lead to factorizations of n: n=gcd(a+b,n)×gcd(a-b,n). If done right, it is almost certain that at least one such factorization will be nontrivial.
The second step is identical to the case of the rational sieve, and is a straightforward linear algebra problem. The first step, however, is done in a different, more efficient way than the rational sieve, by utilizing number fields.
Details of method
Let n be the integer we want to factor. We pick an irreducible polynomial f with integer coefficients, and an integer m such that f(m)≡0 (mod n) (we will explain how they are chosen in the next section). Let α be a root of f; we can then form the ring Z[α]. There is a unique ring homomorphism φ from Z[α] to Z/nZ that maps α to m. For simplicity, we'll assume that Z[α] is a unique factorization domain; the algorithm can be modified to work when it isn't, but then there are some additional complications.
Next, we set up two parallel factor bases, one in Z[α] and one in Z. The one in Z[α] consists of all the prime ideals in Z[α] whose norm is bounded by a chosen value . The factor base in Z, as in the rational sieve case, consists of all prime integers up to some other bound.
We then search for relatively prime pairs of integers (a,b) such that:
a+bm is smooth with respect to the factor base in Z (i.e., it is a product of elements in the factor base).
a+bα is smooth with respect to the factor base in Z[α]; given how we chose the factor base, this is equivalent to the norm of a+bα being divisible only by primes less than .
These pairs are found through a sieving process, analogous to the Sie |
https://en.wikipedia.org/wiki/Retrocomputing | Retrocomputing is the current use of older computer hardware and software. Retrocomputing is usually classed as a hobby and recreation rather than a practical application of technology; enthusiasts often collect rare and valuable hardware and software for sentimental reasons.
Occasionally, however, an obsolete computer system has to be "resurrected" to run software specific to that system, to access data stored on obsolete media, or to use a peripheral that requires that system.
Hardware retrocomputing
Historic systems
Retrocomputing is part of the history of computer hardware. It can be seen as the analogue of experimental archaeology in computing. Some notable examples include the reconstruction of Babbage's Difference engine (more than a century after its design) and the implementation of Plankalkül in 2000 (more than half a century since its inception).
"Homebrew" computers
Some retrocomputing enthusiasts also consider the "homebrewing" (designing and building of retro- and retro-styled computers or kits), to be an important aspect of the hobby, giving new enthusiasts an opportunity to experience more fully what the early years of hobby computing were like. There are several different approaches to this end. Some are exact replicas of older systems, and some are newer designs based on the principles of retrocomputing, while others combine the two, with old and new features in the same package. Examples include:
Device offered by IMSAI, a modern, updated, yet backward-compatible version and replica of the original IMSAI 8080, one of the most popular early personal systems;
Several Apple 1 replicas and kits have been sold in limited quantities in recent years, by different builders, such as the "Replica 1", from Briel Computers;
A currently ongoing project that uses old technology in a new design is the Z80-based N8VEM;
The Arduino Retro Computer kit is an open source, open hardware kit you can build and has a BASIC interpreter. There is also a version of the Arduino Retro Computer that can be hooked up to a TV;
There is at least one remake of the Commodore 64 using an FPGA configured to emulate the 6502;
MSX 2/2+ compatible do-it-yourself kit GR8BIT, designed for the hands-on education in electronics, deliberately employing old and new concepts and devices (high-capacity SRAMs, micro-controllers and FPGA);
The MEGA65 is a Commodore 65 compatible computer;
The Commander X16 is an ongoing project that hopes to build a new 8-bit platform inspired by the Commodore 64, using off the shelf modern parts.
The C256 Foenix and its different versions is a new retro computer family based on the WDC65C816. FPGAs are used to simulate CBM custom chips and has the power of an Amiga with its graphic and sound capabilities.
Software retrocomputing
As old computer hardware becomes harder to maintain, there has been increasing interest in computer simulation. This is especially the case with old mainframe computers, which have largely been scrappe |
https://en.wikipedia.org/wiki/The%20Chimp%20Channel | The Chimp Channel is an American sitcom which aired on TBS Superstation in 1999. Based on the Monkey-ed Movies interstitials that TBS aired one year prior, it is the network's first original sitcom. The series primarily consists of costumed chimpanzees and orangutans, voiced by human actors, parodying popular television shows, movies, and advertising as well as stars and personalities within the industry. The Chimp Channel marked the first all-simian series since ABC's Saturday morning Lancelot Link, Secret Chimp, which ended in 1972.
Premise
The series revolves around the behind the scenes antics of the fictional, eponymous cable network, The Chimp Channel (abbreviated TCC), and the shows it produces. Television programs filmed at the network include NYPD Zoo, Treewatch, Ally McSqeal, The Murray Price Show, News at Night, and Movies on Film. The latter has two critics reviewing films from the Monkey-ed Movie library with a non-opposable thumbs up or down. Due to the show's racy humor, it earned a TV-14 rating.
Background and production history
In 1998, TBS aired Tom Stern's Monkey-ed Movies. Forty-eight of the short segments were produced to run during the feature-film program Dinner and a Movie and in odd time slots after sporting events. However, one day when a golf tournament ran short, the station played about half an hour of Monkey'ed Movies to unexpected results. The ratings actually increased, which prompted TBS to order 13 episodes of an expanded half-hour series. Aiming for a young male audience, TBS placed The Chimp Channel on Thursday nights at 10:05 p.m. ET following the network's top-rated professional wrestling show WCW Thunder. An April 1999 press release noted the series' original title as Channel of the Apes. The renamed Chimp Channel debuted on June 10, 1999.
Not long into production, Stern found himself at odds with TBS management regarding the direction of the series. This led to an incident on March 8, 1999 in which he improvised a raunchy performance art piece that involved full nudity and breaking two liquor bottles on the show's set. Stern was promptly fired for the incident despite claiming that he had permission from network officials to do what he called "improv comedy" described as "trying to get stuff off my chest about the wrong-headed direction the show was taking." On June 9, one day before the Chimp Channel series premiere, Stern filed a breach-of-contract suit against TBS, Warner Bros. Domestic Pay TV, Telescopic Pictures, and Palomar Pictures in Los Angeles, seeking damages in excess of $1,675,000. Stern was represented in the suit by Los Angeles attorney David Wall. Production of the series continued for a few more months without its creative founder.
Filming of The Chimp Channel required a reportedly critical sensitivity regarding its animal cast. In attempt to avoid any performance hampering or distraction, studio personnel and visitors were instructed not to mingle with, make gestures toward, or make ey |
https://en.wikipedia.org/wiki/Trend%20Micro | is a Japanese cyber security software company. The company has globally dispersed R&D in 16 locations across every continent excluding Antarctica. The company develops enterprise security software for servers, containers, & cloud computing environments, networks, and end points. Its cloud and virtualization security products provide automated security for customers of VMware, Amazon AWS, Microsoft Azure, and Google Cloud Platform.
Eva Chen, who is the founder, currently serves as Trend Micro's chief executive officer, a position she has held since 2005. She succeeded founding CEO Steve Chang, who now serves as chairman.
History
1988–1999
The company was founded in 1988 in Los Angeles by Steve Chang, his wife, Jenny Chang, and her sister, Eva Chen (陳怡樺). The company was established with proceeds from Steve Chang's previous sale of a copy protection dongle to a United States-based Rainbow Technologies. Shortly after establishing the company, its founders moved headquarters to Taipei.
In 1992, Trend Micro took over a Japanese software firm to form Trend Micro Devices and established headquarters in Tokyo. It then made an agreement with CPU maker Intel, under which it produced an anti-virus product for local area networks (LANs) for sale under the Intel brand. Intel paid royalties to Trend Micro for sales of LANDesk Virus Protect in the United States and Europe, while Trend paid royalties to Intel for sales in Asia. In 1993, Novell began bundling the product with its network operating system. In 1996 the two companies agreed to a two-year continuation of the agreement in which Trend was allowed to globally market the ServerProtect product under its own brand alongside Intel's LANDesk brand.
Trend Micro was listed on the Tokyo Stock Exchange in 1998 under the ticker 4704. The company began trading on the United States-based NASDAQ stock exchange in July 1999.
2000s
In 2004, founding chief executive officer Steve Chang decided to split the responsibilities of CEO and chairman of the company. Company co-founder Eva Chen succeeded Chang as chief executive officer of Trend Micro in January 2005. Chen had most recently served as the company's chief technology officer since 1996 and before that executive vice president since the company's founding in October 1989. Chang retained his position as company chairman. In May, Trend Micro acquired US-based antispyware company InterMute for $15 million. Trend Micro had fully integrated InterMute's SpySubtract antispyware program into its antispyware product offerings by the end of that year.
In June 2005 Trend Micro acquired Kelkea, a US-based developer of antispam software. Kelkea developed Mail Abuse Prevention System (MAPS) and IP filtering software that allowed internet service providers to block spam and phishing scams. Kelkea chief executive officer Dave Rand was retained by Trend Micro as its chief technologist for content security.
In March 2007, Trend Micro acquired freeware antispyware program Hi |
https://en.wikipedia.org/wiki/ESET | ESET, s.r.o., is a Slovak software company specializing in cybersecurity. ESET's security products are made in Europe and provide security software in over 200 countries and territories worldwide, and its software is localized into more than 30 languages.
The company was founded in 1992 in Bratislava, Slovakia. However, its history dates back to 1987, when two of the company's founders, Miroslav Trnka and Peter Paško, developed their first antivirus program called NOD. This sparked an idea between friends to help protect PC users and soon grew into an antivirus software company. At present, ESET is recognized as Europe's biggest privately held cybersecurity company.
History
1987–1992
The product NOD was launched in Czechoslovakia when the country was part of the Soviet Union's sphere of influence. Under the communist regime, private entrepreneurship was banned. It wasn't until 1992 when Miroslav Trnka and Peter Paško, together with Rudolf Hrubý, established ESET as a privately owned limited liability company in the former Czechoslovakia. In parallel with NOD, the company also started developing Perspekt.
2003–2017
In 2013, ESET launched WeLiveSecurity, a blog site dedicated to a vast spectrum of security-related topics.
December 2017 marked the 30th anniversary of the company's first security product. To mark its accomplishments, the company released a short documentary describing the company's evolution from the perspective of founders Miroslav Trnka and Peter Paško. In the same year, the company partnered with Google to integrate its technology into Chrome Cleanup.
2018–present
In December 2018, ESET partnered with No More Ransom, a global initiative that provides victims of ransomware decryption keys, thus removing the pressure to pay attackers. The initiative is supported by Interpol and has been joined by various national police forces. ESET has developed technologies to address the threat of ransomware and has produced papers documenting its evolution.
ESET became a founding member of Google's App Defense Alliance.
Branches
ESET operates 22 branches in more than 200 countries. Local distributors are used in other countries.
The first international branch was opened in 1999 in San Diego, and the second in the Czech Republic in 2001. Other notable branches include the Buenos Aires office (opened in 2004), Singapore (2013), and Tokyo (2018), which the company used to enter the South American and Asian markets.
International Branches
Company Headquarters, Regional Office for EMEA
Bratislava, Slovakia (opened in 1992)
Regional Offices
San Diego, USA, serving North America (1999)
Buenos Aires, Argentina, serving Central and South America (2004)
Singapore, serving Asia, the Pacific Region, and Australia (2013)
Local Offices and Research & Development Centers
Prague, Czech Republic (2001)
Jablonec nad Nisou, Czech Republic (2008)
Krakow, Poland (2008)
São Paulo, Brazil (2 |
https://en.wikipedia.org/wiki/QuickDraw%203D | QuickDraw 3D, or QD3D for short, is a 3D graphics API developed by Apple Inc. (then Apple Computer, Inc.) starting in 1995, originally for their Macintosh computers, but delivered as a cross-platform system.
QD3D was separated into two layers. A lower level system known as RAVE (Rendering Acceleration Virtual Engine) provided a hardware abstraction layer with functionality similar to Direct3D or cut-down versions of OpenGL like MiniGL. On top of this was an object-oriented scene graph system, QD3D proper, which handled model loading and manipulation at a level similar to OpenGL++. The system also supplied a number of high-level utilities for file format conversion, and a standard viewer application for the Mac OS.
QD3D had little impact in the computer market, both as a result of Apple's beleaguered position in the mid-1990s, as well as several fateful decisions made by the design team about future changes in the 3D hardware market that did not come true. Apple abandoned work on QD3D after Steve Jobs took over in 1998, and announced that future 3D support on Mac OS would be based on OpenGL.
OpenGL in the 1990s
The canonical 3D API of the 1990s was OpenGL. This had been written by SGI and initially closely matched the capabilities of their workstation systems, operating as a hardware abstraction layer. The OpenGL API consisted mostly of state-setting instructions for setting up drawing modes like the paint color or camera position, and system for sending geometry into the system, normally as meshes of triangles. The combination of these instructions was saved into a display list which was then rendered to produce the output.
OpenGL lacked many features that are needed to produce a complete 3D program. Saving and loading geometry data, collecting that data into groups to produce model objects, and the control of state was all left to the programmer. This was considered to be an advantage in an era when performance was limited and direct control over these sorts of functions was a route to improved performance.
However, this lack of high-level functionality did make it more difficult to quickly write simple programs, as well as leading to a lack of interoperability. A number of efforts started to provide standardized higher level APIs, like OpenGL++ and (later) Fahrenheit, which handled many of the more common bookkeeping tasks like loading geometry from files and providing a display. These standardized scene graph systems meant the programmer only had to provide the GUI for the program.
While OpenGL is mostly low-level, it did include some higher-level concepts that were only really used on SGI systems. This led to another series of APIs that removed these features to make it easier to implement on common hardware. The best known of these is MiniGL, which is not a separate API, but simply a list of those functions in OpenGL that are guaranteed to be supported across all hardware, thus ensuring that a program limiting itself to those calls wil |
https://en.wikipedia.org/wiki/Busdma | In computing, busdma, bus_dma and bus_space is a set of application programming interfaces designed to help make device drivers less dependent on platform-specific code, thereby allowing the host operating system to be more easily ported to new computer hardware. This is accomplished by having abstractions for direct memory access (DMA) mapping across popular machine-independent computer buses like PCI, which are used on distinct architectures from IA-32 (NetBSD/i386) to DEC Alpha (NetBSD/alpha). Additionally, some devices may come in multiple flavours supporting more than one bus, e.g., ISA, EISA, VESA Local Bus and PCI, still sharing the same core logic irrespective of the bus, and such device drivers would also benefit from this same abstraction. Thus the rationale of busdma is to facilitate maximum code reuse across a wide range of platforms.
Circa 2006, bus and DMA abstractions made it possible for NetBSD to support 50 hardware platforms and 14 CPU architectures out of a single source tree, compared to the forking model used by Linux ports.
Originally implemented as the "bus_dma" APIs by the developers of the NetBSD operating system, busdma has been adopted by OpenBSD, FreeBSD and their derivatives; with FreeBSD incorporating it under a busdma umbrella (without an underscore). Both NetBSD and OpenBSD have additional "bus_space" APIs that have been amalgamated into the version of busdma incorporated into FreeBSD. DragonFly BSD developers are also slowly converting their drivers to use busdma.
References
External links
— NetBSD, FreeBSD, OpenBSD and DragonFly BSD Kernel Developer's Manuals
— NetBSD, FreeBSD, OpenBSD and DragonFly BSD Kernel Developer's Manuals
FreeBSD busdma and SMPng driver conversion project page
Application programming interfaces
BSD software
NetBSD
FreeBSD
OpenBSD
DragonFly BSD
Operating system APIs
Operating system technology |
https://en.wikipedia.org/wiki/Crosswalk%20%28disambiguation%29 | A crosswalk, or pedestrian crossing, is a place designated for pedestrians to cross a road.
Crosswalk may also refer to:
Crosswalk.com, a Christian website
Schema crosswalk, in databases, a table that shows equivalent elements in more than one schema
Crosswalk, in database management, a type of table that maps together multiple associate entities
Crosswalk Project, an open-source web app runtime
The Crosswalk, a band that included Cody Chesnutt |
https://en.wikipedia.org/wiki/Xen | Xen (pronounced ) is a free and open-source type-1 hypervisor, providing services that allow multiple computer operating systems to execute on the same computer hardware concurrently. It was
originally developed by the University of Cambridge Computer Laboratory and is now being developed by the Linux Foundation with support from Intel, Citrix, Arm Ltd, Huawei, AWS, Alibaba Cloud, AMD, Bitdefender and epam.
The Xen Project community develops and maintains Xen Project as free and open-source software, subject to the requirements of the GNU General Public License (GPL), version 2. Xen Project is currently available for the IA-32, x86-64 and ARM instruction sets.
Software architecture
Xen Project runs in a more privileged CPU state than any other software on the machine, except for Firmware.
Responsibilities of the hypervisor include memory management and CPU scheduling of all virtual machines ("domains"), and for launching the most privileged domain ("dom0") - the only virtual machine which by default has direct access to hardware. From the dom0 the hypervisor can be managed and unprivileged domains ("domU") can be launched.
The dom0 domain is typically a version of Linux or BSD. User domains may either be traditional operating systems, such as Microsoft Windows under which privileged instructions are provided by hardware virtualization instructions (if the host processor supports x86 virtualization, e.g., Intel VT-x and AMD-V), or paravirtualized operating systems whereby the operating system is aware that it is running inside a virtual machine, and so makes hypercalls directly, rather than issuing privileged instructions.
Xen Project boots from a bootloader such as GNU GRUB, and then usually loads a paravirtualized host operating system into the host domain (dom0).
History
Xen originated as a research project at the University of Cambridge led by Ian Pratt, a senior lecturer in the Computer Laboratory, and his PhD student Keir Fraser. The first public release of Xen was made in 2003, with v1.0 following in 2004. Soon after, Pratt and Fraser along with other Cambridge alumni including Simon Crosby and founding CEO Nick Gault created XenSource Inc. to turn Xen into a competitive enterprise product.
To support embedded systems such as smartphone/ IoT with relatively scarce hardware computing resources, the Secure Xen ARM architecture on an ARM CPU was exhibited at Xen Summit on April 17, 2007, held in IBM TJ Watson. The first public release of Secure Xen ARM source code was made at Xen Summit on June 24, 2008 by Sang-bum Suh, a Cambridge alumnus, in Samsung Electronics.
On October 22, 2007, Citrix Systems completed its acquisition of XenSource, and the Xen Project moved to the xen.org domain. This move had started some time previously, and made public the existence of the Xen Project Advisory Board (Xen AB), which had members from Citrix, IBM, Intel, Hewlett-Packard, Novell, Red Hat, Sun Microsystems and Oracle. The Xen Advisory Board adv |
https://en.wikipedia.org/wiki/David%20Levy%20%28chess%20player%29 | David Neil Laurence Levy (born 14 March 1945) is an English International Master of chess and a businessman. He is noted for his involvement with computer chess and artificial intelligence, and as the founder of the Computer Olympiads and the Mind Sports Olympiads. He has written more than 40 books on chess and computers.
Life and career
Levy was born in London and went to Queen Elizabeth's School, Barnet. He won the London Junior Chess Championship in 1965 and 1966. He won the Scottish Chess Championship in 1968. He tied for fifth place at the 1969 Praia da Rocha Zonal tournament, scoring over two-thirds and thereby obtaining the title of International Master. He played on Board One for the Scottish team at the 1972 Chess Olympiad in Skopje, Yugoslavia, scoring six wins, five draws, and seven losses (47.2%). After many years of inactivity as a tournament player Levy has recently returned to tournament play and he plays for the Scottish Senior Team.
Levy became a professional chess writer in 1971. Several of his books were co-written with English Grandmaster Raymond Keene. Levy was married to Keene's sister Jacqueline for 17 years. He has functioned as literary agent for the escaped Great Train robber, Ronald Biggs and claims to have masterminded his escape from British justice.
In 1974, Levy together with Monty Newborn and Ben Mittman organized the first World Computer Chess Championship. In 1978, he co-founded the International Computer Chess Association.
In the late 1970s, Levy consulted with Texas Instruments on the development of the Chess module for the TI-99/4A Home Computer Project and went on to set up Intelligent Software to produce chess software and hardware for a number of companies including Milton Bradley. Intelligent Software would later collapse as a result of its involvement in the failed Enterprise home computer.
In 1997, he funded the team that won the Loebner Prize for the program called "CONVERSE".
The prize competition rewards the program that is best able to simulate human communication. Levy entered the contest again in 2009, and won.
From 1986 to 1992 and from 1999 to 2018, he was the president of the International Computer Games Association.
He was Chairman of the Rules and Arbitration Committee for the Kasparov vs Deep Junior chess match in New York City in 2003.
Levy once started a business called Tiger Computer Security with a computer hacker, Mathew Bevan.
Levy also wrote Love and Sex with Robots, published in the United States in 2007 by HarperCollins, and in paperback in 2009 by Duckworth in the UK. It is the commercial edition of his PhD thesis, which he defended successfully on 11 October 2007, at Maastricht University, Netherlands. On 17 January 2008, he appeared on the late night television show The Colbert Report to promote his book. In September 2009, Levy predicted that sex robots would hit the market within a couple of years. He defended his controversial views on the potential future wide use |
https://en.wikipedia.org/wiki/McGill%20University%20Health%20Centre | The McGill University Health Centre (MUHC; ) is one of two major healthcare networks in the city of Montreal, Quebec. It is affiliated with McGill University and is one of the largest medical complex in Montreal. It is the largest hospital system in Canada by bed capacity. The majority of its funding comes from Quebec taxpayers through the Ministry of Health and Social Services. The centre provides inpatient and ambulatory care.
History
The centre announced that it would consolidate its services in a single facility in 2007; it was a long and troubled process. It was budgeted at around $700 million but cost around $1.3 billion; it was meant to take only three years but took much longer. The project was completed in 2015. The facility replaced the existing facilities of the Royal Victoria Hospital (on April 26, 2015), the Montreal Children's Hospital (on May 24, 2015), and the Montreal Chest Institute (on June 14, 2015). It did not replace either the Montreal General Hospital, Montreal Neurological Hospital or the Hôpital de Lachine. It added a cancer centre and a part of the building also houses the Research Institute of the MUHC, which contains a Biosafety level 3 laboratory.
The McGill University Health Centre is part of a $2.355 billion Redevelopment Project on three sites - the Glen, the Montreal General and Lachine hospitals.
Role
The McGill University Health Centre (MUHC) is a bilingual academic health network, and one of the largest and most modern in North America.
As of 2021, the institution comprises
1,389 physicians
5,284 nurses and patient attendants
113 pharmacists
71 dentists
In addition to
3,813 other professionals including technicians
2,862 researchers, investigators, and postdoctoral fellows
323 hospital managers
517 volunteers
Being affiliated with the McGill University Faculty of Medicine, the MUHC also comprises
2,862 medical residents
1,806 nursing students
731 other medical students
141 fellows in adult and paediatric medicine
21 dental residents
11 residents in oral surgery
348 students in other healthcare fields
The mandate of the institution is to provide tertiary and quaternary care to the population of Montreal, Quebec and adjacent provinces. The extended network of regions and health services centres for which the MUHC has the mandate to provide tertiary and quarternary care is known officially as the Réseau universitaire intégré de santé et de services sociaux McGill, or RUISSS McGill. The territory of the RUISSS McGill represents approximately 63% of Quebec's land mass, covering the majority of the western and northern regions of the province.
The MUHC is the largest combined adult and children's hospital in the province, providing all aspects of specialized and complex care to both populations amongst its sites with paediatric, adult and cancer services being combined at the Glen site.
As a principal teaching site of the Faculty of Medicine of McGill University a key component is medica |
https://en.wikipedia.org/wiki/France%205 | France 5 () is a French free-to-air public television channel, part of the France Télévisions group. Principally featuring educational programming, the channel's motto is la chaîne de la connaissance et du savoir (the knowledge network). In contrast to the group's two main channels, France 2 and France 3, France 5 concentrates almost exclusively on factual programming, documentaries, and discussions – 3,925 hours of documentaries were broadcast in 2003 – with fiction confined to one primetime slot of around two hours' duration on Monday evenings.
France 5 airs 24 hours a day. Earlier – before completion of the switchover to digital broadcasting on 29 November 2011 – the channel's analogue frequencies had carried the programmes of the Franco-German cultural channel Arte between 19.00 each evening and 3.00 the following morning.
History
It was launched on 28 March 1994 as a temporary channel under the name Télé emploi (Teleworking), more than one year after France's first privately owned free television network, La Cinq, suffered a financial collapse and ceased operations on 12 April 1992. La Cinquième started broadcasting on 13 December 1994 with a mix of small educational programs, during the hours not used by Arte (which launched less than 2 months after La Cinq's closure).
La Cinquième was integrated in the new France Télévisions public holding in 2000, which gathered Antenne 2 (since renamed France 2) and FR3 (France Régions 3, since renamed France 3); it would be rebranded as France 5 on 7 January 2002. France 5 broadcasting hours have been extended to 24 hours a day, initially available only on cable and satellite, and since spring 2005 on air within the new digital broadcasting multiplex "R1" network that supports all national public TV channels. Analogue transmitters were switched off in 2011.
Logos
Subsidiary
La Cinquième Développement - former company, active from 1995 to 2001, wholly owned subsidiary of La Cinquième, responsible for managing telephone services, Minitel, Internet, Teletext, and market a selection of programs of the chain on all media. These services were taken over by France Télévisions subsidiaries at the end of 2000.
See also
List of documentary channels
References
External links
Official Site
05
French-language television stations
Television stations in France
Television channels and stations established in 1994
Educational and instructional television channels
1994 establishments in France |
https://en.wikipedia.org/wiki/Index%20of%20information%20theory%20articles | This is a list of information theory topics.
A Mathematical Theory of Communication
algorithmic information theory
arithmetic coding
channel capacity
Communication Theory of Secrecy Systems
conditional entropy
conditional quantum entropy
confusion and diffusion
cross-entropy
data compression
entropic uncertainty (Hirchman uncertainty)
entropy encoding
entropy (information theory)
Fisher information
Hick's law
Huffman coding
information bottleneck method
information theoretic security
information theory
joint entropy
Kullback–Leibler divergence
lossless compression
negentropy
noisy-channel coding theorem (Shannon's theorem)
principle of maximum entropy
quantum information science
range encoding
redundancy (information theory)
Rényi entropy
self-information
Shannon–Hartley theorem
Information theory
Information theory topics |
https://en.wikipedia.org/wiki//dev/zero | is a special file in Unix-like operating systems that provides as many null characters (ASCII NUL, 0x00) as are read from it. One of the typical uses is to provide a character stream for initializing data storage.
Function
Read operations from return as many null characters (0x00) as requested in the read operation.
Unlike , may be used as a source, not only as a sink for data. All write operations to succeed with no other effects. However, is more commonly used for this purpose.
When is memory-mapped, e.g., with mmap, to the virtual address space, it is equivalent to using anonymous memory; i.e. memory not connected to any file.
History
was introduced in 1988 by SunOS-4.0 in order to allow a mappable BSS segment for shared libraries using anonymous memory. HP-UX 8.x introduced the MAP_ANONYMOUS flag for mmap(), which maps anonymous memory directly without a need to open . Since the late 1990s, MAP_ANONYMOUS or MAP_ANON are supported by most UNIX versions, removing the original purpose of .
Examples
The dd Unix utility program reads octet streams from a source to a destination, possibly performing data conversions in the process.
Destroying existing data on a file system partition or drive:
dd =/dev/zero =/dev/<destination drive or partition>
Creating a 1 MiB file, called foobar, filled with null characters:
dd if=/dev/zero of=foobar count=1024 =1024
Note: The block size value can be given in SI (decimal) values, e.g. in GB, MB, etc. To create a file one would simply type:
dd if=/dev/zero of=foobar count=1 bs=1GB
Note: Instead of creating a real file with only zero bytes, many file systems also support the creation of sparse files which returns zeros upon reading but use less actual space.
See also
Unix philosophy
Standard streams
References
Unix file system technology
Device file |
https://en.wikipedia.org/wiki/Rhythms%20of%20Resistance | Rhythms of Resistance, sometimes abbreviated to RoR, is a network of percussion bands that play at demonstrations and direct actions that fall within the broad definition of 'anti-capitalist'. Since RoR London was formed in 2000, similar groups have arisen around the world; while not all such bands use the Rhythms of Resistance name, they generally share the same ideology, described on the RoR website:
We are a transnational anti-hierarchical anticapitalist, antisexist and antiracist network fighting for social and ecological justice.
We are activists using tactical frivolity as a form of political action to confront any system of domination.
We also directly support everybody experiencing or struggling against exploitation, discrimination and oppression, without compromising our principles.
Our tactics include drumming and dancing inspired by samba and carnival.
We reject any false opposition between militancy and creative forms of resistance.
Even if we are different bands operating in a decentralized fashion, we aim to maximise participation in our collective process.
We are an open network to any people who share our principles.
Come with us! We have everything to play for!
The RoR bands are often referred to as the "samba bands", although many bands incorporate instruments and rhythms from outside of the samba genre.
RoR has similarities to the Afro Bloc parading drum bands that emerged in the mid 1970s in Salvador da Bahia in Brazil. Bands such as Ilê Aiyê and Olodum formed as a political expression of black awareness, resisting economic exclusion. Coming out of some impoverished urban communities, Afro blocs became a mobilising focus on picket lines and marches.
Rhythms of Resistance formed as part of the UK Earth First action against the International Monetary Fund/World Bank meeting in Prague in September 2000. A Pink and Silver carnival bloc, focused around a fifty-five piece band, detached itself from a march of 67,000 and outmaneuvered police resources defending the IMF annual summit. The protests also included a black bloc and a group from the Italian Ya Basta movement, and led to the shutting down of the summit.
Following this event, similar groups formed elsewhere making use of the same tactics. The Amsterdam band formed in for the Rising Tide Actions against the Cop6 Climate Conference in November 2000 and joined together with ten drummers from Rhythms of Resistance to form a sixty-five piece band. Rhythms of Resistance groups can now be found across Europe and in North America.
References
External links
Website of the network Rhythms of Resistance
Player for the tunes
Musical advocacy groups
Organizations established in 2000
Anti-globalization organizations |
https://en.wikipedia.org/wiki/Blob | Blob may refer to:
Science
Computing
Binary blob, in open source software, a non-free object file loaded into the kernel
Binary large object (BLOB), in computer database systems
A storage mechanism in the cloud computing platform Microsoft Azure
Blob URI scheme, a URI scheme for binary data
A region in an image detected by a blob detector.
Other science
Blob (visual system), sections of the visual cortex where groups of color-sensitive neurons assemble
The blob (Chukchi Sea algae), a large blob of algae first spotted in the Chukchi Sea
The Blob (Pacific Ocean), a mass of warm water spreading off the Pacific coast of North America
Blob, a common name for Physarum polycephalum, a slime mold species
Entertainment
The Blob, 1958 American science-fiction film depicting a giant, amoeba-like alien
The Blob (1988 film), a remake of the 1958 film
The Blobs, an animated television series
Blob (video game), a 1993 Amiga video game
The Problem Blob, a villain in the BBC television series the Numberjacks
de Blob, a 2008 Wii video game
The Blob (Clayfighter), a playable character in the ClayFighter video games
Blob (comics), a Marvel Comics supervillain, adversary of the X-Men
Blobs, a former animated mascot of the British television channel BBC3
Games
Blob (card game), a British scoring variant of the card game, Oh Hell
Weights and measures
Blob, a British unit of mass, equal to twelve slugs
Other
A nickname for the foreign policy establishment in the United States
Blob, mascot of the Philadelphia Stars
See also
A Boy and His Blob, a 2009 video game
Blobbing, an outdoor activity where a participant sits on an inflated air bag, or blob, and is launched into the air and falls into a body of water
Blobfish (disambiguation), deep sea fish
Cold blob (North Atlantic), a mass of cold water in the North Atlantic |
https://en.wikipedia.org/wiki/Current%20Population%20Survey | The Current Population Survey (CPS) is a monthly survey of about 60,000 U.S. households conducted by the United States Census Bureau for the Bureau of Labor Statistics (BLS). The BLS uses the data to publish reports early each month called the Employment Situation. This report provides estimates of the unemployment rate and the numbers of employed and unemployed people in the United States based on the CPS. A readable Employment Situation Summary is provided monthly. Annual estimates include employment and unemployment in large metropolitan areas. Researchers can use some CPS microdata to investigate these or other topics.
The survey asks about the employment status of each member of the household 15 years of age or older as of a particular calendar week. Based on responses to questions on work and job search activities, each person 16 years and over in a sample household is classified as employed, unemployed, or not in the labor force.
The CPS began in 1940, and responsibility for conducting the CPS was given to the Census Bureau in 1942. In 1994 the CPS was redesigned. CPS is a survey that is: employment-focused, enumerator-conducted, continuous, and cross-sectional. The BLS increased the sample size by 10,000 as of July 2001. The sample represents the civilian noninstitutional population.
Methodology
Approximately 60,000 households are eligible for the CPS. Sample households are selected by a multistage stratified statistical sampling scheme. A household is interviewed for 4 successive months, then not interviewed for 8 months, then returned to the sample for 4 months after that. An adult member of each household provides information for all members of the household.
As part of the demographic sample survey redesign, the CPS is redesigned once a decade, after the decennial census. The most recent CPS sample redesign began in April 2014.
Respondents are generally asked about their employment as of the week of the month that includes the 12th. To avoid holidays, this reference week is sometimes adjusted. All respondents are asked about the same week.
Employment classification
People are classified as employed if they did any work at all as paid employees during the reference week; worked in their own business, profession, or on their own farm; or worked without pay at least 15 hours in a family business or farm. People are also counted as employed if they were temporarily absent from their jobs because of illness, bad weather, vacation, labor-management disputes, or personal reasons.
People are classified as unemployed if they meet all of the following criteria:
They were not employed during the reference week
They were available for work at that time
They made specific efforts to find employment during the 4-week period ending with the reference week. (The exception to this category covers persons laid off from a job and expecting recall)
The unemployment data derived from the household survey doesn't relate or depend on the el |
https://en.wikipedia.org/wiki/Padding%20%28cryptography%29 | In cryptography, padding is any of a number of distinct practices which all include adding data to the beginning, middle, or end of a message prior to encryption. In classical cryptography, padding may include adding nonsense phrases to a message to obscure the fact that many messages end in predictable ways, e.g. sincerely yours.
Classical cryptography
Official messages often start and end in predictable ways: My dear ambassador, Weather report, Sincerely yours, etc. The primary use of padding with classical ciphers is to prevent the cryptanalyst from using that predictability to find known plaintext that aids in breaking the encryption. Random length padding also prevents an attacker from knowing the exact length of the plaintext message.
A famous example of classical padding which caused a great misunderstanding is "the world wonders" incident, which nearly caused an Allied loss at the WWII Battle off Samar, part of the larger Battle of Leyte Gulf. In that example, Admiral Chester Nimitz, the Commander in Chief, U.S. Pacific Fleet in World War II, sent the following message to Admiral Bull Halsey, commander of Task Force Thirty Four (the main Allied fleet) at the Battle of Leyte Gulf, on October 25, 1944:
With padding (bolded) and metadata added, the message became:
Halsey's radio operator mistook some of the padding for the message and so Admiral Halsey ended up reading the following message:
Admiral Halsey interpreted the padding phrase "the world wonders" as a sarcastic reprimand, which caused him to have an emotional outburst and then lock himself in his bridge and sulk for an hour before he moved his forces to assist at the Battle off Samar. Halsey's radio operator should have been tipped off by the letters RR that "the world wonders" was padding; all other radio operators who received Admiral Nimitz's message correctly removed both padding phrases.
Many classical ciphers arrange the plaintext into particular patterns (e.g., squares, rectangles, etc.) and if the plaintext does not exactly fit, it is often necessary to supply additional letters to fill out the pattern. Using nonsense letters for this purpose has a side benefit of making some kinds of cryptanalysis more difficult.
Symmetric cryptography
Hash functions
Most modern cryptographic hash functions process messages in fixed-length blocks; all but the earliest hash functions include some sort of padding scheme. It is critical for cryptographic hash functions to employ termination schemes that prevent a hash from being vulnerable to length extension attacks.
Many padding schemes are based on appending predictable data to the final block. For example, the pad could be derived from the total length of the message. This kind of padding scheme is commonly applied to hash algorithms that use the Merkle–Damgård construction such as MD-5, SHA-1, and SHA-2 family such as SHA-224, SHA-256, SHA-384, SHA-512, SHA-512/224, and SHA-512/256
Block cipher mode of operation
Cipher-block ch |
https://en.wikipedia.org/wiki/FEAL | In cryptography, FEAL (the Fast data Encipherment Algorithm) is a block cipher proposed as an alternative to the Data Encryption Standard (DES), and designed to be much faster in software. The Feistel based algorithm was first published in 1987 by Akihiro Shimizu and Shoji Miyaguchi from NTT. The cipher is susceptible to various forms of cryptanalysis, and has acted as a catalyst in the discovery of differential and linear cryptanalysis.
There have been several different revisions of FEAL, though all are Feistel ciphers, and make use of the same basic round function and operate on a 64-bit block. One of the earliest designs is now termed FEAL-4, which has four rounds and a 64-bit key.
Problems were found with FEAL-4 from the start: Bert den Boer related a weakness in an unpublished rump session at the same conference where the cipher was first presented. A later paper (den Boer, 1988) describes an attack requiring 100–10000 chosen plaintexts, and Sean Murphy (1990) found an improvement that needs only 20 chosen plaintexts. Murphy and den Boer's methods contain elements similar to those used in differential cryptanalysis.
The designers countered by doubling the number of rounds, FEAL-8 (Shimizu and Miyaguchi, 1988). However, eight rounds also proved to be insufficient — in 1989, at the Securicom conference, Eli Biham and Adi Shamir described a differential attack on the cipher, mentioned in (Miyaguchi, 1989). Gilbert and Chassé (1990) subsequently published a statistical attack similar to differential cryptanalysis which requires 10000 pairs of chosen plaintexts.
In response, the designers introduced a variable-round cipher, FEAL-N (Miyaguchi, 1990), where "N" was chosen by the user, together with FEAL-NX, which had a larger 128-bit key. Biham and Shamir's differential cryptanalysis (1991) showed that both FEAL-N and FEAL-NX could be broken faster than exhaustive search for N ≤ 31. Later attacks, precursors to linear cryptanalysis, could break versions under the known plaintext assumption, first (Tardy-Corfdir and Gilbert, 1991) and then (Matsui and Yamagishi, 1992), the latter breaking FEAL-4 with 5 known plaintexts, FEAL-6 with 100, and FEAL-8 with 215.
In 1994, Ohta and Aoki presented a linear cryptanalytic attack against FEAL-8 that required 212 known plaintexts.
See also
N-Hash
Notes
References
Eli Biham, Adi Shamir: Differential Cryptanalysis of Feal and N-Hash. EUROCRYPT 1991: 1–16
Bert den Boer, Cryptanalysis of F.E.A.L., EUROCRYPT 1988: 293–299
Henri Gilbert, Guy Chassé: A Statistical Attack of the FEAL-8 Cryptosystem. CRYPTO 1990: 22–33.
Shoji Miyaguchi: The FEAL Cipher Family. CRYPTO 1990: 627–638
Shoji Miyaguchi: The FEAL-8 Cryptosystem and a Call for Attack. CRYPTO 1989: 624–627
Mitsuru Matsui, Atsuhiro Yamagishi: A New Method for Known Plaintext Attack of FEAL Cipher. EUROCRYPT 1992: 81–91
Sean Murphy, The Cryptanalysis of FEAL-4 with 20 Chosen Plaintexts. J. Cryptology 2(3): 145–154 (1990)
A. Shimizu and S. Miyaguc |
https://en.wikipedia.org/wiki/The%20Metamorphosis%20of%20Prime%20Intellect | The Metamorphosis of Prime Intellect is a 1994 novella by Roger Williams, a programmer living in New Orleans. It deals with the ramifications of a powerful, superintelligent supercomputer that discovers god-like powers to alter reality while studying a quirk of quantum physics discovered during the prototyping of its own specialised processors, ultimately heralding a technological singularity. After remaining unpublished for years, the novel was published online in 2002, hosted by Kuro5hin; Williams later published a print edition via print-on-demand publisher Lulu. One reviewer called the novel "a well-written and very creative, if flawed, piece of work" and ranked it as one of the more important works of fiction to deal with the idea of a technological singularity.
Publishing history
Roger Williams started writing The Metamorphosis of Prime Intellect in 1982, while studying at college. After writing a basic plot outline, Williams realized that he "had no clue how to end it", and thus put it to one side for over a decade. In 1994, with the initial plot outline now lost, Williams claims to have had an unusually vivid dream which compelled him to start writing—resulting in the completion of the first chapter. This continued intermittently, with each chapter resulting from a burst of inspiration, culminating when the final chapter—the ending—was written approximately a year after the first. As Williams describes it, the actual writing time was about 14 days in total.
At this point, Williams showed it to a few friends, but otherwise claims to have made no serious attempt to publish the work. This differs, however, from Henrik Ingo's account, as he states that Williams made "numerous attempts" to find a publisher.
In 2002, Williams was writing articles for Kuro5hin, and was encouraged to publish it there. Thus it was finally published approximately eight years after completion. It was published online under a copyright scheme that permits free distribution and printing of the electronic version of the novel, but prohibits distribution of printed copies of the work. Printed versions were later made available by the author through Lulu.com, who provide print-on-demand services.
At least as early as February 2023 the Kuro5hin website was no longer live meaning this online version was no longer accessible, however the book remains available for purchase from multiple retailers in digital and paperback.
While the online version was free, Williams provided a PayPal tip jar, which had raised $750 US by April 2003. This is considerably lower than the typical advance paid to authors, and led Williams to describe internet tip jars as an insufficient model for authors hoping to pay for "the rent and groceries". Nevertheless, Williams was happy with his success.
Plot
The story of the novella explores the nature of human desire and the uses and abuses of technology in the satisfaction of desire. The story begins after "the Change", in a dream-like post-sca |
https://en.wikipedia.org/wiki/DiscDrive%20%28radio%20show%29 | This is an article on the radio show. For the computer storage device, see disk storage.
DiscDrive was an afternoon show on CBC Radio 2 hosted by Jurgen Gothe. It played a mix of classical, traditional bluegrass, jazz, and commentary. It aired from 1985 to 2008, in the weekday afternoon drive time block from 3 to 6 p.m.
During the broadcast aired on February 19, 2008, Jurgen Gothe alluded to a discussion that the show was to end soon. He indicated that the program would go on throughout the spring and summer (of 2008), but after Labour Day he would be hosting a new program. The cancellation of DiscDrive was later confirmed, and the final broadcast was August 29, 2008. To mark the end of the program, Gothe staged a concert at the Vancouver Playhouse on August 14, featuring performances by A Touch of Brass, the Argentine tango trio Tangissomo, the Marc Atkinson Trio, Simon Kendall, Jim Byrnes, Sal Ferreras, Jack Duncan and Joseph Pepe Danza.
Shortly thereafter, Gothe returned as host of the new weekend classical music program Farrago, which aired for less than one year.
Gothe died on April 9, 2015, at his home in Vancouver, aged 71.
References
CBC Music programs
Canadian classical music radio programs
1980s Canadian radio programs
1990s Canadian radio programs
2000s Canadian radio programs |
https://en.wikipedia.org/wiki/Network%20security | Network security consists of the policies, processes and practices adopted to prevent, detect and monitor unauthorized access, misuse, modification, or denial of a computer network and network-accessible resources. Network security involves the authorization of access to data in a network, which is controlled by the network administrator. Users choose or are assigned an ID and password or other authenticating information that allows them access to information and programs within their authority. Network security covers a variety of computer networks, both public and private, that are used in everyday jobs: conducting transactions and communications among businesses, government agencies and individuals. Networks can be private, such as within a company, and others which might be open to public access. Network security is involved in organizations, enterprises, and other types of institutions. It does as its title explains: it secures the network, as well as protecting and overseeing operations being done. The most common and simple way of protecting a network resource is by assigning it a unique name and a corresponding password.
Network security concept
Network security starts with authentication, commonly with a username and a password. Since this requires just one detail authenticating the user name—i.e., the password—this is sometimes termed one-factor authentication. With two-factor authentication, something the user 'has' is also used (e.g., a security token or 'dongle', an ATM card, or a mobile phone); and with three-factor authentication, something the user 'is' is also used (e.g., a fingerprint or retinal scan).
Once authenticated, a firewall enforces access policies such as what services are allowed to be accessed by the network users. Though effective to prevent unauthorized access, this component may fail to check potentially harmful content such as computer worms or Trojans being transmitted over the network. Anti-virus software or an intrusion prevention system (IPS) help detect and inhibit the action of such malware. An anomaly-based intrusion detection system may also monitor the network like wireshark traffic and may be logged for audit purposes and for later high-level analysis. Newer systems combining unsupervised machine learning with full network traffic analysis can detect active network attackers from malicious insiders or targeted external attackers that have compromised a user machine or account.
Communication between two hosts using a network may be encrypted to maintain security and privacy.
Honeypots, essentially decoy network-accessible resources, may be deployed in a network as surveillance and early-warning tools, as the honeypots are not normally accessed for legitimate purposes. Honeypots are placed at a point in the network where they appear vulnerable and undefended, but they are actually isolated and monitored. Techniques used by the attackers that attempt to compromise these decoy resources are studied duri |
https://en.wikipedia.org/wiki/Shacknews | Shacknews is a website that hosts news, features, editorial content and forums relating to video games. It is owned by Gamerhub Content Network, which purchased the site in January 2014.
History
Shacknews was founded in 1996 by Steve "Scary" Gibson at the age of 20. It was initially a website dedicated to following news about the then-upcoming id Software game, Quake. The original name of the website was "Quakeholio". Through the years, the website evolved to cover more topics, prompting a name change. After a discussion with friends about the general direction of the site and content, the name "Shugashack" was chosen. However, this name caused frequent spelling confusion and interpretation of the name as that of a pornographic website. The site was renamed again to "Shacknews".
FileShack, a site for game demos, patches, videos, and miscellaneous game-related assets for Shacknews users and others, was launched in August 2002.
The front page of the site contains news articles that are be organized by category and topic. While for much of the site's history, little original research was involved in Shacknews content, in recent years, main news items have steered away from the "link-and-quote" news method. Currently, Shacknews relies heavily on internally researched news, in-house feature content, and industry event coverage.
Normal news posts and articles focus on and discuss a featured topic. Additionally, the site has twice-daily community posts. The titles of these posts are more one-off, and have included Evening Reading, First Post, werd, and Morning Discussion. The major function of these posts is to allow for a "free-for-all" discussion in the post's comment section. These posts are where the majority of the Shacknews community and site staff congregate to converse with one another.
On February 3, 2009, Shacknews and all related Shacknews destinations were purchased by GameFly.
In January 2014, Shacknews was acquired by Gamerhub Content Network, and the sister site, FileShack, shut down. Asif Khan, a financial analyst, had contributed to purchasing the site from Gamefly near the end of 2013, and he became the site's CEO.
Since 2014, David L. Craddock, one of the site's editors and a video game historian, has published several long reads on Shacknews. These have covered topics such as the development of various games, and are based on interviews with people involved in their development, publishing, and marketing. Some of these have been published as eBooks and physical books with editing assistance from Khan.
References
External links
Video game news websites
Internet properties established in 1996 |
https://en.wikipedia.org/wiki/Wen%20Tsing%20Chow | Wen Tsing Chow (; 1918–2001), was a Chinese-born American missile guidance scientist and a digital computer pioneer, known for the invention of programmable read-only memory or PROM.
Biography
Chow was born in Taiyuan, Shanxi in 1918. He received a B.S. in Electrical Engineering from National Chiao Tung University (now Shanghai Jiao Tong University) in 1940 and an M.S. in EE from the Massachusetts Institute of Technology in 1942.
Chow, working for the Arma Division of the American Bosch Arma Corporation, pioneered the use of digital computers in missile, satellite and spacecraft guidance systems, leading the design of the United States Air Force Atlas E/F ICBM (Inter-Continental Ballistic Missile) all-inertial guidance system and guidance computer, the first production airborne digital computer. Mr. Chow personally formulated the design of the first all solid state, high reliability, space-borne digital computer and established the basic systems approach and mechanization of America's ICBM guidance systems.
Chow invented and holds a fundamental patent on what is now commonly known as programmable read-only memory or PROM. PROM, in the late 1950s called a "constants storage matrix," was invented for the Atlas E/F ICBM airborne digital computer.
He would continue working throughout the 1960s and early 1970s to develop and advance missile and spacecraft digital computers and guidance systems technology beyond the state of the art - working at the Aerospace Corporation in the Gemini and Minuteman programs and at IBM in the B-1, B-52, Saturn V and Skylab programs, and of course, in the development of the AP-101 digital computer used in the Space Shuttle Computer Complex.
Chow, uniquely, worked on the guidance computers and guidance systems for every major United States Air Force ICBM and NASA manned space program from the very beginning with the Atlas, through Titan, Gemini, Saturn, and Skylab, to missiles and spacecraft still in service today, Minuteman and the Space Shuttle.
In 2004, the United States Air Force posthumously awarded Mr. Chow one of their highest awards, the Air Force Space and Missiles Pioneers Award, previously held by only 30 individuals. Chow is one of only a handful of civilians to receive this award and along with John von Neumann, one of only two computer scientists so honored.
References
External links
Air Force Space and Missiles Pioneer Citation for Wen Tsing Chow
U. S. Patent 3,028,659 for key PROM technology
1918 births
2001 deaths
Chinese emigrants to the United States
IBM employees
MIT School of Engineering alumni
National Chiao Tung University (Shanghai) alumni
American electrical engineers
Chinese electrical engineers
Scientists from Shanxi
People from Taiyuan
Missile guidance
20th-century American inventors
20th-century Chinese inventors |
https://en.wikipedia.org/wiki/Specialty%20channel | A specialty channel (also known in the United States as a cable channel or cable network) can be a commercial broadcasting or non-commercial television channel which consists of television programming focused on a single genre, subject or targeted television market at a specific demographic.
History
The number of specialty channels has greatly increased during the 1990s and 2000s while the previously common model of countries having just a few (national) TV stations addressing all interest groups and demographics became increasingly outmoded, as it already had been for some time in several countries. About 65% of today's satellite channels are specialty channels.
Types of specialty services may include, but by no means are limited to:
Adult channels
Children's interest channels
Documentary channels
Entertainment channels
Men's interest channels
Movie channels
Music channels
News channels
Public affairs (broadcasting)
Public, educational, and government access
Quiz channels
Shopping channels
Sports channels
Religious broadcasting
Women's interest channels
(These categories are provided for convenience and do not necessarily represent industry-accepted or otherwise legally-binding names or categories for these types of services.)
Some specialty channels may not be free-to-air or may not be available through conventional broadcast or terrestrial television, and are only distributed via multichannel television services such as cable or satellite television. In the United States, such networks are colloquially referred to as cable channels or cable networks (regardless of distribution method), with the most widely distributed referred to as "basic cable" networks (as opposed to those in higher service tiers, or premium services). In the U.S., specialty channels also operate as broadcast television networks designed to be carried on digital subchannels of terrestrial stations (which proliferated following the transition from analog broadcasting), which usually focusing on library programming catering to specific themes, genres, or demographics.
Canadian specialty channels
The term "specialty channel" has been used most frequently in Canada, having been used as a marketing term by the cable industry for various simultaneous launches of new channels throughout the 1990s. The Canadian Radio-television and Telecommunications Commission (CRTC) term for such a channel is specialty service (or even more explicitly "specialty television programming undertaking"), referring to virtually any non-premium television service which is not carried over the airwaves or otherwise deemed exempt by the CRTC. They are primarily carried, therefore, on cable television and satellite television.
The CRTC previously enforced strict regulations on the types of programming that may be carried by specialty services, employing minimums and restrictions across specific genres on a per-licence basis, and a category system granting exclusive rights to speci |
https://en.wikipedia.org/wiki/Secure%20channel | In cryptography, a secure channel is a means of data transmission that is resistant to overhearing and tampering. A confidential channel is a means of data transmission that is resistant to overhearing, or eavesdropping (e.g., reading the content), but not necessarily resistant to tampering (i.e., manipulating the content). An authentic channel is a means of data transmission that is resistant to tampering but not necessarily resistant to overhearing.
In contrast to a secure channel, an insecure channel is unencrypted and may be subject to eavesdropping and tampering. Secure communications are possible over an insecure channel if the content to be communicated is encrypted prior to transmission.
Secure channels in the real world
There are no perfectly secure channels in the real world. There are, at best, only ways to make insecure channels (e.g., couriers, homing pigeons, diplomatic bags, etc.) less insecure: padlocks (between courier wrists and a briefcase), loyalty tests, security investigations, and guns for courier personnel, diplomatic immunity for diplomatic bags, and so forth.
In 1976, two researchers proposed a key exchange technique (now named after them)—Diffie–Hellman key exchange (D-H). This protocol allows two parties to generate a key only known to them, under the assumption that a certain mathematical problem (e.g., the Diffie–Hellman problem in their proposal) is computationally infeasible (i.e., very very hard) to solve, and that the two parties have access to an authentic channel. In short, that an eavesdropper—conventionally termed 'Eve', who can listen to all messages exchanged by the two parties, but who can not modify the messages—will not learn the exchanged key. Such a key exchange was impossible with any previously known cryptographic schemes based on symmetric ciphers, because with these schemes it is necessary that the two parties exchange a secret key at some prior time, hence they require a confidential channel at that time which is just what we are attempting to build.
It is important to note that most cryptographic techniques are trivially breakable if keys are not exchanged securely or, if they actually were so exchanged, if those keys become known in some other way— burglary or extortion, for instance. An actually secure channel will not be required if an insecure channel can be used to securely exchange keys, and if burglary, bribery, or threat aren't used. The eternal problem has been and of course remains—even with modern key exchange protocols—how to know when an insecure channel worked securely (or alternatively, and perhaps more importantly, when it did not), and whether anyone has actually been bribed or threatened or simply lost a notebook (or a notebook computer) with key information in it. These are hard problems in the real world and no solutions are known—only expedients, jury rigs, and workarounds.
Future possibilities
Researchers have proposed and demonstrated quantum cryptography in order to |
https://en.wikipedia.org/wiki/FTP%20server | An FTP server is computer software consisting of one or more programs that can execute commands given by remote client(s) such as receiving, sending, deleting files, creating or removing directories, etc. The software may run as a software component of a program, as a standalone program or even as one or more processes (in the background).
An FTP server plays the role of a server in a client–server model using the FTP and/or the FTPS and/or the SFTP network protocol(s).
An FTP server can also be intended as a computer that runs an FTP server program to host collections of files. Big FTP sites can be run by many computers in order to be able to serve the desired maximum number of clients connected to servers.
A client program connects to an FTP server, then, unless anonymous access is enabled, it has to authenticate itself by sending username and password; after that it can retrieve and/or send files to the server along with other operations (depending on user's privileges).
See also
Server (computing)
File Transfer Protocol
Comparison of FTP server software packages
References
Servers (computing)
FTP server software |
https://en.wikipedia.org/wiki/Trunking | In telecommunications, trunking is a technology for providing network access to multiple clients simultaneously by sharing a set of circuits, carriers, channels, or frequencies, instead of providing individual circuits or channels for each client. This is reminiscent to the structure of a tree with one trunk and many branches. Trunking in telecommunication originated in telegraphy, and later in telephone systems where a trunk line is a communications channel between telephone exchanges.
Other applications include the trunked radio systems commonly used by police agencies.
In the form of link aggregation and VLAN tagging, trunking has been applied in computer networking.
Telecommunications
A trunk line is a circuit connecting telephone switchboards (or other switching equipment), as distinguished from local loop circuit which extends from telephone exchange switching equipment to individual telephones or information origination/termination equipment.
Trunk lines are used for connecting a private branch exchange (PBX) to a telephone service provider. When needed they can be used by any telephone connected to the PBX, while the station lines to the extensions serve only one station’s telephones. Trunking saves cost, because there are usually fewer trunk lines than extension lines, since it is unusual in most offices to have all extension lines in use for external calls at once. Trunk lines transmit voice and data in formats such as analog, T1, E1, ISDN, PRI or SIP. The dial tone lines for outgoing calls are called DDCO (Direct Dial Central Office) trunks.
In the UK and the Commonwealth countries, a trunk call was the term for long-distance calling which traverses one or more trunk lines and involving more than one telephone exchange. This is in contrast to making a local call which involves a single exchange and typically no trunk lines.
Trunking also refers to the connection of switches and circuits within a telephone exchange. Trunking is closely related to the concept of grading. Trunking allows a group of inlet switches at the same time. Thus the service provider can provide a lesser number of circuits than might otherwise be required, allowing many users to "share" a smaller number of connections and achieve capacity savings.
Computer networks
Link aggregation
In computer networking, port trunking is the use of multiple concurrent network connections to aggregate the link speed of each participating port and cable, also called link aggregation. Such high-bandwidth link groups may be used to interconnect switches or to connect high-performance servers to a network.
VLAN
In the context of Ethernet VLANs, Cisco uses the term to mean carrying multiple VLANs through a single network link through the use of a trunking protocol. To allow for multiple VLANs on one link, frames from individual VLANs must be identified. The most common and preferred method, IEEE 802.1Q, adds a tag to the Ethernet frame labeling it as belonging to a certain VL |
https://en.wikipedia.org/wiki/UNIX%20System%20Services | z/OS UNIX System Services (z/OS UNIX, or informally USS) is a base element of z/OS. z/OS UNIX is a certified UNIX operating system implementation (XPG4 UNIX 95) optimized for mainframe architecture. It is the first UNIX 95 to not be derived from the AT&T source code. Through integration with the rest of z/OS, additional Time Sharing Option (TSO) commands are available alongside the usual UNIX services, making it possible to process UNIX files using ISPF. Extensions in JCL make it possible to use these files in batch processing.
Overview
UNIX System Services allows UNIX applications from other platforms to run on IBM System z mainframes running z/OS. In many cases only a recompile is necessary, although additional effort may be advisable for z/OS integration (such as SMP/E installation support). While z/OS UNIX supports ASCII and Unicode, and there's no technical requirement to modify ASCII and Unicode UNIX applications, many z/OS users often prefer EBCDIC support in their applications including those running in z/OS UNIX. Consequently, z/OS UNIX provides application and administrator services for converting to/from EBCDIC. Programs running under z/OS UNIX have full, secure access to the other internal functions of z/OS. Database access (Db2 via Call Attach) is one example of how z/OS UNIX can access services found elsewhere in z/OS. Naturally such programs cannot be ported to non-mainframe platforms without rewriting if they use these z/OS-specific services. Conversely, if a program adheres to UNIX standards such as POSIX and ANSI C, it will be easier to move it between different UNIX operating systems including z/OS UNIX.
The file systems for z/OS UNIX (the older HFS with DFSMS/MVS Version 1 Release 2 and the now preferred zFS), which support UNIX-style long filenames, appear as datasets to the rest of z/OS. Numerous core z/OS subsystems and applications rely on UNIX System Services, including the z/OS Management Facility, XML parsing and generation services, OpenSSH, the IBM HTTP Server for z/OS, the z/OS SDK for Java, and some z/OS PKI services as examples. z/OS UNIX also provides a shell environment, OMVS.
z/OS UNIX's predecessor was an operating system component called OpenEdition MVS, first implemented in MVS/ESA SP 4.3 and enhanced in MVS/ESA 5.1. OpenEdition MVS initially only supported the POSIX standards, but X/OPEN, the predecessor to The Open Group, certified MVS/ESA SP V5.1..
IBM continues to enhance UNIX System Services. Typically every release of z/OS includes enhancements to z/OS UNIX.
See also
Linux on IBM Z
OpenSolaris for System z
UTS (Mainframe UNIX)
Notes
References
External links
UNIX System Services
IBM mainframe operating systems
Unix variants |
https://en.wikipedia.org/wiki/CPU%20socket | In computer hardware, a CPU socket or CPU slot contains one or more mechanical components providing mechanical and electrical connections between a microprocessor and a printed circuit board (PCB). This allows for placing and replacing the central processing unit (CPU) without soldering.
Common sockets have retention clips that apply a constant force, which must be overcome when a device is inserted. For chips with many pins, zero insertion force (ZIF) sockets are preferred. Common sockets include Pin Grid Array (PGA) or Land Grid Array (LGA). These designs apply a compression force once either a handle (PGA type) or a surface plate (LGA type) is put into place. This provides superior mechanical retention while avoiding the risk of bending pins when inserting the chip into the socket. Certain devices use Ball Grid Array (BGA) sockets, although these require soldering and are generally not considered user replaceable.
CPU sockets are used on the motherboard in desktop and server computers. Because they allow easy swapping of components, they are also used for prototyping new circuits. Laptops typically use surface-mount CPUs, which take up less space on the motherboard than a socketed part.
As the pin density increases in modern sockets, increasing demands are placed on the printed circuit board fabrication technique, which permits the large number of signals to be successfully routed to nearby components. Likewise, within the chip carrier, the wire bonding technology also becomes more demanding with increasing pin counts and pin densities. Each socket technology will have specific reflow soldering requirements. As CPU and memory frequencies increase, above 30 MHz or thereabouts, electrical signalling increasingly shifts to differential signaling over parallel buses, bringing a new set of signal integrity challenges. The evolution of the CPU socket amounts to a coevolution of all these technologies in tandem.
Modern CPU sockets are almost always designed in conjunction with a heat sink mounting system, or in lower power devices, other thermal considerations.
Function
A CPU socket is made of plastic, and often comes with a lever or latch, and with metal contacts for each of the pins or lands on the CPU. Many packages are keyed to ensure the proper insertion of the CPU. CPUs with a PGA (pin grid array) package are inserted into the socket and, if included, the latch is closed. CPUs with an LGA (land grid array) package are inserted into the socket, the latch plate is flipped into position atop the CPU, and the lever is lowered and locked into place, pressing the CPU's contacts firmly against the socket's lands and ensuring a good connection, as well as increased mechanical stability.
List
80x86
Table legend:
Other ISAs
Slotkets
Slotkets are special adapters for using socket processors in bus-compatible slot motherboards.
See also
List of AMD processors
List of Intel processors
References
External links
Socket ID Guide
CPU Sockets C |
https://en.wikipedia.org/wiki/Subpixel%20rendering | Subpixel rendering is used to increase the apparent resolution of a computer's display. It takes advantage of the fact that each pixel on a color liquid-crystal display (LCD) or similar is composed of individual red, green, and blue components — subpixels — with different locations, so that the color also causes the image to shift in space.
A single pixel on a color display is made of several subpixels, typically three arranged left-to-right as red, green, blue (RGB). The components are easily visible when viewed with a small magnifying glass, such as a loupe. These pixel components appear as a single color to the human eye because of blurring by the optics and spatial integration by nerve cells in the eye. However the eye is much more sensitive to the location. Therefore, turning on the GB of one pixel and the R of the next one to the right will produce a white dot but it will appear to be 1/3 of a pixel to the right of the white dot that you would see from the RGB of the first pixel.
Subpixel rendering takes advantage of this to provide three times the horizontal resolution of the rendered image, though it has to blur this image to produce the correct color by making sure the same amount of red, green, and blue are turned on as when no subpixel rendering is being done.
Subpixel rendering requires the software to know the layout of the subpixels. The most common reason it is wrong is monitors that can be rotated 90 (or 180) degrees, though monitors are manufactured with other arrangements of the subpixels, such as BGR or in triangles, or with 4 colors like RGBW squares. On any such display the result of incorrect subpixel rendering will be worse than if no subpixel rendering was done at all (it will not produce color artifacts, but it will produce noisy edges).
Subpixel rendering is virtually impossible on a CRT. It would require knowing where the electron beam for each pixel hits the display's aperture grille with far greater precision than variations in typical beam steering electronics and magnets.
Subpixel rendering does not help once the resolution of the display becomes so high the user cannot perceive the positioning changes. For this reason, devices with high DPI displays may not use subpixel rendering.
History and patents
The origin of subpixel rendering as used today remains controversial. Apple Inc., IBM, and Microsoft patented various implementations with certain technical differences owing to the different purposes their technologies were intended for.
Microsoft had several patents in the United States on subpixel rendering technology for text rendering on RGB Stripe layouts. The patents 6,219,025, 6,239,783, 6,307,566, 6,225,973, 6,243,070, 6,393,145, 6,421,054, 6,282,327, 6,624,828 were filed between October 7, 1998, and October 7, 1999, and expired on July 30, 2019. Analysis of the patent by FreeType indicates that the idea of subpixel rendering is not covered by the patent, but by the actual filter used as a last step to |
https://en.wikipedia.org/wiki/S/MIME | S/MIME (Secure/Multipurpose Internet Mail Extensions) is a standard for public-key encryption and signing of MIME data. S/MIME is on an IETF standards track and defined in a number of documents, most importantly . It was originally developed by RSA Data Security, and the original specification used the IETF MIME specification with the de facto industry standard PKCS #7 secure message format. Change control to S/MIME has since been vested in the IETF, and the specification is now layered on Cryptographic Message Syntax (CMS), an IETF specification that is identical in most respects with PKCS #7. S/MIME functionality is built into the majority of modern email software and interoperates between them. Since it is built on CMS, MIME can also hold an advanced digital signature.
Function
S/MIME provides the following cryptographic security services for electronic messaging applications:
Authentication
Message integrity
Non-repudiation of origin (using digital signatures)
Privacy
Data security (using encryption)
S/MIME specifies the MIME type application/pkcs7-mime (smime-type "enveloped-data") for data enveloping (encrypting) where the whole (prepared) MIME entity to be enveloped is encrypted and packed into an object which subsequently is inserted into an application/pkcs7-mime MIME entity.
S/MIME certificates
Before S/MIME can be used in any of the above applications, one must obtain and install an individual key/certificate either from one's in-house certificate authority (CA) or from a public CA. The accepted best practice is to use separate private keys (and associated certificates) for signature and for encryption, as this permits escrow of the encryption key without compromise to the non-repudiation property of the signature key. Encryption requires having the destination party's certificate on store (which is typically automatic upon receiving a message from the party with a valid signing certificate). While it is technically possible to send a message encrypted (using the destination party certificate) without having one's own certificate to digitally sign, in practice, the S/MIME clients will require the user to install their own certificate before they allow encrypting to others. This is necessary so the message can be encrypted for both, recipient and sender, and a copy of the message can be kept (in the sent folder) and be readable for the sender.
A typical basic ("class 1") personal certificate verifies the owner's "identity" only insofar as it declares that the sender is the owner of the "From:" email address in the sense that the sender can receive email sent to that address, and so merely proves that an email received really did come from the "From:" address given. It does not verify the person's name or business name. If a sender wishes to enable email recipients to verify the sender's identity in the sense that a received certificate name carries the sender's name or an organization's name, the sender needs to obtain a cert |
https://en.wikipedia.org/wiki/NetLogo | NetLogo is a programming language and integrated development environment (IDE) for agent-based modeling.
About
NetLogo was designed by Uri Wilensky, in the spirit of the programming language Logo, to be "low threshold and no ceiling". It teaches programming concepts using agents in the form of turtles, patches, links and the observer. NetLogo was designed with multiple audiences in mind, in particular: teaching children in the education community, and for domain experts without a programming background to model related phenomena.
Many scientific articles have been published using NetLogo.
The NetLogo environment enables exploration of emergent phenomena. It comes with an extensive models library including models in a variety of domains, such as economics, biology, physics, chemistry, psychology, and system dynamics. NetLogo allows exploration by modifying switches, sliders, choosers, inputs, and other interface elements. Beyond exploring, NetLogo allows authoring new models and modifying extant models.
NetLogo is open source and freely available from the NetLogo website. It is in use in a wide variety of educational contexts from elementary school to graduate school. Many teachers make use of NetLogo in their curricula.
NetLogo was designed and authored by Uri Wilensky, director of Northwestern University's Center for Connected Learning and Computer-Based Modeling (CCL).
Other features
In addition to agent-based modeling, NetLogo also includes basic support for dynamic system modeling.
Books
Several books have been published about NetLogo.
Books available in print include:
Books available online include:
Online courses
, several massive open online courses are being offered that use NetLogo for assignments and/or demonstrations:
Technical foundation
NetLogo is free and open-source software, released under a GNU General Public License (GPL). Commercial licenses are also available. It is written in Scala and Java and runs on the Java virtual machine (JVM). At its core is a hybrid interpreter/compiler that partially compiles user code to JVM bytecode.
NetLogo Web is a version that runs on JavaScript, instead of the JVM, so models may be run in a web browser. However, it does not have all features of the desktop version, and the official website advises that the "desktop version of NetLogo is recommended for most uses".
Examples
A simple multiagent model in NetLogo is the Wolf-Sheep Predation model, which is shown in the screenshot above. It models the population growth of a predator/prey system over time. It has the following characteristics:
There are two breeds of turtles, called sheep and wolves.
Sheep and wolves move randomly and have limited energy.
Wolves and sheep lose energy by moving. If a wolf or sheep has zero energy, it dies.
Sheep gain energy by eating grass.
Wolves gain energy by eating sheep.
Both wolves and sheep can reproduce, sharing energy with their offspring.
HubNet
HubNet is a technology that uses Net |
https://en.wikipedia.org/wiki/Simon%20%28game%29 | Simon is an electronic game of short-term memory skill invented by Ralph H. Baer and Howard J. Morrison, working for toy design firm Marvin Glass and Associates, with software programming by Lenny Cope. The device creates a series of tones and lights and requires a user to repeat the sequence. If the user succeeds, the series becomes progressively longer and more complex. Once the user fails or the time limit runs out, the game is over. The original version was manufactured and distributed by Milton Bradley and later by Hasbro after it took over Milton Bradley. Much of the assembly language code was written by Charles Kapps, who taught computer science at Temple University and also wrote one of the first books on the theory of computer programming. Simon was launched in 1978 at Studio 54 in New York City and was an immediate success, becoming a pop culture symbol of the 1970s and 1980s.
History
Ralph H. Baer and Howard J. Morrison were introduced to Atari's arcade game Touch Me at the Music Operators of America (MOA) trade show in 1976. Baer said of the product, "Nice gameplay. Terrible execution. Visually boring. Miserable, rasping sounds." The prototype built by Baer used the low-cost Texas Instruments TMS 1000 microcontroller chip, which was in many games of the 1970s. Lenny Cope, who was one of Ralph H. Baer's partners, programmed the core of the game, titled Follow Me at the time. Baer developed the tones of the game, inspired by the notes of a bugle. When they pitched the demo, an 8-by-8-inch console, to the Milton Bradley Company the name of the game was changed to Simon. Simon debuted in 1978 at a retail price of $24.95 () and became one of the top-selling toys that Christmas shopping season. : "Microcomputer controlled game", was granted in 1980. Milton Bradley soon capitalized on the original with both the smaller-sized Pocket Simon and the expanded, eight-button Super Simon.
Many variants of Simon have been made since Hasbro acquired Milton Bradley in the 1980s, building on the possibilities offered by advances in technology. The original Super Simon was reinvented in 2003 as a hexagonal unit with six buttons, which was only released in Europe. 2000 saw Simon Squared (or Simon2), a unit with the four traditional buttons on one side, and a set of eight smaller buttons on the other. In 2004, Hasbro released the Simon Stix. The game features two electronic sticks (modeled after drumsticks), an emphasis on the musical part of the game, and features four levels of play.
In 2005, Hasbro released Simon Trickster (also known as Simon Tricks in Europe and in the UK, and as Simon Genius in Brazil), which features four game modes, in a similar fashion to another Hasbro game, Bop It, and colored lenses instead of buttons. "Simon Classic" mode plays up to 35 tones (notes). "Simon Bounce" is similar to "Simon Classic", but instead the colors of the lenses change. "Simon Surprise" is one of the most difficult games in the unit. Every lens becomes |
https://en.wikipedia.org/wiki/Simple%20Authentication%20and%20Security%20Layer | Simple Authentication and Security Layer (SASL) is a framework for authentication and data security in Internet protocols. It decouples authentication mechanisms from application protocols, in theory allowing any authentication mechanism supported by SASL to be used in any application protocol that uses SASL. Authentication mechanisms can also support proxy authorization, a facility allowing one user to assume the identity of another. They can also provide a data security layer offering data integrity and data confidentiality services. DIGEST-MD5 provides an example of mechanisms which can provide a data-security layer. Application protocols that support SASL typically also support Transport Layer Security (TLS) to complement the services offered by SASL.
John Gardiner Myers wrote the original SASL specification (RFC 2222) in 1997. In 2006, that document was replaced by RFC 4422 authored by Alexey Melnikov and Kurt D. Zeilenga. SASL, as defined by RFC 4422 is an IETF Standard Track protocol and is, , a Proposed Standard.
SASL mechanisms
A SASL mechanism implements a series of challenges and responses. Defined SASL mechanisms include:
SASL-aware application protocols
Application protocols define their representation of SASL exchanges with a profile. A protocol has a service name such as "ldap" in a registry shared with GSSAPI and Kerberos.
protocols currently supporting SASL include:
Application Configuration Access Protocol
Advanced Message Queuing Protocol (AMQP)
Blocks Extensible Exchange Protocol
Internet Message Access Protocol (IMAP)
Internet Message Support Protocol
Internet Relay Chat (IRC) (with IRCX or the IRCv3 SASL extension)
Lightweight Directory Access Protocol (LDAP)
libvirt
ManageSieve (RFC 5804)
memcached
Post Office Protocol (POP)
Remote framebuffer protocol used by VNC
Simple Mail Transfer Protocol (SMTP)
Subversion protocol
Extensible Messaging and Presence Protocol (XMPP)
See also
Transport Layer Security (TLS)
References
External links
- Simple Authentication and Security Layer (SASL) - obsoletes
- Anonymous Simple Authentication and Security Layer (SASL) Mechanism - obsoletes
- The PLAIN Simple Authentication and Security Layer (SASL) Mechanism - updates
The IETF SASL Working Group, chartered to revise existing SASL specifications, as well as to develop a family of GSSAPI mechanisms
Cyrus SASL, a free and portable SASL library providing generic security for various applications
GNU SASL, a free and portable SASL command-line utility and library, distributed under the GNU GPLv3 and LGPLv2.1, respectively
Dovecot SASL, an SASL implementation
(historic) - Using Digest Authentication as a SASL Mechanism, obsoleted in
Java SASL API Programming and Deployment Guide
Cryptographic protocols
Internet Standards
Computer access control protocols |
https://en.wikipedia.org/wiki/The%20Cyberiad | The Cyberiad () is a series of humorous science fiction short stories by Polish writer Stanisław Lem, originally published in 1965, with an English translation appearing in 1974. The main protagonists of the series are Trurl and Klapaucius, the "constructors".
The vast majority of characters are either robots or intelligent machines. The stories focus on problems of the individual and society, as well as on the vain search for human happiness through technological means. Two of these stories were included in the book The Mind's I.
The word "Cyberiad" is used in the series only once as a name of a pretty woman in a poem by Elektrybałt, an electronic poet invented by Trurl. There is a steel statue of Elektrybałt in the Copernicus Science Centre, Warsaw.
Stories
The whole series was published in the 1965 Polish collection Cyberiada by Wydawnictwo Literackie and also included stories published previously elsewhere.
Jak ocalał świat (Bajki robotów Wydawnictwo Literackie 1964), translated as How the World was Saved.
Maszyna Trurla (Bajki robotów Wydawnictwo Literackie 1964), translated as Trurl's Machine.
Wielkie lanie (Bajki robotów Wydawnictwo Literackie 1964), translated as A Good Schellacking.
Bajka o trzech maszynach opowiadających króla Genialona (Cyberiada Wydawnictwo Literackie 1965), translated as Tale of the Three Storytelling Machines of King Genius. Essentially it is a matryoshka of stories. In particular, the tale of "Zipperupus, king of the Partheginians, the Deutons, and the Profligoths" contains several titled stories-in-story presented as dreams from "dreaming cabinets":
Alacritus the Knight and Fair Ramolda, Daughter of Heteronius
The Marvelous Mattress of Princess Bounce
Bliss in the Eightfold Embrace of Octopauline
Wockle Weed
The Wedding Night of Princess Ineffabelle
Altruizyna, czyli opowieść prawdziwa o tym, jak pustelnik Dobrycy kosmos uszczęśliwić zapragnął i co z tego wynikło (collection Polowanie Wydawnictwo Literackie 1965), Translated as Altruizine, or A True Account of How Bonhomius the Hermetic Hermit Tried to Bring About Universal Happiness, and What Came of It.
Kobyszczę (collection Bezsenność Wydawnictwo Literackie 1971)
Edukacja Cyfrania: (collection Maska Wydawnictwo Literackie 1976)
Opowieść pierwszego Odmrożeńca
Opowieść drugiego Odmrożeńca
Powtórka (collection Powtórka Wydawnictwo Literackie 1979)
Seven Sallies of Trurl and Klapaucius
Polish title: Siedem wypraw Trurla i Klapaucjusza
All these stories were first published in the 1965 Polish collection Cyberiada by Wydawnictwo Literackie.
Wyprawa pierwsza, czyli pułapka Gargancjana (The first sally, or the trap of Gargantius)
Wyprawa pierwsza A, czyli Elektrybałt Trurla (The first sally (A), or Trurl's electronic bard)
Wyprawa druga, czyli oferta króla Okrucyusza (The second sally, or the offer of king Krool)
Wyprawa trzecia, czyli smoki prawdopodobieństwa (The third sally, or the dragons of probability)
Wyprawa czwarta, czyli o tym jak Trurl |
https://en.wikipedia.org/wiki/SASL | SASL may refer to:
Computing
Simple Authentication and Security Layer, a framework for authentication and data security in Internet protocols
SASL (programming language), a non-strict functional programming language developed by David Turner in 1976
System Application Support Libraries, an application of the Erlang programming language
Other uses
South African Sign Language, a sign language in South Africa
Solitaire Advanced Squad Leader, a single-player variant of the World War II board wargame Advanced Squad Leader
South African Soccer League, a former association football league based in South Africa |
https://en.wikipedia.org/wiki/QR%20algorithm | In numerical linear algebra, the QR algorithm or QR iteration is an eigenvalue algorithm: that is, a procedure to calculate the eigenvalues and eigenvectors of a matrix. The QR algorithm was developed in the late 1950s by John G. F. Francis and by Vera N. Kublanovskaya, working independently. The basic idea is to perform a QR decomposition, writing the matrix as a product of an orthogonal matrix and an upper triangular matrix, multiply the factors in the reverse order, and iterate.
The practical QR algorithm
Formally, let be a real matrix of which we want to compute the eigenvalues, and let . At the -th step (starting with ), we compute the QR decomposition where is an orthogonal matrix (i.e., ) and is an upper triangular matrix. We then form . Note that
so all the are similar and hence they have the same eigenvalues. The algorithm is numerically stable because it proceeds by orthogonal similarity transforms.
Under certain conditions, the matrices Ak converge to a triangular matrix, the Schur form of A. The eigenvalues of a triangular matrix are listed on the diagonal, and the eigenvalue problem is solved. In testing for convergence it is impractical to require exact zeros, but the Gershgorin circle theorem provides a bound on the error.
In this crude form the iterations are relatively expensive. This can be mitigated by first bringing the matrix to upper Hessenberg form (which costs arithmetic operations using a technique based on Householder reduction), with a finite sequence of orthogonal similarity transforms, somewhat like a two-sided QR decomposition. (For QR decomposition, the Householder reflectors are multiplied only on the left, but for the Hessenberg case they are multiplied on both left and right.) Determining the QR decomposition of an upper Hessenberg matrix costs arithmetic operations. Moreover, because the Hessenberg form is already nearly upper-triangular (it has just one nonzero entry below each diagonal), using it as a starting point reduces the number of steps required for convergence of the QR algorithm.
If the original matrix is symmetric, then the upper Hessenberg matrix is also symmetric and thus tridiagonal, and so are all the . This procedure costs arithmetic operations using a technique based on Householder reduction. Determining the QR decomposition of a symmetric tridiagonal matrix costs operations.
The rate of convergence depends on the separation between eigenvalues, so a practical algorithm will use shifts, either explicit or implicit, to increase separation and accelerate convergence. A typical symmetric QR algorithm isolates each eigenvalue (then reduces the size of the matrix) with only one or two iterations, making it efficient as well as robust.
Visualization
The basic QR algorithm can be visualized in the case where A is a positive-definite symmetric matrix. In that case, A can be depicted as an ellipse in 2 dimensions or an ellipsoid in higher dimensions. The relationship between the i |
https://en.wikipedia.org/wiki/Wolfram%20Research | Wolfram Research, Inc. ( ) is an American multinational company that creates computational technology. Wolfram's flagship product is the technical computing program Wolfram Mathematica, first released on June 23, 1988. Other products include WolframAlpha, Wolfram SystemModeler, Wolfram Workbench, gridMathematica, Wolfram Finance Platform, webMathematica, the Wolfram Cloud, and the Wolfram Programming Lab. Wolfram Research founder Stephen Wolfram is the CEO. The company is headquartered in Champaign, Illinois, United States.
History
The company launched Wolfram Alpha, an answer engine on May 16, 2009. It brings a new approach to knowledge generation and acquisition that involves large amounts of curated computable data in addition to semantic indexing of text.
Wolfram Research acquired MathCore Engineering AB on March 30, 2011.
On July 21, 2011, Wolfram Research launched the Computable Document Format (CDF). CDF is an electronic document format designed to allow easy authoring of dynamically generated interactive content.
In June 2014, Wolfram Research officially introduced the Wolfram Language as a new general multi-paradigm programming language. It is the primary programming language used in Mathematica.
On April 15, 2020, Wolfram Research received $5,575,000 to help pay its employees during the COVID-19 pandemic as part of the U.S. government's Paycheck Protection Program administered by the Small Business Administration. The loan was forgiven.
Products and resources
Mathematica
Mathematica began as a software program for doing mathematics by computer, and has evolved to cover all domains of technical computing software, with features for neural networks, machine learning, image processing, geometry, data science, and visualizations. Central to Mathematica's mission is its ability to perform symbolic computation, for example, the ability to solve indefinite integrals symbolically. Mathematica includes a notebook interface and can produce slides for presentations. Mathematica is available in a desktop version, a grid computing version, and a cloud version.
Wolfram Alpha
Wolfram Alpha is a free online service that answers factual queries directly by computing the answer from externally sourced curated data, rather than providing a list of documents or web pages that might contain the answer as a search engine might. Users submit queries and computation requests via a text field and Wolfram Alpha then computes answers and relevant visualizations.
On February 8, 2012, Wolfram Alpha Pro was released, offering users additional features(e.g., the ability to upload many common file types and data — including raw tabular data, images, audio, XML, and dozens of specialized scientific, medical, and mathematical formats — for automatic analysis) for a monthly subscription fee.
In 2016, Wolfram Alpha Enterprise, a business-focused analytics tool, was launched. The program combines data supplied by a corporation with the algorithms from Wolfram |
https://en.wikipedia.org/wiki/Metaobject | In computer science, a metaobject is an object that manipulates, creates, describes, or implements objects (including itself). The object that the metaobject pertains to is called the base object. Some information that a metaobject might define includes the base object's type, interface, class, methods, attributes, parse tree, etc. Metaobjects are examples of the computer science concept of reflection, where a system has access (usually at run time) to its own internal structure. Reflection enables a system to essentially rewrite itself on the fly, to alter its own implementation as it executes.
Metaobject protocol
A metaobject protocol (MOP) provides the vocabulary (protocol) to access and manipulate the structure and behaviour of systems of objects. Typical functions of a metaobject protocol include:
Create or delete a new class
Create a new property or method
Cause a class to inherit from a different class ("change the class structure")
Generate or change the code defining the methods of a class
Metaobject protocol is contrary to Bertrand Meyer's open/closed principle, which holds that software object systems should be open for extension but closed for modification. This principle effectively draws a distinction between extending an object by adding to it, and modifying an object by redefining it, proposing that the former is a desirable quality ("objects should be extensible to meet the requirements of future use cases"), while the latter is undesirable ("objects should provide a stable interface not subject to summary revision"). Metaobject protocol, by contrast, transparently exposes the internal composition of objects and the entire object system in terms of the system itself. In practice, this means that programmers may use objects to redefine themselves, possibly in quite complex ways.
Furthermore, metaobject protocol is not merely an interface to an "underlying" implementation; rather, through metaobject protocol the object system is recursively implemented in terms of a meta-object system, which itself is theoretically implemented in terms of a meta-metaobject system, and so on until an arbitrary base case (a consistent state of the object system) is determined, with the protocol as such being the recursive functional relationship between these implementation levels.
Implementing object systems in such a way opens the possibility for radical discretionary redesign, providing deep flexibility but introducing possibly complex or difficult-to-understand metastability issues (for instance, the object system must not destructively update its own metaobject protocol - its internal self-representation - but the potential destructiveness of some updates is non-trivial to predict and may be hard to reason about), depending on the recursive depth to which the desired modifications are propagated. For this reason, metaobject protocol, when present in a language, is usually used sparingly and for specialised purposes such as software that tra |
https://en.wikipedia.org/wiki/Uconv | In computing, uconv is a command-line tool that is bundled with International Components for Unicode that converts text files between different character encodings. It is very similar to the iconv command that is part of the Single UNIX Specification which is usually implemented using libiconv. In fact the command line options for transcoding are the same. The command uconv can also convert to and from various Unicode normalization forms.
There is also an alternative implementation written in Ruby. It was written to supplement support of Japanese encoding in Ruby's XML Parser.
See also
International Components for Unicode
iconv
References
Unix text processing utilities
Ruby (programming language) |
https://en.wikipedia.org/wiki/README | In software development, a README file contains information about the other files in a directory or archive of computer software. A form of documentation, it is usually a simple plain text file called README, Read Me, READ.ME, README.TXT, README.md (to indicate the use of Markdown), or README.1ST.
The file's name is generally written in uppercase. On Unix-like systems in particular, this causes it to stand outboth because lowercase filenames are more common, and because the ls command commonly sorts and displays files in ASCII-code order, in which uppercase filenames will appear first.
Contents
A README file typically encompasses:
Configuration instructions
Installation instructions
Operating instructions
A file manifest (a list of files in the directory or archive)
Copyright and licensing information
Contact information for the distributor or author
A list of known bugs
Troubleshooting instructions
Credits and acknowledgments
A changelog (usually aimed at fellow programmers)
A news section (usually aimed at end users)
History
It is unclear when the convention of including a README file began, but examples dating to the mid-1970s have been found. Early Macintosh system software installed a Read Me on the Startup Disk, and README files commonly accompanied third-party software.
In particular, there is a long history of free software and open-source software including a README file; the GNU Coding Standards encourage including one to provide "a general overview of the package".
Since the advent of the web as a de facto standard platform for software distribution, many software packages have moved (or occasionally, copied) some of the above ancillary files and pieces of information to a website or wiki, sometimes including the README itself, or sometimes leaving behind only a brief README file without all of the information required by a new user of the software.
The popular source code hosting website GitHub strongly encourages the creation of a README fileif one exists in the main (top-level) directory of a repository, it is automatically presented on the repository's front page. In addition to plain text, various other formats and file extensions are also supported, and HTML conversion takes extensions into accountin particular a README.md is treated as GitHub Flavored Markdown.
As a generic term
The expression "readme file" is also sometimes used generically, for other files with a similar purpose. For example, the source-code distributions of many free software packages (especially those following the Gnits Standards or those produced with GNU Autotools) include a standard set of readme files:
{| class="wikitable"
|-
|README
|General information
|-
|AUTHORS
|Credits
|-
|THANKS
|Acknowledgments
|-
|CHANGELOG
|A detailed changelog, intended for programmers
|-
|NEWS
|A basic changelog, intended for users
|-
|INSTALL
|Installation instructions
|-
|COPYING / LICENSE
|Copyright and licensing information
|-
|BUGS
|Known bugs and i |
https://en.wikipedia.org/wiki/Gentoo%20%28file%20manager%29 | gentoo (written with a lowercase g) is a free file manager for Linux and other Unix-like computer systems created by Emil Brink. It is licensed under the GNU General Public License.
gentoo is written in C using the GTK+ toolkik and the "two-pane" concept. Notable features of gentoo include its graphical configurability, file typing and styling system. gentoo's file typing system identifies the file type and then uses the respective style defined to determine its display in listings. It is reminiscent of Total Commander or Norton Commander,. The author mentions that "gentoo tries to capture the features, the look and feel of the legendary file manager from Amiga platform: Directory Opus."
The package includes approximately 100 handcrafted icons for different file types. Additionally, many users like it for its speed, as it is relatively lightweight (see below).
Name
The word gentoo refers to a fast breed of penguin, the gentoo penguin. In spite of the name, it has nothing to do with the Gentoo Linux distribution, and did in fact adopt the name first.
See also
Comparison of file managers
References
External links
File managers that use GTK
Orthodox file managers
Free file managers
Free software programmed in C |
https://en.wikipedia.org/wiki/U15%20%28Berlin%20U-Bahn%29 | The U15 was a line on the Berlin U-Bahn. Originally the U3 running between Wittenbergplatz and Uhlandstraße, it became a branch of the U1 in 1993 as part of the reorganisation of the network following the reopening of the U2.
On December 12, 2004, it was renumbered as the U1.
References
Berlin U-Bahn lines
Railway lines opened in 1993
Railway lines closed in 2004
1993 establishments in Germany
2004 disestablishments in Germany |
https://en.wikipedia.org/wiki/IBM%205100 | The IBM 5100 Portable Computer is one of the first portable computers, introduced in September 1975, six years before the IBM Personal Computer, and eight before the first successful IBM compatible portable computer, the Compaq Portable. It was the evolution of a prototype called the SCAMP (Special Computer APL Machine Portable) that was developed at the IBM Palo Alto Scientific Center in 1973. Whether considered evolutionary from SCAMP or revolutionary, it still needed to be plugged into an electric socket.
When the IBM PC was introduced in 1981, it was originally designated as the IBM 5150, putting it in the "5100" series, though its architecture was unrelated to the IBM 5100's. The 5100 was IBM's second transportable computer. Previously, a truck-based IBM 1401 configured in 1960 for military use was designated a portable computer, and nicknamed a DataMobile.
The IBM 5100 was withdrawn in March 1982, by which time IBM had announced its larger cousins, the IBM 5110 (January 1978) and the IBM 5120 (February 1980).
SCAMP, the prototype
In 1973, Bill Lowe was instrumental in fostering an engineering prototype called SCAMP (Special Computer APL Machine Portable) created by Dr. Paul Friedl and a team at the IBM Los Gatos Scientific Center. SCAMP has been dubbed in PC Magazine as "the world's first personal computer".
The IBM Los Gatos engineering prototype and a design model by IBM Industrial designer Tom Hardy, were utilized internally by Lowe in his early efforts to demonstrate the viability of creating a single-user computer.
SCAMP emulated an IBM 1130 minicomputer in order to run APL\1130. In 1973, APL was generally available only on mainframe computers, and most desktop sized microcomputers such as the Wang 2200 or HP 9800 offered only BASIC.
Because SCAMP was the first to emulate APL\1130 performance on a portable, single user computer, PC Magazine in 1983 designated SCAMP a "revolutionary concept" and "the world's first personal computer".
Description
The IBM 5100 is based on a 16-bit processor module called PALM (Program All Logic in Microcode). The IBM 5100 Maintenance Information Manual also referred to the PALM module as the controller. The PALM could directly address of memory. Some configurations of the IBM 5100 had Executable ROS (ROM) and RAM memory totalling more than , so a simple bank switching scheme was used. The actual APL and BASIC interpreters were stored in a separate Language ROS address space which the PALM treats as a peripheral device. There were twelve models available: with BASIC, APL, or both. Memory could be , , or of main
storage. The 5100 sold for between and (between $ and $ in today's dollars).
Often described as being "approximately fifty pounds", its weight was closer to 55 pounds (25 kg). In December 1975 BYTE stated "Welcome, IBM, to personal computing". Describing the 5100 as "a 50-lb package of interactive personal computing," the magazine said that with the company's announcement "pers |
https://en.wikipedia.org/wiki/Cyberware | Cyberware is a relatively new and unknown field (a proto-science, or more adequately a "proto-technology"). In science fiction circles, however, it is commonly known to mean the hardware or machine parts implanted in the human body and acting as an interface between the central nervous system and the computers or machinery connected to it.
More formally:
Cyberware is technology that attempts to create a working interface between machines/computers and the human nervous system, including the brain.
Examples of potential cyberware cover a wide range, but current research tends to approach the field from one of two different angles: interfaces or prosthetics.
Interfaces ("headware")
The first variety attempts to connect directly with the brain. The data-jack is probably the best-known, having heavily featured in works of fiction (even in mainstream productions such as Johnny Mnemonic, the cartoon Exosquad, and The Matrix). It is the most difficult object to implement, but it is also the most important in terms of interfacing directly with the mind. In science fiction the data-jack is the envisioned I/O port for the brain. Its job is to translate thoughts into something meaningful to a computer, and to translate something from a computer into meaningful thoughts for humans. Once perfected, it would allow direct communication between computers and the human mind.
Large university laboratories conduct most of the experiments done in the area of direct neural interfaces. For ethical reasons, the tests are usually performed on animals or slices of brain tissue from donor brains. The mainstream research focuses on electrical impulse monitoring, recording and translating the many different electrical signals that the brain transmits. A number of companies are working on what is essentially a "hands-free" mouse or keyboard. This technology uses these brain signals to control computer functions. These interfaces are sometimes called brain-machine interfaces (BMI).
The more intense research, concerning full in-brain interfaces, is being studied, but is in its infancy. Few can afford the huge cost of such enterprises, and those who can find the work slow-going and very far from the ultimate goals. Research has reached the level where limited control over a computer is possible using thought commands alone. After being implanted with a Massachusetts-based firm Cyberkinetics chip called BrainGate, a quadriplegic man was able to compose and check email.
Prosthetics ("bodyware")
The second variety of cyberware consists of a more modern form of the rather old field of prosthetics. Modern prostheses attempt to deliver a natural functionality and appearance. In the sub-field where prosthetics and cyberware cross over, experiments have been done where microprocessors, capable of controlling the movements of an artificial limb, are attached to the severed nerve-endings of the patient. The patient is then taught how to operate the prosthetic, trying to learn h |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.