source
stringlengths
32
199
text
stringlengths
26
3k
https://en.wikipedia.org/wiki/AWS%20%28disambiguation%29
AWS is Amazon Web Services, a cloud computing and web services provider. AWS or Aws may also refer to: Technology Advanced Wireless Services, spectrum band used for mobile services in America Apple Workgroup Server Automatic Warning System, a train safety system used on railways Autonomous Web Services, via the World Wide Web Automatic weather station Organisations Alien Workshop, a skateboard company American Welding Society Association of Women Surgeons Austria Wirtschaftsservice Gesellschaft, an Austrian bank AWS Shopper, a German manufacturer of microcar Solidarity Electoral Action (Akcja Wyborcza Solidarność), a political party coalition in Poland Military Aircraft Warning Service, a former civilian service of the United States Army Ground Observer Corps Affordable Weapon System, a US Navy cruise missile system in a shipping container American War Standard, a plan for the electronic color code of small radio parts Arctic Warfare Suppressed, an internally suppressed bolt-action sniper rifle Autonomous Weapon System aka Lethal autonomous weapon Music A Wilhelm Scream, a punk rock band from Massachusetts A Wilhelm Scream (EP), a 2009 record by the band American Wind Symphony AWS (band), Hungarian metal band Other uses South Awyu language (ISO 639-3 code: aws) Afrikaanse Woordelys en Spelreëls, the Afrikaans Word List and Spelling Rules publication All-women shortlist Associate Writer to the Signet, a membership level in the Society of Writers to Her Majesty's Signet Banu Aws (Sons of Aws), a tribe in Mohammad's era Alternative work schedule, a kind of flexitime used in certain U.S. government agencies Alcohol withdrawal syndrome
https://en.wikipedia.org/wiki/Laurie%20Spiegel
Laurie Spiegel (born September 20, 1945) is an American composer. She has worked at Bell Laboratories, in computer graphics, and is known primarily for her electronic-music compositions and her algorithmic composition software Music Mouse. She also plays the guitar and lute. Spiegel is seen by some as a pioneer of the New York new-music scene. She withdrew from this scene in the early 1980s, believing that its focus had shifted from artistic process to product. While she continues to support herself through software development, Spiegel aims to use technology in music as a means of furthering her art rather than as an end in itself. In her words, "I automate whatever can be automated to be freer to focus on those aspects of music that can't be automated. The challenge is to figure out which is which." Spiegel's realization of Johannes Kepler's Harmonices Mundi was chosen for the opening track on the "Sounds of Earth" section of the golden record placed on board the Voyager spacecraft in 1977. Another work, titled "Sediment", was included in the 2012 film The Hunger Games. She has been inducted into the National Women's Hall of Fame. Education Spiegel's early musical experiences were largely self-directed, beginning with the mandolin, guitar, and banjo she had as a child, which she learned to play by ear. She taught herself Western music notation at the age of 20, after which she began writing down her compositions. Spiegel attended Shimer College through the school's early entrance program, which allows students to enter college without having completed high school. She subsequently attended Oxford University, initially through Shimer's Oxford study abroad program, under which students spend a year continuing the Great Books core curriculum in Oxford while taking tutorials from Oxford. After receiving her BA degree in the Social Sciences from Shimer in 1967, Spiegel stayed in Oxford an additional year, commuting to London to study guitar, theory and composition with John W. Duarte. After moving to New York, where she briefly worked in social sciences research and documentary film, she went on to study composition with Jacob Druckman, Vincent Persichetti and Hall Overton at the Juilliard School from 1969 to 1972, privately with Emmanuel Ghent, then she relocated along with Druckman, to whom she was composer's assistant, to Brooklyn College, completing her MA in Music Composition there in 1975 as well as pursuing research in early American music under the direction of H. Wiley Hitchcock. Career Best known for her use of interactive and algorithmic logic as part of the compositional process, Spiegel worked with Buchla and Electronic Music Laboratories synthesizers and subsequently many early, often experimental and prototype-level music and image generation systems, including GROOVE system (1973–1978), Alles Machine (1977) and Max Mathews's RTSked and John R. Pierce tunings (1984, later known as the Bohlen–Pierce scale) at Bell Labs, the
https://en.wikipedia.org/wiki/Research%20Natural%20Area
Research Natural Area is a designation for certain protected areas in the United States. Research Natural Areas (RNAs) are part of a nationwide network of ecological areas set aside for both research and education. The network includes areas managed by many Federal agencies. The United States Forest Service and other agencies establish these areas to typify certain types of important forest, shrubland, grassland, aquatic, geological, alpine, or similar environments that have unique characteristics of scientific interest. The areas "contain important ecological and scientific values and are managed for minimum human disturbance". The first RNA was established on the Coronado National Forest in Arizona in 1927. The Bureau of Land Management is another agency that designates and manages Research Natural Areas. According to the Bureau of Land Management, the objectives of the RNA program are "(1) To preserve examples of all significant natural ecosystems for comparison with those influenced by man; (2) to provide educational and research areas for ecological and environmental studies; and (3) to preserve gene pools of typical and endangered plants and animals. Research natural areas are intended to represent the full array of North American ecosystems with their biological communities, habitats, natural phenomena, and geological and hydrological formations." The Bureau of Land Management manages 192 RNAs with a total area of . , the US Forest Service managed 289 established RNAs and more than 300 candidate RNAs. , there were 571 US Forest Service RNAs (Note that these are not complete lists): These areas are established under the Organic Act of 1897. Areas designated as Research Natural Areas are primarily located inside National Forests. References Protected areas of the United States
https://en.wikipedia.org/wiki/BlackBerry%20Limited
BlackBerry Limited (formerly Research In Motion) is a Canadian software company specializing in cybersecurity. Founded in 1984, it originally developed the BlackBerry brand of interactive pagers, smartphones and tablets. In 2016, it transitioned to a cybersecurity enterprise software and services company under CEO John S. Chen. Its products are used by various businesses, car manufacturers, and government agencies to prevent hacking and ransomware attacks. They include the BlackBerry Cylance, the QNX real-time operating system; BlackBerry Enterprise Server (BlackBerry Unified Endpoint Manager), and a Unified Endpoint Management (UEM) platform. History 1984–2001: early years and growth Research In Motion Limited was founded in March 1984 by Mike Lazaridis and Douglas Fregin. At the time, Lazaridis was an engineering student at the University of Waterloo while Fregin was an engineering student at the University of Windsor. In 1988, RIM became the first wireless data technology developer in North America and the first company outside Scandinavia to develop connectivity products for Mobitex wireless packet-switched data communications networks. Mobitex's wireless data transport also became RIM's first wireless data service that powered the Blackberry and Palm devices until it was phased out. In 1990, RIM introduced the DigiSync Film KeyKode Reader. In 1991, it introduced the first Mobitex protocol converter. In 1992, it introduced the first Mobitex point-of-sale solution, a protocol converter box that interfaced with existing point-of-sale terminal equipment to enable wireless communication. In 1993, it introduced the RIMGate, the first general-purpose Mobitex X.25 gateway. In the same year, it launched Ericsson Mobidem AT and Intel wireless modems containing RIM modem firmware. In 1994, it introduced the first Mobitex mobile point-of-sale terminal. In the same year, it received the Emmy Award for Technical Innovation and the KPMG High Technology Award. In 1995, it introduced Freedom, the first Type II PCMCIA radio modem for Mobitex. In 1995, RIM was financed by Canadian institutional and venture capital investors through a private placement in the privately held company. Working Ventures Canadian Fund Inc. led the first venture round with a C$5,000,000 investment with the proceeds being used to complete the development of RIM's two-way paging system hardware and software. A total of C$30,000,000 in pre-IPO financing was raised by the company prior to its initial public offering on the Toronto Stock Exchange in January 1998 under the symbol RIM. In 1996, RIM introduced the Interactive Pager, the first two-way messaging pager, and the RIM 900 OEM radio modem. The company developed the pager prototype with the support of Intel Corporation. The company worked with RAM Mobile Data and Ericsson to turn the Ericsson-developed Mobitex wireless data network into a two-way paging and wireless e-mail network. Pivotal in this development was the release of
https://en.wikipedia.org/wiki/French%20post%20offices%20abroad
The French post offices abroad were a global network of post offices in foreign countries established by France to provide mail service where the local services were deemed unsafe or unreliable. They were generally set up in cities with some sort of French commercial interest. They started appearing in the early 19th century, reached their heyday at the beginning of the 20th century, then started closing down in the 1910s and 1920s, with the office at Kwangchou in China holding out until the 1940s. Offices abroad: French post offices in China French post offices in Crete French post offices in Egypt French post offices in the Ottoman Empire French post offices in Zanzibar Sources Stanley Gibbons Ltd: various catalogues AskPhil – Glossary of Stamp Collecting Terms Encyclopaedia of Postal History Rossiter, Stuart & John Flower. The Stamp Atlas. London: Macdonald, 1986. External links https://web.archive.org/web/20100918014836/http://www.rpsl.org.uk/indochina/index.html Postage stamps of France Postal system of France
https://en.wikipedia.org/wiki/8SVX
8-Bit Sampled Voice (8SVX) is an audio file format standard developed by Electronic Arts for the Amiga computer series. It is a data subtype of the IFF file container format. It typically contains linear pulse-code modulation (LPCM) digital audio. Description The 8SVX subtype stores 8-bit audio data within chunks contained within an IFF file container. 8SVX subtypes can exist alone within IFF file containers (audio only), or can be multiplexed together with other IFF subtypes, such as video animation streams. Metadata about the 8SVX data stream is contained in separate descriptor chunks that come prior to the main data body chunk. Sample rate, volume and compression type are described in a VHDR chunk. Various other chunks are available to describe the name, author and copyright. 8SVX supports features such as attack, release and section repeat, which are useful for storage of musical instrument samples. An example layout of an audio-only 8SVX IFF audio file: Encoding The majority of 8SVX data streams are encoded using uncompressed linear PCM streams. Optionally, Fibonacci-delta lossy data compression is also available, resulting in a 50% compression ratio at the cost of decreased fidelity. Multi-byte values are stored in big-endian format, the native byte order for the Motorola 68000 family. Support IFF-8SVX encoded audio was the default audio format for the Commodore Amiga. Most audio programs for the Amiga supported the format. AmigaOS 3.0 introduced a multimedia framework using the datatype subsystem that included an 8SVX decoder (8SVX.datatype). Many sound editing programs and music tracker programs of the late 1980s and early 1990s supported the format. It is still a common format for cross-platform audio editing programs (such as Sound eXchange). 8SVX support is also available to modern programs via libavcodec (and the related ffdshow codec package) as well as via libsndfile. Legacy The Commodore Amiga computer series never received native hardware support for 16-bit digital audio before the decline of the platform. As such, the related 16SVX and MAUD subtypes never saw wide adoption. Apple Computer developed a separate subtype known as AIFF which included support for 16-bit samples and additional compression types. It superseded 8SVX as the dominant audio subtype for IFF files. Microsoft and IBM co-developed the RIFF file container and the related WAVE audio subtype for Windows. Both formats are heavily influenced by the IFF/8SVX container format, but like AIFF, were extended to support higher bit-depths and additional compression types. See also AIFF IFF File Format WAV Paula, the digital audio processor for the Commodore Amiga computer References External links IFF file container and subtypes at Multimedia Wiki IFF chunk registry at Amigan Software Amiga file formats Computer file formats AmigaOS
https://en.wikipedia.org/wiki/Dynamic%20time%20warping
In time series analysis, dynamic time warping (DTW) is an algorithm for measuring similarity between two temporal sequences, which may vary in speed. For instance, similarities in walking could be detected using DTW, even if one person was walking faster than the other, or if there were accelerations and decelerations during the course of an observation. DTW has been applied to temporal sequences of video, audio, and graphics data — indeed, any data that can be turned into a one-dimensional sequence can be analyzed with DTW. A well-known application has been automatic speech recognition, to cope with different speaking speeds. Other applications include speaker recognition and online signature recognition. It can also be used in partial shape matching applications. In general, DTW is a method that calculates an optimal match between two given sequences (e.g. time series) with certain restriction and rules: Every index from the first sequence must be matched with one or more indices from the other sequence, and vice versa The first index from the first sequence must be matched with the first index from the other sequence (but it does not have to be its only match) The last index from the first sequence must be matched with the last index from the other sequence (but it does not have to be its only match) The mapping of the indices from the first sequence to indices from the other sequence must be monotonically increasing, and vice versa, i.e. if are indices from the first sequence, then there must not be two indices in the other sequence, such that index is matched with index and index is matched with index , and vice versa We can plot each match between the sequences and as a path in a matrix from to , such that each step is one of . In this formulation, we see that the number of possible matches is the Delannoy number. The optimal match is the match that satisfies all the restrictions and the rules and that has the minimal cost, where the cost is computed as the sum of absolute differences, for each matched pair of indices, between their values. The sequences are "warped" non-linearly in the time dimension to determine a measure of their similarity independent of certain non-linear variations in the time dimension. This sequence alignment method is often used in time series classification. Although DTW measures a distance-like quantity between two given sequences, it doesn't guarantee the triangle inequality to hold. In addition to a similarity measure between the two sequences, a so called "warping path" is produced. By warping according to this path the two signals may be aligned in time. The signal with an original set of points X(original), Y(original) is transformed to X(warped), Y(warped). This finds applications in genetic sequence and audio synchronisation. In a related technique sequences of varying speed may be averaged using this technique see the average sequence section. This is conceptually very similar to the Ne
https://en.wikipedia.org/wiki/Standard%20Delay%20Format
Standard Delay Format (SDF) is an IEEE standard for the representation and interpretation of timing data for use at any stage of an electronic design process. It finds wide applicability in design flows, and forms an efficient bridge between dynamic timing verification and static timing analysis. It was originally developed as an OVI standard, and later modified into the IEEE format. Technically only the SDF version 4.0 onwards are IEEE formats. It is an ASCII format that is represented in a tool and language independent way and includes path delays, timing constraint values, interconnect delays and high level technology parameters. It has usually two sections: one for interconnect delays and the other for cell delays. SDF format can be used for back-annotation as well as forward-annotation. See also: VITAL, EDIF References IEC 61523-3:2004 External links https://web.archive.org/web/20130524215112/http://www.eda.org/sdf/ Standard Delay Format Specification in PDF format, version 3.0 (1995) EDA file formats IEEE DASC standards IEC standards
https://en.wikipedia.org/wiki/DG
DG may refer to: Arts and entertainment Death Grips, an American experimental hip hop group DG (character), in the science fiction series Tin Man Dial Global, a radio network Dragon Gate, a Japanese professional wrestling promotion Drain Gang, a Swedish rap group Business and organizations Data General, a minicomputer manufacturer Desnoes & Geddes (DG, D&G), a Jamaican brand of soft drinks Delta Gamma, a women's sorority Deutsche Grammophon, a classical music record label DG Flugzeugbau, a German airplane manufacturer Dial Global, a radio network Dolce & Gabbana, an Italian luxury fashion design Dollar General, an American variety store (NYSE ticker DG) Places DG postcode area, the Dumfries and Galloway postcode area in Scotland Danilovgrad, a municipality in Montenegro, abbreviated DG on car plates German-speaking Community of Belgium () Diego Garcia, exceptionally reserved ISO 3166-1 alpha-2 country code for Science, mathematics and technology Decigram, a unit of measure Dentate gyrus, a brain structure Diglyme, an organic solvent Directional gyroscope, a heading indicator use in aircraft Discontinuous Galerkin method, a numerical method Distributed generation of energy Transport Cebgo, formerly South East Asian Airlines, SEAir, Tigerair Philippines; IATA code Dindigul Junction railway station, Tamil Nadu, India; Indian Railways station code Other uses DG (footballer) (born 2001), Douglas da Silva Teixeira, Brazilian footballer Dei gratia (Latin: "By the grace of God"), abbreviated D.G. Deo gratias (Latin: "Thanks be to God"). Dangerous goods Differential geometry Director-general Directorate-General, a type of specialised administrative body in the European Union
https://en.wikipedia.org/wiki/Rational%20ClearCase
Rational ClearCase is a family of computer software tools that supports software configuration management (SCM) of source code and other software development assets. It also supports design-data management of electronic design artifacts, thus enabling hardware and software co-development. ClearCase includes revision control and forms the basis for configuration management at large and medium-sized businesses, accommodating projects with hundreds or thousands of developers. It is developed by IBM. ClearCase supports two configuration management models: UCM (Unified Change Management) and base ClearCase. UCM provides an out-of-the-box model while base ClearCase provides a basic infrastructure (UCM is built on base ClearCase). Both can be configured to support a wide variety of needs. ClearCase can accommodate large binary files, large numbers of files, and large repository sizes. It supports branching, labeling, and versioning of directories. It uses the MultiVersion File System (MVFS) which is a virtual file system that displays specific versions of data stored. In particular, it supports dynamic views which can show an arbitrary combination of local and remote files. History ClearCase was developed by Atria Software and first released in 1992 on Unix and later on Windows. Some of the Atria developers had worked on an earlier, similar system: DSEE (Domain Software Engineering Environment) from Apollo Computer. After Hewlett-Packard bought Apollo Computer in 1989, those developers left to form Atria. Atria later merged with Pure Software to form PureAtria. That firm was acquired by Rational Software, which was purchased by IBM in 2003. IBM continues to develop and market ClearCase. In September 2016, IBM announced a strategic partnership with HCL Technologies that will allow for accelerated development. DSEE introduced many concepts that were adopted by ClearCase. The Apollo Domain file system allowed special handler programs to intervene during file access. DSEE made use of this feature to invisibly substitute a versioned copy when a particular file was opened. With the versioning specification resident in the user environment, all accesses to versioned files were redirected, including such mundane accesses as printing, viewing in a generic text editor etc. DSEE relied heavily on a file that described all the software modules and their dependencies. The file had to be generated manually, which was a major impediment to its use in large systems. However, once generated, it enabled DSEE to calculate the optimum way to perform a build, re-using all modules that had previously been processed and whose version specifications matched the specifications for the build. DSEE also introduced the "version spec," which was called a "thread." This was a list of possible versions that could be in the user environment or in a build. A major innovation was the use of build signatures and software release signatures in the thread. The items in a thread m
https://en.wikipedia.org/wiki/SOB
Sob is a verb meaning cry. Sob and SoB may refer to: Souls on Board (sometimes POB for People on Board), used in Aviation communication Seventeen or Bust, a distributed computing project Special Operations Battalion (SOB), an elite unit of the Croatian army Society of Old Brooklynites Son of a bitch, an insult or curse phrase, sometimes abbreviated as "S.O.B." Sons of Ben (MLS supporters association) Suggested opening bid, a term in auctioneering Super Oralloy Bomb, a nuclear bomb Senate Office Building Dirksen Senate Office Building Russell Senate Office Building Hart Senate Office Building S.O.B. Hill, a mountain in Utah, United States Start of business, the start of the business day (as opposed to EOB) Biomedicine Shortness of breath or dyspnea, a medical symptom relating to breathing difficulties Sobralia, a genus of orchids Super optimal broth, a bacterial growth medium Music and media S.O.B. (film), a 1981 film by Blake Edwards The Sob, a 1953 Turkish film S.O.B. (band), a Japanese grindcore band "S.O.B." (song), by Nathaniel Rateliff and the Night Sweats, 2015 Scars on Broadway, an American rock band Sons of Batman, a group of characters in a Batman comics miniseries Sons of Butcher (band), a Canadian rock band Sons of Butcher (TV show), a cartoon inspired by the band Sounds of Blackness, an American vocal and instrumental ensemble SOB's, Sounds of Brazil, a live music venue in New York City Styles of Beyond, an underground rap group Switched-On Bach, an album by Walter Carlos S.O.B.s, an episode of the television series Arrested Development Transport Seaford-Oyster Bay Expressway Balaton Airport, SOB in IATA code South Bend station (Amtrak), SOB is its station code Southbourne railway station, SOB in Network Rail code Stadt-Omnibus Bern, a former public transport operator in Bern, Switzerland Südostbahn, a railway company in Switzerland SüdostBayernBahn, a railway company in southern Germany MS Spirit of Britain, a Dover - Calais P&O Ferry Toponyms Sob (river), a river in Siberia, a tributary of the Ob Sob River, a river in Ukraine, a tributary of Southern Bug
https://en.wikipedia.org/wiki/Internet2
Internet2 is a not-for-profit United States computer networking consortium led by members from the research and education communities, industry, and government. The Internet2 consortium administrative headquarters are located in Ann Arbor, Michigan, with offices in Washington, D.C., and Emeryville, California. As of November 2013, Internet2 has over 500 members including 251 institutions of higher education, 9 partners and 76 members from industry, over 100 research and education networks or connector organizations, and 67 affiliate members. Internet2 operates the Internet2 Network, an Internet Protocol network using optical fiber that delivers network services for research and education, and provides a secure network testing and research environment. In late 2007, Internet2 began operating its newest dynamic circuit network, the Internet2 DCN, an advanced technology that allows user-based allocation of data circuits over the fiber-optic network. The Internet2 Network, through its regional network and connector members, connects over 60,000 U.S. educational, research, government and "community anchor" institutions, from primary and secondary schools to community colleges and universities, public libraries and museums to health care organizations. The Internet2 community develops and deploys network technologies for the future of the Internet. These technologies include large-scale network performance measurement and management tools, secure identity and access management tools and capabilities such as scheduling high-bandwidth, high-performance circuits. Internet2 members serve on several advisory councils, collaborate in a variety of working groups and special interest groups, gather at spring and fall member meetings, and are encouraged to participate in the strategic planning process. History As the Internet gained in public recognition and popularity, universities were among the first institutions to outgrow the Internet's bandwidth limitations because of the data transfer requirements faced by academic researchers who needed to collaborate with their colleagues. Some universities wanted to support high-performance applications like data mining, medical imaging and particle physics. This resulted in the creation of the very-high-performance Backbone Network Service, or vBNS, developed in 1995 by the National Science Foundation (NSF) and MCI for supercomputers at educational institutions. After the expiration of the NSF agreement, vBNS largely transitioned to providing service to the government. As a result, the research and education community founded Internet2 to serve its networking needs. The Internet2 Project was originally established by 34 university researchers in 1996 under the auspices of EDUCOM (later EDUCAUSE), and was formally organized as the not-for-profit University Corporation for Advanced Internet Development (UCAID) in 1997. It later changed its name to Internet2. Internet2 is a registered trademark. The Internet2
https://en.wikipedia.org/wiki/Qian%20Xuesen
Qian Xuesen, or Hsue-shen Tsien (; 11 December 1911 – 31 October 2009), was a Chinese aerospace engineer and cyberneticist who made significant contributions to the field of aerodynamics and established engineering cybernetics. Qian received his undergraduate education in mechanical engineering at National Chiao Tung University in Shanghai in 1934 and took a pre-departure transitional year in aircraft design at National Tsinghua University in Beijing. He travelled to the United States in 1935 and attained a master's degree in aeronautical engineering at the Massachusetts Institute of Technology in 1936. Afterward, he joined Theodore von Kármán's group at the California Institute of Technology in 1936, received a doctorate in aeronautics and mathematics there in 1939, and became an associate professor at Caltech in 1943. While at Caltech, Qian helped found NASA's Jet Propulsion Laboratory. He was recruited by the United States Department of Defense and the Department of War to serve in various positions, including as an expert consultant with a colonel rank in 1945. At the same time, he became an associate professor at MIT in 1946, a full professor at MIT in 1947, and a full professor at Caltech in 1949. During the Second Red Scare in the 1950s, the United States federal government accused him of communist sympathies. In 1950, despite protests by his colleagues, he was stripped of his security clearance. He decided to return to China but was then detained at Terminal Island in Los Angeles County, California. After spending five years under house arrest, he was released in 1955 in exchange for the repatriation of American pilots who had been captured during the Korean War. He left the United States in September 1955 on the American President Lines passenger liner SS President Cleveland, arriving in Mainland China via Hong Kong. Upon his return, he helped lead development of the Dongfeng ballistic missile and the Chinese space program. He also played a significant part in the construction and development of China's defense industry system, higher education and research system, rocket force, and a key technology university. For his contributions, he became known as the "Father of Chinese Rocketry", nicknamed the "King of Rocketry". He is recognized as one of the founding fathers of Two Bombs, One Satellite. In 1957, Qian was elected an academician of the Chinese Academy of Sciences. He served as a Vice Chairman of the National Committee of the Chinese People's Political Consultative Conference from 1987 to 1998. He was the cousin of engineer Hsue-Chu Tsien, who was involved in the aerospace industries of both China and the United States. His nephew, Roger Y. Tsien, was the 2008 winner of the Nobel Prize in Chemistry. Early life and education Qian was born in Shanghai International Settlement, with ancestral roots in Lin'an, Hangzhou, in 1911. His parents were Qian Junfu and Zhang Lanjuan. He graduated from the High School Affiliated to Beijin
https://en.wikipedia.org/wiki/Wear%20leveling
Wear leveling (also written as wear levelling) is a technique for prolonging the service life of some kinds of erasable computer storage media, such as flash memory, which is used in solid-state drives (SSDs) and USB flash drives, and phase-change memory. There are several wear leveling mechanisms that provide varying levels of longevity enhancement in such memory systems. The term preemptive wear leveling (PWL) has been used by Western Digital to describe their preservation technique used on hard disk drives (HDDs) designed for storing audio and video data. However, HDDs generally are not wear-leveled devices in the context of this article. Rationale EEPROM and flash memory media have individually erasable segments, each of which can be put through a limited number of erase cycles before becoming unreliable. This is usually around 3,000/5,000 cycles but many flash devices have one block with a specially extended life of 100,000+ cycles that can be used by the Flash memory controller to track wear and movement of data across segments. Erasable optical media such as CD-RW and DVD-RW are rated at up to 1,000 cycles (100,000 cycles for DVD-RAM media). Wear leveling attempts to work around these limitations by arranging data so that erasures and re-writes are distributed evenly across the medium. In this way, no single erase block prematurely fails due to a high concentration of write cycles. In flash memory, a single block on the chip is designed for longer life than the others so that the memory controller can store operational data with less chance of its corruption. Conventional file systems such as FAT, UFS, HFS/HFS+, EXT, and NTFS were originally designed for magnetic disks and as such rewrite many of their data structures (such as their directories) repeatedly to the same area. When these systems are used on flash memory media, this becomes a problem. The problem is aggravated by the fact that some file systems track last-access times, which can lead to file metadata being constantly rewritten in-place. Types There are three basic types of wear leveling mechanisms used in flash memory storage devices: No wear leveling A flash memory storage system with no wear leveling will not last very long if data is written to the flash. Without wear leveling, the underlying flash controller must permanently assign the logical addresses from the operating system (OS) to the physical addresses of the flash memory. This means that every write to a previously written block must first be read, erased, modified, and re-written to the same location. This approach is very time-consuming and frequently written locations will wear out quickly, while other locations will not be used at all. Once a few blocks reach their end of life, such a device becomes inoperable. Dynamic wear leveling The first type of wear leveling is called dynamic wear leveling and it uses a map to link logical block addresses (LBAs) from the OS to the physical flash memory. Each ti
https://en.wikipedia.org/wiki/National%20Space%20Grant%20College%20and%20Fellowship%20Program
The space-grant colleges are educational institutions in the United States that comprise a network of fifty-two consortia formed for the purpose of outer space-related research. Each consortium is based in one of the fifty states, the District of Columbia, or Puerto Rico, and each consists of multiple independent space-grant institutions, with one of the institutions acting as lead. Similar programs include sea-grant colleges (instituted in 1966) and sun-grant colleges (instituted in 2003). Objectives The program claims the following objectives: Establish and maintain a national network of universities with interests and capabilities in aeronautics, outer space, and related fields; Encourage cooperative programs among universities, the aerospace industry, and federal, state, and local governments; Encourage interdisciplinary training, research, and public service programs related to aerospace; Recruit and train U.S. citizens, especially women, underrepresented minorities, and persons with disabilities, for careers in aerospace science and technology; and, Promote a strong science, mathematics, and technology education base from elementary through secondary levels. History The National Space Grant College and Fellowship Program was established in 1988 by the US Congress following the success of similar scholarship opportunities in the oceanic Sea Grant and agricultural Land Grant fields. The catch-all term Space Grant refers back to these previous federal programs. In 1989, the program was given over to NASA, which now administers it in the same way that NOAA administers Sea Grant. The first meeting of the National Council of Space Grant Directors took place from January 16 to January 19, 1990, at the Johns Hopkins Applied Physics Laboratory in Columbia, Maryland, and the Jet Propulsion Laboratory in La Cañada Flintridge, California. In 2011, an effort was made to begin collecting the history of the council. Programs Building Leaders for Advancing Science and Technology (BLAST) BLAST is a three-day summer event that takes place at Virginia's Old Dominion University. This free, STEM-centered event, is available to 8th and 9th graders with a 2.3 GPA or better. BLAST's purpose is to show the students the interesting aspects of STEM to those who are unsure whether they want to be involved in STEM or not. Nationwide Eclipse Ballooning Project The NEBP is a high-altitude, balloon launching program. eXploration Systems and Habitation (X-Hab) 2019 Academic Innovation Challenge The eXploration Systems and Habitation (X-Hab) challenge seeks better the academic and technological abilities of university students. It utilizes hands-on training in the implementation of space habitats for deep space missions. There is an Advanced Exploration Systems (AES) Division where university students can compete to win awards worth $15,000–50,000 for a functional design that the AES finds useful and, or, interesting. John Mather Nobel Scholars Established in 2008,
https://en.wikipedia.org/wiki/National%20Sea%20Grant%20College%20Program
The National Sea Grant College Program is a program of the National Oceanic and Atmospheric Administration (NOAA) within the U.S. Department of Commerce. It is a national network of 34 university-based Sea Grant programs involved in scientific research, education, training, and extension projects geared toward the conservation and practical use of the coasts, Great Lakes, and other marine areas. The program is administered by the National Oceanic and Atmospheric Administration (NOAA) with the national office located in Silver Spring, Maryland. There are Sea Grant programs located in every coastal and Great Lakes state as well as in Puerto Rico and Guam. The program was instituted in 1966 when Congress passed the National Sea Grant College Program Act. Sea Grant programs and colleges are not to be confused with land-grant colleges (a program instituted in 1862), space-grant colleges (instituted in 1988), or sun-grant colleges (instituted in 2003), although an institution may also be in one or more of the other programs concurrently with being a sea-grant institution. History At a 1963 meeting of the American Fisheries Society, a University of Minnesota professor, Athelstan Spilhaus, first suggested the establishment of Sea Grant colleges in universities that wished to develop oceanic work. The name "Sea Grant" was chosen to draw a parallel with the land-grant college program that was funded by grants of western lands to the states by the 1862 Morrill Land Grant Act of 1862. Early in the legislative process, there was consideration of leases of offshore parcels of ocean and sea bottom to fund the program by John A. Knauss and bill sponsor Claiborne Pell much like the 1862 land grants, but that plan was eventually scrapped in favor of direct congressional appropriation for the program. The 1966 Act allowed the National Science Foundation (NSF) authority to initiate and support education, research, and extension by: Encouraging and developing programs consisting of instruction, practical demonstrations, publications, and otherwise, by sea grant colleges and other suitable institutes, laboratories, and public and private agencies through marine advisory programs with the object of imparting useful information to person currently employed or interested in the various fields related to the development of marine resources, the scientific community, and the general public. Signing of the 1966 Sea Grant College and Program Act into law by President Lyndon B. Johnson was on October 15, 1966, as Public Law 89-688. The only major subsequent change to the Sea Grant Act was with a 1970 Reorganization Plan, whereby the Office of Sea Grant was transferred from the National Science Foundation to the newly organized National Oceanic and Atmospheric Administration, where it still resides today. Participating institutions Institutions involved with the program include: Pacific region Oregon State University University of Washington University of California
https://en.wikipedia.org/wiki/Digital%20cable
Digital cable is the distribution of cable television using digital data and video compression. The technology was first developed by General Instrument. By 2000, most cable companies offered digital features, eventually replacing their previous analog-based cable by the mid 2010s. During the late 2000s, broadcast television converted to the digital HDTV standard, which was incompatible with existing analog cable systems. In addition to providing high-definition video, digital cable systems provide more services such as pay-per-view programming, cable internet access and cable telephone services. Most digital cable signals are encrypted, which reduced the incidence of cable television piracy which occurred in analog systems. History In 1990, General Instrument (acquired by Motorola and now owned by ARRIS Group) demonstrated that it was possible to use digital compression to deliver high quality HDTV in a standard 6 MHz television channel. Using the same technology General Instrument (GI) demonstrated the digital transmission of multiple high quality standard definition programs in a 6 MHz cable channel. In the 1990s, cable providers began to invest heavily in this new multi-channel digital TV technology to expand the number of channels and services available to subscribers. Increased competition and programming choices from direct-broadcast satellite services such as DirecTV, Dish Network, and PrimeStar caused cable providers to seek new ways to provide more programming. Customers were increasingly interested in more channels, pay-per-view programming, digital music services, and high speed internet services. By 2000, most cable providers in the US were offering some form of digital cable TV to their customers. Digital cable technology has allowed cable providers to compress video channels so that they take up less bandwidth and to offer two-way communication capabilities. This has enabled providers to offer more channels, video-on-demand services that don't require a separate telephone line, telephone services, high speed internet services, and interactive television services. Digital cable implements error correction to ensure the integrity of the received signal and uses a secure digital distribution system (i.e. a secure encrypted signal to prevent eavesdropping and theft of service.) Most digital cable providers use QAM for video services and DOCSIS standards for data services. Some providers have also begun to roll out video services using IPTV or Switched video. Channels Digital cable technology can allow many TV channels to occupy the frequency space that would normally be occupied by a single analog cable TV channel. The number of channels placed on a single analog frequency depends on the compression used. Many cable providers are able to fit about 10 digital SD channels or 2 digital HD channels on a single analog channel frequency. Some providers are able to squeeze more channels onto a single frequency with higher compression,
https://en.wikipedia.org/wiki/ACH%20Network
In the United States, the ACH Network is the national automated clearing house (ACH) for electronic funds transfers established in the 1960s and 1970s. It processes financial transactions for consumers, businesses, and federal, state, and local governments. ACH processes large volumes of credit and debit transactions in batches. ACH credit transfers include direct deposit for payroll, Social Security and other benefit payments, tax refunds, and vendor payments. ACH direct debit transfers include consumer payments on insurance premiums, mortgage loans, and other kinds of bills. The rules and regulations that govern the ACH network are established by National Automated Clearinghouse Association (Nacha). In 2018, the network processed 23billion transactions with a total value of $51.2trillion. Credit card payments are handled by separate networks. The Federal Reserve's FedACH and The Clearing House Payments Company's Electronic Payments Network (EPN) are the two ACH operators in the United States. History The ideas leading to the ACH arose in the late 1960s. One early predecessor was a US federal initiative used to help United States Air Force personnel get their paychecks on time. The success of this initiative led to an expansion to other employees and the government adopted it as a major payroll standard. Separately in 1968 a group of check clearinghouse associations set up The Special Committee on Paperless Entries (SCOPE) to build an automated payment system after concerns for the number of checks being cleared for payrolls. This led to the first ACH association, formed in California in 1972. Other regional ACH associations followed. The difficulty in compliance between different organizations led them to join to form National Automated Clearinghouse Association (Nacha) in 1974. Nacha consolidated and added new rules which led to ACH. As computer and telecommunication technology advanced over the next few years, the system continued to develop. By 1978, electronic funds transfers were available. From the late 1980s through to the 2000s, the system continued to develop with a number of enhancements. In 2001, there was a major reorganization of Nacha which led to financial institutions insured by the Federal Deposit Insurance Corporation becoming direct members making it much easier for the ACH network to be used by banks; that same year internet payments also went into effect which would go on to be a big part of ACH payments. Uses of the ACH payment system Bank treasury management departments sell this service to business and government customers Business-to-business payments Direct debit payment of consumer bills such as mortgages, loans, utilities, insurance premiums, rents, and any other regular payment Direct deposit of payroll, Social Security and other government payments, and tax refunds E-commerce payments Federal, state, and local tax payments Non-immediate transfer of funds between accounts at different financial instit
https://en.wikipedia.org/wiki/Train%2048
Train 48 was a Canadian improvised soap opera, broadcast on Global Television Network and CH from 2003 until 2005. The series was based on the format of an Australian television program called Going Home. Characters Liz Irwin-Gallo (Krista Sutton) (seasons 1–2, guest season 3) is a marketing expert and a senior executive at TWC, a large pharmaceutical company. After two divorces and a series of unsuccessful relationships, including with fellow passenger Randy, Liz has decided to have a child through in vitro fertilization. The character was shot in the finale of the second season and was in a coma for much of season 3, with Sutton not listed on the main cast, but was a recurring character late in the season with the character recovering from a brain injury induced by the shooting. Johnny McLaughlin (Paul Braunstein) is a "hoser". A handyman and construction foreman, he is often between jobs. He attempted to start his own business making sheds, "Sheds for Brains", before returning to construction and later, employment as a heavy-duty industrial cleaner in a downtown building. A loyal family man, he's been married for 18 years to Charity and has three daughters. He loves hockey, fishing, beer, and his family. Randy Ko (Paul Sun-Hyung Lee) is a mild-mannered engineer and scifi geek. After being dumped by his fiancée, Agnes, he enters a relationship with Liz but they broke up when she decided to have a baby through artificial insemination. Peter (Pete) Subramani (Raoul Bhaneja) is an ambitious stockbroker with questionable ethics and a succession of questionable get rich quick schemes. Lucas West (Andrew Kenneth Martin) works as a producer for a small advertising agency but has ambitions to be a film maker. He comes from a wealthy family that made its money in the oil business in Alberta but is estranged from them. He's had a string of girlfriends, some of whom are married to other men. Brenda Murphy (Lisa Merchant) comes from a small town out west and works as a clerical worker in downtown Toronto at Dubbin & Wallace, an insurance company. She's "forty-something", and has worked in the same department for 16 years. Her husband, Doug, died in a bicycle accident in August 2003. She has three children. A church-goer, Brenda means well but has traditional views which often get her into trouble with the other passengers, as does her love of gossip. Dana Davin (Joanne Boland) is an aspiring songwriter, music store employee, and activist who is returning to university and has come out as a lesbian. Zach Eisler (Joe Dinicol) is a junior graphic designer on an entertainment newspaper. Shannon Lamarche (Ingrid Hart) has worked her way up from hair salon student to expert aesthetician and co-owner of a "man spa", a chic salon in Yorkville. After Darnell in New York, Gavin, her mother's nurse, a brief dalliance with Lucas West, and, a brief affair with Anwar the weatherman, she has found true love. Dies late in season 3 from necrotizing fascii
https://en.wikipedia.org/wiki/Average%20absolute%20deviation
The average absolute deviation (AAD) of a data set is the average of the absolute deviations from a central point. It is a summary statistic of statistical dispersion or variability. In the general form, the central point can be a mean, median, mode, or the result of any other measure of central tendency or any reference value related to the given data set. AAD includes the mean absolute deviation and the median absolute deviation (both abbreviated as MAD). Measures of dispersion Several measures of statistical dispersion are defined in terms of the absolute deviation. The term "average absolute deviation" does not uniquely identify a measure of statistical dispersion, as there are several measures that can be used to measure absolute deviations, and there are several measures of central tendency that can be used as well. Thus, to uniquely identify the absolute deviation it is necessary to specify both the measure of deviation and the measure of central tendency. The statistical literature has not yet adopted a standard notation, as both the mean absolute deviation around the mean and the median absolute deviation around the median have been denoted by their initials "MAD" in the literature, which may lead to confusion, since in general, they may have values considerably different from each other. Mean absolute deviation around a central point The mean absolute deviation of a set {x1, x2, ..., xn} is The choice of measure of central tendency, , has a marked effect on the value of the mean deviation. For example, for the data set {2, 2, 3, 4, 14}: Mean absolute deviation around the mean The mean absolute deviation (MAD), also referred to as the "mean deviation" or sometimes "average absolute deviation", is the mean of the data's absolute deviations around the data's mean: the average (absolute) distance from the mean. "Average absolute deviation" can refer to either this usage, or to the general form with respect to a specified central point (see above). MAD has been proposed to be used in place of standard deviation since it corresponds better to real life. Because the MAD is a simpler measure of variability than the standard deviation, it can be useful in school teaching. This method's forecast accuracy is very closely related to the mean squared error (MSE) method which is just the average squared error of the forecasts. Although these methods are very closely related, MAD is more commonly used because it is both easier to compute (avoiding the need for squaring) and easier to understand. For the normal distribution, the ratio of mean absolute deviation from the mean to standard deviation is . Thus if X is a normally distributed random variable with expected value 0 then, see Geary (1935): In other words, for a normal distribution, mean absolute deviation is about 0.8 times the standard deviation. However, in-sample measurements deliver values of the ratio of mean average deviation / standard deviation for a given Gaussian sample
https://en.wikipedia.org/wiki/Ki%20no%20Tsurayuki
was a Japanese author, poet and court noble of the Heian period. He is best known as the principal compiler of the Kokin Wakashū, also writing its Japanese Preface, and as a possible author of the Tosa Diary, although this was published anonymously. He is well known for his waka poetry and is counted as one of the Thirty-six Poetry Immortals selected by Fujiwara no Kintō; his poetry was included also in the Hyakunin Isshu. As a courtier, he served as Governor of Tosa (930-935), Vice Governor of Kaga (917-923) and Vice Governor of Mino Province (918-923). Biography Tsurayuki was born in either 866 or 872, the son of Ki no Mochiyuki and a court dancer of the naikyoubou (内教坊), whose name is unknown. He had the childhood name of Akokuso (阿古久曽). In the 890s he became a poet of waka, short poems composed in Japanese. In 905, under the order of Emperor Daigo, he was one of four poets selected to compile the Kokin Wakashū (Kokinshu), the first imperially-sponsored anthology (chokusen-shū) of waka poetry. After holding a few offices in Kyoto, he was appointed governor of Tosa Province and stayed there from 930 until 935. Later he was presumably appointed governor of Suō Province, since it was recorded that he held a waka party (Utaai) at his home in Suo. He was one of the editors of the Kokin Wakashū. Tsurayuki wrote one of two prefaces to Kokin Wakashū; the other is in Chinese. His preface was the first critical essay on waka. He wrote of its history from its mythological origin to his contemporary waka, which he grouped into genres, referred to some major poets and gave a bit of harsh criticism to his predecessors like Ariwara no Narihira. Tsurayuki composed well over 500 poems throughout his life, including those included in the Kokinshu, imperial collections, and personal collections. One of his waka was included in the Hyakunin Isshu, which was compiled in the 13th century by Fujiwara no Teika. His death is recorded in the Sanjurokunin kasenden (三十六人歌仙伝) as potentially being May 18, 945, but it is also reported as being on June 30 of the same year. Works Kokin Wakashu In 905, under the order of Emperor Daigo, Ki no Tsurayuki was one of four poets selected to compile the Kokin Wakashū (Kokinshū), the first imperially-sponsored anthology (chokusenshū) of waka. Tsurayuki was the chief editor of the Kokinshu, and was the author of its Japanese preface (https://en.wikipedia.org/wiki/Kokin_Wakash%C5%AB). Tosa Nikki In addition to the Kokinshu, Tsurayuki's supposed major literary work (though it was published anonymously) was the Tosa Nikki (土佐日記, "Tosa Diary"), which was written using kana. The Tosa Diary is written from the perspective of a woman who embarks on a 55 day trip in 934 to the capital (present day Kyoto), where Tsurayuki had been the provincial governor, from her home in the Tosa Province. The female perspective of the diary, though written by a man (Tsurayuki), has sparked debate among scholars regarding gender and social status
https://en.wikipedia.org/wiki/Assignment%20%28computer%20science%29
In computer programming, an assignment statement sets and/or re-sets the value stored in the storage location(s) denoted by a variable name; in other words, it copies a value into the variable. In most imperative programming languages, the assignment statement (or expression) is a fundamental construct. Today, the most commonly used notation for this operation is x = expr (originally Superplan 1949–51, popularized by Fortran 1957 and C). The second most commonly used notation is x := expr (originally ALGOL 1958, popularised by Pascal). Many other notations are also in use. In some languages, the symbol used is regarded as an operator (meaning that the assignment statement as a whole returns a value). Other languages define assignment as a statement (meaning that it cannot be used in an expression). Assignments typically allow a variable to hold different values at different times during its life-span and scope. However, some languages (primarily strictly functional languages) do not allow that kind of "destructive" reassignment, as it might imply changes of non-local state. The purpose is to enforce referential transparency, i.e. functions that do not depend on the state of some variable(s), but produce the same results for a given set of parametric inputs at any point in time. Modern programs in other languages also often use similar strategies, although less strict, and only in certain parts, in order to reduce complexity, normally in conjunction with complementing methodologies such as data structuring, structured programming and object orientation. Semantics An assignment operation is a process in imperative programming in which different values are associated with a particular variable name as time passes. The program, in such model, operates by changing its state using successive assignment statements. Primitives of imperative programming languages rely on assignment to do iteration. At the lowest level, assignment is implemented using machine operations such as MOVE or STORE. Variables are containers for values. It is possible to put a value into a variable and later replace it with a new one. An assignment operation modifies the current state of the executing program. Consequently, assignment is dependent on the concept of variables. In an assignment: The expression is evaluated in the current state of the program. The variable is assigned the computed value, replacing the prior value of that variable. Example: Assuming that a is a numeric variable, the assignment a := 2*a means that the content of the variable a is doubled after the execution of the statement. An example segment of C code: int x = 10; float y; x = 23; y = 32.4f; In this sample, the variable x is first declared as an int, and is then assigned the value of 10. Notice that the declaration and assignment occur in the same statement. In the second line, y is declared without an assignment. In the third line, x is reassigned the value of 23. Finally, y is assigned
https://en.wikipedia.org/wiki/Ontario%20Provincial%20Highway%20Network
The Provincial Highway Network consists of all the roads in Ontario maintained by the Ministry of Transportation of Ontario (MTO), including those designated as part of the King's Highway, secondary highways, and tertiary roads. Components of the system—comprising of roads and 2,880 bridges —range in scale from Highway 401, the busiest highway in North America, to unpaved forestry and mining access roads. The longest highway is nearly long, while the shortest is less than a kilometre. Some roads are unsigned highways, lacking signage to indicate their maintenance by the MTO; these may be remnants of highways that are still under provincial control whose designations were decommissioned, roadway segments left over from realignment projects, or proposed highway corridors. Predecessors to today's modern highways include the foot trails and portages used by indigenous peoples in the time before European settlement. Shortly after the creation of the Province of Upper Canada in 1791, the new government under John Graves Simcoe built overland military roads to supplement water-based transportation, including Yonge Street and Dundas Street. At the time, road construction was under the control of the township and county governments. Local township roads were financed and constructed through a statute labour system that required landowners to make improvements in lieu of taxes. Private companies constructed corduroy and later plank roads and charged tolls in the second half of the 19th century. The rising popularity of the bicycle led to the formation of Ontario Good Roads Association, which advocated for the improvement of roads and recreation as the automobile rose to prominence. By the early 20th century, the province had taken interest in road improvement and began funding it through counties. The increasing adoption of the automobile resulted in the formation of the Department of Public Highways of Ontario (DPHO) in 1916. The passing of the Canada Highways Act in 1919 resulted in the establishment of a provincial network of highways. The DPHO assigned internal highway numbers to roads in the system, and in 1925, the numbers were signposted along the roads and marked on maps. In 1930, provincial highways were renamed King's Highways and the familiar crown route markers created. The DPHO was also renamed the Department of Highways (DHO). The 1930s saw several major depression relief projects built by manual labour, including the first inter-city divided highway in North America along the Middle Road, which would become the Queen Elizabeth Way in 1939. In 1937, the DHO merged with the Department of Northern Development, extending the highway network into the Canadian Shield and Northern Ontario. Significant traffic engineering and surveying through the war years, during which construction came to a near standstill, led to the planning and initial construction of controlled-access highways. The 400-series highways were built beginning in the late 194
https://en.wikipedia.org/wiki/B%C3%BCchi%20automaton
In computer science and automata theory, a deterministic Büchi automaton is a theoretical machine which either accepts or rejects infinite inputs. Such a machine has a set of states and a transition function, which determines which state the machine should move to from its current state when it reads the next input character. Some states are accepting states and one state is the start state. The machine accepts an input if and only if it will pass through an accepting state infinitely many times as it reads the input. A non-deterministic Büchi automaton, later referred to just as a Büchi automaton, has a transition function which may have multiple outputs, leading to many possible paths for the same input; it accepts an infinite input if and only if some possible path is accepting. Deterministic and non-deterministic Büchi automata generalize deterministic finite automata and nondeterministic finite automata to infinite inputs. Each are types of ω-automata. Büchi automata recognize the ω-regular languages, the infinite word version of regular languages. They are named after the Swiss mathematician Julius Richard Büchi, who invented them in 1962. Büchi automata are often used in model checking as an automata-theoretic version of a formula in linear temporal logic. Formal definition Formally, a deterministic Büchi automaton is a tuple A = (Q,Σ,δ,q0,F) that consists of the following components: Q is a finite set. The elements of Q are called the states of A. Σ is a finite set called the alphabet of A. δ: Q × Σ → Q is a function, called the transition function of A. q0 is an element of Q, called the initial state of A. F⊆Q is the acceptance condition. A accepts exactly those runs in which at least one of the infinitely often occurring states is in F. In a (non-deterministic) Büchi automaton, the transition function δ is replaced with a transition relation Δ that returns a set of states, and the single initial state q0 is replaced by a set I of initial states. Generally, the term Büchi automaton without qualifier refers to non-deterministic Büchi automata. For more comprehensive formalism see also ω-automaton. Closure properties The set of Büchi automata is closed under the following operations. Let and be Büchi automata and be a finite automaton. Union: There is a Büchi automaton that recognizes the language Proof: If we assume, w.l.o.g., is empty then is recognized by the Büchi automaton Intersection: There is a Büchi automaton that recognizes the language Proof: The Büchi automaton recognizes where By construction, is a run of automaton A' on input word w if is run of A on w and is run of B on w. is accepting and is accepting if r' is concatenation of an infinite series of finite segments of 1-states (states with third component 1) and 2-states (states with third component 2) alternatively. There is such a series of segments of r' if r' is accepted by A'. Concatenation: There is a Büchi automaton th
https://en.wikipedia.org/wiki/Neuroevolution%20of%20augmenting%20topologies
NeuroEvolution of Augmenting Topologies (NEAT) is a genetic algorithm (GA) for the generation of evolving artificial neural networks (a neuroevolution technique) developed by Kenneth Stanley and Risto Miikkulainen in 2002 while at The University of Texas at Austin. It alters both the weighting parameters and structures of networks, attempting to find a balance between the fitness of evolved solutions and their diversity. It is based on applying three key techniques: tracking genes with history markers to allow crossover among topologies, applying speciation (the evolution of species) to preserve innovations, and developing topologies incrementally from simple initial structures ("complexifying"). Performance On simple control tasks, the NEAT algorithm often arrives at effective networks more quickly than other contemporary neuro-evolutionary techniques and reinforcement learning methods. Algorithm Traditionally, a neural network topology is chosen by a human experimenter, and effective connection weight values are learned through a training procedure. This yields a situation whereby a trial and error process may be necessary in order to determine an appropriate topology. NEAT is an example of a topology and weight evolving artificial neural network (TWEANN) which attempts to simultaneously learn weight values and an appropriate topology for a neural network. In order to encode the network into a phenotype for the GA, NEAT uses a direct encoding scheme which means every connection and neuron is explicitly represented. This is in contrast to indirect encoding schemes which define rules that allow the network to be constructed without explicitly representing every connection and neuron, allowing for more compact representation. The NEAT approach begins with a perceptron-like feed-forward network of only input neurons and output neurons. As evolution progresses through discrete steps, the complexity of the network's topology may grow, either by inserting a new neuron into a connection path, or by creating a new connection between (formerly unconnected) neurons. Competing conventions The competing conventions problem arises when there is more than one way of representing information in a phenotype. For example, if a genome contains neurons A, B and C and is represented by [A B C], if this genome is crossed with an identical genome (in terms of functionality) but ordered [C B A] crossover will yield children that are missing information ([A B A] or [C B C]), in fact 1/3 of the information has been lost in this example. NEAT solves this problem by tracking the history of genes by the use of a global innovation number which increases as new genes are added. When adding a new gene the global innovation number is incremented and assigned to that gene. Thus the higher the number the more recently the gene was added. For a particular generation if an identical mutation occurs in more than one genome they are both given the same number, beyond that h
https://en.wikipedia.org/wiki/Little%20Barford
Little Barford is a hamlet and civil parish in the Borough of Bedford in Bedfordshire, England about northeast of the county town of Bedford. The 2011 census combines other data for Little Barford with Wyboston, Chawston and Colesden civil parish but its population is separately shown as 44. Little Barford Power Station is to the north of the village. Geography Little Barford is south of St Neots, west of Cambridge and north of Central London. The East Coast Mainline railway runs north–south through the parish. Area The civil parish covers an area of about . The parish's northern boundary is the A428 road, Cambridgeshire is to the east and Central Bedfordshire to the south. The River Great Ouse forms its western boundary. Landscape The hamlet lies within the Bedfordshire and Cambridgeshire Claylands as designated by Natural England. Bedford Borough Council classifies the local landscape as the Great Ouse Clay Valley around and west of the village, and the eastern part of the parish as within the Biggin Wood Clay Vale that extends northwards from Central Bedfordshire. The surrounding area is mostly arable farmland. The northernmost part of the parish between the railway and the Great Ouse is taken up by the power station and industrial units. Cottages A row of four Grade II listed 18th century rough cast, thatched cottages stand by the roadside. One is detached, the remainder are terraced. At the northern end of the hamlet are four pairs of late 19th century semi-detached estate cottages having yellow brickwork with red brick dressings. Elevation The hamlet is above sea level. The whole parish is mainly low lying and flat, although the land does rise to in the southeast corner of the parish. Geology and soil type The hamlet lies mainly on first and second terrace river gravel and glacial gravel. Boulder clay is to the east. Alluvium borders the Great Ouse. Underlying these superficial deposits and also in part at the surface is Oxford clay and Kellaways beds. Around the hamlet the soil has low fertility, is freely draining and slightly acid with a loamy texture. The eastern part of the parish has highly fertile, lime-rich loamy and clayey soils with impeded drainage. By the Great Ouse are loamy and clayey floodplain soils with naturally high groundwater. The night sky and light pollution Light pollution is the level of radiance (night lights) shining up into the night sky. The Campaign to Protect Rural England (CPRE) divides the level of night sky brightness into 9 bands with band 1 being the darkest i.e. with the lowest level of light pollution and band 9 the brightest and most polluted. Little Barford in bands 6 and 7 is adversely affected by lighting at the power station and Arlington Road Industrial Estate. The night sky is darker looking southeast. Public footpaths The sole public footpath runs from opposite the last cottage at the northern end of the hamlet to the Ouse and onwards to link with the Ouse Valley Way.
https://en.wikipedia.org/wiki/D.C.%20Sniper%3A%2023%20Days%20of%20Fear
D.C. Sniper: 23 Days of Fear (also known as Sniper: 23 Days of Fear in Washington D.C.) is a 2003 TV movie created by USA Network based on the Beltway sniper attacks of 2002. The films chronicles the period when John Allen Muhammad (played by Bobby Hosea) and Lee Boyd Malvo (played by Trent Cameron) went on a serial killing spree in October 2002 in Virginia, Washington, D.C., and Maryland, all parts of the Washington Metropolitan Area, the entire area of which was held in a "grip of terror." Plot In October 2002, Chief Charles Moose (played by Charles S. Dutton) of the Montgomery County Police Department, heads an effort to track down those responsible for a recent string of murders in Montgomery County, Maryland. Unable to give anything but small pieces of information at various press conferences held during the 23 dark days, Moose finds himself vilified and derided in many corners as ineffectual and incompetent. Indeed, quite a few newspapers outside the area targeted by snipers came right out and called for Moose's resignation. But the chief's dogged persistence ultimately paid off and — in the sort of twist that a professional writer of thrillers might dismiss as inconceivable — the two men arrested for the carnage turned out to be the archetypal "least likely suspects." Release D.C. Sniper: 23 Days of Fear originally aired on the USA Network on October 17, 2003, just as John Allen Muhammad and John Lee Malvo's murder trials were getting underway. References External links 2003 films 2003 television films Films scored by Mark Snow Films about snipers Crime films based on actual events Films directed by Tom McLoughlin Films set in 2002 Films set in Montgomery County, Maryland Films set in Virginia Films set in Washington, D.C. USA Network original films Films set in Frederick County, Maryland
https://en.wikipedia.org/wiki/VEGAS%20algorithm
The VEGAS algorithm, due to G. Peter Lepage, is a method for reducing error in Monte Carlo simulations by using a known or approximate probability distribution function to concentrate the search in those areas of the integrand that make the greatest contribution to the final integral. The VEGAS algorithm is based on importance sampling. It samples points from the probability distribution described by the function so that the points are concentrated in the regions that make the largest contribution to the integral. The GNU Scientific Library (GSL) provides a VEGAS routine. Sampling method In general, if the Monte Carlo integral of over a volume is sampled with points distributed according to a probability distribution described by the function we obtain an estimate The variance of the new estimate is then where is the variance of the original estimate, If the probability distribution is chosen as then it can be shown that the variance vanishes, and the error in the estimate will be zero. In practice it is not possible to sample from the exact distribution g for an arbitrary function, so importance sampling algorithms aim to produce efficient approximations to the desired distribution. Approximation of probability distribution The VEGAS algorithm approximates the exact distribution by making a number of passes over the integration region while histogramming the function f. Each histogram is used to define a sampling distribution for the next pass. Asymptotically this procedure converges to the desired distribution. In order to avoid the number of histogram bins growing like with dimension d the probability distribution is approximated by a separable function: so that the number of bins required is only Kd. This is equivalent to locating the peaks of the function from the projections of the integrand onto the coordinate axes. The efficiency of VEGAS depends on the validity of this assumption. It is most efficient when the peaks of the integrand are well-localized. If an integrand can be rewritten in a form which is approximately separable this will increase the efficiency of integration with VEGAS. See also Las Vegas algorithm Monte Carlo integration Importance sampling References Monte Carlo methods Computational physics Statistical algorithms Variance reduction
https://en.wikipedia.org/wiki/Jonathan%20Schaeffer
Jonathan Herbert Schaeffer (born 1957) is a Canadian researcher and professor at the University of Alberta and the former Canada Research Chair in Artificial Intelligence. He led the team that wrote Chinook, the world's strongest American checkers player, after some relatively good results in writing computer chess programs. He is involved in the University of Alberta GAMES group developing computer poker systems. Schaeffer is also a member of the research group that created Polaris, a program designed to play the Texas Hold'em variant of poker. He is a Founder of Onlea, which produces online learning experiences. Early life Born in Toronto, Ontario, he received a Bachelor of Science degree in 1979 from the University of Toronto. He received a Master of Mathematics degree in 1980 and a Ph.D. in 1986 from the University of Waterloo. Schaeffer reached national master strength in chess while in his early 20s, but has played little competitive chess since that time. Draughts: Chinook Chinook is the first computer program to win the world champion title in a competition against humans. In 1990 it won the right to play in the human World Championship by being second to Marion Tinsley in the US Nationals. At first the American Checkers Federation and English Draughts Association were against the participation of a computer in a human championship. When Tinsley resigned his title in protest, the ACF and EDA created the new title Man vs. Machine World Championship, and competition proceeded. Tinsley won with four wins to Chinook's two. In a rematch, Chinook was declared the Man-Machine World Champion in checkers in 1994 in a match against Marion Tinsley after six drawn games, and Tinsley's withdrawal due to pancreatic cancer. While Chinook became the world champion, it had never defeated the best checkers player of all time, Tinsley, who was significantly superior to even his closest peer. The championship continued with Chinook defending its title against Don Lafferty when it lost one game, won one and drew 18. After the match, Jonathan Schaeffer decided not to let Chinook compete anymore, but instead try to solve checkers. It was rated at 2814 Elo. In 2007, after 18 years of computation, he proved through a weak solution that checkers always results in a draw if neither player makes a mistake. The solution involved 1014 calculations from endgame positions with fewer than 10 pieces on the board. Poker: Polaris Schaeffer is a member and, until 2004, leader of the computer poker research group at the University of Alberta, which has developed several strong computer programs for playing Texas hold 'em poker. The earliest and most general of these is Poki, which uses Monte Carlo simulation to choose actions during a game. More recently, the group has focused on the two-player (Heads-Up) variant, and has developed a series of programs that approximate Nash equilibrium strategies for the game. Several of these programs (such as Poki, SparBot and
https://en.wikipedia.org/wiki/FM
FM or Fm may refer to: Technology and computing Frequency modulation, the carrying of information over an electromagnetic wave by varying its frequency. Its most common applications are: FM broadcasting, used primarily to broadcast music and speech at VHF frequencies via radio FM broadcast band, frequency band used for FM broadcasting .fm, country-code top-level domain (ccTLD) of the Federated States of Micronesia FM Towns, one of a line of personal computers by Fujitsu Formula Mazda, SCCA Class designation FM Nissan FM platform, a "front midships" automobile layout Volvo FM, heavy truck range Adobe FrameMaker, an Adobe software Science and medicine fm, the symbol for femtometre, a unit of length also called a fermi fM, the symbol for femtomolar, a unit of molar concentration Fermium, a chemical element with symbol Fm FM (chemotherapy), a chemotherapy regimen using Fludarabine and Mitoxantrone Family medicine, a medical specialty devoted to comprehensive health care for people of all ages Fibromyalgia, a chronic syndrome with widespread pain etc. Film and television FM (film), 1978 FM (TV channel), American TV network FM (American TV series), an American sitcom broadcast from 1989–1990 FM (British TV series), British sitcom broadcast in 2009 Music Musical artists FM (British band), a British AOR/hard rock band FM (Canadian band), a Canadian progressive rock band FM Static, a Canadian pop punk duo F.M. Einheit (born 1958), German musician and ex-member of Einstürzende Neubauten Far East Movement, a hip hop/electropop quartet based in Los Angeles Other uses in music FM (soundtrack), the original soundtrack to the 1978 film FM "FM (No Static at All)", a 1978 song by Steely Dan from the soundtrack FM Records, a record label "FM" (short for "frequent mutilation"), a song by The Slits FM (EP), a 2015 extended play by South Korean girl group Crayon Pop FM!, a 2018 album by Vince Staples F minor, chords and scale FM (album), a 2006 album by the Turkish band Replikas Businesses and organizations FM Global, a mutual insurance company FM Records, a record label Fannie Mae, a colloquial name for the United States' Federal National Mortgage Association Future Movement, a Lebanese political party Franklin & Marshall College or F&M, a private liberal arts college in Lancaster, Pennsylvania, U.S. Shanghai Airlines (IATA airline designator), an airline headquartered in Shanghai People F.M. Einheit (born 1958), German musician, ex-member of Einstürzende Neubauten FM-2030 (1930–2000), transhumanist philosopher Ferdinand Marcos (1917–1989), Filipino politician and former President of the Philippines Places Federated States of Micronesia (ISO 3166-1 country code, FIPS PUB 10-4 territory code, and postal code) Province of Fermo, a province of Italy (ISO 3166-2:IT code) Flower Mound, Texas, a city in the Dallas-Fort Worth Metroplex Titles FIDE Master, a title awarded by the world chess governing body Field ma
https://en.wikipedia.org/wiki/Dartford%20warbler
The Dartford warbler (Curruca undata) is a typical warbler from the warmer parts of western Europe and northwestern Africa. It is a small warbler with a long thin tail and a thin pointed bill. The adult male has grey-brown upperparts and is dull reddish-brown below except for the centre of the belly which has a dirty white patch. It has light speckles on the throat and a red eye-ring. The sexes are similar but the adult female is usually less grey above and paler below. Its breeding range lies west of a line from southern England to the heel of Italy (southern Apulia). The Dartford warbler is usually resident all year in its breeding range, but there is some limited migration. Taxonomy and systematics The Dartford warbler was first described in 1776 by the Welsh naturalist, Thomas Pennant. He introduced the English name and based his description on two specimens that had been obtained by the ornithologist John Latham from Bexley Heath, near Dartford in Kent. In 1783 Latham included the warbler in his A General Synopsis of Birds but did not coin the binomial name, Sylvia dartfordiensis, until the supplement to his work was published in 1787. However, in 1783, before the publication of Latham's supplement, the Dutch naturalist Pieter Boddaert introduced the name, Motacilla undata, based on a coloured plate of "Le Pitte-chou, de Provence" in Edmé-Louis Daubenton's Planches enluminées d'histoire naturelle. The specific epithet undata is from Medieval Latin undatus meaning "with wavy markings". The type locality is Provence in France. This species probably forms a superspecies with Tristram's warbler and this in turn seems close to Marmora's warbler and the Balearic warbler. Altogether, this group of typical warblers bears a resemblance to the wrentit, the only species of Sylviidae from the Americas. However, the wrentit is less closely related to the genus Sylvia than to the parrotbills. Its visual similarity to the Dartford warbler group is an example of convergent evolution. Three subspecies are recognised: C. u. dartfordiensis (Latham, 1787) – south England and north-west France C. u. toni (Hartert, 1909) – north-west Africa C. u. undata (Boddaert, 1783) – Iberian Peninsula and south France Description The Dartford warbler is a small, , passerine bird, distinguished by its long tail compared with that of other warblers. Its plumage comprises unobtrusive and muted tones, which blend in with the dry dead plants, old wood or sunny greyish wood found in its preferred habitats. Like many typical warblers, the Dartford warbler has distinct male and female plumages. The male has a grey back and head, reddish underparts, and a red eye. The reddish throat is spotted with white. The sides are a dull greyish tone, being more clear about the abdomen. In some populations males have bluish-grey or brownish-grey backs and heads. The female is paler below, especially on the throat, and a browner grey above. The female's throat also has white spots, altho
https://en.wikipedia.org/wiki/System%207
System 7, codenamed "Big Bang", and later (as of version 7.6) also known as Mac OS 7, is a graphical user interface-based operating system for Macintosh computers and is part of the classic Mac OS series of operating systems. It was introduced on May 13, 1991, by Apple Computer It succeeded System 6, and was the main Macintosh operating system until it was succeeded by Mac OS 8 in 1997. Current for more than six years, System 7 was the longest-lived major version series of the classic Macintosh operating system (to date, only Mac OS X had a longer lifespan). Features added with the System 7 release included virtual memory, personal file sharing, QuickTime, QuickDraw 3D, and an improved user interface. With the release of version 7.6 in 1997, Apple officially renamed the operating system "Mac OS", a name that had first appeared on System 7.5.1's boot screen. System 7 was developed for Macs that used the Motorola 680x0 line of processors, but was ported to the PowerPC after Apple adopted the new processor in 1994 with the introduction of the Power Macintosh. Development The development of the Macintosh system software up to System 6 followed a fairly smooth progression with the addition of new features and relatively small changes and upgrades over time. Major additions were fairly limited. Some perspective on the scope of the changes can be seen by examining the official system documentation, Inside Macintosh. This initially shipped in three volumes, adding another to describe the changes introduced with the Mac Plus, and another for the Mac II and Mac SE. These limited changes meant that the original Macintosh system remained largely as it was when initially introduced. That is, the machine was geared towards a single user and task running on a floppy disk based machine of extremely limited RAM. However, many of the assumptions of this model were no longer appropriate. Most notable among these was the single-tasking model, the replacement of which had first been examined in 1986's "Switcher" and then replaced outright with MultiFinder in System 5. Running MultiFinder normally required a larger amount of RAM and a hard drive, but these became more common by the late 1980s. While additions had been relatively limited, so had fixes to some of the underlying oddities of the system architecture. For instance, to support a limited form of multitasking, the original Mac OS supported small co-resident programs known as desk accessories, which had to be installed into the system using special tools. If the system were able to support multiple tasks, this one-off solution would no longer be needed — desk accessories could simply be small programs, placed anywhere. Yet, as MultiFinder was still optional, such a step had not been taken. Numerous examples of this sort of problem could be found throughout the system. Finally, the widespread adoption of hard drives and local area networks led to any number of new features being requested from users and de
https://en.wikipedia.org/wiki/Desk%20accessory
A desk accessory (DA) in computing is a small transient or auxiliary application that can be run concurrently in a desktop environment with any other application on the system. Early examples, such as Sidekick and Macintosh desk accessories, used special programming models to provide a small degree of multitasking on systems that initially did not have any other multitasking ability. Personal information managers Early personal information managers, such as Norton Desktop and Borland's Sidekick, provided pop-up calculator, alarm, calendar and other functions for single-tasking operating systems like MS-DOS using terminate-and-stay-resident techniques. Apple Macintosh Introduced in 1984 as part of the operating system for the Apple Macintosh computer, a Desk Accessory (DA) was a piece of software written as a device driver, conforming to a particular programming model. The purpose of this model was to permit very small helper-type applications to be run concurrently with any other application on the system. This provided a small degree of multitasking on a system that initially did not have any other multitasking ability. DAs were implemented as a special class of driver. It was installed in the driver queue, and given time periodically and co-operatively as a result of the host application calling SystemTask() within its main loop. A DA was permitted to have a user interface as long as it was confined to one main window. A special window frame with black title bar and rounded corners was reserved for the use of DAs so that the user could distinguish it from the windows of the hosting application. Typical early DAs included the Calculator and Alarm Clock. The control panel, Chooser, and Scrapbook were initially implemented as DAs. Third-party DAs such as spelling checkers could be purchased. It was considered hard to write a DA, especially early on when there was little in the way of developer tools. However, since on the early Mac OS drivers did not have any special privileges, writing a DA was, with practice, no more difficult than any other application. A special Font/DA Mover utility was used to change the configuration of DAs. Because DAs were not installed or launched in the same way that applications were, the user could not drag and drop DAs into or out of the system. They resided in the System file's 'DRVR' resources, like actual drivers, though they could be installed in any file whose resources were loaded into the memory, and were stored in "suitcases" when not installed in the system file. If installed within a separate application, such as MacWrite, their functionality would be accessible only when that application was running. That is, a desk accessory installed as a resource within an application would appear on the Apple menu as a desk accessory only when that application was active. It could then be activated while the application was run and would then disappear when the application was terminated through the Quit function
https://en.wikipedia.org/wiki/System%207%20%28disambiguation%29
System 7 or System/7 may refer to: Computing IBM System/7, a minicomputer developed by IBM; premiered in 1971 Operating systems: Macintosh System 7, the Apple operating system introduced in 1991 Operating System/7, the UNIVAC operating system introduced in 1974 System Manager 7, a successor to Multiuser DOS by Datapac Version 7 Unix, otherwise known as (Unix Time Sharing) System 7; released in 1979 Other System 7 (album), eponymous, debut studio album of System 7; released in 1991 System 7 (band), a British electronic music duo Signalling System 7, telephony signaling protocols System 7 Napoleonics, tabletop miniatures wargaming STS-7 (Space Transportation System-7), the Space Shuttle mission See also Series 7 (disambiguation)
https://en.wikipedia.org/wiki/Kwun%20Tong%20line
The Kwun Tong line () is a rapid transit line of the MTR network in Hong Kong, coloured green on the MTR map. Starting at Whampoa in Hung Hom and ending at Tiu Keng Leng in Tseung Kwan O, Sai Kung, the route has 17 stations and takes 35 minutes to complete. The Kwun Tong line is one of the busiest railway lines on the network connecting the central and the eastern portions of Kowloon via Wong Tai Sin. The line is mostly underground, but includes a lengthy elevated section, and runs generally in an east-west direction. During the morning rush hour, the Kwun Tong line utilises 33 trains running at 29tph (trains per hour) to achieve a route capacity of 85,000 pphpd. Opened on 1 October 1979 as the first urban railway line in Hong Kong and the first operated by the Mass Transit Railway Corporation (MTRC), the Kwun Tong line operates over much of the original section of the "Modified Initial System", from Shek Kip Mei to Kwun Tong station, which it is named after. The line has seen the most changes in alignment of all the MTR lines, the most recent in 2016. It has crossed Victoria Harbour to serve Hong Kong Island using two separate routes in its history, though other lines have since taken over the harbour crossings; as such, the current route lies entirely within Kowloon except for Tiu Keng Leng in the New Territories. It has interchanges with four other lines: the Tsuen Wan line at Yau Ma Tei- Prince Edward station, the East Rail line at Kowloon Tong station, the Tuen Ma line at Ho Man Tin and Diamond Hill station, and the Tseung Kwan O line at Yau Tong and Tiu Keng Leng station. Route map History Initial section The Kwun Tong line operates over the majority of the track used by the "Modified Initial System", and can so be said to be the first MTR line to enter service. It was predated only by the suburban East Rail line, which at the time was in the process of being electrified and upgraded to a commuter service as part of the Kowloon–Canton Railway. Construction was approved in November 1975 under the administration of Governor Murray MacLehose, and service commenced on 1 October 1979. The line initially ran between Shek Kip Mei station and Kwun Tong station, and each train consisted of four cars. The trains were expanded to consist of 6 cars and 8 cars later respectively. It was extended to the south twice: firstly to Tsim Sha Tsui on 31 December 1979, and secondly to Central station on 12 February 1980 (named Chater at the time), crossing Victoria Harbour through the first underwater rail tunnel in Hong Kong and completing the original Modified Initial System plan. When the Tsuen Wan line started service in May 1982, it took over the section of the Modified Initial System south of Argyle (present-day Mong Kok). At that point, Waterloo (present-day Yau Ma Tei) station became the terminus of the newly christened Kwun Tong line (until then, the line had no official name), and both Argyle and Prince Edward stations became cross-platform inter
https://en.wikipedia.org/wiki/K%26R
K&R may refer to: Kernighan and Ritchie (Brian Kernighan and Dennis Ritchie) The C Programming Language (book), a book written by Brian Kernighan and Dennis Ritchie K&R C, the original dialect of the C programming language, introduced by the first edition of the book K&R indent style, used in the book K&R Insurance, a kidnap and ransom insurance
https://en.wikipedia.org/wiki/Apple%20Computer%2C%20Inc.%20v.%20Microsoft%20Corp.
Apple Computer, Inc. v. Microsoft Corporation, 35 F.3d 1435 (9th Cir. 1994), was a copyright infringement lawsuit in which Apple Computer, Inc. (now Apple Inc.) sought to prevent Microsoft and Hewlett-Packard from using visual graphical user interface (GUI) elements that were similar to those in Apple's Lisa and Macintosh operating systems. The court ruled that, "Apple cannot get patent-like protection for the idea of a graphical user interface, or the idea of a desktop metaphor [under copyright law]...". In the midst of the Apple v. Microsoft lawsuit, Xerox also sued Apple alleging that Mac's GUI was heavily based on Xerox's. The district court dismissed Xerox's claims without addressing whether Apple's GUI infringed Xerox's. Apple lost all claims in the Microsoft suit except for the ruling that the trash can icon and folder icons from Hewlett-Packard's NewWave windows application were infringing. The lawsuit was filed in 1988 and lasted four years; the decision was affirmed on appeal in 1994, and Apple's appeal to the U.S. Supreme Court was denied. Background Apple Inc. had agreed to license certain parts of its GUI to Microsoft for use in Windows 1.0, but when Microsoft made changes in Windows 2.0 adding overlapping windows and other features found in the Macintosh GUI, Apple filed suit. Apple added additional claims to the suit when Microsoft released Windows 3.0. Apple claimed the "look and feel" of the Macintosh operating system, taken as a whole, was protected by copyright, and that each element of the interface (such as the existence of windows on the screen, the rectangular appearance of windows, windows could be resized, overlap, and have title bars) was not as important as all these elements taken together. After oral arguments, the court insisted on an analysis of specific GUI elements that Apple claimed were infringements. Apple listed 189 GUI elements; the court decided that 179 of these elements had been licensed to Microsoft in the Windows 1.0 agreement and most of the remaining 10 elements were not copyrightable —either they were unoriginal to Apple, or they were the only possible way of expressing a particular idea. Midway through the suit, Xerox filed a lawsuit against Apple claiming Apple had infringed copyrights Xerox held on its GUI's. Xerox had invited the Macintosh design team to view their GUI computers at the PARC research lab; these visits had been very influential on the development of the Macintosh GUI. Xerox's lawsuit appeared to be a defensive move to ensure that if Apple v. Microsoft established that "look and feel" was copyrightable, then Xerox would be the primary beneficiary, rather than Apple. The Xerox case was dismissed, for a variety of legal reasons, most notably that Xerox "had waited too long to file a copyright infringement case and had to resort to a weaker charge of unfair competition". Court case The district court ruled that it would require a standard of "virtual identity" between Windows and th
https://en.wikipedia.org/wiki/Backhaul%20%28broadcasting%29
In the context of broadcasting, backhaul refers to uncut program content that is transmitted point-to-point to an individual television station or radio station, broadcast network or other receiving entity where it will be integrated into a finished TV show or radio show. The term is independent of the medium being used to send the backhaul, but communications satellite transmission is very common. When the medium is satellite, it is called a wildfeed. Backhauls are also referred to sometimes as clean feeds, being clean in the sense that they lack any of the post-production elements that are added later to the feed's content (i.e. on-screen graphics, voice-overs, bumpers, etc.) during the integration of the backhaul feed into a finished show. In live sports production, a backhaul is used to obtain live game footage (usually for later repackaging in highlights shows) when an off-air source is not readily available. In this instance the feed that is being obtained contains all elements except for TV commercials or radio ads run by the host network's master control. This is particularly useful for obtaining live coverage of post-game press conferences or extended game highlights (melts), since the backhaul may stay up to feed these events after the network has concluded their broadcast. Electronic news gathering, including live via satellite interviews, reporters' live shots, and sporting events are all examples of radio or television content that is backhauled to a station or network before being made available to the public through that station or network. Cable TV channels, particularly public, educational, and government access (PEG) along with (local origination) channels, may also backhauled to cable headends before making their way to the subscriber. Finished network feeds are not considered backhauls, even if local insertion is used to modify the content prior to final transmission. There exists a dedicated group of enthusiasts who use TVRO (TV receive-only) gear such as satellite dishes to peek in on backhaul signals that are available on any of the dozens of broadcast satellites that are visible from almost any point on Earth. In its early days, their hobby was strengthened by the fact that most backhaul was analog and in the clear (unscrambled or unencrypted) which made for a vast smorgasbord of free television available for the technically inclined amateur. In recent years, full-time content and cable channels have added encryption and conditional access, and occasional signals are steadily becoming digital, which has had a deleterious effect on the hobby. Some digital signals remain freely accessible (sometimes using Ku band dishes as small as one meter) under the international DVB-S standard or the US Motorola-proprietary Digicipher system. The small dishes may either be fixed (much like DBS antennas), positioned using a rotor (usually DiSEqC-standard) or may be toroidal in design (twin toroidal reflectors focus the incoming sig
https://en.wikipedia.org/wiki/Ghost%20in%20the%20Shell%202%3A%20Innocence
Ghost in the Shell 2: Innocence, known in Japan as just , is a 2004 Japanese animated cyberpunk film written and directed by Mamoru Oshii. The film serves as a standalone sequel to Oshii's 1995 film Ghost in the Shell and is loosely based on the manga by Masamune Shirow. The film was co-produced by Production I.G and Studio Ghibli for Tokuma Shoten, Nippon Television Network, Dentsu, Buena Vista Home Entertainment, Toho and D-Rights, and distributed by Toho. It was released in Japan on March 6, 2004, and was later released in the US on September 17, 2004, by Go Fish Pictures. The film's US box office performance had exceeded Go Fish Pictures' expectation. Innocence received the 2004 Nihon SF Taisho Award. It was also in competition at the 2004 Cannes Film Festival, making it the first and only anime in history to compete for the Palme d'Or. The soundtrack for the film was released under the name Innocence O.S.T. and a related novel called Ghost in the Shell: Innocence - After the Long Goodbye was released on February 29, 2004. Plot The story is loosely based on Ghost in the Shell manga chapter "Robot Rondo" (with elements of "Phantom Fund"). Opening in 2032, Public Security Section 9 cybernetic operative Batou is teamed with Togusa, an agent with very few cybernetic upgrades, following the events of Ghost in the Shell. After a series of deaths due to malfunctioning gynoids—doll-like sex robots—Section 9 is asked to investigate. As the gynoids all malfunctioned without clear cause, the deaths are believed to be premeditated murders; Batou and Togusa are sent to investigate possible terrorist or political motives. Additionally, the most recent gynoid's remains show they all contained an illegal "ghost". Section 9 concludes human sentience is being artificially duplicated onto the dolls illegally, making the robots more lifelike, and possibly acting as a motive in the murders. Called to a homicide scene, information warfare/technology specialist Ishikawa explains the victim is Jack Walkson, a consignment officer at gynoid company LOCUS SOLUS, who may have been killed by the Yakuza. A previous Yakuza boss was recently killed by a gynoid, so Ishikawa concludes Walkson was held responsible and killed in an act of revenge. Batou and Togusa enter a Yakuza bar to question the current boss, only to be threatened by the bar occupants. Batou opens fire, killing and wounding numerous gang members, including the cyborg that murdered Walkson. The current boss then admits his predecessor was somehow involved in LOCUS SOLUS, but insists he doesn't know how. Entering a store on his way home, Batou is then seemingly warned by the Major and shot in the arm by an unseen assailant. Caught in a firefight, Batou nearly kills the store owner in confusion, but is subdued when Ishikawa appears. While having his damaged arm replaced, Batou is informed by Ishikawa that his e-brain was hacked, causing him to shoot himself and attack the store occupants. Ishikawa explain
https://en.wikipedia.org/wiki/Database%20schema
The database schema is the structure of a database described in a formal language supported typically by a relational database management system (RDBMS). The term "schema" refers to the organization of data as a blueprint of how the database is constructed (divided into database tables in the case of relational databases). The formal definition of a database schema is a set of formulas (sentences) called integrity constraints imposed on a database. These integrity constraints ensure compatibility between parts of the schema. All constraints are expressible in the same language. A database can be considered a structure in realization of the database language. The states of a created conceptual schema are transformed into an explicit mapping, the database schema. This describes how real-world entities are modeled in the database. "A database schema specifies, based on the database administrator's knowledge of possible applications, the facts that can enter the database, or those of interest to the possible end-users." The notion of a database schema plays the same role as the notion of theory in predicate calculus. A model of this "theory" closely corresponds to a database, which can be seen at any instant of time as a mathematical object. Thus a schema can contain formulas representing integrity constraints specifically for an application and the constraints specifically for a type of database, all expressed in the same database language. In a relational database, the schema defines the tables, fields, relationships, views, indexes, packages, procedures, functions, queues, triggers, types, sequences, materialized views, synonyms, database links, directories, XML schemas, and other elements. A database generally stores its schema in a data dictionary. Although a schema is defined in text database language, the term is often used to refer to a graphical depiction of the database structure. In other words, schema is the structure of the database that defines the objects in the database. In an Oracle Database system, the term "schema" has a slightly different connotation. Ideal requirements for schema integration The requirements listed below influence the detailed structure of schemas that are produced. Certain applications will not require that all of these conditions are met, but these four requirements are the most ideal. Overlap preservation Each of the overlapping elements specified in the input mapping is also in a database schema relation. Extended overlap preservation Source-specific elements that are associated with a source’s overlapping elements are passed through to the database schema. Normalization Independent entities and relationships in the source data should not be grouped together in the same relation in the database schema. In particular, source specific schema elements should not be grouped with overlapping schema elements, if the grouping co-locates independent entities or relationships. Minimality If any element
https://en.wikipedia.org/wiki/Telecommunication%20circuit
A telecommunication circuit is a path in a telecommunications network used to transmit information. Circuits have evolved from generally being built on physical connections between individual hardware cables, as in an analog phone switch, to virtual circuits established over packet switching networks. Definitions A telecommunication circuit may be defined as follows: The complete path between two terminals over which one-way or two-way communications may be provided. An electronic path between two or more points, capable of providing a single or multiple communication channels. An electronic closed-loop path among two or more points used for signal transfer. In operational terms, a telecommunication circuit may be capable of transmitting information in only one direction (simplex circuit), or it may be bi-directional (duplex circuit). Bi-directional circuits may support half-duplex operation, when only one end of the channel transmits at any one time, or they may support full-duplex operation, when independent simultaneous transmission occurs in both directions. Applications Originally, telecommunication circuits transmitted analog signals. Radio stations used them as studio transmitter links (STLs) or as remote pickup unit (RPU) for sound reproduction, sometimes as a backup to other means. Later lines were digital, used in pair-gain applications, such as carrier systems, or in enterprise data networks. A leased line, private circuit, or dedicated circuit, is a circuit that is dedicated to only one use and is typically not switched at a central office. The opposite is a switched circuit, which can be connected to different paths in a switching center or telephone exchange. Plain old telephone service (POTS) and ISDN telephone lines are switched circuits. On certain packet switching telecommunication circuits, a virtual circuit may be created, while sharing the physical circuit. References Further reading Telecommunications equipment
https://en.wikipedia.org/wiki/List%20of%20mathematical%20logic%20topics
This is a list of mathematical logic topics. For traditional syllogistic logic, see the list of topics in logic. See also the list of computability and complexity topics for more theory of algorithms. Working foundations Peano axioms Giuseppe Peano Mathematical induction Structural induction Recursive definition Naive set theory Element (mathematics) Ur-element Singleton (mathematics) Simple theorems in the algebra of sets Algebra of sets Power set Empty set Non-empty set Empty function Universe (mathematics) Axiomatization Axiomatic system Axiom schema Axiomatic method Formal system Mathematical proof Direct proof Reductio ad absurdum Proof by exhaustion Constructive proof Nonconstructive proof Tautology Consistency proof Arithmetization of analysis Foundations of mathematics Formal language Principia Mathematica Hilbert's program Impredicative Definable real number Algebraic logic Boolean algebra (logic) Dialectica space categorical logic Model theory Finite model theory Descriptive complexity theory Model checking Trakhtenbrot's theorem Computable model theory Tarski's exponential function problem Undecidable problem Institutional model theory Institution (computer science) Non-standard analysis Non-standard calculus Hyperinteger Hyperreal number Transfer principle Overspill Elementary Calculus: An Infinitesimal Approach Criticism of non-standard analysis Standard part function Set theory Forcing (mathematics) Boolean-valued model Kripke semantics General frame Predicate logic First-order logic Infinitary logic Many-sorted logic Higher-order logic Lindström quantifier Second-order logic Soundness theorem Gödel's completeness theorem Original proof of Gödel's completeness theorem Compactness theorem Löwenheim–Skolem theorem Skolem's paradox Gödel's incompleteness theorems Structure (mathematical logic) Interpretation (logic) Substructure (mathematics) Elementary substructure Skolem hull Non-standard model Atomic model (mathematical logic) Prime model Saturated model Existentially closed model Ultraproduct Age (model theory) Amalgamation property Hrushovski construction Potential isomorphism Theory (mathematical logic) Complete theory Vaught's test Morley's categoricity theorem Stability spectrum Morley rank Stable theory Forking extension Strongly minimal theory Stable group Tame group o-minimal theory Weakly o-minimal structure C-minimal theory Spectrum of a theory Vaught conjecture Model complete theory List of first-order theories Conservative extension Elementary class Pseudoelementary class Strength (mathematical logic) Differentially closed field Exponential field Ax–Grothendieck theorem Ax–Kochen theorem Peano axioms Non-standard model of arithmetic First-order arithmetic Second-order arithmetic Presburger arithmetic Wilkie's theorem Functional predicate T-schema Back-and-forth method Barwise compactness theorem Skolemization Lindenbaum–Tarski algebra Löb's theorem Arithmetical set Definable set Ehrenfeucht–Fraïssé game Herbrand int
https://en.wikipedia.org/wiki/TMC
TMC may stand for: Companies and brands Thinking Machines Corporation, a defunct supercomputer company Toyota Motor Corporation, a Japanese automobile manufacturer Trans Mountain Corporation Transportation Management Center, a division of American shipping company C.H. Robinson Transportation Manufacturing Corporation, a defunct bus manufacturer based in Roswell, New Mexico Triumph Motor Company, a defunct British automotive manufacturer Educational and medical institutions Tehran Monolingual Corpus, a Persian monolingual text corpus, Iran Texas Medical Center, a medical institution Thanjavur Medical College, a golden jubilee college in Tamil Nadu, India Thomson Medical Centre, a private hospital in Novena, Singapore Thurgood Marshall College, a college within University of California, San Diego Thursday Morning Club, a not-for-profit organization in Madison, New Jersey Tripura Medical College & Dr. B.R. Ambedkar Memorial Teaching Hospital, India Tucson Medical Center, a not-for-profit community hospital Thomas More College (South Africa), a leading private school in South Africa Washtenaw Technical Middle College Gaming The Legend of Zelda: The Minish Cap, a video game The Mud Connector, a Web portal site covering MUDs, text-based online role-playing games Locations Trece Martires, a city in Cavite, Philippines Television The Movie Channel, a cable television network Télé Monte Carlo, French language television channel in Monaco La7, Italian private television channel formerly known as Telemontecarlo Tagalized Movie Channel, a Philippine cable channel Politics Trinamool Congress, a state political party in West Bengal, India Tamil Maanila Congress, a political party in Tamil Nadu, India Transitional Military Council (disambiguation), interim governments in Chad and Sudan Other uses Taipei Music Center, performing arts and cultural center in Taipei, Taiwan The Mandela Catalogue, analog horror web series Thomas & Mack Center, an indoor arena on the campus of the University of Nevada, Las Vegas opened in 1983 Traffic Message Channel, a technology for delivering traffic and travel information to drivers Travel management company, managing corporate travel Transportation Materiel Command, a unit of the United States Army until 1962 Trapeziometacarpal joint, a joint in the thumb Tmcft (TMC, tmc) (thousand million cubic feet), a volume measurement of water Run TMC, a high-scoring trio of teammates in the National Basketball Association from 1989 to 1991
https://en.wikipedia.org/wiki/Mode%2013h
Mode 13h is the standard 256-color mode on VGA graphics hardware introduced in 1987 with the IBM PS/2. It has a resolution of 320×200 pixels. It was used extensively in computer games and art/animation software of the late 1980s and early to mid-1990s. "13h" refers to the number of the mode in the VGA BIOS. The "h" stands for hexadecimal. Mode 13h provides programmers with a linear 320x200 block of video memory, where each byte represents one pixel. This allows ease of programming at the expense of access to other useful features of the VGA hardware. Given the aspect ratio of a 320×200 resolution screen for use on a 4:3 display, Mode 13h does not have square pixels. Technical details Mode 13h is something of a curiosity, because the VGA is a planar device from a hardware perspective, and not suited to chunky graphics operation. The VGA has 256 KiB of video memory consisting of 4 banks of 64 KiB, known as planes (or 'maps' in IBM's documentation). Planar memory arrangement splits the pixels horizontally into groups of four. For any given byte in the PC's 64 KiB video memory aperture, four pixels can be accessed on screen by selecting the required plane(s). This is more complicated for the programmer, but allows access to all of the available video memory and other benefits (see Mode X). Mode 13h, however, allows the programmer to access the VGA in "chunky" (linear) fashion, where each consecutive address in the aperture represents a consecutive pixel on screen. Planes need not be manipulated to select the correct pixels to modify or read. This is achieved using the VGA 'Chain 4' setting in which the lowest two bits of the 16 bit aperture address are used to select the plane to write to/read from. The upshot is that the programmer has a simple access model; the downside is that now only 214 (16 KiB) of the video plane's address space are accessible (the other 2 bits being used to select planes) so only a total of 64 KiB is addressable (4 × 16 KiB). The upper 48 KiB of each plane remains unused in this mode. Unlike SVGA and higher color depths, Mode 13h, part of the MCGA set of video modes, has 18 bits of color, 6 per RGB channel. The color palette data is not stored in memory per se, but rather modified, read and written to by low level I/O port calls, to the DAC registers. The palette data was often copied from the low level ports one at a time to a spot in memory after the screen data so the colors could be saved along with the pixels they belonged to. A mode 13h screen capture is merely 64,000 bytes of values from 0 to 255, each an index of a color from a known palette of multiple colors. If the color palette is jumbled or totally wrong, a green leaf will appear blue, or any other possible color, depending on what that particular color was set for, which particular one out of 256. Due to the way chunks of memory are allocated in such an environment, the bytes in segment 0xA000 from offset 64000 to 64768, can be written to with no neg
https://en.wikipedia.org/wiki/Multiple%20buffering
In computer science, multiple buffering is the use of more than one buffer to hold a block of data, so that a "reader" will see a complete (though perhaps old) version of the data, rather than a partially updated version of the data being created by a "writer". It is very commonly used for computer display images. It is also used to avoid the need to use dual-ported RAM (DPRAM) when the readers and writers are different devices. Description An easy way to explain how multiple buffering works is to take a real-world example. It is a nice sunny day and you have decided to get the paddling pool out, only you can not find your garden hose. You'll have to fill the pool with buckets. So you fill one bucket (or buffer) from the tap, turn the tap off, walk over to the pool, pour the water in, walk back to the tap to repeat the exercise. This is analogous to single buffering. The tap has to be turned off while you "process" the bucket of water. Now consider how you would do it if you had two buckets. You would fill the first bucket and then swap the second in under the running tap. You then have the length of time it takes for the second bucket to fill in order to empty the first into the paddling pool. When you return you can simply swap the buckets so that the first is now filling again, during which time you can empty the second into the pool. This can be repeated until the pool is full. It is clear to see that this technique will fill the pool far faster as there is much less time spent waiting, doing nothing, while buckets fill. This is analogous to double buffering. The tap can be on all the time and does not have to wait while the processing is done. If you employed another person to carry a bucket to the pool while one is being filled and another emptied, then this would be analogous to triple buffering. If this step took long enough you could employ even more buckets, so that the tap is continuously running filling buckets. In computer science the situation of having a running tap that cannot be, or should not be, turned off is common (such as a stream of audio). Also, computers typically prefer to deal with chunks of data rather than streams. In such situations, double buffering is often employed. Double buffering Petri net The Petri net in the illustration shows how double buffering works. Transitions W1 and W2 represent writing to buffer 1 and 2 respectively while R1 and R2 represent reading from buffer 1 and 2 respectively. At the beginning only the transition W1 is enabled. After W1 fires, R1 and W2 are both enabled and can proceed in parallel. When they finish, R2 and W1 proceed in parallel and so on. So after the initial transient where W1 fires alone, this system is periodic and the transitions are enabled – always in pairs (R1 with W2 and R2 with W1 respectively). Double buffering in computer graphics In computer graphics, double buffering is a technique for drawing graphics that shows no (or less) stutter, tearing, and other
https://en.wikipedia.org/wiki/Point%20of%20presence
A point of presence (PoP) is an artificial demarcation point or network interface point between communicating entities. A common example is an ISP point of presence, the local access point that allows users to connect to the Internet with their Internet service provider (ISP). A PoP typically houses servers, routers, network switches, multiplexers, and other network interface equipment, and is typically located in a data center. ISPs typically have multiple PoPs. PoPs are often located at Internet exchange points and colocation centres. In the US, this term became important during the court-ordered breakup of the Bell Telephone system. A point of presence was a location where a long-distance carrier (IXC) could terminate services and provide connections into a local telephone network (LATA). See also Content delivery network Web cache Meet-me room References Telecommunications infrastructure
https://en.wikipedia.org/wiki/Image%20retrieval
An image retrieval system is a computer system used for browsing, searching and retrieving images from a large database of digital images. Most traditional and common methods of image retrieval utilize some method of adding metadata such as captioning, keywords, title or descriptions to the images so that retrieval can be performed over the annotation words. Manual image annotation is time-consuming, laborious and expensive; to address this, there has been a large amount of research done on automatic image annotation. Additionally, the increase in social web applications and the semantic web have inspired the development of several web-based image annotation tools. The first microcomputer-based image database retrieval system was developed at MIT, in the 1990s, by Banireddy Prasaad, Amar Gupta, Hoo-min Toong, and Stuart Madnick. A 2008 survey article documented progresses after 2007. All image retrieval systems as of 2021 were designed for 2D images, not 3D ones. Search methods Image search is a specialized data search used to find images. To search for images, a user may provide query terms such as keyword, image file/link, or click on some image, and the system will return images "similar" to the query. The similarity used for search criteria could be meta tags, color distribution in images, region/shape attributes, etc. Image meta search - search of images based on associated metadata such as keywords, text, etc. Content-based image retrieval (CBIR) – the application of computer vision to the image retrieval. CBIR aims at avoiding the use of textual descriptions and instead retrieves images based on similarities in their contents (textures, colors, shapes etc.) to a user-supplied query image or user-specified image features. List of CBIR Engines - list of engines which search for images based image visual content such as color, texture, shape/object, etc. Image collection exploration - search of images based on the use of novel exploration paradigms. Data scope It is crucial to understand the scope and nature of image data in order to determine the complexity of image search system design. The design is also largely influenced by factors such as the diversity of user-base and expected user traffic for a search system. Along this dimension, search data can be classified into the following categories: Archives - usually contain large volumes of structured or semi-structured homogeneous data pertaining to specific topics. Domain-Specific Collection - this is a homogeneous collection providing access to controlled users with very specific objectives. Examples of such a collection are biomedical and satellite image databases. Enterprise Collection - a heterogeneous collection of images that is accessible to users within an organization's intranet. Pictures may be stored in many different locations. Personal Collection - usually consists of a largely homogeneous collection and is generally small in size, accessible primarily to its
https://en.wikipedia.org/wiki/Reusability
In computer science and software engineering, reusability is the use of existing assets in some form within the software product development process; these assets are products and by-products of the software development life cycle and include code, software components, test suites, designs and documentation. The opposite concept of reusability is leverage, which modifies existing assets as needed to meet specific system requirements. Because reuse implies the creation of a , it is preferred over leverage. Subroutines or functions are the simplest form of reuse. A chunk of code is regularly organized using modules or namespaces into layers. Proponents claim that objects and software components offer a more advanced form of reusability, although it has been tough to objectively measure and define levels or scores of reusability. The ability to reuse relies in an essential way on the ability to build larger things from smaller parts, and being able to identify commonality among those parts. Reusability is often a required characteristic of platform software. Reusability brings several aspects to software development that do not need to be considered when reusability is not required. Reusability implies some explicit management of build, packaging, distribution, installation, configuration, deployment, maintenance and upgrade issues. If these issues are not considered, software may appear to be reusable from design point of view, but will not be reused in practice. Software reusability more specifically refers to design features of a software element (or collection of software elements) that enhance its suitability for reuse. Many reuse design principles were developed at the WISR workshops. Candidate design features for software reuse include: Adaptable Brief: small size Consistency Correctness Extensibility Fast Flexible Generic Localization of volatile (changeable) design assumptions (David Parnas) Modularity Orthogonality Parameterization Simple: low complexity Stability under changing requirements Consensus has not yet been reached on this list on the relative importance of the entries nor on the issues which make each one important for a particular class of applications. See also Code reuse References Source code Software quality
https://en.wikipedia.org/wiki/Multimethodology
Multimethodology or multimethod research includes the use of more than one method of data collection or research in a research study or set of related studies. Mixed methods research is more specific in that it includes the mixing of qualitative and quantitative data, methods, methodologies, and/or paradigms in a research study or set of related studies. One could argue that mixed methods research is a special case of multimethod research. Another applicable, but less often used label, for multi or mixed research is methodological pluralism. All of these approaches to professional and academic research emphasize that monomethod research can be improved through the use of multiple data sources, methods, research methodologies, perspectives, standpoints, and paradigms. The term multimethodology was used starting in the 1980s and in the 1989 book Multimethod Research: A Synthesis of Styles by John Brewer and Albert Hunter. During the 1990s and currently, the term mixed methods research has become more popular for this research movement in the behavioral, social, business, and health sciences. This pluralistic research approach has been gaining in popularity since the 1980s. Multi and mixed methods research designs There are four broad classes of research studies that are currently being labeled "mixed methods research": Quantitatively driven approaches/designs in which the research study is, at its core, a quantitative study with qualitative data/method added to supplement and improve the quantitative study by providing an added value and deeper, wider, and fuller or more complex answers to research questions; quantitative quality criteria are emphasized but high quality qualitative data also must be collected and analyzed; Qualitatively driven approaches/designs in which the research study is, at its core, a qualitative study with quantitative data/method added to supplement and improve the qualitative study by providing an added value and deeper, wider, and fuller or more complex answers to research questions; qualitative quality criteria are emphasized but high quality quantitative data also must be collected and analyzed; Interactive or equal status designs in which the research study equally emphasizes (interactively and through integration) quantitative and qualitative data, methods, methodologies, and paradigms. This third design is often done through the use of a team composed of an expert in quantitative research, an expert in qualitative research, and an expert in mixed methods research to help with dialogue and continual integration. In this type of mixed study, quantitative and qualitative and mixed methods quality criteria are emphasized. This use of multiple quality criteria is seen in the concept of multiple validities legitimation. Here is a definition of this important type of validity or legitimation: Multiple validities legitimation "refers to the extent to which the mixed methods researcher successfully addresses and resolve
https://en.wikipedia.org/wiki/Tangerine%20Computer%20Systems
Tangerine Computer Systems was a British microcomputer company founded in 1979 by Dr. Paul Johnson, Mark Rainer and Nigel Penton Tilbury in St. Ives, Cambridgeshire. The very first product was the successful TAN1648 VDU kit which received much acclaim in the technical press. The home computer market was beginning to move, albeit slowly, and it was essential to establish a presence. Development and expansion was imperative. It was decided that the latter two partners would relinquish their involvement in order to focus on their consultancy work. Barry Muncaster became involved operationally and the company moved to new premises in Ely, Cambridgeshire. The company was later renamed, and was known as Oric Products International. Early years: Microtan 65 Tangerine produced one of the first 6502-based kit computers, the Microtan 65. It had a 3U form factor, a small amount of memory (RAM), a video character generator and UHF modulator for use with a TV set, and a simple latch for entering hex data from a keypad, and the computer was designed to be expandable. The manual came with a one-kilobyte listing of Conway's Game of Life. An optional expansion board could be built with a UART, more memory and BASIC ROMs. Additional expansion boards became available later, offering more RAM, dedicated serial and parallel I/O boards, etc. After the Microtan 65, Tangerine planned to build a desktop machine and managed to get as far as selling the design for the Microtan 2 also known as Tangerine Tiger to HH Electronics, better known for building amplifiers. They released it as the HH Tiger, but it was not a commercial success. Tantel Several Prestel machines were sold, under the general designation of Tantel: AlphaTantel (1981) – Prestel adapter with a full keyboard and a phone connection jack. Connection to a TV was through a RF modulator, but there was also a RGB output. Built in modem was 1200 baud. It was possible to connect a tape recorder to save data and a printer. New Tantel Adaptor (1981) – Prestel adapter Tantel Data Adaptor (1984) – Data adaptor for use with the Prestel system, including a data tape recorder 1983 onwards: The Oric family With the success of the ZX Spectrum Tangerine's backers suggested a home computer and Tangerine formed Oric Products International Ltd to develop and release the Oric-1 in 1983. A series of Oric computers (including the Atmos) followed through to 1987. On 13 October 1983 the factory of Kenure Plastics in Berkshire, where the Oric-1 was manufactured, burnt to the ground. The factory was rebuilt, minus a considerable stock of bits (including 15,000 old ROMs) that went to make up the Oric-1. In the meantime production was said to have restarted within 24 hours in a new factory. Just a day later, a neighbouring warehouse went up in flames. Police were said at the time to suspect that the arsonist got the wrong place first time round. It was about this time, too, that Tansoft moved to co-exist with Oric Research at
https://en.wikipedia.org/wiki/Chown
The command , an abbreviation of change owner, is used on Unix and Unix-like operating systems to change the owner of file system files, directories. Unprivileged (regular) users who wish to change the group membership of a file that they own may use . The ownership of any file in the system may only be altered by a super-user. A user cannot give away ownership of a file, even when the user owns it. Similarly, only a member of a group can change a file's group ID to that group. The command is available as a separate package for Microsoft Windows as part of the UnxUtils collection of native Win32 ports of common GNU Unix-like utilities. The command has also been ported to the IBM i operating system. See also chgrp chmod takeown References External links chown manual page The chown Command by The Linux Information Project (LINFO) Operating system security Standard Unix programs Unix SUS2008 utilities IBM i Qshell commands
https://en.wikipedia.org/wiki/Association%20%28object-oriented%20programming%29
In object-oriented programming, association defines a relationship between classes of objects that allows one object instance to cause another to perform an action on its behalf. This relationship is structural, because it specifies that objects of one kind are connected to objects of another and does not represent behaviour. In generic terms, the causation is usually called "sending a message", "invoking a method" or "calling a member function" to the controlled object. Concrete implementation usually requires the requesting object to invoke a method or member function using a reference or pointer to the memory location of the controlled object. The objects that are related via the association are considered to act in a role with respect to the association, if object's current state in the active situation allows the other associated objects to use the object in the manner specified by the role. A role can be used to distinguish two objects of the same class when describing its use in the context of the association. A role describes the public aspects of an object with respect to an association. The ends of the association can have all the characteristics of a property: They can have a multiplicity, expressed by a lower and an upper limit in the form of "lowerLimit..upperLimit". You can have a name. You can declare a visibility. You can specify whether the end of the association is ordered and / or unique. See also Aggregation Object composition References Object-oriented programming
https://en.wikipedia.org/wiki/All%20Things%20Considered
All Things Considered (ATC) is the flagship news program on the American network National Public Radio (NPR). It was the first news program on NPR, premiering on May 3, 1971. It is broadcast live on NPR affiliated stations in the United States, and worldwide through several different outlets, formerly including the NPR Berlin station in Germany. All Things Considered and Morning Edition were the highest rated public radio programs in the United States in 2002 and 2005. The show combines news, analysis, commentary, interviews, and special features, and its segments vary in length and style. ATC airs weekdays from 4:00 p.m. to 6:00 p.m. Eastern Time (live) or Pacific Time (recorded with some updates; in Hawaii it airs as a fully recorded program) or from 3:00 p.m. to 5:00 p.m. Central Time. A weekend version of ATC, Weekend All Things Considered, airs on Saturdays and Sundays. Background ATC programming combines news, analysis, commentary, interviews, and special features broadcast live daily from 4:00 p.m. to 6:00 p.m. Eastern Time (3 to 5 p.m. Central Time) (20:00 to 22:00 UTC), and is re-fed with updates until 10 p.m. ET (9 p.m. CT) or 7 p.m. PT (02:00 UTC). Broadcasts run about 105 minutes with local content interspersed in between to complete two hours. In 2005, ATC aired on over 560 radio stations and reached an audience of approximately 12 million listeners each weekday, making it the third most listened to radio program in the United States after The Rush Limbaugh Show and Morning Edition. In September 2010, All Things Considered had an average quarter-hour audience of 1.8 million. ATC is co-hosted by rotating cast of regular anchors; current hosts include Ari Shapiro, Ailsa Chang, Mary Louise Kelly, and Juana Summers. The first broadcast of ATC was fed to about 90 radio stations on May 3, 1971, with host Robert Conley. During the first week, these stations were not allowed to broadcast the feed "live" but could record it for later broadcast. The first story was about the march on Washington, D.C., and the growing anti–Vietnam War protests taking place there. NPR chose to place its inaugural daily newscast at the afternoon commute timeslot instead of the morning because many of its affiliates at that time did not sign on for the day until mid-morning or afterward. It was not until 1979, by which time most affiliates had expanded their broadcast days to begin at 6 a.m. or earlier, that NPR premiered Morning Edition. Weekend All Things Considered (WATC) is a one-hour version of the show that premiered in 1974 and is broadcast on Saturdays and Sundays at 5 p.m. ET, currently hosted by Scott Detrow. ATC was excluded from the NPR deal with Sirius Satellite Radio so as not to compete with local stations airing the show. To coordinate the choice of interview partners in cultural coverage between ATC and other NPR shows (as of 2010: Morning Edition, the weekend editions, Talk of the Nation, and Tell Me More), NPR set up a "dibs list" system ar
https://en.wikipedia.org/wiki/Conceptual%20graph
A conceptual graph (CG) is a formalism for knowledge representation. In the first published paper on CGs, John F. Sowa used them to represent the conceptual schemas used in database systems. The first book on CGs applied them to a wide range of topics in artificial intelligence, computer science, and cognitive science. Research branches Since 1984, the model has been developed along three main directions: a graphical interface for first-order logic, a diagrammatic calculus of logics, and a graph-based knowledge representation and reasoning model. Graphical interface for first-order logic In this approach, a formula in first-order logic (predicate calculus) is represented by a labeled graph. A linear notation, called the Conceptual Graph Interchange Format (CGIF), has been standardized in the ISO standard for common logic. The diagram above is an example of the display form for a conceptual graph. Each box is called a concept node, and each oval is called a relation node. In CGIF, this CG would be represented by the following statement: In CGIF, brackets enclose the information inside the concept nodes, and parentheses enclose the information inside the relation nodes. The letters x and y, which are called coreference labels, show how the concept and relation nodes are connected. In CLIF, those letters are mapped to variables, as in the following statement: As this example shows, the asterisks on the coreference labels and in CGIF map to existentially quantified variables in CLIF, and the question marks on and map to bound variables in CLIF. A universal quantifier, represented in CGIF, would be represented in CLIF. Reasoning can be done by translating graphs into logical formulas, then applying a logical inference engine. Diagrammatic calculus of logics Another research branch continues the work on existential graphs of Charles Sanders Peirce, which were one of the origins of conceptual graphs as proposed by Sowa. In this approach, developed in particular by Dau , conceptual graphs are conceptual diagrams rather than graphs in the sense of graph theory, and reasoning operations are performed by operations on these diagrams. Graph-based knowledge representation and reasoning model Key features of GBKR, the graph-based knowledge representation and reasoning model developed by Chein and Mugnier and the Montpellier group , can be summarized as follows: All kinds of knowledge (ontology, rules, constraints and facts) are labeled graphs, which provide an intuitive and easily understandable means to represent knowledge. Reasoning mechanisms are based on graph notions, basically the classical notion of graph homomorphism; this allows, in particular, to link basic reasoning problems to other fundamental problems in computer science (e.g., problems concerning conjunctive queries in relational databases, or constraint satisfaction problems). The formalism is logically founded, i.e., it has a semantics in first-order logic and the in
https://en.wikipedia.org/wiki/Modeling%20language
A modeling language is any artificial language that can be used to express data, information or knowledge or systems in a structure that is defined by a consistent set of rules. The rules are used for interpretation of the meaning of components in the structure Programing language. Overview A modeling language can be graphical or textual. Graphical modeling languages use a diagram technique with named symbols that represent concepts and lines that connect the symbols and represent relationships and various other graphical notation to represent constraints. Textual modeling languages may use standardized keywords accompanied by parameters or natural language terms and phrases to make computer-interpretable expressions. An example of a graphical modeling language and a corresponding textual modeling language is EXPRESS. Not all modeling languages are executable, and for those that are, the use of them doesn't necessarily mean that programmers are no longer required. On the contrary, executable modeling languages are intended to amplify the productivity of skilled programmers, so that they can address more challenging problems, such as parallel computing and distributed systems. A large number of modeling languages appear in the literature. Type of modeling languages Graphical types Example of graphical modeling languages in the field of computer science, project management and systems engineering: Behavior Trees are a formal, graphical modeling language used primarily in systems and software engineering. Commonly used to unambiguously represent the hundreds or even thousands of natural language requirements that are typically used to express the stakeholder needs for a large-scale software-integrated system. Business Process Modeling Notation (BPMN, and the XML form BPML) is an example of a Process Modeling language. C-K theory consists of a modeling language for design processes. DRAKON is a general-purpose algorithmic modeling language for specifying software-intensive systems, a schematic representation of an algorithm or a stepwise process, and a family of programming languages. EXPRESS and EXPRESS-G (ISO 10303-11) is an international standard general-purpose data modeling language. Extended Enterprise Modeling Language (EEML) is commonly used for business process modeling across a number of layers. Flowchart is a schematic representation of an algorithm or a stepwise process. Fundamental Modeling Concepts (FMC) modeling language for software-intensive systems. IDEF is a family of modeling languages, which include IDEF0 for functional modeling, IDEF1X for information modeling, IDEF3 for business process modeling, IDEF4 for Object-Oriented Design and IDEF5 for modeling ontologies. Jackson Structured Programming (JSP) is a method for structured programming based on correspondences between data stream structure and program structure. LePUS3 is an object-oriented visual Design Description Language and a formal specification la
https://en.wikipedia.org/wiki/IEEE%20802.1X
IEEE 802.1X is an IEEE Standard for port-based network access control (PNAC). It is part of the IEEE 802.1 group of networking protocols. It provides an authentication mechanism to devices wishing to attach to a LAN or WLAN. IEEE 802.1X defines the encapsulation of the Extensible Authentication Protocol (EAP) over wired IEEE 802 networks and over 802.11 wireless networks, which is known as "EAP over LAN" or EAPOL. EAPOL was originally specified for IEEE 802.3 Ethernet, IEEE 802.5 Token Ring, and FDDI (ANSI X3T9.5/X3T12 and ISO 9314) in 802.1X-2001, but was extended to suit other IEEE 802 LAN technologies such as IEEE 802.11 wireless in 802.1X-2004. The EAPOL was also modified for use with IEEE 802.1AE ("MACsec") and IEEE 802.1AR (Secure Device Identity, DevID) in 802.1X-2010 to support service identification and optional point to point encryption over the internal LAN segment. Overview 802.1X authentication involves three parties: a supplicant, an authenticator, and an authentication server. The supplicant is a client device (such as a laptop) that wishes to attach to the LAN/WLAN. The term 'supplicant' is also used interchangeably to refer to the software running on the client that provides credentials to the authenticator. The authenticator is a network device that provides a data link between the client and the network and can allow or block network traffic between the two, such as an Ethernet switch or wireless access point; and the authentication server is typically a trusted server that can receive and respond to requests for network access, and can tell the authenticator if the connection is to be allowed, and various settings that should apply to that client's connection or setting. Authentication servers typically run software supporting the RADIUS and EAP protocols. In some cases, the authentication server software may be running on the authenticator hardware. The authenticator acts like a security guard to a protected network. The supplicant (i.e., client device) is not allowed access through the authenticator to the protected side of the network until the supplicant's identity has been validated and authorized. With 802.1X port-based authentication, the supplicant must initially provide the required credentials to the authenticator - these will have been specified in advance by the network administrator and could include a user name/password or a permitted digital certificate. The authenticator forwards these credentials to the authentication server to decide whether access is to be granted. If the authentication server determines the credentials are valid, it informs the authenticator, which in turn allows the supplicant (client device) to access resources located on the protected side of the network. Protocol operation EAPOL operates over the data link layer, and in Ethernet II framing protocol has an EtherType value of 0x888E. Port entities 802.1X-2001 defines two logical port entities for an authenticated port—the "controll
https://en.wikipedia.org/wiki/Berkeley%20Open%20Infrastructure%20for%20Network%20Computing
The Berkeley Open Infrastructure for Network Computing (BOINC, pronounced – rhymes with "oink") is an open-source middleware system for volunteer computing (a type of distributed computing). Developed originally to support SETI@home, it became the platform for many other applications in areas as diverse as medicine, molecular biology, mathematics, linguistics, climatology, environmental science, and astrophysics, among others. The purpose of BOINC is to enable researchers to utilize processing resources of personal computers and other devices around the world. BOINC development began with a group based at the Space Sciences Laboratory (SSL) at the University of California, Berkeley, and led by David P. Anderson, who also led SETI@home. As a high-performance volunteer computing platform, BOINC brings together 34,236 active participants employing 136,341 active computers (hosts) worldwide, processing daily on average 20.164 PetaFLOPS (it would be the 21st largest processing capability in the world compared with an individual supercomputer). The National Science Foundation (NSF) funds BOINC through awards SCI/0221529, SCI/0438443 and SCI/0721124. Guinness World Records ranks BOINC as the largest computing grid in the world. BOINC code runs on various operating systems, including Microsoft Windows, macOS, Android, Linux, and FreeBSD. BOINC is free software released under the terms of the GNU Lesser General Public License (LGPL). History BOINC was originally developed to manage the SETI@home project. The original SETI client was a non-BOINC software exclusively for SETI@home. It was one of the first volunteer computing projects, and not designed with a high level of security. As a result, some participants in the project attempted to cheat the project to gain "credits", while others submitted entirely falsified work. BOINC was designed, in part, to combat these security breaches. The BOINC project started in February 2002, and its first version was released on April 10, 2002. The first BOINC-based project was Predictor@home, launched on June 9, 2004. In 2009, AQUA@home deployed multi-threaded CPU applications for the first time, followed by the first OpenCL application in 2010. As of 15 August 2022, there are 33 projects on the official list. There are also, however, BOINC projects not included on the official list. Each year, an international BOINC Workshop is hosted to increase collaboration among project administrators. In 2021, the workshop was hosted virtually. While not affiliated with BOINC officially, there have been several independent projects that reward BOINC users for their participation, including Charity Engine (sweepstakes based on processing power with prizes funded by private entities who purchase computational time of CE users), Bitcoin Utopia (now defunct), and Gridcoin (a blockchain which mints coins based on processing power). Design and structure BOINC is software that can exploit the unused CPU and GPU cycles on com
https://en.wikipedia.org/wiki/SuperCollider
SuperCollider is an environment and programming language originally released in 1996 by James McCartney for real-time audio synthesis and algorithmic composition. Since then it has been evolving into a system used and further developed by both scientists and artists working with sound. It is a dynamic programming language providing a framework for acoustic research, algorithmic music, interactive programming and live coding. Originally released under the terms of the GPL-2.0-or-later in 2002, and from version 3.4 under GPL-3.0-or-later, SuperCollider is free and open-source software. Architecture Starting with version 3, the SuperCollider environment has been split into two components: a server, scsynth; and a client, sclang. These components communicate using OSC (Open Sound Control). The SC language combines the object-oriented structure of Smalltalk and features from functional programming languages with a C-family syntax. The SC Server application supports simple C and C++ plugin APIs, making it easy to write efficient sound algorithms (unit generators), which can then be combined into graphs of calculations. Because all external control in the server happens via OSC, it is possible to use it with other languages or applications. The SuperCollider synthesis server (scsynth) SuperCollider's sound generation is bundled into an optimised command-line executable (named scsynth). In most cases it is controlled from within the SuperCollider programming language, but it can be used independently. The audio server has the following features: Open Sound Control access Simple ANSI C and C++11 plugin APIs Supports any number of input and output channels, including massively multichannel setups Gives access to an ordered tree structure of synthesis nodes which define the order of execution Bus system which allows dynamically restructuring the signal flow Buffers for writing and reading Calculation at different rates depending on the needs: audio rate, control rate, demand rate Supernova, an independent implementation of the Server architecture, adds multi-processor support through explicit parallel grouping of synthesis nodes. The SuperCollider programming language (sclang) The SuperCollider programming language is a dynamically typed, garbage-collected, single-inheritance object-oriented and functional language similar to Smalltalk, with a syntax similar to Lisp or the C programming language. Its architecture strikes a balance between the needs of realtime computation and the flexibility and simplicity of an abstract language. Like many functional languages, it implements functions as first-class objects, which may be composed. Functions and methods can have default argument values and variable length argument lists and can be called with any order of keyword arguments. Closures are lexical, and scope is both lexical and dynamic. Further features typical of functional languages are supported, including creation of closures via partial
https://en.wikipedia.org/wiki/Browser%20exploit
A browser exploit is a form of malicious code that takes advantage of a flaw or vulnerability in an operating system or piece of software with the intent to breach browser security to alter a user's browser settings without their knowledge. Malicious code may exploit ActiveX, HTML, images, Java, JavaScript, and other Web technologies and cause the browser to run arbitrary code. Symptoms Users whose web browsers have fallen victim of a successful browser exploit may find their homepage, search page, and/or favorites have been changed. Other signs include Internet settings options within the browser being altered, access being blocked to specific functions, and the redirection of incorrectly typed URL prefixes. Prevention There are multiple ways users can protect their web browsers from falling victim to a browser exploit. Such things include installing firewall software, keeping software updated, being cautious when downloading files, and not opening email attachments from unknown sources. Notable browser exploits JailbreakMe is a series of browser based exploits used to jailbreak Apple's iOS mobile operating system. It uses an exploit in the browser's PDF parser to execute unauthorised code and gain access to the underlying operating system. See also Browser security Internet security References External links Online Threats - Browser exploits | WOT Web of Trust Web security exploits Web browsers
https://en.wikipedia.org/wiki/RELAX%20NG
In computing, RELAX NG (REgular LAnguage for XML Next Generation) is a schema language for XML—a RELAX NG schema specifies a pattern for the structure and content of an XML document. A RELAX NG schema is itself an XML document but RELAX NG also offers a popular compact, non-XML syntax. Compared to other XML schema languages RELAX NG is considered relatively simple. It was defined by a committee specification of the OASIS RELAX NG technical committee in 2001 and 2002, based on Murata Makoto's RELAX and James Clark's TREX, and also by part two of the international standard ISO/IEC 19757: Document Schema Definition Languages (DSDL). ISO/IEC 19757-2 was developed by ISO/IEC JTC 1/SC 34 and published in its first version in 2003. Schema examples Suppose we want to define an extremely simple XML markup scheme for a book: a book is defined as a sequence of one or more pages; each page contains text only. A sample XML document instance might be: <book> <page>This is page one.</page> <page>This is page two.</page> </book> XML syntax A RELAX NG schema can be written in a nested structure by defining a root element that contains further element definitions, which may themselves contain embedded definitions. A schema for our book in this style, using the full XML syntax, would be written: <element name="book" xmlns="http://relaxng.org/ns/structure/1.0"> <oneOrMore> <element name="page"> <text/> </element> </oneOrMore> </element> Nested structure becomes unwieldy with many sublevels and cannot define recursive elements, so most complex RELAX NG schemas use references to named pattern definitions located separately in the schema. Here, a "flattened schema" defines precisely the same book markup as the previous example: <grammar xmlns="http://relaxng.org/ns/structure/1.0"> <start> <element name="book"> <oneOrMore> <ref name="page"/> </oneOrMore> </element> </start> <define name="page"> <element name="page"> <text/> </element> </define> </grammar> Compact syntax RELAX NG compact syntax is a non-XML format inspired by extended Backus–Naur form and regular expressions, designed so that it can be unambiguously translated to its XML counterpart, and back again, with one-to-one correspondence in structure and meaning, in much the same way that Simple Outline XML (SOX) relates to XML. It shares many features with the syntax of DTDs. Here is the compact form of the above schema: element book { element page { text }+ } With named patterns, this can be flattened to: start = element book { page+ } page = element page { text } A compact RELAX NG parser will treat these two as the same pattern. Comparison with W3C XML Schema Although the RELAX NG specification was developed at roughly the same time as the W3C XML Schema specification, the latter was arguably better known and more widely implemented in both open-source and proprietary XML parsers and e
https://en.wikipedia.org/wiki/ZZT
ZZT is a 1991 action-adventure puzzle video game and game creation system developed and published by Potomac Computer Systems for MS-DOS. It was later released as freeware in 1997. It is an early game allowing user-generated content using object-oriented programming. Players control a smiley face to battle various creatures and solve puzzles in different grid-based boards in a chosen world. It has four worlds where players explore different boards and interact with objects such as ammo, bombs, and scrolls to reach the end of the game. It includes an in-game editor, allowing players to develop worlds using the game's scripting language, ZZT-OOP. The game was designed by mechanical engineering student Tim Sweeney in roughly six to nine months. It was built from a text editor conceived in 1989 to build a better editor for Pascal, after he disliked editors that came with other programming languages. During development, he experimented with adding creatures and characters. He built boards that grew into worlds and refined the editor he used to create his own gameswhile studying at university. Initially, he made the game for himself, but after positive reception from his friends and neighbors, and seeing the potential for making a profit by releasing the game under shareware, he released it publicly. He marketed the game by distributing it across shareware vendors and bulletin board systems, earning money through mail orders for registered worlds. ZZT was a commercial success, with around 4,000–5,000 copies by 2009. The game received mixed reception. Much of the positive reception focused on the gameplay, editor, and the community it developed. Criticisms focused on the game's graphical and audio limitations and perceived unfair difficulty. The sequel is Super ZZT (1991). Other ZZT worlds were published later as Best of ZZT (1992) and ZZT's Revenge (1992). ZZTs success led Sweeney to change his company's name to Epic MegaGames, and focus on competing as a video game company using shareware to distribute commercial games. Epic MegaGames later developed other successful games including Jill of the Jungle (1992) and Unreal (1998), using lessons from ZZTs success by focusing on developing the editor and engine to allow others to more easily make games. The game inspired one of the earliest active modding communities, which has grown by making new worlds, editing tools, and source ports, and inspiring some in the community to pursue a career in the video game industry. Gameplay ZZT is a top-down action-adventure puzzle video game and game creation system. Players control a white smiley face on a navy blue rectangle that can move around in four directions. Players can interact with objects by touching or shooting at them. Touching ammo containers, gems, torches, and keys adds them to your status bar. Boxes of ammo allow players to fire bullets at objects, gems increase health and are used as currency, torches light up a small area around the player in da
https://en.wikipedia.org/wiki/Restart
Restart may refer to: Computing Reboot (computing), the act of restarting a computer Reset (computing), bringing the system to normal condition or an initial state Music Restart (band), a Brazilian band Restart (Restart album), Restart's debut studio album Restart (Newsboys album), 2013 Restart (Bilal song), 2010 Restart (Nnadia Chan album), 2003 "Restart" (Kotoko song), 2012 The Restarts, a street punk band from London, England Other Restart (training course), a program in the United Kingdom for people who were long-term unemployed Re:START, a temporary shopping mall in Christchurch, New Zealand Restart, a novel by Gordon Korman. Restart (group), Iranian opposition group based in California, U.S. See also Booting Reboot (disambiguation) Reset (disambiguation) Restarter (disambiguation) Start (disambiguation)
https://en.wikipedia.org/wiki/Removable%20media
In computing, a removable media is a data storage media that is designed to be readily inserted and removed from a system. Most early removable media, such as floppy disks and optical discs, require a dedicated read/write device (i.e. a drive) to be installed in the computer, while others, such as USB flash drives, are plug-and-play with all the hardware required to read them built into the device, so only need a driver software to be installed in order to communicate with the device. Some removable media readers/drives are integrated into the computer case, while others are standalone devices that need to be additionally installed or connected. Examples of removable media that require a dedicated reader drive include: Optical discs, e.g. Blu-rays (both standard and UHD versions), DVDs, CDs Flash memory-based memory cards, e.g. CompactFlash, Secure Digital, Memory Stick Magnetic storage media Floppy and Zip disks (now obsolete) Disk packs (now obsolete) Magnetic tapes (now obsolete) Paper data storage, e.g. punched cards, punched tapes (now obsolete) Examples of removable media that are standalone plug-and-play devices that carry their own reader hardwares include: USB flash drives Portable storage devices Dedicated external solid state drives (SSD) Enclosured mass storage drives, i.e. modified hard disk drives (HDD)/internal SSDs Peripheral devices that have integrated data storage capability Digital cameras Mobile devices such as smartphones, tablets and handheld game consoles Portable media players Other external or dockable peripherals that have expandable removable media capabilities, usually via a USB port or memory card reader USB hubs Wired or wireless printers Network routers, access points and switches Using removable media can pose some computer security risks, including viruses, data theft and the introduction of malware. History The earliest form of removable media, punched cards and tapes, predates the electronic computer by centuries, with the Jacquard loom of 1801 using interlinked cards to control the machine. This followed a loom made by Basile Bouchon in 1725 that used paper tape for its instructions. Punched tape was later used in Colossus, the first electronic computer. Magnetic tape was developed in the early 20th century in Germany, based on magnetic wire recording invented by Valdemar Poulsen in 1898. In 1951, the UNIVAC I was released, using magnetic tape to store data. 8-inch floppy disks were commercially introduced in 1971 by IBM, with them being compacted to -inch by Shugart Associates in 1976. At the same time Compact Cassettes started being used to store data, being popular in the late 1970s and 1980s for holding data for personal computers. In 1982, the -inch floppy disk became commonplace, with its introduction for the Apple Macintosh and Amiga. The CD-ROM was introduced in 1985, providing much higher capacity than a floppy disk, however could not be written to. This was resolved in 1990
https://en.wikipedia.org/wiki/Oberon%20%28operating%20system%29
The Oberon System is a modular, single-user, single-process, multitasking operating system written in the programming language Oberon. It was originally developed in the late 1980s at ETH Zurich. The Oberon System has an unconventional visual text user interface (TUI) instead of a conventional command-line interface (CLI) or graphical user interface (GUI). This TUI was very innovative in its time and influenced the design of the Acme text editor for the Plan 9 from Bell Labs operating system. The latest version of the Oberon System, Project Oberon 2013, is still maintained by Niklaus Wirth and several collaborators, but older ETH versions of the system have been orphaned. The system also evolved into the multi-process, symmetric multiprocessing (SMP) capable A2 (formerly Active Object System (AOS), then Bluebottle), with a zooming user interface (ZUI). History The Oberon operating system was originally developed as part of the NS32032-based Ceres workstation project. It was written almost entirely (and in the 2013 version entirely is valid) in the Oberon programming language. The basic system was designed and implemented by Niklaus Wirth and Jürg Gutknecht and its design and implementation is fully documented in their book "Project Oberon". The user Interface and programmers reference is found in Martin Reiser's book "The Oberon System". The Oberon System was later extended and ported to other hardware platforms by a team at ETH Zurich and there was recognition in popular magazines. Wirth and Gutknecht (although being active computer science professors) refer to themselves as 'part-time programmers' in the book Project Oberon. In late 2013, a few months before his 80th birthday, Wirth published a second edition of Project Oberon. It details implementing the Oberon System using a reduced instruction set computer (RISC) CPU of his own design realized on a Xilinx field-programmable gate array (FPGA) board. It was presented at the symposium organized for his 80th birthday at ETH Zurich. In the meantime, several emulators for this version were implemented. According to Josef Templ, a former member of the developer group at Swiss Federal Institute of Technology in Zurich and later member of the Institut für Systemsoftware of Johannes Kepler University Linz, where one forked version (V4) was maintained, the genealogy of the different versions of the Oberon System is this: User interface Oberon has a text user interface (TUI), which is very different from a terminal user interface. It combines the point and click convenience of a graphical user interface (GUI) with the linguistic strength of a command-line interface (CLI) and is closely tied to the naming conventions of the Oberon language. Text appearing almost anywhere on a screen can be edited and used as command input. Commands are activated by a middle-mouse click on a text fragment of the form Module.Command (optionally followed by parameters, which are terminated by ~). A command is defi
https://en.wikipedia.org/wiki/List%20of%20TCP%20and%20UDP%20port%20numbers
This is a list of TCP and UDP port numbers used by protocols for operation of network applications. The Transmission Control Protocol (TCP) and the User Datagram Protocol (UDP) only need one port for duplex, bidirectional traffic. They usually use port numbers that match the services of the corresponding TCP or UDP implementation, if they exist. The Internet Assigned Numbers Authority (IANA) is responsible for maintaining the official assignments of port numbers for specific uses. However, many unofficial uses of both well-known and registered port numbers occur in practice. Similarly, many of the official assignments refer to protocols that were never or are no longer in common use. This article lists port numbers and their associated protocols that have experienced significant uptake. Table legend Well-known ports The port numbers in the range from 0 to 1023 (0 to 210 − 1) are the well-known ports or system ports. They are used by system processes that provide widely used types of network services. On Unix-like operating systems, a process must execute with superuser privileges to be able to bind a network socket to an IP address using one of the well-known ports. Registered ports The range of port numbers from 1024 to 49151 (210 to 214 + 215 − 1) are the registered ports. They are assigned by IANA for specific service upon application by a requesting entity. On most systems, registered ports can be used without superuser privileges. Dynamic, private or ephemeral ports The range 49152–65535 (215 + 214 to 216 − 1) contains dynamic or private ports that cannot be registered with IANA. This range is used for private or customized services, for temporary purposes, and for automatic allocation of ephemeral ports. Note See also Port (computer networking) Internet protocol suite List of IP numbers Lists of network protocols Comparison of file transfer protocols References and notes Further reading External links Computing-related lists Internet-related lists Lists of network protocols Transmission Control Protocol
https://en.wikipedia.org/wiki/Datapoint
Datapoint Corporation, originally known as Computer Terminal Corporation (CTC), was a computer company based in San Antonio, Texas, United States. Founded in July 1968 by Phil Ray and Gus Roche, its first products were, as the company's initial name suggests, computer terminals intended to replace Teletype machines connected to time sharing systems. Background From the mid-1960s, engineers (John) Phil Ray and Austin O. "Gus" Roche were working for General Dynamics Dynatronic Division in Florida, as part of a computing contract team for NASA to enable President John F. Kennedy's vision of putting a man on the moon. At the time, mainframe computers were large room-filling pieces of equipment, for which data was input using dumb and noisy Teletype terminals. On the advice of one of his tutors from the University of Texas, Ray and Roche decided to develop a quieter and smaller input device based on using a television set screen. Early history After finding San Antonio, Texas based backers, the pair incorporated Computer Terminal Corporation (CTC) there in July 1968. Based on a newly developed Texas Instruments chip, with a casing designed by an industrial designer in New York City, the company had developed three prototype Datapoint 3300 (deriving the name from the existing competitor Teletype Model 33), it was an immediate hit at that year's computer show. However, the resultant orders left the company with several major problems: they had no production capacity, and no money with which to build a factory. Furthermore, the TI chips within the 3300 were so unreliable, they needed replacing every 30 minutes. In October 1969, the company raised US$4 million through an initial public offering (IPO). This enabled them to outsource production short term to a series of local manufacturers, including a motorcycle helmet manufacturer, before constructing their own manufacturing premises. Once in production, they began sourcing chips from both TI and Intel, running up large debts from both. The Datapoint 3300 sold very successfully for a number of years. It was later rebadged as the DEC VT06 and the HP 2600A. Datapoint 2200 and the Intel 8008 Having raised $4 million in an IPO in August 1969, selling over the counter for $8, by August 1970 the shares were selling at $45. Ray and Roche wanted to develop a new, more intelligent terminal, and employed a trio of engineers who knew each other from their interests in Amateur Radio: Victor Poor, Harry Pyle and undergraduate Jonathan Schmidt. The company began development on the Datapoint 2200, its most popular product and credited by some historians as the invention of the personal computer. While working his notice from Maryland-based Frederick Electronics during the 1969 Thanksgiving holiday, Poor and Pyle developed the underlying instruction set architecture of the processor on a living room floor. This enabled Phil Ray and Gus Roche to design and develop the mass-produced programmable 2200, which could lo
https://en.wikipedia.org/wiki/RPR
RPR may refer to: Computing RPR FOM, a distributed computer simulation standard Science RPR problem diagnosis, to find the cause of IT problems RPRD2 gene, which encodes the KIAA0460 protein RprA RNA, a gene Flopristin or RPR 132552A, an antibiotic Rapid plasma reagin, a screening test for syphilis Politics Rally for the Republic, or Rassemblement pour la République, a defunct French political party Republican Party of Russia – People's Freedom Party, or RPR-PARNAS, a Russian political party Romanian People's Republic or Republica Populară Romînă, 1947-1965 Transportation Swami Vivekananda Airport, India, IATA code Redlands Passenger Rail, previous name of Arrow, California, United States Other Registered Professional Reporter, a US certification Resilient Packet Ring or IEEE 802.17, a data traffic protocol Ruger Precision Rifle, an American bolt-action rifle See also
https://en.wikipedia.org/wiki/Function%20key
A function key is a key on a computer or terminal keyboard that can be programmed so as to cause an operating system command interpreter or application program to perform certain actions, a form of soft key. On some keyboards/computers, function keys may have default actions, accessible on power-on. Function keys on a terminal may either generate short fixed sequences of characters, often beginning with the escape character (ASCII 27), or the characters they generate may be configured by sending special character sequences to the terminal. On a standard computer keyboard, the function keys may generate a fixed, single byte code, outside the normal ASCII range, which is translated into some other configurable sequence by the keyboard device driver or interpreted directly by the application program. Function keys may have abbreviations or pictographic representations of default actions printed on/besides them, or they may have the more common "F-number" designations. History The Singer/Friden 2201 Flexowriter Programmatic, introduced in 1965, had a cluster of 13 function keys, labeled F1 to F13 to the right of the main keyboard. Although the Flexowriter could be used as a computer terminal, this electromechanical typewriter was primarily intended as a stand-alone word processing system. The interpretation of the function keys was determined by the programming of a plugboard inside the back of the machine. Soft keys date to avionics multi-function displays of military planes of the late 1960s/early 1970s, such as the Mark II avionics of the F-111D (first ordered 1967, delivered 1970–73). In computing use, they were found on the HP 9810A calculator (1971) and later models of the HP 9800 series, which featured 10 programmable keys in 5×2 block (2 rows of 5 keys) at the top left of the keyboard, with paper labels. The HP 9830A (1972) was an early desktop computer, and one of the earliest specifically computing uses. HP continued its use of function keys in the HP 2640 (1975), which used screen-labeled function keys, placing the keys close to the screen, where labels could be displayed for their function. NEC's PC-8001, introduced in 1979, featured five function keys at the top of the keyboard, along with a numeric keypad on the right-hand side of the keyboard. Their modern use may have been popularized by IBM keyboards: first the IBM 3270 terminals, then the IBM PC. IBM use of function keys dates to the IBM 3270 line of terminals, specifically the IBM 3277 (1972) with 78-key typewriter keyboard or operator console keyboard version, which both featured 12 programmed function (PF) keys in a 3×4 matrix at the right of the keyboard. Later models replaced this with a numeric keypad, and moved the function keys to 24 keys at the top of the keyboard. The original IBM PC keyboard (PC/XT, 1981) had 10 function keys (F1–F10) in a 2×5 matrix at the left of the keyboard; this was replaced by 12 keys in 3 blocks of 4 at the top of the keyboard in the Model
https://en.wikipedia.org/wiki/Inheritance%20%28genetic%20algorithm%29
In genetic algorithms, inheritance is the ability of modeled objects to mate, mutate (similar to biological mutation), and propagate their problem solving genes to the next generation, in order to produce an evolved solution to a particular problem. The selection of objects that will be inherited from in each successive generation is determined by a fitness function, which varies depending upon the problem being addressed. The traits of these objects are passed on through chromosomes by a means similar to biological reproduction. These chromosomes are generally represented by a series of genes, which in turn are usually represented using binary numbers. This propagation of traits between generations is similar to the inheritance of traits between generations of biological organisms. This process can also be viewed as a form of reinforcement learning, because the evolution of the objects is driven by the passing of traits from successful objects which can be viewed as a reward for their success, thereby promoting beneficial traits. Process Once a new generation is ready to be created, all of the individuals that have been successful and have been chosen for reproduction are randomly paired together. Then the traits of these individuals are passed on through a combination of crossover and mutation. This process follows these basic steps: Pair off successful objects for mating. Determine randomly a crossover point for each pair. Switch the genes after the crossover point in each pair. Determine randomly if any genes are mutated in the child objects. After following these steps, two child objects will be produced for every pair of parent objects used. Then, after determining the success of the objects in the new generation, this process can be repeated using whichever new objects were most successful. This will usually be repeated until either a desired generation is reached or an object that meets a minimum desired result from the fitness function is found. While crossover and mutation are the common genetic operators used in inheritance, there are also other operators such as regrouping and colonization-extinction. Example Assume these two strings of bits represent the traits being passed on by two parent objects: Object 1: 1100011010110001 Object 2: 1001100110011001 Now, consider that the crossover point is randomly positioned after the fifth bit: Object 1: 11000 | 11010110001 Object 2: 10011 | 00110011001 During crossover, the two objects will swap all of the bits after the crossover point, leading to: Object 1: 11000 | 00110011001 Object 2: 10011 | 11010110001 Finally, mutation is simulated on the objects by there being zero or more bits flipped randomly. Assuming the tenth bit for object 1 is mutated, and the second and seventh bits are mutated for object 2, the final children produced by this inheritance would be: Object 1: 1100000111011001 Object 2: 1101110010110001 See also Artificial intelligence Bioinformatic
https://en.wikipedia.org/wiki/Boxing%20%28computer%20science%29
In computer science, boxing (a.k.a. wrapping) is the transformation of placing a primitive type within an object so that the value can be used as a reference. Unboxing is the reverse transformation of extracting the primitive value from its wrapper object. Autoboxing is the term for automatically applying boxing and/or unboxing transformations as needed. Boxing Boxing's most prominent use is in Java where there is a distinction between reference and value types for reasons such as runtime efficiency and syntax and semantic issues. In Java, a can only store values of type . One might desire to have a of , but this is not directly possible. Instead Java defines primitive wrapper classes corresponding to each primitive type: and , and , and , etc. One can then define a using the boxed type and insert values into the list by boxing them as objects. (Using generic parameterized types introduced in J2SE 5.0, this type is represented as .) On the other hand, C# has no primitive wrapper classes, but allows boxing of any value type, returning a generic reference. In Objective-C, any primitive value can be prefixed by a to make an out of it (e.g. or ). This allows for adding them in any of the standard collections, such as an . Haskell has little or no notion of reference type, but still uses the term "boxed" for the runtime system's uniform pointer-to-tagged union representation. The boxed object is always a copy of the value object, and is usually immutable. Unboxing the object also returns a copy of the stored value. Repeated boxing and unboxing of objects can have a severe performance impact, because boxing dynamically allocates new objects and unboxing (if the boxed value is no longer used) then makes them eligible for garbage collection. However, modern garbage collectors such as the default Java HotSpot garbage collector can more efficiently collect short-lived objects, so if the boxed objects are short-lived, the performance impact may not be severe. In some languages, there is a direct equivalence between an unboxed primitive type and a reference to an immutable, boxed object type. In fact, it is possible to substitute all the primitive types in a program with boxed object types. Whereas assignment from one primitive to another will copy its value, assignment from one reference to a boxed object to another will copy the reference value to refer to the same object as the first reference. However, this will not cause any problems, because the objects are immutable, so there is semantically no real difference between two references to the same object or to different objects (unless you look at physical equality). For all operations other than assignment, such as arithmetic, comparison, and logical operators, one can unbox the boxed type, perform the operation, and re-box the result as needed. Thus, it is possible to not store primitive types at all. Autoboxing Autoboxing is the term for getting a reference type out of a value type
https://en.wikipedia.org/wiki/ECTS
ECTS may refer to: Elementary cognitive tasks, from psychometrics Engine coolant temperature sensor European Calcified Tissue Society European Computer Trade Show European Credit Transfer and Accumulation System, a higher education standard ECTS grading scale a misspelling of ETCS, the European Train Control System
https://en.wikipedia.org/wiki/Fifth%20Generation%20Computer%20Systems
The Fifth Generation Computer Systems (FGCS; ) was a 10-year initiative begun in 1982 by Japan's Ministry of International Trade and Industry (MITI) to create computers using massively parallel computing and logic programming. It aimed to create an "epoch-making computer" with supercomputer-like performance and to provide a platform for future developments in artificial intelligence. FGCS was ahead of its time, and its excessive ambitions led to commercial failure. However, on a theoretical level, the project spurred the development of concurrent logic programming. The term "fifth generation" was intended to convey the system as being advanced: In the history of computing hardware, there were four "generations" of computers. Computers using vacuum tubes were called the first generation; transistors and diodes, the second; integrated circuits, the third; and those using microprocessors, the fourth. Whereas previous computer generations had focused on increasing the number of logic elements in a single CPU, the fifth generation, it was widely believed at the time, would instead turn to massive numbers of CPUs to gain performance. Background In the late 1960s until the early 1970s, there was much talk about "generations" of computer hardware, then usually organized into three generations. First generation: Thermionic vacuum tubes. Mid-1940s. IBM pioneered the arrangement of vacuum tubes in pluggable modules. The IBM 650 was a first-generation computer. Second generation: Transistors. 1956. The era of miniaturization begins. Transistors are much smaller than vacuum tubes, draw less power, and generate less heat. Discrete transistors are soldered to circuit boards, with interconnections accomplished by stencil-screened conductive patterns on the reverse side. The IBM 7090 was a second-generation computer. Third generation: Integrated circuits (silicon chips containing multiple transistors). 1964. A pioneering example is the ACPX module used in the IBM 360/91, which, by stacking layers of silicon over a ceramic substrate, accommodated over 20 transistors per chip; the chips could be packed together onto a circuit board to achieve unprecedented logic densities. The IBM 360/91 was a hybrid second- and third-generation computer. Omitted from this taxonomy is the "zeroth-generation" computer based on metal gears (such as the IBM 407) or mechanical relays (such as the Mark I), and the post-third-generation computers based on Very Large Scale Integrated (VLSI) circuits. There was also a parallel set of generations for software: First generation: Machine language. Second generation: Low-level programming languages such as Assembly language. Third generation: Structured high-level programming languages such as C, COBOL and FORTRAN. Fourth generation: "Non-procedural" high-level programming languages (such as object-oriented languages). Throughout these multiple generations up to the 1970s, Japan built computers following U.S. and British leads
https://en.wikipedia.org/wiki/On%20the%20fly
On the fly is a phrase used to describe something that is being changed while the process that the change affects is ongoing. It is used in the automotive, computer, and culinary industries. In cars, on the fly can be used to describe the changing of the cars configuration while it is still driving. Processes that can occur while the car is still driving include switching between two wheel drive and four wheel drive on some cars and opening and closing the roof on some convertible cars. In computing, on the fly CD writers can read from one CD and write the data to another without saving it on a computer's memory. Switching programs or applications on the fly in multi-tasking operating systems means the ability to switch between native and/or emulated programs or applications that are still running and running in parallel while performing their tasks or processes, but without pausing, freezing, or delaying any, or other unwanted events. Switching computer parts on the fly means computer parts are replaced while the computer is still running. It can also be used in programming to describe changing a program while it is still running. In restaurants and other places involved in the preparation of food, the term is used to indicate that an order needs to be made right away. Colloquial usage In colloquial use, "on the fly" means something created when needed. The phrase is used to mean: something that was not planned ahead changes that are made during the execution of same activity: ex tempore, impromptu. Automotive usage In the automotive industry, the term refers to the circumstance of performing certain operations while a vehicle is driven by the engine and moving. In reference to four-wheel drive vehicles, this term describes the ability to change from two to four-wheel drive while the car is in gear and moving. In some convertible models, the roof can be folded electrically on the fly, whereas in other cases the car must be stopped. In harvesting machines, newer monitoring systems let the driver track the quality of the grain, while enabling them to adjust the rotor speed on the fly as harvesting progresses. Computer usage In multitasking computing an operating system can handle several programs, both native applications or emulated software, that are running independent, parallel, together in the same time in the same device, using separated or shared resources and/or data, executing their tasks separately or together, while a user can switch on the fly between them or groups of them to use obtained effects or supervise purposes, without waste of time or waste of performance. In operating systems using GUI very often it is done by switching from an active window (or an object playing similar role) of a particular software piece to another one but of another software. A computer can compute results on the fly, or retrieve a previously stored result. It can mean to make a copy of a removable media (CD-ROM, DVD, etc.) directly, without fir
https://en.wikipedia.org/wiki/List%20of%20the%20most%20popular%20names%20in%20the%201910s%20in%20the%20United%20States
These are the most popular given names in the United States for all years of the 1910s. Data from the Social Security Administration. 1910 Males John William James Robert Joseph Charles; George (tie) Edward Frank Henry ----- Females Mary Helen Margaret Dorothy; Ruth (tie) Anna Mildred Elizabeth Alice Ethel ----- 1911 Males John William James Joseph Charles; Robert (tie) George Frank Edward Walter ----- Females Mary Helen Dorothy; Margaret (tie) Ruth Anna Mildred Elizabeth Marie Gladys ----- 1912 Males John William James Robert George Joseph Charles Frank Edward Thomas; Walter (tie) Females Mary Helen Dorothy Ruth Margaret Anna Mildred Frances Elizabeth Marie 1913 Males John William James Robert Joseph Charles George Frank Edward Thomas Females Mary Helen Dorothy Margaret Ruth Mildred Elizabeth Anna Marie Florence 1914 Males John William James Robert Joseph George Charles Frank Edward Walter Females Mary Helen Dorothy Margaret Ruth Mildred Anna Elizabeth Evelyn Marie 1915 Males John William James Robert Joseph Charles George Edward Frank Thomas Females Mary Helen Dorothy Margaret Ruth Anna Mildred Evelyn Virginia Elizabeth 1916 Males John William James Robert Charles George Joseph Edward Frank Walter Females Mary Helen Margaret Dorothy Ruth Mildred Anna Frances Elizabeth Marie 1917 Males John William James Robert Joseph George Charles Edward Frank Thomas; Walter (tie) Females Mary Helen Dorothy Margaret Ruth Anna Frances Elizabeth Mildred Marie 1918 Males John William Robert James Joseph Charles George Edward Frank Thomas Females Mary Helen Dorothy Margaret Ruth Frances Virginia Anna Mildred Elizabeth 1919 Males John William James Robert Charles Joseph George Edward Frank Thomas Females Mary Helen Dorothy Margaret Ruth Virginia Elizabeth Mildred Frances Anna References http://www.ssa.gov/OACT/babynames/index.html Most Popular 1000 Names of the 1910s from the Social Security Administration 1910s 1910s in the United States
https://en.wikipedia.org/wiki/AppleLink
AppleLink was the name of both Apple Computer's online service for its dealers, third-party developers, and users, and the client software used to access it. Prior to the commercialization of the Internet, AppleLink was a popular service for Mac and Apple IIGS users. The service was offered from about 1986 to 1994 to various groups, before being superseded by their short-lived eWorld and finally today's multiple Apple websites. Early years The original AppleLink, which went online in 1985, was a service available only to Apple employees and dealers, and shortly thereafter to Apple University Consortium members. Apple's consumer 800 number in fact touted this fact, promoting your dealer as the place to turn for help because of his access to AppleLink. In the late 1980s the service was also opened up to software developers, who could use it both as an end-user support system as well as a conduit to Apple development for questions and suggestions. AppleLink used client software written in Pascal under contract to Apple by Pete Burnight/Central Coast Software. The program extended the desktop metaphor of the Macintosh Finder to encompass the areas on the remote server site. These were displayed as folders and files just as local folders and files were. In addition, there was a set of public bulletin boards, and the ability to use email via the service—although initially only between AppleLink users. File transfer for drivers and system software was another important role, and for this Apple created the AppleLink Package format to combine and compress the two forks of a Macintosh file into one for storage and sending. Apple also developed their Communications Control Language (CCL) for AppleLink, a language still used in a very similar form for today's Macintosh modem scripts. The "back end" of the AppleLink system was hosted on General Electric's Information Services (GEIS) (division) Mark III time-sharing mainframes and worldwide communications network. AppleLink translated the user's GUI actions into a transaction protocol that was sent into a server program running on the mainframe. The connection used a proprietary error free sliding window protocol called EF3. Later versions upgraded this protocol to GEIS' "EFX" (Error-Free eXtended) protocol. GEIS charged fairly substantial prices to Apple, both for maintaining the service (about $30 million a year per year), and the end-users' connection fees of about $15 an hour during business hours (which Apple rebilled to the users). For some years prior to the discontinuation of the system, GEIS maintained a number of employees on the campus of Apple's Cupertino, California, headquarters, co-located with the Apple On-Line Services organization. Repeated attempts to negotiate a lower cost failed, and Apple management chafed at paying for a service that had no obviously measurable income. Eventually Apple approached Steve Case of Quantum Computer Services, who ran a somewhat similar system for users o
https://en.wikipedia.org/wiki/Digital%20video%20recorder
A digital video recorder (DVR) is an electronic device that records video in a digital format to a disk drive, USB flash drive, SD memory card, SSD or other local or networked mass storage device. The term includes set-top boxes with direct to disk recording, portable media players and TV gateways with recording capability, and digital camcorders. Personal computers are often connected to video capture devices and used as DVRs; in such cases the application software used to record video is an integral part of the DVR. Many DVRs are classified as consumer electronic devices; such devices may alternatively be referred to as personal video recorders (PVRs), particularly in Canada. Similar small devices with built-in (~5 inch diagonal) displays and SSD support may be used for professional film or video production, as these recorders often do not have the limitations that built-in recorders in cameras have, offering wider codec support, the removal of recording time limitations and higher bitrates. History Hard-disk-based digital video recorders The first working DVR prototype was developed in 1998 at Stanford University Computer Science department. The DVR design was a chapter of Edward Y. Chang's PhD dissertation, supervised by Professors Hector Garcia-Molina and Jennifer Widom. Two design papers were published 2017 VLDB conference, and 1999 ICDE conference. The prototype was developed in 1998 at Pat Hanrahan's CS488 class: Experiments in Digital Television, and the prototype was demoed to industrial partners including SONY, Intel, and Apple. Consumer digital video recorders ReplayTV and TiVo were launched at the 1999 Consumer Electronics Show in Las Vegas, Nevada. Microsoft also demonstrated a unit with DVR capability, but this did not become available until the end of 1999 for full DVR features in Dish Network's DISHplayer receivers. TiVo shipped their first units on March 31, 1999. ReplayTV won the "Best of Show" award in the video category with Netscape co-founder Marc Andreessen as an early investor and board member, but TiVo was more successful commercially. Ad Age cited Forrester Research as saying that market penetration by the end of 1999 was "less than 100,000". Legal action by media companies forced ReplayTV to remove many features such as automatic commercial skip and the sharing of recordings over the Internet, but newer devices have steadily regained these functions while adding complementary abilities, such as recording onto DVDs and programming and remote control facilities using PDAs, networked PCs, and Web browsers. In contrast to VCRs, hard-disk based digital video recorders make "time shifting" more convenient and also allow for functions such as pausing live TV, instant replay, chasing playback (viewing a recording before it has been completed) and skipping over advertising during playback. Many DVRs use the MPEG format for compressing the digital video. Video recording capabilities have become an essential part of the m
https://en.wikipedia.org/wiki/Desktop%20metaphor
In computing, the desktop metaphor is an interface metaphor which is a set of unifying concepts used by graphical user interfaces to help users interact more easily with the computer. The desktop metaphor treats the computer monitor as if it is the top of the user's desk, upon which objects such as documents and folders of documents can be placed. A document can be opened into a window, which represents a paper copy of the document placed on the desktop. Small applications called desk accessories are also available, such as a desk calculator or notepad, etc. The desktop metaphor itself has been extended and stretched with various implementations of desktop environments, since access to features and usability of the computer are usually more important than maintaining the 'purity' of the metaphor. Hence one can find trash cans on the desktop, as well as disks and network volumes (which can be thought of as filing cabinets—not something normally found on a desktop). Other features such as menu bars or taskbars have no direct counterpart on a real-world desktop, though this may vary by environment and the function provided; for instance, a familiar wall calendar can sometimes be displayed or otherwise accessed via a taskbar or menu bar belonging to the desktop. History The desktop metaphor was first introduced by Alan Kay, David C. Smith, and others at Xerox PARC in 1970 and elaborated in a series of innovative software applications developed by PARC scientists throughout the ensuing decade. The first computer to use an early version of the desktop metaphor was the experimental Xerox Alto, and the first commercial computer that adopted this kind of interface was the Xerox Star. The use of window controls to contain related information predates the desktop metaphor, with a primitive version appearing in Douglas Engelbart's "Mother of All Demos", though it was incorporated by PARC in the environment of the Smalltalk language. One of the first desktop-like interfaces on the market was a program called Magic Desk I. Built as a cartridge for the Commodore 64 home computer in 1983, a very primitive GUI presented a low resolution sketch of a desktop, complete with telephone, drawers, calculator, etc. The user made their choices by moving a sprite depicting a hand pointing by using the same joystick the user may have used for video gaming. Onscreen options were chosen by pushing the fire button on the joystick. The Magic Desk I program featured a typewriter graphically emulated complete with audio effects. Other applications included a calculator, rolodex organiser, and a terminal emulator. Files could be archived into the drawers of the desktop. A trashcan was also present. The first computer to popularise the desktop metaphor, using it as a standard feature over the earlier command-line interface was the Apple Macintosh in 1984. The desktop metaphor is ubiquitous in modern-day personal computing; it is found in most desktop environments of modern op
https://en.wikipedia.org/wiki/Bitzi
Bitzi was a website, operating from 2001 to 2013, where volunteers shared reports about any kind of digital file, with identifying metadata, commentary, and other ratings. Information contributed and rated by volunteers was compiled into the Bitpedia data set and reference work, described by Bitzi as a "digital media encyclopedia". The Bitpedia was published through the Bitzi website and web services under an open content license (Creative Commons Attribution - Share Alike 2.0). Bitzi's standards and services have been adopted by a number of popular peer-to-peer file sharing systems. Bitzi was sponsored by a metadata publishing company of the same name based in San Francisco. History Founded by Gordon Mohr with Mike Linksvayer the Bitzi service launched in 2001. The Bitzi website shut down on 31 December 2013. Technology At Bitzi, files are identified by applying a strong hash function to their contents, which gives a distinct "fingerprint" for each file. Bitzi calls the combination of standard hash functions used by its system "bitprints." An open source downloadable tool, the Bitcollider, calculates file hashes and extracts intrinsic metadata from common media file types to assist user contributions. Bitzi cross-references multiple Uniform Resource Identifiers (URIs) for files, primarily URIs used by peer-to-peer file-sharing networks and/or clients: sha1:/magnet: (Gnutella) (Grokster) tree:tiger: (DC++, Shareaza). ed2k: (eDonkey2000, eMule, OverNet) Other file-specific metadata is also collected, such as file size, alternate filenames, audio/video encoding details, user ratings, and free-form comments or descriptions. Users can displace flawed information with new contributions. Data about specific files can also be programmatically retrieved via a REST-style XML Web Services. Relationship to peer-to-peer networks Bitzi originated several popular standards in the peer-to-peer file sharing sphere, including the Magnet URI scheme and Tiger tree hashes. Many peer-to-peer file sharing programs, including LimeWire/Frostwire, older Bearshare versions, and older Shareaza versions, offer an option to look up local files or network search results at Bitzi for more information about their contents or quality. As peer-to-peer file sharing networks are often plagued by mislabeled or corrupt files, Bitzi can sometimes provide additional confidence that a file is as expected, before a user begins a long download. In such a role, Bitzi serves to ameliorate some common attacks on peer-to-peer networks. References External links The bitzi.com website List of applications using Bitzi Bitcollider project at SourceForge Knowledge markets
https://en.wikipedia.org/wiki/Trinity%20Broadcasting%20Network
The Trinity Broadcasting Network (TBN; legally Trinity Broadcasting of Texas, Inc.) is an international Christian-based broadcast television network and the world's largest religious television network. TBN was headquartered in Costa Mesa, California, until March 3, 2017, when it sold its highly visible office park, Trinity Christian City. The broadcaster retained its studios in nearby Tustin. Auxiliary studio facilities are located in Irving, Hendersonville, Gadsden, Decatur, Miami and Orlando, Tulsa and New York City. TBN has characterized itself as broadcasting programs hosted by a diverse group of ministries from Evangelical, traditional Protestant and Catholic denominations, non-profit charities, Messianic Jewish and Christian media personalities. TBN also offers a wide range of original programming, faith-based films, and political opinion commentary from various distributors. TBN owns and operates six broadcast networks, each reaching separate demographics. In addition to the main TBN network, TBN owns TBN Inspire, Smile, TBN Enlace (how TBN distributes Enlace in the United States), TBN Salsa and Positiv. It also owns several other religious networks outside the United States, including international versions of its five U.S. networks. Matt Crouch is currently TBN's president and head of operations. History The Trinity Broadcasting Network was co-founded as the Trinity Broadcasting Systems in 1973 by Paul Crouch, an Assemblies of God minister, and his spouse Jan Crouch. TBN began its broadcasting activities by renting time on the independent station KBSA (now UniMás owned-and-operated station KFTR-DT) in Ontario, California. After that station was sold, he began buying two hours a day of programming time on KLXA-TV in Fontana, California, in early 1974. That station was put up for sale shortly afterward. Paul Crouch then placed a bid to buy the station for $1 million and raised $100,000 for a down payment. After many struggles, the Crouches managed to raise the down payment and took over the station outright, with the station becoming KTBN-TV in 1977 and its city of license being reassigned to TBN's original homebase, Santa Ana, in 1983. Initially, the station ran Christian programs for about six hours a day, expanding its programming to 12 hours a day by 1975, and began selling time to other Christian organizations to supplement its local programming. The station eventually instituted a 24-hour schedule in 1978. The fledgling network was so weak in its first days, that, according to Crouch in his autobiography, Hello World!, it almost went bankrupt after just two days on the air. TBN began national distribution through cable television providers in 1978. The ministry, which became known as the Trinity Broadcasting Network, gained national distribution via communications satellite in 1982. The network was a member of the National Religious Broadcasters association until 1990. In 1977, the ministry purchased KPAZ-TV in Phoenix, Arizon
https://en.wikipedia.org/wiki/TBN
TBN may mean: Media Trinity Broadcasting Network, an international religious television network The Baseball Network, former US television network The Buffalo News, a newspaper Other The Barter Network, a commercial trading network Toronto Bicycling Network, a recreational cycling organization Total base number, an indicator for the alkalinity of lubricant oil Trombone, a musical instrument Waynesville-St. Robert Regional Airport
https://en.wikipedia.org/wiki/Tricorder
A tricorder is a fictional handheld sensor that exists in the Star Trek universe. The tricorder is a multifunctional hand-held device that can perform environmental scans, data recording, and data analysis; hence the word "tricorder" to refer to the three functions of sensing, recording, and computing. In Star Trek stories the devices are issued by the fictional Starfleet organization. The original physical prop for the tricorder was designed by Wah Chang and appeared in The Man Trap in 1966, the first Star Trek episode to air. Types The tricorder of the 23rd century, as seen in Star Trek: The Original Series, is a black, rectangular device with a top-mounted rotating hood, two opening compartments, and a shoulder strap. The top pivots open, exposing a small screen and control buttons. Three main variants appear in shows. The standard tricorder is a general-purpose device used primarily to scout unfamiliar areas, make detailed examination of living things, and record and review technical data. The medical tricorder is used by doctors to help diagnose diseases and collect bodily information about a patient; the key difference between this and a standard tricorder is a detachable hand-held high-resolution scanner stored in a compartment of the tricorder when not in use. The engineering tricorder is fine-tuned for starship engineering purposes. There are also many other lesser-used varieties of special-use tricorders. The ship's medical variant employs a detachable "sensor probe" stored in the bottom compartment when not in use. The probe, although originally thought to have been fashioned from a spare salt shaker, was actually scratch-built for the show; the conical Danish salt shakers were set dressing, used as laser scalpels. The 24th-century version introduced in Star Trek: The Next Generation is a small, gray, hand-held model with a flip-out panel to allow for a larger screen. This design was later refined with a slightly more angular appearance that was seen in most Next Generation–era movies as well as later seasons of Star Trek: Deep Space Nine and Voyager. In the post-Next Generation-era (Star Trek: Nemesis and Star Trek: Elite Force II ), a newer tricorder was introduced. It is flatter, with a small flap that opens on top and a large touchscreen interface. Production The tricorder prop for the original Star Trek series was designed and built by Wah Ming Chang, who created several futuristic props under contract. Some of his designs are considered to have been influential on later, real-world consumer electronics devices. For instance, his communicator inspired cell phone inventor Martin Cooper's desire to create his own form of mobile communication device. Many other companies followed this example and life-sized replicas remain popular collectibles today. The tricorder in The Next Generation was initially inspired by the HP-41C scientific calculator. "Real" tricorders Software exists to make hand-held devices simulate a tricorder.
https://en.wikipedia.org/wiki/Visual%20Basic%20Extension
In computer programming, a Visual Basic Extension (commonly abbreviated VBX) or custom control, was the component model used in Microsoft Visual Basic versions 1.0 to 3.0. It is still supported in the 16-bit version of version 4.0, but was made obsolete by OCXs. VBX has also been supported in the first, 16-bit version of Borland Delphi. The VBX specification was created so that developers could extend Visual Basic. Since at the time Visual Basic was an interpreted language with limited speed, VBXs, which were most often written in C or C++, provided the ability to create very powerful and efficient user interface sub-components which could be assembled into applications using Visual Basic as the glue. Visual Basic 4.0 added support for the newer OLE Controls or OCXs (which became ActiveX controls), implicitly deprecating VBX controls. OCXs are based on the Component Object Model. Each version of Visual Basic has come with many controls. By incorporating VBX controls into Visual Basic, Microsoft spawned the first commercially viable market for reusable software components. Third parties have created a large market of custom controls for resale. VBX components were called "custom controls" because the Microsoft development team envisioned them as visual "control" components like a toggle switch or button. Enterprising third party software component developers saw opportunities to use the VBX specification for non-visual components, such as components for creating ZIP files, manipulating low-level system settings and communicating with TCP/IP. References https://web.archive.org/web/20091002232037/http://www.ascilite.org.au/ajet/ajet10/nelson.html Microsoft application programming interfaces
https://en.wikipedia.org/wiki/Document%20warehouse
In the field of data warehouses, a document warehouse is a software framework for analysis, sharing, and reuse of unstructured data, such as textual or multimedia documents. This is different from data warehouses that focuses on structured data, such as tabularized sales reports. On the other hand, Document Warehouse for SAP is also a FileNet's commercial software that enables SAP's business applications to access document images stored by FileNet. See also Decision support References Data warehousing
https://en.wikipedia.org/wiki/Eigenface
An eigenface () is the name given to a set of eigenvectors when used in the computer vision problem of human face recognition. The approach of using eigenfaces for recognition was developed by Sirovich and Kirby and used by Matthew Turk and Alex Pentland in face classification. The eigenvectors are derived from the covariance matrix of the probability distribution over the high-dimensional vector space of face images. The eigenfaces themselves form a basis set of all images used to construct the covariance matrix. This produces dimension reduction by allowing the smaller set of basis images to represent the original training images. Classification can be achieved by comparing how faces are represented by the basis set. History The eigenface approach began with a search for a low-dimensional representation of face images. Sirovich and Kirby showed that principal component analysis could be used on a collection of face images to form a set of basis features. These basis images, known as eigenpictures, could be linearly combined to reconstruct images in the original training set. If the training set consists of M images, principal component analysis could form a basis set of N images, where N < M. The reconstruction error is reduced by increasing the number of eigenpictures; however, the number needed is always chosen less than M. For example, if you need to generate a number of N eigenfaces for a training set of M face images, you can say that each face image can be made up of "proportions" of all the K "features" or eigenfaces: Face image1 = (23% of E1) + (2% of E2) + (51% of E3) + ... + (1% En). In 1991 M. Turk and A. Pentland expanded these results and presented the eigenface method of face recognition. In addition to designing a system for automated face recognition using eigenfaces, they showed a way of calculating the eigenvectors of a covariance matrix such that computers of the time could perform eigen-decomposition on a large number of face images. Face images usually occupy a high-dimensional space and conventional principal component analysis was intractable on such data sets. Turk and Pentland's paper demonstrated ways to extract the eigenvectors based on matrices sized by the number of images rather than the number of pixels. Once established, the eigenface method was expanded to include methods of preprocessing to improve accuracy. Multiple manifold approaches were also used to build sets of eigenfaces for different subjects and different features, such as the eyes. Generation A set of eigenfaces can be generated by performing a mathematical process called principal component analysis (PCA) on a large set of images depicting different human faces. Informally, eigenfaces can be considered a set of "standardized face ingredients", derived from statistical analysis of many pictures of faces. Any human face can be considered to be a combination of these standard faces. For example, one's face might be composed of the average face
https://en.wikipedia.org/wiki/Crawler
Crawler may refer to: Web crawler, a computer program that gathers and categorizes information on the World Wide Web A first-instar nymph of a scale insect that has legs and walks around before it attaches itself and becomes stationary Crawler (BEAM) in robotics A type of crane on tracks "Crawlers" (Into the Dark), an episode of the second season of Into the Dark The Crawler, an episode of the cartoon Extreme Ghostbusters Crawler (album), an album by IDLES Crawler (band), a British rock band Crawlers (band), a British rock band A fictional creature in the video game Fable III A fictional creature in the movie The Descent See also Bottom crawler, an underwater exploration and recovery vehicle Crawl (disambiguation) Crawler-transporter, a large tracked vehicle used by NASA to transport spacecraft Nightcrawler (Lumbricus terrestris), an annelid worm also called "common earthworm" and "dew worm"
https://en.wikipedia.org/wiki/IBM%20Personal%20Computer%20AT
The IBM Personal Computer AT (model 5170, abbreviated as IBM AT or PC/AT) was released in 1984 as the fourth model in the IBM Personal Computer line, following the IBM PC/XT and its IBM Portable PC variant. It was designed around the Intel 80286 microprocessor. Name IBM did not specify an expanded form of "AT" on the machine, press releases, brochures or documentation, but some sources expand the term as "Advanced Technology", including at least one internal IBM document. History IBM's 1984 introduction of the AT was seen as an unusual move for the company, which typically waited for competitors to release new products before producing its own models. At $4,000–6,000, it was only slightly more expensive than considerably slower IBM models. The announcement surprised rival executives, who admitted that matching IBM's prices would be difficult. No major competitor showed a comparable computer at COMDEX Las Vegas that year. Features The AT is IBM PC compatible, with the most significant difference being a move to the 80286 processor from the 8088 processor of prior models. Like the IBM PC, the AT supported an optional math co-processor chip, the Intel 80287, for faster execution of floating point operations. In addition, it introduces the AT bus, later known as the ISA bus, a 16-bit bus with backwards compatibility with 8-bit PC-compatible expansion cards. The bus also offered fifteen IRQs and seven DMA channels, expanded from eight IRQs and four DMA channels for the PC, achieved by adding another 8259A IRQ controller and another 8237A DMA controller. Some IRQ and DMA channels are used by the motherboard and not exposed on the expansion bus. Both dual IRQ and DMA chipsets are cascading which shares the primary pair. In addition to these chipsets, Intel 82284 Clock Driver and Ready Interface and Intel 82288 Bus Controller are to support the microprocessor. The 24-bit address bus of the 286 expands RAM capacity to 16 MB. PC DOS 3.0 was included with support for the new AT features, including preliminary kernel support for networking (which was fully supported in a later version 3.x release). The motherboard includes a battery-backed real-time clock (RTC) using the Motorola MC146818. This was an improvement from the PC, which required setting the clock manually or installing an RTC expansion card. The RTC also included a 1024 Hz timer (on IRQ 8), a much finer resolution than the 18 Hz timer on the PC. In addition to keeping the time, the RTC includes 50 bytes of CMOS memory which is used to store software-adjustable BIOS parameters. A disk-based BIOS setup program which saved to this memory took the place of the DIP switches used to set system settings on PCs. Most AT clones have the setup program in ROM rather than on disk. Storage The standard floppy drive was upgraded to a 1.2 MB  inch floppy disk drive (15 sectors of 512 bytes, 80 tracks, two sides), which stored over three times as much data as the 360 KB PC floppy disk, but had compa
https://en.wikipedia.org/wiki/Eli%20Biham
Eli Biham () is an Israeli cryptographer and cryptanalyst who is a professor at the Technion - Israel Institute of Technology Computer Science department. From 2008 to 2013, Biham was the dean of the Technion Computer Science department, after serving for two years as chief of CS graduate school. Biham invented (publicly) differential cryptanalysis, for which he received his Ph.D., while working under Adi Shamir. It had been invented before by a team at IBM during their Data Encryption Standard work; the National Security Agency told IBM to keep the discovery secret. Contributions to cryptanalysis differential cryptanalysis - publicly invented during his Ph.D. studies under Adi Shamir Attacking all triple modes of operation. impossible differential cryptanalysis - joint work with Adi Shamir and Alex Biryukov Breaking (together with Lars Knudsen) the ANSI X9.52 CBCM mode (few days before the final standardization) Breaking the GSM security mechanisms (with Elad Barkan and Nathan Keller) Co-invention of related-key attacks. Differential Fault Analysis - joint work with Adi Shamir New cryptographic primitives Biham has taken part in the design of several new cryptographic primitives: Serpent (with Ross Anderson and Lars Knudsen), a block cipher which was one of the final five contenders to become the Advanced Encryption Standard Tiger (with Ross Anderson), a hash function fast on 64-bit machines, and Py (with Jennifer Seberry), one of a family of fast stream ciphers (see article for more detail on their cryptanalytic status). SHAvite-3 (with Orr Dunkelman), a hash function which was one of the 14 semifinalists in the NIST hash function competition. References External links Eli Biham's homepage at Technion 1960 births Living people Biham Eli International Association for Cryptologic Research fellows Israeli computer scientists Biham Eli Israeli Jews Modern cryptographers
https://en.wikipedia.org/wiki/Delphi%20%28software%29
Delphi is a general-purpose programming language and a software product that uses the Delphi dialect of the Object Pascal programming language and provides an integrated development environment (IDE) for rapid application development of desktop, mobile, web, and console software, currently developed and maintained by Embarcadero Technologies. Delphi's compilers generate native code for Microsoft Windows, macOS, iOS, Android and Linux (x64). Delphi includes a code editor, a visual designer, an integrated debugger, a source code control component, and support for third-party plugins. The code editor features Code Insight (code completion), Error Insight (real-time error-checking), and refactoring. The visual forms designer has the option of using either the Visual Component Library (VCL) for pure Windows development or the FireMonkey (FMX) framework for cross-platform development. Database support is a key feature and is provided by FireDAC (Database Access Components). Delphi is known for its fast compilation speed, native code, and developer productivity. Delphi was originally developed by Borland as a rapid application development tool for Windows as the successor of Turbo Pascal. Delphi added full object-oriented programming to the existing language, and the language has grown to support generics, anonymous methods, closures, and native Component Object Model (COM) support. Delphi and its C++ counterpart, C++Builder, are interoperable and jointly sold under the name RAD Studio. There are Professional, Enterprise, and Architect editions, with the higher editions having more features at a higher price. There is also a free-of-charge Community edition, with most of the features of Professional, but restricted to users and companies with low revenue. Features Delphi supports rapid application development (RAD). Prominent features are a visual designer and two application frameworks, VCL for Windows and FireMonkey (FMX) for cross-platform development. Delphi uses the Pascal-based programming language Object Pascal created by Anders Hejlsberg for Borland (now IDERA) as the successor to Turbo Pascal. It supports native cross-compilation to many platforms including Windows, Linux, iOS and Android. To better support development for Microsoft Windows and interoperate with code developed with other software development tools, Delphi supports independent interfaces of Component Object Model (COM) with reference counted class implementations, and support for many third-party components. Interface implementations can be delegated to fields or properties of classes. Message handlers are implemented by tagging a method of a class with the integer constant of the message to handle. Database connectivity is extensively supported through VCL database-aware and database access components. Later versions have included upgraded and enhanced runtime library routines, some provided by the community group FastCode. Characteristics Delphi uses a strongly typed
https://en.wikipedia.org/wiki/Third-party%20software%20component
In computer programming, a third-party software component is a reusable software component developed to be either freely distributed or sold by an entity other than the original vendor of the development platform. The third-party software component market thrives because many programmers believe that component-oriented development improves the efficiency and the quality of developing custom applications. Common third-party software includes macros, bots, and software/scripts to be run as add-ons for popular developing software. In the case of operating systems such as Windows XP, Vista or Seven, there are applications installed by default, such as Windows Media Player or Internet Explorer. See also Middleware Enterprise Java Beans VCL / CLX KParts (KDE) Video-game third-party developers Third-party source Online all programming languages and their third party libraries includes a guide. References Component-based software engineering Computer programming
https://en.wikipedia.org/wiki/List%20of%20programs%20broadcast%20by%20the%20Hellenic%20Broadcasting%20Corporation
The following is a list of programs broadcast by ERT (Ellinikí Radiofonía Tileórasi) television stations. ERT was the Greek television network before the launch of ANT1 and Mega Channel in 1989 as well as ERT2. Children's Alvin and the Chipmunks (dubbed in Greek) (1993) [ΕΤ-2] Andy Pandy Animals Antics – [ET-1] Big World For Little People – educational program for children that deals with human interaction, natural science and mathematics; [ET-3] Cartoons (dubbed in Greek) (1991–1993) – [ET-1] Carussel (dubbed in Greek) – Mexico (1993–1998) [ET-2] [ET-1] CatDog (dubbed in Greek) (broadcast in 2002) [ET-1] Cebollitas (dubbed in Greek) – Argentina (2000–2006) [ET-1] Chiquititas (dubbed in Greek) – Argentina (2002–2004) [ET-1] Code Lyoko (dubbed in Greek) – French Franny's Feet (dubbed in Greek) – Canadian/American animated cartoon series for children; [ET-1] Hallo Spencer (1997) (dubbed in Greek with German subtitles); [ET-1] Hey Duggee – [ΕΡΤ-2] The Intrepids – [ET-2] Kaput and Zösky – [ET-1] Kouklomegaloi kai Polysporia/FraPaidia me MarmEllada – children's TV show; hosted by Sofianos family (2002–2005); [ET-1] Little Nicholas (dubbed in Greek) – French/Indian animated cartoon series for children; (2012–2013) [ET-1] Lucky Luke (dubbed in Greek) – (1995–2002) [ET-1] Maggie and the Ferocious Beast – [ET-3] Μathimatika Αsteria – children's program designed to teach mathematics in a fun way; [ET-3] The Number Crew – educational program designed to help teach kids aged 5–7 math; [ET-3] The Noddy Shop (dubbed in Greek) – (1999–2001) [ET-1] Ntrin Stop – children's game show; hosted by Fotis Petridis and Olga Michalopoulou (1999–2003); [ET-1] Ouranio Toxo – program aimed at pre-schoolers aged 4–6, to prepare them for school; teaches them about the alphabet, numbers, drawing, colours, weather and much more. Hosted by Christos Demopoulos, airs Monday – Friday at 9am; (1986–2013) [ET-1] Pablo the Little Red Fox – [ET-1] Paidi TV – children's TV show; hosted by Giota Militsi (2002–2003); [ET-1] Pop kai pame/Takat tak – children's game shows; hosted by Adam Arnold (2001–2005) (2001–2003 Pop kai pame and 2004–2005 Takat tak); [ET-1] Punky Brewster (1994) [ET-1] Ratz – [ET-1] The Raccoons (Greek dubbed version, 1990–1994; subtitled version, 1997) [ET-2] The Smurfs (dubbed in Greek) (1984-1999) [ET-1] Sesame Street (dubbed in Greek) (1992–2000) [ET-1] Story of the Alps: My Annette – (dubbed in Greek) (1995-1997) [ET-1] Superhero Kindergarten – [ΕΡΤ-2] Teletubbies (2003) – [ET-1] Tweenies – [ET-1] Under the Umbrella Tree (dubbed in Greek) (1990) [ET-2] Unser Charly (dubbed in Greek) (2000-2001) [ET-1] The Zig and Zag Show (dubbed in Greek) (broadcast in 2003) [NET] 1,000 hromata tou Hristou (1,000 χρώματα του Χρήστου) (2019–2022) [ΕΡΤ-2] Tilekpedaifsi (2020) [ΕΡΤ-2] Yoho Ahoy (dubbed in Greek) – (2002-2003) [ET-1] Tom and Jerry – [ET-1] Disney Zone – Disne
https://en.wikipedia.org/wiki/Simplex%20algorithm
In mathematical optimization, Dantzig's simplex algorithm (or simplex method) is a popular algorithm for linear programming. The name of the algorithm is derived from the concept of a simplex and was suggested by T. S. Motzkin. Simplices are not actually used in the method, but one interpretation of it is that it operates on simplicial cones, and these become proper simplices with an additional constraint. The simplicial cones in question are the corners (i.e., the neighborhoods of the vertices) of a geometric object called a polytope. The shape of this polytope is defined by the constraints applied to the objective function. History George Dantzig worked on planning methods for the US Army Air Force during World War II using a desk calculator. During 1946 his colleague challenged him to mechanize the planning process to distract him from taking another job. Dantzig formulated the problem as linear inequalities inspired by the work of Wassily Leontief, however, at that time he didn't include an objective as part of his formulation. Without an objective, a vast number of solutions can be feasible, and therefore to find the "best" feasible solution, military-specified "ground rules" must be used that describe how goals can be achieved as opposed to specifying a goal itself. Dantzig's core insight was to realize that most such ground rules can be translated into a linear objective function that needs to be maximized. Development of the simplex method was evolutionary and happened over a period of about a year. After Dantzig included an objective function as part of his formulation during mid-1947, the problem was mathematically more tractable. Dantzig realized that one of the unsolved problems that he had mistaken as homework in his professor Jerzy Neyman's class (and actually later solved), was applicable to finding an algorithm for linear programs. This problem involved finding the existence of Lagrange multipliers for general linear programs over a continuum of variables, each bounded between zero and one, and satisfying linear constraints expressed in the form of Lebesgue integrals. Dantzig later published his "homework" as a thesis to earn his doctorate. The column geometry used in this thesis gave Dantzig insight that made him believe that the Simplex method would be very efficient. Overview The simplex algorithm operates on linear programs in the canonical form maximize subject to and with the coefficients of the objective function, is the matrix transpose, and are the variables of the problem, is a p×n matrix, and . There is a straightforward process to convert any linear program into one in standard form, so using this form of linear programs results in no loss of generality. In geometric terms, the feasible region defined by all values of such that and is a (possibly unbounded) convex polytope. An extreme point or vertex of this polytope is known as basic feasible solution (BFS). It can be shown that for a linear progr
https://en.wikipedia.org/wiki/Environment%20variable
An environment variable is a user-definable value that can affect the way running processes will behave on a computer. Environment variables are part of the environment in which a process runs. For example, a running process can query the value of the TEMP environment variable to discover a suitable location to store temporary files, or the HOME or USERPROFILE variable to find the directory structure owned by the user running the process. They were introduced in their modern form in 1979 with Version 7 Unix, so are included in all Unix operating system flavors and variants from that point onward including Linux and macOS. From PC DOS 2.0 in 1982, all succeeding Microsoft operating systems, including Microsoft Windows, and OS/2 also have included them as a feature, although with somewhat different syntax, usage and standard variable names. Design In all Unix and Unix-like systems, as well as on Windows, each process has its own separate set of environment variables. By default, when a process is created, it inherits a duplicate run-time environment of its parent process, except for explicit changes made by the parent when it creates the child. At the API level, these changes must be done between running fork and exec. Alternatively, from command shells such as bash, a user can change environment variables for a particular command invocation by indirectly invoking it via env or using the ENVIRONMENT_VARIABLE=VALUE <command> notation. A running program can access the values of environment variables for configuration purposes. Shell scripts and batch files use environment variables to communicate data and preferences to child processes. They can also be used to store temporary values for reference later in a shell script. However, in Unix, non-exported variables are preferred for this as they don't leak outside the process. In Unix, an environment variable that is changed in a script or compiled program will only affect that process and possibly child processes. The parent process and any unrelated processes will not be affected. Similarly, changing or removing a variable's value inside a DOS or Windows batch file will change the variable for the duration of COMMAND.COMor CMD.EXE's existence, respectively. In Unix, the environment variables are normally initialized during system startup by the system init startup scripts, and hence inherited by all other processes in the system. Users can, and often do, augment them in the profile script for the command shell they are using. In Microsoft Windows, each environment variable's default value is stored in the Windows Registry or set in the AUTOEXEC.BAT file. On Unix, a setuid program is given an environment chosen by its caller, but it runs with different authority from its caller. The dynamic linker will usually load code from locations specified by the environment variables $LD_LIBRARY_PATH and $LD_PRELOAD and run it with the process's authority. If a setuid program did this, it would be insecure
https://en.wikipedia.org/wiki/Jakob%20Nielsen%20%28usability%20consultant%29
Jakob Nielsen (born 5 October 1957) is a Danish web usability consultant, human–computer interaction researcher, and co-founder of Nielsen Norman Group. He was named the “guru of Web page usability” in 1998 by The New York Times and the “king of usability” by Internet Magazine. Background Jakob Nielsen was born 5 October 1957 in Copenhagen, Denmark. He holds a Ph.D. in 1988 in human–computer interaction from the Technical University of Denmark. Nielsen's earlier affiliations include Bellcore (now known as Telcordia Technologies, formally Bell Communications Research), teaching at the Technical University of Denmark, and the IBM User Interface Institute at the Thomas J. Watson Research Center. Career Sun Microsystems From 1994 to 1998, he was a Sun Microsystems Distinguished Engineer. He was hired to make heavy-duty enterprise software easier to use, since large-scale applications had been the focus of most of his projects at the phone company and IBM. But luckily the job definition of a Distinguished Engineer is "you're supposed to be the world's leading expert in your field, so you figure out what would be most important for the company for you to work on." Therefore, Dr. Nielsen ended up spending most of his time at Sun defining the emerging field of web usability. He was the usability lead for several design rounds of Sun's website and intranet (SunWeb), including the original SunWeb design in 1994. Nielsen Norman Group After his regular articles on his website about usability research attracted media attention, he co-founded usability consulting company Nielsen Norman Group (NN/g) of Fremont, California in 1998 with fellow usability expert Donald Norman. The company's vision is to help designers and other companies move toward more human-centered products and internet interactions, as experts and pioneers in the field of usability. Other activities Nielsen is on the editorial board of Morgan Kaufmann Publishers' book series in Interactive Technologies. Nielsen writes a fortnightly newsletter, Alertbox, on web design matters and has published several books on the subject of web design. Contributions Nielsen founded the usability engineering movement for efficient and affordable improvements of user interfaces and he has invented several usability methods, including heuristic evaluation. He holds more than a thousand United States patents, mainly on ways of improving usability for technology. In the early 1990s, Nielsen popularized the principle that five test users per usability test session is enough, allowing numerous tests at various stages of the development process. His argument is that "elaborate usability tests are a waste of resources." Once it is found that a few people are totally confused by a home page, little is gained by watching more people suffer through the same flawed design." Jakob's law Users will anticipate what an experience will be like, based on their mental models of prior experiences on websites. When ma
https://en.wikipedia.org/wiki/Download
In computer networks, download means to receive data from a remote system, typically a server such as a web server, an FTP server, an email server, or other similar systems. This contrasts with uploading, where data is sent to a remote server. A download is a file offered for downloading or that has been downloaded, or the process of receiving such a file. Definition Downloading generally transfers entire files for local storage and later use, as contrasted with streaming, where the data is used nearly immediately, while the transmission is still in progress, and which may not be stored long-term. Websites that offer streaming media or media displayed in-browser, such as YouTube, increasingly place restrictions on the ability of users to save these materials to their computers after they have been received. Downloading in computer networks involves retrieving data from a remote system, like a web server, FTP server, or email server, unlike uploading where data is sent to a remote server. A download can refer to a file made available for retrieval or one that has been received, encompassing the entire process of obtaining such a file. Downloading is not the same as data transfer; moving or copying data between two storage devices would be data transfer, but receiving data from the Internet or BBS is downloading. Copyright Downloading media files involves the use of linking and framing Internet material, and relates to copyright law. Streaming and downloading can involve making copies of works that infringe on copyrights or other rights, and organizations running such websites may become vicariously liable for copyright infringement by causing others to do so. Open hosting servers allows people to upload files to a central server, which incurs bandwidth and hard disk space costs due to files generated with each download. Anonymous and open hosting servers make it difficult to hold hosts accountable. Taking legal action against the technologies behind unauthorized "file sharing" has proven successful for centralized networks like Napster, and untenable for decentralized networks like Gnutella or BitTorrent. The leading YouTube audio-ripping site agreed to shut down after being sued by a huge coalition of recording labels. Downloading and streaming relates to the more general usage of the Internet to facilitate copyright infringement also known as "software piracy". As overt static hosting to unauthorized copies of works (i.e. centralized networks) is often quickly and uncontroversially rebuffed, legal issues have in recent years tended to deal with the usage of dynamic web technologies (decentralized networks, trackerless BitTorrents) to circumvent the ability of copyright owners to directly engage particular distributors and consumers. Litigations in European Union In Europe, the Court of Justice of the European Union (CJEU) has ruled that it is legal to create temporary or cached copies of works (copyrighted or otherwise) online. The rul
https://en.wikipedia.org/wiki/System%20requirements
To be used efficiently, all computer software needs certain hardware components or other software resources to be present on a computer. These prerequisites are known as (computer) system requirements and are often used as a guideline as opposed to an absolute rule. Most software defines two sets of system requirements: minimum and recommended. With increasing demand for higher processing power and resources in newer versions of software, system requirements tend to increase over time. Industry analysts suggest that this trend plays a bigger part in driving upgrades to existing computer systems than technological advancements. A second meaning of the term system requirements, is a generalisation of this first definition, giving the requirements to be met in the design of a system or sub-system. Recommended system requirements Often manufacturers of games will provide the consumer with a set of requirements that are different from those that are needed to run a software. These requirements are usually called the recommended requirements. These requirements are almost always of a significantly higher level than the minimum requirements, and represent the ideal situation in which to run the software. Generally speaking, this is a better guideline than minimum system requirements in order to have a fully usable and enjoyable experience with that software. Hardware requirements The most common set of requirements defined by any operating system or software application is the physical computer resources, also known as hardware, A hardware requirements list is often accompanied by a hardware compatibility list (HCL), especially in case of operating systems. An HCL lists tested, compatible, and sometimes incompatible hardware devices for a particular operating system or application. The following sub-sections discuss the various aspects of hardware requirements. Architecture All computer operating systems are designed for a particular computer architecture. Most software applications are limited to particular operating systems running on particular architectures. Although architecture-independent operating systems and applications exist, most need to be recompiled to run on a new architecture. See also a list of common operating systems and their supporting architectures. Processing power The power of the central processing unit (CPU) is a fundamental system requirement for any software. Most software running on x86 architecture define processing power as the model and the clock speed of the CPU. Many other features of a CPU that influence its speed and power, like bus speed, cache, and MIPS are often ignored. This definition of power is often erroneous, as different makes and models of CPUs at similar clock speed often have different throughput speeds. Memory All software, when run, resides in the random access memory (RAM) of a computer. Memory requirements are defined after considering demands of the application, operating system, supp
https://en.wikipedia.org/wiki/Artificial%20neuron
An artificial neuron is a mathematical function conceived as a model of biological neurons in a neural network. Artificial neurons are the elementary units of artificial neural networks. The artificial neuron receives one or more inputs (representing excitatory postsynaptic potentials and inhibitory postsynaptic potentials at neural dendrites) and sums them to produce an output (or , representing a neuron's action potential which is transmitted along its axon). Usually, each input is separately weighted (representing the synaptic weight), and the sum is often added to a term known as a bias (loosely corresponding to the threshold potential), before being passed through a non-linear function known as an activation function or transfer function. The transfer functions usually have a sigmoid shape, but they may also take the form of other non-linear functions, piecewise linear functions, or step functions. They are also often monotonically increasing, continuous, differentiable and bounded. Non-monotonic, unbounded and oscillating activation functions with multiple zeros that outperform sigmoidal and ReLU-like activation functions on many tasks have also been recently explored. The thresholding function has inspired building logic gates referred to as threshold logic; applicable to building logic circuits resembling brain processing. For example, new devices such as memristors have been extensively used to develop such logic in recent times. The artificial neuron transfer function should not be confused with a linear system's transfer function. Simple artificial neurons, such as the McCulloch–Pitts model, are sometimes described as "caricature models", since they are intended to reflect one or more neurophysiological observations, but without regard to realism. Artificial neurons can also refer to artificial cells in neuromorphic engineering that are similar to natural physical neurons. Basic structure For a given artificial neuron k, let there be m + 1 inputs with signals x0 through xm and weights wk0 through wkm. Usually, the x0 input is assigned the value +1, which makes it a bias input with wk0 = bk. This leaves only m actual inputs to the neuron: from x1 to xm. The output of the kth neuron is: Where (phi) is the transfer function (commonly a threshold function). The output is analogous to the axon of a biological neuron, and its value propagates to the input of the next layer, through a synapse. It may also exit the system, possibly as part of an output vector. It has no learning process as such. Its transfer function weights are calculated and threshold value are predetermined. McCulloch–Pitts (MCP) neuron Depending on the specific model used they may be called a semi-linear unit, Nv neuron, binary neuron, linear threshold function, or McCulloch–Pitts (MCP) neuron. The original MCP neural network are described in detail in Chapter 3 of . In short, a MCP neuron is a restricted kind of artificial neuron, operating in discrete ti
https://en.wikipedia.org/wiki/Server%20Message%20Block
Server Message Block (SMB) is a communication protocol mainly used by Microsoft Windows equipped computers normally used to share files, printers, serial ports, and miscellaneous communications between nodes on a network. SMB implementation consists of two vaguely named Windows services: "Server" (ID: LanmanServer) and "Workstation" (ID: LanmanWorkstation). It uses NTLM or Kerberos protocols for user authentication. It also provides an authenticated inter-process communication (IPC) mechanism. SMB was originally developed in 1983 by Barry A. Feigenbaum at IBM and intended to provide shared access to files and printers across nodes on a network of systems running IBM's OS/2. In 1987, Microsoft and 3Com implemented SMB in LAN Manager for OS/2, at which time SMB used the NetBIOS service atop the NetBIOS Frames protocol as its underlying transport. Later, Microsoft implemented SMB in Windows NT 3.1 and has been updating it ever since, adapting it to work with newer underlying transports: TCP/IP and NetBT. SMB over QUIC was introduced in Windows Server 2022. In 1996, Microsoft published a version of SMB 1.0 with minor modifications under the Common Internet File System (CIFS ) moniker. CIFS was compatible with even the earliest incarnation of SMB, including LAN Manager's. It supports symbolic links, hard links, and larger file size, but none of the features of SMB 2.0 and later. Microsoft's proposal, however, remained an Internet Draft and never achieved standard status. Microsoft has since discontinued use of the CIFS moniker but continues developing SMB and making subsequent specifications publicly available. Samba is a free software reimplementation of the SMB protocol and the Microsoft extensions to it. Features Server Message Block (SMB) enables file sharing, printer sharing, network browsing, and inter-process communication (through named pipes) over a computer network. SMB serves as the basis for Microsoft's Distributed File System implementation. SMB relies on the TCP and IP protocols for transport. This combination allows file sharing over complex, interconnected networks, including the public Internet. The SMB server component uses TCP port 445. SMB originally operated on NetBIOS over IEEE 802.2 - NetBIOS Frames or NBF - and over IPX/SPX, and later on NetBIOS over TCP/IP (NetBT), but Microsoft has since deprecated these protocols. On NetBT, the server component uses three TCP or UDP ports: 137 (NETBIOS Name Service), 138 (NETBIOS Datagram Service), and 139 (NETBIOS Session Service). In Microsoft Windows, two vaguely named Windows services implement SMB. The "Server" service (ID: LanmanServer) is in charge of serving shared resources. The "Workstation" service (ID: LanmanWorkstation) maintains the computer name and helps access shared resources on other computers. SMB uses the Kerberos protocol to authenticate users against Active Directory on Windows domain networks. On simpler, peer-to-peer networks, SMB uses the NTLM protocol. Wind
https://en.wikipedia.org/wiki/Mona%20Lisa%20Overdrive
Mona Lisa Overdrive is a science fiction novel by American-Canadian writer William Gibson, published in 1988. It is the final novel of the cyberpunk Sprawl trilogy, following Neuromancer and Count Zero, taking place eight years after the events of the latter. The novel was nominated for the Nebula Award for Best Novel, the Hugo Award for Best Novel, and the Locus Award for Best Science Fiction Novel in 1989. Plot Taking place eight years after the events of Count Zero and fifteen years after Neuromancer, the story is formed from several interconnecting plot threads, and also features characters from Gibson's previous works (such as Molly Millions, the razor-fingered mercenary from Neuromancer). Thread one: concerns Mona, a teen prostitute who has a more-than-passing resemblance to famed Simstim superstar Angie Mitchell. Mona is hired by shady individuals for a "gig" which later turns out to be part of a plot to abduct Angie. Thread two: focuses on a young Japanese girl named Kumiko, daughter of a Yakuza boss sent to London to keep her safe while her father engages in a gang war with other top Yakuza leaders. In London, she is cared for by one of her father's retainers, who is also a powerful member of the London Mob. She meets Molly Millions (having altered her appearance and now calling herself "Sally Shears", in order to conceal her identity from hostile parties who are implied to be pursuing her), who takes the girl under her wing. Thread three: follows a reclusive artist named Slick Henry, who lives in a place named Factory in the Dog Solitude; a large, poisoned expanse of deserted factories and dumps, probably in southern New Jersey. Slick Henry is a convicted car thief whose punishment consisted of having his short-term memory erased every five minutes, leading to continuous confusion and dissociation. Following the end of his sentence, he spends his days creating large robotic sculptures and periodically suffers episodes of time loss, returning to consciousness afterward with no memory of what he did during the blackout. He is hired by an acquaintance to look after the comatose "Count" (Bobby Newmark from the second novel, Count Zero, who has hooked himself into a super-capacity cyber-bio harddrive called an Aleph). A theoretical "Aleph" would have the RAM capacity to literally contain all of reality, enough that a memory construct of a person would contain the complete personality of the individual and allow it to learn, grow and act independently. The final plotline follows Angela Mitchell, famous simstim star, and the girl from the second Sprawl novel Count Zero. Angie, thanks to brain manipulations by her father when she was a child, has always had the ability to access cyberspace directly (without a cyberspace deck), but drugs provided by her production company Sense/Net have severely impeded this ability. The plot culminates when Angie and Bobby "upload" their consciousness into the Aleph, on the verge of visiting an alien ar
https://en.wikipedia.org/wiki/MIP
MIP may refer to: Science Mixed integer programming, linear programming where some variables are constrained to be integers Minimum Ionizing Particle, in particle physics Maximum intensity projection, a computer visualization method Molecularly imprinted polymer, polymers processed using the molecular imprinting technique with affinity to a chosen 'template molecule' Moon Impact Probe, of the Indian lunar satellite Chandrayaan-1 Model Intercomparison Project; see Coupled model intercomparison project MIP, an interactive proof system complexity class; see Interactive proof system Mars ISPP Precursor, a test payload intended to be flown on the cancelled Mars Surveyor 2001 Lander Biology Macrophage inflammatory protein, in biology Maximum Inspiratory Pressure, the maximum inspiratory pressure is the highest atmospheric pressure developed during inspiration against an occluded airway Methylation induced premeiotically Mid-inguinal point Molecular Inversion Probe Mitochondrial intermediate peptidase, an enzyme MIP (gene), a gene in humans Microprotein, a small protein encoded from small open reading frames Arts and entertainment Mipcom, a TV and entertainment market museum in progress, a non-commercial art association, based in Vienna Most Improved Player (disambiguation) Business Mortgage insurance premium Managers in Partnership (MIP), a British trade union for healthcare managers Managing Intellectual Property, a monthly magazine specialized in intellectual property law and business Master in Ingegneria della Produzione (MIP), Politecnico di Milano School of Management Mint in Package; a collectors' abbreviation; see Mint condition Government and politics Macroeconomic Imbalance Procedure, a European Union economic governance procedure Minor In Possession in U.S. civil law Multilateral Interoperability Programme, a consortium of 29 NATO and Non-NATO nations Popular Independent Movement (), a Luxembourgian political party in the 1960s Technology Male Iron Pipe, a plumbing pipe connection to an FIP (Female Iron Pipe); see National Pipe Thread Mega-frame Initialization Packet, an MPEG-2 Transport Stream packet used for synchronization in DVB-T single-frequency networks Memory in Pixel, a type of liquid crystal display Mipmap (multum in parvo, much in little space), a sequences of images, each of which is a progressively lower resolution than the previous Mobile IP, an IP protocol extension to provide mobility in the Internet See also MIPS (disambiguation)
https://en.wikipedia.org/wiki/CLX
The acronym CLX can refer to a number of things: 160 in Roman numerals Cargolux, an airline using the ICAO code CLX Clorox stock ticker CLX (Common Lisp), a Common Lisp computer library CLX Communications, a telecommunications and cloud communications platform as a service company, based in Stockholm, Sweden Component Library for Cross Platform (CLX), a cross-platform visual component-based framework