source
stringlengths
32
199
text
stringlengths
26
3k
https://en.wikipedia.org/wiki/List%20comprehension
A list comprehension is a syntactic construct available in some programming languages for creating a list based on existing lists. It follows the form of the mathematical set-builder notation (set comprehension) as distinct from the use of map and filter functions. Overview Consider the following example in set-builder notation. or often This can be read, " is the set of all numbers "2 times " SUCH THAT is an ELEMENT or MEMBER of the set of natural numbers (), AND squared is greater than ." The smallest natural number, x = 1, fails to satisfy the condition x2>3 (the condition 12>3 is false) so 2 ·1 is not included in S. The next natural number, 2, does satisfy the condition (22>3) as does every other natural number. Thus x consists of 2, 3, 4, 5... Since the set consists of all numbers "2 times x" it is given by S = {4, 6, 8, 10,...}. S is, in other words, the set of all even numbers greater than 2. In this annotated version of the example: is the variable representing members of an input set. represents the input set, which in this example is the set of natural numbers is a predicate expression acting as a filter on members of the input set. is an output expression producing members of the new set from members of the input set that satisfy the predicate expression. braces indicate that the result is a set the vertical bar is read as "SUCH THAT". The bar and the colon ":" are used interchangeably. commas separate the predicates and can be read as "AND". A list comprehension has the same syntactic components to represent generation of a list in order from an input list or iterator: A variable representing members of an input list. An input list (or iterator). An optional predicate expression. And an output expression producing members of the output list from members of the input iterable that satisfy the predicate. The order of generation of members of the output list is based on the order of items in the input. In Haskell's list comprehension syntax, this set-builder construct would be written similarly, as: s = [ 2*x | x <- [0..], x^2 > 3 ] Here, the list [0..] represents , x^2>3 represents the predicate, and 2*x represents the output expression. List comprehensions give results in a defined order (unlike the members of sets); and list comprehensions may generate the members of a list in order, rather than produce the entirety of the list thus allowing, for example, the previous Haskell definition of the members of an infinite list. History The existence of related constructs predates the use of the term "List Comprehension". The SETL programming language (1969) has a set formation construct which is similar to list comprehensions. E.g., this code prints all prime numbers from 2 to : print([n in [2..N] | forall m in {2..n - 1} | n mod m > 0]); The computer algebra system AXIOM (1973) has a similar construct that processes streams. The first use of the term "comprehension" for such constructs was in Rod Burstall
https://en.wikipedia.org/wiki/DNO
DNO may refer to: The Danish Nurses' Organization, a trade union for nurses in Denmark Director of Naval Ordnance, British Admiralty administration Distribution network operator, companies licensed to distribute electricity in Great Britain Dno (air base), a former air base in Russia located 4 km south of Dno DNO ASA, a Norwegian oil company Drang nach Osten, a term coined in the 19th century to designate German expansion into Slavic lands Dutch National Opera, based in Amsterdam, the Netherlands
https://en.wikipedia.org/wiki/Static%20VAR%20compensator
A static VAR compensator (SVC) is a set of electrical devices for providing fast-acting reactive power on high-voltage electricity transmission networks. SVCs are part of the flexible AC transmission system device family, regulating voltage, power factor, harmonics and stabilizing the system. A static VAR compensator has no significant moving parts (other than internal switchgear). Prior to the invention of the SVC, power factor compensation was the preserve of large rotating machines such as synchronous condensers or switched capacitor banks. The SVC is an automated impedance matching device, designed to bring the system closer to unity power factor. SVCs are used in two main situations: Connected to the power system, to regulate the transmission voltage ("transmission SVC") Connected near large industrial loads, to improve power quality ("industrial SVC") In transmission applications, the SVC is used to regulate the grid voltage. If the power system's reactive load is capacitive (leading), the SVC will use thyristor controlled reactors to consume VARs from the system, lowering the system voltage. Under inductive (lagging) conditions, the capacitor banks are automatically switched in, thus providing a higher system voltage. By connecting the thyristor-controlled reactor, which is continuously variable, along with a capacitor bank step, the net result is continuously variable leading or lagging power. In industrial applications, SVCs are typically placed near high and rapidly varying loads, such as arc furnaces, where they can smooth flicker voltage. Description Principle Typically, an SVC comprises one or more banks of fixed or switched shunt capacitors or reactors, of which at least one bank is switched by thyristors. Elements which may be used to make an SVC typically include: Thyristor-controlled reactor (TCR), where the reactor may be air- or iron-cored Thyristor-switched capacitor (TSC) Harmonic filter(s) Mechanically switched capacitors or reactors (switched by a circuit breaker) By means of phase angle modulation switched by the thyristors, the reactor may be variably switched into the circuit and so provide a continuously variable VAR injection (or absorption) to the electrical network. In this configuration, coarse voltage control is provided by the capacitors; the thyristor-controlled reactor is to provide smooth control. Smoother control and more flexibility can be provided with thyristor-controlled capacitor switching. The thyristors are electronically controlled. Thyristors, like all semiconductors, generate heat and deionized water is commonly used to cool them. Chopping reactive load into the circuit in this manner injects undesirable odd-order harmonics and so banks of high-power filters are usually provided to smooth the waveform. Since the filters themselves are capacitive, they also export MVARs to the power system. More complex arrangements are practical where precise voltage regulation is required. Voltage regu
https://en.wikipedia.org/wiki/William%20Wulf
William Allan Wulf (December 8, 1939 – March 10, 2023) was an American computer scientist notable for his work in programming languages and compilers. Early life and education Born in Chicago, Wulf attended the University of Illinois Urbana–Champaign, receiving a Bachelor of Science (B.S.) in engineering physics in 1961 and an Master of Science (M.S.) in electrical engineering in 1963. He then achieved the first Doctor of Philosophy (Ph.D.) in computer science from the University of Virginia in 1968. Career In 1970, while at Carnegie Mellon University (CMU), he designed the BLISS programming language and developed a groundbreaking optimizing compiler for it. From 1971 to 1975, as part of CMUs C.mmp project, he worked on an operating system (OS) microkernel named Hydra which is capability-based, object-oriented, and designed to support a wide range of possible OSs to run on it. With his wife Anita K. Jones, Wulf was a founder and vice president of Tartan Laboratories, a compiler technology company, in 1981. Wulf served as president of the National Academy of Engineering from 1996 to 2007. He chaired the Computer Science and Telecommunications Board of the National Research Council from 1992 to 1996. Prior to that, he served as Assistant Director of the US National Science Foundation's Computer and Information Science and Engineering (CISE) directorate from 1988-1990. During this time, he played a major role in making the NSFnet computer network (as it was known at that time) available to the public as the Internet, for which he received the ACM Policy Award in 2017. He served on the Council of the ACM, on the board of directors of CRDF Global, and was a reviewing editor of Science. In 1994 he was inducted as a Fellow of the ACM. In 2007 Wulf was awarded the honor of delivering the prestigious Charles P. Steinmetz Lecture at Union College. He was elected to the American Philosophical Society that same year. Wulf's research also included computer architecture, computer security, and hardware-software codesign. Wulf ended his career at the University of Virginia by resigning on Tuesday, June 19, 2012, in protest of the forced resignation of former President Teresa A. Sullivan, in what he called, "the worst example of corporate governance I have ever seen. After widespread challenges from the faculty, student body, alumni, and the national academic community; and in the face of a direct threat from the Governor of Virginia that he would replace the entire board if they did not resolve the conflict, Sullivan was unanimously rehired some two weeks later. Personal life and death Wulf was married to Anita K. Jones, an Emeritus Professor of Computer Science at the University of Virginia. William Wulf died in Charlottesville, Virginia, on March 10, 2023, at the age of 83. Publications Wulf, W. A., "Programming Without the GOTO", Proceedings of the Internationale Federation of Information Processing, Ljubljana, Yugoslavia, August 1971. Wulf,
https://en.wikipedia.org/wiki/Rorschach%20test
The Rorschach test is a projective psychological test in which subjects' perceptions of inkblots are recorded and then analyzed using psychological interpretation, complex algorithms, or both. Some psychologists use this test to examine a person's personality characteristics and emotional functioning. It has been employed to detect underlying thought disorder, especially in cases where patients are reluctant to describe their thinking processes openly. The test is named after its creator, Swiss psychologist Hermann Rorschach. The Rorschach can be thought of as a psychometric examination of pareidolia, the active pattern of perceiving objects, shapes, or scenery as meaningful things to the observer's experience, the most common being faces or other pattern of forms that are not present at the time of the observation. In the 1960s, the Rorschach was the most widely used projective test. Although the Exner Scoring System (developed since the 1960s) claims to have addressed and often refuted many criticisms of the original testing system with an extensive body of research, some researchers continue to raise questions. The areas of dispute include the objectivity of testers, inter-rater reliability, the verifiability and general validity of the test, bias of the test's pathology scales towards greater numbers of responses, the limited number of psychological conditions which it accurately diagnoses, the inability to replicate the test's norms, its use in court-ordered evaluations, and the proliferation of the ten inkblot images, potentially invalidating the test for those who have been exposed to them. History The use of interpreting "ambiguous designs" to assess an individual's personality is an idea that goes back to Leonardo da Vinci and Botticelli. Interpretation of inkblots was central to a game, Gobolinks, from the late 19th century. The Rorschach test, however, was the first systematic approach of this kind. After studying 300 mental patients and 100 control subjects, in 1921 Hermann Rorschach wrote his book Psychodiagnostik, which was to form the basis of the inkblot test. After experimenting with several hundred inkblots which he drew himself, he selected a set of ten for their diagnostic value. Although he had served as Vice President of the Swiss Psychoanalytic Society, Rorschach had difficulty in publishing the book and it attracted little attention when it first appeared. Rorschach died the following year. It has been suggested that Rorschach's use of inkblots may have been influenced by German doctor Justinus Kerner who, in 1857, had published a popular book of poems, each of which was inspired by an accidental inkblot. French psychologist Alfred Binet had also experimented with inkblots as a creativity test, and, after the turn of the century, psychological experiments where inkblots were utilized multiplied, with aims such as studying imagination and consciousness. In 1927, the newly founded publishing house purchased Rorschac
https://en.wikipedia.org/wiki/Subhash%20Kak
Subhash Kak is an Indian-American computer scientist and historical revisionist. He is the Regents Professor of Computer Science Department at Oklahoma State University–Stillwater, an honorary visiting professor of engineering at Jawaharlal Nehru University, and a member of the Indian Prime Minister's Science, Technology and Innovation Advisory Council (PM-STIAC). Kak has published on the history of science, the philosophy of science, ancient astronomy, and the history of mathematics. Kak has also published on archaeoastronomy, and advocated the idea of Indigenous Aryans. Many scholars have rejected his theories on these topics in entirety, and his writings have been heavily criticized. In 2019, the Government of India awarded him the Padma Shri, the fourth highest civilian award in India, for his contributions on the history of mathematics, science, ancient astronomy and philosophy of science. Early life and education Kak was born to Ram Nath Kak, a government veterinary doctor and Sarojini Kak in Srinagar, India. His brother is the computer scientist Avinash Kak and his sister is the literary theorist Jaishree Odin. Kak received a Bachelor of Engineering degree from the Regional Engineering College, Srinagar (now the National Institute of Technology, Srinagar) and a Ph.D. from the Indian Institute of Technology Delhi in 1970. Academic career During 1975–1976, Kak was a visiting faculty at Imperial College, London, and a guest researcher at Bell Laboratories, Murray Hill. In 1977, he was a visiting researcher at the Tata Institute of Fundamental Research, Bombay. In 1979, he joined Louisiana State University, Baton Rouge, where he was appointed the Donald C. and Elaine T. Delaune Distinguished Professor of Electrical and Computer Engineering. In 2007, he joined the Computer Science department at Oklahoma State University–Stillwater. Kak proposed an efficient three-layer feed-forward neural network architecture and developed four corner classification algorithms for training it. Despite being criticized for scalability issues; it gained the attention of the electronic hardware community. Kak has argued that there are limits to artificial intelligence and that it cannot match biological intelligence. Kak has been critical of the generalization of the quantum computing to commercial scale; he argues error correction is a significant challenge for scalability although it's fundamental to multi-purpose computing. Kak is the Regents Professor of Computer Science Department at Oklahoma State University–Stillwater and an honorary visiting professor of engineering at Jawaharlal Nehru University. He is also an honorary visiting professor of media studies at Jawaharlal Nehru University. On 28 August 2018, he was appointed member of the Prime Minister's Science, Technology and Innovation Advisory Council (PM-STIAC) in India. Indology Indigenous Aryanism Kak primarily advocates for an autochthonous origin of the Indo-Aryans from Punjab ("Indigenous
https://en.wikipedia.org/wiki/Computer%20Fraud%20and%20Abuse%20Act
The Computer Fraud and Abuse Act of 1986 (CFAA) is a United States cybersecurity bill that was enacted in 1986 as an amendment to existing computer fraud law (), which had been included in the Comprehensive Crime Control Act of 1984. Prior to computer-specific criminal laws, computer crimes were prosecuted as mail and wire fraud, but the applying law was often insufficient. The original 1984 bill was enacted in response to concern that computer-related crimes might go unpunished. The House Committee Report to the original computer crime bill characterized the 1983 techno-thriller film WarGames—in which a young teenager (played by Matthew Broderick) from Seattle breaks into a U.S. military supercomputer programmed to predict possible outcomes of nuclear war and unwittingly almost starts World War III—as "a realistic representation of the automatic dialing and access capabilities of the personal computer." The CFAA was written to extend existing tort law to intangible property, while, in theory, limiting federal jurisdiction to cases "with a compelling federal interest—i.e., where computers of the federal government or certain financial institutions are involved or where the crime itself is interstate in nature", but its broad definitions have spilled over into contract law (see "Protected Computer", below). In addition to amending a number of the provisions in the original section 1030, the CFAA also criminalized additional computer-related acts. Provisions addressed the distribution of malicious code and denial-of-service attacks. Congress also included in the CFAA a provision criminalizing trafficking in passwords and similar items. Since then, the Act has been amended a number of times—in 1989, 1994, 1996, in 2001 by the USA PATRIOT Act, 2002, and in 2008 by the Identity Theft Enforcement and Restitution Act. With each amendment of the law, the types of conduct that fell within its reach were extended. In January 2015, President Barack Obama proposed expanding the CFAA and the RICO Act in his Modernizing Law Enforcement Authorities to Combat Cyber Crime proposal. DEF CON organizer and Cloudflare researcher Marc Rogers, Senator Ron Wyden, and Representative Zoe Lofgren stated opposition to this on the grounds it would make many regular Internet activities illegal, and moved further away from what they were trying to accomplish with Aaron's Law. Protected computers The only computers, in theory, covered by the CFAA are defined as "protected computers". They are defined under section to mean a computer: exclusively for the use of a financial institution or the United States Government, or any computer, when the conduct constituting the offense affects the computer's use by or for the financial institution or the government; or which is used in or affecting interstate or foreign commerce or communication, including a computer located outside the United States that is used in a manner that affects interstate or foreign commerce or communica
https://en.wikipedia.org/wiki/Second%20audio%20program
Second audio program (SAP), also known as secondary audio programming, is an auxiliary audio channel for analog television that can be broadcast or transmitted both over-the-air and by cable television. Used mostly for audio description or other languages, SAP is part of the multichannel television sound (MTS) standard originally set by the National Television Systems Committee (NTSC) in 1984 in the United States. The NTSC video format and MTS are also used in Canada and Mexico. Usage SAP is often used to provide audio tracks in languages other than the native language included in the program. In the United States, this is sometimes used for Spanish-language audio (especially during sports telecasts), often leading to the function being referred to facetiously as the "Spanish audio program". Likewise, some Spanish-language programs may, in rare cases, offer English on SAP. Some stations may relay NOAA Weather Radio services, or, particularly in the case of PBS stations, a local National Public Radio (NPR) sister station, on the audio channel when SAP is not being used. In Canada, parliamentary and public affairs channel CPAC similarly uses SAP to carry both English and French-language audio. SAP is also a means of distribution for audio description of programs for the visually impaired. Under the Twenty-First Century Communications and Video Accessibility Act of 2010, top U.S. television networks and cable networks have been gradually required to broadcast quotas of audio described programming per-quarter, Since May 26, 2015, broadcasters have been required under the Act to provide dictations on SAP of any "emergency information" displayed in a textual format outside of the Emergency Alert System and newscasts. Frequencies MTS features, including stereo and SAP, travel on subcarriers of the video carrier, much like color for television. It is not carried on the audio carrier in the manner of stereo sound for an FM radio broadcast, however, as it only has a frequency deviation of ±25 kHz, whereas regular FM broadcasting has a deviation of ±75 kHz. The SAP subcarrier is located at 78.670 kHz, which is five times the 15.734 kHz corresponding to the MTS pilot signal. In turn, the MTS pilot is locked to the horizontal sync frequency of the video carrier for stability. The SAP channel contains mono audio which has been dbx-encoded for noise reduction, to improve the signal-to-noise ratio. The SAP audio has a bandpass from 60 Hz to 12 kHz, which is less than the "regular" audio channel which runs from 50 Hz to 15 kHz. Though not technically a SAP channel, television stations can also broadcast a "PRO" (professional) audio subcarrier which is used to communicate with station personnel, particularly those engaged in electronic news gathering. This one-way audio channel allows individuals at the television station to send messages to people located away from the station, and is frequently employed during on-location newscasts as the foldback channel to
https://en.wikipedia.org/wiki/BBC%20America
BBC America is an American basic cable network that is jointly owned by BBC Studios and AMC Networks. The channel primarily airs sci-fi and action series and films, as well as selected programs from the BBC (such as its nature documentary series). Unlike the BBC's domestic channels in the United Kingdom, BBC America does not receive funding from the British licence fee (which is the principal funding for the BBC's channels within the United Kingdom), as the BBC cannot fund any of its channels that are available outside the United Kingdom. Consequently, BBC America operates as a commercial-supported channel and accepts traditional advertising. It is also funded by television subscription fees. As of September 2018, BBC America is available to about 80.9 million television households (87.8% of pay television customers) in the United States. History BBC America was launched on March 29, 1998, presenting a mixture of comedy, drama and lifestyle programs from BBC Television and other British television broadcasters including ITV and Channel 4. In the channel's early days, it focused on repeats of popular lifestyle shows such as Changing Rooms and Ground Force. BBC America's head of television programming later stated that the channel needed to establish a niche, since non-British viewers found the lifestyle shows appealing. Most of the newer programs appeared as part of BBC America's evening schedule. The network removed the British soap opera EastEnders from its schedule in 2003 due to low ratings; however, the program's removal from BBC America provoked complaints from viewers, which caught media attention. After CEO Paul Lee was appointed president of ABC Family in 2009, the network appointed Bill Hilary from Comedy Central to serve as its chief executive officer. Hilary appointed Kathryn Mitchell to the new position of general manager. Under Hilary's tenure, BBC America was restructured; it moved its main offices to New York City and had its programming budget increase substantially. The channel was led by Garth Ancier, who served as the president of BBC Worldwide Americas from February 2007 until 2010, when he was succeeded by former MTV Networks executive Herb Scannell. Meanwhile, Perry Simon serves as general manager. In 2014, AMC Networks acquired a 49.9% equity stake in BBC America for $200 million, and replaced Discovery Communications as its managing partner. As part of the deal, AMC Networks also took on the responsibility of negotiating U.S. broadcast and advertising sales for the BBC World News channel. AMC Networks CEO Josh Sapan stated that the deal gave his company "a powerful collection of networks that are among the most critically acclaimed, with distinct dramas and other potent content that creates a deep connection with viewers", while BBC Worldwide CEO Tim Davie considered it "an opportunity to grow the creative quality and ambition from an already high base." On Sunday, April 25, 2021, BBC America unveiled a redesigned l
https://en.wikipedia.org/wiki/Alopex
Alopex may refer to: Alopex lagopus, a taxonomic synonym for the Arctic fox, Vulpes lagopus ALOPEX a correlation-based machine learning algorithm Alopex (Teenage Mutant Ninja Turtles), a character in the Teenage Mutant Ninja Turtles franchise Alopex () ancient Greek for fox
https://en.wikipedia.org/wiki/Shelf%20%28computing%29
The Shelf is an interface feature in NeXTSTEP and OPENSTEP, and is used as a repository to store links to commonly used files, directories and programs, and as a temporary "holding" place to move/copy files and directories around in the file system hierarchy. In macOS, items may be dragged onto the sidebar area of the Finder, but these do not behave as placeholders and cannot be manipulated in the below manner. The dynamics of the Shelf in file system operations can be illustrated by comparison with the metaphor used in the Microsoft Windows operating systems. In order to move a file the following steps may be taken: the window containing the source folder is opened the window containing the destination folder is opened the desired file in the source directory is dragged to the destination folder With the NeXT operating systems, in addition to moving files by dragging them from window to window, the following method can be used: the user navigates to the source directory the file is dragged to the Shelf the user navigates to the destination directory the file is dragged from the Shelf to the destination directory Note that the file, when dragged to the Shelf, has not moved anywhere and is not changed in any way. The Shelf icon is merely a placeholder for the file. In moving the placeholder off the shelf, the actual action occurs. The NeXT functionality builds upon this concept by allowing the destination directory to be put on the Shelf as well, and the file can be merely dragged to the destination directory icon. The process is similar to the Microsoft Windows functionality of copying or cutting file system objects (a file or files, a folder or folders, or a combination of both) to the clipboard; the objects are not copied or removed from their original location until the paste operation to the new location is completed. The Shelf concept, though older, is more powerful in that the file system objects, their sources and destinations are persistent and available as long as they are on the Shelf (in the Windows cut, copy, and paste metaphor the objects and locations persist until one copy/move operation is complete or until something else is placed in the clipboard). Since Shelf icons are 'placeholders' of sorts, icons can be put on the Shelf representing commonly used directories, and commonly used programs can be put on the Shelf as well. The NeXTSTEP and OPENSTEP file management application (called FileViewer and run by the Workspace Manager) also allowed users to have different shelves associated with particular directories. Users simply opened a new browser rooted in a particular subdirectory, and that browser window would show the corresponding shelf, allowing users to have many different shelves based on whatever folder hierarchy they happened to be using to organize their files. See also Miller columns Clipboard (computing) References NeXT Graphical user interface elements
https://en.wikipedia.org/wiki/Predictive%20dialer
A predictive dialer dials a list of telephone numbers and connects answered dials to people making calls, often referred to as agents. Predictive dialers use statistical algorithms to minimize the time that agents spend waiting between conversations, while minimizing the occurrence of someone answering when no agent is available. When dialing numbers one at a time, there are two sources of delay. First, only some fraction of dials are answered; for example, if 1 out of 3 dials are answered, a predictive dialer might dial 3 lines every time an agent becomes available. Second, even dials that are answered take some time before being picked up. If it typically takes 10 seconds for someone to pick up, and conversations typically last 90 seconds, a predictive dialer might start dialing at 80 seconds. A predictive dialer does so by discarding all tones and only passing "Hellos" from the lead to the agent. Dialing one number at a time, only when an agent is available, typically keeps agents utilized for 40 minutes per hour (33% idle time). Predictive dialing can increase utilization to 57 minutes per hour (5% idle time). Predictive dialers may be standalone hardware devices, cloud-based, or they may be integrated in software with call center or contact center platforms. A cloud-based predictive dialer provides the functionalities of a traditional predictive dialer without requiring installation on the local device and reduces the dependency on the hardware. The dialer can often also perform less aggressive dialing modes such as, power, progressive, or preview dialing. Call progress analysis Call Progress Analysis (CPA), also called Call Progress Detection (CPD), is a generic term for signal processing algorithms that operate on audio during call setup. The goal of CPA is to determine the nature of the callee or the outcome of call setup to an external network (traditional or IP). Specifically, when a call or session is being established, the caller or initiator is interested in knowing if someone answered, if the line is busy, etc. When the caller is an automated application, such as an automated dialer or message broadcasting system, CPA algorithms are used to perform the classification automatically. Modern CPA methods in combination with automated outbound dialing applications, ensure fast and accurate automated call classification, which automatically translate in better efficiency of agents and higher quality customer interactions. Traditional Call Progress Analysis implementations rely on simple rule-based algorithms which provide sub-optimal accuracy of speed of processing, leading to significant inefficiencies in operations and issues in meeting compliance from new government regulations. Using statistical models based on neural networks to represent the potential outcomes of an outbound call attempt, companies can meet and exceed these regulations. This approach identifies much better the call patterns that represent call progress even
https://en.wikipedia.org/wiki/Stored%20procedure
A stored procedure (also termed proc, storp, sproc, StoPro, StoredProc, StoreProc, sp, or SP) is a subroutine available to applications that access a relational database management system (RDBMS). Such procedures are stored in the database data dictionary. Uses for stored procedures include data-validation (integrated into the database) or access-control mechanisms. Furthermore, stored procedures can consolidate and centralize logic that was originally implemented in applications. To save time and memory, extensive or complex processing that requires execution of several SQL statements can be saved into stored procedures, and all applications call the procedures. One can use nested stored procedures by executing one stored procedure from within another. Stored procedures may return result sets, i.e., the results of a SELECT statement. Such result sets can be processed using cursors, by other stored procedures, by associating a result-set locator, or by applications. Stored procedures may also contain declared variables for processing data and cursors that allow it to loop through multiple rows in a table. Stored-procedure flow-control statements typically include IF, WHILE, LOOP, REPEAT, and CASE statements, and more. Stored procedures can receive variables, return results or modify variables and return them, depending on how and where the variable is declared. Implementation Stored procedures are similar to user-defined functions (UDFs). The major difference is that UDFs can be used like any other expression within SQL statements, whereas stored procedures must be invoked using the CALL statement. CALL procedure(...) or EXECUTE procedure(...) The exact and correct implementation of stored procedures varies from one database system to the other. Most major database vendors support them in some form. Depending on the database system, stored procedures can be implemented in a variety of programming languages, for example SQL, Java, C, or C++. Stored procedures written in non-SQL languages may or may not execute SQL statements themselves. The increasing adoption of stored procedures led to the introduction of procedural elements to the SQL language in the SQL:1999 and SQL:2003 standards in the part SQL/PSM. That made SQL an imperative programming language. Most database systems offer proprietary and vendor-specific extensions, exceeding SQL/PSM. A standard specification for Java stored procedures exists as well as SQL/JRT. Comparison with static SQL Overhead Because stored procedure statements are stored directly in the database, they may remove all or part of the compiling overhead that is typically needed in situations where software applications send inline (dynamic) SQL queries to a database. (However, most database systems implement statement caches and other methods to avoid repetitively compiling dynamic SQL statements.) Also, while they avoid some pre-compiled SQL, statements add to the complexity of creating an optimal executio
https://en.wikipedia.org/wiki/Principles%20of%20Compiler%20Design
Principles of Compiler Design, by Alfred Aho and Jeffrey Ullman, is a classic textbook on compilers for computer programming languages. Both of the authors won the 2020 Turing award for their work on compilers. It is often called the "green dragon book" and its cover depicts a knight and a dragon in battle; the dragon is green, and labeled "Complexity of Compiler Design", while the knight wields a lance and a shield labeled "LALR parser generator" and "Syntax Directed Translation" respectively, and rides a horse labeled "Data Flow Analysis". The book may be called the "green dragon book" to distinguish it from its successor, Aho, Sethi & Ullman's Compilers: Principles, Techniques, and Tools, which is the "red dragon book". The second edition of Compilers: Principles, Techniques, and Tools added a fourth author, Monica S. Lam, and the dragon became purple; hence becoming the "purple dragon book." The book also contains the entire code for making a compiler. The back cover offers the original inspiration of the cover design: The dragon is replaced by windmills, and the knight is Don Quixote. The book was published by Addison-Wesley, . The acknowledgments mention that the book was entirely typeset at Bell Labs using troff on the Unix operating system, little of which had, at that time, been seen outside the Laboratories. References 1977 non-fiction books Compiler construction Computer science books Addison-Wesley books Engineering textbooks
https://en.wikipedia.org/wiki/Computer%20mathematics
Computer mathematics may refer to: Automated theorem proving, the proving of mathematical theorems by a computer program Symbolic computation, the study and development of algorithms and software for manipulating mathematical expressions and other mathematical objects Computational science, constructing numerical solutions and using computers to analyze and solve scientific and engineering problems Theoretical computer science, collection of topics of computer science and mathematics that focuses on the more abstract and mathematical aspects of computing
https://en.wikipedia.org/wiki/Open-source%20software
Open-source software (OSS) is computer software that is released under a license in which the copyright holder grants users the rights to use, study, change, and distribute the software and its source code to anyone and for any purpose. Open-source software may be developed in a collaborative, public manner. Open-source software is a prominent example of open collaboration, meaning any capable user is able to participate online in development, making the number of possible contributors indefinite. The ability to examine the code facilitates public trust in the software. Open-source software development can bring in diverse perspectives beyond those of a single company. A 2008 report by the Standish Group stated that adoption of open-source software models has resulted in savings of about $60 billion per year for consumers. Open-source code can be used for studying and allows capable end users to adapt software to their personal needs in a similar way user scripts and custom style sheets allow for web sites, and eventually publish the modification as a fork for users with similar preferences, and directly submit possible improvements as pull requests. History In the early days of computing, such as the 1950s and into the 1960s, programmers and developers shared software to learn from each other and evolve the field of computing. For example, Unix included the operating system source code for users. Eventually, the open-source notion moved to the wayside of commercialization of software in the years 1970–1980. However, academics still often developed software collaboratively. Examples are Donald Knuth in 1979 with the TeX typesetting system and Richard Stallman in 1983 with the GNU operating system. In 1997, Eric Raymond published The Cathedral and the Bazaar, a reflective analysis of the hacker community and free-software principles. The paper received significant attention in early 1998, and was one factor in motivating Netscape Communications Corporation to release their popular Netscape Communicator Internet suite as free software. This source code subsequently became the basis behind SeaMonkey, Mozilla Firefox, Thunderbird and KompoZer. Netscape's act prompted Raymond and others to look into how to bring the Free Software Foundation's free software ideas and perceived benefits to the commercial software industry. They concluded that FSF's social activism was not appealing to companies like Netscape, and looked for a way to rebrand the free software movement to emphasize the business potential of sharing and collaborating on software source code. The new term they chose was "open source", which was soon adopted by Bruce Perens, publisher Tim O'Reilly, Linus Torvalds, and others. The Open Source Initiative was founded in February 1998 to encourage the use of the new term and evangelize open-source principles. While the Open Source Initiative sought to encourage the use of the new term and evangelize the principles it adhered to, commercial
https://en.wikipedia.org/wiki/PSK31
PSK31 or "Phase Shift Keying, 31 Baud", also BPSK31 and QPSK31, is a popular computer-sound card-generated radioteletype mode, used primarily by amateur radio operators to conduct real-time keyboard-to-keyboard chat, most often using frequencies in the high frequency amateur radio bands (near-shortwave). PSK31 is distinguished from other digital modes in that it is specifically tuned to have a data rate close to typing speed, and has an extremely narrow bandwidth, allowing many conversations in the same bandwidth as a single voice channel. This narrow bandwidth makes better use of the RF energy in a very narrow space thus allowing relatively low-power equipment (5 watts) to communicate globally using the same skywave propagation used by shortwave radio stations. History PSK31 was developed and named by English amateur radio operator Peter Martinez (call sign G3PLX) and introduced to the wider amateur radio community in December 1998. The 31 baud BPSK modulation system used in PSK31 was introduced by Pawel Jalocha (SP9VRC) in his SLOWBPSK program written for Motorola's EVM radio. Instead of the traditional frequency-shift keying, the information is transmitted by patterns of polarity-reversals (sometimes called 180-degree phase shifts). PSK31 was enthusiastically received, and its usage spread rapidly worldwide, lending a new popularity and tone to the on-air conduct of digital communications. Due to the efficiency of the mode, it became, and still remains, especially popular with operators whose circumstances do not permit the installation of large antenna systems, the use of high power, or both. Use and implementation A PSK31 operator typically uses a single-sideband (SSB) transceiver connected to the sound card of a computer running PSK31 software. When the operator enters a message for transmission, the software produces an audio tone that sounds, to the human ear, like a continuous whistle with a slight warble. This sound is then fed through either a microphone jack (using an intermediate resistive attenuator to reduce the sound card's output power to microphone levels) or an auxiliary connection into the transceiver, from which it is transmitted. From the perspective of the transmitter, the sound amounts to little more than somebody whistling into the microphone. However, the software rapidly shifts the phase of the audio signal between two states (hence the name "phase-shift keying"), forming the character codes. These phase shifts serve the same function as the two tones used in traditional RTTY and similar systems. To decode PSK31, the audio whistle received from the transceiver's headphone output is fed into a computer sound card's audio input, and software decodes it. The software displays the decoded text. Because PSK31 was developed for use through a computer's sound card, many programs have since been created to use the same technology for other modes, such as RTTY, Hellschreiber, and Olivia MFSK. So, once it has been set
https://en.wikipedia.org/wiki/Uptime
Uptime is a measure of system reliability, expressed as the percentage of time a machine, typically a computer, has been working and available. Uptime is the opposite of downtime. It is often used as a measure of computer operating system reliability or stability, in that this time represents the time a computer can be left unattended without crashing, or needing to be rebooted for administrative or maintenance purposes. Conversely, long uptime may indicate negligence, because some critical updates can require reboots on some platforms. Records In 2005, Novell reported a server with a 6-year uptime. Although that might sound unusual, that is actually common when servers are maintained under an industrial context and host critical applications such as banking systems. Netcraft maintains the uptime records for many thousands of web hosting computers. A server running Novell NetWare has been reported to have been shut down after 16 years of uptime due to a failing hard disk. A Cisco router has been reported to have been running continuously for 21 years as of 2018. As of April, 11, 2023, the uptime has increased to 26 years, 25 weeks, 1 day, 1 hour, and 8 minutes. Determining system uptime Microsoft Windows Windows Task Manager Some versions of Microsoft Windows include an uptime field in Windows Task Manager, under the "Performance" tab. The format is D:HH:MM:SS (days, hours, minutes, seconds). systeminfo The output of the systeminfo command includes a "System Up Time" or "System Boot Time" field. C:\>systeminfo | findstr "Time:" System Up Time: 0 days, 8 hours, 7 minutes, 19 seconds The exact text and format is dependent on the language and locale. The time given by systeminfo is not reliable. It does not take into account time spent in sleep or hibernation. Thus, the boot time will drift forward every time the computer sleeps or hibernates. NET command The NET command with its STATISTICS sub-command provides the date and time the computer started, for both the NET STATISTICS WORKSTATION and NET STATISTICS SERVER variants. The command NET STATS SRV is shorthand for NET STATISTICS SERVER. The exact text and date format is dependent on the configured language and locale. C:\>NET STATISTICS WORKSTATION | findstr "since" Statistics since 8/31/2009 8:52:29 PM Windows Management Instrumentation (WMI) Uptime can be determined via Windows Management Instrumentation (WMI), by querying the LastBootUpTime property of the Win32_OperatingSystem class. At the command prompt, this can be done using the wmic command: C:\>wmic os get lastbootuptime LastBootUpTime 20110508161751.822066+060 The timestamp uses the format yyyymmddhhmmss.nnn, so in the above example, the computer last booted up on 8 May 2011 at 16:17:51.822. The text "LastBootUpTime" and the timestamp format do not vary with language or locale. WMI can also be queried using a variety of application programming interfaces, including VBScript or PowerShell. Uptime.exe
https://en.wikipedia.org/wiki/Crash%20%28computing%29
In computing, a crash, or system crash, occurs when a computer program such as a software application or an operating system stops functioning properly and exits. On some operating systems or individual applications, a crash reporting service will report the crash and any details relating to it (or give the user the option to do so), usually to the developer(s) of the application. If the program is a critical part of the operating system, the entire system may crash or hang, often resulting in a kernel panic or fatal system error. Most crashes are the result of a software bug. Typical causes include accessing invalid memory addresses, incorrect address values in the program counter, buffer overflow, overwriting a portion of the affected program code due to an earlier bug, executing invalid machine instructions (an illegal or unauthorized opcode), or triggering an unhandled exception. The original software bug that started this chain of events is typically considered to be the cause of the crash, which is discovered through the process of debugging. The original bug can be far removed from the code that actually triggered the crash. In early personal computers, attempting to write data to hardware addresses outside the system's main memory could cause hardware damage. Some crashes are exploitable and let a malicious program or hacker execute arbitrary code, allowing the replication of viruses or the acquisition of data which would normally be inaccessible. Application crashes An application typically crashes when it performs an operation that is not allowed by the operating system. The operating system then triggers an exception or signal in the application. Unix applications traditionally responded to the signal by dumping core. Most Windows and Unix GUI applications respond by displaying a dialogue box (such as the one shown to the right) with the option to attach a debugger if one is installed. Some applications attempt to recover from the error and continue running instead of exiting. An application can also contain code to crash after detecting a severe error. Typical errors that result in application crashes include: attempting to read or write memory that is not allocated for reading or writing by that application (e.g., segmentation fault, x86-specific general protection fault) attempting to execute privileged or invalid instructions attempting to perform I/O operations on hardware devices to which it does not have permission to access passing invalid arguments to system calls attempting to access other system resources to which the application does not have permission to access attempting to execute machine instructions with bad arguments (depending on CPU architecture): divide by zero, operations on denormal number or NaN (not a number) values, memory access to unaligned addresses, etc. Crash to desktop A "crash to desktop" is said to occur when a program (commonly a video game) unexpectedly quits, abruptly taking the user back t
https://en.wikipedia.org/wiki/Bus%20error
In computing, a bus error is a fault raised by hardware, notifying an operating system (OS) that a process is trying to access memory that the CPU cannot physically address: an invalid address for the address bus, hence the name. In modern use on most architectures these are much rarer than segmentation faults, which occur primarily due to memory access violations: problems in the logical address or permissions. On POSIX-compliant platforms, bus errors usually result in the SIGBUS signal being sent to the process that caused the error. SIGBUS can also be caused by any general device fault that the computer detects, though a bus error rarely means that the computer hardware is physically broken—it is normally caused by a bug in software. Bus errors may also be raised for certain other paging errors; see below. Causes There are at least three main causes of bus errors: Non-existent address Software instructs the CPU to read or write a specific physical memory address. Accordingly, the CPU sets this physical address on its address bus and requests all other hardware connected to the CPU to respond with the results, if they answer for this specific address. If no other hardware responds, the CPU raises an exception, stating that the requested physical address is unrecognized by the whole computer system. Note that this only covers physical memory addresses. Trying to access an undefined virtual memory address is generally considered to be a segmentation fault rather than a bus error, though if the MMU is separate, the processor cannot tell the difference. Unaligned access Most CPUs are byte-addressable, where each unique memory address refers to an 8-bit byte. Most CPUs can access individual bytes from each memory address, but they generally cannot access larger units (16 bits, 32 bits, 64 bits and so on) without these units being "aligned" to a specific boundary (the x86 platform being a notable exception). For example, if multi-byte accesses must be 16 bit-aligned, addresses (given in bytes) at 0, 2, 4, 6, and so on would be considered aligned and therefore accessible, while addresses 1, 3, 5, and so on would be considered unaligned. Similarly, if multi-byte accesses must be 32-bit aligned, addresses 0, 4, 8, 12, and so on would be considered aligned and therefore accessible, and all addresses in between would be considered unaligned. Attempting to access a unit larger than a byte at an unaligned address can cause a bus error. Some systems may have a hybrid of these depending on the architecture being used. For example, for hardware based on the IBM System/360 mainframe, including the IBM System z, Fujitsu B8000, RCA Spectra, and UNIVAC Series 90, instructions must be on a 16-bit boundary, that is, execution addresses must start on an even byte. Attempts to branch to an odd address results in a specification exception. Data, however, may be retrieved from any address in memory, and may be one byte or longer depending on the instruction. CPUs
https://en.wikipedia.org/wiki/Data%20Protection%20Directive
The Data Protection Directive, officially Directive 95/46/EC, enacted in October 1995, was a European Union directive which regulated the processing of personal data within the European Union (EU) and the free movement of such data. The Data Protection Directive was an important component of EU privacy and human rights law. The principles set out in the Data Protection Directive were aimed at the protection of fundamental rights and freedoms in the processing of personal data. The General Data Protection Regulation, adopted in April 2016, superseded the Data Protection Directive and became enforceable on 25 May 2018. Context The right to privacy is a highly developed area of law in Europe. All the member states of the Council of Europe (CoE) are also signatories of the European Convention on Human Rights (ECHR). Article 8 of the ECHR provides a right to respect for one's "private and family life, his home and his correspondence", subject to certain restrictions. The European Court of Human Rights has given this article a very broad interpretation in its jurisprudence. In 1973, American scholar Willis Ware published Records, Computers, and the Rights of Citizens, a report that was to be influential on the directions these laws would take. In 1980, in an effort to create a comprehensive data protection system throughout Europe, the Organisation for Economic Co-operation and Development (OECD) issued its "Recommendations of the Council Concerning Guidelines Governing the Protection of Privacy and Trans-Border Flows of Personal Data". The seven principles governing the OECD's recommendations for protection of personal data were: Notice—data subjects should be given notice when their data is being collected; Purpose—data should only be used for the purpose stated and not for any other purposes; Consent—data should not be disclosed without the data subject's consent; Security—collected data should be kept secure from any potential abuses; Disclosure—data subjects should be informed as to who is collecting their data; Access—data subjects should be allowed to access their data and make corrections to any inaccurate data Accountability—data subjects should have a method available to them to hold data collectors accountable for not following the above principles. The OECD Guidelines, however, were non-binding, and data privacy laws still varied widely across Europe. The United States, meanwhile, while endorsing the OECD's recommendations, did nothing to implement them within the United States. However, the first six principles were incorporated into the EU Directive. In 1981, the Members States of the Council of Europe adopted the Convention for the Protection of Individuals with regard to Automatic Processing of Personal Data (Convention 108) to implement Article 8 of the ECHR. Convention 108 obliges the signatories to enact legislation concerning the automatic processing of personal data, and was modernised and reinforced in 2018 to become
https://en.wikipedia.org/wiki/Pattern%20matching
In computer science, pattern matching is the act of checking a given sequence of tokens for the presence of the constituents of some pattern. In contrast to pattern recognition, the match usually has to be exact: "either it will or will not be a match." The patterns generally have the form of either sequences or tree structures. Uses of pattern matching include outputting the locations (if any) of a pattern within a token sequence, to output some component of the matched pattern, and to substitute the matching pattern with some other token sequence (i.e., search and replace). Sequence patterns (e.g., a text string) are often described using regular expressions and matched using techniques such as backtracking. Tree patterns are used in some programming languages as a general tool to process data based on its structure, e.g. C#, F#, Haskell, ML, Python, Ruby, Rust, Scala, Swift and the symbolic mathematics language Mathematica have special syntax for expressing tree patterns and a language construct for conditional execution and value retrieval based on it. Often it is possible to give alternative patterns that are tried one by one, which yields a powerful conditional programming construct. Pattern matching sometimes includes support for guards. History Early programming languages with pattern matching constructs include COMIT (1957), SNOBOL (1962), Refal (1968) with tree-based pattern matching, Prolog (1972), SASL (1976), NPL (1977), and KRC (1981). Many text editors support pattern matching of various kinds: the QED editor supports regular expression search, and some versions of TECO support the OR operator in searches. Computer algebra systems generally support pattern matching on algebraic expressions. Primitive patterns The simplest pattern in pattern matching is an explicit value or a variable. For an example, consider a simple function definition in Haskell syntax (function parameters are not in parentheses but are separated by spaces, = is not assignment but definition): f 0 = 1 Here, 0 is a single value pattern. Now, whenever f is given 0 as argument the pattern matches and the function returns 1. With any other argument, the matching and thus the function fail. As the syntax supports alternative patterns in function definitions, we can continue the definition extending it to take more generic arguments: f n = n * f (n-1) Here, the first n is a single variable pattern, which will match absolutely any argument and bind it to name n to be used in the rest of the definition. In Haskell (unlike at least Hope), patterns are tried in order so the first definition still applies in the very specific case of the input being 0, while for any other argument the function returns n * f (n-1) with n being the argument. The wildcard pattern (often written as _) is also simple: like a variable name, it matches any value, but does not bind the value to any name. Algorithms for matching wildcards in simple string-matching situations have be
https://en.wikipedia.org/wiki/Type%20inference
Type inference refers to the automatic detection of the type of an expression in a formal language. These include programming languages and mathematical type systems, but also natural languages in some branches of computer science and linguistics. Nontechnical explanation Types in a most general view can be associated to a designated use suggesting and restricting the activities possible for an object of that type. Many nouns in language specify such uses. For instance, the word leash indicates a different use than the word line. Calling something a table indicates another designation than calling it firewood, though it might be materially the same thing. While their material properties make things usable for some purposes, they are also subject of particular designations. This is especially the case in abstract fields, namely mathematics and computer science, where the material is finally only bits or formulas. To exclude unwanted, but materially possible uses, the concept of types is defined and applied in many variations. In mathematics, Russell's paradox sparked early versions of type theory. In programming languages, typical examples are "type errors", e.g. ordering a computer to sum values that are not numbers. While materially possible, the result would no longer be meaningful and perhaps disastrous for the overall process. In a typing, an expression is opposed to a type. For example, , , and are all separate terms with the type for natural numbers. Traditionally, the expression is followed by a colon and its type, such as . This means that the value is of type . This form is also used to declare new names, e.g. , much like introducing a new character to a scene by the words "detective Decker". Contrary to a story, where the designations slowly unfold, the objects in formal languages often have to be defined with their type from very beginning. Additionally, if the expressions are ambiguous, types may be needed to make the intended use explicit. For instance, the expression might have a type but could also be read as a rational or real number or even as a plain text. As a consequence, programs or proofs can become so encumbered with types, that it is desirable to deduce them from the context. This can be possible by collecting the uses of untyped expression (including undefined names). If, for instance, a yet undefined name n is used in an expression , one could conclude, that n is at least a number. The process of deducing the type from an expression and its context is type inference. In general not only objects, but also activities have types and may be introduced simply by their use. For a Star Trek story, such a unknown activity could be "beaming", which for sake of the story's flow is just executed and never formally introduced. Nevertheless one can deduce its type (transport) following what happens. Additionally, both objects and activities can be constructed from their parts. In such a setting, type inference cannot onl
https://en.wikipedia.org/wiki/Type%20signature
In computer science, a type signature or type annotation defines the inputs and outputs for a function, subroutine or method. A type signature includes the number, types, and order of the arguments contained by a function. A type signature is typically used during overload resolution for choosing the correct definition of a function to be called among many overloaded forms. Examples C/C++ In C and C++, the type signature is declared by what is commonly known as a function prototype. In C/C++, a function declaration reflects its use; for example, a function pointer with the signature would be called as: char c; double d; int retVal = (*fPtr)(c, d); Erlang In Erlang, type signatures may be optionally declared, as: -spec function_name(type1(), type2(), ...) -> out_type(). For example: -spec is_even(number()) -> boolean(). Haskell A type signature in Haskell generally takes the following form: functionName :: arg1Type -> arg2Type -> ... -> argNType Notice that the type of the result can be regarded as everything past the first supplied argument. This is a consequence of currying, which is made possible by Haskell's support for first-class functions; this function requires two inputs where one argument is supplied and the function is "curried" to produce a function for the argument not supplied. Thus, calling , where , yields a new function that can be called to produce . The actual type specifications can consist of an actual type, such as , or a general type variable that is used in parametric polymorphic functions, such as , or , or . So we can write something like: Since Haskell supports higher-order functions, functions can be passed as arguments. This is written as: This function takes in a function with type signature and returns data of type out. Java In the Java virtual machine, internal type signatures are used to identify methods and classes at the level of the virtual machine code. Example: The method is represented in bytecode as . The signature of the method looks like this: public static void main(String[] args); And in the disassembled bytecode, it takes the form of The method signature for the method contains three modifiers: indicates that the method can be called by any object. indicates that the method is a class method. indicates that the method has no return value. Signature A function signature consists of the function prototype. It specifies the general information about a function like the name, scope and parameters. Many programming languages use name mangling in order to pass along more semantic information from the compilers to the linkers. In addition to mangling, there is an excess of information in a function signature (stored internally to most compilers) which is not readily available, but may be accessed. Understanding the notion of a function signature is an important concept for all computer science studies. Modern object orientation techniques make use of interfaces, which ar
https://en.wikipedia.org/wiki/Type%20variable
In type theory and programming languages, a type variable is a mathematical variable ranging over types. Even in programming languages that allow mutable variables, a type variable remains an abstraction, in the sense that it does not correspond to some memory locations. Programming languages that support parametric polymorphism make use of universally quantified type variables. Languages that support existential types make use of existentially quantified type variables. For example, the following OCaml code defines a polymorphic identity function that has a universally quantified type, which is printed by the interpreter on the second line: # let id x = x;; val id : 'a -> 'a = <fun> In mathematical notation, the type of the function id is , where is a type variable. See also System F Type theory Functional programming Dependently typed programming
https://en.wikipedia.org/wiki/S60%20%28software%20platform%29
The S60 Platform (formerly Series 60 User Interface) was a software platform for smartphones that runs on top of the Symbian operating system. It was created by Nokia based on the 'Pearl' user interface from Symbian Ltd. It was introduced at COMDEX in November 2001 and first shipped with the Nokia 7650 smartphone. The platform has since seen 5 updated editions. Series 60 was renamed to S60 in November 2005. In 2008, the Symbian Foundation was formed to consolidate all the assets of different Symbian platforms (S60, UIQ, MOAP), making it open source. In 2009, based on the code base of S60, the first iteration of the platform since the creation of Symbian Foundation was launched as S60 5th Edition, or Symbian^1, on top of Symbian OS 9.4 as its base. Subsequent iterations were named Symbian^2 (Japanese market only) and Symbian^3. The S60 software was a multivendor standard for smartphones that supports application development in Java MIDP, C++, Python and Adobe Flash. Its API was called Avkon UI. S60 consists of a suite of libraries and standard applications, such as telephony, personal information manager (PIM) tools, and Helix-based multimedia players. It was intended to power fully featured modern phones with large colour screens, which are commonly known as smartphones. Originally, the most distinguishing feature of S60 phones was that they allowed users to install new applications after purchase. Unlike a standard desktop platform, however, the built-in apps are rarely upgraded by the vendor beyond bug fixes. New features are only added to phones while they are being developed rather than after public release. Certain buttons are standardized, such as a menu key, a four way joystick or d-pad, left and right soft keys and a clear key. S60 was mainly used by Nokia but they also licensed it to a few other manufacturers, including Lenovo, LG Electronics, Panasonic, Samsung, Sendo, Siemens Mobile, Sony Ericsson, Solstice and Vertu. Sony Ericsson notably was the main vendor using the competing UIQ Symbian interface. In addition to the manufacturers the community includes: Software integration companies such as Sasken, Elektrobit, Teleca, Digia, Mobica, Atelier.tm Semiconductor companies Texas Instruments, STMicroelectronics, Broadcom, Sony, Freescale Semiconductor, Samsung Electronics Operators such as Vodafone and Orange who develop and provide S60-based mobile applications and services Software developers and independent software vendors (ISVs). Editions There have been four major releases of S60: "Series 60" (2001), "Series 60 Second Edition" (2002), "S60 3rd Edition" (2005) and "S60 5th Edition" (2008). Each release had an updated version called Feature Pack, sometimes known as relay. Each runs on a different Symbian version. Series 60 1st Edition The devices' display resolution was fixed to 176×208. The Siemens SX1 meanwhile had 176×220. Version 0.9 was the original and first shipped with Nokia 7650. Version 1.2 (marketed as Feature
https://en.wikipedia.org/wiki/3G
3G is the third generation of wireless mobile telecommunications technology. It is the upgrade over 2G, 2.5G, GPRS and 2.75G Enhanced Data Rates for GSM Evolution networks, offering faster data transfer, and better voice quality. This network was superseded by 4G, and later on by 5G. This network is based on a set of standards used for mobile devices and mobile telecommunications use services and networks that comply with the International Mobile Telecommunications-2000 (IMT-2000) specifications by the International Telecommunication Union. 3G finds application in wireless voice telephony, mobile Internet access, fixed wireless Internet access, video calls and mobile TV. 3G telecommunication networks support services that provide an information transfer rate of at least 144 kbit/s. Later 3G releases, often denoted 3.5G and 3.75G, also provide mobile broadband access of several Mbit/s to smartphones and mobile modems in laptop computers. This ensures it can be applied to wireless voice calls, mobile Internet access, fixed wireless Internet access, video calls and mobile TV technologies. A new generation of cellular standards has appeared approximately every tenth year since 1G systems were introduced in 1979 and the early to mid-1980s. Each generation is characterized by new frequency bands, higher data rates and non–backward-compatible transmission technology. The first commercial 3G networks were introduced in mid-2001. Overview Several telecommunications companies marketed wireless mobile Internet services as 3G, indicating that the advertised service was provided over a 3G wireless network. However, 3G services have largely been supplanted in marketing by 4G and 5G services in most areas of the world. Services advertised as 3G are required to meet IMT-2000 technical standards, including standards for reliability and speed (data transfer rates). To meet the IMT-2000 standards, a system must provide peak data rates of at least 144 kbit/s. However, many services advertised as 3G provide higher speed than the minimum technical requirements for a 3G service. Subsequent 3G releases, denoted 3.5G and 3.75G, provided mobile broadband access of several Mbit/s for smartphones and mobile modems in laptop computers. 3G branded standards: The UMTS (Universal Mobile Telecommunications System) system, standardized by 3GPP in 2001, was used in Europe, Japan, China (with a different radio interface) and other regions predominated by GSM (Global Systems for Mobile) 2G system infrastructure. The cell phones are typically UMTS and GSM hybrids. Several radio interfaces are offered, sharing the same infrastructure: The original and most widespread radio interface is called W-CDMA (Wideband Code Division Multiple Access). The TD-SCDMA radio interface was commercialized in 2009 and only offered in China. The latest UMTS release, HSPA+, can provide peak data rates up to 56 Mbit/s in the downlink in theory (28 Mbit/s in existing services) and 22 Mbit/s in the u
https://en.wikipedia.org/wiki/2G
2G is a short notation for second-generation cellular network, a group of technology standards employed for cellular networks. 2G was commercially launched on the GSM standard in Finland by Radiolinja (now part of Elisa Oyj) in 1991. After 2G was launched, the previous mobile wireless network systems were retroactively dubbed 1G. While radio signals on 1G networks are analog, radio signals on 2G networks are digital, though both systems use digital signaling to connect cellular radio towers to the rest of the mobile network system. The most common 2G technology was the time-division multiple access (TDMA)-based GSM standard, used in most of the world outside Japan. In North America, Digital AMPS (IS-54 and IS-136) and cdmaOne (IS-95) were dominant, but GSM was also used. In Japan the ubiquitous system was Personal Digital Cellular (PDC) though another, Personal Handy-phone System (PHS), also existed. Three primary benefits of 2G networks over their 1G predecessors were: Digitally encrypted phone conversations, at least between the mobile phone and the cellular base station but not necessarily in the rest of the network. Significantly more efficient use of the radio frequency spectrum enabling more users per frequency band. Data services for mobile, starting with SMS text messages then expanding to Multimedia Messaging Service (MMS). With General Packet Radio Service (GPRS), 2G offers a theoretical maximum transfer speed of 40 kbit/s (5 kB/s). With EDGE (Enhanced Data Rates for GSM Evolution), there is a theoretical maximum transfer speed of 384 kbit/s (48 kB/s). Evolution 2.5G (GPRS) 2.5G ("second and a half generation") is used to describe 2G-systems that have implemented a packet-switched domain in addition to the circuit-switched domain. It does not necessarily provide faster service because bundling of timeslots is used for circuit-switched data services (HSCSD) as well. 2.75G (EDGE) GPRS networks evolved to EDGE networks with the introduction of 8PSK encoding. While the symbol rate remained the same at 270.833 samples per second, each symbol carried three bits instead of one. Enhanced Data rates for GSM Evolution (EDGE), Enhanced GPRS (EGPRS), or IMT Single Carrier (IMT-SC) is a backward-compatible digital mobile phone technology that allows improved data transmission rates, as an extension on top of standard GSM. EDGE was deployed on GSM networks beginning in 2003, initially by AT&T in the United States. 2.875G (EDGE Evolution) Phase-out 2G, understood as GSM and CdmaOne, has been superseded by newer technologies such as 3G (UMTS / CDMA2000), 4G (LTE / WiMAX) and 5G (5G NR). However, 2G networks were still available in most parts of the world, while notably excluding the majority of carriers in North America, East Asia, and Australasia. Many modern LTE-enabled devices have the ability to fall back to 2G for phone calls, necessary especially in rural areas where later generations have not yet been implemented. In some pla
https://en.wikipedia.org/wiki/Commodore%201540
The Commodore 1540 (also known as the VIC-1540) introduced in 1982 is the companion floppy disk drive for the VIC-20 home computer. It uses single-sided 5¼" floppy disks, on which it stores roughly of data utilizing Commodore's GCR data encoding scheme. Because of the low price of both the VIC-20 and the 1540, this combination was the first computer with a disk drive to be offered on the US market for less than although the combination of the Commodore 64 and 1541 would prove more enduring. The 1540 is an "intelligent peripheral" in that it has its own MOS Technology 6502 CPU (just like its VIC-20 host) and the resident Commodore DOS on board in ROM – contrary to almost all other home computer systems of the time, where the DOS was loaded from a boot floppy and was executed on the computer's CPU. Due to a timing conflict with the C64's video chip, the C64 doesn't work properly with the 1540. The better-known 1541 is mechanically and nearly electronically identical to the 1540 but has a revised ROM that permits it to work with the C64 by slowing the drive down slightly. However, it is possible to revert the 1541 into 1540 mode with a Commodore BASIC software command (OPEN 15,8,15, "UI-" : CLOSE 15) to permit better speed when used with a VIC-20. The 1540 is relatively rare. While cheaper than most other drives of the day, it was more expensive than the VIC-20 computer itself, and the disk media was also still relatively pricey. Also, the relatively small memory of the VIC meant that the faster program loading times of the drive did not gain more than a few seconds compared to tape media. Thirdly, almost all commercial software for the VIC-20 was sold on cartridge or cassette tape media, giving low incentive to buy a floppy drive. The C64 followed close on the heels of the VIC-20, quickly discontinuing the 1540. Most 1540s still in existence were modified with a 1541 ROM so it would work with a C64. Unmodified 1540s are now considered collector's items. The launch price in Germany was (approximate The US-American version is named and the German version References VIC-20 CBM floppy disk drives
https://en.wikipedia.org/wiki/FRS
FRS may also refer to: Government and politics Facility Registry System, a centrally managed Environmental Protection Agency database that identifies places of environmental interest in the United States Family Resources Survey, a survey to collect information on the incomes and circumstances of households in Great Britain Federal Reserve System, the central bank of the United States Fire and Rescue Service, an organization that provides predominantly emergency firefighting services for a specific geographic area Fisheries Research Services, a now-defunct agency of the Executive agencies of the Scottish Government Forum of Social Republicans (French: ), now the Christian Democratic Party, a French political party Science Fellow of the Royal Society, an award and fellowship granted by the Royal Society of London to individuals the society judges to have made a "substantial contribution to the improvement of natural knowledge" Filtered Rayleigh scattering, a diagnostic technique which measures velocity, temperature, and pressure Forward recoil scattering, an ion beam analysis technique in materials science to obtain elemental concentration depth profiles in thin films Fragment separator, an ion-optical device used to focus and separate products from the collision of relativistic ion beams with thin targets Framingham Risk Score, a gender-specific algorithm used to estimate the 10-year cardiovascular risk of an individual. Free radical scavenger, a synonym for antioxidant Free radical substitution, a substitution reaction involving free radicals as a reactive intermediate Technology Family Radio Service, an improved walkie-talkie radio system in the United States File Replication Service, a Microsoft Windows Server service for distributing shared files and Group Policy objects Frame Relay Switch, a standardized wide-area network technology that specifies the physical and data link layers of digital telecommunications channels using a packet switching methodology Freely redistributable software, software that anyone is free to redistribute Functional requirement Specification, defines a function of a system or its components Facial recognition system, a computer application capable of identifying or verifying a person from a digital image or a video frame from a video source Transportation Ffestiniog Railway Society, a 1 ft 11 1⁄2 in (597 mm) narrow gauge heritage railway, located in Gwynedd, Wales Flandre Air, (ICAO: FRS) was a French regional airline Förde Reederei Seetouristik, a German shipping company Forsinard railway station, in northern Scotland Franconia–Springfield station, Amtrak station code FRS Full-rigged ship, a sailing vessel's sail plan with three or more masts, all of them square-rigged Mundo Maya International Airport, (IATA: FRS, formerly Flores International Airport) in Flores, Guatemala Scion FR-S, an automobile Other uses Fixed repeating schedule, a production scheduling methodology Fleet Repl
https://en.wikipedia.org/wiki/Algebraic%20data%20type
In computer programming, especially functional programming and type theory, an algebraic data type (ADT) is a kind of composite type, i.e., a type formed by combining other types. Two common classes of algebraic types are product types (i.e., tuples and records) and sum types (i.e., tagged or disjoint unions, coproduct types or variant types). The values of a product type typically contain several values, called fields. All values of that type have the same combination of field types. The set of all possible values of a product type is the set-theoretic product, i.e., the Cartesian product, of the sets of all possible values of its field types. The values of a sum type are typically grouped into several classes, called variants. A value of a variant type is usually created with a quasi-functional entity called a constructor. Each variant has its own constructor, which takes a specified number of arguments with specified types. The set of all possible values of a sum type is the set-theoretic sum, i.e., the disjoint union, of the sets of all possible values of its variants. Enumerated types are a special case of sum types in which the constructors take no arguments, as exactly one value is defined for each constructor. Values of algebraic types are analyzed with pattern matching, which identifies a value by its constructor or field names and extracts the data it contains. Algebraic data types were introduced in Hope, a small functional programming language developed in the 1970s at the University of Edinburgh. Examples One of the most common examples of an algebraic data type is the singly linked list. A list type is a sum type with two variants, Nil for an empty list and Cons x xs for the combination of a new element x with a list xs to create a new list. Here is an example of how a singly linked list would be declared in Haskell: data List a = Nil | Cons a (List a) or data [] a = [] | a : [a] Cons is an abbreviation of construct. Many languages have special syntax for lists defined in this way. For example, Haskell and ML use [] for Nil, : or :: for Cons, respectively, and square brackets for entire lists. So Cons 1 (Cons 2 (Cons 3 Nil)) would normally be written as 1:2:3:[] or [1,2,3] in Haskell, or as 1::2::3::[] or [1,2,3] in ML. For a slightly more complex example, binary trees may be implemented in Haskell as follows: data Tree = Empty | Leaf Int | Node Int Tree Tree or data BinaryTree a = BTNil | BTNode a (BinaryTree a) (BinaryTree a) Here, Empty represents an empty tree, Leaf represents a leaf node, and Node organizes the data into branches. In most languages that support algebraic data types, it is possible to define parametric types. Examples are given later in this article. Somewhat similar to a function, a data constructor is applied to arguments of an appropriate type, yielding an instance of the data type to which the type constructor belongs. For example, the data constructor Le
https://en.wikipedia.org/wiki/List%20of%20Chobits%20characters
This is a list of the main characters from the manga and anime series Chobits by Clamp. The series tells the story of Hideki Motosuwa, who finds an abandoned persocom, or personal computer with human form, that he names Chi after the only word she initially can speak. As the series progresses, together they explore the mysteries of Chi's origin and questions about the relationships between humans and persocoms. Main characters Hideki Motosuwa is a 19-year-old (18 in the anime) repeat student (ronin) attempting to get into a university by studying at Seki cram school, which his parents have sent him to. They have also severed his allowance, which means he has to work at Yorokonde (My Pleasure), an izakaya, to make ends meet. One night, while walking home from a convenience store, he finds a persocom lying in the trash outside his apartment building. He takes her home, and after much searching, switches her on and finds that the only word she can utter is "Chi". Because of this, he gives her the name "Chi" and takes her into his care, doing his best to protect and teach her. Hideki could best be described as "a fish out of water." Living in the country his entire life, Hideki has trouble adjusting to his new life in Tokyo. Due to this, he is prone to outbursts of confusion and frustration. Also, because of the time spent by himself on his family's farm (which in the anime is revealed to be in Hokkaidō), Hideki has a habit of thinking out loud and talking to himself. This leads to some very entertaining moments between himself and the women he meets. Like many young men, Hideki has an active libido, possessing many pornographic magazines which he refers to as "okazu" ("side dish"). (In the English manga translation, Chi refers to one of the magazines as a "tasty side dish"; in the anime's American translation, Chi refers to these magazines as "Hideki's yummies"). When he moves to Tokyo, Hideki desires a persocom so that he can finally surf the internet for pornography. It should be stated however that he isn't the only character who thinks this way. In fact Shinbo himself proudly admits that Plum (Sumomo in the anime) is also equipped to perform these tasks when he first meets Hideki. Despite his shortcomings and occasional moments of awkwardness with the bustling environment of Tokyo, Hideki is a genuinely kind and honest person, whose habit of thinking of others' well-being before his own can cause many problems. When a friend is in trouble, especially Chi, he is always there to help. Although many characters in the series often tease him for being a "nice guy", mostly by saying that's why he's a virgin, they turn to him for advice with their problems, knowing he will always hear them out and do his best to assist them. In an extra 27th episode included on the Chobits anime DVD, it is shown that Hideki is finally admitted to college after studying for a year at the prep school. Chi is a "chobit," a technologically advanced persocom rumo
https://en.wikipedia.org/wiki/Think%20aloud%20protocol
A think-aloud (or thinking aloud) protocol is a method used to gather data in usability testing in product design and development, in psychology and a range of social sciences (e.g., reading, writing, translation research, decision making, and process tracing). Description Think-aloud protocols involve participants thinking aloud as they are performing a set of specified tasks. Participants are asked to say whatever comes into their mind as they complete the task. This might include what they are looking at, thinking, doing, and feeling. This gives observers insight into the participant's cognitive processes (rather than only their final product), to make thought processes as explicit as possible during task performance. In a formal research protocol, all verbalizations are transcribed and then analyzed. In a usability testing context, observers are asked to take notes of what participants say and do, without attempting to interpret their actions and words, and especially noting places where they encounter difficulty. Test sessions may be completed on participants own devices or in a more controlled setting. Sessions are often audio- and video-recorded so that developers can go back and refer to what participants did and how they reacted. History The think-aloud method was introduced in the usability field by Clayton Lewis while he was at IBM, and is explained in Task-Centered User Interface Design: A Practical Introduction by Lewis and John Rieman. The method was developed based on the techniques of protocol analysis by K. Ericsson and H. Simon. However, there are some significant differences between the way Ericsson and Simon propose that protocols be conducted and how they are actually conducted by usability practitioners, as noted by Ted Boren and Judith Ramey. These differences arise from the specific needs and context of usability testing; practitioners should be aware of these differences and adjust their method to meet their needs while still collecting valid data. For example, they may need to prompt for additional information more often than Ericsson and Simon would allow, but should take care not to influence what participants say and do. Process A typical procedure of think-aloud protocols would include: Design the study and write the guide: Determine the number and type of participant for the study. Generally 5 participants would be sufficient. The next step is to write a guide that ask the participants to complete the tasks intended with clear step-by-step instructions. In the script, there should be reminders to participants to say their thoughts out when performing tasks. Recruit participants: The team should set up a screener for eligibility of participants. After contacting the person of interest and setting up meeting details such as time and location, the team could also provide additional information to help participant better prepare for the activity. Conduct think-aloud protocol: After stating the purpose and a
https://en.wikipedia.org/wiki/Paper%20prototyping
In human–computer interaction, paper prototyping is a widely used method in the user-centered design process, a process that helps developers to create software that meets the user's expectations and needs – in this case, especially for designing and testing user interfaces. It is throwaway prototyping and involves creating rough, even hand-sketched, drawings of an interface to use as prototypes, or models, of a design. While paper prototyping seems simple, this method of usability testing can provide useful feedback to aid the design of easier-to-use products. This is supported by many usability professionals. History Paper prototyping started in the mid-1980s and then became popular in the mid-1990s, when companies such as IBM, Honeywell, Microsoft, and others, started using the technique in developing their products. Today, paper prototyping is used widely in user-centered design by usability professionals. More recently, digital paper prototyping has been advocated by companies like Pidoco due to advantages in terms of collaboration, flexibility, and cost. Benefits Paper prototyping saves time and money, since it enables developers to test product interfaces (from software and websites to cell phones and microwave ovens) before they write code or begin development. This also allows for easy and inexpensive modification to existing designs, which makes this method useful in the early phases of design. Using paper prototyping allows the entire creative team to be involved in the process, which eliminates the chance of someone with key information not being involved in the design process. Another benefit of paper prototyping is that users feel more comfortable being critical of the mock-up because it doesn't have a polished look. There are different methods of paper prototyping, each of them showing several benefits regarding the communication within the development team, as well as the quality of the product being developed. In the development team, paper prototypes can serve as a visual specification of the graphical user interface – and by this means, assure the quality of the software. Prototyping forces a more complete design of the user interface to be captured. In team meetings, it provides a communication base between the team members. Testing prototypes at an early stage of development helps to identify software usability problems even before any code is written. The costs and annoyances of later changes are reduced, the support burden is lowered, and the overall quality of the software or website is increased. Paper prototyping is a quick way to generate digital ideas by sketching on paper. In tight VPC workshop, quick ideas need to be explored and evaluated. Paper prototyping is usually the preferred tool to generate ideas visually and to evaluate them within the team and with target customers. Drawbacks Despite many benefits as a quick and easy way to receive feedback on initial design ideas, this method also has certain draw
https://en.wikipedia.org/wiki/Deep%20Blue%20versus%20Kasparov%2C%201996%2C%20Game%201
Deep Blue–Kasparov, 1996, Game 1 is a famous chess game in which a computer played against a human being. It was the first game played in the 1996 Deep Blue versus Garry Kasparov match, and the first time that a chess-playing computer defeated a reigning world champion under normal chess tournament conditions (in particular, standard time control; in this case 40 moves in two hours). Overview Deep Blue was a computer developed by IBM to beat grandmaster Garry Kasparov, the top chess player in the world at the time according to Elo ratings. Playing White, Deep Blue won this first game in the match on February 10, 1996, in Philadelphia, Pennsylvania. Kasparov rebounded over the next five games, winning three and drawing two, to soundly beat the machine in the 1996 match. The game White: Deep Blue Black: Kasparov Opening: Sicilian Defense, Alapin Variation (ECO B22) 1. e4 c5 2. c3 It is more common to play 2.Nf3, but Kasparov has deep experience with that line, so White's opening book goes in a different direction. The IBM team determined the opening moves played by Deep Blue. 2... d5 3. exd5 Qxd5 4. d4 Nf6 5. Nf3 Bg4 6. Be2 e6 7. h3 Bh5 8. 0-0 Nc6 9. Be3 cxd4 10. cxd4 Bb4 (diagram) A more common move here is Be7. This was a new approach by Kasparov, developing the bishop in an unusual way. If 11.Nc3 Qa5 12.Qb3 then the game transposes into a game Kasparov previously played against Kramnik. The merit of the new move is debated. After this move, the computer left its opening book and began calculating its moves. 11. a3 Ba5 12. Nc3 Qd6 13. Nb5 Qe7 14. Ne5 Bxe2 15. Qxe2 0-0 16. Rac1 Rac8 17. Bg5 Black now has a problem with the pinned knight on f6. 17... Bb6 18. Bxf6 gxf6 Kasparov avoids ...Qxf6 because White would gain with 19.Nd7 by forking the Queen and Rook. Note that Kasparov's king is now far more exposed. 19. Nc4! Black cannot take the d4-pawn due to Qg4+. 19... Rfd8 20. Nxb6! axb6 21. Rfd1 f5 22. Qe3!? Two questionable moves by Deep Blue. 21.Qg4+ Kh8 22.Rcd1 was better, bringing White's queen and to attack Black's . 22... Qf6 (diagram) 23. d5! This type of pawn sacrifice is typical of Kasparov's style of play. Kasparov commented that he might have played 23.d5 himself in this position, since it hurts Black's pawn structure and opens up the board, and Black's exposed king suggests that there is probably a way to exploit the result. Kasparov has been attacking White's d-pawn, and the computer wisely decides to advance it for an attack instead of trying to defend it. 23... Rxd5 24. Rxd5 exd5 25. b3! Kh8 Kasparov attempts to prepare a counterattack by preparing to move his rook to the g-, but it will not work. Burgess suggests that 25...Ne7 26.Rxc8+ would have been better, though White would still have some advantage. Keene suggests that 25...Rd8! 26.Qxb6 Rd7 was Black's best try, strengthening his passed d-pawn and . 26. Qxb6 Rg8 27. Qc5 Black was threatening 27...Qg5 forking g2 and the white rook. 27... d4? 28. Nd6 f4 29. Nxb7 T
https://en.wikipedia.org/wiki/Read-copy-update
In computer science, read-copy-update (RCU) is a synchronization mechanism that avoids the use of lock primitives while multiple threads concurrently read and update elements that are linked through pointers and that belong to shared data structures (e.g., linked lists, trees, hash tables). Whenever a thread is inserting or deleting elements of data structures in shared memory, all readers are guaranteed to see and traverse either the older or the new structure, therefore avoiding inconsistencies (e.g., dereferencing null pointers). It is used when performance of reads is crucial and is an example of space–time tradeoff, enabling fast operations at the cost of more space. This makes all readers proceed as if there were no synchronization involved, hence they will be fast, but also making updates more difficult. Name and overview The name comes from the way that RCU is used to update a linked structure in place. A thread wishing to do this uses the following steps: create a new structure, copy the data from the old structure into the new one, and save a pointer to the old structure, modify the new, copied, structure, update the global pointer to refer to the new structure, sleep until the operating system kernel determines that there are no readers left using the old structure, for example, in the Linux kernel, by using , once awakened by the kernel, deallocate the old structure. So the structure is read concurrently with a thread copying in order to do an update, hence the name "read-copy update". The abbreviation "RCU" was one of many contributions by the Linux community. Other names for similar techniques include passive serialization and MP defer by VM/XA programmers and generations by K42 and Tornado programmers. Detailed description A key property of RCU is that readers can access a data structure even when it is in the process of being updated: RCU updaters cannot block readers or force them to retry their accesses. This overview starts by showing how data can be safely inserted into and deleted from linked structures despite concurrent readers. The first diagram on the right depicts a four-state insertion procedure, with time advancing from left to right. The first state shows a global pointer named that is initially , colored red to indicate that it might be accessed by a reader at any time, thus requiring updaters to take care. Allocating memory for a new structure transitions to the second state. This structure has indeterminate state (indicated by the question marks) but is inaccessible to readers (indicated by the green color). Because the structure is inaccessible to readers, the updater may carry out any desired operation without fear of disrupting concurrent readers. Initializing this new structure transitions to the third state, which shows the initialized values of the structure's fields. Assigning a reference to this new structure to transitions to the fourth and final state. In this state, the structure i
https://en.wikipedia.org/wiki/CCF
CCF can refer to: Computing Confidential Consortium Framework, a free and open source blockchain infrastructure framework developed by Microsoft Customer Care Framework, a Microsoft product Finance Credit conversion factor converts the amount of a free credit line and other off-balance-sheet transactions to its credit exposure equivalent, i.e. an Exposure at default Common contractual fund, an Irish collective investment scheme Crédit Commercial de France, a defunct French bank, now part of HSBC Health care Congestive cardiac failure Organisations Cambodian Children's Fund, a charity organisation Center for Consumer Freedom, a food industry advocacy group Cheetah Conservation Fund, a Namibian wildlife conservation organization China Carbon Forum, a non-profit organization promoting climate change stakeholder dialogue China Computer Federation, Chinese association for computing professionals Combined Cadet Force, British-government sponsored organisation for youth military training Politics Conservative Christian Fellowship, an organisation allied with the British Conservative Party Conservative Collegiate Forum, the United Kingdom Conservative party's student organisation from 1986 to 1998 Co-operative Commonwealth Federation, historic Canadian political party Sport Cyprus Cycling Federation, a Cypriot bicycling organization Transport Canadian Car and Foundry, a Canadian manufacturer of rail cars and buses CCF, IATA code for Carcassonne Salvaza Airport in France Other Cartoon Cartoon Fridays, a former programming block from Cartoon Network Centum cubic feet (100 cubic feet), an American standard unit of measurement for the volume of water or natural gas, sometimes capitalized as "Ccf" (from the Roman numeral for 100, C; see also Therm) Christ's Commission Fellowship, a non-denominational evangelical Christian church based in the Philippines Concentrated Complete Fertiliser, a product of Imperial Chemical Industries (ICI); see History of fertilizer Concentric crater fill, a geologic feature observed in impact craters on Mars Congress for Cultural Freedom, an anti-Communist advocacy group, dissolved in 1979 Corpus Carminum Færoensium, a collection of Faroese folk ballads
https://en.wikipedia.org/wiki/Timeline%20of%20hypertext%20technology
This article presents a timeline of hypertext technology, including "hypermedia" and related human–computer interaction projects and developments from 1945 on. The term hypertext is credited to the author and philosopher Ted Nelson. See also Graphical user interface, Multimedia; also Paul Otlet and Henri La Fontaine's Mundaneum, a massively cross-referenced card index system established in 1910. 1940s 1941 Jorge Luis Borges' "The Garden of Forking Paths" 1945 Memex (concept by Vannevar Bush) 1960s 1960 Project Xanadu (concept) 1962 Marshall McLuhan's The Gutenberg Galaxy uses the term surfing 1967 Hypertext Editing System (HES) by Andries van Dam and Ted Nelson at Brown University 1968 FRESS (File Retrieval and Editing System, successor to HES) NLS (oN-Line System) 1970s 1972 ZOG 1973 Xerox Alto desktop 1976 PROMIS 1978 Aspen Movie Map 1979 PERQ 1980s 1980 ENQUIRE (not released) 1981 Electronic Document System (EDS, aka Document Presentation System) Kussmaul Encyclopedia Xerox Star desktop 1982 Guide 1983 Knowledge Management System (KMS, successor to ZOG) TIES (The Interactive Encyclopedia System, later HyperTies) 1984 NoteCards 1985 Intermedia (successor to FRESS and EDS) Symbolics Document Examiner (Symbolics workstations) 1986 Texinfo TextNet (a network-based approach to text handling) Neptune (a hypertext system for CAD applications) 1987 Macromedia Authorware Canon Cat ("Leap" function, interface) HyperCard Knowledge Navigator (concept described by former Apple Computer CEO John Sculley in his book Odyssey) Storyspace 1988 Microcosm (hypermedia system) (University of Southampton) 1989 Macromedia Director Information Management: a proposal, Tim Berners-Lee, CERN 1990s 1990 DynaText World Wide Web Hyperland (BBC documentary written by Douglas Adams) ToolBook HyTelnet WinHelp 1991 Gopher AmigaGuide 1995 Wiki 1996 Hyperwire (Kinetix) 1998 Everything2 XML 1999 RSS 2000s 2001 Wikipedia 2014 OpenXanadu, an implementation of Project Xanadu 2019 Gemini, a lightweight complement to the Web Hypertext Hypertext Hypertext es:Hipertexto#Historia
https://en.wikipedia.org/wiki/Bitboard
A bitboard is a specialized bit array data structure commonly used in computer systems that play board games, where each bit corresponds to a game board space or piece. This allows parallel bitwise operations to set or query the game state, or determine moves or plays in the game. Bits in the same bitboard relate to each other by the rules of the game, often forming a game position when taken together. Other bitboards are commonly used as masks to transform or answer queries about positions. Bitboards are applicable to any game whose progress is represented by the state of, or presence of pieces on, discrete spaces of a gameboard, by mapping of the space states to bits in the data structure. Bitboards are a more efficient alternative board representation to the traditional mailbox representation, where each piece or space on the board is an array element. Bitboards are especially effective when the associated bits of various related states on the board fit into a single word or double word of the CPU architecture, so that single bitwise operators like AND and OR can be used to build or query game states. Among the computer game implementations that use bitboards are chess, checkers, othello and word games. The scheme was first employed in checkers programs in the 1950s, and since the mid-1970s has been the de facto standard for game board representation in computer automatons. Description A bitboard, a specialized bit field, is a format that packs multiple related boolean variables into the same machine word, typically representing a position on a board game, or state of a game. Each bit represents a space; when the bit is positive, a property of that space is true. Bitboards allow the computer to answer some questions about game state with one bitwise operation. For example, if a chess program wants to know if the white player has any pawns in the center of the board (center four squares) it can just compare a bitboard for the player's pawns with one for the center of the board using a bitwise AND operation. If there are no center pawns then the result will be all zero bits (i.e. equal to zero). Multiple bitboards may represent different properties of spaces over the board, and special or temporary bitboards (like temporary variables) may represent local properties or hold intermediate collated results. The efficacy of bitboards is augmented by two other properties of the implementation. First, bitboards are fast to incrementally update, such as flipping the bits at the source and destination positions in a bitboard for piece location when a piece is moved. Second, bitmaps representing static properties like all spaces attacked by each piece type for every position on a chessboard can be pre-collated and stored in a table, so that answering a question like "what are the legal moves of a knight on space e4?" can be answered by a single memory fetch. Bitfield implementations take advantage of the presence of fullword (32-bit or 64
https://en.wikipedia.org/wiki/IBM%20PCjr
The IBM PCjr (pronounced "PC junior") was a home computer produced and marketed by IBM from March 1984 to May 1985, intended as a lower-cost variant of the IBM PC with hardware capabilities better suited for video games, in order to compete more directly with other home computers such as the Apple II and Commodore 64. It retained the IBM PC's 8088 CPU and BIOS interface, but provided enhanced graphics and sound, ROM cartridge slots, built-in joystick ports, and an infrared wireless keyboard. The PCjr supported expansion via "sidecar" modules, which could be attached to the side of the unit. Despite widespread anticipation, the PCjr was ultimately unsuccessful in the market. It was only partially IBM compatible, limiting support for IBM's software library, its chiclet keyboard was widely criticized for its poor quality, expandability was limited, and it was initially offered with a maximum of of RAM, insufficient for many PC programs. Models The PCjr came in two models: 4860-004 - 64 KB of memory, priced at US$669 () 4860-067 - 128 KB of memory and a 360 KB, 5.25-inch floppy disk drive, priced at US$1,269 () The PCjr was manufactured for IBM in Lewisburg, Tennessee by Teledyne. A related machine, the IBM JX, was sold in the Japan, Australia and New Zealand markets. Hardware The PCjr chassis is made entirely of plastic, unlike the all-steel chassis of the IBM PC. A 5.25" front bay allows the installation of a 180/360K floppy disk drive. The internal floppy drive was a half-height Qume 5.25" unit; IBM also used these drives in the PC Portable, but the PCjr units were specially equipped with a small fan to prevent overheating since the computer did not have a case fan. Cartridges The front of the PCjr exposes a pair of cartridge slots in which the user can insert software on ROM cartridges, as was common with other home computers. When a ROM cartridge is inserted, the machine automatically restarts and boots off of the ROM, without requiring the user to manually reboot. Cartridges can also replace the system BIOS and other firmware. A number of patches from various vendors are included on a single "combo-cartridge", licensed and sold by PC Enterprises, to support add-on hardware, bypass certain limitations of design, and keep up with changing OS requirements. Processor Like the IBM PC, the PCjr uses an Intel 8088 clocked at 4.77 MHz. Despite using the same CPU and clock speed, performance is often inferior to the PC, because access to system RAM is delayed by wait states added by the Video Gate Array to synchronize shared access to RAM between the CPU and the video hardware. IBM claimed that an average of two wait states are added, but the designers of the Tandy 1000, a clone of the PCjr, claimed that six was a more accurate figure. This delay only applies to software resident in the first 64 KB or 128 KB of RAM inside the system unit itself, and not to programs or data located in ROM - including software on ROM cartridges plugged i
https://en.wikipedia.org/wiki/TI-99/4A
The TI-99/4 and TI-99/4A are home computers released by Texas Instruments in 1979 and 1981, respectively. The TI-99 series competed against major home computers such as the Apple II, TRS-80, and the later Atari 400/800 series and VIC-20. Based on the Texas Instruments TMS9900 microprocessor originally used in minicomputers, the TI-99/4 was the first 16-bit home computer. The associated video display controller provides color graphics and sprite support which were only comparable with those of the Atari 400 and 800 released a month after the TI-99/4. The calculator-style keyboard of the TI-99/4 was cited as a weak point, and TI's reliance on ROM cartridges and their practice of limiting developer information to select third parties resulted in a lack of software for the system. The TI-99/4A was released in June 1981 to address some of these issues with a simplified internal design, full-travel keyboard, improved graphics, and a unique expansion system. At half the price of the original model, sales picked up significantly and TI supported the 4A with peripherals, including a speech synthesizer and a "Peripheral Expansion System" box to contain hardware add-ons. TI released developer information and tools, but the insistence on remaining sole publisher continued to starve the platform of software. The 1981 US launch of the TI-99/4A followed Commodore's VIC-20 by several months. Commodore CEO Jack Tramiel did not like TI's predatory pricing in the mid-1970s and retaliated with a price war by repeatedly lowering the price of the VIC-20 and forcing TI to do the same. By late 1982, TI was dominating the U.S. home computer market, shipping 5,000 computers a day from their factory in Lubbock, Texas. By 1983, the 99/4A was selling at a loss for under . Even with the increased user base created by the heavy discounts, Texas Instruments lost US$330 million in the third quarter of 1983 and announced the discontinuation of the TI-99/4A in October 1983. Production ended in March 1984. The TI-99/4 was intended to fit in the middle of a planned range of TI-99 computers, none of which were released, but prototypes and documentation have been found after the TI-99/4A was discontinued. Features The TI-99/4A is a self-contained console with the motherboard, ROM cartridge slot, and full-travel keyboard in the same case. The power supply is external. An RF modulator allows the use of a television as a monitor. Lowercase letters are displayed as small caps, rather than separate glyphs. TI BASIC, an ANSI-compliant BASIC interpreter based on Dartmouth BASIC, is built-in and includes support for graphics, sound, and file system access. Later versions of the 99/4A, identified by (C)1983 TEXAS INSTRUMENTS V2.2 on the title page, prevent the use of unlicensed ROM cartridges from third-party manufacturers such as Atarisoft. Peripherals include a 5¼" floppy disk drive and controller, an RS-232 card with two serial ports and one parallel port, a P-code card for Pascal su
https://en.wikipedia.org/wiki/Fairlight%20CMI
The Fairlight CMI (short for Computer Musical Instrument) is a digital synthesizer, sampler, and digital audio workstation introduced in 1979 by Fairlight. It was based on a commercial licence of the Qasar M8 developed by Tony Furse of Creative Strategies in Sydney, Australia. It was one of the earliest music workstations with an embedded sampler and is credited for coining the term sampling in music. It rose to prominence in the early 1980s and competed with the Synclavier from New England Digital. History Origins: 1971–1979 In the 1970s, Kim Ryrie, then a teenager, had an idea to develop a build-it-yourself analogue synthesizer, the ETI 4600, for the magazine he founded, Electronics Today International (ETI). Ryrie was frustrated by the limited number of sounds that the synthesizer could make. After his classmate, Peter Vogel, graduated from high school and had a brief stint at university in 1975, Ryrie asked Vogel whether he would be interested in making "the world's greatest synthesizer" based on the recently announced microprocessor. He recalled: "We had long been interested in computers I built my first computer when I was about 12 and it was obvious to me that combining digital technology with music synthesis was the way to go." In December 1975, Ryrie and Vogel formed a home business to manufacture digital synthesizers. They named the business Fairlight after the hydrofoil ferry passing before Ryrie's grandmother's home in Sydney Harbour. The two planned to design a digital synthesizer that could create sounds reminiscent of acoustic instruments (physical modelling synthesis). They initially planned to make an analogue synthesizer that was digitally controlled, as the competing Moog synthesizer was difficult to control. After six months, the pair met the Motorola consultant Tony Furse. In association with the Canberra School of Electronic Music, Furse built a digital synthesizer using two 8-bit Motorola 6800 microprocessors, and the light pen and some of the graphics that would later become part of the Fairlight CMI. However, it was only able to create exact harmonic partials, sounding sterile and inexpressive. Vogel and Ryrie licensed Furse's design, mainly for its processing power, and decided to use microprocessor technology instead of analogue synthesis. Over the next year, they built what Ryrie called a "research design", the bulky, expensive, and unmarketable eight-voice QASAR M8 synthesizer, which included a 2×2×4-foot processing box and a keyboard. Sampling By 1978, Vogel and Ryrie were making "interesting" but unrealistic sounds. Hoping to learn how to synthesize an instrument by studying the harmonics of real instruments, Vogel recorded about a second of a piano piece from a radio broadcast. He discovered that by playing the recording back at different pitches, it sounded much more realistic than a synthesized piano sound. He recalled in 2005: Vogel and Ryrie coined the term sampling to describe this process. With the Fai
https://en.wikipedia.org/wiki/CSIRAC
CSIRAC (; Commonwealth Scientific and Industrial Research Automatic Computer), originally known as CSIR Mk 1, was Australia's first digital computer, and the fifth stored program computer in the world. It is the oldest surviving first-generation electronic computer (the Zuse Z4 at the Deutsches Museum is older, but was electro-mechanical, not electronic), and was the first in the world to play digital music. After being exhibited at Melbourne Museum for many years, it was relocated to Scienceworks in 2018 and is now on permanent display in the Think Ahead gallery. A comprehensive source of information about the CSIRA collection, its contributors and related topics is available from Museums Victoria on their Collections website. History The CSIRAC was constructed by a team led by Trevor Pearcey and Maston Beard, working in large part independently of similar efforts across Europe and the United States, and ran its first test program (multiplication of numbers) sometime in November 1949. In restricted operation from late 1950, publicly demonstrated and operational in 1951. Design The machine was fairly representative of first-generation valve-driven computer designs. It used mercury acoustic delay lines as its primary data storage, with a typical capacity of 768 20-bit words, supplemented by a parallel disk-type device with a total 4096-word capacity and an access time of 10 milliseconds. Its memory clock ran at 1000 Hz, and the control unit, synchronized to the clock, took two cycles to execute an instruction (later the speed was doubled to one cycle per instruction). The bus (termed the "digit trunk" in their design) is unusual compared to most computers in that it was serial—it transferred one bit at a time. Most of CSIRAC's approximately 2000 valves were of the types 6SN7, 6V6, EA50 diodes and KT66. George Semkiw later redesigned the drum-read electronics to use germanium transistors. Input to the machine was performed in the form of punched wide, 12-track paper tape, after experiments with punch cards proved unsatisfactory. The machine was controlled through a console which allowed programs to be stepped through one instruction at a time, and featured CRT displays which showed the contents of registers. Output was through a standard teleprinter or to punch tape. The instruction set supported the basic set of arithmetic and logical operations, as well as conditional and relative jumps (making it possible to write a library of subroutines). Instructions consisted of three components: a 5-bit "destination" P1-P5, a 5-bit "source" P6-P10, and a 10-bit "address" P11-P20. For instructions that used the main store, the six bits P15-P20 selected one of the 64 logical delay lines. Bits P11-P14 determined the time at which 20 bits of data were written to or extracted from the delay line, and thus represented address of a word within the selected delay line. There were 32 destination gates and 32 source gates; the 10 address bits identified a da
https://en.wikipedia.org/wiki/Information%20Technology%20Lokam
Information Technology Lokam (or IT Lokam) is a monthly computer magazine in Malayalam. It was started in 2001 by Infofriend Publications, in Kozhikode. The magazine is the first venture of the company. In September 2005, IT Lokam inaugurated a CD edition, like its competitor, Info Kairali. The IT Lokam CD mainly contains free software programs and games, useful articles, demo and shareware programs, wallpapers, drivers, skins, etc. References External links Official website 2001 establishments in Kerala Computer magazines published in India Monthly magazines published in India Magazines established in 2001 Malayalam-language magazines
https://en.wikipedia.org/wiki/Surf
Surf or SURF may refer to: Commercial products Surf (detergent), a brand of laundry detergent made by Unilever Computers and software "Surfing the Web", slang for exploring the World Wide Web surf (web browser), a lightweight web browser for Unix-like systems Surf (video game), a 2020 video game included with Microsoft Edge SURF, an acronym for "Speeded up robust features", a computer vision algorithm Education Summer Undergraduate Research Fellowship, a common summer immersion experience in higher education which supplement research activities that occur during the academic year Music Surf (Roddy Frame album), a 2002 album released by Roddy Frame Surf (Donnie Trumpet & The Social Experiment album), a 2015 album by Donnie Trumpet & The Social Experiment Surf music, a genre of popular music associated with surf culture "Surf" (Mac Miller song), a 2020 song by Mac Miller Places Surf, California, unincorporated community in Santa Barbara County near Lompoc Surf station, passenger rail station served by Amtrak's Pacific Surfliner Sanford Underground Research Facility, national laboratory for experiments conducted deep underground, in South Dakota Popular culture Surf culture, the culture surrounding the sport of surfing Ships , the name of more than one United States Navy ship Sports Surfing, a surface water sport The Atlantic City Surf, a defunct professional baseball team that played in Atlantic City, New Jersey, in the United States from 1998 to 2008 Other uses Surf zone, the foreshore region where approaching ocean surface waves get taller and break to form foamy/bubbly surface known as surf SURF (Stanford US-Russia Forum) Scottish Urban Regeneration Forum, an urban renewal body Sanford Underground Research Facility, the underground laboratory SURFnet, the university computing organisation in the Netherlands See also Surfing (disambiguation) Serfdom Surf and turf Surf City (disambiguation) Surf clam (disambiguation)
https://en.wikipedia.org/wiki/UVC
UVC may refer to: Science and technology Ultraviolet C, a subtype of ultraviolet light Universal Virtual Computer, a concept in digital archiving Umbilical venous catheter or umbilical vein catheter, a type of umbilical line in neonatal medicine USB video device class, for connecting video cameras Far Ultraviolet Camera/Spectrograph, one of the experiments deployed on the lunar surface by the Apollo 16 astronauts Other uses United Volleyball Club, a volleyball club in the Philippines. Uniform Vehicle Code, a set of US traffic laws Unidad de Valor Constante, a former currency created by the "Ley de Valores" of Ecuador in 1993
https://en.wikipedia.org/wiki/PIC16x84
The PIC16C84, PIC16F84 and PIC16F84A are 8-bit microcontrollers of which the PIC16C84 was the first introduced in 1993 and hailed as the first PIC microcontroller to feature a serial programming algorithm and EEPROM memory. It is a member of the PIC family of controllers, produced by Microchip Technology. The memory architecture makes use of bank switching. Software tools for assembler, debug and programming were only available for the Microsoft Windows operating system. Description The PIC16x84 is a microcontroller in the PIC family of controllers produced by Microchip Technology (originally named " Arizona Microchip"). It was Microchip's first microcontroller that utilised "EEPROM" memory technology for the program memory. The use of "EEPROM" technology for program memory has now been disused in favour of "FLASH" memory that is considerably cheaper to manufacture, releases less toxins into the atmosphere and is much more reliable than "EEPROM". Both "EEPROM" and "FLASH" utilise similar forms of "floating gate" technologies to operate. The device features one 8-bit timer, and 13 I/O pins. The PIC16x84 became popular in many hobbyist applications because it uses a serial programming algorithm that lends itself to very simple programmers. Additionally, the PIC16C84 uses EEPROM memory, so it is easy to erase and requires no special tools to do so. The PIC16F84 and its updated version, the PIC16F84A both utilised FLASH program memory. The PIC16C84, PIC16C84A, PIC16F84 and the PIC16F84A all contain an additional 64 Bytes of EEPROM addressed from the "DATA" memory map. This additional memory is intended for use as "user data", hence the reason it can only be addressed from the "DATA" memory mapping. F-version The PIC16F84/PIC16F84A is an improved version of the PIC16C84, and almost completely compatible, with better program security and using flash memory instead of EEPROM memory for program memory. The PIC16F84/PIC16F84A has 68 bytes of RAM whilst the PIC16C84 has 36 bytes. Since the two chips are so similar they are often referred to by the term PIC16x84 (x is used as a wildcard when referring to chips). History The PIC16C84 was introduced in 1993 and has been hailed as the first PIC microcontroller to feature a serial programming algorithm and EEPROM memory (it was preceded by the Motorola MC68HC805B6 and MC68HC805C4 along with the MC68HC11E2 with serial bootloader and EEPROM program storage released in the late 1980s). These chips lend themselves to hobby use: only a simple and cheap programmer is required to program, erase and reprogram the chip. As PIC16C84 supplies became limited due to it being discontinued, the PIC16F84 became popular as it is an almost drop-in replacement. New programming software was needed as the programming algorithm is different but the programming hardware required was the same. Even later (1998) Microchip Technology introduced the improved PIC16F84A which allowed for faster clock speeds (up to 20 MHz),
https://en.wikipedia.org/wiki/BBC%20Radio%205
BBC Radio 5 may refer to: BBC Radio 5 (former), a BBC Radio network from 1990 to 1994 BBC Radio 5 Live, a BBC Radio network since 1994 BBC Radio 5 Sports Extra (until 2022 BBC Radio 5 Live Sports Extra), a BBC digital radio service
https://en.wikipedia.org/wiki/Audio/modem%20riser
The audio/modem riser (AMR) is a riser expansion slot found on the motherboards of some Pentium III, Pentium 4, Duron, and Athlon personal computers. It was designed by Intel to interface with chipsets and provide analog functionality, such as sound cards and modems, on an expansion card. Technology Physically, it has two rows of 23 pins, making 46 pins total. Three drawbacks of AMR are that it eliminates one PCI slot, it is not plug and play, and it does not allow for hardware accelerated cards (only software-based). Technologically, it has been superseded by the Advanced Communications Riser (ACR) and Intel's own communications and networking riser (CNR). However, riser technologies in general never really took off. Modems generally remained as PCI cards while audio and network interfaces were integrated on to motherboards. See also Advanced Communications Riser (ACR) GeoPort Mobile Daughter Card References Further reading Motherboard expansion slot
https://en.wikipedia.org/wiki/Cray%20X-MP
The Cray X-MP was a supercomputer designed, built and sold by Cray Research. It was announced in 1982 as the "cleaned up" successor to the 1975 Cray-1, and was the world's fastest computer from 1983 to 1985 with a quad-processor system performance of 800 MFLOPS. The principal designer was Steve Chen. Description The X-MP's main improvement over the Cray-1 was that it was a shared-memory parallel vector processor, the first such computer from Cray Research. It housed up to four CPUs in a mainframe that was nearly identical in outside appearance to the Cray-1. The X-MP CPU had a faster 9.5 nanosecond clock cycle (105 MHz), compared to 12.5 ns for the Cray-1A. It was built from bipolar gate-array integrated circuits containing 16 emitter-coupled logic gates each. The CPU was very similar to the Cray-1 CPU in architecture, but had better memory bandwidth (with two read ports and one write port to the main memory instead of only one read/write port) and improved chaining support. Each CPU had a theoretical peak performance of 200 MFLOPS. The X-MP initially supported 2 million 64-bit words (16 MB) of main memory in 16 banks, respectively. The main memory was built from 4 Kbit bipolar SRAM ICs. CMOS memory versions of the Cray-1M were renamed Cray X-MP/1s. This configuration was first used for Cray Research's UNIX port. In 1984, improved models of the X-MP were announced, consisting of one, two, and four-processor systems with 4 and 8 million word configurations. The top-end system was the X-MP/48, which contained four CPUs with a theoretical peak system performance of over 800 MFLOPS and 8 million words of memory. The CPUs in these models introduced vector gather/scatter memory reference instructions to the product line. The amount of main memory supported was increased to a maximum of 16 million words, depending on the model. The main memory was built from bipolar or MOS SRAM ICs, depending on the model. The system initially ran the proprietary Cray Operating System (COS) and was object-code compatible with the Cray-1. A UNIX System V derivative initially named CX-OS and finally renamed UNICOS ran through a guest operating system facility. UNICOS became the main OS from 1986 onwards. The DOE ran the Cray Time Sharing System OS instead. See the Software section for the Cray-1 for a more detailed elaboration of software (language compiler, assembler, OSes, and applications) as X-MPs and 1s were mostly compatible. Extended Architecture series Cray Research announced the X-MP Extended Architecture series in 1986. The EA series CPU had an 8.5 ns clock cycle (117 MHz), and was built from macrocell array and gate array ICs. The EA series extended the width of the A and B registers to 32 bits and performed 32-bit address arithmetic, increasing the amount of memory theoretically addressable to 2 billion words. The largest configuration produced was 64 million words of MOS SRAM in 64 banks. For compatibility with existing software written for the Cray-1 a
https://en.wikipedia.org/wiki/B%C3%A9zier%20surface
Bézier surfaces are a species of mathematical spline used in computer graphics, computer-aided design, and finite element modeling. As with Bézier curves, a Bézier surface is defined by a set of control points. Similar to interpolation in many respects, a key difference is that the surface does not, in general, pass through the central control points; rather, it is "stretched" toward them as though each were an attractive force. They are visually intuitive and, for many applications, mathematically convenient. History Bézier surfaces were first described in 1962 by the French engineer Pierre Bézier who used them to design automobile bodies. Bézier surfaces can be of any degree, but bicubic Bézier surfaces generally provide enough degrees of freedom for most applications. Equation A given Bézier surface of degree (n, m) is defined by a set of (n + 1)(m + 1) control points ki,j where i = 0, ..., n and j = 0, ..., m. It maps the unit square into a smooth-continuous surface embedded within the space containing the ki,j s – for example, if the ki,j s are all points in a four-dimensional space, then the surface will be within a four-dimensional space. A two-dimensional Bézier surface can be defined as a parametric surface where the position of a point p as a function of the parametric coordinates u, v is given by: evaluated over the unit square, where is a basis Bernstein polynomial, and is a binomial coefficient. Some properties of Bézier surfaces: A Bézier surface will transform in the same way as its control points under all linear transformations and translations. All u = constant and v = constant lines in the (u, v) space, and – in particular – all four edges of the deformed (u, v) unit square are Bézier curves. A Bézier surface will lie completely within the convex hull of its control points, and therefore also completely within the bounding box of its control points in any given Cartesian coordinate system. The points in the patch corresponding to the corners of the deformed unit square coincide with four of the control points. However, a Bézier surface does not generally pass through its other control points. Generally, the most common use of Bézier surfaces is as nets of bicubic patches (where m = n = 3). The geometry of a single bicubic patch is thus completely defined by a set of 16 control points. These are typically linked up to form a B-spline surface in a similar way as Bézier curves are linked up to form a B-spline curve. Simpler Bézier surfaces are formed from biquadratic patches (m = n = 2), or Bézier triangles. Bézier surfaces in computer graphics Bézier patch meshes are superior to triangle meshes as a representation of smooth surfaces. They require fewer points (and thus less memory) to represent curved surfaces, are easier to manipulate, and have much better continuity properties. In addition, other common parametric surfaces such as spheres and cylinders can be well approximated by relatively small number
https://en.wikipedia.org/wiki/International%20Energy%20Agency
The International Energy Agency (IEA) is a Paris-based autonomous intergovernmental organisation, established in 1974, that provides policy recommendations, analysis and data on the entire global energy sector. The 31 member countries and 13 association countries of the IEA represent 75% of global energy demand. The IEA was set up under the framework of the Organisation for Economic Co-operation and Development (OECD) in the aftermath of the 1973 oil crisis to respond to physical disruptions in global oil supplies, provide data and statistics about the global oil market and energy sector, promote energy savings and conservation, and establish international technical collaboration on innovation and research. Since its founding, the IEA has also coordinated use of the oil reserves that its members are required to hold. In subsequent decades, the IEA's role expanded to cover the entire global energy system, encompassing traditional fuels such as gas, and coal as well as cleaner and fast-growing energy sources and technologies including renewable energy sources; solar photovoltaics, wind power, biofuels as well as nuclear power, and hydrogen, and the critical minerals needed for these technologies. The core activity of the IEA is providing policy advice to its 31 member states, as well as to its 13 Associated countries, which include Argentina, Brazil, China, India, Indonesia, Kenya, Senegal, South Africa, Ukraine, Singapore, Thailand, Egypt and Morocco to support their energy security and advance their transition to clean energy. The Agency publishes policy recommendations and solutions to help all countries ensure secure, affordable and sustainable energy, as well as analysis, roadmaps, policy reviews, detailed data on more than 150 countries. Recently, it has focused in particular on supporting global efforts to accelerate clean energy transition, mitigate climate change, and reach net zero emissions. As the COVID-19 pandemic set off a global health and economic crisis in early 2020, the IEA called on governments to ensure that their economic recovery plans focus on clean energy investments in order to create the conditions for a sustainable recovery and long-term structural decline in carbon emissions. In May 2021, the IEA published a roadmap for the global energy sector to reach net zero emissions by 2050 and to prevent global temperatures from rising above 1.5 °C. All IEA member countries have signed the Paris Agreement which aims to limit warming to 1.5 °C, and two thirds of IEA member governments have made commitments to emission neutrality by 2050. In March 2022, the IEA's Ministerial Meeting gave the agency a broader mandate to focus on the clean energy transition. The IEA has been criticised for historically undervaluing the role of renewable energy sources such as photovoltaics in future energy systems as well as under-estimating their declining cost. Environmental groups, investors and companies have been urging the IEA to do more
https://en.wikipedia.org/wiki/Loader
Loader can refer to: Loader (equipment) Loader (computing) LOADER.EXE, an auto-start program loader optionally used in the startup process of Microsoft Windows ME Loader (surname) Fast loader Speedloader Boot loader LOADER.COM (aka "NEWLDR"), a multi-boot loader shipping with various Digital Research, Novell, IMS, Caldera, etc. DOS-based operating systems like Multiuser DOS and DR-DOS LOADER.SYS, part of a LOADER.COM installation (see above) Clapper loader (on a film crew, also simply known as "loader") A loader, a member of a crew responsible for handling and loading ammunition, such as on a howitzer or tank crew Autoloader, an automated replacement for a crewer loader See also Loder LOADR NEWLDR Load (disambiguation)
https://en.wikipedia.org/wiki/MacLife
MacLife (stylized as Mac|Life) is an American monthly magazine published by Future US. It focuses on the Macintosh personal computer and related products, including the iPad and iPhone. It was sold as a print product on newsstands, but is now exclusively a digital–only product distributed through Magazines Direct, or through the Mac|Life app which can be obtained via the App Store. Between September 1996 and February 2007, the magazine was known as MacAddict. History MacLife is one of two successor magazines to the defunct CD-ROM Today. First published in 1993 by Imagine Publishing (now Future US), CD-ROM Today was targeted at both Windows and Macintosh users, and each issue shipped with a CD-ROM of shareware and demo programs. In August 1996, CD-ROM Today ceased publication, with two magazines taking its place: MacAddict for Macintosh users, and boot (now Maximum PC) for Windows users. As was the case with CD-ROM Today, MacAddict'''s discs included shareware and demo programs, but also came with other added features, such as staff videos and previews of content inside the magazine's hard copy. The MacAddict website was updated daily with news relevant to Apple products. MacAddict also had a mascot, a stick-figure named Max. By 1998, MacAddict surpassed Macworld as the Macintosh magazine with the highest consumer newsstand spending due to its $7.99 cover price. In February 2007, MacAddict was relaunched as MacLife. The new magazine is physically larger than the old magazine and was focused on the creativity of Mac users, and no longer comes with a CD-ROM. In April 2023, MacLife issued its last print edition and switched to a digital-only format. The magazine can still be read and enjoyed on a Mac, iPad, iPhone or other non-Apple digital devices. In Germany, a magazine of the same name but with no association is published by Falkemedia from Kiel (). Reviewing system From 1996 to mid-2002, there were four rating icons, which depicted Max. There was "Blech" (the lowest), "Yeah, Whatever" (a mediocre product), "Spiffy" (a solid yet not perfect product), and "Freakin' Awesome" (the highest). From 2002 to 2009, it was replaced with a more conventional five-point system. In 2010, MacLife'' adopted a 10-point system that included half stars. References External links – official site Archived MacAddict magazines on the Internet Archive Computer magazines published in the United States Monthly magazines published in the United States Macintosh magazines Macintosh websites Magazines established in 1996 Magazines published in the San Francisco Bay Area
https://en.wikipedia.org/wiki/Dining%20philosophers%20problem
In computer science, the dining philosophers problem is an example problem often used in concurrent algorithm design to illustrate synchronization issues and techniques for resolving them. It was originally formulated in 1965 by Edsger Dijkstra as a student exam exercise, presented in terms of computers competing for access to tape drive peripherals. Soon after, Tony Hoare gave the problem its present form. Problem statement Five philosophers dine together at the same table. Each philosopher has his own plate at the table. There is a fork between each plate. The dish served is a kind of spaghetti which has to be eaten with two forks. Each philosopher can only alternately think and eat. Moreover, a philosopher can only eat his spaghetti when he has both a left and right fork. Thus two forks will only be available when his two nearest neighbors are thinking, not eating. After an individual philosopher finishes eating, he will put down both forks. The problem is how to design a regimen (a concurrent algorithm) such that no philosopher will starve; i.e., each can forever continue to alternate between eating and thinking, assuming that no philosopher can know when others may want to eat or think (an issue of incomplete information). Problems The problem was designed to illustrate the challenges of avoiding deadlock, a system state in which no progress is possible. To see that a proper solution to this problem is not obvious, consider a proposal in which each philosopher is instructed to behave as follows: think unless the left fork is available; when it is, pick it up; think unless the right fork is available; when it is, pick it up; when both forks are held, eat for a fixed amount of time; put the left fork down; put the right fork down; repeat from the beginning. However, each philosopher will think for an undetermined amount of time and may end up holding a left fork thinking, staring at the right side of the plate, unable to eat because there is no right fork, until he starves. Resource starvation, mutual exclusion and livelock are other types of sequence and access problems. Solutions Dijkstra's solution Dijkstra's solution uses one mutex, one semaphore per philosopher and one state variable per philosopher. This solution is more complex than the resource hierarchy solution. This is a C++20 version of Dijkstra's solution with Tanenbaum's changes: #include <chrono> #include <iostream> #include <mutex> #include <random> #include <semaphore> #include <thread> constexpr const size_t N = 5; // number of philosophers (and forks) enum class State { THINKING = 0, // philosopher is THINKING HUNGRY = 1, // philosopher is trying to get forks EATING = 2, // philosopher is EATING }; size_t inline left(size_t i) { // number of the left neighbor of philosopher i, for whom both forks are available return (i - 1 + N) % N; // N is added for the case when i - 1 is negative } size_t inline right(size_t i) {
https://en.wikipedia.org/wiki/State%20%28computer%20science%29
In information technology and computer science, a system is described as stateful if it is designed to remember preceding events or user interactions; the remembered information is called the state of the system. The set of states a system can occupy is known as its state space. In a discrete system, the state space is countable and often finite. The system's internal behaviour or interaction with its environment consists of separately occurring individual actions or events, such as accepting input or producing output, that may or may not cause the system to change its state. Examples of such systems are digital logic circuits and components, automata and formal language, computer programs, and computers. The output of a digital circuit or deterministic computer program at any time is completely determined by its current inputs and its state. Digital logic circuit state Digital logic circuits can be divided into two types: combinational logic, whose output signals are dependent only on its present input signals, and sequential logic, whose outputs are a function of both the current inputs and the past history of inputs. In sequential logic, information from past inputs is stored in electronic memory elements, such as flip-flops. The stored contents of these memory elements, at a given point in time, is collectively referred to as the circuit's state and contains all the information about the past to which the circuit has access. Since each binary memory element, such as a flip-flop, has only two possible states, one or zero, and there is a finite number of memory elements, a digital circuit has only a certain finite number of possible states. If N is the number of binary memory elements in the circuit, the maximum number of states a circuit can have is 2N. Program state Similarly, a computer program stores data in variables, which represent storage locations in the computer's memory. The contents of these memory locations, at any given point in the program's execution, is called the program's state. A more specialized definition of state is used for computer programs that operate serially or sequentially on streams of data, such as parsers, firewalls, communication protocols and encryption. Serial programs operate on the incoming data characters or packets sequentially, one at a time. In some of these programs, information about previous data characters or packets received is stored in variables and used to affect the processing of the current character or packet. This is called a stateful protocol and the data carried over from the previous processing cycle is called the state. In others, the program has no information about the previous data stream and starts fresh with each data input; this is called a stateless protocol. Imperative programming is a programming paradigm (way of designing a programming language) that describes computation in terms of the program state, and of the statements which change the program state. Changes of st
https://en.wikipedia.org/wiki/Parallel%20transport
In geometry, parallel transport (or parallel translation) is a way of transporting geometrical data along smooth curves in a manifold. If the manifold is equipped with an affine connection (a covariant derivative or connection on the tangent bundle), then this connection allows one to transport vectors of the manifold along curves so that they stay parallel with respect to the connection. The parallel transport for a connection thus supplies a way of, in some sense, moving the local geometry of a manifold along a curve: that is, of connecting the geometries of nearby points. There may be many notions of parallel transport available, but a specification of one — one way of connecting up the geometries of points on a curve — is tantamount to providing a connection. In fact, the usual notion of connection is the infinitesimal analog of parallel transport. Or, vice versa, parallel transport is the local realization of a connection. As parallel transport supplies a local realization of the connection, it also supplies a local realization of the curvature known as holonomy. The Ambrose–Singer theorem makes explicit this relationship between the curvature and holonomy. Other notions of connection come equipped with their own parallel transportation systems as well. For instance, a Koszul connection in a vector bundle also allows for the parallel transport of vectors in much the same way as with a covariant derivative. An Ehresmann or Cartan connection supplies a lifting of curves from the manifold to the total space of a principal bundle. Such curve lifting may sometimes be thought of as the parallel transport of reference frames. Parallel transport on a vector bundle Let M be a smooth manifold. Let E→M be a vector bundle with covariant derivative ∇ and γ: I→M a smooth curve parameterized by an open interval I. A section of along γ is called parallel if By example, if is a tangent space in a tangent bundle of a manifold, this expression means that, for every in the interval, tangent vectors in are "constant" (the derivative vanishes) when an infinitesimal displacement from in the direction of the tangent vector is done. Suppose we are given an element e0 ∈ EP at P = γ(0) ∈ M, rather than a section. The parallel transport of e0 along γ is the extension of e0 to a parallel section X on γ. More precisely, X is the unique part of E along γ such that Note that in any given coordinate patch, (1) defines an ordinary differential equation, with the initial condition given by (2). Thus the Picard–Lindelöf theorem guarantees the existence and uniqueness of the solution. Thus the connection ∇ defines a way of moving elements of the fibers along a curve, and this provides linear isomorphisms between the fibers at points along the curve: from the vector space lying over γ(s) to that over γ(t). This isomorphism is known as the parallel transport map associated to the curve. The isomorphisms between fibers obtained in this way will, in gen
https://en.wikipedia.org/wiki/Cyberport
Cyberport is a business park in Southern District, Hong Kong consisting of four office buildings, a hotel, and a retail entertainment complex. It describes itself as a digital technology community with over 1,800 (800 on-site and 1,000 off-site) digital and technology companies, The Cyberport project has courted controversy since its inception because of the government's bypassing the open-tender process in awarding the project to real estate developer Richard Li Tzar-Kai, and also because of its reliance on "ancillary residential" revenue. In recent years, Cyberport has developed as a digital technology flagship and key incubator for entrepreneurship in Hong Kong. Digital technology community Six start-ups from Cyberport, namely GOGOX, KLOOK, WeLab, TNG, Animoca Brands and ZA International, have emerged as "unicorns", spanning categories of fintech, smart living, digital entertainment and esports. Cyberport is managed by Hong Kong Cyberport Management Company Limited, which is wholly owned by the Hong Kong SAR Government. In operation since 2004, Cyberport focuses on fintech, smart living, digital entertainment and esports, AI and big data, blockchain, and cybersecurity. Cyberport is currently home to the largest Fintech community in Hong Kong with around 400 Fintech companies. Bowtie, a member of the Cyberport community, was authorised by the Hong Kong Insurance Authority to become Hong Kong's first virtual insurer under the Fast Track system in November 2018. Zhong An and WeLab, who are also members of Cyberport, were two of the initial eight financial institutions and firms to be granted a virtual bank license. To promote the development of esports in Hong Kong, Cyberport launched the opening of Hong Kong's largest professional-grade esports venue on 16 July 2019 located in the Cyberport Arcade. It also launched the Esports Industry Facilitation Scheme and the Esports Internship Scheme. The former offers cash grants to support industry activities, while the latter provides a cash subsidy for internships in the esports industry. As of February 2022, Cyberport Incubation Programme had incubated and funded over 1,000 technology start-ups companies since its inception in 2005. The Cyberport Macro Fund was announced in 2016 to support local start-ups after seed stage but generally before or around Series A stage of funding. The fund, which received a further injection of HK$200 million by Cyberport and its scope extended to cover Series A and later stage investments, has, as of July 2022, invested in 22 start-ups totalling about HK$168 million, driving a co-investment ratio of 1:8. On 25 March, the Hong Kong SAR Government announced that Simon Chan Sai-ming would succeed Dr George Lam as chairman of the board of directors of the Hong Kong Cyberport Management Company Limited from April 1, 2022, to March 31, 2024. Cyberport currently comprises four phases, providing a total of 119,000 square metres of office space. As of May 2021, the occupa
https://en.wikipedia.org/wiki/Last%20mile%20%28telecommunications%29
The last mile or last kilometer is a phrase widely used in the telecommunications, cable television and internet industries to refer to the final leg of the telecommunications networks that deliver telecommunication services to retail end-users (customers). More specifically, the last mile describes the portion of the telecommunications network chain that physically reaches the end-user's premises. Examples are the copper wire subscriber lines connecting landline telephones to the local telephone exchange; coaxial cable service drops carrying cable television signals from utility poles to subscribers' homes, and cell towers linking local cell phones to the cellular network. The word "mile" is used metaphorically; the length of the last mile link may be more or less than a mile. Because the last mile of a network to the user is conversely the first mile from the user's premises to the outside world when the user is sending data, the term first mile is also alternatively used. The last mile is typically the speed bottleneck in communication networks; its bandwidth effectively limits the amount of data that can be delivered to the customer. This is because retail telecommunication networks have the topology of "trees", with relatively few high capacity "trunk" communication channels branching out to feed many final mile "twigs". The final mile links, being the most numerous and thus the most expensive part of the system, as well as having to interface with a wide variety of user equipment, are the most difficult to upgrade to new technology. For example, telephone trunklines that carry phone calls between switching centers are made of modern optical fiber, but the last mile is typically twisted pair wires, a technology which has essentially remained unchanged for over a century since the original laying of copper phone cables. In recent years, usage of the term "last mile" has expanded outside the communications industries, to include other distribution networks that deliver goods to customers, such as the pipes that deliver water and natural gas to customer premises, and the final legs of mail and package delivery services. The term has also been used to describe education and training providers that more tightly link individuals with job opportunities. Existing delivery system problems The increasing worldwide demand for rapid, low-latency and high-volume communication of information to homes and businesses has made economical information distribution and delivery increasingly important. As demand has escalated, particularly fueled by the widespread adoption of the Internet, the need for economical high-speed access by end-users located at millions of locations has ballooned as well. As requirements have changed, the existing systems and networks that were initially pressed into service for this purpose have proven to be inadequate. To date, although a number of approaches have been tried, no single clear solution to the 'last mile problem'
https://en.wikipedia.org/wiki/ACH
ACH or Ach may refer to: Businesses and organisations Businesses ACH Food Companies, Inc., American subsidiary of Associated British Foods ACH Network, American electronic fund clearing house Automotive Components Holdings, LLC, a Ford-managed temporary company Hospitals Adelaide Children's Hospital, former component of the Women's and Children's Hospital, Adelaide, South Australia Akron Children's Hospital, Akron, Ohio, U.S. Alberta Children's Hospital, Calgary, Alberta, Canada Arkansas Children's Hospital, Little Rock, Arkansas, U.S. Organizations Accion contra el hambre, part of the Action Against Hunger network Association for Computers and the Humanities People Ach (surname) ACH (wrestler), American professional wrestler Albert C. Hardie, Jr (born 1987) Ach., taxonomic author abbreviation for Erik Acharius (1757–1819), Swedish botanist Places Aach (toponymy), a term for "river" in German hydronymy Ach (Ammer), a tributary of the Ammer, Bavaria, Germany Ach (Blau), a tributary of the Blau, Baden-Württemberg, Germany Bregenzer Ach, a tributary of Lake Constance, Austria Dornbirner Ach, a tributary of Lake Constance, Austria Friedberger Ach, a tributary of the Lech, Bavaria, Germany Ach, Iran, a village in Qazvin Province Science Acetone cyanohydrin, an organic compound Acetylcholine (ACh), a neurotransmitter Aluminium chlorohydrate, a group of water-soluble, specific aluminium salts Sports AC Horsens, a Danish football club Allan Cup Hockey, a Canadian senior ice hockey league Transportation Achnashellach railway station, Scotland, National Railway code ACH St. Gallen–Altenrhein Airport, Switzerland, IATA airport code ACH Other uses Acholi dialect, ISO 639 code ach Advanced Combat Helmet, in the United States Army Air changes per hour, a measure of ventilation Analysis of competing hypotheses, a methodology Automated clearing house, a network for processing transactions See also Aboriginal Cultural Heritage Act 2021, or ACH Act, in Western Australia
https://en.wikipedia.org/wiki/Email%20art
Email art refers to artwork created for the medium of email. It includes computer graphics, animations, screensavers, digital scans of artwork in other media, and even ASCII art. When exhibited, Email art can be either displayed on a computer screen or similar type of display device, or the work can be printed out and displayed. Email art is an evolution of the networking Mail Art movement and began during the early 1990s. Chuck Welch, also known as Cracker Jack Kid, connected with early online artists and created a net-worker telenetlink. The historical evolution of the term "Email art" is documented in Chuck Welch's Eternal Network: A Mail Art Anthology published and edited by University of Calgary Press. By the end of the 1990s, many mailartists, aware of increasing postal rates and cheaper internet access, were beginning the gradual migration of collective art projects towards the web and new, inexpensive forms of digital communication. The Internet facilitated faster dissemination of Mail Art calls (invitations), Mail Art blogs and websites have become commonly used to display contributions and online documentation, and an increasing number of projects include an invitation to submit Email art digitally, either as the preferred channel or as an alternative to sending contributions by post. In 2006, Ramzi Turki received an e-mail containing a scanned work of Belgian artist Luc Fierens, so he sent this picture to about 7000 e-mail addresses artists seeking their interactions in order to acquire about 200 contributions and answers. See also Cyberculture Digital art Fax art Internet art Mail art References Computer networking Email Internet art New media art Digital art
https://en.wikipedia.org/wiki/LACNIC
LACNIC (Latin America and Caribbean Network Information Centre; , ) is the regional Internet registry for the Latin American and Caribbean regions. LACNIC provides number resource allocation and registration services that support the global operation of the Internet. It is a not-for-profit, membership-based organisation whose members include Internet Service Providers, and similar organisations. Functions LACNIC's main functions are: Allocating IPv4 and IPv6 address space, and Autonomous System Numbers Maintaining the public Whois Database for the Latin American and Caribbean region Reverse DNS delegations Representing the interests of the Latin American and Caribbean Internet community on the global stage Formation Since 1993, academic organizations in Latin America like ENRED – Foro de Redes de América Latina y el Caribe, discussed the need of a register for Latin America, independent from the influence of the United States. In 1998 during an ENRED meeting in Panama including NIC-MX, this theme was discussed and they learned that another group formed by commercial organizations like CABASE (Cámara Argentina de Base de Datos y Servicio en Línea) and ECOM-LAC (Latin America and Caribbean Federation for Internet and Electronic Commerce), were also discussing the idea of a Latin American registry. On January 30, 1998, Ira Magaziner, then the senior adviser to President Clinton for policy development, released a discussion paper, known as the "green paper" after the DNS root authority incident. A revised version known as the "white paper" was released on June 5. This paper proposed a new organization to handle internet resources (that later became ICANN). The International Forum for the White Paper organized four meetings, the final one in Buenos Aires was attended by South American communication organizations. Those organizations joined by ECOM-LAC, argued that Latin American IPs addresses could be handled by a local entity and the agreement for creation of LACNIC was signed in Santiago, Chile, on August 22, 1999, during the second ICANN meeting. An Interim Board was defined with six members: AHCIET (Ibero American Association of Research Centers and Telecommunication Companies), Raimundo Beca CABASE (Argentine Chamber of Databases and Online Services), Jorge Plano, later substituted by Oscar Messano CGI.br, José Luis Ribeiro ENRED (Network Forum for Latin America and the Caribbean), Julian Dunayerich; later substituted by Raul Echeverria NIC.mx (NIC Mexico), German Valdez ECOM-LAC, Fabio Marinho The agreement to form LACNIC was submitted on August 26, 1999, to Esther Dyson, then Chair of ICANN Interim Board, and the business plan was presented to ARIN. Statutes were created and it was decided that LACNIC headquarters would be in Montevideo, with technicians and equipment in São Paulo, at the NIC.br premises.LACNIC was established in 2001, with administrative offices in Montevideo, Uruguay, and technical facilities provided by CGI
https://en.wikipedia.org/wiki/AFRINIC
AFRINIC (African Network Information Centre) is the regional Internet registry (RIR) for Africa. Its headquarters are in Ebene, Mauritius. Before AFRINIC was formed, IP addresses (IPv6 and IPv4) for Africa were distributed by the Asia-Pacific Network Information Centre (APNIC), the American Registry for Internet Numbers (ARIN), and the RIPE NCC. ICANN provisionally recognised AFRINIC on 11 October 2004. The registry became operational on 22 February 2005. ICANN gave it final recognition in April 2005. Organisational structure Board of directors The AFRINIC Board consists of a nine-member Board of Directors. Six of the directors are elected to represent the different sub-regions, while two directors are elected to serve on the Board-based solely on competency as opposed to regional representation. The last seat on the Board is filled by the Chief Executive Officer. Elections are held at each AFRNIC Annual General Meeting (AGMM), which is conducted around May/June every year. Voting takes place both on-site at these meetings and prior to the meeting via online voting. Council of Elders The AFRINIC Council of Elders consists of six former AFRINIC chairpersons. They fulfill an advisory role and harness all their experience leading the organisation as former Chairs. The Members of the AFRINIC Council of Elders are: Dr. Nii N. Quaynor (AFRINIC Board of Trustees Chair: 2001–2004) Mr Pierre S. Dandjinou (AFRINIC Chair: 2004–2008) Board Member: 2004-2010 Dr. Viv Padayatchy (AFRINIC Chair: 2008–2011) - Board Member: - 2005-2011 Mrs. Maimouna Ndeye Diop Diagne (AFRINIC Chair: 2011–2012) - Board Member: 2010-2013 Dr. Christian Bope AFRINIC Staff AFRINIC staff carries out the daily operations of the organization. The Staff is structured in nine departments: CEO's Office, HR and Administration, Research and Innovation, Finance and Accounting, External Relations, Communication and Public Relations, Member Services, IT and Engineering, and Capacity Building. These divisions encompass all AFRINIC activities, including that of acting as a central source of information for Members. AFRINIC's open policy development process also invites stakeholders interested in Internet number resources from around the world (but mostly the African region) to participate. These include representatives from governments, regulators, educators, media, the technical community, civil society, and other not-for-profit organisations. Public Policy Meeting Each year, AFRINIC conducts two public policy meetings. These give the community the chance to come together for policy development, information sharing, and networking. The first Public Policy Meeting of each year is known as the Africa Internet Summit (AIS), and the second is held as a standalone meeting. The meetings are held in various locations throughout Africa. Controversies and scandals AFRINIC has been at the center of several organizational controversies in the past five years. Corruption A form
https://en.wikipedia.org/wiki/Extraction
Extraction may refer to: Science and technology Biology and medicine Comedo extraction, a method of acne treatment Dental extraction, the surgical removal of a tooth from the mouth Computing and information science Data extraction, the process of retrieving data out of data sources Information extraction Knowledge extraction The process of reversing data compression, a.k.a. decompression The process of choosing elements from a source document, in linguistics Other uses in science and technology Root extraction, in mathematics, the computation of a th root Extraction (chemistry), the separation of a substance from a matrix Primary extraction, the act of removing a spent cartridge from the chamber of a firearm Fragrance extraction, the process of obtaining fragrant oils and compounds from raw materials Resource extraction, the process of locating, acquiring and selling any resource Petroleum extraction, the process of recovering petroleum from the ground Ancestry or origin of a person Arts and entertainment Extraction (2015 film), an American thriller film starring Kellan Lutz, Bruce Willis, and Gina Carano Extraction (2020 film), an American action film starring Chris Hemsworth Extraction (album), by guitarist Greg Howe "Extraction" (The Shield), an episode of the television series Dead Space: Extraction, a video game prequel to the 2008 game Dead Space "Extractions", a song by BS 2000 on the 2001 album Simply Mortified Dirty Bomb (video game), formerly known as Extraction Tom Clancy's Rainbow Six Extraction, 2022 video game Other uses Extraction (military), the removal of someone from a hostile area to an area occupied by either friendly personnel or within friendly control Literary extract, in which a portion of a written work is published A part of the anatomy of a stroke, in rowing See also Extract Extracted, a 2012 American science fiction thriller
https://en.wikipedia.org/wiki/Drop%20%28telecommunication%29
In a communications network, a drop is the portion of a device directly connected to the internal station facilities, such as toward a telephone switchboard, toward a switching center, or toward a telephone exchange. A drop can also be a wire or cable from a pole or cable terminus to a building, in which case it may be referred to as a downlead. These cables may be reinforced to withstand the tension (due to gravity and weather) of an aerial drop (i.e., hanging in air), as in "messenger" type RG-6 coaxial cable, which is reinforced with a steel messenger wire along its length. Sources Federal Standard 1037C, in support of MIL-STD-188 External links Drop Wire Information https://web.archive.org/web/20140419020353/http://www.dropwire.co.uk/ Telecommunications equipment
https://en.wikipedia.org/wiki/Carnegie%20Mellon%20School%20of%20Computer%20Science
The School of Computer Science (SCS) at Carnegie Mellon University in Pittsburgh, Pennsylvania, US is a school for computer science established in 1988. It has been consistently ranked among the top computer science programs over the decades. As of 2022 U.S. News & World Report ranks the graduate program as tied for second with Stanford University and University of California, Berkeley. It is ranked second in the United States on Computer Science Open Rankings, which combines scores from multiple independent rankings. In the past 15 years, researchers from Carnegie Mellon's School of Computer Science have made developments in the fields of algorithms, artificial intelligence, computer networks, distributed systems, parallel processing, programming languages, computational biology, robotics, language technologies, human–computer interaction and software engineering. History In July 1965, Allen Newell, Herbert A. Simon, and Alan J. Perlis, in conjunction with the faculty from the Graduate School of Industrial Administration (GSIA, renamed Tepper School of Business in 2004), staff from the newly formed Computation Center, and key administrators created the Computer Science Department, one of the first such departments in the nation. Their mission statement was "to cultivate a course of study leading to the PhD degree in computer science, a program that would exploit the new technology and assist in establishing a discipline of computer science." The educational program, formally accepted in October 1965, drew its first graduate students from several existing academic disciplines: mathematics, electrical engineering, psychology, and the interdisciplinary Systems and Communications Sciences program in the Graduate School of Industrial Administration. The department was housed within the Mellon College of Science. With support from Newell, Simon, Nico Haberman, Provost Angel Jordan and President Richard Cyert, the computer science department began a two-year status as a "floating" department in the early months of 1986. Then, the Department began to grow, both academically and financially. In 1988, the School of Computer Science was established, among the first such schools in the country. The Computer Science Department was the original department within the school. Structure in the 1970s During the 1970s the Computer Science Department offered only a PhD study program, with no master's degree as an intermediate step. The PhD program required a minimum of six years of residency. It was called the "do or die" program among the graduate students, because a student could not drop a PhD and receive a master's degree. It had quickly focused on computer networking, operating systems (Hydra, Accent, Mach), and robotics. SCS today Organizational units Computational Biology Department (CBD) Computer Science Department (CSD) Human–Computer Interaction Institute (HCII) Software and Societal Systems Department (S3D): Its research and education foc
https://en.wikipedia.org/wiki/Elite%20%28video%20game%29
Elite is a space trading video game. It was written and developed by David Braben and Ian Bell and originally published by Acornsoft for the BBC Micro and Acorn Electron computers in September 1984. Elites open-ended game model, and revolutionary 3D graphics led to it being ported to virtually every contemporary home computer system and earned it a place as a classic and a genre maker in gaming history. The game's title derives from one of the player's goals of raising their combat rating to the exalted heights of "Elite". Elite was one of the first home computer games to use wire-frame 3D graphics with hidden-line removal. It added graphics and twitch gameplay aspects to the genre established by the 1974 game Star Trader. Another novelty was the inclusion of The Dark Wheel, a novella by Robert Holdstock which gave players insight into the moral and legal codes to which they might aspire. The Elite series is among the longest-running video game franchises. The first game was followed by the sequels Frontier: Elite II in 1993, and Frontier: First Encounters in 1995, which introduced Newtonian physics, realistic star systems and seamless freeform planetary landings. A third sequel, Elite Dangerous, began crowdfunding in 2012 and was launched on 16 December 2014, following a period of semi-open testing; it received a paid-for expansion season, Horizons, on 15 December 2015. Elite proved hugely influential, serving as a model for other games including Wing Commander: Privateer, Grand Theft Auto, EVE Online, Freelancer, the X series and No Man's Sky. Non-Acorn versions were each first published by Firebird and Imagineer. Subsequently, Frontier Developments has claimed the game to be a "Game by Frontier" to be part of its own back catalogue and all the rights to the game have been owned by David Braben. Gameplay The player initially controls the character "Commander Jameson", though the name can be changed each time the game is saved. The player starts at Lave Station with 100 credits and a lightly armed trading ship, a Cobra Mark III. Most of the ships that the player encounters are similarly named after snakes or other reptiles. Credits can be accumulated through a number of means. These include piracy, trade, military missions, bounty hunting and asteroid mining. The money generated by these enterprises allows the player to upgrade their ship with enhancements such as better weapons, increased cargo capacity, an automated docking system, an extra energy bank and more. In the game universe, stars have single planets, each with a space station in its orbit. Stars are always separated by interstellar distances which cannot be crossed using the ship's sublight engines. Travel between stars is accomplished by hyperspace jumps, and is constrained to those within range of the limited fuel capacity (a maximum of 7 light years) of the ship's hyperdrive. Sublight travel uses no fuel. Fuel can be replenished after docking with a space station, which
https://en.wikipedia.org/wiki/The%20Interactive%20Encyclopedia%20System
The Interactive Encyclopedia System, or TIES, was a hypertext system developed in the University of Maryland Human-Computer Interaction Lab by Ben Shneiderman in 1983. The earliest versions of TIES ran in DOS text mode, using the cursor arrow keys for navigating through information. A later version of HyperTIES for the Sun workstation was developed by Don Hopkins using the NeWS window system, with an authoring tool based on UniPress's Gosling Emacs text editor. HyperTIES The TIES program has evolved into the HyperTIES commercial product, sold by the Cognetics Corporation. HyperTIES has a small feature set and has touch-screen support which makes it optimal for public displays and information kiosks. As for navigation types, only reference links are supported, which can be either text or graphics. The mouse pointer also highlights anchors when passing over them. External links Hypertext Research: The Development of HyperTIES HyperTIES Hypermedia Browser and Emacs Authoring Tool for NeWS, overview article at Don Hopkins' website Designing to facilitate browsing: A look back at the Hyperties workstation browser, by Ben Shneiderman, Catherine Plaisant, Rodrigo Botafogo, Don Hopkins, William Weiland Hypertext Computer-related introductions in 1983 Editing software
https://en.wikipedia.org/wiki/.am
.am is the internet country code top-level domain (ccTLD) for Armenia. The Armenia Network Information Centre is managed by the Internet Society of Armenia and is headquartered in Yerevan. Regulation The registry for .am is operated by ISOC-AM, the local chapter of the Internet Society. Regulatory notes: Any person in the world can register a .am, .com.am, .net.am, .org.am domain for a fee. Each domain name is subject to review. Generally, each review takes about 2 or 3 working days. Armenian law prohibits its domain names from being used for spam, pornography, or terrorism sites. The AM-NIC was moved over to IPv6 address compatibility in line with the global DNS. Unicode compatible names will not be instituted at AM-NIC until all issues related to IPv6 are resolved. Usage in domain hacks Domains within the .am ccTLD enjoy popularity due to the connection to AM radio (similar to the .fm and .tv ccTLDs), and the ability to form English words ending in "am" — e.g. the mobile photo sharing service Instagram uses the Armenian domain name Instagr.am. The live video streaming service Stre.am uses the TLD to form their operative keyword, "Stream". American music artist and producer will.i.am uses the domain for his website. Such unconventional usage of TLDs in domain names are called domain hacks. Second top-level domain In 2014, a new Armenian top-level domain name was added, intended for Armenian language domain names. The TLD is .հայ (“hy”), from the ethnonym for Armenian. It launched in 2016. Partnerships Armenia is a member of the Council of European National Top Level Domain Registries and RIPE NCC. See also Internet in Armenia Mass media in Armenia References External links IANA .am whois information Armenia Network Information Centre List of .AM registrars .AM accredited registrar .AM accredited registrar, .AM domain marketplace Internet in Armenia Mass media in Armenia Country code top-level domains Council of European National Top Level Domain Registries members Computer-related introductions in 1994 sv:Toppdomän#A
https://en.wikipedia.org/wiki/Adobe%20Authorware
Adobe Authorware (previously Macromedia Authorware, originally Authorware) was an elearning authoring tool with its own interpreted, flowchart-based, graphical programming language. Authorware was used for creating interactive elearning programs that could integrate a range of multimedia content, particularly electronic educational technology (also called e-learning) applications. The flowchart model differentiated Authorware from other authoring tools, such as Adobe Flash and Adobe Director, which rely on a visual stage, time-line and script structure. History Authorware was originally produced by Authorware Inc., founded in 1987 by Dr Michael Allen. Allen had contributed to the development of the PLATO computer-assisted instruction system during the 1970s that was developed jointly by the University of Illinois and Control Data Corporation. CDC, where Allen was Director of Advanced Instructional Systems R&D, invested heavily in the development of an expansive library of interactive "courseware" to run on the PLATO system. PLATO courseware was developed in a unique programming language called TUTOR, which provided structures unique to learning interactions, such as answer judging and branching. TUTOR programmers were able to build powerful scripted systems that included simulations and games. However, despite the intent to create a special-purpose programming language usable by educators, and the decades of work that went into it, the TUTOR language was not easy to learn. Authorware started as a system aimed at addressing the problems in TUTOR. Originally titled Course of Action (CoA), it was a Macintosh-only program, but produced runtime programs for both the Mac and DOS. Authorware came later and produced DOS runtime versions until the release of the first Microsoft Windows version. Authorware used a visual interface with icons, representing essential components of the interactive learning experience. "Authors" placed icons along a “flowline” to create a sequence of events. Icons represented such components as Display—put something on the screen, Question—ask the learner for a response, Calc—perform a calculation, read data, and/or store data, and Animate—move something around on the screen. By simply placing the icons in sequence and adjusting their properties, authors could instantly see the structure of program they were creating and, most importantly, run it to see what learners would see. On-screen changes were easy to make, even while the program was running. Authorware became a rapid success in the marketplace, obtaining more than 80% of the market in about three years. Authorware Inc. merged with MacroMind/Paracomp in 1992 to form Macromedia. In December 2005, Adobe and Macromedia merged, under the Adobe Systems name. The most recent version of Authorware is 7.02; version 7 was released in 2003. The Authorware player has some issues with Internet Explorer 7 and later under Windows Vista due to Protected mode as well as runtime er
https://en.wikipedia.org/wiki/Canon%20Cat
Canon Cat is a task-dedicated desktop computer released by Canon Inc. in 1987 at the price of U.S. $1,495. On the surface, it was not unlike dedicated word processors popular in the late 1970s to early 1980s, but it was far more powerful, and incorporated many unique ideas for data manipulation. Description Canon Cat is primarily the creation of Jef Raskin, the originator of the Macintosh project at Apple. After leaving the company in 1982 and founding Information Appliance, Inc., he began designing a new computer closer to his original vision of an inexpensive, utilitarian "people's computer". Information Appliance developed the Swyft computer prototype, and then developed and licensed it to Canon as the Cat. BYTE in 1987 described the Cat as "a spiritual heir to the Macintosh". It features a text user interface, not making use of any pointer (mouse), icons, or graphics. All data are seen as a long "stream" of text broken into several pages. Instead of using a traditional command-line interface or menu system, the Cat makes use of its special keyboard, with commands being activated by holding down a "Use Front" key and pressing another key. The Cat also uses special "Leap keys", which, when held down, allows the user to incrementally search for strings of characters. The machine's hardware consists of a 9-inch (229 mm) black-and-white monitor (80 x 24 character display, 672 x 344 resolution), a single 3½-inch 256 KB floppy disk drive, and an IBM Selectric–compatible keyboard. It uses a Motorola 68000 CPU (like the Macintosh, Lisa, Atari ST, and Amiga) running at 5 MHz, has 256 KB of RAM, and an internal 300/1200 bit/s modem. Setup and user preference data are stored in 8 KB of non-volatile (battery backed-up) RAM. The Cat's array of I/O interfaces encompasses one Centronics parallel port, one RS-232C serial port (DB-25), and two RJ11 telephone jacks for the modem loop. The total weight of the system is 17 pounds (7.7 kg). An extensive range of application software is built into 256 KB of ROM: standard office suite programs, communications, a 90,000-word spelling dictionary, and user programming toolchains for Forth and assembly language. A text-only machine, the Cat includes graphics routines in ROM as well as connectors for a mouse or other pointing device that are never used. There is a defunct software project called Archy, initiated by Raskin, to develop a similar yet even more capable system for current computing systems. The project was designed to eventually replace current software user interfaces. See also Jupiter Ace, a British home computer of the early 1980s that also used Forth References External links CanonCat.org Canon Cat Computer User Manual Canon Cat Computer Reference Manual Canon Cat Computer Repair Manual Promotional Video for Leap and Canon Cat (If broken link: video is titled "Leap Technology") SwyftCard Quick Reference Card (for the earlier SwyftCard, giving the Apple II somewhat similar functionali
https://en.wikipedia.org/wiki/Open%20system
Open system may refer to: Technical term Open system (computing), one of a class of computers and associated software that provides some combination of interoperability, portability and open software standards, particularly Unix and Unix-like systems Open system (systems theory), in the natural and social sciences, a process that exchanges material, energy, people, capital or information with its environment Open system (thermodynamics), in thermodynamics and physics, a system where matter and energy can enter or leave, in contrast to a closed system where energy can enter or leave but matter can not Open system (control theory), a feedforward system that does not have any feedback loop to control its output in a control system Open system, in management science a system that is capable of self-maintenance on the basis of throughput of resources from the environment Open and closed system in social science Open system of learning, where information is sourced from multiple sources Open government, system Open-system environment reference model, one of the first reference models for enterprise architecture Open Distribution system (Example: Peer-to-peer file sharing system) Proper name Open Systems Interconnection (IT company) Open Systems Accounting Software (product), an accounting and business software Open Systems International, supplier of open automation solutions for utilities in the electric, oil & gas, transport, and water industries Open Systems AG, a company headquartered in Zurich, Switzerland Open-source software, system (product) Open publication, system (non-standard term) The system in which the exchange of both mass and energy take place through surrounding environment
https://en.wikipedia.org/wiki/Quantel%20Paintbox
The Quantel Paintbox was a dedicated computer graphics workstation for composition of broadcast television video and graphics. Produced by the British production equipment manufacturer Quantel (which, via a series of mergers, is now part of Grass Valley), its design emphasized the studio workflow efficiency required for live news production. At a price of per unit, they were used primarily by large TV networks such as NBC, while in the UK, Peter Claridge's company CAL Videographics was the first commercial company to purchase one. Following its initial launch in 1981, the Paintbox revolutionised the production of television graphics. History Artist Martin Holbrook worked with Quantel's development team to develop the artist-oriented functionality and user interface, which remained virtually unchanged throughout the life of the product; their Patented pressure-sensitive pen and tablet elevated it from a computer into a real artist's tool. The real time, broadcast quality, 24 bit Quantel "Paint Box" as it was then known, was launched at NAB in Las Vegas in May 1981. The PAL DPB-7001 and the NTSC DPB-7000 , were literally just digital paint machines, with stencils/layers introduced 9 months later and font and text functions implemented by Pepper Howard in 1983. The hardware was readily-available off-the-shelf components, supported by Programmable Array Logic ICs which were custom-programmed by Quantel. Combining the latest hardware, custom software which had solved usable digital paint issues and an artist-friendly familiar way of creating artwork that required no computer knowledge meant that the Paintbox was an instant success. The Paintbox became the global industry standard digital studio and image manipulation tool for the next fifteen years, bringing digital art and graphics onto everyone's TV screens. The second generation V-Series Paintbox was released in 1989 as a modernized and more compact and affordable model; with internal hardware improvements, better tablet, upgraded keyboard and a cordless stylus. In 1985, Quantel found a way to vastly increase the framestore capacity, enabling them to create a high resolution print quality Paintbox, which revolutionized the photo manipulation industry five years before Photoshop was introduced and led to Quantel's high profile lawsuit against Adobe for using the Paintbox's patented features. In the late 1980s, Quantel filed a patent infringement lawsuit the Spaceward Graphics Matisse system for creating a machine which was marketed as a cheaper version of the Paintbox system. They won the case against Spaceward at the High Court in London in 1990, but lost the case against Adobe in 1997, who were able to demonstrate that Dick Shoup's 1970's Superpaint had introduced particular features before Quantel's Paintbox. Though Adobe didn't yet exist as a company when Paintbox was launched, demonstrating prior use to the court meant that Quantel's Patents became invalid and the case was thrown out
https://en.wikipedia.org/wiki/Coleco%20Adam
The Coleco Adam is a home computer and expansion device for the ColecoVision by American toy and video game manufacturer Coleco. The Adam was an attempt to follow on the success of the company's ColecoVision video game console. It was available as Expansion Module #3 for the ColecoVision, converting it into a home computer, and as a standalone unit. As such, it had the benefit of being entirely compatible with all ColecoVision games and peripherals. The computer came with 64 KB of memory, a tape drive for a proprietary medium called Digital Data Packs, a daisy wheel printer, and productivity applications, along with two DDPs for SmartBASIC and Buck Rogers: Planet of Zoom Super Game. It was released in October 1983 with the initial price of $700. Although its presentation and concept were positively received, the Adam was heavily criticized upon launch for numerous defects in early units. Users lodged complaints about the reliability of the disk and tape drives. In particular, when the computer starts up, the power supply emits an electromagnetic pulse so strong that it can scramble or destroy data on storage media left inside the drives or near the computer. Other design flaws were noted, such as the Digital Data Pack tapes being corrupted by the drives they are stored in and the peculiar decision to require users to plug the machine into a printer for power. About 60 percent of Adam owners returned their units because of the defects. Besides the defects, the Adam also suffered from store availability issues, with Coleco having shipped only 95,000 units rather than the goal of 500,000 by the end of 1983. Many of the quality issues were resolved, but the Adam's reputation was permanently damaged and, in spite of price reductions, its sales negatively impacted, with Coleco reporting a loss of over $258 million. The Adam was discontinued in January 1985, and Coleco never recovered from the losses incurred, discontinuing its ColecoVision shortly afterward and finally declared itself bankrupt in 1988. Several computer journalists consider the Coleco Adam computer to be one of the worst ever. Despite its failures, it has gained a following among enthusiasts, who continue to develop hardware and software for it. History Coleco announced the Adam at the Summer Consumer Electronics Show (CES) in June 1983, and executives predicted sales of 500,000 by Christmas 1983. From the time of the computer's introduction to the time of its shipment, the price increased, from to . The Adam announcement received favorable press coverage. Competitors such as Commodore and Atari almost immediately announced similar computer/printer bundles. The company engaged in an extensive marketing campaign, with television commercials for "boys age 8 to 16 and their fathers ... the two groups that really fuel computer purchases", and print advertisements in nontechnical publications like Time and People. The Boston Phoenix, observing that Adam's $600 price was comparable to t
https://en.wikipedia.org/wiki/Create%2C%20read%2C%20update%20and%20delete
In computer programming, create, read, update, and delete (CRUD) are the four basic operations of persistent storage. CRUD is also sometimes used to describe user interface conventions that facilitate viewing, searching, and changing information using computer-based forms and reports. History The term was likely first popularized by James Martin in his 1983 book Managing the data-base environment. Conceptual Data can be put in a location/area of a storage mechanism. The fundamental feature of a storage location is that its content is both readable and updatable. Before a storage location can be read or updated it needs to be created; that is allocated and initialized with content. At some later point, the storage location may need to be destructed; that is finalized and deallocated. Together these four operations make up the basic operations of storage management known as CRUD: Create, Read, Update and Delete. Use cases Databases The acronym CRUD refers to the major operations which are implemented by databases. Each letter in the acronym can be mapped to a standard Structured Query Language (SQL) statement. Although relational databases are a common persistence layer in software applications, numerous other persistence layers exist. CRUD functionality can for example be implemented with document databases, object databases, XML databases, text files, or binary files. Some big data systems do not implement UPDATE, but have only a timestamped INSERT (journaling), storing a completely new version of the object each time. RESTful APIs The acronym CRUD also appears in the discussion of RESTful APIs. Each letter in the acronym may be mapped to a Hypertext Transfer Protocol (HTTP) method: In HTTP, the GET (read), PUT (create and update), POST (create - if we don't have `id` or `uuid`), and DELETE (delete) methods are CRUD operations as they have storage management semantics, meaning that they let user agents directly manipulate the states of target resources. The POST method, on the other hand, is a process operation that has target-resource-specific semantics which typically exceed the scope of CRUD operations. User Interface CRUD is also relevant at the user interface level of most applications. For example, in address book software, the basic storage unit is an individual contact entry. As a bare minimum, the software must allow the user to: Create, or add new entries Read, retrieve, search, or view existing entries Update, or edit existing entries Delete, deactivate, or remove existing entries Because these operations are so fundamental, they are often documented and described under one comprehensive heading such as "contact management" or "document management" in general. Other variations Other variations of CRUD include: ABCD (add, browse, change, delete) CRUDL (create, read, update, delete, list) BREAD (browse, read, edit, add, delete) DAVE (delete, add, view, edit) CRAP (create, replicate, append, process) S
https://en.wikipedia.org/wiki/NPA
NPA may refer to: Organizations and companies National Parks Association (disambiguation) National People's Action, a community organizing network in the United States National Pasta Association, an association for the United States pasta industry National Pawnbrokers Association, an American-based trade association National Physicians Alliance, a multi-specialty medical organization in the United States National Ploughing Association, organises the National Ploughing Championships in Ireland National Postdoctoral Association, an organization for postdoctoral researchers in the United States National Prescription Administrators, a former pharmaceutical company in the United States Natural Philosophy Alliance, a group that believes mainstream physics and cosmology are wrong Network Printing Alliance, a group of printer manufacturers Network Professional Association, an advocate for the international network computing professional in the United States Northern Pacific Airways, a US low-cost airline Norwegian People's Aid, a non-governmental labour and humanitarian organization in Norway National Pharmacy Association of the UK Government and politics National Pacification Army, Chinese anti-Kuomintang warlord army formed in 1926 National park authority, a term used in the United Kingdom for the legal body in charge of a national park National Party of Australia, an Australian political party National People's Army, the armed forces of the former German Democratic Republic (GDR) Sardar Vallabhbhai Patel National Police Academy, Indian national institute for training of Indian Police Service (IPS) officers National Police Agency (disambiguation) National Port Authority, Liberia National Production Authority, a former agency of the United States government for defense mobilization from 1950 to 1953 National Prosecuting Authority, a governing body of the Republic of South Africa New Anticapitalist Party (Nouveau Parti Anticapitaliste), a far-left French political party New People's Army, armed wing of the Communist Party of the Philippines Nigerian Ports Authority, a government agency that governs and operates the ports of Nigeria No party affiliation (disambiguation), a voter or politician who is not affiliated with any political party Non-Party Affiliate, a politician who is not affiliated with any political party Non-Partisan Association, a civic political party in Vancouver, British Columbia Note Printing Australia, a subsidiary of the Reserve Bank of Australia Transnet National Ports Authority of South Africa NPA Satellite Mapping, a UK remote sensing company Non-Prosecution Agreement, a way to avoid prosecution Science and medicine N-Propyl-L-arginine, a chemical substance 1-N-Naphthylphthalamic acid, an auxin plant hormone transport inhibitor Nanopascal (nPa), a unit of pressure Nasopharyngeal airway, a tube that is designed to be inserted into the nasal passageway Niemann–Pick disease type A, a genetic d
https://en.wikipedia.org/wiki/West%20Midlands%20Metro
The West Midlands Metro is a light-rail/tram system in the county of West Midlands, England. The network has 33 stops with a total of of track; it currently consists of a single route, Line 1, which operates between the cities of Birmingham and Wolverhampton via the towns of Bilston, West Bromwich and Wednesbury, on a mixture of former railway lines and urban on-street running. The system is owned by the public body Transport for West Midlands, and operated by Midland Metro Limited, a company wholly owned by the West Midlands Combined Authority. The tram system was launched on 30 May 1999 as Midland Metro, partly using the disused Birmingham Snow Hill to Wolverhampton Low Level Line. The line originally terminated at Birmingham Snow Hill station at the edge of the city centre, but following an extension opened in December 2015 it now serves the central core of Birmingham, including the principal regional mainline station, Birmingham New Street. Following further extensions the line now terminates at Edgbaston Village since 2022. At the other end of the line, an extension to Wolverhampton station was opened on 17 September 2023. Construction of a new Line 2 & 3 from Wednesbury to Brierley Hill was approved in March 2019, started in February 2020 and was intended to be completed for the 2022 Commonwealth Games, but has been delayed; it is currently expected to be completed at least to Dudley by 2025. A branch in Birmingham to Curzon Street – a planned High Speed 2 interchange – and on to Digbeth, is also under construction as of 2023. There are also proposals to extend the system further towards Solihull or Chelmsley Wood and Birmingham Airport. History Birmingham once had an extensive tram network run by Birmingham Corporation Tramways. However, as in most British cities, the network was wound down and closed by the local authority, with the last tram running in 1953. 1984 proposals There had been proposals for a light rail or Metro system in Birmingham and the Black Country put forward as early as the 1950s and 1960s, ironically at a time when some of the region's lines and services were beginning to be cut back. However, serious inquiry into the possibility started in 1981 when the West Midlands County Council and the West Midlands Passenger Transport Executive formed a joint planning committee to look at light rail as a means of solving the conurbation's congestion problems. In the summer of 1984 they produced a report entitled "Rapid Transit for the West Midlands" which set out ambitious proposals for a £500 million network of ten light rail routes which would be predominantly street running, but would include some underground sections in Birmingham city centre. One of the proposed routes would have used part of the existing line as far as West Bromwich. The scheme suffered from several drawbacks, one being that three of the proposed routes, from Birmingham to Sutton Coldfield, Shirley, and Dorridge would take over existing railways, and
https://en.wikipedia.org/wiki/Allen%20Newell
Allen Newell (March 19, 1927 – July 19, 1992) was an American researcher in computer science and cognitive psychology at the RAND Corporation and at Carnegie Mellon University's School of Computer Science, Tepper School of Business, and Department of Psychology. He contributed to the Information Processing Language (1956) and two of the earliest AI programs, the Logic Theorist (1956) and the General Problem Solver (1957) (with Herbert A. Simon). He was awarded the ACM's A.M. Turing Award along with Herbert A. Simon in 1975 for their contributions to artificial intelligence and the psychology of human cognition. Early studies Newell completed his bachelor's degree in physics from Stanford in 1949. He was a graduate student at Princeton University from 1949 to 1950, where he did mathematics. Due to his early exposure to an unknown field known as game theory and the experiences from the study of mathematics, he was convinced that he would prefer a combination of experimental and theoretical research to pure mathematics. In 1950, he left Princeton and joined the RAND Corporation in Santa Monica where he worked for "a group that was studying logistics problems of the Air Force" (Simon). His work with Joseph Kruskal led to the creation of two theories: A Model for Organization Theory and Formulating Precise Concepts in Organization Theory. Newell eventually earned his PhD from the now Tepper School of Business at Carnegie Mellon with Herbert Simon serving as his advisor. Afterwards, Newell "turned to the design and conduct of laboratory experiments on decision making in small groups" (Simon). He was dissatisfied, however, with the accuracy and validity of their findings produced from small-scale laboratory experiments. He joined with fellow RAND teammates John Kennedy, Bob Chapman, and Bill Biel at an Air Force Early Warning Station to study organizational processes in flight crews. They received funding from the Air Force in 1952 to build a simulator that would enable them to examine and analyze the interactions in the cockpit related to decision-making and information-handling. From these studies, Newell came to believe that information processing is the central activity in organizations. Artificial intelligence In September 1954, Newell enrolled in a seminar where Oliver Selfridge "described a running computer program that learned to recognize letters and other patterns" (Simon). This was when Allen came to believe that systems may be created and contain intelligence and have the ability to adapt. With this in mind, Allen, after a couple of months, wrote in 1955 The Chess Machine: An Example of Dealing with a Complex Task by Adaptation, which "outlined an imaginative design for a computer program to play chess in humanoid fashion" (Simon). His work came to the attention of economist (and future nobel laureate) Herbert A. Simon, and, together with programmer J. C. Shaw, they developed the first true artificial intelligence program, the Logic T
https://en.wikipedia.org/wiki/Alan%20Newell
Alan Newell may refer to: Alan C. Newell (born 1941), Irish/American mathematician Alan Newell (English computer scientist), professor at Dundee University See also Allen Newell (1927–1992), researcher in computer science and cognitive psychology
https://en.wikipedia.org/wiki/Hot%20swapping
Hot swapping is the replacement or addition of components to a computer system without stopping, shutting down, or rebooting the system; hot plugging describes the addition of components only. Components which have such functionality are said to be hot-swappable or hot-pluggable; likewise, components which do not are cold-swappable or cold-pluggable. Most desktop computer hardware, such as CPUs and memory, are only cold-pluggable. However, it is common for mid to high-end servers and mainframes to feature hot-swappable capability for hardware components, such as CPU, memory, PCIe, SATA and SAS drives. An example of hot swapping is the express ability to pull a Universal Serial Bus (USB) peripheral device, such as a thumb drive, external hard disk drive (HDD), mouse, keyboard, or printer out of a computer's USB slot or peripheral hub without ejecting it first. Most smartphones and tablets with tray-loading holders can interchange SIM cards without powering down the system. Dedicated digital cameras and camcorders usually have readily accessible memory card and battery compartments for quick changing with only minimal interruption of operation. Batteries can be cycled through by recharging reserve batteries externally while unused. Many cameras and camcorders feature an internal memory to allow capturing when no memory card is inserted. Rationale Hot swapping is used whenever it is desirable to change the configuration or repair a working system without interrupting its operation. It may simply be for convenience of avoiding the delay and nuisance of shutting down and then restarting complex equipment or because it is essential for equipment, such as a server, to be continuously active. Hot swapping may be used to add or remove peripherals or components, to allow a device to synchronize data with a computer, and to replace faulty modules without interrupting equipment operation. A machine may have dual power supplies, each adequate to power the machine; a faulty one may be hot-swapped. Important cards such as disk controllers or host adapters may be designed with redundant paths in order for these to be replaceable in case of failure without necessitating interruption of associated computer system operation. System considerations Machines that support hot swapping need to be able to modify their operation for the changed configuration, either automatically on detecting the change, or by user intervention. All electrical and mechanical connections associated with hot-swapping must be designed so that neither the equipment nor the user can be harmed while hot-swapping. Other components in the system must be designed so that the removal of a hot-swappable component does not interrupt operation. Mechanical design Protective covering plates, shields, or bezels may be used on either the removable components or the main device itself to prevent operator contact with live powered circuitry, to provide antistatic protection for components being ad
https://en.wikipedia.org/wiki/List%20of%20programs%20broadcast%20by%20Cartoon%20Network
This is a list of television programs currently or formerly broadcast by Cartoon Network in the United States. The network was launched on October 1, 1992, and airs mainly animated programming, ranging from action to animated comedy. In its early years, Cartoon Network's programming was predominantly made up of reruns of Looney Tunes, Tom and Jerry, and Hanna-Barbera shows (most notably Scooby-Doo, Yogi Bear, Quick Draw McGraw, The Flintstones, The Jetsons, and Jonny Quest). Cartoon Network's first original series was The Moxy Show and the late-night satirical animated talk show Space Ghost Coast to Coast. The What a Cartoon! series of showcase shorts brought the creation of many Cartoon Network original series collectives branded as "Cartoon Cartoons" in 1995. Cartoon Network has also broadcast several feature films, mostly animated or containing animated sequences, under its "Cartoon Theater" block, later renamed "Flicks". Current programming Original programming Cartoon Network Studios Warner Bros. Animation Hanna-Barbera Studios Europe Repeats of ended programming Cartoon Network Studios Warner Bros. Animation Cartoonito programming Original programming Acquired programming Upcoming programming Original programming Cartoon Network Studios Hanna-Barbera Studios Europe Acquired programming American co-productions Canadian co-productions European co-productions Preschool Former programming An asterisk (*) indicates that the program initially aired as a Cartoon Network program. A double-asterisk (**) indicates that the program became a Boomerang program. A triple-asterisk (***) indicates that the program became an Adult Swim/Toonami program. A quadruple-asterisk (****) indicates that the program became an HBO Max program. Original programming Cartoon Network Warner Bros. Animation Hanna-Barbera Hanna-Barbera Studios Europe Live-action and live-action/animated series Anthology series Miniseries Short series Programming from Hanna-Barbera Programming from Warner Bros. Animation Acquired programming Canadian co-productions European co-productions Animated Anime Live-action and live-action/animated series Adult Swim programming Cartoonito programming Acquired programming Former specials Programming blocks Current programming blocks Former programming blocks Pilots Short format This is a list of pilot episodes on Cartoon Network, along with their premiere dates for each. Picked up Not picked up Long format This is a list of pilot movies on Cartoon Network, along with their status and premiere dates for each. See also List of Cartoon Network films List of programs broadcast by Cartoonito List of programs broadcast by Adult Swim List of programs broadcast by Boomerang List of programs broadcast by Toonami List of Cartoon Network Studios productions Hanna-Barbera Studios Europe filmography Notes References Programs Cartoon Network original programming Lists of television series by netwo
https://en.wikipedia.org/wiki/GPB
GPB may refer to: Gazprombank, a Russian bank Georgia Public Broadcasting, the public broadcast network in the American state of Georgia Georgian Public Broadcaster, the national public broadcaster of the nation of Georgia Global power barometer Glossopharyngeal breathing Google Protocol Buffers, a method of serializing structured data GrandPooBear, a video game streamer Granny Peace Brigade, an American peace organizations Gravity Probe B, a satellite-based test of general relativity Guarapuava Airport, in Brazil
https://en.wikipedia.org/wiki/Georgia%20Public%20Broadcasting
Georgia Public Broadcasting (GPB) is a state network of PBS member television stations and NPR member radio stations serving the U.S. state of Georgia. It is operated by the Georgia Public Telecommunications Commission, an agency of the Georgia state government which holds the licenses for most of the PBS and NPR member stations licensed in the state. The broadcast signals of the nine television stations and 19 radio stations cover almost all of the state, as well as parts of Alabama, Florida, North Carolina, South Carolina and Tennessee. The network's headquarters and primary radio and television production facilities are located on 14th Street in Midtown Atlanta, just west of the Downtown Connector in the Home Park neighborhood. The facility and GPB are also a major part of Georgia's film and television industry, and in addition to commercial production occurring at the GPB facilities, some production companies also rent production offices from GPB. History Establishing the network On May 23, 1960, the University of Georgia signed on WGTV, the second educational television station in Georgia (after Atlanta's WETV, later WPBA, now WABE-TV). From 1960 to 1964, in a separate initiative, the Georgia Board of Education launched four educational television stations across the state, aimed at providing in-school instruction. This evolved into the Georgia Educational Television Network, which aired Board of Education-produced classes for schools and evening programming from WGTV. WGTV moved its transmitter to Stone Mountain in 1969, adding Atlanta to its coverage area. In November 1980, Governor George Busbee proposed the consolidation of WGTV with the state's network of transmitters into a new Georgia Public Telecommunications Council and also called for said body to negotiate to buy WETV from the Atlanta Board of Education. The Georgia state senate approved the bill, but it stalled in the House of Representatives due to the objections of Athens-area members and those involved with the UGA station. After the legislative session ended, Governor Busbee revived the proposal by executive order. On January 1, 1982, the new council took operational control of WGTV, and the combined service rebranded as Georgia Public Television; by June 1982, after the expiration of remaining program contracts, WGTV was fully incorporated into the network, and UGA's role was reduced to program supplier. Growth into radio In February 1985, the GPTC entered into public radio, launching stations serving Macon, Columbus and Valdosta in the first year. These formed the nuclei of Peach State Public Radio, which provided the first public radio services to much of Georgia; previously, only Atlanta and Savannah had public stations. During the 1980s and 1990s, stations that had been operated by other educational institutions and community groups became affiliated with the network. The service was renamed Georgia Public Radio in 2001. In 1995, the GPTC began using "Georgia Publi
https://en.wikipedia.org/wiki/MetaPost
MetaPost refers to both a programming language and the interpreter of the MetaPost programming language. Both are derived from Donald Knuth's Metafont language and interpreter. MetaPost produces vector graphic diagrams from a geometric/algebraic description. The language shares Metafont's declarative syntax for manipulating lines, curves, points and geometric transformations. However, Metafont is set up to produce fonts, in the form of image files (in .gf format) with associated font metric files (in .tfm format), whereas MetaPost produces EPS, SVG, or PNG files The output of Metafont consists of the fonts at a fixed resolution in a raster-based format, whereas MetaPost's output is vector-based graphics (lines, Bézier curves) Metafont output is monochrome, whereas MetaPost uses RGB or CMYK colors. The MetaPost language can include text labels on the diagrams, either strings from a specified font, or anything else that can be typeset with TeX. Starting with version 1.8, Metapost allows floating-point arithmetic with 64 bits (default: 32 bit fixed-point arithmetic) Many of the limitations of MetaPost derive from features of Metafont. For instance, MetaPost does not support all features of PostScript. Most notably, paths can have only one segment (so that regions are simply connected), and regions can be filled only with uniform colours. PostScript level 1 supports tiled patterns and PostScript 3 supports Gouraud shading. Availability and usage MetaPost is distributed with many distributions of the TeX and Metafont framework, for example, it is included in the MiKTeX and the TeX Live distributions. The encapsulated postscript produced by Metapost can be included in LaTeX, ConTeXt, and TeX documents via standard graphics inclusion commands. The encapsulated postscript output can also be used with the PDFTeX engine, thus directly giving PDF. This ability is implemented in ConTeXt and in the LaTeX graphics package, and can be used from plain TeX via the supp-pdf.tex macro file. ConTeXt and LuaTeX supports the inclusion of MetaPost code within the input file. Inclusion of MetaPost code in LaTeX is also possible by using LaTeX-packages, for example gmp or mpgraphics. Examples This is a single file example.mp which when processed by the MetaPost interpreter (via the command mpost on Linux) produces three eps files example.1, example.2, example.3. These are pictured on the right. transform pagecoords; pagecoords:=identity scaled 10mm shifted (100mm,150mm); beginfig (1) fill ((0,0)--(2,0)--(2,1)--(1,1)--(1,2)--(0,2)--cycle) transformed pagecoords withcolor green; draw ((2,0)..(2,1)..(1,1)..(1,2)..(0,2)) transformed pagecoords; drawarrow ((0,0)--(2,2)) transformed pagecoords; endfig; beginfig (2) draw (for i=0 upto 7: dir (135i)-- endfor cycle) transformed pagecoords; endfig; pagecoords:=identity scaled 15mm shifted (100mm,150mm); beginfig (3); % declare paths to be used path p[],p[]t; % set up
https://en.wikipedia.org/wiki/Anime%20Network
Anime Network was an American video on demand (VOD) network dedicated to anime owned by AMC Networks. History The network was launched in North America in late 2002 and is marketed to multi system operators (MSOs) as both a free and subscription Video On Demand (VOD) programming service. Anime Network also provides online streaming of its anime for North America via its website with free full-length preview episodes for non-members, more episodes for members, and all online titles available for subscribers. In October 2007, Anime Network was launched on DirecTV On Demand, initially on channel 1801 at launch. As of the fall of 2015, it can now be found on channel 1889. The Anime Network ceased broadcast of its linear 24/7 network on January 1, 2008; it continues to support a VOD service and online player on its main website. On September 1, 2009, A.D. Vision had sold off the Anime Network to Valkyrie Media Partners LLC as part of the dissolution of the company and the reorganization of its assets. In June 2017, Section23 Films announced it would discontinue the Anime Network Online streaming service and focus solely on Anime Network's pay television and subscription VOD services. HIDIVE, LLC, a new company which is not affiliated with Anime Network, Section23, or Sentai Filmworks, acquired Anime Network Online's assets and spun them off into a new streaming service called HIDIVE. On June 20, 2017, Anime Network Inc. announced it acquired the "www.theanimenetwork.com" domain name from Anime Network Online. Availability and distribution At launch, the Anime Network was only available by Video On Demand to Comcast subscribers in the Philadelphia area with 1.2 million customers with more providers choosing to carry the service later on. The linear service launched on June 30, 2004, to complement the VOD service. Comcast Communications previously carried the channel on video on demand since its launch. On May 29, 2009, Comcast discontinued carriage of the service in both its free package and its pay-per-view digital television package which, however, are periodically included among Comcast Xfinity's Top Picks. 24/7 linear service From June 30, 2004, to January 4, 2008, Anime Network offered a 24/7 channel available to pay television providers. The service was only carried by small cable companies. On January 4, 2008, Anime Network officially announced that the 24/7 service would cease operations. The network will still operate as a VOD provider. Video On Demand The Anime Network's content is distributed through video on demand in two formats. Sony announced at E3 2010 that the PlayStation Network would begin carrying The Anime Network. Anime Network (UK) On June 3, 2007, Anime Network was launched in the United Kingdom as a programming block on the Propeller TV satellite channel, produced by A.D. Vision. It was broadcast as a 2-hour block with 4 different anime series and aired from 8:00pm to 10:00pm, 7 days a week. The block was remov
https://en.wikipedia.org/wiki/Jurong%20Island
{ "type": "ExternalData", "service": "geoshape", "ids": "Q2383204", "title": "Jurong Island" } Jurong Island is an island located to the southwest of the main island of Singapore. It was formed from the amalgamation of seven offshore islands, the islands of Pulau Ayer Chawan, Pulau Ayer Merbau, Pulau Merlimau, Pulau Pesek, Pulau Pesek Kechil (also called Terumbu Pesek), Pulau Sakra (which was a previous merger of Pulau Sakra and Pulau Bakau), Pulau Seraya, Pulau Meskol, Pulau Mesemut Laut, Pulau Mesemut Darat and Anak Pulau. This was done through Singapore's land reclamation efforts. Land reclamation on Jurong Island was completed on 25 September 2009, 20 years earlier than scheduled. Pulau Buaya was joined to Jurong Island via reclamation in 2010. Jurong Island forms a land area of about from an initial area of less than , and is the largest of Singapore's outlying islands. History The island was previously a stop on a maritime trade route. Jurong Island used to be a maze of island hideouts frequented by raiding pirates. Later in the early 19th century, entrepreneurs trading gambier altered the area for large scale settlement. In 1848 a British surveyor, John Turnbull Thomson, mentioned there were numerous Chinese and Malay villages in Jurong. The outlying islands of Pulau Ayer Chawan, Pulau Ayer Merbau, Pulau Merlimau and Pulau Seraya used to house fishing communities comprising small villages up to the 1960s. The villagers lived in Malay-style wooden stilt houses on the palm-fringed islands. Between the late-1960s and early-1970s, three big oil companies planned to house their facilities on Pulau Ayer Chawan for Esso, Pulau Merlimau for Singapore Refinery Company and Pulau Pesek for Mobil Oil. The Government of Singapore then took the opportunity to grow the petrochemical industry to improve economic growth. It proved a success when the petroleum industry took off in the 1970s. By the 1980s, after a decade of rapid industrialisation, industrial land was growing scarce on Singapore mainland. The idea of joining the southern islands off Jurong to form one colossal island to create more industrial land was therefore conceived. In 1991, JTC Corporation (formerly Jurong Town Corporation) was appointed the agent of the Jurong Island project. JTC planned and coordinated with various government agencies in providing the necessary infrastructure and services to the island. Physical land reclamation began in 1995, and Jurong Island was officially opened on 14 October 2000 by then Prime Minister Goh Chok Tong. From the 9.91 km² land area of the original seven islets, as of completion of the land reclamation on 25 September 2009, Jurong Island currently has a total land area of 30 km². Penta-Ocean Construction was the major contractor and reclamation was completed 20 years ahead of schedule. Petrochemical industries Today, Jurong Island is home to many companies such as Arkema, Arlanxeo, Afton Chemical, Air Liquide, Asahi Kasei, BASF, Ce
https://en.wikipedia.org/wiki/SEM
SEM or Sem can refer to: Computing Search engine marketing, promoting websites by increasing their visibility in search engine results Security event manager, a security log tool used on data networks Economics and management Stock Exchange of Mauritius, the principal stock exchange of the island country of Mauritius Strategic enrollment management, an element of planning for new growth at a university or college Energy Single Electricity Market, the integration of the electricity sectors of the Republic of Ireland and Northern Ireland in a single market Strategic energy management, continuous improvement process for energy efficiency based on behavioral change Mathematics and statistics Structural equation modeling, in data analysis Simultaneous equations model, in econometrics Standard error of the mean in statistics Places Craig Field (Alabama), US airport, IATA code Sem, Ariège, France Sem, Norway Science and technology Scanning electron microscope Space Experiment Module, of the Freestar experiment on space shuttle Columbia Synthesizer expansion module in an Oberheim Electronics synthesizer Sistema Eléctrico de Magallanes, an electrical power grid of Chile SEM, instrument (seismometer) of the Surface Science Platform (Kazachok) mission ExoMars 2022 A protecting group in organic chemistry, see 2-(trimethylsilyl)ethoxymethyl chloride (SEM-Cl) Schools, organisations, societies The School of Ecclesiastic Music the school of Byzantine Music in Mount Lebanon Swedish Evangelical Mission Society for Ethnomusicology, an ethnomusicology in Indiana, US Science and Engineering Magnet, a school in Dallas, Texas, US S.E.M. Group civil engineering company based in South Australia Steam Engine Makers' Society, early British trade union Shandong Engineering Machinery, owned by Caterpillar Inc. State Secretariat for Migration (Switzerland) People Shem or Sem, one of the sons of Noah, as depicted in the Bible Sem (given name), a usually masculine given name Sem (artist) (1863–1934), French caricaturist Elise Sem (1870–1950), Norwegian barrister, women's activist, and sports official Ingebjørg Dahl Sem (born 1938), Norwegian Christian writer Ingebjørg Sem (1931–2009), Norwegian actress John Sem (born 1973), Papua New Guinean Olympic boxer Niels Arntzen Sem (1782–1859), Norwegian politician Suy Sem (born 1947), Cambodian politician Other Super Étendard Modernisé, a French fighter aircraft Self-evaluation maintenance theory, a psychological theory of self New Austrian tunnelling method or Sequential Excavation Method Sports and exercise medicine
https://en.wikipedia.org/wiki/NTLDR
NTLDR (abbreviation of NT loader) is the boot loader for all releases of Windows NT operating system from 1993 with the release of Windows NT 3.1 up until Windows XP and Windows Server 2003. From Windows Vista onwards it was replaced by the BOOTMGR bootloader. NTLDR is typically run from the primary storage device, but it can also run from portable storage devices such as a CD-ROM, USB flash drive, or floppy disk. NTLDR can also load a non NT-based operating system given the appropriate boot sector in a file. NTLDR requires, at the minimum, the following two files to be on the system volume: , the main boot loader itself , required for booting an NT-based OS, detects basic hardware information needed for successful boot An additional important file is boot.ini, which contains boot configuration (if missing, NTLDR will default to \Windows on the first partition of the first hard drive). NTLDR is launched by the volume boot record of system partition, which is typically written to the disk by the Windows or command. History Windows NT was originally designed for ARC-compatible platforms, relying on its boot manager support and providing only osloader.exe, a loading program accepting ordinary command-line arguments specifying Windows directory partition, location or boot parameters, which is launched by ARC-compatible boot manager when a user chooses to start specific Windows NT operating system. However, because the x86 lacked any of the ARC support, the additional layer was added specifically for that platform: custom boot manager code presenting text menu allowing the user to choose from one or more operating system and its options configured in boot.ini configuration file, prepended by special StartUp module which is responsible for some preparations such as switching the CPU to the protected mode. When a user chooses an operating system from the boot menu, the following command-line arguments are then passed to the part of the osloader.exe common to all processor architectures: In Windows releases starting from Windows Vista and Windows Server 2008, NTLDR has been split off back to its boot manager and system loader parts: the Windows Boot Manager and winload.exe. The boot manager part has been completely rewritten; it no longer uses boot.ini as a configuration file, although the bootcfg utility for modifying boot.ini is still present for the case of multi-boot configurations with Windows versions up to Windows XP and Windows Server 2003. Command-line interface The bootsect.exe utility program in the Windows PE tools has options /nt52 (NTLDR) and /nt60 (Vista and up) to store a NTLDR or Vista boot record in the first sector of a specified partition. The command can be used for FAT and NTFS based file systems. It replaces the FixFAT and FixNTFS tools. Example The following example applies the NTLDR compatible master boot code to the D: volume: C:\>bootsect /nt52 D: Startup process When a PC is powered on its BIOS follows the confi
https://en.wikipedia.org/wiki/Beige%20box
In consumer computer products, a beige box is a standard personal computer (PC). It has come to be used as a term of derision implying conservative or dated aesthetics and unremarkable specifications. The term is ultimately derived from the style of many early personal computers and dedicated word processors, which were usually beige or similar colors like off white or ecru. IBM's early desktop computers (e.g. IBM Personal Computer, IBM PC/AT) were beige, and box-shaped, and most manufacturers of clones followed suit. As IBM and its imitators came to dominate the industry, these features became standards of desktop computer design. Early Macintosh models were a beige color (specifically Pantone 453). Although Apple switched to a desaturated gray they called “Platinum” in 1987, users began to refer to them as “beige” following the introduction of the brightly colored iMac in 1998 and the Blue and White G3 in 1999. It eventually became a standard term to identify any previous Old World Macintosh, such as the “Beige G3.” While the original Commodore 64 was a deeper brown (specifically, RAL 1019), its second revision in 1986, the C64C, was beige. The German-exclusive minor revision of the original form factor the following year, sometimes referred to as the C64G, combined the new beige color of the C64C with the original larger size case. The term is also sometimes used to distinguish generic PCs from models made by "name brands" such as Compaq, Dell, or HP. In the early years of these companies, most of their units were beige as well. More recently, as name-brand manufacturers have moved away from beige (typically switching to black, dark gray, and silver-colored cases), inexpensive generic cases became more distinct as "beige boxes". Today, the term "white box" has largely replaced this usage. See also White box (computer hardware) References Personal computers Computer enclosure
https://en.wikipedia.org/wiki/Rog-O-Matic
Rog-O-Matic is a bot developed in 1981 to play and win the video game Rogue, by four graduate students in the Computer Science Department at Carnegie-Mellon University in Pittsburgh: Andrew Appel, Leonard Hamey, Guy Jacobson and Michael Loren Mauldin. Described as a "belligerent expert system", Rog-O-Matic performs well when tested against expert Rogue players, even winning the game. Because all information in Rogue is communicated to the player via ASCII text, Rog-O-Matic has automatic access to the same information a human player has. The program is still the subject of some scholarly interest; a 2005 paper said: Notes References External links Game artificial intelligence Expert systems
https://en.wikipedia.org/wiki/Honeypot%20%28computing%29
In computer terminology, a honeypot is a computer security mechanism set to detect, deflect, or, in some manner, counteract attempts at unauthorized use of information systems. Generally, a honeypot consists of data (for example, in a network site) that appears to be a legitimate part of the site which contains information or resources of value to attackers. It is actually isolated, monitored, and capable of blocking or analyzing the attackers. This is similar to police sting operations, colloquially known as "baiting" a suspect. The main use for this network decoy is to distract potential attackers from more important information and machines on the real network, learn about the forms of attacks they can suffer, and examine such attacks during and after the exploitation of a honeypot. It provides a way to prevent and see vulnerabilities in a specific network system. A honeypot is a decoy used to protect a network from present or future attacks. Types Honeypots can be differentiated based on if they are physical or virtual: Physical honeypots: real machine with its own IP address, this machine simulates behaviors modeled by the system. Many times this modality is not used as much as the high price of acquiring new machines, their maintenance and the complication affected by configuring specialized hardware Virtual honeypots: the use of these types of honeypot allow one to install and simulate hosts on the network from different operating systems, but in order to do so, it is necessary to simulate the TCP/IP of the target operating system. This modality is more frequent. Honeypots can be classified based on their deployment (use/action) and based on their level of involvement. Based on deployment, honeypots may be classified as: production honeypots research honeypots Production honeypots are easy to use, capture only limited information, and are used primarily by corporations. Production honeypots are placed inside the production network with other production servers by an organization to improve their overall state of security. Normally, production honeypots are low-interaction honeypots, which are easier to deploy. They give less information about the attacks or attackers than research honeypots. Research honeypots are run to gather information about the motives and tactics of the black hat community targeting different networks. These honeypots do not add direct value to a specific organization; instead, they are used to research the threats that organizations face and to learn how to better protect against those threats. Research honeypots are complex to deploy and maintain, capture extensive information, and are used primarily by research, military, or government organizations. Based on design criteria, honeypots can be classified as: pure honeypots high-interaction honeypots low-interaction honeypots Pure honeypots are full-fledged production systems. The activities of the attacker are monitored by using a bug tap that has been
https://en.wikipedia.org/wiki/Courts%20Malaysia
COURTS MALAYSIA, is a consumer electronics and furniture retailer in Malaysia with a network of 46 stores nationwide and a staff strength of 1,100. Courts has been operating in Malaysia for 37 years. History Incorporated on 23 July 1986 as a private limited company under the name Courts Furnishers (Malaysia) Sdn. Bhd, Courts changed its name to Courts Sdn Bhd on 30 June 1997. Courts subsequently converted into a public company and assumed its present name, Courts Sdn Bhd on 1 July 1997. Courts was listed on the Main Board of the Kuala Lumpur Stock Exchange (now known as Bursa Malaysia Securities Berhad) on October 11, 2000, and delisted again from bursa Malaysia 3 September 2007 markets following the completion of ownership engagement with Malaysia Retail Group. Courts is now located in Malaysia, Singapore, and Indonesia. 2019, Japanese electronics retailer Nojima Corp acquired Courts Asia. Nojima is listed on the Tokyo Stock Exchange. Like, Courts Asia, it is an electrical appliance retailer, boasting more than 8000 employees and a market capitalisation of S$1.4 billion. Re-branding Courts has been undergoing various stages of a re-branding initiatives since 2006. Phase 1 of the reformatting of its urban stores adopting the ‘refreshed’ look and feel with an expanded product range in contemporary furnishing and in electronics. Courts entered Phase 2 of its re-branding with the launch of Courts Megastore in December 2007. Phase 2, also includes the refurbishment of the rest of the Courts outlets scheduled to be completed in March 2009. Courts Megastore and Latest Developments Courts Malaysia opened its first 'Big-Box' Megastore in Malaysia in the early second half of 2013. It is located at the ground and first floor of the 8trium building in Sri Damansara. This 108,000 sq ft store is the ultimate one-stop lifestyle destination built on a retail proposition to provide choice, value, experience, service and payment options of credit, cash or credit card to consumers. The concept of launching Courts Megastore targets the urban consumer's evolving tastes and needs as well as appealing to a broader market which may not be aware of Courts. Courts Megastore offers many first-in-Malaysia concepts such as Hot off the Docks, Countdown Corner, Dr. Digital, Sleep Clinic, made-to-order and many more. 2015 saw the re-launched of Courts online alongside the nationwide brand campaign “Senang Sehaja, Courts Ada” where local comedian Harith Iskander was appointed the first Courts brand ambassador in 20 years. Agency created the campaign was DVG Web Sdn Bhd. 2017, COURTS celebrated their 30 Years Anniversary in Malaysia. Singer Liyana Fizi was featured in their music video titled Konfetti. Harith Iskander was also featured on the store's promotions. 2020, New COurts Store opened in Kota Marudu on 27th February 2022 2022 First Flagship Nojima Store grand opening at Lalaport, Bukit Bintang, Kualalumpur, Malaysia Grand Opening of Courts Miri at Binta
https://en.wikipedia.org/wiki/List%20of%20sports%20announcers
This is a list of sports announcers and sports commentators. Those television and radio networks included must have national exposure, not regional. American football Troy Aikman – Fox 2001–2021, ESPN 2022–present Kenny Albert – Fox 1994–present Jason Bell – BBC Sport 2015–2022, ITV Sport 2022–, Chris Berman – ESPN 1985–present Allen Bestwick – ESPN 2014–present Rocky Boiman – BBC Radio 5 Live 2012–13 & 2015–2017, Absolute Radio 2013–2014 Terry Bradshaw – CBS 1984–1994, Fox 1994–present Greg Brady – BBC Radio 5 Live 2006–13, Talksport 2017–present Tim Brant – ABC 1982–present James Brown – Fox 1994–2006, CBS 2006–present, Showtime 2008–present Jack Buck – CBS 1964–96 Joe Buck – Fox 1994–97, 2001–2021, ESPN 2022–present Mike Carlson – Channel 5 1997–2009, BBC Sport 2009–13 & 2015–2022, Channel 4 2010–2015, Talksport 2017–present Mark Chapman – BBC Sport 2012–13 & 2015–2021 Cris Collinsworth – HBO 1989–2008, NBC 1990–98, NBC 2006–present, Fox 1998–2006, NFL Network 2006–08, Showtime 2008–present Nat Coombs – Channel 5 2006–2009, Channel 4 2012–15, BBC Sport 2015–2022, Talksport 2017–present Lee Corso – ESPN 1987–present Howard Cosell – ESPN/ABC 1970–1983 David Croft – BBC Radio 5 Live 2007 Randy Cross – CBS 1989–1993, 1998–present, NBC 1994–97 Al DeRogatis – NBC 1966–1975 Dan Dierdorf – CBS 1985–1986, 1999–2014, ABC 1987–1999 Tony Dungy – NBC 2009–present Dick Enberg – NBC 1975–1999, CBS 2000–09 Boomer Esiason – ESPN/ABC 1998–2000, Fox Sports Net 2001, CBS 2002–present Darren Fletcher – BBC Radio 5 Live 2011–13 & 2015–present, Absolute Radio 2013–2014, BBC Sport 2012–13, ITV Sport 2022–present Dan Fouts – CBS 1988–1993, 2008–present, ESPN/ABC 1997–2008 Chris Fowler – ESPN/ABC, ESPN/ABC, ABC 1986–present Frank Gifford – CBS 1965–70, ABC 1971–99 Jerry Glanville – Fox 1994–98, CBS 1999–2003 Curt Gowdy – ABC 1960–1964, 1982–1983, NBC 1965–79 Jon Gruden – ESPN/ABC 2009–2017 Kevin Harlan – Fox 1994–97, CBS 1998–present Rodney Harrison – NBC 2009–present Kirk Herbstreit – ESPN/ABC, ESPN/ABC, ABC 1996–present Jake Humphrey – BBC Sport 2008–2011 Gary Imlach – Channel 4 1987–1997, ITV Sport 2005–2008, Channel 4 2010 Keith Jackson – ABC 1966–2006 Ron Jaworski – ESPN/ABC 2007–2011 Jimmy Johnson – Fox 1994–95, 2002–present Daryl Johnston – Fox 2001–present Vernon Kay – Channel 4 2013–15 Danny Kelly – Channel 4 2011 Howie Long – Fox 1994–present Bill Maas – Fox 1995–2005 John Madden – CBS 1981–1994, Fox 1994–2002, ABC 2002–06, NBC 2006–09 Paul Maguire – NBC 1976–79, 1988–97, ESPN 1998–2005 Dan Marino – CBS 2002–present Curt Menefee – Fox 2006–present Don Meredith – ABC 1970–73, 1977–85, NBC 1974–76 Al Michaels – ABC 1986–2006, NBC 2006–present Matt Millen – CBS 1992–1993, Fox 1994–2001 Colin Murray – Channel 5 2004–2006, Channel 4 2013–14 Brent Musburger – CBS 1973–1990, ESPN, ABC 1990–2016 Jim Nantz – CBS 1985–present Merlin Olsen – NBC 1977–1989, CBS 1990–1991 Mike Patrick – ESPN 1987–present Jonathan
https://en.wikipedia.org/wiki/Web%20application
A web application (or web app) is application software that is accessed using a web browser. Web applications are delivered on the World Wide Web to users with an active network connection. History In earlier computing models like client-server, the processing load for the application was shared between code on the server and code installed on each client locally. In other words, an application had its own pre-compiled client program which served as its user interface and had to be separately installed on each user's personal computer. An upgrade to the server-side code of the application would typically also require an upgrade to the client-side code installed on each user workstation, adding to the support cost and decreasing productivity. In addition, both the client and server components of the application were usually tightly bound to a particular computer architecture and operating system and porting them to others was often prohibitively expensive for all but the largest applications (Nowadays, native apps for mobile devices are also hobbled by some or all of the foregoing issues). In 1995, Netscape introduced a client-side scripting language called JavaScript, allowing programmers to add some dynamic elements to the user interface that ran on the client side. So instead of sending data to the server in order to generate an entire web page, the embedded scripts of the downloaded page can perform various tasks such as input validation or showing/hiding parts of the page. In 1999, the "web application" concept was introduced in the Java language in the Servlet Specification version 2.2. [2.1?]. At that time both JavaScript and XML had already been developed, but Ajax had still not yet been coined and the XMLHttpRequest object had only been recently introduced on Internet Explorer 5 as an ActiveX object. Applications like Gmail started to make their client sides more and more interactive since early 2000s. A web page script is able to contact the server for storing/retrieving data without downloading an entire web page. The practice became known as Ajax in 2005. "Progressive web apps", the term coined by designer Frances Berriman and Google Chrome engineer Alex Russell in 2015, refers to apps taking advantage of new features supported by modern browsers, which initially run inside a web browser tab but later can run completely offline and can be launched without entering the app URL in the browser. Structure Traditional PC applications consist only of 1 tier, which resides on the client machine, but web applications lend themselves to a multi-tiered approach by nature. Though many variations are possible, the most common structure is the three-tiered application. In its most common form, the three tiers are called presentation, application and storage. A web browser is the first tier (presentation), an engine using some dynamic Web content technology (such as ASP, CGI, ColdFusion, Dart, JSP/Java, Node.js, PHP, Python or Ruby on Rails
https://en.wikipedia.org/wiki/Firebird%20%28database%20server%29
Firebird is an open-source SQL relational database management system that supports Linux, Microsoft Windows, macOS and other Unix platforms. The database forked from Borland's open source edition of InterBase in 2000 but the code has been largely rewritten since Firebird 1.5. History Within a week of the InterBase 6.0 source being released by Borland on 25 July 2000, the Firebird project was created on SourceForge. Firebird 1.0 was released for Linux, Microsoft Windows and Mac OS X on 11 March 2002, with ports to Solaris, FreeBSD 4, HP-UX over the next two months. Work on porting the codebase from C to C++ began in 2000. On 23 February 2004, Firebird 1.5 was released, which was the first stable release of the new codebase. Version 1.5 featured an improved query optimizer, SQL-92 conditional expressions, SQL:1999 savepoints and support for explicit locking. Firebird 2.0 was released on 12 November 2006, adding support for 64-bit architectures, tables nested in FROM clauses, and programmable lock timeouts in blocking transactions. The previous stable release was version 2.1.6, which added new features including procedural triggers, recursive queries, and support for SQL:2003 MERGE statements. Firebird 2.5 introduced new features like improved multithreading, regular expression syntax and the ability to query remote databases. The most recent stable version is Firebird 3.0, released 19 April 2016, with focus in performance and security. A major re-architecture of the code allowed total support to SMP machines when using the SuperServer version. Through the Google Summer of Code 2013 work has begun on integrating Firebird as a replacement for HSQLDB in LibreOffice Base. Mozilla Firefox name conflict In April 2003, the Mozilla Organization announced a rename of its web browser from Phoenix to Firebird after a trademark dispute with Phoenix Technologies. This decision caused concern within the Firebird database project due to the assumption that users and Internet search engines would be confused by a database and a web browser both using the name Firebird. The Mozilla developers issued a statement, making clear that their software package was called "Mozilla Firebird", not "Firebird". The statement also said that the Mozilla Firebird name was a project codename. The dispute was resolved on 9 February 2004, when Mozilla changed the name of its browser to Mozilla Firefox, thus ending the conflict. Main features Full support for stored procedures and triggers Full ACID compliant transactions Referential integrity Multi Generational Architecture (sometimes called MVCC) Support for External Functions (UDFs) SQL activity can send asynchronous notification events to clients Third-party tools, including GUI administrative tools and replication tools Careful writes - fast recovery, no need for transaction logs Many access methods: native/API, dbExpress/FireDAC drivers, ODBC, OLE DB, .NET provider, JDBC native type 4 driver, Python module, PH
https://en.wikipedia.org/wiki/The%20Rockford%20Files
The Rockford Files is an American detective drama television series starring James Garner that aired on the NBC network from September 13, 1974, to January 10, 1980. Garner portrays Los Angeles private investigator Jim Rockford, with Noah Beery Jr. in the supporting role of his father, Joseph "Rocky" Rockford, a retired truck driver. The show was created by Roy Huggins and Stephen J. Cannell. Huggins had created the television show Maverick (1957–1962), which starred Garner, and he wanted to create a similar show in a modern-day detective setting. In 2002, The Rockford Files was ranked No. 39 on TV Guides 50 Greatest TV Shows of All Time. Premise Producers Roy Huggins and Stephen J. Cannell devised the Rockford character as a departure from typical television detectives, essentially Bret Maverick as a modern detective. In the series storyline, James Scott "Jim" Rockford had served time in California's San Quentin Prison in the 1960s due to a wrongful conviction. After five years, he was pardoned (not paroled, a distinction frequently mentioned in plot points). His jobs as a private investigator barely allow him to maintain his weathered mobile home (which doubles as his office) in a parking lot on a beach in Malibu, California. In early episodes of the first season, Rockford's trailer is located in a parking lot alongside the highway at 2354 Beach Boulevard (Pacific Coast Highway), Malibu, and near the ocean; for the rest of the series, the trailer is at Paradise Cove (address 29 Cove Road), adjacent to a pier and a restaurant ("The Sand Castle", now known as the "Paradise Cove Beach Cafe"). In the television movies from 1994 to 1999, Rockford is still living and working at the same Paradise Cove location, but in a much newer trailer that has been extensively enlarged and remodeled. In contrast to sharp-dressed, pugnacious television private eyes of the time, Rockford wears casual, off-the-rack clothing and tries to avoid physical altercations. He can hold his own in a one-on-one fistfight, but is frequently overpowered when ambushed or outnumbered, often from behind. But he almost always winds up figuring out what's going on, catching the bad guys/gals, and usually exacting revenge by the end of the episode, with some notable exceptions. He is experienced, observant, tenacious, quick-thinking, and has a faculty for impersonation and accents (usually Southern, drawing on Garner's Oklahoma background). He rarely carries his Colt Detective Special revolver, for which he has no permit and usually stores in a cookie jar; and prefers to talk his way out of trouble. He works on cold cases, missing persons investigations, and low-budget insurance scams, repeatedly stating that he does not handle "open cases" to avoid trouble with the police. (This self-imposed rule was relaxed in later seasons, after "trouble with the police" became a frequent plot device.) Rockford has been a private investigator since 1968–69, and his fee, when he can collect
https://en.wikipedia.org/wiki/LOL
LOL, or lol, is an initialism for laughing out loud and a popular element of Internet slang. It was first used almost exclusively on Usenet, but has since become widespread in other forms of computer-mediated communication and even face-to-face communication. It is one of many initialisms for expressing bodily reactions, in particular laughter, as text, including initialisms for more emphatic expressions of laughter such as LMAO ("laughing my ass off") and ROFL or ROTFL ("rolling on the floor laughing"). The list of acronyms "grows by the month" (said Peter Hershock in 2003), and they are collected along with emoticons and smileys into folk dictionaries that are circulated informally amongst users of Usenet, IRC, and other forms of (textual) computer-mediated communication. These initialisms are controversial, and several authors recommend against their use, either in general or in specific contexts such as business communications. The Oxford English Dictionary first listed LOL in March 2011. Origins In the 1980s, Wayne Pearson was reportedly the first person to have used LOL while responding to a friend's joke in a pre-Internet digital chat room called Viewline. Instead of writing "hahaha," as he had done before when he found something humorous, Pearson instead typed "LOL" to symbolize extreme laughter. Analysis Silvio Laccetti (professor of humanities at Stevens Institute of Technology) and Scott Molski, in their essay entitled The Lost Art of Writing, are critical of the terms, predicting reduced chances of employment for students who use such slang, stating that, "Unfortunately for these students, their bosses will not be 'lol' when they read a report that lacks proper punctuation and grammar, has numerous misspellings, various made-up words, and silly acronyms." Fondiller and Nerone in their style manual assert that "professional or business communication should never be careless or poorly constructed" whether one is writing an electronic mail message or an article for publication, and warn against the use of smileys and abbreviations, stating that they are "no more than e-mail slang and have no place in business communication". Linguist John McWhorter stated, "Lol is being used in a particular way. It's a marker of empathy. It's a marker of accommodation. We linguists call things like that pragmatic particles…" Pragmatic particles are the words and phrases utilized to alleviate the awkward areas in casual conversation, such as oh in "Oh, I don't know" and uh when someone is thinking of something to say. McWhorter stated that lol is utilized less as a reaction to something that is hilarious, but rather as a way to lighten the conversation. Frank Yunker and Stephen Barry, in a study of online courses and how they can be improved through podcasting, have found that these slang terms, and emoticons as well, are "often misunderstood" by students and are "difficult to decipher" unless their meanings are explained in advance. They single out
https://en.wikipedia.org/wiki/Windows%20Script%20Host
The Microsoft Windows Script Host (WSH) (formerly named Windows Scripting Host) is an automation technology for Microsoft Windows operating systems that provides scripting abilities comparable to batch files, but with a wider range of supported features. This tool was first provided on Windows 95 after Build 950a on the installation discs as an optional installation configurable and installable by means of the Control Panel, and then a standard component of Windows 98 (Build 1111) and subsequent and Windows NT 4.0 Build 1381 and by means of Service Pack 4. The WSH is also a means of automation for Internet Explorer via the installed WSH engines from IE Version 3.0 onwards; at this time VBScript became means of automation for Microsoft Outlook 97. The WSH is also an optional install provided with a VBScript and JScript engine for Windows CE 3.0 and following and some third-party engines including Rexx and other forms of Basic are also available. It is language-independent in that it can make use of different Active Scripting language engines. By default, it interprets and runs plain-text JScript (.JS and files) and VBScript (.VBS and files). Users can install different scripting engines to enable them to script in other languages, for instance PerlScript. The language independent filename extension WSF can also be used. The advantage of the Windows Script File (.WSF) is that it allows multiple scripts ("jobs") as well as a combination of scripting languages within a single file. WSH engines include various implementations for the Rexx, BASIC, Perl, Ruby, Tcl, PHP, JavaScript, Delphi, Python, XSLT, and other languages. Windows Script Host is distributed and installed by default on Windows 98 and later versions of Windows. It is also installed if Internet Explorer 5 (or a later version) is installed. Beginning with Windows 2000, the Windows Script Host became available for use with user login scripts. Usage Windows Script Host may be used for a variety of purposes, including logon scripts, administration and general automation. Microsoft describes it as an administration tool. WSH provides an environment for scripts to run – it invokes the appropriate script engine and provides a set of services and objects for the script to work with. These scripts may be run in GUI mode (WScript.exe) or command line mode (CScript.exe), or from a COM object (wshom.ocx), offering flexibility to the user for interactive or non-interactive scripts. Windows Management Instrumentation is also scriptable by this means. The WSH, the engines, and related functionality are also listed as objects which can be accessed and scripted and queried by means of the VBA and Visual Studio object explorers and those for similar tools like the various script debuggers, e.g. Microsoft Script Debugger, and editors. WSH implements an object model which exposes a set of Component Object Model (COM) interfaces. So in addition to ASP, IIS, Internet Explorer, CScript and WScript,
https://en.wikipedia.org/wiki/RSM
RSM may refer to: Companies RSM Global, worldwide accountancy & professional services network RSM US, a tax, accounting and consulting firm based in Chicago RSM UK RSM Singapore RSM Tenon, defunct professional services network in the UK RSM Robson Rhodes, defunct professional services network in the UK Reliable Source Music, an independent production music company Renault Samsung Motors, a South Korean motor manufacturer Institutions Rotterdam School of Management (RSM Erasmus University), the business school of Erasmus University, Rotterdam Royal School of Mines, an institute of higher education in London, England Royal Society of Medicine, a British medical society Rajalakshmi School of Management, a Business School based in Chennai, Tamil Nadu, India. Russian School of Mathematics, an after-school mathematical education program based in Massachusetts Other organizations Radioamaterski Sojuz na Makedonija, an amateur radio organization in the Republic of Macedonia Religious Order of the Sisters of Mercy, an order of Catholic women Rete degli Studenti Medi, Italian High School and VET student union Russian Socialist Movement, a socialist political party in Russia Rassemblement Saint-Martinois, a political party in Saint Martin Military Regimental sergeant major a warrant officer appointment in the British and most Commonwealth armies and the Royal Marines Resolute Support Mission, a NATO mission in Afghanistan from the end of 2014 Radio Squadron, Mobile, signal intelligence squadrons of the United States Air Force Security Service Places Rancho Santa Margarita, California, a city in California Republic of San Marino, a country in southern Europe Vehicle registration plates of San Marino Science and technology Rating scale model, a model used in Item Response Theory Reciprocal Space Map, a crystallographic method Response surface methodology, sequential experimentation for improvement and for finding an optimal response Reynolds stress model, a method in Computational fluid dynamics for modeling turbulence Route Switch Module, a component of a powered device that provides network routing services RNA-targeting small molecule drugs, a class of organic compounds with traditional drug properties Other uses R. Stevie Moore (born 1952), American singer-songwriter Rolling Stones Mobile Studio, a mobile recording studio
https://en.wikipedia.org/wiki/IPhoto
iPhoto is a discontinued digital photograph manipulation software application developed by Apple Inc. It was included with every Mac computer from 2002 to 2015, when it was replaced with Apple's Photos application. Originally sold as part of the iLife suite of digital media management applications, iPhoto is able to import, organize, edit, print and share digital photos. History iPhoto was announced at Macworld 2002, during which Steve Jobs (then-CEO of Apple) also announced that macOS would be the default operating system on new Macs, and revealed new iMac and iBook models. On March 7, 2012, Apple CEO Tim Cook announced an iOS-native version of iPhoto alongside the third-generation iPad. On June 27, 2014, Apple announced that they would cease development of iPhoto and work on a transition to their new Photos app. On February 5, 2015 Apple included a preview of Photos with a beta release of OS X Yosemite. On April 8, 2015, Apple released OS X Yosemite 10.10.3, which included the new Photos app. iPhoto and Aperture were discontinued and removed from the Mac App Store, but can still be downloaded by users who previously purchased them. macOS Mojave 10.14 was the last version of macOS to officially support iPhoto. However, a third party application allows iPhoto to run on macOS Catalina and newer. Features macOS version iPhoto is designed to allow the importing of pictures from digital cameras, local storage devices such as USB flash drives, CDs, DVDs and hardrives to a user's iPhoto Library. Almost all digital cameras are recognized without additional software. iPhoto supports most common image file formats, including several Raw image formats. iPhoto also supports videos from cameras, but editing is limited to trimming clips. After photos are imported, they can be titled, labeled, sorted and organized into groups (known as "events"). Individual photos can be edited with basic image manipulation tools, such as a red-eye filter, contrast and brightness adjustments, cropping and resizing tools, and other basic functions. iPhoto did not, however, provide the comprehensive editing functionality of programs such as Apple's own Aperture, Adobe's Photoshop, Album) or GIMP. iPhoto offers numerous options for sharing photos. Photo albums can be made into dynamic slideshows, with the option to add music imported from iTunes. Photos can be shared via iMessage, Mail, Facebook, Flickr and Twitter. Creating and sharing iCloud Photostreams is possible as well. iPhoto is also able to sync photo albums to any iPod with a color display. These iPods might also have an audio/video output to allow photos to be played back, along with music, on any modern television. Additionally, photos can be printed to a local printer, or, in certain markets, be sent over the internet to Kodak for professional printing. iPhoto users can order a range of products, including standard prints, posters, cards, calendars, and 100-page hardcover or softcover volumes, although such