source
stringlengths
32
199
text
stringlengths
26
3k
https://en.wikipedia.org/wiki/ARS%2B%2B
ARS++ was developed in 2002 for the book Undiluted Programming to demonstrate ARS based programming in a real world context. ARS++ is used in the book to implement an A++ interpreter and an XML Database System. Even the implementation of ARS++ in C was used to demonstrate ARS-based programming. Principally the programming language Scheme would have been perfectly all right to be used for these demonstration programs if Scheme had included primitives supporting network programming, database programming plus a few others required in this context. The definition of the programming language Scheme does not include those primitives however, leaving this issue up to the implementors of the language. The consequence is that there exist many Scheme implementations, almost none of them covering all areas essential for real application programs. This is the historical background that gave birth to the development of ARS++. ARS++ explained by its name The name ARS++, being an acronym for ARS + Scheme + Extensions, indicates that ARS++ has a lot to do with Scheme but that it is not equivalent to Scheme. The first part of the name refers to the core of the language, which is nothing else but A++, i.e. Abstraction + Reference + Synthesis. The second part in the name of ARS++ stands for the primitive functions that are imported from Scheme, giving ARS++ almost the same functionality as Scheme. Primitive functions are those, that are not and cannot be defined as lambda abstractions because they are representing a functionality that can only be provided by the underlying operating system or the hardware. The third part of the name refers to primitive functions that are not defined in R5RS (the official definition of the programming language Scheme) but are nevertheless important for real world programming like functions allowing to work with regular expressions, with TCP/IP networks, with embedded databases (e.g. Berkeley DB, GNU database manager) and a few that provide an interface with the operating system. From a practical point of view a Scheme implementation that includes support for regular expressions, databases, sockets and also provides an interface to the operating system can be called an ARS++ language. See also ARS-based programming A++ Educational programming language References Educational programming languages
https://en.wikipedia.org/wiki/Socialist%20Solidarity%20Network
The Socialist Solidarity Network was a grouping of socialists in the United Kingdom most of whom were former members of the Socialist Party. They supported the Socialist Alliance in England and support the Scottish Socialist Party in Scotland. The SSN's supporters included Lesley Mahmood, a long-time member of the Militant group, and Phil Hearse, former editor of Socialist Outlook. In 2002, the Network absorbed the Socialist Democracy Group. In 2003, launched the publication Socialist Resistance together with the International Socialist Group. The group ceased to function in the 2000s, as SR transformed itself into the USFI section in Britain, effectively re-absorbing some people such as Phil Hearse, while others such as Lesley Mahmood and Mandy Baker were no longer involved. References External links Socialist Resistance Defunct Trotskyist organisations in the United Kingdom Scottish Socialist Party 2000s disestablishments in the United Kingdom
https://en.wikipedia.org/wiki/StuffIt
StuffIt is a discontinued family of computer software utilities for archiving and compressing files. Originally produced for the Macintosh, versions for Microsoft Windows, Linux (x86), and Sun Solaris were later created. The proprietary compression format used by the StuffIt utilities is also termed StuffIt. In December 2019, Smith Micro Software, the product's most-recent owner and developer, officially announced that StuffIt had reached its end-of-life and that StuffIt products would no longer be developed. One last update did come out in December 2020 after the launch of the Apple M1 architecture to support that and Intel Mac systems through a universal binary of the program. Overview StuffIt was originally developed in the summer of 1987 by Raymond Lau, who was then a student at Stuyvesant High School in New York City. It combined the fork-combining capabilities of utilities such as MacBinary with newer compression algorithms similar to those used in ZIP. Compared to existing utilities on the Mac, notably PackIt, StuffIt offered "one step" operation and higher compression ratios. By the fall of 1987 StuffIt had largely replaced PackIt in the Mac world, with many software sites even going so far as to convert existing PackIt archives to save more space. StuffIt soon became very popular and Aladdin Systems was formed to market it (the last shareware release by Lau was version 1.5.1). They split the product line in two, offering StuffIt Classic in shareware and StuffIt Deluxe as a commercial package. Deluxe added a variety of additional functions, including additional compression methods and integration into the Mac Finder to allow files to be compressed from a "Magic Menu", or seamlessly browse inside and edit compressed files without expanding them using "True Finder Integration". StuffIt was upgraded several times, and Lau removed himself from direct development as major upgrades to the "internal machinery" were rare. Because new features and techniques appeared regularly on the Macintosh platform, the shareware utility Compact Pro emerged as a competitor to StuffIt in the early 1990s. A major competitive upgrade followed, accompanied by the release of the freeware StuffIt Expander, to make the format more universally readable, as well as the shareware StuffIt Lite which made it easier to produce. Prior to this anyone attempting to use the format needed to buy StuffIt, making Compact Pro more attractive. This move was a success, and Compact Pro subsequently fell out of use. Several other Mac compression utilities appeared and disappeared during the 1990s, but none became a real threat to StuffIt's dominance. The only ones to see any widespread use were special-purpose "disk expanders" like DiskDoubler and SuperDisk!, which served a different niche. Apparently as a side-effect, StuffIt once again saw few upgrades. The file format changed in a number of major revisions, leading to incompatible updates. PC-based formats long surpassed th
https://en.wikipedia.org/wiki/Inline%20function
In the C and C++ programming languages, an inline function is one qualified with the keyword inline; this serves two purposes: It serves as a compiler directive that suggests (but does not require) that the compiler substitute the body of the function inline by performing inline expansion, i.e. by inserting the function code at the address of each function call, thereby saving the overhead of a function call. In this respect it is analogous to the register storage class specifier, which similarly provides an optimization hint. The second purpose of inline is to change linkage behavior; the details of this are complicated. This is necessary due to the C/C++ separate compilation + linkage model, specifically because the definition (body) of the function must be duplicated in all translation units where it is used, to allow inlining during compiling, which, if the function has external linkage, causes a collision during linking (it violates uniqueness of external symbols). C and C++ (and dialects such as GNU C and Visual C++) resolve this in different ways. Example An inline function can be written in C or C++ like this: inline void swap(int *m, int *n) { int tmp = *m; *m = *n; *n = tmp; } Then, a statement such as the following: swap(&x, &y); may be translated into (if the compiler decides to do the inlining, which typically requires optimization to be enabled): int tmp = x; x = y; y = tmp; When implementing a sorting algorithm doing lots of swaps, this can increase the execution speed. Standard support C++ and C99, but not its predecessors K&R C and C89, have support for inline functions, though with different semantics. In both cases, inline does not force inlining; the compiler is free to choose not to inline the function at all, or only in some cases. Different compilers vary in how complex a function they can manage to inline. Mainstream C++ compilers like Microsoft Visual C++ and GCC support an option that lets the compilers automatically inline any suitable function, even those not marked as inline functions. However, simply omitting the inline keyword to let the compiler make all inlining decisions is not possible, since the linker will then complain about duplicate definitions in different translation units. This is because inline not only gives the compiler a hint that the function should be inlined, it also has an effect on whether the compiler will generate a callable out-of-line copy of the function (see storage classes of inline functions). Nonstandard extensions GNU C, as part of the dialect gnu89 that it offers, has support for inline as an extension to C89. However, the semantics differ from both those of C++ and C99. armcc in C90 mode also offers inline as a non-standard extension, with semantics different from gnu89 and C99. Some implementations provide a means by which to force the compiler to inline a function, usually by means of implementation-specific declaration specifiers: Microsoft Visual C++: __forc
https://en.wikipedia.org/wiki/Valgrind
Valgrind () is a programming tool for memory debugging, memory leak detection, and profiling. Valgrind was originally designed to be a freely licensed memory debugging tool for Linux on x86, but has since evolved to become a generic framework for creating dynamic analysis tools such as checkers and profilers. Overview Valgrind is in essence a virtual machine using just-in-time compilation techniques, including dynamic recompilation. Nothing from the original program ever gets run directly on the host processor. Instead, Valgrind first translates the program into a temporary, simpler form called intermediate representation (IR), which is a processor-neutral, static single assignment form-based form. After the conversion, a tool (see below) is free to do whatever transformations it would like on the IR, before Valgrind translates the IR back into machine code and lets the host processor run it. Valgrind recompiles binary code to run on host and target (or simulated) CPUs of the same architecture. It also includes a GDB stub to allow debugging of the target program as it runs in Valgrind, with "monitor commands" that allow querying the Valgrind tool for various information. A considerable amount of performance is lost in these transformations (and usually, the code the tool inserts); usually, code run with Valgrind and the "none" tool (which does nothing to the IR) runs at 20% to 25% of the speed of the normal program. Tools Memcheck There are multiple tools included with Valgrind (and several external ones). The default (and most used) tool is Memcheck. Memcheck inserts extra instrumentation code around almost all instructions, which keeps track of the validity (all unallocated memory starts as invalid or "undefined", until it is initialized into a deterministic state, possibly from other memory) and addressability (whether the memory address in question points to an allocated, non-freed memory block), stored in the so-called V bits and A bits respectively. As data is moved around or manipulated, the instrumentation code keeps track of the A and V bits, so they are always correct on a single-bit level. In addition, Memcheck replaces the standard C memory allocator with its own implementation, which also includes memory guards around all allocated blocks (with the A bits set to "invalid"). This feature enables Memcheck to detect off-by-one errors where a program reads or writes outside an allocated block by a small amount. The problems Memcheck can detect and warn about include the following: Use of uninitialized memory Reading/writing memory after it has been free'd Reading/writing off the end of malloc'd blocks Memory leaks The price of this is lost performance. Programs running under Memcheck usually run 20–30 times slower than running outside Valgrind and use more memory (there is a memory penalty per allocation). Thus, few developers run their code under Memcheck (or any other Valgrind tool) all the time. They most commonly use such
https://en.wikipedia.org/wiki/Splint%20%28programming%20tool%29
Splint, short for Secure Programming Lint, is a programming tool for statically checking C programs for security vulnerabilities and coding mistakes. Formerly called LCLint, it is a modern version of the Unix lint tool. Splint has the ability to interpret special annotations to the source code, which gives it stronger checking than is possible just by looking at the source alone. Splint is used by gpsd as part of an effort to design for zero defects. Splint is free software released under the terms of the GNU General Public License. Main development activity on Splint stopped in 2010. According to the CVS at SourceForge, as of September 2012 the most recent change in the repository was in November 2010. A Git repository at GitHub has more recent changes, starting in July 2019. Example #include <stdio.h> int main() { char c; while (c != 'x'); { c = getchar(); if (c = 'x') return 0; switch (c) { case '\n': case '\r': printf("Newline\n"); default: printf("%c",c); } } return 0; } Splint's output: <nowiki> Variable c used before definition Suspected infinite loop. No value used in loop test (c) is modified by test or loop body. Assignment of int to char: c = getchar() Test expression for if is assignment expression: c = 'x' Test expression for if not boolean, type char: c = 'x' Fall through case (no preceding break) </nowiki> Fixed source: #include <stdio.h> int main() { int c = 0; // Added an initial assignment definition. while (c != 'x') { c = getchar(); // Corrected type of c to int if (c == 'x') // Fixed the assignment error to make it a comparison operator. return 0; switch (c) { case '\n': case '\r': printf("Newline\n"); break; // Added break statement to prevent fall-through. default: printf("%c",c); break; //Added break statement to default catch, out of good practice. } } return 0; } See also Buffer overflow Memory debugger Software testing List of tools for static code analysis References External links Static program analysis tools Free memory management software Cross-platform software Free software testing tools Software using the GPL license
https://en.wikipedia.org/wiki/Left%20shift
Left shift may refer to: Left shift (medicine), a medical term similar to blood shift Logical left shift, a computer operation Arithmetic left shift, a computer operation Left Shift key, a key on a computer keyboard Left Shift (political group) (aka Linksruck), a former Trotskyist group in Germany Left shift (quality assurance), thinking about quality earlier in the product development lifecycle See also Right shift (disambiguation)
https://en.wikipedia.org/wiki/Frontend%20and%20backend
In software engineering, the terms frontend and backend (sometimes written as back end or back-end) refer to the separation of concerns between the presentation layer (frontend), and the data access layer (backend) of a piece of software, or the physical infrastructure or hardware. In the client–server model, the client is usually considered the frontend and the server is usually considered the backend, even when some presentation work is actually done on the server itself. Introduction In software architecture, there may be many layers between the hardware and end user. The front is an abstraction, simplifying the underlying component by providing a user-friendly interface, while the back usually handles data storage and business logic. In telecommunication, the front can be considered a device or service, while the back is the infrastructure that supports provision of service. A rule of thumb is that the client-side (or "frontend") is any component manipulated by the user. The server-side (or "backend") code usually resides on the server, often far removed physically from the user. Software definitions In content management systems, the terms frontend and backend may refer to the end-user facing views of the CMS and the administrative views, respectively. In speech synthesis, the frontend refers to the part of the synthesis system that converts the input text into a symbolic phonetic representation, and the backend converts the symbolic phonetic representation into actual sounds. In compilers, the frontend translates a computer programming source code into an intermediate representation, and the backend works with the intermediate representation to produce code in a computer output language. The backend usually optimizes to produce code that runs faster. The frontend/backend distinction can separate the parser section that deals with source code and the backend that generates code and optimizes. Some designs, such as GCC, offer choices between multiple frontends (parsing different source languages) or backends (generating code for different target processors). Some graphical user interface (GUI) applications running in a desktop environment are implemented as a thin frontend for underlying command-line interface (CLI) programs, to save the user from learning the special terminology and memorizing the commands. Web development as an example Another way to understand the difference between the two is to understand the knowledge required of a frontend vs. a backend software developer. The list below focuses on web development as an example. Both Version control tools such as Git, Mercurial, or Subversion File transfer tools and protocols such as FTP or rsync Frontend focused Markup and web languages such as HTML, CSS, JavaScript, and ancillary libraries commonly used in those languages such as Sass or jQuery Asynchronous request handling and AJAX Single-page applications (with frameworks like React, Angular or Vue.js) Web per
https://en.wikipedia.org/wiki/Drag%20and%20drop
In computer graphical user interfaces, drag and drop is a pointing device gesture in which the user selects a virtual object by "grabbing" it and dragging it to a different location or onto another virtual object. In general, it can be used to invoke many kinds of actions, or create various types of associations between two abstract objects. As a feature, drag-and-drop support is not found in all software, though it is sometimes a fast and easy-to-learn technique. However, it is not always clear to users that an item can be dragged and dropped, or what is the command performed by the drag and drop, which can decrease usability. Actions The basic sequence involved in drag and drop is: Move the pointer to the object Press, and hold down, the button on the mouse or other pointing device, to "grab" the object "Drag" the object to the desired location by moving the pointer to this one "Drop" the object by releasing the button Dragging requires more physical effort than moving the same pointing device without holding down any buttons. Because of this, a user cannot move as quickly and precisely while dragging (see Fitts' law). However, drag-and-drop operations have the advantage of thoughtfully chunking together two operands (the object to drag, and the drop location) into a single action. Extended dragging and dropping (as in graphic design) can stress the mousing hand. A design problem appears when the same button selects and drags items. Imprecise movement can cause an attempt to select an object to register as a dragging motion. Another problem is that the target of the dropping can be hidden under other objects. The user would have to stop the dragging, make both the source and the target visible and start again. In classic Mac OS the top-of-screen menu bar served as a universal "drag cancel" target. This issue has been dealt with in Mac OS X with the introduction of Exposé. In Mac OS Drag and drop, called click and drag at the time, was used in the original Macintosh to manipulate files (for example, copying them between disks or folders). System 7 added the ability to open a document in an application by dropping the document icon onto the application's icon. Apple added "Macintosh Drag and Drop" to System 7.5, extending "click and drag" to common clipboard operations like copying or moving textual content within a document. Content could also be dragged into the filesystem to create a "clipping file" which could then be stored and reused. Files could also be dropped on application windows, for example to enclose a document in an email, or add an image to a word processor document. For most of its history Mac OS has used a single button mouse with the button covering a large portion of the top surface of the mouse. This may mitigate the ergonomic concerns of keeping the button pressed while dragging. In OS/2 The Workplace Shell of OS/2 uses dragging and dropping extensively with the secondary mouse button, leaving the primary on
https://en.wikipedia.org/wiki/Point%20and%20click
Point and click are one of the actions of a computer user moving a pointer to a certain location on a screen (pointing) and then pressing a button on a mouse or other pointing device (click). An example of point and click is in hypermedia, where users click on hyperlinks to navigate from document to document. User interfaces, for example graphical user interfaces, are sometimes described as "point-and-click interfaces", often to suggest that they are very easy to use, requiring that the user simply point to indicate their wishes. Describing software this way implies that the interface can be controlled solely through a pointing device with little or no input from the keyboard, as with many graphical user interfaces. In some systems, such as Internet Explorer, moving the pointer over a link (or other GUI control) and waiting for a split-second will cause a tooltip to be displayed. Single click A single click or "click" is the act of pressing a computer mouse button once without moving the mouse. Single clicking is usually a primary action of the mouse. Single clicking, by default in many operating systems, selects (or highlights) an object while double-clicking executes or opens the object. The single-click has many advantages over double click due to the reduced time needed to complete the action. The single-click or one-click phrase has also been used to apply to the commercial field as a competitive advantage. The slogan "single click" or "one-click" commonly advertises services' ease of use. On icons By default on most computer systems, for a person to select a certain software function, they will have to click on the left button. An example of this can be a person clicking on an icon. Similarly, clicking on the right button will present the user with a text menu to select more actions. These actions can range from open, explore, properties, etc. In terms of entertainment software, point-and-click interfaces are common input methods, usually offering a 'menu' or 'icon bar' interface that functions expectedly. In other games, the character explores different areas within the game world. To move to another area, the player will move the cursor to one point of the screen, where the cursor will turn into an arrow. Clicking will then move the player to that area. On text In many text processing programs, such as web browsers or word processors, clicking on text moves the cursor to that location. Clicking and holding the left button allows users to highlight the selected text. This enables more options to edit or use the text. Double click A double click is most commonly used with a computer mouse when the pointer is placed over an icon or object and the button is quickly pressed twice without moving the mouse. Fitts's Law Fitts's law can be used to quantify the time required to perform a point-and-click action. where: is the average time taken to complete a single movement. represents the start/stop time of the device and stands
https://en.wikipedia.org/wiki/LAN%20party
A LAN party is a social gathering of participants with personal computers or compatible game consoles, where a local area network (LAN) connection is established between the devices using a router or switch, primarily for the purpose of playing multiplayer video games together. LAN party events differ significantly from LAN gaming centers and Internet cafes in that LAN parties generally require participants to bring your own computer (BYOC) and are not permanent installations, often taking place in general-use venues or residences. The size of these networks may vary from as few as two people to very large gatherings of a hundred or more. Small parties can form spontaneously and take advantage of common household networking equipment, but larger ones typically require more planning, equipment and preparation, even dedicated gaming servers. As of 2020, the world record for the size of a LAN party is 22,810 visitors, set at DreamHack, in Jönköping, Sweden. Small parties Usually, smaller LAN parties consist of people bringing their computers over to each other's houses to host and play multiplayer games. These are sometimes established between small groups of friends, and hosted at a central location or one that is known to all participants. Such events are often organized quickly with little planning, and some overnight events, with some stretching into days (or even weeks). Because of the small number of players, games are usually played on small levels and/or against bots. A small LAN party requires either a network switch, with enough ports to accommodate all the players, or if all the computers have Wi-Fi capability, an ad hoc network may be set up. This allows two or more computers to connect over a wireless connection, thereby eliminating the need for a wired network, a fair amount of power, and suitable surfaces for all the computers. Providing refreshments is often also a duty of the host, though guests are usually asked to contribute. In larger parties where participants may not all know each other personally, an entry fee may even be charged. Another tradition of some small groups is to purchase large amounts of fast food for consumption over many days. Many LAN participants will also bring food or drink to consume over the course of the party—though they can be held at any hour, many LAN parties begin late in the evening and run through the next morning, making energy drinks a popular choice. When some of the participants cannot be present or when merging a few LAN parties together, VPN software such as Hamachi can be used to arrange computers over the Internet so they appear to be on the same LAN. Normally, the host will host the games but sometimes at very small LAN parties (e.g. 2 or 3 people) all participants will connect to an online internet server and add a word in front of their name to tell everyone else that they are a clan or group. At bigger LANs (e.g. 5 or more people) the host or a friend of the host will use a spare
https://en.wikipedia.org/wiki/Public-access%20television
Public-access television (sometimes called community-access television) is traditionally a form of non-commercial mass media where the general public can create content television programming which is narrowcast through cable television specialty channels. Public-access television was created in the United States between 1969 and 1971 by the Federal Communications Commission (FCC), under Chairman Dean Burch, based on pioneering work and advocacy of George Stoney, Red Burns (Alternate Media Center), and Sidney Dean (City Club of NY). Public-access television is often grouped with public, educational, and government access television channels, under the acronym PEG. Distinction from PBS In the United States, the Public Broadcasting Service (PBS) produces public television, offering an educational television broadcasting service of professionally produced, highly curated content. It is not public-access television, and has no connection with cable-only PEG television channels. Although non-commercial educational television bears some resemblance to the E of PEG, PBS bears little resemblance to public-access television. PBS generally does not offer local programming content. Instead, it broadcasts content produced for a national audience distributed via satellites. There is no generally accepted right of access for citizens to use broadcast studio facilities of PBS member stations, nor right of access by community content producers to the airwaves stewarded by these television stations outside of some universities or technical colleges such as Milwaukee's Milwaukee Area Technical College, which owns the area's two PBS member stations and offers students the limited ability (within FCC guidelines) to produce their own programs to air on a public television station for television production experience. These qualities are in stark contrast to PEG channel content, which is mostly locally produced, especially in conjunction with local origination studio facilities. And in the case of the P, public-access television, the facilities and channel capacity are uncurated free-speech zones available to anyone for free or little cost. Since 53% to 60% of public television's revenues come from private membership donations and grants, most stations solicit individual donations by methods including fundraising, pledge drives, or telethons which can disrupt regularly scheduled programming. PBS is also funded by the federal government of the United States. PEG channels are generally funded by cable television companies through revenues derived from cable television franchise fees, member fees, grants and contributions. PEG-TV Public, educational, and government access television (also PEG-TV, PEG channel, PEGA, local-access television) refers to three different cable television narrowcasting and specialty channels. Public-access television was created in the United States between 1969 and 1971 by the Federal Communications Commission (FCC) and has since been
https://en.wikipedia.org/wiki/UNIVAC%20Solid%20State
The UNIVAC Solid State was a magnetic drum-based solid-state computer announced by Sperry Rand in December 1958 as a response to the IBM 650. It was one of the first computers offered for sale to be (nearly) entirely solid-state, using 700 transistors, and 3000 magnetic amplifiers (FERRACTOR) for primary logic, and 20 vacuum tubes largely for power control. It came in two versions, the Solid State 80 (IBM-style 80-column cards) and the Solid State 90 (Remington-Rand 90-column cards). In addition to the "80/90" designation, there were two variants of the Solid State the SS I 80/90 and the SS II 80/90. The SS II series included two enhancements the addition of 1,280 words of core memory and support for magnetic tape drives. The SS I had only the standard 5,000-word drum memory described in this article and no tape drives. The memory drum had a regular access speed AREA and a FAST ACCESS AREA. 4,000 words of memory had one set of R/W heads to access. The programmer was required to keep track of what words of memory where under the R/W heads and available to be read or written. At worst the program would have to wait for a full revolution of the drum to access the required memory locations. However 1,000 words of memory had 4 sets of R/W heads requiring only a 90 degree turn of the drum to access the required words. Programming required that any function that changed the contents of a memory location had first to transfer the contents of the affected word from the drum to a static register. There were 3 of these registers A X L, to add the values contained in drum memory locations the programmer would transfer the contents of the specific drum location to register A, then the second operand would be copied to the X register. The ADD INSTRUCTION WOULD BE EXECUTED leaving the result in the X register. The contents of the X register would then be written back to the appropriate word on the drum. Both variants included a card reader, a card punch, and the line printer described in this article. The only "console" was a 10-key adding machine-type keypad, from which the operator would enter the commands to boot the computer. That keypad was also used by programmers in the debugging process. There was no Operating System as we have come to know them in recent years; every program was completely self-contained, including the boot loader that initiated execution. All programs were loaded from punched cards; even on the SS II, with its tape drives, there was no ability to launch programs from those drives. The SS II, including two tape drives, weighed about . Architecture The UNIVAC Solid State was a two-address, bi-quinary coded decimal computer using signed 10-digit words. Main memory storage was provided by a 5000-word magnetic drum spinning at 17,667 RPM in a helium atmosphere. For efficiency, programmers had to take into account drum latency, the time required for a specific data item, once written, to rotate to where it could be read. Tec
https://en.wikipedia.org/wiki/TVO
TVO (stylized in all lowercase as tvo), formerly known as TVOntario, is a publicly funded English-language educational television network and media organization serving the Canadian province of Ontario. It operates flagship station CICA-DT (channel 19) in Toronto, which also relays programming across portions of Ontario through eight rebroadcast stations. All pay television (cable, satellite, IPTV) providers throughout Ontario are required to carry TVO on their basic tier, and programming can be streamed for free online within Canada. TVO is operated by the Ontario Educational Communications Authority (OECA), a Crown corporation owned by the Government of Ontario, which since 2022 has done business as the TVO Media Education Group (or TVO.me). TVO.me also operates TVO Today, TVO ILC, TVO Learn, and TVOKids. Governance, funding and other responsibilities TVO is governed by a volunteer board of directors, and supported by a network of regional councillors from across the province. TVO also reports to the Ontario legislature through the Minister of Education, in accordance with the Ontario Educational Communications Authority Act. Instead of following the model of the federally owned Canadian Broadcasting Corporation (CBC)'s television services, which shows commercial advertisements, TVO chose a commercial-free model similar to the Public Broadcasting Service (PBS) in the United States (in fact, various TVO productions wound up being aired on PBS stations). This model was later emulated by provincial educational broadcasters Télé-Québec in Quebec and Knowledge Network in British Columbia. The majority of TVO's funding is provided by the Government of Ontario through the Ministry of Education, which provides $39 million annually, with additional funding provided by charitable donations. TVO is also responsible for over-the-air broadcasts of the Ontario Legislative Assembly in some remote Northern Ontario communities that do not receive cable television access to the Ontario Parliament Network. In 2002, the Ministry of Education transferred responsibility for the Independent Learning Centre—the agency which provides distance education at the elementary and secondary school level—to TVO. TVO used to operate TFO (Télévision française de l'Ontario), a separate but similar network for Franco-Ontarian audiences. Before the launch of TFO, TVO aired French-language programming on Sundays. Even after TFO's launch, TVO and TFO swapped programming on Sundays well into the 1990s. TFO was separated from TVO and was incorporated under the newly formed GroupeMédia TFO, a separate Crown corporation of the Government of Ontario, in 2007. In 2017 and 2018, TVO launched four regional "hubs", featuring journalism on issues in the various regions of Ontario, on its website. Hubs are currently based in Thunder Bay for the Northwestern Ontario region, Sudbury for Northeastern Ontario, Kingston for Eastern Ontario, and London for Southwestern Ontario. In 2019, the se
https://en.wikipedia.org/wiki/Noovo
Noovo is a Canadian French-language terrestrial television network owned by the Bell Media subsidiary of BCE Inc. The network has five owned-and-operated and three affiliated stations throughout Quebec. It can also be seen over-the-air in some bordering markets in the provinces of Ontario and New Brunswick, and in some other parts of Canada on cable television or direct broadcast satellite. The network was launched in 1986 as Télévision Quatre-Saisons (TQS), and was known by that name until Remstar, which had bought the network in 2008, renamed it V on August 31, 2009. It was the namesake and flagship property of V Media Group (now known as Remstar Media Group), a separate company majority-owned by Remstar owner Maxime Rémillard (partially through Remstar). V was acquired by Bell Media in May 2020, after which it was renamed Noovo on August 31, 2020. The name "Noovo" is a stylized phonetic spelling of "nouveau", the French word for "new". History In 1968, the Canadian Radio-television and Telecommunications Commission (CRTC) first expressed interest in the establishment of a third French-language commercial television service in the province of Quebec to compete with Télévision de Radio-Canada and the loose association of independent stations that eventually became TVA. However the CRTC did not call on applications for licences. In 1972, the CRTC said it was prepared to receive licence applications in order to authorize a third commercial television service in Quebec, although it was not until 1974 when the CRTC granted licences to Télé Inter-Cité Québec Ltée. to operate TV stations in Montreal (channel 29) and in Quebec City (channel 2). Télé Inter-Cité found itself unable to launch the network due to materials shortages and delays in equipment delivery; the CRTC granted a time extension to 1976. Civitas Corp., owner of several radio stations in Quebec and a denied applicant for the same channels a year earlier, filed to buy Téle Inter-Cité, but the CRTC denied the purchase and noted that the proposal to reduce local programming commitments substantially altered the original accepted application. Unable to go forward due to what it called "economic reasons", the firm surrendered the licences for revocation in 1976. On November 15, 1984, the CRTC launched another call for applications in response to a bid from Cogeco. In 1985, it held public hearings in Montreal to examine competing applications from partners Cogeco Inc. (60.3%) and Moffat Communications (39.7%), and another application by the Pouliot family, owners of Montreal's CTV affiliate, CFCF-TV and radio stations CFCF (later CINW, now defunct) and CFQR-FM (now CKBE-FM). Both applications applied to launch television stations in Montreal and Quebec City. On September 6 of that year, the CRTC approved the application of the Pouliot family and its company, Réseau de Télévision Quatre-Saisons Inc., noting its existing facilities in Montreal and more realistic revenue projections compared
https://en.wikipedia.org/wiki/SSH%20File%20Transfer%20Protocol
In computing, the SSH File Transfer Protocol (also known as Secure File Transfer Protocol or SFTP) is a network protocol that provides file access, file transfer, and file management over any reliable data stream. It was designed by the Internet Engineering Task Force (IETF) as an extension of the Secure Shell protocol (SSH) version 2.0 to provide secure file transfer capabilities, and is seen as a replacement of File Transfer Protocol (FTP) due to superior security. The IETF Internet Draft states that, even though this protocol is described in the context of the SSH-2 protocol, it could be used in a number of different applications, such as secure file transfer over Transport Layer Security (TLS) and transfer of management information in VPN applications. This protocol assumes that it is run over a secure channel, such as SSH, that the server has already authenticated the client, and that the identity of the client user is available to the protocol. Capabilities Compared to the SCP protocol, which only allows file transfers, the SFTP protocol allows for a range of operations on remote files which make it more like a remote file system protocol. An SFTP client's extra capabilities include resuming interrupted transfers, directory listings, and remote file removal. There is also support for all UNIX file types, including symbolic links. SFTP attempts to be more platform-independent than SCP; with SCP, for instance, the expansion of wildcards specified by the client is up to the server, whereas SFTP's design avoids this problem. While SCP is most frequently implemented on Unix platforms, SFTP servers are commonly available on most platforms. In SFTP, the file transfer can be easily terminated without terminating a session like other mechanisms do. SFTP is not FTP run over SSH, but rather a new protocol designed from the ground up by the IETF SECSH working group. It is sometimes confused with Simple File Transfer Protocol. The protocol itself does not provide authentication and security; it expects the underlying protocol to secure this. SFTP is most often used as subsystem of SSH protocol version 2 implementations, having been designed by the same working group. It is possible, however, to run it over SSH-1 (and some implementations support this) or other data streams. Running an SFTP server over SSH-1 is not platform-independent as SSH-1 does not support the concept of subsystems. An SFTP client willing to connect to an SSH-1 server needs to know the path to the SFTP server binary on the server side. Uploaded files may be associated with their basic attributes, such as time stamps. This is an advantage over the common FTP protocol. History and development The Internet Engineering Task Force (IETF) working group "Secsh" that was responsible for the development of the Secure Shell version 2 protocol (RFC 4251) also attempted to draft an extension of that standard for secure file transfer functionality. Internet Drafts were created that succe
https://en.wikipedia.org/wiki/Information%20art
Information art, which is also known as informatism or data art, is an emerging art form that is inspired by and principally incorporates data, computer science, information technology, artificial intelligence, and related data-driven fields. The information revolution has resulted in over-abundant data that are critical in a wide range of areas, from the Internet to healthcare systems. Related to conceptual art, electronic art and new media art, informatism considers this new technological, economical, and cultural paradigm shift, such that artworks may provide social commentaries, synthesize multiple disciplines, and develop new aesthetics. Realization of information art often take, although not necessarily, interdisciplinary and multidisciplinary approaches incorporating visual, audio, data analysis, performance, and others. Furthermore, physical and virtual installations involving informatism often provide human-computer interaction that generate artistic contents based on the processing of large amounts of data. Background Information art has a long history as visualization of qualitative and quantitative data forms a foundation in science, technology, and governance. Information design and informational graphics, which has existed before computing and the Internet, are closely connected with this new emergent art movement. An early example of informatism the 1970 exhibition organized called "Information" at the Museum of Modern Art in New York City (curated by Kynaston McShine). This is the time when conceptual art has emerged as a leading tendency in the United States and internationally. At the same time arose the activities of Experiments in Art and Technology known as E.A.T. Contemporary practices Information art are manifested using a variety of data sources such as photographs, census data, video clips, search engine results, digital painting, network signals, and others. Often, such data are transformed, analyzed, and interpreted in order to convey concepts and develop aesthetics. When dealing with big data, artists may use statistics and machine learning to seek meaningful patterns that drive audio, visual, and other forms of representations. Recently, informatism is used in interactive and generative installations that are often dynamically linked with data and analytical pipelines. See also Examples The Tempestry Project Warming stripes Climate spiral Related subjects Algorithmic art Climate change art Computer art Conceptual art Data visualization Digital art Experiments in Art and Technology Generative art Knowledge visualization Post-conceptual art Roy Ascott Software art Systems art Systems thinking References Further reading Alan Liu (2004). "The Laws of Cool: Knowledge Work and the Culture of Information", University of Chicago Press Kenneth R. Allan, "Understanding Information," in Michael Corris (ed.), Conceptual Art, Theory, Myth, and Practice (Cambridge: Cambridge University Press, 2004), 1
https://en.wikipedia.org/wiki/GLORIAD
GLORIAD (Global Ring Network for Advanced Application Development) is a high-speed computer network used to connect scientific organizations in Russia, China, United States, the Netherlands, Korea and Canada. India, Singapore, Vietnam, and Egypt were added in 2009. GLORIAD is sponsored by the US National Science Foundation, a consortium of science organizations and Ministries in Russia, the Chinese Academy of Sciences, the Ministry of Science and Technology of Korea, the Canadian CANARIE network, the national research network in The Netherlands SURFnet and has some telecommunications services donated by Tyco Telecommunications. GLORIAD provides bandwidth of up to 10 Gbit/s via OC-192 links, e.g. between KRLight in Korea and the Pacific NorthWest GigaPOP in the United States. The previous version of the network, "Little GLORIAD", was completed in mid-2004, and it connected Chicago, Hong Kong, Beijing, Novosibirsk, Moscow, Amsterdam and Chicago again. For this network, a direct computer link was drawn between Russia and China for the first time in history. References National Science Foundation (2003): United States, Russia, China link up first Global-Ring Network for advanced science and education cooperation. Retrieved January 13, 2004 from https://www.sciencedaily.com/releases/2004/01/040102092834.htm Paul, J. (2003): New network to link U.S., Russia, China. Retrieved January 13, 2004 from http://apnews.excite.com/article/20031223/D7VK5LOG2.html External links GLORIAD Academic computer network organizations
https://en.wikipedia.org/wiki/Gamut
In color reproduction, including computer graphics and photography, the gamut, or color gamut , is a certain complete subset of colors. The most common usage refers to the subset of colors that can be accurately represented in a given circumstance, such as within a given color space or by a certain output device. Another sense, less frequently used but still correct, refers to the complete set of colors found within an image at a given time. In this context, digitizing a photograph, converting a digitized image to a different color space, or outputting it to a given medium using a certain output device generally alters its gamut, in the sense that some of the colors in the original are lost in the process. Introduction The term gamut was adopted from the field of music, where in middle age Latin "gamut" meant the entire range of musical notes of which musical melodies are composed; Shakespeare's use of the term in The Taming of the Shrew is sometimes attributed to the author / musician Thomas Morley. In the 1850s, the term was applied to a range of colors or hue, for example by Thomas de Quincey, who wrote "Porphyry, I have heard, runs through as large a gamut of hues as marble." In color theory, the gamut of a device or process is that portion of the color space that can be represented, or reproduced. Generally, the color gamut is specified in the hue–saturation plane, as a system can usually produce colors over a wide intensity range within its color gamut; for a subtractive color system (such as used in printing), the range of intensity available in the system is for the most part meaningless without considering system-specific properties (such as the illumination of the ink). When certain colors cannot be expressed within a particular color model, those colors are said to be out of gamut. A device that can reproduce the entire visible color space is an unrealized goal within the engineering of color displays and printing processes. Modern techniques allow increasingly good approximations, but the complexity of these systems often makes them impractical. While processing a digital image, the most convenient color model used is the RGB model. Printing the image requires transforming the image from the original RGB color model to the printer's CMYK color model. During this process, the colors from the RGB model which are out of gamut must be somehow converted to approximate values within the CMYK model. Simply trimming only the colors which are out of gamut to the closest colors in the destination space would burn the image. There are several algorithms approximating this transformation, but none of them can be truly perfect, since those colors are simply out of the target device's capabilities. This is why identifying the colors in an image that are out of gamut in the target color space as soon as possible during processing is critical for the quality of the final product. It is also important to remember that there are colors inside the
https://en.wikipedia.org/wiki/TVA%20%28Canadian%20TV%20network%29
TVA is a Canadian French-language terrestrial television network, owned by Groupe TVA, a publicly traded subsidiary of Quebecor Media. Headquartered in Montreal, the network only has terrestrial stations in Quebec. However, parts of New Brunswick and Ontario are within the broadcast ranges of TVA stations, and two TVA stations operate rebroadcasters in New Brunswick. Since becoming a national network in 1998, it has been available on cable television across Canada. TVA is short for Téléviseurs associés (roughly translated to "Associated Telecasters"). This reflects the network's roots as a cooperative. Overview TVA traces its roots to 1963, when CJPM-TV in Chicoutimi, a station only a few months old and in need of revenue, began sharing programs with the largest privately owned francophone station in Canada, CFTM-TV in Montreal. They were joined by CFCM-TV in Quebec City in 1964 after CFCM lost its Radio-Canada affiliation to newly-launched CBVT. While the three stations shared programs for many years, it was not until September 12, 1971, that the informal link became a proper network, TVA, with CFTM as the flagship station. The network began the first private French-language network news service in Canada in 1972. Between 1973 and 1983, seven more stations joined the network. When the network was formally organized in 1971, its affiliates ran it as a cooperative, much like CTV operated for many years. In 1982, the cooperative became a corporation with the station owners as shareholders. For many years, TVA's schedule was very similar to that of what CTV offered before Baton Broadcasting took over the network in that it did not have what could be called a main schedule aside from news. For instance, Pathonic Communications, which owned the TVA affiliates in Quebec City, Sherbrooke, Trois-Rivières and Rimouski and provided programming to the affiliates in Rivière-du-Loup and Carleton, offered programming that was different from that offered on CFTM. The differences were enough that Sherbrooke's CHLT-TV, whose over-the-air signal reached Montreal, was carried on Montreal cable systems well into the 1990s. However, CFTM dominated the network to an even greater extent that Toronto's CFTO-TV dominated CTV, contributing as much of 90% of the network's programming. In 1989, Télé-Metropole, which owned CFTM and CJPM, bought out Pathonic. The other station owners sold the outstanding shares of the network in 1992. Nine years later, Quebecor became owner of TVA. TVA also owns Le Canal Nouvelles (LCN), Canada's only private French-language headline-news channel. When TVA completes its broadcast day, the TVA stations simulcast LCN until TVA's next broadcast day begins. As well, the company owns a magazine publishing division unit, a film production and distribution house, and a number of other Internet and cable properties, many of which are often used to cross-promote TVA series and events. In 1998, the Canadian Radio-television and Telecommunicat
https://en.wikipedia.org/wiki/Empirical%20orthogonal%20functions
In statistics and signal processing, the method of empirical orthogonal function (EOF) analysis is a decomposition of a signal or data set in terms of orthogonal basis functions which are determined from the data. The term is also interchangeable with the geographically weighted Principal components analysis in geophysics. The i th basis function is chosen to be orthogonal to the basis functions from the first through i − 1, and to minimize the residual variance. That is, the basis functions are chosen to be different from each other, and to account for as much variance as possible. The method of EOF analysis is similar in spirit to harmonic analysis, but harmonic analysis typically uses predetermined orthogonal functions, for example, sine and cosine functions at fixed frequencies. In some cases the two methods may yield essentially the same results. The basis functions are typically found by computing the eigenvectors of the covariance matrix of the data set. A more advanced technique is to form a kernel out of the data, using a fixed kernel. The basis functions from the eigenvectors of the kernel matrix are thus non-linear in the location of the data (see Mercer's theorem and the kernel trick for more information). See also Blind signal separation Multilinear PCA Multilinear subspace learning Nonlinear dimensionality reduction Orthogonal matrix Signal separation Singular spectrum analysis Transform coding Varimax rotation References and notes Further reading Bjornsson Halldor and Silvia A. Venegas "A manual for EOF and SVD analyses of climate data", McGill University, CCGCR Report No. 97-1, Montréal, Québec, 52pp., 1997. David B. Stephenson and Rasmus E. Benestad. "Environmental statistics for climate researchers". (See: "Empirical Orthogonal Function analysis") Christopher K. Wikle and Noel Cressie. "A dimension reduced approach to space-time Kalman filtering", Biometrika 86:815-829, 1999. Donald W. Denbo and John S. Allen. "Rotary Empirical Orthogonal Function Analysis of Currents near the Oregon Coast", "J. Phys. Oceanogr.", 14, 35-46, 1984. David M. Kaplan "Notes on EOF Analysis" Spatial analysis Statistical signal processing
https://en.wikipedia.org/wiki/Intervision
Intervision may refer to: the Intervision Network, the Eastern European equivalent of the Eurovision Network the Intervision Song Contest organised by the Intervision Network Intervision (album), a 1997 album by Jimi Tenor Intervision, a brief orchestral composition by Dmitri Shostakovich commissioned by the Intervision Network in 1971 for use in its broadcasts. See also Eurovision (disambiguation)
https://en.wikipedia.org/wiki/Message%20queue
In computer science, message queues and mailboxes are software-engineering components typically used for inter-process communication (IPC), or for inter-thread communication within the same process. They use a queue for messaging – the passing of control or of content. Group communication systems provide similar kinds of functionality. The message queue paradigm is a sibling of the publisher/subscriber pattern, and is typically one part of a larger message-oriented middleware system. Most messaging systems support both the publisher/subscriber and message queue models in their API, e.g. Java Message Service (JMS). Remit and ownership Message queues implement an asynchronous communication pattern between two or more processes/threads whereby the sending and receiving party do not need to interact with the message queue at the same time. Messages placed onto the queue are stored until the recipient retrieves them. Message queues have implicit or explicit limits on the size of data that may be transmitted in a single message and the number of messages that may remain outstanding on the queue. Remit Many implementations of message queues function internally within an operating system or within an application. Such queues exist for the purposes of that system only. Other implementations allow the passing of messages between different computer systems, potentially connecting multiple applications and multiple operating systems. These message queuing systems typically provide resilience functionality to ensure that messages do not get "lost" in the event of a system failure. Examples of commercial implementations of this kind of message queuing software (also known as message-oriented middleware) include IBM MQ (formerly MQ Series) and Oracle Advanced Queuing (AQ). There is a Java standard called Java Message Service, which has several proprietary and free software implementations. Real-time operating systems (RTOSes) such as VxWorks and QNX encourage the use of message queuing as the primary inter-process or inter-thread communication mechanism. This can result in integration between message passing and CPU scheduling. Early examples of commercial RTOSes that encouraged a message-queue basis to inter-thread communication also include VRTX and pSOS+, both of which date to the early 1980s. The Erlang programming language uses processes to provide concurrency; these processes communicate asynchronously using message queuing. Ownership The message queue software can be either proprietary, open source or a mix of both. It is then run either on premise in private servers or on external cloud servers (message queuing service). Proprietary options have the longest history, and include products from the inception of message queuing, such as IBM MQ, and those tied to specific operating systems, such as Microsoft Message Queuing (MSMQ). Cloud service providers also provide their proprietary solutions such as Amazon Simple Queue Service (SQS), Storm
https://en.wikipedia.org/wiki/Message-oriented%20middleware
Message-oriented middleware (MOM) is software or hardware infrastructure supporting sending and receiving messages between distributed systems. MOM allows application modules to be distributed over heterogeneous platforms and reduces the complexity of developing applications that span multiple operating systems and network protocols. The middleware creates a distributed communications layer that insulates the application developer from the details of the various operating systems and network interfaces. APIs that extend across diverse platforms and networks are typically provided by MOM. This middleware layer allows software components (applications, Enterprise JavaBeans, servlets, and other components) that have been developed independently and that run on different networked platforms to interact with one another. Applications distributed on different network nodes use the application interface to communicate. In addition, by providing an administrative interface, this new, virtual system of interconnected applications can be made fault tolerant and secure. MOM provides software elements that reside in all communicating components of a client/server architecture and typically support asynchronous calls between the client and server applications. MOM reduces the involvement of application developers with the complexity of the master-slave nature of the client/server mechanism. Middleware categories Remote procedure call or RPC-based middleware Object request broker or ORB-based middleware Message-oriented middleware or MOM-based middleware All these models make it possible for one software component to affect the behavior of another component over a network. They are different in that RPC- and ORB-based middleware create systems of tightly coupled components, whereas MOM-based systems allow for a loose coupling of components. In an RPC- or ORB-based system, when one procedure calls another, it must wait for the called procedure to return before it can do anything else. In these synchronous messaging models, the middleware functions partly as a super-linker, locating the called procedure on a network and using network services to pass function or method parameters to the procedure and then to return results. Advantages Central reasons for using a message-based communications protocol include its ability to store (buffer), route, or transform messages while conveying them from senders to receivers. Another advantage of messaging provider mediated messaging between clients is that by adding an administrative interface, you can monitor and tune performance. Client applications are thus effectively relieved of every problem except that of sending, receiving, and processing messages. It is up to the code that implements the MOM system and up to the administrator to resolve issues like interoperability, reliability, security, scalability, and performance. Asynchronicity Using a MOM system, a client makes an API call to send a message to a des
https://en.wikipedia.org/wiki/Tumbler%20%28Project%20Xanadu%29
In the design of the Xanadu computer system, a tumbler is an address of any range of content or link or a set of ranges or links. According to Gary Wolf in Wired, the idea of tumblers was that "the address would not only point the reader to the correct machine, it would also indicate the author of the document, the version of the document, the correct span of bytes, and the links associated with these bytes." Tumblers were created by Roger Gregory and Mark Miller. They were used in the Xanadu FEBE (Front End - Back End) protocol in a manner similar to the use of URIs between web browsers and servers. Concept and implementation The idea behind tumblers comes from transfinite numbers. A tumbler is a unique numerical address of any interesting artifact. It resembles an IP address, but is much larger and has much more detailed structure. The structure looks like this. 1. < node >.0. < user >.0. < document >.0. < element > The leading "1." is used in order to mark the start of a new address. The individual fields of the address are divided with ".0." so that they can be arbitrarily long. Each < element > has the format "n. n. ... . n", a hierarchy of subaddresses. The last element denotes the type of data the tumbler refers to, for example: Text/Bytes Links Bitmaps etc. The 9287th byte of this version of the document would be 1.2368.792.6.0.6974.383.1988.352.0.75.2.0.1.9287 and the 356th link would be 0.2.356 on the end instead. A tumbler can be issued only once and never changes. The type of structure can grow at will, so the address space is infinite. Nelson also introduces the concepts of direction and a "span", which is a part of a document that is semantically meaningful for the document. For example, one can speak of "2 chapters back" or "300 bytes forward". See also Purple Numbers, a proposal to address paragraphs in Web pages. XPointer Cross-reference Impure name References External links (Xanadu project wiki, restructured in August 2005) Identifiers Hypertext Ted Nelson
https://en.wikipedia.org/wiki/CKCO-DT
CKCO-DT (channel 13) is a television station in Kitchener, Ontario, Canada. Part of the CTV Television Network, it is owned and operated by network parent Bell Media alongside London-based CTV 2 station CFPL-DT, although the two stations maintain separate operations. CKCO-DT's studios are located at 864 King Street West in Kitchener (across from the Grand River Hospital and Ion rapid transit light rail station adjacent to the Waterloo border), and its transmitter is located at Baden Tower between Snyders Road East and Highway 7 in Baden, just west of the Kitchener city limits. History The station first signed on the air at 6 p.m. on March 1, 1954. Its signal transmitted from the Baden Tower (a transmitter on Baden Hill), near Baden, just west of Kitchener. The transmitter has become one of the most identifiable landmarks in the area. Originally, like all privately owned television stations in Canada from 1953 to 1959, CKCO was an affiliate of the CBC; it became an affiliate of CTV in 1963. The station increased its transmitter power in the early 1960s to reach London, from which Kitchener then received CBC affiliate programs on CFPL-TV. CKCO was originally owned by Central Ontario Television, a consortium that included the Famous Players theatre chain (now owned by Cineplex Entertainment) and businessman Carl Arthur Pollock, president of the family-owned television manufacturer Electrohome, although his broadcast holdings – which also included radio stations CFCA-FM and CKKW – were operated by a separate company. At one time, CKCO was owned by CAP Communications, whose name was taken from Pollock's initials; a corporate reorganization in 1970 placed the stations directly under the ownership of Electrohome, which also acquired control of CKCO when Canadian broadcasting laws required domestic ownership of stations, ending the involvement of American-owned Famous Players, which at the time was owned by Paramount Pictures' parent company Gulf + Western (the latter was acquired by the original Viacom). CKCO would become the first station in Canada to provide closed captioning for all of its local newscasts, in 1988. In the 1990s, Baton Broadcasting had owned competing local stations in southwestern Ontario (CFPL-TV in London, CHWI-TV in Windsor, CKNX-TV in Wingham). A deal between Electrohome and Baton in 1996 resulted in each company owning 50% of these stations, as well as CKCO-TV, among other Canadian stations. The following year, another deal gave Baton control over CKCO-TV, while CHUM Limited took control over the other southwestern Ontario stations (which presently operate as owned-and-operated stations of the CTV Two television system). CTVglobemedia reacquired CFPL, CHWI, and CKNX in 2007 as a result of a takeover of CHUM Limited. In 1998, Baton changed its name to CTV Inc. after becoming the sole owner of CTV, ending the decades of cooperative ownership of the network. In 2000, BCE purchased CTV Inc. and combined it with NetStar Communic
https://en.wikipedia.org/wiki/List%20of%20radio%20stations%20in%20North%20Carolina
The following is a list of FCC-licensed radio stations in the U.S. state of North Carolina, which can be sorted by their call signs, frequencies, cities of license, licensees, and programming formats. List of radio stations Defunct WBIG WCRY WDJD-LP WGIV WGSB WGTL WGTM (Spindale, North Carolina) WGTM (Wilson, North Carolina) WJBX WJOS WJPI WJSL-LP WLTT WMBL WOOW WPTP-LP WQNX WRDK WSHP-LP WSPF WTOW WTRQ WVBS WVOT WVSP WWIL WWNG See also North Carolina media List of newspapers in North Carolina List of television stations in North Carolina Media of cities in North Carolina: Asheville, Charlotte, Durham, Fayetteville, Greensboro, High Point, Raleigh, Wilmington, Winston-Salem References Bibliography External links (Directory ceased in 2017) North Carolina Association of Broadcasters Asheville Radio Museum (est. 2001) Carolinas Chapter of the Antique Wireless Association Images North Carolina
https://en.wikipedia.org/wiki/Universal%20Network%20Objects
Universal Network Objects (UNO) is the component model used in the OpenOffice.org and LibreOffice computer software application suites. It is interface-based and designed to offer interoperability between different programming languages, object models and machine architectures, on a single machine, within a LAN or over the Internet. Users can implement or access UNO components from any programming language for which a language binding exists. Complete UNO language bindings exist for C++ (compiler-dependent), Java, Object REXX, Python, and Tcl. Bindings allowing access, but not writing, to components exist for StarOffice Basic, OLE Automation and the .NET Common Language Infrastructure. In particular, this API is used by macros. Universal Network Objects operate within the UNO Runtime Environment (URE). The Apache OpenOffice version of UNO is released under the terms Apache License (Version 2) as free and open source software. UNO for function-calling Examples: an external program can export an ODT file as a PDF file, or import and convert a DOCX, calling LibreOffice by the UNO interface. Another external program can access a cell and formulas from LibreOffice Calc file. Application examples: Docvert, JODConverter, unoConv. UNO for Add-Ons Programmers can write and integrate their own UNO components to OpenOffice/LibreOffice. Those components can be added to the LibreOffice menus and toolbars; they are called "Add-Ons". The Add-Ons can extend the functionality of LibreOffice. The integration of new components is supported by some tools and services. The three main steps are as follows: Register the new components within LibreOffice. This can be accomplished using the tool unopkg. Integrate the new components as services. The ProtocolHandler and JobDispatch services assist you. Change the user interface (menus or toolbars). This can be done almost automatically by writing an XML text file that describes the changes. Application example: jOpenDocument. References External links Apache OpenOffice UNO Development Kit project page Overview and technical details Java overview-summary OpenOffice.org Software Development Kit ODF Toolkit: Transition Steps Developer's Guide LibreOffice see unoexe and unopkg Inside LibreOffice: Universal Network Objects Language bridges (native for Java and Python) UNO for Object REXX UNO for PHP (written in C++) UNO for FreePascal/Delphi maybe orphaned Object-oriented programming OpenOffice
https://en.wikipedia.org/wiki/Software%20agent
In computer science, a software agent is a computer program that acts for a user or another program in a relationship of agency. The term agent is derived from the Latin agere (to do): an agreement to act on one's behalf. Such "action on behalf of" implies the authority to decide which, if any, action is appropriate. Some agents are colloquially known as bots, from robot. They may be embodied, as when execution is paired with a robot body, or as software such as a chatbot executing on a computer, such as a mobile device, e.g. Siri. Software agents may be autonomous or work together with other agents or people. Software agents interacting with people (e.g. chatbots, human-robot interaction environments) may possess human-like qualities such as natural language understanding and speech, personality or embody humanoid form (see Asimo). Related and derived concepts include intelligent agents (in particular exhibiting some aspects of artificial intelligence, such as reasoning), autonomous agents (capable of modifying the methods of achieving their objectives), distributed agents (being executed on physically distinct computers), multi-agent systems (distributed agents that work together to achieve an objective that could not be accomplished by a single agent acting alone), and mobile agents (agents that can relocate their execution onto different processors). Concepts The basic attributes of an autonomous software agent are that agents: are not strictly invoked for a task, but activate themselves, may reside in wait status on a host, perceiving context, may get to run status on a host upon starting conditions, do not require interaction of user, may invoke other tasks including communication. The term "agent" describes a software abstraction, an idea, or a concept, similar to OOP terms such as methods, functions, and objects. The concept of an agent provides a convenient and powerful way to describe a complex software entity that is capable of acting with a certain degree of autonomy in order to accomplish tasks on behalf of its host. But unlike objects, which are defined in terms of methods and attributes, an agent is defined in terms of its behavior. Various authors have proposed different definitions of agents, these commonly include concepts such as persistence (code is not executed on demand but runs continuously and decides for itself when it should perform some activity) autonomy (agents have capabilities of task selection, prioritization, goal-directed behavior, decision-making without human intervention) social ability (agents are able to engage other components through some sort of communication and coordination, they may collaborate on a task) reactivity (agents perceive the context in which they operate and react to it appropriately). Distinguishing agents from programs All agents are programs, but not all programs are agents. Contrasting the term with related concepts may help clarify its meaning. Franklin & Graesser (1997)
https://en.wikipedia.org/wiki/Prairie%20Public%20Radio
Prairie Public is a network of ten North Dakota radio stations. It is a service of Prairie Public Broadcasting, in association with North Dakota State University in Fargo. Prairie Public maintains active studios in Grand Forks, Fargo, and Bismarck. It provides National Public Radio (NPR) news and programming, local and regional news, and two distinct music formats: the News and Classical network, and the adult album alternative formatted Roots, Rock, and Jazz network. Programming Prairie Public produces and broadcasts Main Street, a weekday interview show hosted by Ashley Thornberg and Alicia Hegland-Thorpe, Dakota Datebook, Into the Music with Mike Olson, Prebys on Classics, and Why?, hosted by UND philosophy professor Dr. Jack Weinstein. Prairie Public is also the distributor for The Thomas Jefferson Hour. Prairie Public offers news programming on weekday mornings and afternoons from its newsrooms in Bismarck and Fargo. It also airs news from NPR. Prairie Public is a member station of National Public Radio, airing programs such as All Things Considered, and also carries programming from Public Radio International (such as The World) and American Public Media, as well as from Public Radio Exchange (such as This American Life). Prairie Public's radio network offers two programming services. The primary News and Classical network originating from KCND in Bismarck is carried on most stations, and split into eastern and western schedules. The adult album alternative formatted Roots, Rock, and Jazz network originating from KFJM in Grand Forks has gradually expanded its programming to additional stations since its launch in 2002. KDSU in Fargo carries a combination of both networks, airing Roots, Rock and Jazz programming when the rest of the main network airs classical music. News and Classical network The primary network of Prairie Public airs classical music, news, talk, and weekend specialty shows, including jazz. Roots, Rock, and Jazz network KFJM originates Prairie Public's second music format, a mixture of adult album alternative, blues, folk, and jazz. The network is rebroadcast full-time on KPPR Williston and the HD-2 channel of Prairie Public's other full-power News and Classical stations. KDSU of Fargo broadcasts the network midday weekdays and overnights. Stations Prairie Public has 10 full power stations and 5 low-power translators broadcasting across North Dakota, northwest Minnesota, and eastern Montana. HD Radio Prairie Public's full power stations broadcast HD Radio signals, adding full-digital simulcasts of their analog channel, plus the Roots, Rock, and Jazz network on subchannel "HD-2" of the News and Classical stations. Cable systems Shaw Cable's Winnipeg system carried Prairie Public's News and Classical service at 107.9 FM (via KUND-FM), until Shaw discontinued FM distribution in 2012. Prairie Public's News and Classical network is carried on MTS Ultimate TV across Manitoba, on channel 733. History Prairie Public w
https://en.wikipedia.org/wiki/VAG
VAG or vag may refer to: Freiburger Verkehrs AG, the municipal transport company of the city of Freiburg, Germany IBM VisualAge Generator, a platform-independent programming code generator The Vancouver Art Gallery in Vancouver, British Columbia, Canada A slang term for vagina, pronounced "vadge" Vag, a dramatic and humorous skit, part of Tamasha musical theatre of Maharashtra Versova-Andher-Ghatkopar, a line on the Mumbai Metro in India Vereinigte Astronomische Gesellschaft, United Astronomical Society Volkswagen Aktiengesellschaft, the German name for the automaker Volkswagen Group VAG Rounded, a text font designed for Volkswagen Group in 1979 Våg, an old Scandinavian unit of mass Verkehrs-Aktiengesellschaft Nürnberg (VAG or VAGN), the municipal transport company of the city of Nuremberg, Germany See also Vaj (disambiguation)
https://en.wikipedia.org/wiki/Electronic%20document
An electronic document is any electronic media content (other than computer programs or system files) that is intended to be used in either an electronic form or as printed output. Originally, any computer data were considered as something internal — the final data output was always on paper. However, the development of computer networks has made it so that in most cases it is much more convenient to distribute electronic documents than printed ones. The improvements in electronic visual display technologies made it possible to view documents on a screen instead of printing them (thus saving paper and the space required to store the printed copies). However, using electronic documents for the final presentation instead of paper has created the problem of multiple incompatible file formats. Even plain text computer files are not free from this problem — e.g. under MS-DOS, most programs could not work correctly with UNIX-style text files (see newline), and for non-English speakers, the different code pages always have been a source of trouble. Even more problems are connected with complex file formats of various word processors, spreadsheets, and graphics software. To alleviate the problem, many software companies distribute free file viewers for their proprietary file formats (one example is Adobe's Acrobat Reader). The other solution is the development of standardized non-proprietary file formats (such as HTML and OpenDocument), and electronic documents for specialized uses have specialized formats – the specialized electronic articles in physics use TeX or PostScript. See also Digital era governance Digital library Digital media Ebook Electronic paper Electronic publishing Paperless office E-government External links What is a digital document Digital Imaging Frequent Questions From ETSI Word processors
https://en.wikipedia.org/wiki/Network%20congestion
Network congestion in data networking and queueing theory is the reduced quality of service that occurs when a network node or link is carrying more data than it can handle. Typical effects include queueing delay, packet loss or the blocking of new connections. A consequence of congestion is that an incremental increase in offered load leads either only to a small increase or even a decrease in network throughput. Network protocols that use aggressive retransmissions to compensate for packet loss due to congestion can increase congestion, even after the initial load has been reduced to a level that would not normally have induced network congestion. Such networks exhibit two stable states under the same level of load. The stable state with low throughput is known as congestive collapse. Networks use congestion control and congestion avoidance techniques to try to avoid collapse. These include: exponential backoff in protocols such as CSMA/CA in 802.11 and the similar CSMA/CD in the original Ethernet, window reduction in TCP, and fair queueing in devices such as routers and network switches. Other techniques that address congestion include priority schemes which transmit some packets with higher priority ahead of others and the explicit allocation of network resources to specific flows through the use of admission control. Network capacity Network resources are limited, including router processing time and link throughput. Resource contention may occur on networks in several common circumstances. A wireless LAN is easily filled by a single personal computer. Even on fast computer networks, the backbone can easily be congested by a few servers and client PCs. Denial-of-service attacks by botnets are capable of filling even the largest Internet backbone network links, generating large-scale network congestion. In telephone networks, a mass call event can overwhelm digital telephone circuits, in what can otherwise be defined as a denial-of-service attack. Congestive collapse Congestive collapse (or congestion collapse) is the condition in which congestion prevents or limits useful communication. Congestion collapse generally occurs at choke points in the network, where incoming traffic exceeds outgoing bandwidth. Connection points between a local area network and a wide area network are common choke points. When a network is in this condition, it settles into a stable state where traffic demand is high but little useful throughput is available, during which packet delay and loss occur and quality of service is extremely poor. Congestive collapse was identified as a possible problem by 1984. It was first observed on the early Internet in October 1986, when the NSFNET phase-I backbone dropped three orders of magnitude from its capacity of 32 kbit/s to 40 bit/s, which continued until end nodes started implementing Van Jacobson and Sally Floyd's congestion control between 1987 and 1988. When more packets were sent than could be handled by intermedia
https://en.wikipedia.org/wiki/The%207th%20Guest
The 7th Guest is an interactive movie puzzle adventure game, produced by Trilobyte and originally released by Virgin Interactive Entertainment in April 1993. It is one of the first computer video games to be released only on CD-ROM. The 7th Guest is a horror story told from the unfolding perspective of the player, as an amnesiac. The game received press attention for making live action video clips a core part of its gameplay, for its then-unprecedented amount of pre-rendered 3D graphics, and for its adult content. The game was very successful, with over two million copies sold. It, alongside Myst, is widely regarded as a killer app that accelerated the sales of CD-ROM drives. The 7th Guest has subsequently been re-released on Apple's app store for various systems such as the Mac. Bill Gates called The 7th Guest "the new standard in interactive entertainment". The game has been ported in various formats to different systems, with Trilobyte mentioning the potential for a third entry in the series. Gameplay The game is played by wandering through a mansion, solving logic puzzles and watching videos that further the story. The main antagonist, Henry Stauf, is an ever-present menace, taunting the player with clues, mocking the player as they fail his puzzles ("We'll all be dead by the time you solve this!"), and expressing displeasure when the player succeeds ("Don't think you'll be so lucky next time!"). A plot of manipulation and sin is gradually played out, in flashback, by actors through film clips as the player progresses between rooms by solving twenty-one puzzles of shifting nature and increasing difficulty. The first puzzles most players encounter are either one where players must select the right interconnected letters inside the lens of a telescope to form a coherent sentence; or a relatively simple cake puzzle, where the player has to divide the cake evenly into six pieces, each containing the same number of decorations. Other puzzles include mazes, chess problems, logical deductions, Simon-style pattern-matching, word manipulations, and an difficult game of Ataxx, similar to Reversi. For players who need help or cannot solve a particular puzzle, there is a hint book in the library of the house. The first two times the book is consulted about a puzzle, the book gives clues about how to solve the puzzle; for the third time, the book completes the puzzle for the player so that the player can proceed. After each puzzle, the player is shown a video clip of part of the plot, if the hint book was consulted three times, the player does not get to view the clip. The hint book can be used for all but the final puzzle. The 7th Guest was one of the first games for the PC platform to be available only on CD-ROM, since it was too large to be distributed on floppy disks. Computer Gaming World reported with amazement in 1993, "not only does Guest consume an entire CD-ROM ... it actually requires TWO". Removing some of the large movies and videos was n
https://en.wikipedia.org/wiki/Educational%20research
Educational research refers to the systematic collection and analysis of data related to the field of education. Research may involve a variety of methods and various aspects of education including student learning, interaction, teaching methods, teacher training, and classroom dynamics. Educational researchers generally agree that research should be rigorous and systematic. However, there is less agreement about specific standards, criteria and research procedures. As a result, the value and quality of educational research has been questioned. Educational researchers may draw upon a variety of disciplines including psychology, economics, sociology, anthropology, and philosophy. Methods may be drawn from a range of disciplines. Conclusions drawn from an individual research study may be limited by the characteristics of the participants who were studied and the conditions under which the study was conducted. General characteristics Gary Anderson outlined ten aspects of educational research: Attempt to discover cause and effect. Research involves gathering new data from primary or first-hand sources or using existing data for a new purpose. Research is based upon observable experience or empirical evidence. Research demands accurate observation and description. Research generally employs carefully designed procedures and rigorous analysis. Research emphasizes the development of generalizations, principles or theories that will help in understanding, prediction and/or control. Research requires expertise—familiarity with the field; competence in methodology; technical skill in collecting and analyzing the data. Research attempts to find an objective, unbiased solution to the problem and takes great pains to validate the procedures employed. Research is a deliberate and unhurried activity which is directional but often refines the problem or ques Approaches There are different approaches to educational research. One is a basic approach, also referred to as an academic research approach. Another approach is applied research or a contract research approach. These approaches have different purposes which influence the nature of the respective research. Basic approach Basic, or academic research focuses on the search for truth or the development of educational theory. Researchers with this background "design studies that can test, refine, modify, or develop theories". Generally, these researchers are affiliated with an academic institution and are performing this research as part of their graduate or doctoral work. Applied approach The pursuit of information that can be directly applied to practice is aptly known as applied or contractual research. Researchers in this field are trying to find solutions to existing educational problems. The approach is much more utilitarian and pragmatic as it strives to find information that will directly influence practice. Applied researchers are often commissioned by a sponsor and are responsible fo
https://en.wikipedia.org/wiki/J2STask
In computers, J2STask is a software product that was developed by J2S to streamline the workflow of people who process digital raster graphics (or Encapsulated PostScript) files. It is designed to automatically complete repetitive operations on large numbers of image files stored on a local hard disk, or on an ftp server (J2S, n.d.). It requires Mac OS X v10.2 or newer, and costs 1,200 EUR (MacMinute, 2004). References MacMinute. (2004). J2STask 1.0 image workflow manager released. Retrieved January 14, 2004. J2S. (n.d.). J2STask: task manager for image worklow. Retrieved January 14, 2004. Graphics software
https://en.wikipedia.org/wiki/Dot%20pitch
Dot pitch (sometimes called line pitch, stripe pitch, or phosphor pitch) is a specification for a computer display, computer printer, image scanner, or other pixel-based devices that describe the distance, for example, between dots (sub-pixels) on a display screen. In the case of an RGB color display, the derived unit of pixel pitch is a measure of the size of a triad plus the distance between triads. Dot pitch may be measured in linear units (with smaller numbers meaning higher resolution), usually millimeters (mm), or as a rate, for example, dots per inch (with a larger number meaning higher resolution). Closer spacing produces a sharper image (as there are more dots in a given area). However, other factors may affect image quality, including: Undocumented or inadequately documented measurement method, complicated by ignorance of the existence of different methods Confusion of pixels and subpixels Element spacing varying across screen area (e.g., widening in corners compared to center) Differing pixel geometries Differing image and pixel aspect ratios Miscellanea such as Kell factor or interlaced video The exact difference between horizontal and diagonal dot pitch varies with the design of the monitor (see pixel geometry and widescreen), but a typical entry-level 0.28 mm (diagonal) monitor has a horizontal pitch of 0.24 or 0.25 mm, and a good quality 0.26 mm (diagonal) unit has a horizontal pitch of 0.22 mm. The above dot pitch measurement does not apply to aperture grille displays. Such monitors use continuous vertical phosphor bands on the screen, so the vertical distance between scan lines is limited only by the video input signal's vertical resolution and the thickness of the electron beam, so there is no vertical 'dot pitch' on such devices. Aperture grille only has horizontal 'dot pitch', or otherwise known as 'stripe pitch'. Common dot pitch sizes References External links PPI calculator – Shows dot pitch Pixels Per Inch PPI Calculator – Determines the number of pixels per inch of your display Megapixel Calculator – Identifies aspect ratio and displays photo and video storage requirements for different formats at a given megapixel number Length Television technology
https://en.wikipedia.org/wiki/CAIR
CAIRz or Cair may refer to: Acronyms Carboxyaminoimidazole ribotide, a biochemical intermediate nucleotide Centre for Artificial Intelligence and Robotics, Indian national defence laboratory Clean Air Interstate Rule, US environmental regulation Council on American–Islamic Relations, American Muslim advocacy organization Geography Cair, an alternate spelling of caer, a Welsh placename element referring to strongholds Čair Municipality, municipality of Skopje, Republic of Macedonia Fictional Cair Andros, a fictional island in Tolkien's fiction Cair Paravel, a castle from C. S. Lewis' The Chronicles of Narnia
https://en.wikipedia.org/wiki/Alexandre%20Julliard
Alexandre Julliard (born 1970) is a computer programmer who is best known as the project leader for Wine, a compatibility layer to run Microsoft Windows programs on Unix-like operating systems. Julliard studied computer science at the Swiss Federal Institute of Technology in Lausanne. He spent most of the 1990s working on embedded systems. He now works full-time on Wine for CodeWeavers. Julliard enjoys astronomy and lives in Lausanne, Switzerland. References External links OSNews interview, October 2001 - Interview with WINE's Alexandre Julliard - posted by Eugenia Loli-Queru on Mon 29th Oct 2001 17:25 UTC Wine HQ - WWN Issue #348:Wine Newsletter interview regarding Wine 1.0 release (6/18/2008) Wine HQ - WWN Issue #336:Wine Newsletter Interview with Mr. Alexandre Julliard (12/24/2007) YouTube video of: WINE Conference 2007 Keynote speech by Alexandre Julliard(October 7, 2007) 1970 births Living people Swiss computer programmers Date of birth missing (living people) Place of birth missing (living people) Chief technology officers Free software programmers Linux people
https://en.wikipedia.org/wiki/CrossOver%20%28software%29
CrossOver is a Microsoft Windows compatibility layer available for Linux, macOS, and ChromeOS. This compatibility layer enables many Windows-based applications to run on Linux operating systems, macOS, or ChromeOS. CrossOver is developed by CodeWeavers and based on Wine, an open-source Windows compatibility layer. CodeWeavers modifies the Wine source code, applies compatibility patches, adds configuration tools that are more user-friendly, automated installation scripts, and provides technical support. All changes made to the Wine source code are covered by the LGPL and publicly available. CodeWeavers maintains an online database listing how well various Windows applications perform under CrossOver. Versions CrossOver Linux CrossOver Linux is the original version of CrossOver. It aims to properly integrate with the GNOME and KDE desktop environments so that Windows applications will run seamlessly on Linux distributions. Before version 6, it was called CrossOver Mac Office. CrossOver Linux was originally offered in Standard and Professional editions. CrossOver Linux Standard was designed for a single user account on a machine. CrossOver Linux Professional provided enhanced deployment and management features for corporate users and multiple user accounts per machine. With the release of CrossOver Linux 11 in 2012, these different editions merged into a single CrossOver Linux product. CrossOver Mac In 2005 Apple announced a transition from PowerPC to Intel processors in their computers, which allowed CodeWeavers to develop a Mac OS X version of CrossOver Office called 'CrossOver Mac' CrossOver Mac was released on January 10, 2007. With the release of CrossOver Mac 7 on June 17, 2008, CrossOver Mac was divided into Standard and Pro editions like CrossOver Linux. The Standard version included six months of support and upgrades, while the Pro version included one year of support and upgrades, along with a free copy of CrossOver Games. With the release of CrossOver Mac 11 in 2012 these different editions were all merged into a single CrossOver Mac product. In 2019, macOS Catalina went 64-bit only and eliminated support for 32-bit programs and libraries. In December 2019 Codeweavers released CrossOver 19, providing support for 32 bit Windows applications on an operating system with no 32 bit libraries solving this problem. The technique, known as "wine32on64", requires using modified LLVM to build additional thunk code that allows running 32-bit programs in a 64-bit wine. In early June 2023, CodeWeavers announced early stages of DirectX 12 support on macOS would be available in CrossOver 23. At WWDC 2023, Apple announced the Game Porting Toolkit based on CrossOver to bring Windows games to macOS. Apple did not collaborate with CodeWeavers on this toolkit. In September 2023, CodeWeavers released version 23.5 of Crossover which supports D3DMetal from the Game Porting Toolkit as well as the GStreamer media framework. Component's versions details A
https://en.wikipedia.org/wiki/Meeting%20People%20Is%20Easy
Meeting People Is Easy is a 1998 British documentary film by Grant Gee that follows the English rock band Radiohead on the world tour for their 1997 album OK Computer. It received positive reviews and was nominated for a Grammy Award for Best Music Film at the 42nd Annual Grammy Awards in 2000. It sold more than half a million copies on VHS and DVD. Summary Meeting People Is Easy documents the promotion and tour for Radiohead's third album, OK Computer, which began on 22 May 1997 in Barcelona, Spain. The film comprises footage of Radiohead working on music, filming music videos and promotional material, giving interviews and performing. It includes footage of the filming of the "No Surprises" music video and the failed studio session for the song "Man of War", and a performance of "Karma Police" on the Late Show with David Letterman. The documentary captures the band members' stress during the tour, which the bassist, Colin Greenwood, later said was the lowest point of Radiohead's career. The journalist Alex Ross described the film as "a kind of counterstrike against the music press, recording scores of pointless interviews with dead-tired members of the band". Production According to the director, Grant Gee, Radiohead sat in hotel suites for days giving interviews. To record each interview, Gee "[ran] around, leaving a microphone in one room, going and filming something in another". He placed surveillance cameras in the band's dressing room, which Gee said foreshadowed the rise of reality television such as Big Brother: "We were doing it in a slightly more arty way, but it's the same ... Radiohead Big Brother is what I think of that film in a way." Radiohead's co-manager, Chris Hufford, said the film was "psychologically honest" and that he found it depressing to watch: "Seeing that going on where there should have been pride and joy. I knew they were readdressing how they looked to themselves, each other and the outside world." The drummer, Philip Selway, said the film was the result of how Grant perceived the period, and that other times on the tour were "much lighter". In 2020, the Radiohead singer, Thom Yorke, wrote: "I've never really watched this since it was completed. I couldn't because it would send me back down a mental hole that would take me days to recover from. But now skimming through it looks kind of funny, sad and alarming at the same time. I still recognise us all. But would have had some strong words for myself at this point." Release Meeting People Is Easy was released in the UK on VHS on 30 November 1998, and on DVD on 12 June 2000. It was released in both formats on 18 May 1999 in the United States. It was the first DVD released by Radiohead's record label, EMI. Several television channels broadcast Meeting People Is Easy after its release. In the UK, Channel 4 broadcast the film on 6 May 1999. In the US, MTV broadcast a premiere of the film on 16 May 1999, and the Sundance Channel broadcast the documentary nine time
https://en.wikipedia.org/wiki/Paranoid%20Android
"Paranoid Android" is a song by English alternative rock band Radiohead, released as the lead single from their third studio album, OK Computer (1997), on 26 May 1997. The lyrics were written by singer Thom Yorke following an unpleasant experience in a Los Angeles bar. The song is over six minutes long and contains four sections. The name is taken from Marvin the Paranoid Android from the science fiction series The Hitchhiker's Guide to the Galaxy. "Paranoid Android" charted at number three on the UK Singles Chart, Radiohead's highest-charting position in the UK to date. It received acclaim, with critics comparing it to the songs "Happiness Is a Warm Gun" by the Beatles and "Bohemian Rhapsody" by Queen. It has appeared regularly on lists of the best songs of all time, including NMEs and Rolling Stones respective 500 Greatest Songs of All Time lists. Its animated music video, directed by Magnus Carlsson, was placed on heavy rotation on MTV, although the network censored portions containing nudity in the US. At the 1998 Brit Awards, the song was nominated for Best British Single. The track has been covered by artists in a variety of genres. It was included in the 2008 Radiohead: The Best Of. Writing and recording As with many other OK Computer tracks, "Paranoid Android" was recorded in St Catherine's Court, a 15th-century mansion near the village of St Catherine, near Bath, Somerset. It was produced by Nigel Godrich. Inspired by the through-composed structure of the Beatles' 1968 song "Happiness Is a Warm Gun", Radiohead fused parts from three different songs. Other inspirations included Queen's "Bohemian Rhapsody" and the work of the Pixies. The first version was over 14 minutes long and included a long Hammond organ outro performed by Jonny Greenwood. The guitarist Ed O'Brien said: "We'd be pissing ourselves while we played. We'd bring out the glockenspiel and it would be really, really funny." The singer, Thom Yorke, sarcastically referred to this version as "a Pink Floyd cover". Greenwood said later that the organ solo was "hard to listen to without clutching the sofa for support". Godrich said: "Nothing really happened with the outro. It just spun and spun and it got very Deep Purple and went off." An early extended version was included on the 2019 compilation MiniDiscs [Hacked]. Influenced by the editing of the Beatles' Magical Mystery Tour, Radiohead shortened the song to six and a half minutes, with the organ solo replaced with a shorter guitar outro. The bassist, Colin Greenwood, said the band "felt like irresponsible schoolboys ... Nobody does a six-and-a-half-minute song with all these changes. It's ridiculous." For the ending, Yorke recorded himself shouting gibberish into a Dictaphone. Godrich edited the parts together with tape. He said: "It’s a very hard thing to explain, but it’s all on 24-track and it runs through ... I was very pleased with myself. I sort of stood there and said, 'You guys have no idea what I’ve just done.'
https://en.wikipedia.org/wiki/Touch%20%28disambiguation%29
Touch is one of the sensations processed by the somatosensory system. Touch may also refer to: Places Touch (river), in France Touch House, a mansion near Stirling, Scotland Computing and technology touch (command), a computer program HTC Touch, a touchscreen phone iPod Touch, a portable media player, PDA, and Wi-Fi platform Multi-touch, capability of a flat screen to detect touch gestures Ubuntu Touch, an interface Arts, entertainment, and media Comics and manga Touch (manga), a 1985 manga and anime series by Mitsuru Adachi Touch, a 6-issue comic book series published under the DC Focus imprint Film Touch (1997 film), by Paul Schrader Touch (2005 film), by director Isshin Inudō Touch (2022 film), by Justin Burquist The Touch (1971 film), by Ingmar Bergman The Touch (2002 film), by Peter Pau Television Touch (American TV series), a 2012–2013 American thriller television series Touch (South Korean TV series), a 2020 South Korean television series Literature Touch, a 1987 book by Elmore Leonard Touch, a 2014 book by Natalia Jaster Touch, a 2022 book by Olaf Olafsson Music Groups Touch (1960s band), an American rock band Touch (1980s band), a 1980s American rock band Touch (girl group), the original name of Spice Girls Touch (South Korean group), Korean boy band Albums Touch (Amerie album), or the title song Touch, by Brian Howe Touch (Con Funk Shun album), 1980 Touch (Delirious? album), or the title song Touch (Eurythmics album), 1983 Touch (July Talk album), 2016 Touch (Laura Branigan album), or the title song Touch (NEWS album), 2005 Touch (Noiseworks album), or the title song Touch (Sarah McLachlan album), or the title song Touch (The Supremes album), or the title song Touch, by Touch (1960s band) Touch (EP), or the title song, by Miss A Touch, by Dave Grohl, a soundtrack album from Paul Schrader's film Touch Songs "Touch" (Amerie song), 2005 "Touch", a song by Cigarettes After Sex from the 2019 album Cry "Touch", a song by Daft Punk from the 2013 album Random Access Memories "Touch: (Earth, Wind & Fire song), 1984 "Touch" (Little Mix song), 2016 "Touch" (Natasha Bedingfield song), 2010 "Touch" (Noiseworks song), 1988 "Touch" (NCT 127 song), 2018 "Touch" (Omarion song), 2004 "Touch", 1966 song by the band The Outsiders "Touch" (Pia Mia song), 2015 "Touch" (Shift K3Y song), 2014 "Touch" (Sori song), 2018 "Touch" (The Supremes song), 1971 "Touch" (Tea Party song), 2000 "Touch / Yume no Tsuzuki", 2005 song by Younha Other arts, entertainment, and media Touch (ballet), a ballet by David Parsons Touch FM Touch Music, an audio-visual publishing company based in the UK Touch! Generations, a video-game brand Sports Touch football (disambiguation) Touch (rugby), an area of a rugby field Touch (sport), a sport derived from rugby football Other uses Consoling touch, a social behaviour Touch, a characteristic of a tangent, in geometry Touch, a clothing line by Alyssa Milano Touch typ
https://en.wikipedia.org/wiki/Touch%20%28command%29
In computing, touch is a command used to update the access date and/or modification date of a computer file or directory. It is included in Unix and Unix-like operating systems, TSC's FLEX, Digital Research/Novell DR DOS, the AROS shell, the Microware OS-9 shell, and ReactOS. The command is also available for FreeDOS and Microsoft Windows. Overview In its default usage, it is the equivalent of creating or opening a file and saving it without any change to the file contents. touch avoids opening, saving, and closing the file. Instead it simply updates the dates associated with the file or directory. An updated access or modification date can be important for a variety of other programs such as backup utilities or the make command-line interface programming utility. Typically these types of programs are only concerned with files which have been created or modified after the program was last run. The touch command can also be useful for quickly creating files for programs or scripts that require a file with a specific name to exist for successful operation of the program, but do not require the file to have any specific content. The Single Unix Specification (SUS) specifies that touch should change the access times, modification times, or both, for a file. The file is identified by a pathname supplied as a single argument. It also specifies that if the file identified does not exist, the file is created and the access and modification times are set as specified. If no new timestamps are specified, touch uses the current time. History A touch utility first appeared in Version 7 AT&T UNIX. Today, the command is available for a number of different operating systems, including many Unix and Unix-like systems, DOS, Microsoft Windows and the classic Mac OS. The version of touch bundled in GNU coreutils was written by Paul Rubin, Arnold Robbins, Jim Kingdon, David MacKenzie, and Randy Smith. The command is available as a separate package for Microsoft Windows as part of the UnxUtils collection of native Win32 ports of common GNU Unix-like utilities. The FreeDOS version was developed by Kris Heidenstrom and is licensed under the GPL. DR DOS 6.0 and KolibriOS include an implementation of the command. The command has also been ported to the IBM i operating system. See also System time List of Unix commands References Further reading External links examples showing how to use touch Standard Unix programs Unix SUS2008 utilities Plan 9 commands Inferno (operating system) commands ReactOS commands IBM i Qshell commands
https://en.wikipedia.org/wiki/Bloomberg%20L.P.
Bloomberg L.P. is a privately held financial, software, data, and media company headquartered in Midtown Manhattan, New York City. It was co-founded by Michael Bloomberg in 1981, with Thomas Secunda, Duncan MacMillan, Charles Zegar, and a 12% ownership investment by Bank of America through their brokerage subsidiary Merrill Lynch. Bloomberg L.P. provides financial software tools and enterprise applications such as analytics and equity trading platform, data services, and news to financial companies and organizations through the Bloomberg Terminal (via its Bloomberg Professional Service), its core revenue-generating product. Bloomberg L.P. also includes a news agency (Bloomberg News), a global television network (Bloomberg Television), websites, radio stations (Bloomberg Radio), subscription-only newsletters, and two magazines: Bloomberg Businessweek and Bloomberg Markets. The company has 176 locations and nearly 20,000 employees. History In 1981, Salomon Brothers was acquired, and Michael Bloomberg, a general partner, was given a $10million partnership settlement. Bloomberg, having designed in-house computerized financial systems for Salomon, used his $10million partnership buyout to start Innovative Market Systems (IMS). Bloomberg developed and built his own computerized system to provide real-time market data, financial calculations and other financial analytics to Wall Street firms. The Market Master terminal, later called the Bloomberg Terminal, was released to market in December 1982. Merrill Lynch became the first customer, purchasing 20 terminals and a 30% equity stake in the company for $30million in exchange for a five-year restriction on marketing the terminal to Merrill Lynch's competitors. Merrill Lynch released IMS from this restriction in 1984. In 1986, the company renamed itself Bloomberg L.P. (limited partnership). Bloomberg launched Bloomberg Business News, later Bloomberg News, in 1990, with Matthew Winkler as editor-in-chief. Bloomberg.com was first established on September 29, 1993, as a financial portal with information on markets, currency conversion, news and events, and Bloomberg Terminal subscriptions. In late 1996, Bloomberg bought back one-third of Merrill Lynch's 30 percent stake in the company for $200million, valuing the company at $2billion. In 2008, facing losses during the financial crisis, Merrill Lynch agreed to sell its remaining 20 percent stake in the company back to Bloomberg Inc., majority-owned by Michael Bloomberg, for a reported $4.43billion, valuing Bloomberg L.P. at approximately $22.5billion. Bloomberg L.P. has remained a private company since its founding; the majority of which is owned by billionaire Michael Bloomberg. To run for the position of Mayor of New York against Democrat Mark Green in 2001, Bloomberg gave up his position of CEO and appointed Lex Fenwick as CEO in his stead. in 2012, Peter Grauer became the chairman of the company, a role he still holds. In 2008, Fenwick became the CE
https://en.wikipedia.org/wiki/Object%20%28IBM%20i%29
On many computing platforms everything is a file, but in contrast in IBM i everything is an object. Overview IBM i objects share similarities with objects in object-oriented programming, but there are differences as well. There are similarities in that when storage is allocated for something, that something is of a specific type, and only a specific set of programs are allowed to act upon that object. There are differences in that IBM i objects cannot be inherited, and the set of object types is fixed, and only IBM has the ability to create new ones. The number of object types is huge and a small subset of them are available to users. The human readable form of the object type is always a three to six character mnemonic preceded by an asterisk. What follows is a short list of the more commonly used objects and their mnemonics: *LIB: Library (where everything below, except directories and stream files, is stored; libraries cannot exist within other libraries). *PGM: Program (for compiled languages: CL, RPG-IV, C, C++, COBOL, etc. and there are no interface restrictions between the languages). *MODULE: Module (linkable into a program from a compiled language above and here too there are no restrictions on linkability between languages). *SRVPGM: Service program (dynamic set of one or more modules, akin to a DLL file in Microsoft Windows). *BNDDIR: Binding directory (holds a list of modules and service programs and is used when creating programs). *CMD: Command (an object used for calling programs that allows users to prompt for their parameters; can be created with the Command Definition language). See Control Language for more information. *MENU: Menu (accessed with the GO command). *FILE: File (IBM i files can be used for data, input/output devices, and source code, depending on sub type). *DTAARA: Data area (small bits of storage used to store tiny items of data for fast access). *DIR: Directory (part of the Integrated File System that is equivalent to Unix and Microsoft Windows hierarchical file systems). *STMF: Stream file (traditional file that would be familiar to most Unix and Microsoft Windows users and only stored in directories). *JRN & *JRNRCV: Journal and journal receiver (used to journal changes to files, data areas, and stream files). *USRPRF: User profile (allows users to sign-on to the system). *JOBD: Job description (used when submitting/starting jobs). *SBSD: Subsystem description (used when starting subsystems; this is the place where user jobs run). *JOBQ: Job queue (used to queue up batch jobs to run in a subsystem). *LIND: Line description (communications line: Ethernet, token ring, etc.). *CTLD: Controller description (communications controller for lines, workstations, etc.). *DEVD: Device description (communications device for lines, workstations, printers tape drives, etc.) *DTAQ: Data queue (used to queue up data entries for fast retrieval by other jobs). *MSGQ: Message queue (used to send message
https://en.wikipedia.org/wiki/Lou%20Montulli
Louis J. Montulli II (best known as Lou Montulli) is a computer programmer who is well known for his work in producing web browsers. In 1991 and 1992, he co-authored a text web browser called Lynx, with Michael Grobe and Charles Rezac, while he was at the University of Kansas. This web browser was one of the first available and is still in use today. Career In 1994, he became a founding engineer of Netscape Communications and programmed the networking code for the first versions of the Netscape web browser. He was also responsible for several browser innovations, such as HTTP cookies, the blink element, server push and client pull, HTTP proxying, and encouraging the implementation of animated GIFs into the browser. While at Netscape, he also was a founding member of the HTML working group at the W3C and was a contributing author of the HTML 3.2 specification. He is one of only six inductees in the World Wide Web Hall of Fame announced at the First International Conference on the World-Wide Web in 1994. In 1998, he became a founding engineer of Epinions which is now a Shopping.com company. In 2002, he was named to the MIT Technology Review TR100 as one of the top 100 innovators in the world under the age of 35. In 2004, he became co-founder and CEO of Memory Matrix, which was acquired by Shutterfly Inc. in May 2005. Montulli served as Vice President of Client Engineering at Shutterfly through the summer of 2007. In 2008, he became co-founder of Zetta.net, a cloud storage company. In 2015, he joined JetInsight as co-founder and CTO. In 2022, he was recognized among Hidden Heroes for his significant technology contribution, including the HTTP cookie and Lynx. Ongoing projects While working on the Netscape browser, Montulli built the Fishcam, one of the earliest live image websites (ie. live as in broadcasting), famously built into early versions of the Netscape browser as the Fishcam Easter egg. The company Netscape hosted this fishcam until long after they were no longer Netscape. After a short hiatus, in 2009 it found a new host; it is still one of the longest (nearly) continuously running live websites. See also List of programmers References External links Where cookies come from - DominoPower Magazine. HNSource History Montulli's first conversion to HTML page for WWW-VL Living people Year of birth missing (living people) Place of birth missing (living people) Netscape people University of Kansas alumni Computer programmers
https://en.wikipedia.org/wiki/AEI
AEI may refer to: Adelaide Educational Institution of South Australia Aei Latin-script trigraph AEI Music Network Inc. (Audio Environments Incorporated), which created the "Foreground Music" industry in 1971 Albert Einstein Institute, the Max Planck Institute for Gravitational Physics, Germany Albert Einstein Institution, an organisation involved in non-violent methods of political resistance based in the US Alliance for European Integration, the ruling coalition in Moldova since the July 2009 election American Enterprise Institute, a conservative think tank Architectural Engineering Institute Archive of European Integration Associated Electrical Industries, large company in the UK (now defunct) Automatic Equipment Identification, as used by the railroad industry Average Earnings Index, British labour market measure Average Earnings Index (horse racing), American horse racing measure Atomic Energy Organization See also EAI (disambiguation)
https://en.wikipedia.org/wiki/ECos
The Embedded Configurable Operating System (eCos) is a free and open-source real-time operating system intended for embedded systems and applications which need only one process with multiple threads. It is designed to be customizable to precise application requirements of run-time performance and hardware needs. It is implemented in the programming languages C and C++ and has compatibility layers and application programming interfaces for Portable Operating System Interface (POSIX) and The Real-time Operating system Nucleus (TRON) variant µITRON. eCos is supported by popular SSL/TLS libraries such as wolfSSL, thus meeting all standards for embedded security. Design eCos was designed for devices with memory sizes in the range of a few tens or several hundred kilobytes, or for applications with real-time requirements. eCos runs on a wide variety of hardware platforms, including ARM, CalmRISC, FR-V, Hitachi H8, IA-32, Motorola 68000, Matsushita AM3x, MIPS, NEC V850, Nios II, PowerPC, SPARC, and SuperH. The eCos distribution includes RedBoot, an open source application that uses the eCos hardware abstraction layer to provide bootstrap firmware for embedded systems. History eCos was initially developed in 1997 by Cygnus Solutions which was later bought by Red Hat. In early 2002, Red Hat ceased development of eCos and laid off the staff of the project. Many of the laid-off staff continued to work on eCos and some formed their own companies providing services for the software. In January 2004, at the request of the eCos developers, Red Hat agreed to transfer the eCos copyrights to the Free Software Foundation in October 2005, a process finally completed in May 2008. Non-free versions The eCosPro real-time operating system is a commercial fork of eCos created by eCosCentric which incorporates proprietary software components. It is claimed as a "stable, fully tested, certified and supported version", with additional features that are not released as free software. On Pi Day 2017, eCosCentric announced they had ported eCosPro to all of the Raspberry Pi models, with demonstrations at the Embedded World trade fair in Nuremberg (Germany) and releases free for non-commercial uses to follow. See also Comparison of open-source operating systems References External links ARM operating systems Embedded operating systems Free software operating systems MIPS operating systems Real-time operating systems
https://en.wikipedia.org/wiki/Preboot%20Execution%20Environment
In computing, the Preboot eXecution Environment, PXE (most often pronounced as pixie, often called PXE Boot/pixie boot.) specification describes a standardized client–server environment that boots a software assembly, retrieved from a network, on PXE-enabled clients. On the client side it requires only a PXE-capable network interface controller (NIC), and uses a small set of industry-standard network protocols such as DHCP and TFTP. The concept behind the PXE originated in the early days of protocols like BOOTP/DHCP/TFTP, and it forms part of the Unified Extensible Firmware Interface (UEFI) standard. In modern data centers, PXE is the most frequent choice for operating system booting, installation and deployment. Overview Since the beginning of computer networks, there has been a persistent need for client systems which can boot appropriate software images, with appropriate configuration parameters, both retrieved at boot time from one or more network servers. This goal requires a client to use a set of pre-boot services, based on industry standard network protocols. Additionally, the Network Bootstrap Program (NBP) which is initially downloaded and run must be built using a client firmware layer (at the device to be bootstrapped via PXE) providing a hardware independent standardized way to interact with the surrounding network booting environment. In this case the availability and subjection to standards are a key factor required to guarantee the network boot process system interoperability. One of the first attempts in this regard was bootstrap loading using TFTP standard RFC 906, published in 1984, which established the 1981 published Trivial File Transfer Protocol (TFTP) standard RFC 783 to be used as the standard file transfer protocol for bootstrap loading. It was followed shortly after by the Bootstrap Protocol standard RFC 951 (BOOTP), published in 1985, which allowed a disk-less client machine to discover its own IP address, the address of a TFTP server, and the name of an NBP to be loaded into memory and executed. BOOTP implementation difficulties, among other reasons, eventually led to the development of the Dynamic Host Configuration Protocol standard RFC 2131 (DHCP) published in 1997. The pioneering TFTP/BOOTP/DHCP approach fell short, as at the time, it did not define the required standardized client side of the provisioning environment. The Preboot Execution Environment (PXE) was introduced as part of the Wired for Management framework by Intel and is described in the specification published by Intel and SystemSoft. PXE version 2.0 was released in December 1998, and the update 2.1 was made public in September 1999. The PXE environment makes use of several standard client‑server protocols including DHCP and TFTP (now defined by the 1992 published RFC 1350). Within the PXE schema the client side of the provisioning equation is an integral part of the PXE standard and it is implemented either as a Network Interface Card (NIC
https://en.wikipedia.org/wiki/Round-off%20error
In computing, a roundoff error, also called rounding error, is the difference between the result produced by a given algorithm using exact arithmetic and the result produced by the same algorithm using finite-precision, rounded arithmetic. Rounding errors are due to inexactness in the representation of real numbers and the arithmetic operations done with them. This is a form of quantization error. When using approximation equations or algorithms, especially when using finitely many digits to represent real numbers (which in theory have infinitely many digits), one of the goals of numerical analysis is to estimate computation errors. Computation errors, also called numerical errors, include both truncation errors and roundoff errors. When a sequence of calculations with an input involving any roundoff error are made, errors may accumulate, sometimes dominating the calculation. In ill-conditioned problems, significant error may accumulate. In short, there are two major facets of roundoff errors involved in numerical calculations: The ability of computers to represent both magnitude and precision of numbers is inherently limited. Certain numerical manipulations are highly sensitive to roundoff errors. This can result from both mathematical considerations as well as from the way in which computers perform arithmetic operations. Representation error The error introduced by attempting to represent a number using a finite string of digits is a form of roundoff error called representation error. Here are some examples of representation error in decimal representations: Increasing the number of digits allowed in a representation reduces the magnitude of possible roundoff errors, but any representation limited to finitely many digits will still cause some degree of roundoff error for uncountably many real numbers. Additional digits used for intermediary steps of a calculation are known as guard digits. Rounding multiple times can cause error to accumulate. For example, if 9.945309 is rounded to two decimal places (9.95), then rounded again to one decimal place (10.0), the total error is 0.054691. Rounding 9.945309 to one decimal place (9.9) in a single step introduces less error (0.045309). This can occur, for example, when software performs arithmetic in x86 80-bit floating-point and then rounds the result to IEEE 754 binary64 floating-point. Floating-point number system Compared with the fixed-point number system, the floating-point number system is more efficient in representing real numbers so it is widely used in modern computers. While the real numbers are infinite and continuous, a floating-point number system is finite and discrete. Thus, representation error, which leads to roundoff error, occurs under the floating-point number system. Notation of floating-point number system A floating-point number system is characterized by integers: : base or radix : precision : exponent range, where is the lower bound and is the upper bound
https://en.wikipedia.org/wiki/CTV%20Northern%20Ontario
CTV Northern Ontario, formerly known as MCTV, is a system of four television stations in Northern Ontario, Canada, owned and operated by the CTV Television Network, a division of Bell Media. These stations are: CICI - Greater Sudbury (flagship station) CKNY - North Bay CHBX - Sault Ste. Marie CITO - Timmins Since 2005, all four stations refer to themselves on-air as simply CTV instead of their call letters; however, they remain legally licensed as separate stations, and continue to have common local programming. Station information and history is discussed on each station's own page. History Background Each of the four cities served by the CTV Northern Ontario system saw the launch of a locally owned television station in the 1950s: Sudbury's CKSO-TV was launched by the owners of the Sudbury Star in 1953, Sault Ste. Marie's CJIC-TV was launched by Hyland Broadcasting in 1955, North Bay's CKGN-TV was launched by Gerry Alger and Gerry Stanton in 1955, and Timmins's CFCL-TV was launched by J. Conrad Lavigne in 1956. All four stations were CBC Television affiliates at the time, as CTV did not exist until 1961. Each station continued to operate separately until 1970 when applications were filed with Canadian Radio-television and Telecommunications Commission to launch a second television station in Sudbury; the application process ultimately resulted in a major realignment in Sudbury, North Bay and Timmins. Cambrian Broadcasting, the Sudbury station's owners at this time, acquired the North Bay station and launched a repeater of CKSO in Timmins, serving as the new CTV affiliate in all three cities, while Lavigne launched new stations CKNC-TV in Sudbury and CHNB-TV in North Bay, and retained the CBC affiliation. Although Hyland Broadcasting was one of the original applicants for a new Sudbury station, CJIC remained unaffected by the final outcome at the time. Through the 1970s, however, the North Bay and Timmins markets proved too small to support competition between multiple stations; although the Sudbury stations were nominally profitable on their own, the losses in North Bay and Timmins left both companies nearly bankrupt by 1980. MCTV As a result of the stations' precarious financial situation, the CRTC permitted Northern Cable, the region's primary cable television provider, to purchase both companies. Northern Cable formed Mid-Canada Communications as a holding company for the six stations, operating them under a twinstick model. The CRTC explicitly stated that it intended this to be only a temporary arrangement, to end as soon as the CBC could afford to directly acquire MCTV's CBC affiliates. At this time, CKSO-TV adopted the new callsign CICI, and its repeater in Timmins became a new standalone station, CITO-TV. All six stations were referred to on air as Mid-Canada Television, or MCTV for short; the station pairs were distinguished from each other by use of their network affiliation (i.e., "MCTV-CTV" and "MCTV-CBC"). As well, MCTV
https://en.wikipedia.org/wiki/Edit%20decision%20list
An edit decision list or EDL is used in the post-production process of film editing and video editing. The list contains an ordered list of reel and timecode data representing where each video clip can be obtained in order to conform the final cut. EDLs are created by offline editing systems, or can be paper documents constructed by hand such as shot logging. These days, linear video editing systems have been superseded by non-linear editing (NLE) systems which can output EDLs electronically to allow autoconform on an online editing system – the recreation of an edited programme from the original sources (usually video tapes) and the editing decisions in the EDL. They are also often used in the digital video editing world, so rather than referring to reels they can refer to sequences of images stored on disk. Some formats, such as CMX3600, can represent simple editing decisions only. Final Cut Pro XML, the Advanced Authoring Format (AAF), and AviSynth scripts are relatively advanced file formats that can contain sophisticated EDLs. B-Roll Linear editing systems cannot dissolve between clips on the same video tape. Hence, one of these clips will need to be dubbed onto a new video tape. EDLs designate these occurrences by marking such dissolves' source reels as B-roll of "b-reels". For example, the EDL will change the 8th character of the reel name to the letter B. However, sometimes editors will (confusingly) use the letter B to designate time code breaks on a video tape. If there is broken time code on a video tape, there will be two (or more) instances of a particular time code on the video tape. When re-capturing, it can be ambiguous as to which timecode is the right one. The letter B may indicate that the right time code is from the second set of timecode on the video tape. Incompatibilities and potential problems EDL formats such as CMX, GVG, Sony, Final Cut Pro, and Avid are similar but can differ in small (but important) ways. Particular attention should be paid to reel naming convention. On the Avid, reel names can be up to 32 characters, but user should be aware that these EDLs don't adhere to online editing machine control specifications. These are used by systems that have modified the import/export code to handle file-based workflows as tape acquisition formats wane. On FCP, in CMX3600 format, only eight characters are allowed. Particular attention should be paid towards b-reels. If the EDL handles dissolves to the same reel, reel names should be limited to 7 characters since the 8th character may be replaced. EDLs can use either drop-frame (DF) or non drop-frame timecode (NDF), running at 24fps (non drop-frame only), 25fps (non drop-frame only), and 30fps (drop-frame and non drop-frame). Overall, EDLs are still commonly used as some systems do not support other more robust formats such as AAF and XML. Systems known to support EDL to some extent Almost any professional editing system and many others support some form of XM
https://en.wikipedia.org/wiki/SM-1420
The SM-1420 (CM-1420) is a 16 bit DEC PDP-11/45 minicomputer clone, and the successor to SM-4 in Soviet Bloc countries. Under the direction of Minpribor it was produced in the Soviet Union and Bulgaria from 1983 onwards, and is more than twice as fast as its predecessor. Its closest western counterpart is the DEC PDP-11/45, which means that the Soviet technology trailed 11 years behind compared to the Digital Equipment Corporation equivalent machine. The standard package includes 256 KiB MOS memory, two RK-06 disks, two TU-10 decks, CM-6315 barrel or DZM-180 dot-matrix printer from Mera Blonie (Poland), VT52 compatible or VTA-2000-15 (BTA 2000-15) VT100 compatible terminals from Mera Elzab. See also History of computing in the Soviet Union List of Soviet computer systems SM EVM References External links CIA reference aid on Soviet mainframes and minicomputers Minicomputers Soviet computer systems PDP-11
https://en.wikipedia.org/wiki/ENEA%20AB
Enea AB is a global information technology company with its headquarters in Kista, Sweden that provides real-time operating systems and consulting services. Enea, which is an abbreviation of Engmans Elektronik Aktiebolag, also produces the OSE operating system. History Enea was founded 1968 by Rune Engman as Engmans Elektronik AB. Their first product was an operating system for a defence computer used by the Swedish Air Force. During the 1970s the firm developed compiler technology for the Simula programming language. During the early days of the European Internet-like connections, Enea employee Björn Eriksen connected Sweden to EUnet using UUCP, and registered enea as the first Swedish domain in April 1983. The domain was later converted to the internet domain enea.se when the network was switched over to TCP and the Swedish top domain .se was created in 1986. Products OSE The ENEA OSE real-time operating system first released in 1985. The Enea multi core family of real-time operating systems was first released in 2009. The Enea Operating System Embedded (OSE) is a family of real-time, microkernel, embedded operating system created by Bengt Eliasson for ENEA AB, which at the time was collaborating with Ericsson to develop a multi-core system using Assembly, C, and C++. Enea OSE Multicore Edition is based on the same microkernel architecture. The kernel design that combines the advantages of both traditional asymmetric multiprocessing (AMP) and symmetric multiprocessing (SMP). Enea OSE Multicore Edition offers both AMP and SMP processing in a hybrid architecture. OSE supports many processors, mainly 32-bit. These include the ColdFire, ARM, PowerPC, and MIPS based system on a chip (SoC) devices. The Enea OSE family features three OSs: OSE (also named OSE Delta) for processors by ARM, PowerPC, and MIPS, OSEck for various DSP's, and OSE Epsilon for minimal devices, written in pure assembly (ARM, ColdFire, C166, M16C, 8051). OSE is a closed-source proprietarily licensed software released on 20 March 2018. OSE uses events (or signals) in the form of messages passed to and from processes in the system. Messages are stored in a queue attached to each process. A link handler mechanism allows signals to be passed between processes on separate machines, over a variety of transports. The OSE signalling mechanism formed the basis of an open-source inter-process kernel design project named LINX. Linux Enea Linux provides an open, cross-development tool chain and runtime environment based on the Yocto Project embedded Linux configuration system. Hypervisor Enea Hypervisor is also based on OSE microkernel technology and runs Enea OSE applications and takes as guests Linux Operating System and optionally semiconductor specific executive environments for bare-metal speed packet processing Optima Enea Optima development tool suite for developing, debugging, and profiling embedded systems software The Element The Element middleware software for high-
https://en.wikipedia.org/wiki/Mechanical%20calculator
A mechanical calculator, or calculating machine, is a mechanical device used to perform the basic operations of arithmetic automatically, or (historically) a simulation such as an analog computer or a slide rule. Most mechanical calculators were comparable in size to small desktop computers and have been rendered obsolete by the advent of the electronic calculator and the digital computer. Surviving notes from Wilhelm Schickard in 1623 reveal that he designed and had built the earliest of the modern attempts at mechanizing calculation. His machine was composed of two sets of technologies: first an abacus made of Napier's bones, to simplify multiplications and divisions first described six years earlier in 1617, and for the mechanical part, it had a dialed pedometer to perform additions and subtractions. A study of the surviving notes shows a machine that would have jammed after a few entries on the same dial, and that it could be damaged if a carry had to be propagated over a few digits (like adding 1 to 999). Schickard abandoned his project in 1624 and never mentioned it again until his death 11 years later in 1635. Two decades after Schickard's supposedly failed attempt, in 1642, Blaise Pascal decisively solved these particular problems with his invention of the mechanical calculator. Co-opted into his father's labour as tax collector in Rouen, Pascal designed the calculator to help in the large amount of tedious arithmetic required; it was called Pascal's Calculator or Pascaline. In 1672, Gottfried Leibniz started designing an entirely new machine called the Stepped Reckoner. It used a stepped drum, built by and named after him, the Leibniz wheel, was the first two-motion calculator, the first to use cursors (creating a memory of the first operand) and the first to have a movable carriage. Leibniz built two Stepped Reckoners, one in 1694 and one in 1706. The Leibniz wheel was used in many calculating machines for 200 years, and into the 1970s with the Curta hand calculator, until the advent of the electronic calculator in the mid-1970s. Leibniz was also the first to promote the idea of an Pinwheel calculator. Thomas' arithmometer, the first commercially successful machine, was manufactured two hundred years later in 1851; it was the first mechanical calculator strong enough and reliable enough to be used daily in an office environment. For forty years the arithmometer was the only type of mechanical calculator available for sale until the industrial production of the more successful Odhner Arithmometer in 1890. The comptometer, introduced in 1887, was the first machine to use a keyboard that consisted of columns of nine keys (from 1 to 9) for each digit. The Dalton adding machine, manufactured in 1902, was the first to have a 10 key keyboard. Electric motors were used on some mechanical calculators from 1901. In 1961, a comptometer type machine, the Anita Mk VII from Sumlock comptometer Ltd., became the first desktop mechanical calculato
https://en.wikipedia.org/wiki/Needham%E2%80%93Schroeder%20protocol
The Needham–Schroeder protocol is one of the two key transport protocols intended for use over an insecure network, both proposed by Roger Needham and Michael Schroeder. These are: The Needham–Schroeder Symmetric Key Protocol, based on a symmetric encryption algorithm. It forms the basis for the Kerberos protocol. This protocol aims to establish a session key between two parties on a network, typically to protect further communication. The Needham–Schroeder Public-Key Protocol, based on public-key cryptography. This protocol is intended to provide mutual authentication between two parties communicating on a network, but in its proposed form is insecure. The symmetric protocol Here, Alice initiates the communication to Bob . is a server trusted by both parties. In the communication: and are identities of Alice and Bob respectively is a symmetric key known only to and is a symmetric key known only to and and are nonces generated by and respectively is a symmetric, generated key, which will be the session key of the session between and The protocol can be specified as follows in security protocol notation: Alice sends a message to the server identifying herself and Bob, telling the server she wants to communicate with Bob. The server generates and sends back to Alice a copy encrypted under for Alice to forward to Bob and also a copy for Alice. Since Alice may be requesting keys for several different people, the nonce assures Alice that the message is fresh and that the server is replying to that particular message and the inclusion of Bob's name tells Alice who she is to share this key with. Alice forwards the key to Bob who can decrypt it with the key he shares with the server, thus authenticating the data. Bob sends Alice a nonce encrypted under to show that he has the key. Alice performs a simple operation on the nonce, re-encrypts it and sends it back verifying that she is still alive and that she holds the key. Attacks on the protocol The protocol is vulnerable to a replay attack (as identified by Denning and Sacco). If an attacker uses an older, compromised value for , he can then replay the message to Bob, who will accept it, being unable to tell that the key is not fresh. Fixing the attack This flaw is fixed in the Kerberos protocol by the inclusion of a timestamp. It can also be fixed with the use of nonces as described below. At the beginning of the protocol: Alice sends to Bob a request. Bob responds with a nonce encrypted under his key with the Server. Alice sends a message to the server identifying herself and Bob, telling the server she wants to communicate with Bob. Note the inclusion of the nonce. The protocol then continues as described through the final three steps as described in the original protocol above. Note that is a different nonce from . The inclusion of this new nonce prevents the replaying of a compromised version of since such a message would need to be of the f
https://en.wikipedia.org/wiki/The%20Emperor%27s%20New%20Mind
The Emperor's New Mind: Concerning Computers, Minds and The Laws of Physics is a 1989 book by the mathematical physicist Sir Roger Penrose. Penrose argues that human consciousness is non-algorithmic, and thus is not capable of being modeled by a conventional Turing machine, which includes a digital computer. Penrose hypothesizes that quantum mechanics plays an essential role in the understanding of human consciousness. The collapse of the quantum wavefunction is seen as playing an important role in brain function. Most of the book is spent reviewing, for the scientifically-minded lay-reader, a plethora of interrelated subjects such as Newtonian physics, special and general relativity, the philosophy and limitations of mathematics, quantum physics, cosmology, and the nature of time. Penrose intermittently describes how each of these bears on his developing theme: that consciousness is not "algorithmic". Only the later portions of the book address the thesis directly. Overview Penrose states that his ideas on the nature of consciousness are speculative, and his thesis is considered erroneous by experts in the fields of philosophy, computer science, and robotics. The Emperor's New Mind attacks the claims of artificial intelligence using the physics of computing: Penrose notes that the present home of computing lies more in the tangible world of classical mechanics than in the imponderable realm of quantum mechanics. The modern computer is a deterministic system that for the most part simply executes algorithms. Penrose shows that, by reconfiguring the boundaries of a billiard table, one might make a computer in which the billiard balls act as message carriers and their interactions act as logical decisions. The billiard-ball computer was first designed some years ago by Edward Fredkin and Tommaso Toffoli of the Massachusetts Institute of Technology. Reception Following the publication of the book, Penrose began to collaborate with Stuart Hameroff on a biological analog to quantum computation involving microtubules, which became the foundation for his subsequent book, Shadows of the Mind: A Search for the Missing Science of Consciousness. Penrose won the Science Book Prize in 1990 for The Emperor's New Mind. According to an article in the American Journal of Physics, Penrose incorrectly claims a barrier far away from a localized particle can affect the particle. See also Alan Turing Anathem Church–Turing thesis Mind–body dualism Orchestrated objective reduction Quantum mind Raymond Smullyan Shadows of the Mind "The Emperor's New Clothes" Turing test References 1989 non-fiction books Works about consciousness English-language books English non-fiction books Mathematics books Oxford University Press books Philosophy of artificial intelligence Philosophy of mind literature Popular physics books Quantum mind Science books Turing machine Works by Roger Penrose
https://en.wikipedia.org/wiki/Otway%E2%80%93Rees%20protocol
The Otway–Rees protocol is a computer network authentication protocol designed for use on insecure networks (e.g. the Internet). It allows individuals communicating over such a network to prove their identity to each other while also preventing eavesdropping or replay attacks and allowing for the detection of modification. The protocol can be specified as follows in security protocol notation, where Alice is authenticating herself to Bob using a server S (M is a session-identifier, NA and NB are nonces): Note: The above steps do not authenticate B to A. This is one of the protocols analysed by Burrows, Abadi and Needham in the paper that introduced an early version of Burrows–Abadi–Needham logic. Attacks on the protocol There are a variety of attacks on this protocol currently published. Interception attacks These attacks leave the intruder with the session key and may exclude one of the parties from the conversation. Boyd and Mao observe that the original description does not require that S check the plaintext A and B to be the same as the A and B in the two ciphertexts. This allows an intruder masquerading as B to intercept the first message, then send the second message to S constructing the second ciphertext using its own key and naming itself in the plaintext. The protocol ends with A sharing a session key with the intruder rather than B. Gürgens and Peralta describe another attack which they name an arity attack. In this attack the intruder intercepts the second message and replies to B using the two ciphertexts from message 2 in message 3. In the absence of any check to prevent it, M (or perhaps M,A,B) becomes the session key between A and B and is known to the intruder. Cole describes both the Gürgens and Peralta arity attack and another attack in his book Hackers Beware. In this the intruder intercepts the first message, removes the plaintext A,B and uses that as message 4 omitting messages 2 and 3. This leaves A communicating with the intruder using M (or M,A,B) as the session key. Disruptive attacks This attack allows the intruder to disrupt the communication but does not allow the intruder to gain access to it. One problem with this protocol is that a malicious intruder can arrange for A and B to end up with different keys. Here is how: after A and B execute the first three messages, B has received the key . The intruder then intercepts the fourth message. He resends message 2, which results in S generating a new key , subsequently sent to B. The intruder intercepts this message too, but sends to A the part of it that B would have sent to A. So now A has finally received the expected fourth message, but with instead of . See also Kerberos (protocol) Needham–Schroeder protocol Yahalom (protocol) Wide Mouth Frog protocol References Computer access control protocols Authentication protocols Key transport protocols Symmetric-key cryptography
https://en.wikipedia.org/wiki/Ned%20Lagin
Ned Lagin (born March 17, 1948) is an American artist, photographer, scientist, composer, and keyboardist. Lagin is considered a pioneer in the development and use of minicomputers and personal computers in real-time stage and studio music composition and performance. He is known for his electronic music composition Seastones, for performing with the Grateful Dead, and for his photography and art. Early years Ned Lagin was born in New York City and raised on Long Island in Roslyn Heights, New York. Growing up, Lagin was influenced by classical and jazz music, and the modern music and art cultures of New York City in the 1960s. He started photography with a Kodak Baby Brownie Special at the age of five, and piano lessons and science, natural history, and electronic projects at the age of six. He attended the Wheatley School in Old Westbury, New York, was awarded two National Science Foundation Scholarships, and attended the Massachusetts Institute of Technology with the intention of becoming an astronaut. Lagin received a degree in molecular biology and humanities from MIT in 1971, where he studied with John Harbison, Gregory Tucker, David Epstein, Noam Chomsky, Gian-Carlo Rota, Salvador Luria, and Jerome Lettvin. Chomsky's generative grammar concepts inspired Lagin's thinking about creating generative music forms (1968), and Lettvin connected him to the writings of Norbert Wiener and Warren McCulloch, and more generally to cybernetics. While at MIT, Lagin also completed jazz coursework at the Berklee School of Music. He was deeply influenced by the jazz world in New York City, particularly pianist Bill Evans, whom he met in Boston and saw perform many times in New York and Boston in the late 1960s and early 1970s. During this period, Evans wrote out some of his tunes for Lagin. His piano teachers included Dean Earl, (a Charlie Parker sideman), Ray Santisi (a sideman with Charlie Parker, Stan Getz, Dexter Gordon), and he studied jazz improvisation with Lee Konitz, (who played with Lenny Tristano, early Miles Davis, many others). Lagin played piano in the MIT Concert Jazz Band and MIT Jazz Quintet led by Herb Pomeroy, a sideman with Duke Ellington and Stan Getz. In the autumn of 1971, Lagin began graduate study in composition as an Irving Fine Fellow at Brandeis University, where he studied with Josh Rifkin and Seymour Shifrin. He completed a symphony, a string quartet, jazz big band pieces, and electronic pieces before dropping out and permanently relocating to the Bay Area. Performing with the Grateful Dead In early 1970, Lagin initiated a correspondence with Jerry Garcia after seeing the Grateful Dead at the Boston Tea Party in 1969. In May 1970, he helped facilitate a concert and free live outdoor performance featuring the band at MIT that coincided with the Kent State shootings. That summer, Lagin, at Garcia's invitation, visited San Francisco and contributed piano to "Candyman" during the American Beauty album sessions, played in sev
https://en.wikipedia.org/wiki/Wide%20Mouth%20Frog%20protocol
The Wide-Mouth Frog protocol is a computer network authentication protocol designed for use on insecure networks (the Internet for example). It allows individuals communicating over a network to prove their identity to each other while also preventing eavesdropping or replay attacks, and provides for detection of modification and the prevention of unauthorized reading. This can be proven using Degano. The protocol was first described under the name "The Wide-mouthed-frog Protocol" in the paper "A Logic of Authentication" (1990), which introduced Burrows–Abadi–Needham logic, and in which it was an "unpublished protocol ... proposed by" coauthor Michael Burrows. The paper gives no rationale for the protocol's whimsical name. The protocol can be specified as follows in security protocol notation: A, B, and S are identities of Alice, Bob, and the trusted server respectively and are timestamps generated by A and S respectively is a symmetric key known only to A and S is a generated symmetric key, which will be the session key of the session between A and B is a symmetric key known only to B and S Note that to prevent active attacks, some form of authenticated encryption (or message authentication) must be used. The protocol has several problems: A global clock is required. The server S has access to all keys. The value of the session key is completely determined by A, who must be competent enough to generate good keys. It can replay messages within the period when the timestamp is valid. A is not assured that B exists. The protocol is stateful. This is usually undesired because it requires more functionality and capability from the server. For example, S must be able to deal with situations in which B is unavailable. See also Alice and Bob Kerberos (protocol) Needham–Schroeder protocol Neuman–Stubblebine protocol Otway–Rees protocol Yahalom (protocol) References Computer access control protocols
https://en.wikipedia.org/wiki/Zooming%20user%20interface
In computing, a zooming user interface or zoomable user interface (ZUI, pronounced zoo-ee) is a graphical environment where users can change the scale of the viewed area in order to see more detail or less, and browse through different documents. A ZUI is a type of graphical user interface (GUI). Information elements appear directly on an infinite virtual desktop (usually created using vector graphics), instead of in windows. Users can pan across the virtual surface in two dimensions and zoom into objects of interest. For example, as you zoom into a text object it may be represented as a small dot, then a thumbnail of a page of text, then a full-sized page and finally a magnified view of the page. ZUIs use zooming as the main metaphor for browsing through hyperlinked or multivariate information. Objects present inside a zoomed page can in turn be zoomed themselves to reveal further detail, allowing for recursive nesting and an arbitrary level of zoom. When the level of detail present in the resized object is changed to fit the relevant information into the current size, instead of being a proportional view of the whole object, it's called semantic zooming. Some consider the ZUI paradigm as a flexible and realistic successor to the traditional windowing GUI, being a Post-WIMP interface. History Ivan Sutherland presented the first program for zooming through and creating graphical structures with constraints and instancing, on a CRT in his Sketchpad program in 1962. A more general interface was done by the Architecture Machine Group in the 1970s at MIT. Hand tracking, touchscreen, joystick, and voice control were employed to control an infinite plane of projects, documents, contacts, video and interactive programs. One of the instances of this project was called Spatial Dataland. Another GUI environment of the 70's, which used the zooming idea was Smalltalk at Xerox PARC, which had infinite desktops (only later named such by Apple Computer), that could be zoomed in upon from a birds eye view after the user had recognized a miniature of the window setup for the project. The longest running effort to create a ZUI has been the Pad++ project begun by Ken Perlin, Jim Hollan, and Ben Bederson at New York University and continued at the University of New Mexico under Hollan's direction. After Pad++, Bederson developed Jazz, then Piccolo, and now Piccolo2D at the University of Maryland, College Park, which is maintained in Java and C#. More recent ZUI efforts include Archy by the late Jef Raskin, ZVTM developed at INRIA (which uses the Sigma lens technique), and the simple ZUI of the Squeak Smalltalk programming environment and language. The term ZUI itself was coined by Franklin Servan-Schreiber and Tom Grauman while they worked together at the Sony Research Laboratories. They were developing the first Zooming User Interface library based on Java 1.0, in partnership with Prof. Ben Bederson, University of New Mexico, and Prof. Ken Perlin, New Yor
https://en.wikipedia.org/wiki/Cybergeneration
CyberGeneration is a follow-up to the R. Talsorian's Cyberpunk 2020 role-playing game. CyberGeneration was originally published as a supplement for Cyberpunk, but later re-released as a fully featured game in its own right under the title CyberGeneration Revolution 2.0. It is set in the year 2027, 7 years after the events in Cyberpunk 2020. The game's timeline doesn't correspond with that of the later third edition of Cyberpunk, which makes no mention of any of its contents or setting elements. The game was licensed out in 2004 to Firestorm Ink under freelance writer Jonathan Lavallee though they are no longer the license holders. Overview Cybergeneration (1993), published by R. Talsorian featured a new setting for Cyberpunk 2020 in an alternate version of the year 2027 with a world where the corporations are now the new governments, and where player characters are young adults that have more heroic motives in opposing the Machine. In terms of tone, CyberGeneration differs from its predecessor somewhat, as the player characters take the part of nanotech-enhanced youngsters in an oppressive world ruled by adults who fear and seek to control them. The special powers of the CyberEvolved children give the game a definite superhero flavor. The players can be many different roles such as actors, inventors or motorbike races but they can be many more. A second edition of Cybergeneration was published in 1995. Firestorm Ink licensed the rights to Cybergeneration in 2004. They published several products including Generation Gap (2005) – which R. Talsorian produced a decade earlier but never published - and Researching Medicine (2005) - a MedTech update set in 2027, ending the line with the PDF-only Mile High Dragon (2009). Background The back story revolves around the "Fox Run" incident of 2025, in which a transport of supposed scientific equipment crashed and accidentally released a weaponized nano-virus called the "Carbon Plague". Adult humans infected by it died horribly after the virus rewrote the victims' genetic code and warped their bodies. However, the virus had a different effect on children and teenagers. Since they haven't fully matured, it altered their bodies. This granted them nanotech-enhanced powers, and made them immune carriers if they survived the illness. Society dramatically fears the capabilities of these "CyberEvolved" children, which drives them underground. CyberEvolved, also called the "Changed", also belong to "yogangs" or youth gangs, subcultures that have distinct philosophies, styles, and special skills. The oppressive mega-corporation Arasaka manages to dominate the US Government and gets its candidate David Whindam elected President of the new Incorporated States of America (ISA). Its laissez-faire government works with the corporations directly, becoming their puppet. The "Bureau of Relocation" (BuReloc) is a paramilitary force that runs prison camps for "unproductive" citizens and hunts down the Changed. Mul
https://en.wikipedia.org/wiki/Display%20size
On 2D displays, such as computer monitors and TVs, the display size or viewable image size (VIS) is the physical size of the area where pictures and videos are displayed. The size of a screen is usually described by the length of its diagonal, which is the distance between opposite corners, usually in inches. It is also sometimes called the physical image size to distinguish it from the "logical image size", which describes a screen's display resolution and is measured in pixels. History The size of a screen is usually described by the length of its diagonal, which is the distance between opposite corners, usually in inches. It is also sometimes called the physical image size to distinguish it from the "logical image size," which describes a screen's display resolution and is measured in pixels. The method of measuring screen size by its diagonal was inherited from the method used for the first generation of CRT television, when picture tubes with circular faces were in common use. Being circular, the external diameter of the bulb was used to describe their size. Since these circular tubes were used to display rectangular images, the diagonal measurement of the visible rectangle was smaller than the diameter of the tube due to the thickness of the glass surrounding the phosphor screen (which was hidden from the viewer by the casing and bezel). This method continued even when cathode ray tubes were manufactured as rounded rectangles; it had the advantage of being a single number specifying the size, and was not confusing when the aspect ratio was universally 4:3. In the US, when virtually all TV tubes were 4:3, the size of the screen was given as the true screen diagonal with a V following it (this was a requirement in the US market but not elsewhere). In virtually all other markets, the size of the outer diameter of the tube was given. What was a 27V in the US could be a 28" elsewhere. However the V terminology was frequently dropped in US advertising referring to a 27V as a 27". This was not misleading for the consumer as the seller had to give the actual screen size by law. Flat panel displays by contrast use the actual diagonal of their visible display size, thus the size is the actual size presented to the viewer in all markets. This means that a similarly specified size of display will be larger as a flat panel display compared with a cathode ray tube display. When the common aspect ratio went from 4:3 to 16:9, the new widescreens were labeled with a W in the US. A screen that is approximately the same height as a 27V would be a 32W. Vizio and other US TV manufacturers have introduced even wider screens with a 21:9 aspect ratio in order to match aspect ratios used in cinemas. In order to gauge the relative sizes of these new screens, the screen aspect must be considered. In a commercial market where multiple aspect ratios are being sold, it will always take two numbers to describe the screen size, some combination of diagonal, aspect r
https://en.wikipedia.org/wiki/Iterative%20deepening%20depth-first%20search
In computer science, iterative deepening search or more specifically iterative deepening depth-first search (IDS or IDDFS) is a state space/graph search strategy in which a depth-limited version of depth-first search is run repeatedly with increasing depth limits until the goal is found. IDDFS is optimal, meaning that it finds the shallowest goal. Since it visits all the nodes in the search tree down to depth before visiting any nodes at depth , the cumulative order in which nodes are first visited is effectively the same as in breadth-first search. However, IDDFS uses much less memory. Algorithm for directed graphs The following pseudocode shows IDDFS implemented in terms of a recursive depth-limited DFS (called DLS) for directed graphs. This implementation of IDDFS does not account for already-visited nodes. function IDDFS(root) is for depth from 0 to ∞ do found, remaining ← DLS(root, depth) if found ≠ null then return found else if not remaining then return null function DLS(node, depth) is if depth = 0 then if node is a goal then return (node, true) else return (null, true) (Not found, but may have children) else if depth > 0 then any_remaining ← false foreach child of node do found, remaining ← DLS(child, depth−1) if found ≠ null then return (found, true) if remaining then any_remaining ← true (At least one node found at depth, let IDDFS deepen) return (null, any_remaining) If the goal node is found by DLS, IDDFS will return it without looking deeper. Otherwise, if at least one node exists at that level of depth, the remaining flag will let IDDFS continue. 2-tuples are useful as return value to signal IDDFS to continue deepening or stop, in case tree depth and goal membership are unknown a priori. Another solution could use sentinel values instead to represent not found or remaining level results. Properties IDDFS achieves breadth-first search's completeness (when the branching factor is finite) using depth-first search's space-efficiency. If a solution exists, it will find a solution path with the fewest arcs. Iterative deepening visits states multiple times, and it may seem wasteful. However, if IDDFS explores a search tree to depth , most of the total effort is in exploring the states at depth . Relative to the number of states at depth , the cost of repeatedly visiting the states above this depth is always small. The main advantage of IDDFS in game tree searching is that the earlier searches tend to improve the commonly used heuristics, such as the killer heuristic and alpha–beta pruning, so that a more accurate estimate of the score of various nodes at the final depth search can occur, and the search completes more quickly since it is done in a better order. For example, alpha–beta pruning is most effic
https://en.wikipedia.org/wiki/Decision%20table
Decision tables are a concise visual representation for specifying which actions to perform depending on given conditions. They are algorithms whose output is a set of actions. The information expressed in decision tables could also be represented as decision trees or in a programming language as a series of if-then-else and switch-case statements. Overview Each decision corresponds to a variable, relation or predicate whose possible values are listed among the condition alternatives. Each action is a procedure or operation to perform, and the entries specify whether (or in what order) the action is to be performed for the set of condition alternatives the entry corresponds to. To make them more concise, many decision tables include in their condition alternatives a don't care symbol. This can be a hyphen or blank, although using a blank is discouraged as it may merely indicate that the decision table has not been finished. One of the uses of decision tables is to reveal conditions under which certain input factors are irrelevant on the actions to be taken, allowing these input tests to be skipped and thereby streamlining decision-making procedures. Aside from the basic four quadrant structure, decision tables vary widely in the way the condition alternatives and action entries are represented. Some decision tables use simple true/false values to represent the alternatives to a condition (similar to if-then-else), other tables may use numbered alternatives (similar to switch-case), and some tables even use fuzzy logic or probabilistic representations for condition alternatives. In a similar way, action entries can simply represent whether an action is to be performed (check the actions to perform), or in more advanced decision tables, the sequencing of actions to perform (number the actions to perform). A decision table is considered balanced or complete if it includes every possible combination of input variables. In other words, balanced decision tables prescribe an action in every situation where the input variables are provided. Example The limited-entry decision table is the simplest to describe. The condition alternatives are simple Boolean values, and the action entries are check-marks, representing which of the actions in a given column are to be performed. The following balanced decision table is an example in which a technical support company writes a decision table to enable technical support employees to efficiently diagnose printer problems based upon symptoms described to them over the phone from their clients. This is just a simple example, and it does not necessarily correspond to the reality of printer troubleshooting. Even so, it demonstrates how decision tables can scale to several conditions with many possibilities. Software engineering benefits Decision tables, especially when coupled with the use of a domain-specific language, allow developers and policy experts to work from the same information, the decision tab
https://en.wikipedia.org/wiki/Nick%20Palmer
Nicholas Douglas Palmer (born 5 February 1950) is a British politician, translator and computer scientist. He was the Labour Party Member of Parliament (MP) for Broxtowe in Nottinghamshire from 1997 until he lost the seat at the 2010 general election to Conservative Anna Soubry, by 390 votes. Described by Andrew Roth as "quietly effective", he was Parliamentary Private Secretary (PPS) to the Minister of State, Margaret Beckett, in the Department for Environment, Food and Rural Affairs until April 2005. He then became PPS to the Minister of State, Malcolm Wicks, first in the Department of Trade and Industry, and later in the Department for Business, Enterprise and Regulatory Reform until Wicks stood down in October 2008. Early life Palmer's father was a translator/editor and his mother was a language teacher. He is the cousin of Anthony Palmer, a former Deputy Chief of the Defence Staff. Palmer attended International Schools in Copenhagen and Vienna. He was awarded an MSc at Copenhagen University and a PhD in Mathematics from Birkbeck College, University of London. He also studied at Massachusetts Institute of Technology (MIT) where he researched artificial intelligence and language translation. Professional life Palmer speaks six languages, and has worked as a professional translator of Danish and German for the European Commission and other clients. He was born with a cleft palate and was the first such person to enter Parliament. As a computer scientist, he developed the COMPACT clinical trials package for the Medical Research Council. Joining the Swiss pharmaceutical firm Ciba-Geigy, he became head of Novartis Internet Service when Ciba-Geigy merged with Sandoz to form Novartis. Board and computer wargames Palmer has written three books about board wargames (The Comprehensive Guide to Board Wargaming (1977), The Best of Board Wargaming (1980), and Beyond the Arcade: Adventures and Wargames on Your Computer (1984)). He designed and developed a computer game about the Battle of Britain, named Their Finest Hour. Palmer still attends international conventions, winning the Diplomacy championship at the World Boardgaming Championships in 2007, as well as giving a seminar in 2008 comparing the traits needed to succeed in wargaming to the traits needed to succeed in politics. He co-founded and edited Flagship magazine in 1983, which focused on play-by-mail games. A keen card player, he has represented the House of Commons at bridge. Parliamentary career Palmer joined the Labour Party on his twenty-first birthday and was selected as the Labour candidate for the ultra-safe Conservative seat of Chelsea in the 1983 general election. Prior to contesting Broxtowe, he edited and published a magazine to represent the views of ordinary Labour party members – Grass Roots. Legislation and Committee Work While an MP, he served on a number of Select committees including the European Scrutiny Committee, the Northern Ireland Affairs Committee, and the T
https://en.wikipedia.org/wiki/List%20of%20Mexican%20states%20by%20population
The following table is a list of the 31 federal states of Mexico plus Mexico City, ranked in order of their total population based on data from the last three National Population Census in 2020, 2010 and 2000. See also Mexico States of Mexico Geography of Mexico List of Mexican states by area List of Mexican states by population growth rate Ranked list of Mexican states List of Mexican states by HDI References Population Mexico, population
https://en.wikipedia.org/wiki/FX%20%28TV%20channel%29
FX is an American pay television channel owned by FX Networks, LLC, a subsidiary of the Disney Entertainment business segment and division of The Walt Disney Company. It is based at the Fox Studios lot in Century City, California. FX was originally launched by News Corporation on June 1, 1994, and later became one of the properties that was included in the acquisition of 21st Century Fox by Disney in 2019. The network's original programming aspires to the standards of premium cable channels in regard to mature themes and content, high-quality writing, directing and acting. Sister channels FXM and FXX were launched in 1994 and 2013, respectively. FX also carries reruns of theatrical films and terrestrial-network sitcoms. Advertising-free content was available through the FX+ premium subscription service until it was shut down on August 21, 2019. As of September 2018, FX is available to approximately 89.2 million television households (96.7% of households with cable) in the United States. In addition to the flagship U.S. network, the "FX" name is licensed to a number of related pay television channels in various countries around the world. History 1994–1997: "TV Made Fresh Daily" FX, originally stylized as "fX", launched on June 1, 1994. Broadcasting from a large "apartment" in Manhattan's Flatiron District, fX was one of the first forays into large-scale interactive television. The channel centered on original programming, which was broadcast live every day from the "fX Apartment", and rebroadcasts of classic television shows from the 1960s, 1970s and 1980s, such as Batman, Wonder Woman, Eight Is Enough, Nanny and the Professor and The Green Hornet. fX had two taglines during this period: "TV Made Fresh Daily" and "The World's First Living Television Network". The "f" in the channel's name and logo was rendered in lower-case to portray a type of relaxed friendliness; the stylized "X" represented the channel's roots: the crossing searchlights of the 20th Century Fox logo. The live shows were each mostly focused on one broad topic. Shows included Personal fX (collectibles and antiques), The Pet Department (pets), Under Scrutiny with Jane Wallace (news) and Sound fX (music). The channel's flagship show, Breakfast Time, hosted by Laurie Hibberd and Tom Bergeron and inspired by the British morning show The Big Breakfast, was formatted like an informal magazine show. Breakfast Time and Personal fX would regularly feature the channel's "roving reporters" – which included Suzanne Whang, John Burke and Phil Keoghan – visiting unique places around the United States live via satellite. Other notable fX personalities included Karyn Bryant and Orlando Jones, who were panelists on Sound fX. The channel prided itself on its interactivity with viewers. fX, in 1994, was an early adopter of the internet, embracing e-mail and the World Wide Web as methods of feedback. Most of the shows would feature instant responses to e-mailed questions, and one show, Backc
https://en.wikipedia.org/wiki/Tracing%20garbage%20collection
In computer programming, tracing garbage collection is a form of automatic memory management that consists of determining which objects should be deallocated ("garbage collected") by tracing which objects are reachable by a chain of references from certain "root" objects, and considering the rest as "garbage" and collecting them. Tracing garbage collection is the most common type of garbage collection – so much so that "garbage collection" often refers to tracing garbage collection, rather than other methods such as reference counting – and there are a large number of algorithms used in implementation. Reachability of an object Informally, an object is reachable if it is referenced by at least one variable in the program, either directly or through references from other reachable objects. More precisely, objects can be reachable in only two ways: A distinguished set of roots: objects that are assumed to be reachable. Typically, these include all the objects referenced from anywhere in the call stack (that is, all local variables and parameters in the functions currently being invoked), and any global variables. Anything referenced from a reachable object is itself reachable; more formally, reachability is a transitive closure. The reachability definition of "garbage" is not optimal, insofar as the last time a program uses an object could be long before that object falls out of the environment scope. A distinction is sometimes drawn between syntactic garbage, those objects the program cannot possibly reach, and semantic garbage, those objects the program will in fact never again use. For example: Object x = new Foo(); Object y = new Bar(); x = new Quux(); /* At this point, we know that the Foo object * originally assigned to x will never be * accessed: it is syntactic garbage. */ /* In the following block, y *could* be semantic garbage; * but we won't know until x.check_something() returns * some value -- if it returns at all. */ if (x.check_something()) { x.do_something(y); } System.exit(0); The problem of precisely identifying semantic garbage can easily be shown to be partially decidable: a program that allocates an object X, runs an arbitrary input program P, and uses X if and only if P finishes would require a semantic garbage collector to solve the halting problem. Although conservative heuristic methods for semantic garbage detection remain an active research area, essentially all practical garbage collectors focus on syntactic garbage. Another complication with this approach is that, in languages with both reference types and unboxed value types, the garbage collector needs to somehow be able to distinguish which variables on the stack or fields in an object are regular values and which are references: in memory, an integer and a reference might look alike. The garbage collector then needs to know whether to treat the element as a reference and follow it, or whether it is a primitive value. One common solution is the us
https://en.wikipedia.org/wiki/Robert%20Fano
Roberto Mario "Robert" Fano (11 November 1917 – 13 July 2016) was an Italian-American computer scientist and professor of electrical engineering and computer science at the Massachusetts Institute of Technology. He became a student and working lab partner to Claude Shannon, whom he admired zealously and assisted in the early years of Information Theory. Early life and education Fano was born in Turin, Italy in 1917 to a Jewish family and grew up in Turin. Fano's father was the mathematician Gino Fano, his older brother was the physicist Ugo Fano, and Giulio Racah was a cousin. Fano studied engineering as an undergraduate at the School of Engineering of Torino (Politecnico di Torino) until 1939, when he emigrated to the United States as a result of anti-Jewish legislation passed under Benito Mussolini. He received his S.B. in electrical engineering from MIT in 1941, and upon graduation joined the staff of the MIT Radiation Laboratory. After World War II, Fano continued on to complete his Sc.D. in electrical engineering from MIT in 1947. His thesis, titled "Theoretical Limitations on the Broadband Matching of Arbitrary Impedances", was supervised by Ernst Guillemin. Career Fano's career spans three areas, microwave systems, information theory, and computer science. Fano joined the MIT faculty in 1947 to what was then called the Department of Electrical Engineering. Between 1950 and 1953, he led the Radar Techniques Group at Lincoln Laboratory. In 1954, Fano was made an IEEE Fellow for "contributions in the field of information theory and microwave filters". He was elected to the American Academy of Arts and Sciences in 1958, to the National Academy of Engineering in 1973, and to the National Academy of Sciences in 1978. Fano was known principally for his work on information theory. He developed Shannon–Fano coding in collaboration with Claude Shannon, and derived the Fano inequality. He also invented the Fano algorithm and postulated the Fano metric. In the early 1960s, Fano was involved in the development of time-sharing computers. From 1963 until 1968 Fano served as the founding director of MIT's Project MAC, which evolved to become what is now known as the MIT Computer Science and Artificial Intelligence Laboratory. He also helped to create MIT's original computer science curriculum. In 1976, Fano received the Claude E. Shannon Award for his work in information theory. In 1977 he was recognized for his contribution to the teaching of electrical engineering with the IEEE James H. Mulligan Jr. Education Medal. Fano retired from active teaching in 1984, and died on 13 July 2016 at the age of 98. Bibliography In addition to his work in information theory, Fano also published articles and books about microwave systems, electromagnetism, network theory, and engineering education. His longer publications include: "The Theory of Microwave Filters" and "The Design of Microwave Filters", chapters 9 and 10 in George L. Ragan, ed., Microwave
https://en.wikipedia.org/wiki/MIT%20Computer%20Science%20and%20Artificial%20Intelligence%20Laboratory
Computer Science and Artificial Intelligence Laboratory (CSAIL) is a research institute at the Massachusetts Institute of Technology (MIT) formed by the 2003 merger of the Laboratory for Computer Science (LCS) and the Artificial Intelligence Laboratory (AI Lab). Housed within the Ray and Maria Stata Center, CSAIL is the largest on-campus laboratory as measured by research scope and membership. It is part of the Schwarzman College of Computing but is also overseen by the MIT Vice President of Research. Research activities CSAIL's research activities are organized around a number of semi-autonomous research groups, each of which is headed by one or more professors or research scientists. These groups are divided up into seven general areas of research: Artificial intelligence Computational biology Graphics and vision Language and learning Theory of computation Robotics Systems (includes computer architecture, databases, distributed systems, networks and networked systems, operating systems, programming methodology, and software engineering, among others) In addition, CSAIL hosts the World Wide Web Consortium (W3C). History Computing Research at MIT began with Vannevar Bush's research into a differential analyzer and Claude Shannon's electronic Boolean algebra in the 1930s, the wartime MIT Radiation Laboratory, the post-war Project Whirlwind and Research Laboratory of Electronics (RLE), and MIT Lincoln Laboratory's SAGE in the early 1950s. At MIT, research in the field of artificial intelligence began in late 1950s. Project MAC On July 1, 1963, Project MAC (the Project on Mathematics and Computation, later backronymed to Multiple Access Computer, Machine Aided Cognitions, or Man and Computer) was launched with a $2 million grant from the Defense Advanced Research Projects Agency (DARPA). Project MAC's original director was Robert Fano of MIT's Research Laboratory of Electronics (RLE). Fano decided to call MAC a "project" rather than a "laboratory" for reasons of internal MIT politics – if MAC had been called a laboratory, then it would have been more difficult to raid other MIT departments for research staff. The program manager responsible for the DARPA grant was J. C. R. Licklider, who had previously been at MIT conducting research in RLE, and would later succeed Fano as director of Project MAC. Project MAC would become famous for groundbreaking research in operating systems, artificial intelligence, and the theory of computation. Its contemporaries included Project Genie at Berkeley, the Stanford Artificial Intelligence Laboratory, and (somewhat later) University of Southern California's (USC's) Information Sciences Institute. An "AI Group" including Marvin Minsky (the director), John McCarthy (inventor of Lisp), and a talented community of computer programmers were incorporated into Project MAC. They were interested principally in the problems of vision, mechanical motion and manipulation, and language, which they view as the key
https://en.wikipedia.org/wiki/Battle%20of%20the%20Network%20Stars
Battle of the Network Stars is a series of competitions in which television stars from ABC, CBS and NBC would compete in various sporting events. A total of 19 of these competitions were held between 1976 and 1988, all of which were aired by ABC. In 2003, NBC attempted to revive Battle of the Network Stars with a two-hour special. In 2005, Bravo premiered a revived version of the show named Battle of the Network Reality Stars. Also in 2005, ESPN premiered a short-lived, sports-themed spinoff version of Battle of the Network Stars as Battle of the Gridiron Stars featuring twenty players from the AFC and NFC competing in a variety of tasks that had nothing to do with football. In 2017, ABC revived the series as a summer series which ran from June 29 to September 7, 2017. Broadcast history The first Battle was broadcast on ABC starting in November 1976. The program proved popular and continued for an additional eight and a half years, with subsequent episodes airing approximately every six months until May 1985. One final competition aired in December 1988. NBC tried to revive the competition in August 2003 with Tony Potts as host, but with an intra-network contest consisting of personalities from the NBC family of networks. Typically, episodes were aired twice per calendar year, once during the spring and once during the fall during Nielsen Ratings sweeps weeks. Sports broadcaster Howard Cosell hosted or co-hosted all but one of the first nineteen competitions (he did not host the 1985 edition due to a falling-out with ABC, but he returned for the final edition in 1988), and commented on the action with a semi-serious version of the style for which he was famous. When ABC revived the program as a weekly series in 2017, Mike Greenberg and Joe Tessitore took over the hosting duties. Furthermore, each episode began with a remake of the opening sequence of ABC's Wide World of Sports. Format 1976–1988 All but one of the competitions took place at the sports facilities of Pepperdine University near Malibu, California, the exception being XVIII which was held in Ixtapa, Mexico. Each network was represented by eight or ten of its stars from various series, and one of those people from each team would be elected to serve as the network's team captain. Some of the events were modeled after those used on The Superstars, another Trans World-ABC production that featured athletes from all sports competing against each other for an overall title. Regular events included swimming, kayaking, volleyball, golf, tennis, bowling (on custom-made outdoor lanes), cycling, 3-on-3 football, the baseball dunk, running, and the obstacle course. Also featured as a regular event was a game of "Simon Says", directed by Catskill hotel Grossinger's entertainer Lou Goldstein. Each network received points based on how it performed in the event. After the regular events were over, the lowest scoring network was eliminated from further competition and the two remaining
https://en.wikipedia.org/wiki/DN
DN, dN, or dn may refer to: Science, technology, and mathematics Computing and telecommunications Digital number, the discrete of an analog value sampled by an analog-to-digital converter Directory number in a phone system Distinguished Name, an identifier type in the LDAP protocol Domain name, an identification string used within the Internet Domain Nameserver DOS Navigator, a DOS file manager Mathematics dn (elliptic function), one of Jacobi's elliptic functions Dn, a Coxeter–Dynkin diagram Dn, a dihedral group Other uses in science and technology Decinewton (symbol dN), an SI unit of force Diametre Nominal, the European equivalent of Nominal Pipe Size Diameter of a rolling element bearing in mm multiplied by its speed in rpm Deductive-nomological model, a philosophical model for scientific explanation Double negative T cells, also called CD4−CD8− Diabetic nephropathy Diabetic neuropathy DN Factor, a value used to calculate the correct lubricant for bearings Entertainment Double nil, a bid in the game of Spades (card game) Descriptive chess notation Duke Nukem, a video game character and a game franchise Journalism The Ball State Daily News, the student newspaper of Ball State University in Muncie, Indiana , a Norwegian newspaper , a Swedish newspaper Democracy Now!, the flagship program for the Pacifica Radio network Diário de Notícias, a Portuguese newspaper Places Denmark (WMO country code DN) DN postcode area for Doncaster and surrounding areas, UK Dunedin, New Zealand (commonly abbreviated DN) Dadra and Nagar Haveli, a former union territory of India Other uses DN, IATA code of Dan Air Down (disambiguation) Diebold Nixdorf, American financial technology company Digha Nikaya, a part of the Buddhist Tripitaka International DN, a kind of iceboat Dreadnought, a class of warships Debit note, a commercial document DN, then IATA code for Norwegian Air Argentina DN, then IATA code for Senegal Airlines
https://en.wikipedia.org/wiki/Hough%20transform
The Hough transform is a feature extraction technique used in image analysis, computer vision, and digital image processing. The purpose of the technique is to find imperfect instances of objects within a certain class of shapes by a voting procedure. This voting procedure is carried out in a parameter space, from which object candidates are obtained as local maxima in a so-called accumulator space that is explicitly constructed by the algorithm for computing the Hough transform. The classical Hough transform was concerned with the identification of lines in the image, but later the Hough transform has been extended to identifying positions of arbitrary shapes, most commonly circles or ellipses. The Hough transform as it is universally used today was invented by Richard Duda and Peter Hart in 1972, who called it a "generalized Hough transform" after the related 1962 patent of Paul Hough. The transform was popularized in the computer vision community by Dana H. Ballard through a 1981 journal article titled "Generalizing the Hough transform to detect arbitrary shapes". History It was initially invented for machine analysis of bubble chamber photographs (Hough, 1959). The Hough transform was patented as in 1962 and assigned to the U.S. Atomic Energy Commission with the name "Method and Means for Recognizing Complex Patterns". This patent uses a slope-intercept parametrization for straight lines, which awkwardly leads to an unbounded transform space since the slope can go to infinity. The rho-theta parametrization universally used today was first described in although it was already standard for the Radon transform since at least the 1930s. O'Gorman and Clowes' variation is described in The story of how the modern form of the Hough transform was invented is given in Theory In automated analysis of digital images, a subproblem often arises of detecting simple shapes, such as straight lines, circles or ellipses. In many cases an edge detector can be used as a pre-processing stage to obtain image points or image pixels that are on the desired curve in the image space. Due to imperfections in either the image data or the edge detector, however, there may be missing points or pixels on the desired curves as well as spatial deviations between the ideal line/circle/ellipse and the noisy edge points as they are obtained from the edge detector. For these reasons, it is often non-trivial to group the extracted edge features to an appropriate set of lines, circles or ellipses. The purpose of the Hough transform is to address this problem by making it possible to perform groupings of edge points into object candidates by performing an explicit voting procedure over a set of parameterized image objects (Shapiro and Stockman, 304). Detecting lines The simplest case of Hough transform is detecting straight lines. In general, the straight line can be represented as a point (b, m) in the parameter space. However, vertical lines pose a problem. Th
https://en.wikipedia.org/wiki/Timex%20Sinclair%201000
The Timex Sinclair 1000 (or T/S 1000) was the first computer produced by Timex Sinclair, a joint venture between Timex Corporation and Sinclair Research. It was launched in July 1982, with a US sales price of US$99.95, making it the cheapest home computer at the time; it was advertised as "the first computer under $100". The computer was aimed at regular home users. As purchased, the T/S 1000 was fully assembled and ready to be plugged into home televisions, which served as a video monitor. The T/S 1000 was a slightly modified version of the Sinclair ZX81 with an NTSC RF modulator, for use with North American TVs, instead of PAL for European TVs. The T/S 1000 doubled the onboard RAM from 1 KB to 2 KB; further expandable by 16 KB through the cartridge port. The T/S 1000's casing had slightly more internal shielding but remained the same as Sinclair's, including the membrane keyboard. Just like the ZX81, the T/S 1000 had black-and-white graphics and no sound. It was followed in 1983 by an improved version, the Timex Sinclair 1500 (or T/S 1500) which incorporated the 16 KB RAM expansion and featured a lower price (US$80). However, the T/S 1500 did not achieve market success, given that by this time the marketplace was dominated by Commodore, Radio Shack, Atari and Apple. History Timex claimed to have sold 600,000 T/S 1000s in the US by early 1983, and other companies imported localized versions of British software. It sold for in the US when it debuted, making it the cheapest home computer at the time; it was advertised as "the first computer under $100". This pricing initiated a price war with Commodore International, who quickly reduced the price of its VIC-20 to match and later announced a trade-in program offering $100 for any competing computer toward the purchase of a Commodore 64. Since the T/S 1000 was selling for $49 by this time, many customers bought them for the sole purpose of trading them in for a Commodore 64. Like the Sinclair ZX81, the T/S 1000 used 8K BASIC, a version of Sinclair BASIC (a BASIC dialect), as its primary interface and programming language. To make the membrane keyboard-less cumbersome for program entry, the T/S 1000 used a shortcut system of one-letter "keywords" for most commands (e.g., pressing while the cursor was in "keyword mode" would generate the keyword PRINT). Some keywords required a short sequence of keystrokes (e.g., + would generate the keyword LPRINT). One notable thing about this version of BASIC was that, unlike other versions where it's optional in a program, the LET command was used extensively for data. The T/S 1000 was normally plugged into a regular TV that served as a computer monitor. The computer produced a black-and-white display that consisted of 32 columns and 24 lines. Of those lines, 22 were accessible for display, with two reserved for data entry and error messages. The limited graphics were based on geometric shapes contained within the operating system's non-ASCII character set
https://en.wikipedia.org/wiki/Timex%20Sinclair%202068
The Timex Sinclair 2068 (T/S 2068), released in November 1983, was Timex Sinclair's third and last home computer for the United States market. It was also marketed in Canada, Argentina, Portugal and Poland, as Timex Computer 2068 (TC 2068). History Following Timex's ZX81-based T/S 1000 and T/S 1500, a new series of ZX Spectrum-based machines was created. Initially named T/S 2000 (as reflected on the user manual), the machine evolved into the T/S 2048 prototype, and was eventually released as T/S 2068, with the name chosen mainly for marketing reasons. Advertisements described the T/S 2068 as offering 72K of memory, color, and sound for a price under $200. Like the T/S 1500 was announced as a 40K memory machine (16K RAM + 24K ROM), so the 2068 was announced as a 72K machine (48K RAM + 24K ROM). Although Timex Computer Corporation folded in February 1984, the independent Portuguese division continued to sell the machine in Portugal as the Timex Computer 2068, and Poland until 1989, as the Unipolbrit Komputer 2086. Although the Portuguese-made TC 2068 was also sold in Poland, only the Komputer 2086 was actually made there. Timex of Portugal sold 2 versions of TC 2068: the silver TC 2068 version came with a ZX Spectrum emulator cartridge and a black TC 2068 version sold with TimeWord word processing cartridge plus the Timex RS232 Interface to use TimeWord with a RS232 printer. Strangely the black version came with a silver keyboard template with TimeWord commands to be used with the program. It can be removed because it is not glued to the black keyboard template. Although the T/S 2068's main improvements over the original Spectrum were in areas that had come in for widespread criticism (graphics, sound, keyboard and—to a lesser extent—the lack of joystick ports and cartridge support), it was not used as the basis for the Spectrum's successors. The ZX Spectrum+ (1984) changed the keyboard only, and even the ZX Spectrum+ 128K (announced in May 1985, but not released in the UK until February 1986) retained the original machine's graphical capabilities. However, unlike the UK models, the T/S 2068 was not burdened by the requirement of compatibility with previous models. Related machines Timex Sinclair 2048 A cut-down version of the T/S 2068, based on the T/S 2048 prototype named T/S 2048, was cancelled before entering intended production in 1984. This was due to the commercial failure of the T/S 1500. According to an early Timex Sinclair 2000 computer flyer, it would have 16 KB of RAM, add a Kempston-compatible joystick interface and a two color high resolution mode for 80 column text. The 2048 model number was the intended model number for what finally got named Timex Sinclair 2068. In an interview with Lou Galie, senior vice president of technology at Timex, he tells what he claims to be the real story. Danny Ross, Timex Computer Corporation president, was giving a speech. Lou points: "When Danny announced what was supposed to be the 2048, h
https://en.wikipedia.org/wiki/Miles%20Gordon%20Technology
Miles Gordon Technology, known as MGT, was a small British company, initially specialising in high-quality add-ons for the ZX Spectrum home computer. It was founded in June 1986 in Cambridge, England by Alan Miles and Bruce Gordon, former employees of Sinclair Research, after Sinclair sold the rights for the Spectrum to Amstrad. They moved to Swansea, Wales, in May 1989, became a public company in July 1989 and went into receivership in June 1990. The DISCiPLE and +D As the ZX Spectrum became hugely popular, the lack of a mass storage system became a problem for more serious users. While Sinclair's response, the ZX Interface 1 and ZX Microdrive, was very cheap and technologically innovative, it was also rather limited. Many companies developed interfaces to connect floppy disk drives to the ZX Spectrum, one of the most successful being the Opus Discovery; however, these were all to some degree incompatible with Sinclair's system. MGT's approach was different. It produced two different floppy-disk interfaces for the Spectrum, first the DISCiPLE (marketed by Rockfort Products) and later the cut-down +D interface (marketed by MGT themselves). Both, however, shared certain features: A Shugart-compatible port for connecting one or two floppy diskette drives (the de facto standard created by Shugart Associates) A parallel printer port A "magic button" The latter generated a non-maskable interrupt, freezing any software running on the Spectrum and allowing it to be saved to disk. This made it simple to store tape-based games on disk, to take screenshots and to enter cheat codes. A duplicate expansion connector at the back allowed other peripherals to be daisy chained, although the complexity of the DISCiPLE meant that many would not work correctly. However, the real innovation was in the ROM. Unlike most of the competing systems, this was compatible with the Sinclair's extended ROM, meaning that the same BASIC commands used to operate Microdrives or the ZX Printer now could control floppy disk drives or a standard parallel printer. As well as being BASIC-compatible, it also mimicked the machine code entry points in the ZX Interface 1 - the so-called "hook codes". This meant that any Microdrive-specific software could use floppy disk drives connected to MGT interfaces instead without modification, provided the hook codes were used. The floppy drives simply appeared to Microdrive-aware applications to be very big, fast Microdrives. Sinclair's Microdrive command syntax was so complex that a selling point of many disk interfaces was that their commands were simpler. While loading from tape required a simple: LOAD "progname" the equivalent Microdrive syntax was: LOAD *"m";1;"progname" Given the complexity of entering punctuation on the Spectrum's tiny keyboard, this was cumbersome. In addition to supporting the Sinclair syntax, MGT's code reduced the command to: LOAD d1"progname" Later, MGT produced the Lifetime Drive range of floppy disk
https://en.wikipedia.org/wiki/Kanotix
Kanotix, also referred to as KANOTIX, is an operating system based on Debian, with advanced hardware detection. It can run from an optical disc drive or other media i.e. USB-stick without using a hard disk drive. Kanotix uses KDE Software Compilation as the default desktop environment. Since 2013 the newer releases ship with LXDE as a second lightweight desktop environment. Unlike other similar Linux-distributions Kanotix is a rolling release. Nightly builds are automated builds every night of the latest development code of KANOTIX and with the latest packages from the repositories. The name "Kanotix" is derived from the founder's nickname "Kano". Kanotix's mascot is a fangtooth. Content Kanotix is based on the newest Debian stable (the last published version is Kanotix "Silverfire 2019" based on Debian 10 "Buster". It also provides its own packages and scripts and many backports. Kanotix also provides an optimized kernel with additional patches. Kanotix includes about 1,500 software packages, among others KDE Software Compilation, the default desktop environment and LXDE the lightweight version Amarok, Video Disk Recorder Internet access software, including KDE Network Manager, KPPP dialer, Wireless LAN-driver and NdisWrapper Firefox web browser, Thunderbird mail/news client, Pidgin instant messenger K3b, for CD (and DVD) authoring and backup GIMP, an image-manipulation program, also Inkscape a free and open-source vector graphics editor GParted and other tools for data rescue and system repair Network analysis and administration tools LibreOffice, the office suite (backports) Programming and development tools NTFS-3G used by default Automatic installation of graphic-drivers with nvidia and fglrx-scripts with dkms support. Scripts for additional multimedia support (current mplayer-versions) Wine (Software) updated to newest versions Usage Kanotix is designed for multiple-purpose usage so that it can be used in live mode on different types of media (DVD, hard disk, and USB flash drive) and includes an installation tool for installing Kanotix to the hard drive. The distribution ships with the latest kernel which is carefully patched with fixes and drivers for most modern hardware. Kanotix is an ideal tool for testing, data rescue, or for working and safe surfing and mailing on different machines e.g. in an Internet cafe. Live mode The Live mode allows it to work without any installation. As Kanotix comes with unionfs and aufs-support one can "install" additional packages by using APT (via connection to the Internet). For USB-users the so-called "persistent mode" allows to save data changes back to the USB storage device and the stored data and customized settings can used again on following boots. Using USB flash drive (when supported by BIOS), is of course much faster than booting from CD or DVD. Installation Kanotix can be installed to the hard disk using the (graphical) acritoxinstaller, which, depending on optical drive, ha
https://en.wikipedia.org/wiki/Enigmail
Enigmail is a data encryption and decryption extension for Mozilla Thunderbird and the Postbox that provides OpenPGP public key e-mail encryption and signing. Enigmail works under Microsoft Windows, Unix-like, and Mac OS X operating systems. Enigmail can operate with other mail clients compatible with PGP/MIME and inline PGP such as: Microsoft Outlook with Gpg4win package installed, Gnome Evolution, KMail, Claws Mail, Gnus, Mutt. Its cryptographic functionality is handled by GNU Privacy Guard. In their default configuration, Thunderbird and SeaMonkey provide e-mail encryption and signing using S/MIME, which relies on X.509 keys provided by a centralised certificate authority. Enigmail adds an alternative mechanism where cooperating users can instead use keys provided by a web of trust, which relies on multiple users to endorse the authenticity of the sender's and recipient's credentials. In principle this enhances security, since it does not rely on a centralised entity which might be compromised by security failures or engage in malpractice due to commercial interests or pressure from the jurisdiction in which it resides. Enigmail was first released in 2001 by Ramalingam Saravanan, and since 2003 maintained by Patrick Brunschwig. Both Enigmail and GNU Privacy Guard are free, open-source software. Enigmail with Thunderbird is now the most popular PGP setup. Enigmail has announced its support for the new "pretty Easy privacy" (p≡p) encryption scheme in a joint Thunderbird extension to be released in December 2015. As of June 2016 the FAQ note it will be available in Q3 2016. Enigmail also supports Autocrypt exchange of cryptographic keys since version 2.0. In October 2019, the developers of Thunderbird announced built-in support for encryption and signing based on OpenPGP Thunderbird 78 to replace the Enigmail add-on. The background is a change in the code base of Thunderbird, removing support for legacy add-ons. Since this would require a rewrite from scratch for Enigmail, Patrick Brunschwig instead supports the Thunderbird team in a native implementation in Thunderbird. Enigmail will be maintained for Thunderbird 68 until 6 months after the release of Thunderbird 78. The support of Enigmail for Postbox will be unaffected. See also GNU Privacy Guard OpenPGP References External links Cryptographic software Thunderbird WebExtensions OpenPGP Free email software MacOS security software Windows security software Unix security software MacOS Internet software Windows Internet software Unix Internet software Cross-platform free software
https://en.wikipedia.org/wiki/Sun%20Enterprise
Sun Enterprise is a range of UNIX server computers produced by Sun Microsystems from 1996 to 2001. The line was launched as the Sun Ultra Enterprise series; the Ultra prefix was dropped around 1998. These systems are based on the 64-bit UltraSPARC microprocessor architecture and related to the contemporary Ultra series of computer workstations. Like the Ultra series, they run Solaris. Various models, from single-processor entry-level servers to large high-end multiprocessor servers were produced. The Enterprise brand was phased out in favor of the Sun Fire model line from 2001 onwards. Ultra workstation-derived servers The first UltraSPARC-I-based servers produced by Sun, launched in 1995, are the UltraServer 1 and UltraServer 2. These are server configurations of the Ultra 1 and Ultra 2 workstations respectively. These were later renamed Ultra Enterprise 1 and Ultra Enterprise 2 for consistency with other server models. Later these were joined by the Ultra Enterprise 150, which comprises an Ultra 1 motherboard in a tower-style enclosure with 12 internal disk bays. In 1998, Sun launched server configurations of the UltraSPARC-IIi-based Ultra 5 and Ultra 10 workstations, called the Enterprise Ultra 5S and Enterprise Ultra 10S respectively. Entry-level servers The Sun Enterprise 450 is a rack-mountable entry-level multiprocessor server launched in 1997, capable of up to four UltraSPARC II processors. The Sun Enterprise 250 is a two-processor version launched in 1998. These were later joined by the Enterprise 220R and Enterprise 420R rack-mount servers in 1999. The 220R and 420R models are respectively based on the motherboards of the Ultra 60 and Ultra 80 workstations. The 250 was replaced by the Sun Fire V250, the 450 by the Sun Fire V880. The 220R was superseded by the Sun Fire 280R and the 420R by the Sun Fire V480. Ultra Enterprise XX00 mid-range servers In 1996, Sun replaced the SPARCserver 1000E and SPARCcenter 2000E models with the Ultra Enterprise 3000, 4000, 5000 and 6000 servers. These are multiprocessor servers based on a common hardware architecture incorporating the Gigaplane packet-switched processor/memory bus and UltraSPARC-I or II processors. High availability and fault-tolerance features are included in the X000 systems which are intended for mission-critical applications. The 3000 model is a deskside server configurable with up to six processors and 10 internal disks, while the 4000 is a rack-mount system with up to 14 processors. The 5000 is essentially a 4000 in a rack cabinet and the 6000 is a cabinet-housed data center server with up to 30 processors. In 1999, the Enterprise 3500, 4500, 5500 and 6500 models were announced. These are upgraded X000 systems, with a faster Gigaplane bus (up to 100 MHz, depending on processor clock speed, compared to 83 MHz). The 3500 also differs from the 3000 by having an additional Gigaplane slot resulting in an increased maximum of eight processors. The Enterprise X500 series were re
https://en.wikipedia.org/wiki/Automatic%20meter%20reading
Automatic meter reading (AMR) is the technology of automatically collecting consumption, diagnostic, and status data from water meter or energy metering devices (gas, electric) and transferring that data to a central database for billing, troubleshooting, and analyzing. This technology mainly saves utility providers the expense of periodic trips to each physical location to read a meter. Another advantage is that billing can be based on near real-time consumption rather than on estimates based on past or predicted consumption. This timely information coupled with analysis can help both utility providers and customers better control the use and production of electric energy, gas usage, or water consumption. AMR technologies include handheld, mobile and network technologies based on telephony platforms (wired and wireless), radio frequency (RF), or powerline transmission. Technologies Touch technology With touch-based AMR, a meter reader carries a handheld computer or data collection device with a wand or probe. The device automatically collects the readings from a meter by touching or placing the read probe in close proximity to a reading coil enclosed in the touchpad. When a button is pressed, the probe sends an interrogate signal to the touch module to collect the meter reading. The software in the device matches the serial number to one in the route database, and saves the meter reading for later download to a billing or data collection computer. Since the meter reader still has to go to the site of the meter, this is sometimes referred to as "on-site" AMR. Another form of contact reader uses a standardized infrared port to transmit data. Protocols are standardized between manufacturers by such documents as ANSI C12.18 or IEC 61107. AMR hosting AMR hosting is a back-office solution which allows a user to track their electricity, water, or gas consumption over the Internet. All data is collected in near real-time, and is stored in a database by data acquisition software. The user can view the data via a web application, and can analyze the data using various online analysis tools such as charting load profiles, analyzing tariff components, and verify their utility bill. Radio frequency network Radio frequency based AMR can take many forms. The more common ones are handheld, mobile, satellite and fixed network solutions. There are both two-way RF systems and one-way RF systems in use that use both licensed and unlicensed RF bands. In a two-way or "wake up" system, a radio signal is normally sent to an AMR meter's unique serial number, instructing its transceiver to power-up and transmit its data. The meter transceiver and the reading transceiver both send and receive radio signals. In a one-way "bubble-up" or continuous broadcast type system, the meter transmits continuously and data is sent every few seconds. This means the reading device can be a receiver only, and the meter a transmitter only. Data travels only from the meter transmitte
https://en.wikipedia.org/wiki/List%20of%20radio%20stations%20in%20Nevada
The following is a list of FCC-licensed radio stations in the U.S. state of Nevada, which can be sorted by their call signs, frequencies, cities of license, licensees, and programming formats. List of radio stations Defunct KLME References Nevada
https://en.wikipedia.org/wiki/MacsBug
MacsBug is a low-level (assembly language/machine-level) debugger for the classic Mac OS operating system. MacsBug is an acronym for Motorola Advanced Computer Systems Debugger, as opposed to Macintosh debugger (The Motorola 68000 Microprocessor is imprinted with the MACSS acronym). The original version was developed by Motorola as a general debugger for its 68000 systems. — it was ported to the Mac as a programmer's tool early in the project's development. MacsBug is invoked by hitting the Macintosh's "Programmer's Key" or, as it became later known, the "Interrupt Key" or by pressing "Command-Power". MacsBug offers many commands for disassembling, searching, and viewing data as well as control over processor registers. MacsBug is not installed by default with Mac OS, although every Macintosh since the Macintosh Plus includes a debugger in ROM known as MicroBug. Users who stumble into MacsBug by accident need only to enter G and press return to escape from MacsBug; however, MacsBug is not installed by default, requiring a system extension, so a typical user environment does not include it. However, it was occasionally installed by end users to provide very basic error recovery. As the classic Mac OS lacked memory protection, "hard crashes" where an application crash simply froze the entire system weren't uncommon. With MacsBug installed, instead of an unresponsive system, the user would be dumped into MacsBug, where they could type ES to Exit to Shell (force quit the crashed application and return to the Finder) or RB for ReBoot, which restarted the system. Such recovery efforts were often not successful, with the only alternative a hard reset. In Mac OS versions 7.5 and later, the presence of MacsBug is indicated at startup; it is present if the user sees the text Debugger installed (although, occasionally, this may indicate the presence of another piece of software loaded into the area of memory reserved for the debugger, instead). MacsBug was originally for the Motorola 68000 series of processors only. When Apple introduced the Power Macintosh in 1994, it was followed by an updated MacsBug that supported the PowerPC instruction set and architecture. The last version of MacsBug was 6.6.3, released September 14, 2000. This final version works with all of the machines released in the July–September timeframe of 2000, including the Power Mac G4 (uni- and multi-processor), Power Mac G4 Cube, the iMac family (Ruby, Indigo, Sage, Graphite, and Snow), and the iBook family (Indigo, Key Lime, and Graphite). 6.6.3 includes better support for debugging MP tasks, and fixes some serious bugs in the memory setting commands when used in PCI I/O space. It can also be used in Classic when running under Mac OS X, where it is invoked by pressing "⌘-⏏" (or "⌘-F12" on systems without an Eject key). Mac OS X allows programmers to use familiar MacsBug commands in gdb. This gdb plugin is included with the OS X Developer Tools, located in the directory /usr/lib
https://en.wikipedia.org/wiki/Register%20renaming
In computer architecture, register renaming is a technique that abstracts logical registers from physical registers. Every logical register has a set of physical registers associated with it. When a machine language instruction refers to a particular logical register, the processor transposes this name to one specific physical register on the fly. The physical registers are opaque and cannot be referenced directly but only via the canonical names. This technique is used to eliminate false data dependencies arising from the reuse of registers by successive instructions that do not have any real data dependencies between them. The elimination of these false data dependencies reveals more instruction-level parallelism in an instruction stream, which can be exploited by various and complementary techniques such as superscalar and out-of-order execution for better performance. Problem approach In a register machine, programs are composed of instructions which operate on values. The instructions must name these values in order to distinguish them from one another. A typical instruction might say: “add and and put the result in ”. In this instruction, , and are the names of storage locations. In order to have a compact instruction encoding, most processor instruction sets have a small set of special locations which can be referred to by special names: registers. For example, the x86 instruction set architecture has 8 integer registers, x86-64 has 16, many RISCs have 32, and IA-64 has 128. In smaller processors, the names of these locations correspond directly to elements of a register file. Different instructions may take different amounts of time; for example, a processor may be able to execute hundreds of instructions while a single load from the main memory is in progress. Shorter instructions executed while the load is outstanding will finish first, thus the instructions are finishing out of the original program order. Out-of-order execution has been used in most recent high-performance CPUs to achieve some of their speed gains. Consider this piece of code running on an out-of-order CPU: r1 ≔ m[1024] r1 ≔ r1 + 2 m[1032] ≔ r1 r1 ≔ m[2048] r1 ≔ r1 + 4 m[2056] ≔ r1 The instructions in the final three lines are independent of the first three instructions, but the processor cannot finish r1 ≔ m[2048] until the preceding m[1032] ≔ r1 is done (doing so otherwise would write an incorrect value). This restriction is eliminated by changing the names of some of the registers: r1 ≔ m[1024] r1 ≔ r1 + 2 m[1032] ≔ r1 r2 ≔ m[2048] r2 ≔ r2 + 4 m[2056] ≔ r2 Now the last three instructions can be executed in parallel with the first three. The program will run faster than before via obliterating any stalling due to a false data dependency. Many high-performance CPUs implement this renaming in hardware to achieve additional parallelism. On targets without appropriate data-flow detection good compilers would detect independent instruction sequenc
https://en.wikipedia.org/wiki/Dbx%20%28debugger%29
dbx is a source-level debugger found primarily on Solaris, AIX, IRIX, Tru64 UNIX, Linux and BSD operating systems. It provides symbolic debugging for programs written in C, C++, Fortran, Pascal and Java. Useful features include stepping through programs one source line or machine instruction at a time. In addition to simply viewing operation of the program, variables can be manipulated and a wide range of expressions can be evaluated and displayed. History dbx was originally developed at University of California, Berkeley, by Mark Linton during the years 1981–1984 and subsequently made its way to various vendors who had licensed BSD. Availability Besides being provided to various vendors through BSD, dbx has also found its way into other products: dbx is also available on IBM z/OS systems, in the UNIX System Services component. dbx for z/OS can debug programs written in C and C++, and can also perform machine level debugging. As of z/OS V1R5, dbx is able to debug programs using the DWARF debug format. z/OS V1R6 added support for debugging 64-bit programs. dbx is included as part of the Oracle Solaris Studio product from Oracle Corporation, and is supported on both Solaris and Linux. It supports programs compiled with the Oracle Solaris Studio compilers and GCC. GCC removed support for dbx in release 13. See also Modular Debugger (mdb) GNU Debugger References External links dbx for z/OS dbx for AIX Sun Studio Compilers and Tools for Solaris OS and Linux Sun Studio 12: Debugging a Program With dbx Debuggers Unix programming tools
https://en.wikipedia.org/wiki/GV
GV or gv may refer to: Businesses and organizations GV (company), formerly Google Ventures Aero Flight (IATA airline designator) Globovisión, a Venezuelan news network Golden Village, a movie theater chain in Singapore Grand Valley State University, in Michigan, USw General Union of Public Sector and Transport Workers, former trade union in Germany People Gianni Versace, fashion designer King George V of the United Kingdom Getúlio Vargas, former Brazilian president Gore Vidal, American author Gino Vannelli, Canadian musician Places Grass Valley, California Greenfield Village, a part of The Henry Ford, a national landmark located in Dearborn, Michigan Green Valley (disambiguation) Vehicles Gulfstream V business jet Lockheed GV, an early designation of the KC-130F Hercules aircraft Suzuki Grand Vitara, an automobile Yugo GV, an automobile Other uses GV (nerve agent) Google Voice, a mobile voice application Manx language (ISO 639 alpha-2 code) Ghostview, a front end for PostScript rasteriser software for Linux/X Grapevine viroid, a plant disease GigaVolt See also G5 (disambiguation), including a list of topics named G.V, etc.
https://en.wikipedia.org/wiki/UO
UO may stand for: Arts and entertainment Call of Duty: United Offensive, an expansion pack for the popular first-person shooter computer game, Call of Duty Ultima Online, a graphical massively multiplayer online role-playing game Underoath, an American Christian metalcore band from Tampa, Florida Urge Overkill, an alternative rock band, formed in Chicago, United States Businesses and organizations Businesses Universal Orlando, a theme park resort in Orlando, Florida Urban Outfitters, a publicly traded American company Hong Kong Express Airways (IATA code: UO), an airline based in Hong Kong Universities University of Okara, a public, coeducational university in Okara, Punjab, Pakistan University of Oregon, a public, coeducational research university in Eugene, Oregon, USA University of Osnabrück, public, coeducational university in Osnabrück, Lower Saxony, Germany University of Otago, a public, coeducational university in Dunedin, New Zealand University of Oradea, a public, coeducational university in Oradea, Romania University of Ottawa, a public, coeducational university in Canada University of Oxford, a public, coeducational university in United Kingdom Other organizations Orthodox Union, a kosher certification service which places a "U inside O" logo on food packaging Other uses Unexploded ordnance, explosive weapons that did not explode when they were employed Urate oxidase, an enzyme that catalyzes the oxidation of uric acid to 5-hydroxyisourate See also U of O (disambiguation)
https://en.wikipedia.org/wiki/TL
TL or Tl may refer to: Arts and entertainment Teens' love, Japanese erotic fiction marketed towards women Télé Liban, a Lebanese television network Turn Left (newspaper), Cornell University student publication Language Tl (digraph), a digraph representing a voiceless alveolar lateral affricate in some languages Tagalog language (ISO 639 alpha-2 code: tl) Organisations Airnorth (IATA airline code TL), an airline Public transport in the Lausanne Region, a transport company Teknisk Landsforbund, the Danish Union of Professional Technicians Team Liquid, a professional gaming and eSports team and community website Science and technology Liquidus temperature, the maximum temperature at which crystals can co-exist with the melt Teralitre (Tl or TL), a metric unit of volume or capacity Thallium, symbol Tl, a chemical element Thermoluminescence dating, in geochronology Total length in fish measurement Transmission loss (TL), in acoustics, electronics, optics, and related fields Computing .tl, East Timor's Internet country code top-level domain Transform, clipping, and lighting (T&L), in computer graphics Vehicles Acura TL, a mid-size luxury car TL11, engine for Leyland Tiger Volkswagen TL, a compact car produced in the 1960s and 1970s Other uses Nickname for Tenderloin, San Francisco East Timor (ISO 3166-1 alpha-2 country code) Turkish lira (TL), a currency Thameslink, a train operator in the UK
https://en.wikipedia.org/wiki/EY
EY, Ey, or ey may refer to: Companies Ernst & Young, a global network of financial services firms currently branded EY Eagle Air (Tanzania) (IATA code 1999–2002) Etihad Airways (IATA code since 2003) People Henri Ey, French psychiatrist Elaine Youngs, American beach volleyball player Other uses Ey, an obsolete term for egg. Ey, a Spivak pronoun used in place of "he/she" Ey, exayear, SI unit for year ey (digraph), in languages -ey (disambiguation), an English diminutive suffix East York, Ontario ("EY" in an old logo) Executive Yuan, the executive branch of the Republic of China (Taiwan)
https://en.wikipedia.org/wiki/Henrik%20Wann%20Jensen
Henrik Wann Jensen (born 1969 in Harlev, Jutland, Denmark) is a Danish computer graphics researcher. He is best known for developing the photon mapping technique as the subject of his PhD thesis, but has also done important research in simulating subsurface scattering and the sky. He was awarded an Academy Award (Academy Award for Technical Achievement) in 2004 together with Stephen R. Marschner and Pat Hanrahan for pioneering research in simulating subsurface scattering of light in translucent materials as presented in their paper "A Practical Model for Subsurface Light Transport.". The technique of simulating and scattering light on a subsurface are used by the major render engines used in the Computer Graphic industry like MentalRay or V-Ray. He is a professor emeritus at the Computer Graphics Laboratory at University of California, San Diego and has engineering degrees from the Technical University of Denmark, where he received his PhD. He is also a part of a company Luxion which is focused on simulating physically accurate light properties on a 3D object. External links https://web.archive.org/web/20040110061126/http://graphics.ucsd.edu/~henrik/ Homepage of Henrik Wann Jensen at University of California, San Diego. Interview with Henrik Wann Jensen and Thomas Teger from Luxion. 1970 births Living people Danish computer scientists Computer graphics professionals Academy Award for Technical Achievement winners
https://en.wikipedia.org/wiki/Red%20Orchestra%20%28espionage%29
The Red Orchestra (, ), as it was known in Germany, was the name given by the Abwehr Section III.F to anti-Nazi resistance workers in August 1941. It primarily referred to a loose network of resistance groups, connected through personal contacts, uniting hundreds of opponents of the Nazi regime. These included groups of friends who held discussions that were centred on Harro Schulze-Boysen, Adam Kuckhoff and Arvid Harnack in Berlin, alongside many others. They printed and distributed prohibited leaflets, posters, and stickers, hoping to incite civil disobedience. They aided Jews and resistance to escape the regime, documented the atrocities of the Nazis, and transmitted military intelligence to the Allies. Contrary to legend, the Red Orchestra was neither directed by Soviet communists nor under a single leadership. It was a network of groups and individuals, often operating independently. To date, about 400 members are known by name. The term was also used by the German Abwehr to refer to associated Soviet intelligence networks, working in Belgium, France, United Kingdom and the low countries, that were built up by Leopold Trepper on behalf of the Main Directorate of State Security (GRU). Trepper ran a series of clandestine cells for organising agents. He used the latest technology, in the form of small wireless radios, to communicate with Soviet intelligence. Although the monitoring of the radios' transmissions by the Funkabwehr would eventually lead to the organisation's destruction, the sophisticated use of the technology enabled the organisation to behave as a network, with the ability to achieve tactical surprise and deliver high-quality intelligence, including the warning of Operation Barbarossa. To this day, the German public perception of the "Red Orchestra" is characterised by the vested interest in historical revisionism of the post-war years and propaganda efforts of both sides of the Cold War. Reappraisal For a long time after World War II, only parts of the German resistance to Nazism had been known to the public within Germany and the world at large. This included the groups that took part in the 20 July plot and the White Rose resistance groups. In the 1970s there was a growing interest in the various forms of resistance and opposition. However, no organisations' history was so subject to systematic misinformation, and as little recognised, as those resistance groups centered on Arvid Harnack and Harro Schulze-Boysen. In a number of publications, the groups that these two people represented were seen as traitors and spies. An example of these was Kennwort: Direktor; die Geschichte der Roten Kapelle (Password: Director; The history of the Red Orchestra) written by Heinz Höhne who was a Der Spiegel journalist. Höhne based his book on the investigation by the Lüneburg Public Prosecutor's Office against the General Judge of the Luftwaffe and Nazi apologist Manfred Roeder who was involved in the Harnack and Schulze-Boysen cases dur
https://en.wikipedia.org/wiki/Dot-decimal%20notation
Dot-decimal notation is a presentation format for numerical data. It consists of a string of decimal numbers, using the full stop (dot) as a separation character. A common use of dot-decimal notation is in information technology where it is a method of writing numbers in octet-grouped base-10 (decimal) numbers. In computer networking, Internet Protocol Version 4 (IPv4) addresses are commonly written using the quad-dotted notation of four decimal integers, ranging from 0 to 255 each. IPv4 address In computer networking, the notation is associated with the specific use of quad-dotted notation to represent IPv4 addresses and used as a synonym for dotted-quad notation. Dot-decimal notation is a presentation format for numerical data expressed as a string of decimal numbers each separated by a full stop. For example, the hexadecimal number 0xFF000000 may be expressed in dot-decimal notation as 255.0.0.0. An IPv4 address has 32 bits. For purposes of representation, the bits may be divided into four octets written in decimal numbers, ranging from 0 to 255, concatenated as a character string with full stop delimiters between each number. This octet-grouped dotted-decimal format may more specifically be called "dotted octet" format, or a "dotted quad address". For example, the address of the loopback interface, usually assigned the host name localhost, is 127.0.0.1. It consists of the four octets, written in binary notation: 01111111, 00000000, 00000000, and 00000001. The 32-bit number is represented in hexadecimal notation as 0x7F000001. No formal specification of this textual IP address representation exists. The first mention of this format in RFC documents was in RFC 780 for the Mail Transfer Protocol published May 1981, in which the IP address was supposed to be enclosed in brackets or represented as a 32-bit decimal integer prefixed by a pound sign. A table in RFC 790 (Assigned Numbers) used the dotted decimal format, zero-padding each number to three digits. RFC 1123 (Requirements for Internet Hosts – Application and Support) of October 1989 mentions a requirement for host software to accept “IP address in dotted-decimal ("#.#.#.#") form”, although it notes “[t]his last requirement is not intended to specify the complete syntactic form for entering a dotted-decimal host number”. An IETF draft intended to define textual representation of IP addresses expired without further activity. A popular implementation of IP networking, originating in 4.2BSD, contains a function inet_aton() for converting IP addresses in character string representation to internal binary storage. In addition to the basic four-decimals format and 32-bit numbers, it also supported intermediate syntax forms of octet.24bits (e.g. 10.1234567; for Class A addresses) and octet.octet.16bits (e.g. 172.16.12345; for Class B addresses). It also allowed the numbers to be written in hexadecimal and octal representations, by prefixing them with 0x and 0, respectively. These features
https://en.wikipedia.org/wiki/Daniel%20Robbins%20%28computer%20programmer%29
Daniel Robbins is a computer programmer and consultant best known as the founder and former chief architect of the Gentoo Linux project. In 2008, he launched the Funtoo project, a free Linux distribution based on Gentoo, and he became the project's lead developer and organizer. He works in Albuquerque, New Mexico at Zenoss, and as president for Funtoo Technologies. Biography Formation of Gentoo Linux distribution During his time as a system administrator at the University of New Mexico in Albuquerque, Robbins formed his own distribution Enoch Linux, which was later renamed Gentoo Linux in 2002. However, like many other free software projects, Gentoo struggled to create a business model which would support its key developers. Robbins resigned as Chief Architect on 26 April 2004, citing considerable personal debt, and a desire to spend more time with his family, formed the Gentoo Foundation and transferred all Gentoo intellectual property to it, so that Gentoo is now run as a full community-based model. He did rejoin the project for a short time from August 2006, becoming a developer again in February 2007 and joining the amd64 team but resigned in early March 2007. There have been several high-profile criticisms of the way Gentoo has run since Robbins left, such as: "...since the resignation of Gentoo's founder and benevolent dictator from the project in 2004, the newly established Gentoo Foundation has been battling with lack of clear directions and frequent developer conflicts...", but in mid-July 2007 it emerged that Robbins was still technically the legal president of the Gentoo Foundation: ...I would like to see more fun in Gentoo, and a lot less politics, and in my apparent role as President of the Gentoo Foundation, I may have an opportunity to change things for the better. I will need to look into this more... Funtoo Linux In 2008, Robbins began to work on Funtoo, a project created to allow him to work on extending the technologies originally created for Gentoo. Microsoft Robbins' move to Microsoft, in May 2005, attracted attention within the Linux community, which has historically had a combative relationship with Microsoft. He described his role working for Bill Hilf as "...helping Microsoft to understand Open Source and community-based projects..." However, Robbins resigned less than a year later on 16 January 2006 due to frustrations that he was unable to fully utilize his technical skills in this position. RTLinux Later in 2006, he joined FSMLabs in Socorro, New Mexico, to work on RTLinux. Funtoo Technologies Daniel Robbins is also president of Funtoo Technologies, a consulting firm founded in 2006 and located in Albuquerque, New Mexico. References External links Linux Crazy podcast featuring an interview with Daniel Robbins Living people University of New Mexico staff Anglophone Quebec people American computer programmers Canadian emigrants to the United States Free software programmers Linux people People from Montr
https://en.wikipedia.org/wiki/BlooP%20and%20FlooP
and (Bounded loop and Free loop) are simple programming languages designed by Douglas Hofstadter to illustrate a point in his book Gödel, Escher, Bach. BlooP is a non-Turing-complete programming language whose main control flow structure is a bounded loop (i.e. recursion is not permitted). All programs in the language must terminate, and this language can only express primitive recursive functions. FlooP is identical to BlooP except that it supports unbounded loops; it is a Turing-complete language and can express all computable functions. For example, it can express the Ackermann function, which (not being primitive recursive) cannot be written in BlooP. Borrowing from standard terminology in mathematical logic, Hofstadter calls FlooP's unbounded loops MU-loops. Like all Turing-complete programming languages, FlooP suffers from the halting problem: programs might not terminate, and it is not possible, in general, to decide which programs do. BlooP and FlooP can be regarded as models of computation, and have sometimes been used in teaching computability. BlooP examples The only variables are OUTPUT (the return value of the procedure) and CELL(i) (an unbounded sequence of natural-number variables, indexed by constants, as in the Unlimited Register Machine). The only operators are ⇐ (assignment), + (addition), × (multiplication), < (less-than), > (greater-than) and = (equals). Each program uses only a finite number of cells, but the numbers in the cells can be arbitrarily large. Data structures such as lists or stacks can be handled by interpreting the number in a cell in specific ways, that is, by Gödel numbering the possible structures. Control flow constructs include bounded loops, conditional statements, ABORT jumps out of loops, and QUIT jumps out of blocks. BlooP does not permit recursion, unrestricted jumps, or anything else that would have the same effect as the unbounded loops of FlooP. Named procedures can be defined, but these can call only previously defined procedures. Factorial function Subtraction function This is not a built-in operation and (being defined on natural numbers) never gives a negative result (e.g. 2 − 3 := 0). Note that OUTPUT starts at 0, like all the CELLs, and therefore requires no initialization. FlooP example The example below, which implements the Ackermann function, relies on simulating a stack using Gödel numbering: that is, on previously defined numerical functions PUSH, POP, and TOP satisfying PUSH [N, S] > 0, TOP [PUSH [N, S]] = N, and POP [PUSH [N, S]] = S. Since an unbounded MU-LOOP is used, this is not a legal BlooP program. The QUIT BLOCK instructions in this case jump to the end of the block and repeat the loop, unlike the ABORT, which exits the loop. See also Machine that always halts References External links Dictionary of Programming Languages - BLooP Dictionary of Programming Languages - FLooP The Retrocomputing Museum Portland Pattern Repository: Bloop Floop and Gloop A compiler for
https://en.wikipedia.org/wiki/ColorForth
colorForth is a programming language from the Forth language's creator, Charles H. Moore, developed in the 1990s. The language combines elements of Moore's earlier Forth systems and adds color as a way of indicating how words should be interpreted. Program text is tokenized as it is edited; the compiler operates on the tokenized form, so there is less work at compile-time. An idiosyncratic programming environment, the colors simplify Forth's semantics, speed compiling, and are said to aid Moore's own poor eyesight: colorForth uses different colors in its source code (replacing some of the punctuation in standard Forth) to determine how different words are treated. colorForth was originally developed as the scripting language for Moore's own VLSI CAD program OKAD, with which he develops custom Forth processors. As the language gained utility, he rewrote his CAD program in it, spruced up the environment, and released it to the public. It has since gained a small following, spurred much debate in the Forth community, and sprung offshoots for other processors and operating environments. The language's roots are closer to the Forth machine languages Moore develops for his processors than to the mainstream standardized Forths in more widespread use. The language comes with its own tiny (63K) operating system. Practically everything is stored as source code and compiled when needed. The current colorForth environment is limited to running on Pentium grade PCs with limited support for lowest-common-denominator motherboards, AGP video, disk, and network hardware. Coloring in colorForth has semantic meaning. Red words start a definition and green words are compiled into the current definition. Thus, colorForth would be rendered in standard Forth as: : color forth ; Yellow words are executed. The transition from green, to yellow, and back again can be used while defining words, to transition between compiling words into the current definition, executing words immediately (manipulating the data stack during compilation), and back again (adding the top of the data stack to the current definition) -- in other words, precomputing a value during compilation (a functionality that other languages use macros or optimizing compilers for). Moore developed Forth in the early 1970s and created a series of implementations of the language. In the 1980s he diverged from the standardization of the language, instead continuing to evolve it. He developed a series of Forth-like languages, each extreme in its simplicity: Machine Forth, OKAD, colorForth. Moore has stated that color is only one option for displaying the language. One of Moore's papers on colorForth was printed in black and white, but used italics and other typographical conventions to present source code. References External links Concatenative programming languages Experimental programming languages Forth programming language family
https://en.wikipedia.org/wiki/Jerry%20Nachman
Jerome A. "Jerry" Nachman (February 24, 1946 – January 19, 2004) was the editor-in-chief and vice president of MSNBC cable news network., and former editor of the New York Post. Early years Nachman was born in Red Hook, Brooklyn and raised in Pittsburgh, Pennsylvania. Nachman's parents got a divorce when he was a child so he moved in with his mother and stepfather in Pittsburgh. Nachman attended, but did not graduate from, Youngstown State University for seven years, not taking a single journalism class. He then worked a number of newspaper jobs before moving into broadcasting. Biography Nachman was editor-in-chief of the New York Post from 1989 to 1992, following a stint as a police reporter and political commentator at the Post. Prior to that, he served as news director of New York's NBC station, WNBC-TV, and as Vice President of New York's CBS flagship station, WCBS-TV. He served as the general manager of WRC radio and local television stations in Washington, D.C. Nachman also wrote scripts for television programs and produced the late-night, half-hour political talk show Politically Incorrect hosted by Bill Maher during the 2000 elections. Family Nachman was married to Nancy Cook, but the marriage ended in divorce. His brother's name is Larry and he lived in Staten Island. Awards and honors Nachman was a Peabody Award, Edward R. Murrow Award, and Emmy Award winner and twice served as a Pulitzer Prize juror. Death Nachman died of cancer in 2004 at his home in Hoboken, New Jersey at the age of 57. References 1946 births 2004 deaths Emmy Award winners MSNBC people New York Post people Peabody Award winners People from Red Hook, Brooklyn Writers from Pittsburgh Deaths from cancer in New Jersey Journalists from Pennsylvania 20th-century American journalists American male journalists
https://en.wikipedia.org/wiki/Mipmap
In computer graphics, mipmaps (also MIP maps) or pyramids are pre-calculated, optimized sequences of images, each of which is a progressively lower resolution representation of the previous. The height and width of each image, or level, in the mipmap is a factor of two smaller than the previous level. Mipmaps do not have to be square. They are intended to increase rendering speed and reduce aliasing artifacts. A high-resolution mipmap image is used for high-density samples, such as for objects close to the camera; lower-resolution images are used as the object appears farther away. This is a more efficient way of downfiltering (minifying) a texture than sampling all texels in the original texture that would contribute to a screen pixel; it is faster to take a constant number of samples from the appropriately downfiltered textures. Mipmaps are widely used in 3D computer games, flight simulators, other 3D imaging systems for texture filtering, and 2D and 3D GIS software. Their use is known as mipmapping. The letters MIP in the name are an acronym of the Latin phrase multum in parvo, meaning "much in little". Since mipmaps, by definition, are pre-allocated, additional storage space is required to take advantage of them. They are also related to wavelet compression. Mipmap textures are used in 3D scenes to decrease the time required to render a scene. They also improve image quality by reducing aliasing and Moiré patterns that occur at large viewing distances, at the cost of 33% more memory per texture. Overview Mipmaps are used for: Level of detail (LOD) Improving image quality. Rendering from large textures where only small, discontiguous subsets of texels are used can easily produce Moiré patterns; Speeding up rendering times, either by reducing the number of texels sampled to render each pixel, or increasing the memory locality of the samples taken; Reducing stress on the GPU or CPU. Water surface reflections Origin Mipmapping was invented by Lance Williams in 1983 and is described in his paper Pyramidal parametrics. From the abstract: "This paper advances a 'pyramidal parametric' prefiltering and sampling geometry which minimizes aliasing effects and assures continuity within and between target images." The referenced pyramid can be imagined as the set of mipmaps stacked in front of each other. The origin of the term mipmap is an initialism of the Latin phrase multum in parvo ("much in a small space"), and map, modeled on bitmap. The term pyramids is still commonly used in a GIS context. In GIS software, pyramids are primarily used for speeding up rendering times. Mechanism Each bitmap image of the mipmap set is a downsized duplicate of the main texture, but at a certain reduced level of detail. Although the main texture would still be used when the view is sufficient to render it in full detail, the renderer will switch to a suitable mipmap image (or in fact, interpolate between the two nearest, if trilinear filtering is activated)
https://en.wikipedia.org/wiki/John%20L.%20Hennessy
John Leroy Hennessy (born September 22, 1952) is an American computer scientist, academician and businessman who serves as chairman of Alphabet Inc. Hennessy is one of the founders of MIPS Computer Systems Inc. as well as Atheros and served as the tenth President of Stanford University. Hennessy announced that he would step down in the summer of 2016. He was succeeded as president by Marc Tessier-Lavigne. Marc Andreessen called him "the godfather of Silicon Valley." Along with David Patterson, Hennessy was a recipient of the 2017 Turing Award for their work in developing the reduced instruction set computer (RISC) architecture, which is now used in 99% of new computer chips. Early life and education Hennessy was raised in Huntington, New York, as one of six children. His father was an aerospace engineer, and his mother was a teacher before raising her children. He is of Irish-Catholic descent, with some of his ancestors arriving in America during the potato famine in the 19th century. He earned his bachelor's degree in electrical engineering from Villanova University, and his master's degree and Doctor of Philosophy in computer science from Stony Brook University. Career and research Hennessy became a Stanford faculty member in 1977. In 1981, he began the MIPS project to investigate RISC processors, and in 1984, he used his sabbatical year to found MIPS Computer Systems Inc. to commercialize the technology developed by his research. In 1987, he became the Willard and Inez Kerr Bell Endowed Professor of Electrical Engineering and Computer Science. Hennessy served as director of Stanford's Computer System Laboratory (1989–93), a research center run by Stanford's Electrical Engineering and Computer Science departments. He was chair of the Department of Computer Science (1994–96) and Dean of the School of Engineering (1996–99). In 1999, Stanford President Gerhard Casper appointed Hennessy to succeed Condoleezza Rice as Provost of Stanford University. When Casper stepped down to focus on teaching in 2000, the Stanford Board of Trustees named Hennessy to succeed Casper as president. In 2008, Hennessy earned a salary of $1,091,589 ($702,771 base salary, $259,592 deferred benefits, $129,226 non-tax benefits), the 23rd highest among all American university presidents. Hennessy has served as a board member of Google (later Alphabet Inc.), Cisco Systems, Atheros Communications, and the Gordon and Betty Moore Foundation. He was elected to the American Philosophical Society in 2008. On October 14, 2010, Hennessy was presented a khata by the 14th Dalai Lama before the latter addressed Maples Pavilion. In December 2010, Hennessy coauthored an editorial with Harvard University President Drew Gilpin Faust urging the passage of the DREAM Act; the legislation did not pass the 111th United States Congress. In 2013, Hennessy became a judge for the inaugural Queen Elizabeth Prize for Engineering. He has remained on the judging panel for the subsequent awar
https://en.wikipedia.org/wiki/1763%20in%20science
The year 1763 in science and technology involved some significant events. Astronomy Publication posthumously of Nicolas Louis de Lacaille's Coelum australe stelliferum, cataloguing all his data from the southern hemisphere and including about 10,000 stars and a number of brighter star clusters and nebulae. Publication of Edward Stone's The whole doctrine of parallaxes explained and illustrated by an arithmetical and geometrical construction of the transit of Venus over the sun, June 6th, 1761. Enriched with a new and general method of determining the places where any transit of this planet, and especially that which will be June 3d, 1769, may be best observed. Mathematics December 23 – Thomas Bayes' solution to a problem of "inverse probability" is presented posthumously in his "Essay towards solving a Problem in the Doctrine of Chances" read by Richard Price to the Royal Society, containing a statement of a special case of Bayes' theorem. Medicine Edward Stone publishes his discovery of the medicinal properties of salicylic acid. Awards Copley Medal: Not awarded Births January 31 (bapt.) – John Brinkley, English astronomer (died 1835) May 12 – John Bell, Scottish surgeon (died 1820) May 16 – Louis Nicolas Vauquelin, French chemist (died 1829) August 16 – Giovanni Battista Guglielmini, Bolognese physicist (died 1817) October 27 – William Maclure, Scottish American geologist (died 1840) December 25 – Claude Chappe, French engineer (died 1805) William Higgins, Irish chemist (died 1825) Deaths March 5 – William Smellie, Scottish obstetrician (born 1697) July 11 – Peter Forsskål, Swedish naturalist (born 1732) References 18th century in science 1760s in science