source
stringlengths 31
227
| text
stringlengths 9
2k
|
|---|---|
https://en.wikipedia.org/wiki/Hartle%E2%80%93Hawking%20state
|
The Hartle–Hawking state is a proposal in theoretical physics concerning the state of the universe prior to the Planck epoch. It is named after James Hartle and Stephen Hawking.
According to the Hartle–Hawking proposal, the universe has no origin as we would understand it: before the Big Bang, which happened about 13.8 billion years ago, the universe was a singularity in both space and time. Hartle and Hawking suggest that if we could travel backwards in time towards the beginning of the universe, we would note that quite near what might have been the beginning, time gives way to space so that there is only space and no time.
Technical explanation
More precisely, the Hartle-Hawking state is a hypothetical vector in the Hilbert space of a theory of quantum gravity that describes the wave function of the universe.
It is a functional of the metric tensor defined at a (D − 1)-dimensional compact surface, the universe, where D is the spacetime dimension. The precise form of the Hartle–Hawking state is the path integral over all D-dimensional geometries that have the required induced metric on their boundary. According to the theory, time, as it is currently observed, diverged from a three-state dimension after the universe was in the age of the Planck time.
Such a wave function of the universe can be shown to satisfy, approximately, the Wheeler–DeWitt equation.
See also
Imaginary time
Multiple histories
Signature change
Notes
|
https://en.wikipedia.org/wiki/FG-7142
|
FG-7142 (ZK-31906) is a drug which acts as a partial inverse agonist at the benzodiazepine allosteric site of the GABAA receptor. It has anorectic, anxiogenic and pro-convulsant effects. It also increases release of acetylcholine and noradrenaline, and improves memory retention in animal studies.
|
https://en.wikipedia.org/wiki/Citizen%20Cyberscience%20Centre
|
The Citizen Cyberscience Centre (CCC) is an organization for volunteer computing formed as a partnership between CERN, UNITAR, and the University of Geneva.
In August 2011, a new version of the BOINC-based volunteer computing project LHC@home began simulating the high-energy collisions of protons in CERN's Large Hadron Collider (LHC), with CCC's help.
The Citizen Cyberscience Centre is currently hosted at the UNITAR offices at CERN.
|
https://en.wikipedia.org/wiki/Smart%20toy
|
A smart toy is an interactive toy which effectively has its own intelligence by virtue of on-board electronics. These enable it to learn, behave according to preset patterns, and alter its actions depending upon environmental stimuli and user input. Typically, it can adjust to the abilities of the player. A modern smart toy has electronics consisting of one or more microprocessors or microcontrollers, volatile and/or non-volatile memory, storage devices, and various forms of input–output devices. It may be networked together with other smart toys or a personal computer in order to enhance its play value or educational features. Generally, the smart toy may be controlled by software which is embedded in firmware or else loaded from an input device such as a USB flash drive, Memory Stick or CD-ROM. Smart toys frequently have extensive multimedia capabilities, and these can be utilized to produce a realistic, animated, simulated personality for the toy. Some commercial examples of smart toys are Amazing Amanda, Furby and iDog. The first smart-toy was the Mego Corporation's 2-XL robot (2XL), invented in the 1970s
Common confusions
Smart toys are frequently confused with toys for which it is claimed that children who play with them become smarter. Examples are educational toys that may or may not provide on-board intelligence features. A toy which merely contains a media player for telling the child a story should not be classified as a smart toy even if the player contains its own microprocessor. What best distinguishes a smart toy is the way the on-board intelligence is holistically integrated into the play experience in order to create simulated human-like intelligence or its facsimile.
History
Smart toys have their early roots in clockworks such as those of the eighteenth and nineteenth century cuckoo clocks, music boxes of the nineteenth, and Disney audio-animatronics of the twentieth. Perhaps the biggest early contribution is from novelty and toy maker
|
https://en.wikipedia.org/wiki/MG-RAST
|
MG-RAST is an open-source web application server that suggests automatic phylogenetic and functional analysis of metagenomes. It is also one of the biggest repositories for metagenomic data. The name is an abbreviation of Metagenomic Rapid Annotations using Subsystems Technology.
The pipeline automatically produces functional assignments to the sequences that belong to the metagenome by performing sequence comparisons to databases in both
nucleotide and amino-acid levels. The applications supply phylogenetic and functional assignments of the metagenome being analysed, as well as tools for comparing different metagenomes. It also provides a RESTful API for programmatic access.
The server was created and maintained by Argonne National Laboratory from the University of Chicago. In December 29 of 2016, the system had analyzed 60 terabase-pairs of data from more than 150,000 data sets. Among the analyzed data sets, more than 23,000 are available to the public.
Currently, the computational resources are provided by the DOE Magellan cloud at Argonne National Laboratory, Amazon EC2 Web services, and a number of traditional clusters.
Background
MG-RAST has been developed in an effort to have a free, public resource for the analysis and the storage of metagenome sequence data. The service removes one of the primary bottlenecks in metagenome analysis: the availability of high-performance computing for annotating data.
Metagenomic and metatranscriptomic studies involve the processing of large datasets, and therefore they can require computationally expensive analysis. Nowadays, scientists are able to generate such volumes of data because, in recent years, the sequencing costs have reduced dramatically. This fact has shifted the limiting factor to the computing costs:for instance, a recent study of the University of Maryland, estimated a cost of more than $5 million per terabase using their CLOVR metagenome analysis pipeline. As the size and number of sequence datasets con
|
https://en.wikipedia.org/wiki/Microbiology%20of%20decomposition
|
Microbiology of decomposition is the study of all microorganisms involved in decomposition, the chemical and physical processes during which organic matter is broken down and reduced to its original elements.
Decomposition microbiology can be divided into two fields of interest, namely the decomposition of plant materials and the decomposition of cadavers and carcasses.
The decomposition of plant materials is commonly studied in order to understand the cycling of carbon within a given environment and to understand the subsequent impacts on soil quality. Plant material decomposition is also often referred to as composting. The decomposition of cadavers and carcasses has become an important field of study within forensic taphonomy.
Decomposition microbiology of plant materials
The breakdown of vegetation is highly dependent on oxygen and moisture levels. During decomposition, microorganisms require oxygen for their respiration. If anaerobic conditions dominate the decomposition environment, microbial activity will be slow and thus decomposition will be slow. Appropriate moisture levels are required for microorganisms to proliferate and to actively decompose organic matter. In arid environments, bacteria and fungi dry out and are unable to take part in decomposition. In wet environments, anaerobic conditions will develop and decomposition can also be considerably slowed down. Decomposing microorganisms also require the appropriate plant substrates in order to achieve good levels of decomposition. This usually translates to having appropriate carbon to nitrogen ratios (C:N). The ideal composting carbon-to-nitrogen ratio is thought to be approximately 30:1. As in any microbial process, the decomposition of plant litter by microorganisms will also be dependent on temperature. For example, leaves on the ground will not undergo decomposition during the winter months where snow cover occurs as temperatures are too low to sustain microbial activities.
Decomposition mi
|
https://en.wikipedia.org/wiki/Windows%20MultiPoint%20Server
|
Windows MultiPoint Server is an operating system based on Microsoft Windows Server using Remote Desktop Services technology to host multiple simultaneous independent computing stations or terminals connected to a single computer (multiseat computing). Windows MultiPoint Server 2012 was the final release as an independent SKU and has been superseded by the MultiPoint Services role in Windows Server 2016.
Versions
Windows MultiPoint Server 2010
This version was planned in January 2010 but was released in February 2010 and is based on Windows Server 2008 R2. Its mainstream support ended on July 14, 2015, and extended support ended on July 14, 2020. Multiple stations can be added to a WMS 2010 host computer by connecting a single monitor, USB 2.0 hub, keyboard and mouse for each station. Hardware requirements for MultiPoint stations are non-proprietary, and virtually any multi-monitor video card, mouse, keyboard and monitor that is supported on Windows Server 2008 R2 can be used to build a station.
Windows MultiPoint Server 2010 is available for purchase via either OEMs or academic volume licensing. The Academic version, acquired via Academic Volume Licensing, supports domain join and no licensing restrictions on station count (however, hardware limits still apply), but requires a Windows Server 2008 R2 client access license (CAL) and a Windows MultiPoint Server 2010 CAL per station, while the non-Academic version that is acquired via OEMs is limited to 10 stations maximum and does not support domain join, but only requires a Windows MultiPoint Server 2010 CAL per station and no Windows Server 2008 R2 CALs.
Windows MultiPoint Server 2011
Windows MultiPoint Server 2011, based on Windows Server 2008 R2 SP1, was released to manufacturing on March 10, 2011. Its mainstream support ended on July 12, 2016 and extended support ended on July 13, 2021. New features in Windows MultiPoint Server 2011 include:
The ability to add connect stations and thin clients over the LAN vi
|
https://en.wikipedia.org/wiki/Thoth%20%28operating%20system%29
|
Thoth is a real-time, message passing operating system (OS) developed at the University of Waterloo in Waterloo, Ontario Canada.
History
Thoth was developed at the University of Waterloo in Waterloo, Ontario, Canada. The curriculum at Waterloo includes a Real Time Operating Systems course and an associated "Train lab", where students must develop a real-time operating system (RTOS) to control a model track with multiple trains.
In 1972, the B programming language, a derivative of BCPL, was brought to Waterloo by Stephen C. Johnson while on sabbatical from Bell Labs. A new language derived from B, named Eh, was developed at Waterloo. Thoth was written originally in Eh with some assembly language.
Initial development of Thoth occurred on a Honeywell 6050 computer. It was first run on a Data General Nova 2 in May 1976, and was next ported to a Texas Instruments TI990/10 in August 1976.
In October 1976, the University of Waterloo published Laurence S. Melen's Master's Thesis, titled "A Portable Real-Time Executive, Thoth".
Eh was later upgraded, in part with the addition of data types, and renamed Zed. Thoth was then rewritten in Zed.
One of the early principal developers of Thoth was David Cheriton. Cheriton would go on to develop the Verex kernel, and the V-System OS; both influenced by Thoth. Another early developer was Michael Malcolm, who would later found Waterloo Microsystems, Network Appliances, Inc., Blue Coat Systems, and Kaliedescape, several of whose operating systems are believed to have been derived from or influenced by Thoth.
Certain papers describe DEMOS as the inspiration for Thoth. As prior art Cheriton cited Per Brinch Hansen's RC 4000, then listed Thoth, DEMOS, and Accent together as later developments. Other influences on the development of Thoth included Multics, Data General's RTOS, Honeywell GCLS, and Unix. Later references cite Thoth as the original implementation of its particular use of synchronous message passing and multiprocess pro
|
https://en.wikipedia.org/wiki/Ford%20Instrument%20Company
|
The Ford Instrument Company was a U.S. corporation known for being the primary supplier of fire control Rangekeepers and analog computers for the United States Navy before and during World War II.
A personal blog, Doug Coward's Analog History Museum, includes a page with details for the Ford Instrument Company Computer Mark I that was used after 1939 on WW II naval guns up to 5 inch and anti-aircraft guns. This page has a background stating that the Ford Instrument Company is a subsidiary of Sperry Rand, indicating that the displayed page was supplied by Sperry while operating as Sperry Rand, 1955 and 1978.
|
https://en.wikipedia.org/wiki/Fifth%20metacarpal%20bone
|
The fifth metacarpal bone (metacarpal bone of the little finger or pinky finger) is the most medial and second-shortest of the metacarpal bones.
Surfaces
It presents on its base one facet on its superior surface, which is concavo-convex and articulates with the hamate, and one on its radial side, which articulates with the fourth metacarpal.
On its ulnar side is a prominent tubercle for the insertion of the tendon of the extensor carpi ulnaris muscle.
The dorsal surface of the body is divided by an oblique ridge, which extends from near the ulnar side of the base to the radial side of the head. The lateral part of this surface serves for the attachment of the fourth Interosseus dorsalis; the medial part is smooth, triangular, and covered by the extensor tendons of the little finger.
The palmar surface is similarly divided: Its lateral side (facing the fourth metacarpal) provides the origin for the third palmar interosseus, its medial side contains the insertion of opponens digiti quinti.
Clinical significance
A fracture of the fourth and/or fifth metacarpal bones transverse neck secondary due to axial loading is known as a boxer's fracture. The fifth metacarpal bone is the most common bone to be injured when throwing a punch.
Ossification
The ossification process begins in the shaft during prenatal life, and in the head between 11th and 37th months.
Additional images
See also
Metacarpus
First metacarpal bone
Second metacarpal bone
Third metacarpal bone
Fourth metacarpal bone
|
https://en.wikipedia.org/wiki/Segmentation%20Rules%20eXchange
|
Segmentation Rules eXchange or SRX is an XML-based standard that was maintained by Localization Industry Standards Association, until it became insolvent in 2011, and then by the Globalization and Localization Association (GALA).
SRX provides a common way to describe how to segment text for translation and other language-related processes. It was created when it was realized that TMX was less useful than expected in certain instances due to differences in how tools segment text. SRX is intended to enhance the TMX standard so that translation memory (TM) data that is exchanged between applications can be used more effectively. Having the segmentation rules available that were used when a TM was created increases the usefulness of the TM data.
Implementation difficulties
SRX make use of the ICU Regular Expression syntax, but not all programming languages support all ICU expressions, making implementing SRX in some languages difficult or impossible. Java is an example of this.
Version history
SRX version 1.0 was officially accepted as an OSCAR standard in April 2004.
SRX version 2.0 was officially accepted as an OSCAR standard in April 2008.
SRX forms part of the Open Architecture for XML Authoring and Localization (OAXAL) reference architecture.
|
https://en.wikipedia.org/wiki/Signal%20peptide
|
A signal peptide (sometimes referred to as signal sequence, targeting signal, localization signal, localization sequence, transit peptide, leader sequence or leader peptide) is a short peptide (usually 16-30 amino acids long) present at the N-terminus (or occasionally nonclassically at the C-terminus or internally) of most newly synthesized proteins that are destined toward the secretory pathway.
These proteins include those that reside either inside certain organelles (the endoplasmic reticulum, Golgi or endosomes), secreted from the cell, or inserted into most cellular membranes. Although most type I membrane-bound proteins have signal peptides, the majority of type II and multi-spanning membrane-bound proteins are targeted to the secretory pathway by their first transmembrane domain, which biochemically resembles a signal sequence except that it is not cleaved. They are a kind of target peptide.
Function (translocation)
Signal peptides function to prompt a cell to translocate the protein, usually to the cellular membrane. In prokaryotes, signal peptides direct the newly synthesized protein to the SecYEG protein-conducting channel, which is present in the plasma membrane. A homologous system exists in eukaryotes, where the signal peptide directs the newly synthesized protein to the Sec61 channel, which shares structural and sequence homology with SecYEG, but is present in the endoplasmic reticulum. Both the SecYEG and Sec61 channels are commonly referred to as the translocon, and transit through this channel is known as translocation. While secreted proteins are threaded through the channel, transmembrane domains may diffuse across a lateral gate in the translocon to partition into the surrounding membrane.
Structure
The core of the signal peptide contains a long stretch of hydrophobic amino acids (about 5–16 residues long) that has a tendency to form a single alpha-helix and is also referred to as the "h-region". In addition, many signal peptides begin with a
|
https://en.wikipedia.org/wiki/Optical%20sine%20theorem
|
In optics, the optical sine theorem states that the products of the index, height, and sine of the slope angle of a ray in object space and its corresponding ray in image space are equal. That is:
External links
http://physics.tamuk.edu/~suson/html/4323/aberatn.html#Optical%20Sine
Sine theorem
Physics theorems
|
https://en.wikipedia.org/wiki/Bile%20acid%3Asodium%20symporter
|
This family of proteins are found both in prokaryotes and eukaryotes. In mammals, they are transmembrane proteins with functions in the liver and in the intestine. They are members of the solute carrier family of cotransporter genes which include SLC10A1 and SLC10A2.
SLC10A1 encodes the sodium-taurocholate cotransporting polypeptide (NTCP) expressed in the liver and found on the basolateral membranes. It is involved in the uptake of all types of bile acids from portal blood plasma, a process mediated by the co-transport of Na+. It is also capable of transporting other solutes and is necessary for the entry of hepatitis B and hepatitis D viruses into the hepatocyte.
SLC10A2 encodes the apical sodium bile acid transporter (ASBT) expressed in the small intestine with highest concentrations in the ileum. It is found on the brush border membrane and is also known as the ileal bile acid transporter (IBAT). It is responsible for the initial uptake of bile acids from the intestine as part of the enterohepatic circulation. Inhibition of the intestinal bile acid:sodium cotransporter by elobixibat is under development for the treatment of constipation and irritable bowel syndrome.
|
https://en.wikipedia.org/wiki/Regulation%20to%20Prevent%20and%20Combat%20Child%20Sexual%20Abuse
|
The Regulation to Prevent and Combat Child Sexual Abuse (Child Sexual Abuse Regulation, or CSAR) is a European Union regulation proposed by the European Commissioner for Home Affairs Ylva Johansson on 11 May 2022. The stated aim of the legislation is to prevent child sexual abuse online through the implementation of a number of measures, including the establishment of a framework that would make the detection and reporting of child sexual abuse material (CSAM) by digital platforms a legal requirement within the European Union.
Support for the proposal
Supporters of the regulation include dozens of campaign groups, activists and MEPs, along with departments within the European Commission and European Parliament themselves.
The European Commission's Migration and Home Affairs department argues that voluntary actions by online service providers to detect online child sexual abuse are insufficient. They emphasize that some service providers are less involved in combating such abuse, leading to gaps where abuse can go undetected. Moreover, they highlight that companies can change their policies, making it challenging for authorities to prevent and combat child sexual abuse effectively. The EU currently relies on other countries, primarily the United States, to launch investigations into abuse occurring within the EU, resulting in delays and inefficiencies.
Several bodies within the EU claim the establishment of a centralized organization, the EU Centre on Child Sexual Abuse, would create a single point of contact for receiving reports of child sexual abuse. It is claimed this centralization would streamline the process by eliminating the need to send reports to multiple entities and would enable more efficient allocation of resources for investigation and response.
Proponents also argue for the need to improve the transparency of the process of finding, reporting, and removing online child sexual abuse material. They claim that there is currently limited oversight o
|
https://en.wikipedia.org/wiki/Probably%20approximately%20correct%20learning
|
In computational learning theory, probably approximately correct (PAC) learning is a framework for mathematical analysis of machine learning. It was proposed in 1984 by Leslie Valiant.
In this framework, the learner receives samples and must select a generalization function (called the hypothesis) from a certain class of possible functions. The goal is that, with high probability (the "probably" part), the selected function will have low generalization error (the "approximately correct" part). The learner must be able to learn the concept given any arbitrary approximation ratio, probability of success, or distribution of the samples.
The model was later extended to treat noise (misclassified samples).
An important innovation of the PAC framework is the introduction of computational complexity theory concepts to machine learning. In particular, the learner is expected to find efficient functions (time and space requirements bounded to a polynomial of the example size), and the learner itself must implement an efficient procedure (requiring an example count bounded to a polynomial of the concept size, modified by the approximation and likelihood bounds).
Definitions and terminology
In order to give the definition for something that is PAC-learnable, we first have to introduce some terminology.
For the following definitions, two examples will be used. The first is the problem of character recognition given an array of bits encoding a binary-valued image. The other example is the problem of finding an interval that will correctly classify points within the interval as positive and the points outside of the range as negative.
Let be a set called the instance space or the encoding of all the samples. In the character recognition problem, the instance space is . In the interval problem the instance space, , is the set of all bounded intervals in , where denotes the set of all real numbers.
A concept is a subset . One concept is the set of all patterns of bi
|
https://en.wikipedia.org/wiki/Master%20of%20Mathematics
|
A Master of Mathematics (or MMath) degree is a specific advanced integrated Master's degree for courses in the field of mathematics.
United Kingdom
In the United Kingdom, the MMath is the internationally recognized standard qualification after a four-year course in mathematics at a university.
The MMath programme was set up by most leading universities after the Neumann Report in 1992. It is classed as a level 7 qualification in the Frameworks of Higher Education Qualifications of UK Degree-Awarding Bodies. The UCAS course codes for the MMath degrees start at G100 upwards, most courses taking the codes G101 - G104.
Universities which offer MMath degrees include:
Aberystwyth University
University of Bath
University of Bristol (MSci)
Brunel University
University of Birmingham (MSci)
Cardiff University
University of Cambridge
City University London
University of Central Lancashire
University of Dundee
University of Durham
University of East Anglia
University of Edinburgh
University of Essex
University of Exeter
University of Glasgow
Heriot-Watt University
University of Hull
University of Keele
University of Kent
Lancaster University
University of Leeds
University of Leicester
University of Lincoln
University of Liverpool
Liverpool Hope University
Loughborough University
University of Manchester
Manchester Metropolitan University
Middlesex University (from 2014)
Newcastle University
Northumbria University
University of Nottingham
Nottingham Trent University
Open University (until 2007)
Oxford Brookes University
University of Oxford
University of Plymouth
University of Portsmouth
University of Reading
University of St Andrews
University of Sheffield
University of Southampton
University of Strathclyde
University of Surrey
University of Sussex
Swansea University
University of Warwick
University of York
Notes
Canada
In Canada, the MMath is a graduate degree offered by the University of Waterloo. The length of the MMath degree program is typically between one and two ye
|
https://en.wikipedia.org/wiki/Colipase
|
Colipase, abbreviated CLPS, is a protein co-enzyme required for optimal enzyme activity of pancreatic lipase. It is secreted by the pancreas in an inactive form, procolipase, which is activated in the intestinal lumen by trypsin. Its function is to prevent the inhibitory effect of bile salts on the lipase-catalyzed intraduodenal hydrolysis of dietary long-chain triglycerides.
In humans, the colipase protein is encoded by the CLPS gene.
Protein domain
Colipase is also a family of evolutionarily related proteins.
Colipase is a small protein cofactor needed by pancreatic lipase for efficient dietary lipid hydrolysis. Efficient absorption of dietary fats is dependent on the action of pancreatic triglyceride lipase. Colipase binds to the C-terminal, non-catalytic domain of lipase, thereby stabilising an active conformation and considerably increasing the hydrophobicity of its binding site. Structural studies of the complex and of colipase alone have revealed the functionality of its architecture.
Colipase is a small protein (12K) with five conserved disulphide bonds. Structural analogies have been recognised between a developmental protein (Dickkopf), the pancreatic lipase C-terminal domain, the N-terminal domains of lipoxygenases and the C-terminal domain of alpha-toxin. These non-catalytic domains in the latter enzymes are important for interaction with membrane. It has not been established if these domains are also involved in eventual protein cofactor binding as is the case for pancreatic lipase.
See also
Enterostatin
|
https://en.wikipedia.org/wiki/Weak%20order%20unit
|
In mathematics, specifically in order theory and functional analysis, an element of a vector lattice is called a weak order unit in if and also for all
Examples
If is a separable Fréchet topological vector lattice then the set of weak order units is dense in the positive cone of
See also
Citations
|
https://en.wikipedia.org/wiki/Concurrent%20lines
|
In geometry, lines in a plane or higher-dimensional space are concurrent if they intersect at a single point. They are in contrast to parallel lines.
Examples
Triangles
In a triangle, four basic types of sets of concurrent lines are altitudes, angle bisectors, medians, and perpendicular bisectors:
A triangle's altitudes run from each vertex and meet the opposite side at a right angle. The point where the three altitudes meet is the orthocenter.
Angle bisectors are rays running from each vertex of the triangle and bisecting the associated angle. They all meet at the incenter.
Medians connect each vertex of a triangle to the midpoint of the opposite side. The three medians meet at the centroid.
Perpendicular bisectors are lines running out of the midpoints of each side of a triangle at 90 degree angles. The three perpendicular bisectors meet at the circumcenter.
Other sets of lines associated with a triangle are concurrent as well. For example:
Any median (which is necessarily a bisector of the triangle's area) is concurrent with two other area bisectors each of which is parallel to a side.
A cleaver of a triangle is a line segment that bisects the perimeter of the triangle and has one endpoint at the midpoint of one of the three sides. The three cleavers concur at the center of the Spieker circle, which is the incircle of the medial triangle.
A splitter of a triangle is a line segment having one endpoint at one of the three vertices of the triangle and bisecting the perimeter. The three splitters concur at the Nagel point of the triangle.
Any line through a triangle that splits both the triangle's area and its perimeter in half goes through the triangle's incenter, and each triangle has one, two, or three of these lines. Thus if there are three of them, they concur at the incenter.
The Tarry point of a triangle is the point of concurrency of the lines through the vertices of the triangle perpendicular to the corresponding sides of the triangle's first
|
https://en.wikipedia.org/wiki/MacMach
|
MacMach is a computer operating system from the early 1990s, developed by Carnegie Mellon University. Architecturally, it consists of Berkeley Software Distribution (BSD) 4.3 code running on the Mach microkernel, with the Apple Macintosh System 7 running experimentally as a Mach task. The entire system runs on Macintoshes based on the Motorola 68000 series (68k) family of microprocessors. Its license requires the user to have an AT&T UNIX license, and includes Apple, Inc.'s restriction against further redistribution.
See also
MkLinux
MachTen
A/UX
NeXTSTEP
|
https://en.wikipedia.org/wiki/Timeline%20of%20physical%20chemistry
|
The timeline of physical chemistry lists the sequence of physical chemistry theories and discoveries in chronological order.
Timeline details
See also
Timeline of physics
Timeline of atomic and subatomic physics
Timeline of chemistry
|
https://en.wikipedia.org/wiki/Klaus%20Samelson
|
Klaus Samelson (21 December 1918 – 25 May 1980) was a German mathematician, physicist, and computer pioneer in the area of programming language translation and push-pop stack algorithms for sequential formula translation on computers.
Early life
He was born in Strasbourg, Alsace-Lorraine, and he lived in Breslau in his early childhood years. His elder brother was the mathematician Hans Samelson. Due to political circumstances, he waited until 1946 to study mathematics and physics at the Ludwig Maximilian University of Munich in Munich.
Career
After graduating, he worked briefly as a high school teacher before he returned to university. In 1951, he completed his Doctor of Philosophy (Ph.D.) in physics with Friedrich Bopp (Fritz) with a dissertation on a quantum mechanics problem posed by Arnold Sommerfeld related to unipolar induction.
Samelson became interested in numerical analysis, and when Hans Piloty, an electrical engineer, and Robert Sauer, a professor of mathematics, began working together, he joined and got involved in early computers as a research associate in the Mathematical Institute of the Technical University of Munich.
This changed his scientific career. His first publications came from Sauer's interests dealing with supersonic speed flow and precision problems of digital computations for numerical calculations of eigenvalues.
Soon after, Samelson's strong influence began on the development of Computer Science and Informatics as a new scientific discipline. With Friedrich L. Bauer, who also had Fritz Bopp as his Ph.D. advisor, he studied the structure of programming languages to develop efficient algorithms for their translation and implementation. This research led to bracketed structures and it became clear to Samelson that this principle should govern the translation of programming languages and the run-time system with stack models and block structure. It was a fundamental breakthrough in how computer systems are modeled and designed.
Pilo
|
https://en.wikipedia.org/wiki/Temporal%20logic
|
In logic, temporal logic is any system of rules and symbolism for representing, and reasoning about, propositions qualified in terms of time (for example, "I am always hungry", "I will eventually be hungry", or "I will be hungry until I eat something"). It is sometimes also used to refer to tense logic, a modal logic-based system of temporal logic introduced by Arthur Prior in the late 1950s, with important contributions by Hans Kamp. It has been further developed by computer scientists, notably Amir Pnueli, and logicians.
Temporal logic has found an important application in formal verification, where it is used to state requirements of hardware or software systems. For instance, one may wish to say that whenever a request is made, access to a resource is eventually granted, but it is never granted to two requestors simultaneously. Such a statement can conveniently be expressed in a temporal logic.
Motivation
Consider the statement "I am hungry". Though its meaning is constant in time, the statement's truth value can vary in time. Sometimes it is true, and sometimes false, but never simultaneously true and false. In a temporal logic, a statement can have a truth value that varies in time—in contrast with an atemporal logic, which applies only to statements whose truth values are constant in time. This treatment of truth-value over time differentiates temporal logic from computational verb logic.
Temporal logic always has the ability to reason about a timeline. So-called "linear-time" logics are restricted to this type of reasoning. Branching-time logics, however, can reason about multiple timelines. This permits in particular treatment of environments that may act unpredictably.
To continue the example, in a branching-time logic we may state that "there is a possibility that I will stay hungry forever", and that "there is a possibility that eventually I am no longer hungry". If we do not know whether or not I will ever be fed, these statements can both be true.
|
https://en.wikipedia.org/wiki/Supermicro
|
Super Micro Computer, Inc., dba Supermicro, is an information technology company based in San Jose, California. It has manufacturing operations in the Silicon Valley, the Netherlands and at its Science and Technology Park in Taiwan. Founded on November 1, 1993, Supermicro is one of the largest producers of high-performance and high-efficiency servers. It also provides server management softwares, and storage systems for various markets, including enterprise data centers, cloud computing, artificial intelligence, 5G and edge computing.
Supermicro's stock trades under the ticker symbol SMCI on the Nasdaq exchange. Its fiscal year 2023 revenues were $7.1 billion and employs over 5,000 globally.
History
In 1993, Supermicro began as a 5 person operation run by Charles Liang alongside his wife and company treasurer, Chiu-Chu Liu, known as Sara. Prior to founding Supermicro, Liang earned a B.S. in Electrical Engineering from the National Taiwan University of Science and Technology and a M.S. in Electrical Engineering from the University of Texas at Arlington. Liang holds several patents for server technology and was previously the president and chief design engineer of Micro Center Computer, a motherboard design and manufacturing company, from July 1991 to August 1993.
International expansion, initial public offering
In 1996, the company opened a manufacturing subsidiary, Ablecom, in Taiwan, which is run by Charles's brother, Steve Liang and Bill Liang. Charles Liang and his wife own close to 31 percent of Ablecom, while Steve Liang and other members of the family own close to 50 percent. In 1998, Supermicro opened a subsidiary in the Netherlands.
In 2006, Supermicro pleaded guilty to a felony charge and paid a $150,000 fine due to a violation of a United States embargo against the sale of computer systems to Iran. In a plea agreement, it was acknowledged that Supermicro became aware of the investigation in February 2004 and set up an export-control program that same
|
https://en.wikipedia.org/wiki/Polymer%20capacitor
|
A polymer capacitor, or more accurately a polymer electrolytic capacitor, is an electrolytic capacitor (e-cap) with a solid conductive polymer electrolyte. There are four different types:
Polymer tantalum electrolytic capacitor (Polymer Ta-e-cap)
Polymer aluminium electrolytic capacitor (Polymer Al-e-cap)
Hybrid polymer capacitor (Hybrid polymer Al-e-cap)
Polymer niobium electrolytic capacitors
Polymer Ta-e-caps are available in rectangular surface-mounted device (SMD) chip style. Polymer Al-e-caps and hybrid polymer Al-e-caps are available in rectangular surface-mounted device (SMD) chip style, in cylindrical SMDs (V-chips) style or as radial leaded versions (single-ended).
Polymer electrolytic capacitors are characterized by particularly low internal equivalent series resistances (ESR) and high ripple current ratings. Their electrical parameters have similar temperature dependence, reliability and service life compared to solid tantalum capacitors, but have a much better temperature dependence and a considerably longer service life than aluminium electrolytic capacitors with non-solid electrolytes. In general polymer e-caps have a higher leakage current rating than the other solid or non-solid electrolytic capacitors.
Polymer electrolytic capacitors are also available in a hybrid construction. The hybrid polymer aluminium electrolytic capacitors combine a solid polymer electrolyte with a liquid electrolyte. These types are characterized by low ESR values but have low leakage currents and are insensitive to transients, however they have a temperature-dependent service life similar to non-solid e-caps.
Polymer electrolytic capacitors are mainly used in power supplies of integrated electronic circuits as buffer, bypass and decoupling capacitors, especially in devices with flat or compact design. Thus they compete with MLCC capacitors, but offer higher capacitance values than MLCC, and they display no microphonic effect (such as class 2 and 3 ceramic capacito
|
https://en.wikipedia.org/wiki/ISO-IR-153
|
ISO-IR-153 (ST SEV 358-88) is an 8-bit character set that covers the Russian and Bulgarian alphabets. Unlike the KOI encodings, this encoding lists the Cyrillic letters in their correct traditional order. This has become the basis for ISO/IEC 8859-5 and the Cyrillic Unicode block.
Standards and Naming
The name ISO-IR-153 refers to this set's number in the ISO-IR registry, and marks it as a set which may be used within ISO/IEC 2022.
ISO-IR-153 is a subset of ISO/IEC 8859-5 (synchronised with ECMA-113 since 1988). The ISO-IR-153 documentation cites ST SEV 358-88 as the source standard. While it also cites the earlier GOST 19768-74 (which defines KOI-8 and was conformed to by the first version of ECMA-113, i.e. ISO-IR-111), it does not follow the KOI-8 layout (rather using a close modification of the letter layout from the Main code page) so this appears to be in error. The ISO-IR-153 encoding was intended to replace GOST 19768-74, and is sometimes referred to as GOST-19768-87. This confusion has led to a common misconception that ISO-8859-5 was defined in or based on GOST 19768-74.
Notwithstanding the extents of their accuracy, the IANA lists , and as labels which may be used for the ISO-IR-153 encoding on the Internet, with reference to RFC 1345, which assigns it those labels.
GOST R 34.303-92 includes the ISO-IR-153 code page and dubs it KOI-8 V1 (in addition to using KOI-8 N1 and KOI-8 N2 for two Alternative code page/Code page 866 variants).
Character set
The following table shows the ISO-IR-153 encoding. Each character is shown with its equivalent Unicode code point.
The encoding closely resembles the letter subset of the Cyrillic part of the Main code page, apart from the relocation of the uppercase Ё from 0xF0 to 0xA1. ISO-8859-5 is a superset.
See also
ISO-IR-111
|
https://en.wikipedia.org/wiki/International%20Human%20Epigenome%20Consortium
|
The International Human Epigenome Consortium (IHEC) is a scientific organization, founded in 2010, that helps to coordinate global efforts in the field of Epigenomics. The initial goal was to generate at least 1,000 reference (baseline) human epigenomes from different types of normal and disease-related human cell types.
Structure and funding
IHEC's operations are funded by its full members (national and regional scientific funding agencies), and staffed largely on a volunteer basis by scientists and other experts from participating funding agencies and epigenome mapping projects.
Current IHEC Member Countries
Canadian Institutes for Health Research (Canada)
European Commission (EU)
Federal Ministry of Education and Research, Project Management Agency within the German Aerospace Centre (Germany)
Genome Institute of Singapore (Singapore)
Hong Kong Epigenomics Project (Hong Kong)
Japan Agency for Medical Research and Development (Japan)
National Institute of Health (South Korea)
National Institutes of Health: Roadmap Epigenomics Program and Encyclopedia of DNA Elements (ENCODE) project (USA)
In addition, countries and agencies supportive of IHEC goals are organizations that have not yet made a full financial contribution to the project, but whose members provide time and expertise:
European Institute of Oncology, FIRC Institute of Molecular Oncology Foundation, Italian Institute of Technology, Centre for Genomic Science (Italy)
National Agency of Research (France)
National Health and Medical Research Council (Australia)
UK Funders Alliance: Medical Research Council, Biotechnology and Biological Sciences Research Council, Cancer Research UK, and Wellcome Trust (UK)
Oversight of IHEC is provided by an executive committee, whose members are nominated by Full Member organizations. This committee works closely with an International Scientific Steering Committee, whose members are the scientific leaders of participating projects and other leaders in the
|
https://en.wikipedia.org/wiki/MicrOmega-IR
|
MicrOmega-IR is an infrared hyperspectral microscope that is part of the science payload on board the European Rosalind Franklin rover, tasked to search for biosignatures on Mars. The rover is planned to land on Mars in the mid- or late 2020s. MicrOmega-IR will analyse in situ the powder material derived from crushed samples collected by the rover's core drill.
Development
The MicrOmega mnemonic is derived from its French name Micro observatoire pour la mineralogie, l'eau, les glaces et l'activité; IR stands for infrared. It was developed by France's Institut d'Astrophysique Spatiale at the CNRS. France has also flown MicrOmega on other missions such as the 2011 Fobos-Grunt and the Hayabusa2 MASCOT mobile lander currently exploring asteroid Ryugu. France is also developing a variant called MacrOmega Near-IR Spectrometer for the Martian Moons Exploration (MMX) lander, a Japanese sample-return mission to Mars' moon Phobos.
The Principal Investigator of the MicrOmega-IR for the Rosalind Franklin rover is Jean-Pierre Bibring, a French astronomer and planetary scientist at the Institut d'Astrophysique Spatiale. Co-PIs are astrobiologists Frances Westall and Nicolas Thomas.
MicrOmega was developed by a consortium including:
IAS (Orsay, France)
LESIA (Meudon, France)
CBM (Orléans, France),
University Of Bern (Bern, Switzerland)
Russian Space Research Institute (IKI) (Moscow, Russia)
Overview
MicrOmega-IR is a visible and infrared hyperspectral microscope that is designed to characterize the texture and composition of crushed samples presented to the instrument. Its objective is to study mineral grain assemblages in detail to try to unravel their geological origin, structure and composition, including potential organics. These data will be vital for interpreting past and present geological processes and environments on Mars. Because MicrOmega-IR is an imaging instrument, it can also be used to identify grains that are particularly interesting, and assign them a
|
https://en.wikipedia.org/wiki/Amiskwia
|
Amiskwia is a genus of soft-bodied animals known from fossils of the Middle Cambrian Lagerstätten both in the Burgess Shale in British Columbia, Canada and the Maotianshan shales of Yunnan Province, China. It is interpreted as a member of the clade Gnathifera sensu lato.
Etymology
The scientific name Amiskwia sagittiformis derives from the Cree amiskwi, "beavertai", a name of various objects in Yoho National Park, and from the Latin sagitta ("arrow") and formis ("shape"), in reference to the general appearance of the animal. "Sinica", of A. sinica, refers to that species' origin from China.
Description
Known specimens of Amiskwia vary in length from and in width from . The body was somewhat flattened. The head had a pair of tentacles that emerged from the midline of the head. The tentacles had a relatively thick base and tapered to a point. Along the sides of the trunk were a pair of lateral fins, which were around one third of the total body length. The trunk terminated with a flat, rounded caudal fin. The gut was straight, and ran from the mouth to the anus, which was located on the underside of the body near the caudal fin. Within the mouth is a pair of semi-circular structures, described as "jaws" each bearing 8-10 conical spikes, which increased in size away from the midline of the structure. Two other structures, dubbed the "dorsal plate" and "ventral plate", are also present in the mouth.
Ecology
Amiskwia was likely a freely swimming (nektonic) organism that was either a predator or a scavenger.
History of research
Amiskwia was originally categorized by paleontologist Charles Walcott. Walcott thought he saw three buccal spines in the fossils, and therefore categorized Amiskwia as a chaetognath worm (arrow worm). However, Amiskwia appears to lack the characteristic grasping spines and teeth of other Burgess fossil arrow worms. Later scientists suggested an affinity with the nemerteans (ribbon worms), but the evidence for this was somewhat inadequate.
|
https://en.wikipedia.org/wiki/Eastport%20Historic%20District
|
The Eastport Historic District encompasses the late 19th-century commercial center of the city of Eastport, Maine. Set on a five-block stretch of Water Street, this area was almost completely redeveloped after a major fire in 1886, and many of its buildings are the work of a single architect, Henry Black. The district was listed on the National Register of Historic Places in 1982, and enlarged slightly in 2016.
Description and history
The city of Eastport is located on Moose Island at the far southeastern tip of Maine, with the Canadian province of New Brunswick just across Passamaquoddy Bay. In the mid-19th century, Eastport served as a major regional commercial center, and developed a significant sardine-packing industry in the 1870s. On October 14, 1886, a fire devastated the city center, destroying its central business district and its commercial wharves. In the following year, the city rebuilt its downtown.
The rebuilt area extends along Water Street, the city's east-facing waterfront area, between Sullivan and Key Streets. There are 26 buildings that were built in 1887, with two more added by 1893, and one in 1928–29. Fourteen of the 1887 buildings were designed (or have been attributed to) Henry Black, an architect who had been based for some years in Saint John, New Brunswick. Black arrived in Eastport in February 1887, and was apparently the only architect directly involved in the rebuilding effort. Buildings typifying his work include the Eastport Savings Bank at 41 Water Street, and the Ferris & Son building at 4 Boynton Street, both Italianate in style, one in brick and the other in wood. Other architect-designed buildings in the district include the Peavey Memorial Library (1893, Rotch and Tilden), and the Frontier National Bank at 30 Water Street (1882, Charles Kimball). Most of the district's buildings are brick Italianate structures, although some are wooden, and the library is Romanesque in style. The 2016 enlargement added the former
|
https://en.wikipedia.org/wiki/Council%20for%20Responsible%20Genetics
|
The Council for Responsible Genetics (CRG) was a nonprofit NGO with a focus on biotechnology.
History
The Council for Responsible Genetics was founded in 1983 in Cambridge, Massachusetts.
An early voice concerned about the social and ethical implications of modern genetic technologies, CRG organized a 1985 Congressional Briefing and a 1986 panel of the American Association for the Advancement of Science, both focusing on the potential dangers of genetically engineered biological weapons. Francis Boyle was asked to draft legislation setting limits on the use of genetic engineering, leading to the Biological Weapons Anti-Terrorism Act of 1989.
CRG was the first organization to advance a comprehensive, scientifically based position against human germline engineering. It was also the first to compile documented cases of genetic discrimination, laying the intellectual groundwork for the Genetic Information Nondiscrimination Act of 2008 (GINA).
The organization created both a Genetic Bill of Rights and a Citizen's Guide to Genetically Modified Food. Also notable are CRG's support for the "Safe Seeds Campaign" (for avoiding gene flow from genetically engineered to non-GE seed) and the organization of a US conference on Forensic DNA Databanks and Racial Disparities in the Criminal Justice System. In 2010 CRG led a successful campaign to roll back a controversial student genetic testing program at the University of California, Berkeley. In 2011, CRG led a campaign to successfully enact [CalGINA] in California, which extended genetic privacy and nondiscrimination protections to life, disability and long term care insurance, mortgages, lending and other areas.
CRG issued five anthologies of commentaries:
Rights and Liberties in the Biotech Age edited by Sheldon Krimsky and Peter Shorett
Race and the Genetic Revolution: Science, Myth and Culture
Genetic Explanations: Sense and Nonsense edited by Krimsky and Jeremy Gruber
Biotechnology in our Lives edited by Krim
|
https://en.wikipedia.org/wiki/Microsoft%20Bob
|
Microsoft Bob was a Microsoft software product intended to provide a more user-friendly interface for the Windows 3.1x, Windows 95 and Windows NT operating systems, supplanting the Windows Program Manager. The program was released on March 11, 1995 and discontinued in early 1996. Microsoft Bob presented screens showing a house, with rooms that the user could visit containing familiar objects corresponding to computer applications, such as a desk with pen and paper and a checkbook. Clicking on the pen and paper would open the system's word processor.
A cartoon dog named Rover and other cartoon characters provided guidance using speech balloons.
Upon release, Microsoft Bob was criticized in the media and did not gain wide acceptance with users, which resulted in its discontinuation. Its legacy would be observed in future Microsoft products, notably the use of virtual assistants. The Rover character later reappeared as a Windows XP search companion.
History
Microsoft Bob was released in March 1995 (before Windows 95 was released), although it had been widely publicized under the codename Utopia. The project leader for Bob was Karen Fries, a Microsoft researcher. The design was based on research by professors Clifford Nass and Byron Reeves of Stanford University. Melinda Gates, wife of Bill Gates, was the marketing manager for the product. Microsoft originally purchased the domain name bob.com from Boston-area techie Bob Antia, but later traded it to Bob Kerstein for the windows2000.com domain name.
Applications
Microsoft Bob includes various office-suite programs such as a finance application and a word processor. The user interface was designed to simplify the navigational experience for novice computer users.
Similar to early graphical shells like Jane, the main interface is portrayed as the inside of a house, with different rooms corresponding to common real-world room styles such as a kitchen or family room. Each room contains decorations and furniture, as we
|
https://en.wikipedia.org/wiki/Wired%20Equivalent%20Privacy
|
Wired Equivalent Privacy (WEP) was a severely flawed security algorithm for 802.11 wireless networks. Introduced as part of the original IEEE 802.11 standard ratified in 1997, its intention was to provide data confidentiality comparable to that of a traditional wired network. WEP, recognizable by its key of 10 or 26 hexadecimal digits (40 or 104 bits), was at one time widely used, and was often the first security choice presented to users by router configuration tools.
Subsequent to a 2001 disclosure of a severe design flaw in the algorithm, WEP was never again secure in practice. In the vast majority of cases, Wi-Fi hardware devices relying on WEP security could not be upgraded to secure operation. Some of the design flaws were addressed in WEP2, but WEP2 also proved insecure, and another generation of hardware could not be upgraded to secure operation.
In 2003, the Wi-Fi Alliance announced that WEP and WEP2 had been superseded by Wi-Fi Protected Access (WPA). In 2004, with the ratification of the full 802.11i standard (i.e. WPA2), the IEEE declared that both WEP-40 and WEP-104 have been deprecated. WPA retained some design characteristics of WEP that remained problematic.
WEP was the only encryption protocol available to 802.11a and 802.11b devices built before the WPA standard, which was available for 802.11g devices. However, some 802.11b devices were later provided with firmware or software updates to enable WPA, and newer devices had it built in.
History
WEP was ratified as a Wi-Fi security standard in 1999. The first versions of WEP were not particularly strong, even for the time they were released, due to U.S. restrictions on the export of various cryptographic technologies. These restrictions led to manufacturers restricting their devices to only 64-bit encryption. When the restrictions were lifted, the encryption was increased to 128 bits. Despite the introduction of 256-bit WEP, 128-bit remains one of the most common implementations.
Encryption det
|
https://en.wikipedia.org/wiki/TUM%20School%20of%20Natural%20Sciences
|
The TUM School of Natural Sciences (NAT) is a school of the Technical University of Munich, established in 2022 by the merger of various former departments. As of 2022, it is structured into the Department of Biosciences, the Department of Chemistry, and the Department of Physics. The school is located at the Garching campus.
Department of Chemistry
History
Chairs
As of 2020, the department consists of 24 chairs and institutes:
Analytical Chemistry and Water Chemistry
Inorganic Chemistry
Inorganic and Organometallic Chemistry
Inorganic Chemistry with Focus on New Materials
Construction Chemistry
Construction Chemicals
Biochemistry
Biomolecular NMR-Spectroscopy
Biophysical Chemistry
Biotechnology
Food Chemistry
Macromolecular Chemistry
Medicinal and Bioinorganic Chemistry
Organic Chemistry
Pharmaceutical Radiochemistry
Physical Chemistry
Silicon Chemistry
Synthetic Biotechnology
Technical Electrochemistry
Chemical Technology
Theoretical Chemistry
Department of Physics
History
Physics was one of the founding disciplines of the Polytechnische Schule München in 1868, with the establishment of the Physikalisches Cabinet, later called the Physikalisches Institut. In 1902, the Laboratorium für Technische Physik (technical physics) was founded, spearheaded by Carl von Linde. In 1943, another institute, the Institut für Theoretische Physik (theoretical physics) was founded. In 1965, the three physics institutes were finally combined into the Department of Physics, as it exists today.
The TUM Department of Physics is notable for its operation of research reactors on the Garching campus, the from 1957 to 2000 and the newer Forschungsreaktor München II since 2004.
Research groups
As of 2020, the main research areas the TUM Department of Physics are biophysics, nuclei, particles, astrophysics, and condensed matter. The following research groups currently exist:
Applied Quantum Field Theory
Biomedical Physics
Biomolecular Nano-Technology
Cell
|
https://en.wikipedia.org/wiki/Mixed%20complementarity%20problem
|
Mixed Complementarity Problem (MCP) is a problem formulation in mathematical programming. Many well-known problem types are special cases of, or may be reduced to MCP. It is a generalization of nonlinear complementarity problem (NCP).
Definition
The mixed complementarity problem is defined by a mapping , lower values and upper values .
The solution of the MCP is a vector such that for each index one of the following alternatives holds:
;
;
.
Another definition for MCP is: it is a variational inequality on the parallelepiped .
See also
Complementarity theory
|
https://en.wikipedia.org/wiki/Lipoprotein%20receptor-related%20protein
|
Lipoprotein receptor-related proteins, low density lipoprotein receptor-related proteins (HGNC) or prolow-density lipoprotein receptor-related protein (UniProt), abbreviated LRP, are a group of proteins.
They include:
LRP1
LRP1B
LRP2 (megalin)
LRP3
LRP4
LRP5
LRP6
LRP8, apolipoprotein e receptor
LRP10
LRP11
LRP12
See also
LRPAP1 (low density lipoprotein receptor-related protein associated protein 1)
Lipoprotein receptor-related proteins are co-receptors for Wnt signaling.
Human proteins
|
https://en.wikipedia.org/wiki/The%20Windup%20Girl
|
The Windup Girl is a biopunk science fiction novel by American writer Paolo Bacigalupi. It was his debut novel and was published by Night Shade Books on September 1, 2009. The novel is set in a future Thailand and covers a number of contemporary issues such as global warming and biotechnology.
The Windup Girl was named as the ninth best fiction book of 2009 by TIME magazine. It won the 2010 Nebula Award and the 2010 Hugo Award (tied with The City & the City by China Miéville), both for best novel. The book also won the 2010 Campbell Memorial Award, the 2010 Compton Crook Award and the 2010 Locus Award for best first novel.
Setting
The Windup Girl is set in 23rd-century Thailand. Global warming has raised the levels of world's oceans, carbon fuel sources have become depleted, and manually wound springs are used as energy storage devices. Biotechnology is dominant and megacorporations (called calorie companies) like AgriGen, PurCal and RedStar control food production through 'genehacked' seeds, and use bioterrorism, private armies and economic hitmen to create markets for their products. Frequent catastrophes, such as deadly and widespread plagues and illness, caused by genetically modified crops and mutant pests, ravage entire populations. The natural genetic seed stock of the world's plants has been almost completely supplanted by those that are genetically engineered to be sterile, forcing farmers to buy new seeds from the calorie companies every season.
Thailand is an exception. It maintains its own reserve of genetically viable seeds, fights off engineered plagues and other bioterrorism, and keeps its borders firmly closed against the calorie companies and other foreign biological imports. The capital city of Bangkok is below sea level and is protected from flooding by levees and pumps. The current monarch of Thailand is a child queen who is essentially a figurehead; the three most powerful people in Thailand are the Somdet Chaopraya (regent for the child que
|
https://en.wikipedia.org/wiki/Diversity%20panel
|
A diversity panel is a collection of genetic material or individual samples taken from a diverse population of a certain species. The idea is to illustrate the genetic and phenotypic diversity of the species.
Diversity panels exist for human populations, mouse and other organisms.
Researchers in the area of genetics often use diversity panels in order to reveal genotypes that are linked to certain traits, such as in QTL mapping with Genome-wide association study.
Those study analyze the Gene–environment interaction underneath simple and complex traits.
Examples
Human Genome Diversity Project
The Hybrid Mouse Diversity Panel
Maize NAM population (Nested association mapping)
Arabidopsis thaliana 1001 Genome project
See also
Genetics
Biodiversity
Evolution
|
https://en.wikipedia.org/wiki/Numerology%20%28Ismailism%29
|
Numerology is an element of Ismailis belief; the idea that numbers have religious meanings. The number seven plays a general role in the theology of the Ismā'īliyya, including mystical speculations that there are seven heavens, seven continents, seven orifices in the skull, seven days in a week, seven prophets, and so forth.
Position of the Imam
Old Ismaili doctrine holds that divine revelation had been given in six periods (daur) entrusted to six prophets, also called Natiq (Speaker), who were commissioned to preach a religious law to their respective communities.
For instance, Nasir Khusraw argues that the world of religion was created in six cycles, corresponding to the six days of the week. The seventh day, corresponding to the Sabbath, is the cycle in which the world comes out of darkness and ignorance and “into the light of her Lord” (Quran 39:69), and the people who “laboured in fulfilment of (the Prophets’) command” are rewarded.
While the Natiq was concerned with the rites and outward shape of religion and life, the inner meaning was entrusted to a Wasi (Representative), who would know the secret meaning of all rites and rules and would reveal them to a small circles of initiates.
The Natiq and Wasi are in turn succeeded by a line of seven Imams, who would guard what they received. The seventh and last Imam in any period would then be the Natiq of the next period. The last Imam of the sixth period however would not bring about a new religion or law but would abrogate the law and introduce din Adama al-awwal ("the original religion of Adam"), as practised by Adam and the Angels in paradise before the fall. This would be without cult or law but would consist in all creatures praising the creator and recognizing his unity. This final stage was called Qiyamah.
|
https://en.wikipedia.org/wiki/SMILES%20arbitrary%20target%20specification
|
SMILES arbitrary target specification (SMARTS) is a language for specifying substructural patterns in molecules. The SMARTS line notation is expressive and allows extremely precise and transparent substructural specification and atom typing.
SMARTS is related to the SMILES line notation that is used to encode molecular structures and like SMILES was originally developed by David Weininger and colleagues at Daylight Chemical Information Systems. The most comprehensive descriptions of the SMARTS language can be found in Daylight's SMARTS theory manual, tutorial and examples. OpenEye Scientific Software has developed their own version of SMARTS which differs from the original Daylight version in how the R descriptor (see cyclicity below) is defined.
SMARTS syntax
Atomic properties
Atoms can be specified by symbol or atomic number. Aliphatic carbon is matched by [C], aromatic carbon by [c] and any carbon by [#6] or [C,c]. The wild card symbols *, A and a match any atom, any aliphatic atom and any aromatic atom respectively. Implicit hydrogens are considered to be a characteristic of atoms and the SMARTS for an amino group can be written as [NH2]. Charge is specified by the descriptors + and - as exemplified by the SMARTS [nH+] (protonated aromatic nitrogen atom) and [O-]C(=O)c (deprotonated aromatic carboxylic acid).
Bonds
A number of bond types can be specified: - (single), = (double), # (triple), : (aromatic) and ~ (any).
Connectivity
The X and D descriptors are used to specify the total numbers of connections (including implicit hydrogen atoms) and connections to explicit atoms. Thus [CX4] matches carbon atoms with bonds to any four other atoms while [CD4] matches quaternary carbon.
Cyclicity
As originally defined by Daylight, the R descriptor is used to specify ring membership. In the Daylight model for cyclic systems, the smallest set of smallest rings (SSSR) is used as a basis for ring membership. For example, indole is perceived as a 5-membered r
|
https://en.wikipedia.org/wiki/Kir2.6
|
{{DISPLAYTITLE:Kir2.6}}
The Kir2.6 also known as inward rectifier potassium channel 18 is a protein that in humans is encoded by the KCNJ18 gene. Kir2.6 is an inward-rectifier potassium ion channel.
Function
Inwardly rectifying potassium channels, such as Kir2.6, maintain resting membrane potential in excitable cells and aid in repolarization of cells following depolarization. Kir2.6 is primarily expressed in skeletal muscle and is transcriptionally regulated by thyroid hormone.
Clinical significance
Mutations in this gene have been linked to thyrotoxic periodic paralysis.
|
https://en.wikipedia.org/wiki/Doujin%20soft
|
is software created by Japanese hobbyists or hobbyist groups (referred to as "circles"), more for fun than for profit. The term includes digital , which are essentially the Japanese equivalent of independent video games or fangames (the term "doujin game" also includes things like doujin-made board games and card games).
Doujin soft is considered part of doujin katsudou, for which it accounts for 5% of all doujin works altogether (as of 2015). Doujin soft began with microcomputers in Japan, and spread to platforms such as the MSX and X68000. Since the 1990's, however, they have primarily been made for Microsoft Windows.
Most doujin soft sales occur at doujin conventions such as Comiket, with several that deal with doujin soft or doujin games exclusively such as Freedom Game (which further only allows games distributed for free) and Digital Games Expo. There is also a growing number of specialized internet sites that sell doujin soft. Additionally, more doujin games have been sold as downloads on consoles and PC stores such as Steam in recent years, through publishers such as Mediascape picking them up.
Digital doujin games
Doujin video games, like doujin soft, began with microcomputers in Japan, such as the PC-98 and PC-88, and spread to platforms such as the MSX, FM Towns and X68000. From the 90's to 00's however, they were primarily exclusive to Microsoft Windows. In recent years, more doujin games have been released on mobile platforms and home consoles, as well as other operating systems like macOS and Linux. Though doujin games used to primarily be for home computers, more doujin games have been made available on gaming consoles in recent years. There are also doujin groups that develop software for retro consoles such as the Game Boy and Game Gear.
Like fangames, doujin games frequently use characters from existing games, anime, or manga ("niji sousaku"). These unauthorized uses of characters are generally ignored and accepted by the copyright holders, an
|
https://en.wikipedia.org/wiki/Symptomatic%20treatment
|
Symptomatic treatment, supportive care, supportive therapy, or palliative treatment is any medical therapy of a disease that only affects its symptoms, not the underlying cause. It is usually aimed at reducing the signs and symptoms for the comfort and well-being of the patient, but it also may be useful in reducing organic consequences and sequelae of these signs and symptoms of the disease. In many diseases, even in those whose etiologies are known (e.g., most viral diseases, such as influenza and Rift Valley fever), symptomatic treatment is the only treatment available so far. For more detail, see supportive therapy. For conditions like cancer, arthritis, neuropathy, tendinopathy, and injury, it can be useful to distinguish treatments that are supportive/palliative and cannot alter the natural history of the disease (disease modifying treatments).
Examples
Examples of symptomatic treatments:
Analgesics, to reduce pain
Anti-inflammatory agents, for inflammation caused by arthritis
Antitussives, for cough
Antihistaminics (also known as antihistamines), for allergy
Antipyretics, for fever
Enemas for constipation
Treatments that reduce unwanted side effects from drugs
Uses
When the etiology (the cause, set of causes, or manner of causation of a disease or condition) for the disease is known, then specific treatment may be instituted, but it is generally associated with symptomatic treatment, as well.
Symptomatic treatment is not always recommended, and in fact, it may be dangerous, because it may mask the presence of an underlying etiology which will then be forgotten or treated with great delay. Examples:
Low-grade fever for 15 days or more is sometimes the only symptom of bacteremia by staphylococcus bacteria. Suppressing it by symptomatic treatment will hide the disease from effective diagnosis and treatment with antibiotics. The consequence may be severe (rheumatic fever, nephritis, endocarditis, etc.)
Chronic headache may be caused simply by a cons
|
https://en.wikipedia.org/wiki/Biospeleology
|
Biospeleology, also known as cave biology, is a branch of biology dedicated to the study of organisms that live in caves and are collectively referred to as troglofauna.
Biospeleology as a science
History
The first documented mention of a cave organisms dates back to 1689, with the documentation of the olm, a cave salamander. Discovered in a cave in Slovenia, in the region of Carniola, it was mistaken for a baby dragon and was recorded by Johann Weikhard von Valvasor in his work The Glory of the Duchy of Carniola.
The first formal study on cave organisms was conducted on the blind cave beetle. Found in 1831 by Luka Čeč, an assistant to the lamplighter, when exploring the newly discovered inner portions of the Postojna cave system in southwestern Slovenia. The specimen was turned over to Ferdinand J. Schmidt, who described it in the paper Illyrisches Blatt (1832). He named it Leptodirus Hochenwartii after the donor, and also gave it the Slovene name drobnovratnik and the German name Enghalskäfer, both meaning "slender-necked (beetle)". The article represents the first formal description of a cave animal (the olm, described in 1768, wasn't recognized as a cave animal at the time).
Subsequent research by Schmidt revealed further previously unknown cave inhabitants, which aroused considerable interest among natural historians. For this reason, the discovery of L. hochenwartii (along with the olm) is considered as the starting point of biospeleology as a scientific discipline. Biospeleology was formalized as a science in 1907 by Emil Racoviță with his seminal work Essai sur les problèmes biospéologiques ("Essay on biospeleological problems”).
Subdivisions
Organisms Categories
Cave organisms fall into three basic classes:
Troglobite
Troglobites are obligatory cavernicoles, specialized for cave life. Some can leave caves for short periods, and may complete parts of their life cycles above ground, but cannot live their entire lives outside of a cave environment. Examp
|
https://en.wikipedia.org/wiki/Reta%20Vortaro
|
Reta Vortaro ("Internet Dictionary", often known by the Esperanto short form ReVo) is a general-purpose multilingual Esperanto dictionary for the Internet. Each of the dictionary's headwords is defined in Esperanto, along with additional information, such as example sentences, to help distinguish the subtle shades of meaning that each particular word form may have.
Headwords also have translation equivalents in various national languages. Over 60 percent of the headwords have French, German, Russian Hungarian and/or English translations; over half the words have Dutch, Slovak, Czech, Polish and/or Belarusian translations; and over 30 percent of the words have Spanish, Portuguese and/or Catalan translations.
Italian, Bulgarian, Romanian, Persian, Indonesian, Swedish, Breton and several other languages are also represented, though with smaller numbers of Esperanto headwords. In addition to indices of headwords linking to the various national languages represented, Reta Vortaro also has multiple thematic indexes, a thesaurus and a bibliography of sources consulted.
History
The project was initiated at the end of 1997. The original purpose had been to create an electronic version of the massive dictionary known as Plena Ilustrita Vortaro de Esperanto (1970 and 2002 editions), but due to a lack of interest on the part of the PIV editors, at least partly due to concerns at the time over harming sales of the printed volume — since alleviated, as PIV now has both a reprinted paper edition and a free online version — the Reta Vortaro project changed its focus to creating a separate Internet-use dictionary. One of its major sources is the Plena Vortaro de Esperanto (1930, with the 1953 supplement of Gaston Waringhien).
Crowd-sourced editing
Contributions are added through an e-mail editing server, for which anyone can register. Lexical items are formulated in XML. Editorial discussion takes place via the mailing list. The content and the tools are licensed through the GNU
|
https://en.wikipedia.org/wiki/Markov%27s%20inequality
|
In probability theory, Markov's inequality gives an upper bound for the probability that a non-negative function of a random variable is greater than or equal to some positive constant. It is named after the Russian mathematician Andrey Markov, although it appeared earlier in the work of Pafnuty Chebyshev (Markov's teacher), and many sources, especially in analysis, refer to it as Chebyshev's inequality (sometimes, calling it the first Chebyshev inequality, while referring to Chebyshev's inequality as the second Chebyshev inequality) or Bienaymé's inequality.
Markov's inequality (and other similar inequalities) relate probabilities to expectations, and provide (frequently loose but still useful) bounds for the cumulative distribution function of a random variable.
Statement
If is a nonnegative random variable and , then the probability
that is at least is at most the expectation of divided by :
Let (where ); then we can rewrite the previous inequality as
In the language of measure theory, Markov's inequality states that if is a measure space, is a measurable extended real-valued function, and , then
This measure-theoretic definition is sometimes referred to as Chebyshev's inequality.
Extended version for nondecreasing functions
If is a nondecreasing nonnegative function, is a (not necessarily nonnegative) random variable, and , then
An immediate corollary, using higher moments of supported on values larger than 0, is
Proofs
We separate the case in which the measure space is a probability space from the more general case because the probability case is more accessible for the general reader.
Intuition
where is larger than or equal to 0 as the random variable is non-negative and is larger than or equal to because the conditional expectation only takes into account of values larger than or equal to which r.v. can take.
Hence intuitively , which directly leads to .
Probability-theoretic proof
Method 1:
From the definition of expectation:
H
|
https://en.wikipedia.org/wiki/BLC1
|
BLC1 (Breakthrough Listen Candidate 1) was a candidate SETI radio signal detected and observed during April and May 2019, and first reported on 18 December 2020, spatially coincident with the direction of the Solar System's closest star, Proxima Centauri.
Signal
The apparent shift in its frequency, consistent with the Doppler effect, was suggested to be inconsistent with what would be caused by the movement of Proxima b, a planet of Proxima Centauri. The Doppler shift in the signal was the opposite of what would be expected from the Earth's spin, in that the signal was observed to increase in frequency rather than decrease. Although the signal was detected by Parkes Radio Telescope during observations of Proxima Centauri, due to the beam angle of Parkes Radio telescope, the signal would be more accurately described as having come from within a circle roughly 16 arcminutes (approximately 1/4 of a degree, half the angular width of Earth's moon) in angular diameter, containing Proxima Centauri, so the signal could have originated elsewhere in the Alpha Centauri system. The signal had a frequency of 982.002 MHz.
The radio signal was detected during 30 hours of observations conducted by Breakthrough Listen through the Parkes Observatory in Australia in April and May 2019. As of December 2020, follow-up observations had failed to detect the signal again, a step necessary to confirm that the signal was a technosignature.
Origin
A paper by other astronomers released 10 days before the news report about BLC1 reports the detection of "a bright, long-duration optical flare, accompanied by a series of intense, coherent radio bursts" from Proxima Centauri also in April and May 2019. Their finding has not been put in direct relation to the BLC1 signal by scientists or media outlets as of January 2021 but implies that planets around Proxima Centauri and other red dwarfs are uninhabitable for humans and other currently known organisms.
In February 2021, a new study proposed th
|
https://en.wikipedia.org/wiki/Dose%20fractionation
|
Dose fractionation effects are utilised in the treatment of cancer with radiation therapy. When the total dose of radiation is divided into several, smaller doses over a period of several days, there are fewer toxic effects on healthy cells. This maximizes the effect of radiation on cancer and minimizes the negative side effects. A typical fractionation scheme divides the dose into 30 units delivered every weekday over six weeks.
Background
Experiments in radiation biology have found that as the absorbed dose of radiation increases, the number of cells which survive decreases. They have also found that if the radiation is fractionated into smaller doses, with one or more rest periods in between, fewer cells die. This is because of self-repair mechanisms which repair the damage to DNA and other biomolecules such as proteins. These mechanisms can be over expressed in cancer cells, so caution should be used in using results for a cancer cell line to make predictions for healthy cells if the cancer cell line is known to be resistant to cytotoxic drugs such as cisplatin. The DNA self repair processes in some organisms is exceptionally good; for instance, the bacterium Deinococcus radiodurans can tolerate a 15 000 Gy (1.5 MRad) dose.
In the graph to the right, called a cell survival curve, the dose vs. surviving fraction have been drawn for a hypothetical group of cells with and without a rest time for the cells to recover. Other than the recovery time partway through the irradiation, the cells would have been treated identically.
The human body contains many types of cells, and the human can be killed by the loss of a single type of cell in a vital organ. For many short-term radiation deaths due to what is commonly known as radiation sickness (3 to 30 days after exposure), it is the loss of bone marrow cells (which produce blood cells), and the loss of other cells in the wall of the intestines, that is fatal.
Radiation fractionation as cancer treatment
Fractionatio
|
https://en.wikipedia.org/wiki/Active%20appearance%20model
|
An active appearance model (AAM) is a computer vision algorithm for matching a statistical model of object shape and appearance to a new image. They are built during a training phase. A set of images, together with coordinates of landmarks that appear in all of the images, is provided to the training supervisor.
The model was first introduced by Edwards, Cootes and Taylor in the context of face analysis at the 3rd International Conference on Face and Gesture Recognition, 1998. Cootes, Edwards and Taylor further described the approach as a general method in computer vision at the European Conference on Computer Vision in the same year. The approach is widely used for matching and tracking faces and for medical image interpretation.
The algorithm uses the difference between the current estimate of appearance and the target image to drive an optimization process.
By taking advantage of the least squares techniques, it can match to new images very swiftly.
It is related to the active shape model (ASM). One disadvantage of ASM is that it only uses shape constraints (together with some information about the image structure near the landmarks), and does not take advantage of all the available information – the texture across the target object. This can be modelled using an AAM.
|
https://en.wikipedia.org/wiki/List%20of%20router%20and%20firewall%20distributions
|
This is a list of router and firewall distributions, which are operating systems designed for use as routers and/or firewalls.
See also
List of router firmware projects
Comparison of router software projects
|
https://en.wikipedia.org/wiki/Dioptra
|
A dioptra (sometimes also named dioptre or diopter, from ) is a classical astronomical and surveying instrument, dating from the 3rd century BC. The dioptra was a sighting tube or, alternatively, a rod with a sight at both ends, attached to a stand. If fitted with protractors, it could be used to measure angles.
Use
Greek astronomers used the dioptra to measure the positions of stars; both Euclid and Geminus refer to the dioptra in their astronomical works.
It continued in use as an effective surveying tool. Adapted to surveying, the dioptra is similar to the theodolite, or surveyor's transit, which dates to the sixteenth century. It is a more accurate version of the groma.
There is some speculation that it may have been used to build the Eupalinian aqueduct. Called "one of the greatest engineering achievements of ancient times," it is a tunnel 1,036 meters (4,000 ft) long, "excavated through Mount Kastro on the Greek island of Samos, in the 6th century BCE" during the reign of Polycrates. Scholars disagree, however, whether the dioptra was available that early.
An entire book about the construction and surveying usage of the dioptra is credited to Hero of Alexandria (also known as Heron; a brief description of the book is available online; see Lahanas link, below). Hero was "one of history’s most ingenious engineers and applied mathematicians."
The dioptra was used extensively on aqueduct building projects. Screw turns on several different parts of the instrument made it easy to calibrate for very precise measurements
The dioptra was replaced as a surveying instrument by the theodolite.
See also
Alidade
|
https://en.wikipedia.org/wiki/Tumor%20necrosis%20factor%20superfamily
|
The tumor necrosis factor (TNF) superfamily is a protein superfamily of type II transmembrane proteins containing TNF homology domain and forming trimers. Members of this superfamily can be released from the cell membrane by extracellular proteolytic cleavage and function as a cytokine. These proteins are expressed predominantly by immune cells and they regulate diverse cell functions, including immune response and inflammation, but also proliferation, differentiation, apoptosis and embryogenesis.
The superfamily contains 19 members that bind to 29 members of TNF receptor superfamily. An occurrence of orthologs in invertebrates hints at ancient origin of this superfamily in evolution.
The PROSITE pattern of this superfamily is located in a beta sheet in the central section of the protein that is conserved across all members.
Members
There are 19 family members, numerically classified as TNFSF#, where # denotes the member number, sometimes followed by a letter.
|
https://en.wikipedia.org/wiki/Asymmetric%20cointegration
|
In economics, testing for an asymmetric cointegration relationship among variables implies distinguishing the positive and the negative effects of the error obtained from the cointegration regression. In order to do that, economists usually use the asymmetric cointegration framework proposed by Enders and Siklos in 2001. Asymmetric cointegration comes from the analysis of multivariate combinations arising from the decomposition of the series into positive and negative values of its cumulative sums; see Lardic and Mignon (2008, p. 484).
According to Cook (2006), testing for a potential asymmetric cointegration extends the analysis in further directions to allow for a possibility that other standard cointegration tests may fail to detect an underlying cointegration relationship and therefore, the threshold autoregressive and momentum-threshold cointegration tests are applied to investigate the hypothesis that the cointegration relationship between banking sector development indicators and remittances is of an asymmetric form which splits the speed of adjustments into two parts based on the direction of the equilibrium error.
|
https://en.wikipedia.org/wiki/Phalaenopsis%20%C3%97%20rolfeana
|
Phalaenopsis × rolfeana is a species of orchid native to the Philippines. It is a hybrid of Phalaenopsis equestris and Phalaenopsis sanderiana.
Etymology
The specific epithet rolfeana honours the English botanist Robert Allen Rolfe, who specialised on the study of orchids. He had first described this taxon as Phalaenopsis × hebe, which however was a Nomen illegitimum and thus not accepted.
|
https://en.wikipedia.org/wiki/Solomon%20Mikhlin
|
Solomon Grigor'evich Mikhlin (, real name Zalman Girshevich Mikhlin) (the family name is also transliterated as Mihlin or Michlin) (23 April 1908 – 29 August 1990) was a Soviet mathematician of who worked in the fields of linear elasticity, singular integrals and numerical analysis: he is best known for the introduction of the symbol of a singular integral operator, which eventually led to the foundation and development of the theory of pseudodifferential operators.
Biography
He was born in , Rechytsa District, Minsk Governorate (in present-day Belarus) on 23 April 1908; himself states in his resume that his father was a merchant, but this assertion could be untrue since, in that period, people sometimes lied on the profession of parents in order to overcome political limitations in the access to higher education. According to a different version, his father was a melamed, at a primary religious school (kheder), and that the family was of modest means: according to the same source, Zalman was the youngest of five children. His first wife was Victoria Isaevna Libina: Mikhlin's book is dedicated to her memory. She died of peritonitis in 1961 during a boat trip on Volga. In 1940 they adopted a son, Grigory Zalmanovich Mikhlin, who later emigrated to Haifa, Israel. His second wife was Eugenia Yakovlevna Rubinova, born in 1918, who was his companion for the rest of his life.
Education and academic career
He graduated from a secondary school in Gomel in 1923 and entered the State Herzen Pedagogical Institute in 1925. In 1927 he was transferred to the Department of Mathematics and Mechanics of Leningrad State University as a second year student, passing all the exams of the first year without attending lectures. Among his university professors there were Nikolai Maximovich Günther and Vladimir Ivanovich Smirnov. The latter became his master thesis supervisor: the topic of the thesis was the convergence of double series, and was defended in 1929. Sergei Lvovich Sobo
|
https://en.wikipedia.org/wiki/Interrupt%20request
|
In a computer, an interrupt request (or IRQ) is a hardware signal sent to the processor that temporarily stops a running program and allows a special program, an interrupt handler, to run instead. Hardware interrupts are used to handle events such as receiving data from a modem or network card, key presses, or mouse movements.
Interrupt lines are often identified by an index with the format of IRQ followed by a number. For example, on the Intel 8259 family of programmable interrupt controllers (PICs) there are eight interrupt inputs commonly referred to as IRQ0 through IRQ7. In x86 based computer systems that use two of these PICs, the combined set of lines are referred to as IRQ0 through IRQ15. Technically these lines are named IR0 through IR7, and the lines on the ISA bus to which they were historically attached are named IRQ0 through IRQ15 (although historically as the number of hardware devices increased, the total possible number of interrupts was increased by means of cascading requests, by making one of the IRQ numbers cascade to another set or sets of numbered IRQs, handled by one or more subsequent controllers).
Newer x86 systems integrate an Advanced Programmable Interrupt Controller (APIC) that conforms to the Intel APIC Architecture. These APICs support a programming interface for up to 255 physical hardware IRQ lines per APIC, with a typical system implementing support for only around 24 total hardware lines.
During the early years of personal computing, IRQ management was often of user concern. With the introduction of plug and play devices this has been alleviated through automatic configuration.
Overview
When working with personal computer hardware, installing and removing devices, the system relies on interrupt requests. There are default settings that are configured in the system BIOS and recognized by the operating system. These default settings can be altered by advanced users. Modern plug and play technology has not only reduced the need f
|
https://en.wikipedia.org/wiki/Felgo
|
Felgo (previously V-Play Engine until February 2019) is a cross-platform development tool, based on the Qt framework. It can be used to create mobile apps or games. Felgo apps and games are supported on iOS, Android, Windows Phone, embedded devices and desktop devices. Felgo developers use QML, JavaScript and C++ to create mobile apps and games.
Apps and games built with Felgo use a single code base and work across multiple platforms and screen resolutions. Felgo was founded in 2012 and is based in Vienna, Austria.
Architecture
Felgo is based on the Qt (software) cross-platform development framework that provides abstraction layers for timers, threads, storage, networking and UI rendering on different platforms. Felgo uses Qt as its core and offers components and plugins on top of it, which further simplify the development of apps and games.
Engine Features
Qt Creator IDE
The Qt Creator IDE supports editing QML and JavaScript code with context-sensitive help, code completion of Felgo components, navigation between components and more.
It includes a QML debugger and profiler for debugging custom components and JavaScript functions. It can inspect and change property values and QML code at runtime and is able to measure the time of element creation and binding evaluations.
Declarative Language Features
Felgo apps and games are written in JavaScript and QML, a declarative language that features property bindings, state machines or fluid animations of any property.
Resolution & Aspect Ratio Independence
Felgo is built to handle all possible aspect ratios and resolutions of modern mobile devices. Developers create their game for a logical scene using a content scaling approach.
Native Dialogs and Functionality
Felgo apps and games provide a native look and feel on all platforms. Felgo has abstracted components for displaying native input dialogs and alert boxes without the need of a single native code line.
Felgo Game Network
The Felgo Game Network is a mobi
|
https://en.wikipedia.org/wiki/SGI%20Prism
|
The Silicon Graphics Prism is a series of visualization computer systems developed and manufactured by Silicon Graphics (SGI). Released in April 2005, the Prism's basic system architecture is based on the Altix 3000 servers, but with graphics hardware.
The Prism uses the Linux operating system and the OpenGL software library.
Three models of the SGI Prism are Power, Team and Extreme levels.
The Power level supports two to eight Itanium 2 processors, up to 96 GB of memory and two to four graphics pipelines.
The Team level supports 8 to 16 Itanium 2 processors, up to 192 GB of memory and four to eight graphics pipelines.
The Extreme level supports 16 to 256 Itanium 2 processors, up to 3 TB of memory and 4 to 16 graphics pipelines.
The graphics pipelines for the Prism are ATI FireGL cards based on either the R350 or R420 GPUs.
|
https://en.wikipedia.org/wiki/Cohen%27s%20h
|
In statistics, Cohen's h, popularized by Jacob Cohen, is a measure of distance between two proportions or probabilities. Cohen's h has several related uses:
It can be used to describe the difference between two proportions as "small", "medium", or "large".
It can be used to determine if the difference between two proportions is "meaningful".
It can be used in calculating the sample size for a future study.
When measuring differences between proportions, Cohen's h can be used in conjunction with hypothesis testing. A "statistically significant" difference between two proportions is understood to mean that, given the data, it is likely that there is a difference in the population proportions. However, this difference might be too small to be meaningful—the statistically significant result does not tell us the size of the difference. Cohen's h, on the other hand, quantifies the size of the difference, allowing us to decide if the difference is meaningful.
Uses
Researchers have used Cohen's h as follows.
Describe the differences in proportions using the rule of thumb criteria set out by Cohen. Namely, h = 0.2 is a "small" difference, h = 0.5 is a "medium" difference, and h = 0.8 is a "large" difference.
Only discuss differences that have h greater than some threshold value, such as 0.2.
When the sample size is so large that many differences are likely to be statistically significant, Cohen's h identifies "meaningful", "clinically meaningful", or "practically significant" differences.
Calculation
Given a probability or proportion p, between 0 and 1, its arcsine transformation is
Given two proportions, and , h is defined as the difference between their arcsine transformations. Namely,
This is also sometimes called "directional h" because, in addition to showing the magnitude of the difference, it shows which of the two proportions is greater.
Often, researchers mean "nondirectional h", which is just the absolute value of the directional h:
In R,
|
https://en.wikipedia.org/wiki/William%20C.%20Carter%20Award
|
The William C. Carter Award is a technical award presented annually since 1997 to recognizing an individual who has made a significant contribution to the field of dependable and secure computing throughout his or her PhD dissertation. It is named after, and honors, the late William C. Carter, an important figure in the field. The award is sponsored by IEEE Technical Committee on Fault-Tolerant Computing (TC-FTC) and the IFIP Working Group on Dependable Computing and Fault Tolerance (WG 10.4).
Past recipients
See also
List of computer science awards
|
https://en.wikipedia.org/wiki/Michel%20Deza
|
Michel Marie Deza (27 April 1939 – 23 November 2016) was a Soviet and French mathematician, specializing in combinatorics, discrete geometry and graph theory. He was the retired director of research at the French National Centre for Scientific Research (CNRS), the vice president of the European Academy of Sciences, a research professor at the Japan Advanced Institute of Science and Technology, and one of the three founding editors-in-chief of the European Journal of Combinatorics.
Deza graduated from Moscow University in 1961, after which he worked at the Soviet Academy of Sciences until emigrating to France in 1972. In France, he worked at CNRS from 1973 until his 2005 retirement.
He has written eight books and about 280 academic papers with 75 different co-authors, including four papers with Paul Erdős, giving him an Erdős number of 1.
The papers from a conference on combinatorics, geometry and computer science, held in Luminy, France in May 2007, have been collected as a special issue of the European Journal of Combinatorics in honor of Deza's 70th birthday.
Selected papers
. This paper solved a conjecture of Paul Erdős and László Lovász (in , p. 406) that a sufficiently large family of k-subsets of any n-element universe, in which the intersection of every pair of k-subsets has exactly t elements, has a common t-element set shared by all the members of the family. Manoussakis writes that Deza is sorry not to have kept and framed the US$100 check from Erdős for the prize for solving the problem, and that this result inspired Deza to pursue a lifestyle of mathematics and travel similar to that of Erdős.
. This paper considers functions ƒ from subsets of some n-element universe to integers, with the property that, when A is a small set, the sum of the function values of the supersets of A is zero. The strength of the function is the maximum value t such that all sets A of t or fewer elements have this property. If a family of sets F has the property that it cont
|
https://en.wikipedia.org/wiki/Orbital%20fascia
|
The Orbital Fascia forms the periosteum of the orbit.
It is loosely connected to the bones and can be readily separated from them.
Behind, it is united with the dura mater by processes which pass through the optic foramen and superior orbital fissure, and with the sheath of the optic nerve.
In front, it is connected with the periosteum at the margin of the orbit, and sends off a process which assists in forming the orbital septum.
From it two processes are given off; one to enclose the lacrimal gland, the other to hold the pulley of the Obliquus superior in position.
Anatomy
The orbital fascia consisting of 3 parts:
Periorbita
Considered the periosteum of the bones forming the orbit, and is continuous with dura mater through the superior orbital fissure. It also forms the lacrimal sac.
Bulbar fascia
also known as Tenon's capsule.
It encapsulates the eyeball, forming a narrow space, called the Episcleral space, between the fascia and eyeball. This allows for the movement of the eyeball, while providing a socket that continues posteriorly with the optic nerve and its dural covering. Anteriorly it is attached to the corneoscleral junction.
Orbital septum
The framework that binds the orbital fat pad into the orbit, it also binds the palpebra to the bony orbit.
Other contents of the orbital cavity
Eyeball
Lacrimal gland
Extraocular muscles
Orbital adipose tissue
optic nerve
Oculomotor nerve branches
Trochlear nerve branches
Ophthalmic nerve branches
Abducent nerve branches
Ciliary ganglion
Ophthalmic artery
Ophthalmic veins
|
https://en.wikipedia.org/wiki/Sylvester%20matroid
|
In matroid theory, a Sylvester matroid is a matroid in which every pair of elements belongs to a three-element circuit (a triangle) of the matroid.
Example
The -point line (i.e., the rank 2 uniform matroid on elements, ) is a Sylvester matroid because every pair of elements is a basis and every triple is a circuit.
A Sylvester matroid of rank three may be formed from any Steiner triple system, by defining the lines of the matroid to be the triples of the system. Sylvester matroids of rank three may also be formed from Sylvester–Gallai configurations, configurations of points and lines (in non-Euclidean spaces) with no two-point line. For example, the Fano plane and the Hesse configuration give rise to Sylvester matroids with seven and nine elements respectively, and may be interpreted either as Steiner triple systems or as Sylvester–Gallai configurations.
Properties
A Sylvester matroid with rank must have at least elements; this bound is tight only for the projective spaces over GF(2), of which the Fano plane is an example.
In a Sylvester matroid, every independent set can be augmented by one more element to form a circuit of the matroid.
Sylvester matroids (other than ) cannot be represented over the real numbers (this is the Sylvester–Gallai theorem), nor can they be oriented.
History
Sylvester matroids were studied and named by after James Joseph Sylvester, because they violate the Sylvester–Gallai theorem (for points and lines in the Euclidean plane, or in higher-dimensional Euclidean spaces) that for every finite set of points there is a line containing only two of the points.
|
https://en.wikipedia.org/wiki/Multiple%20gamma%20function
|
In mathematics, the multiple gamma function is a generalization of the Euler gamma function and the Barnes G-function. The double gamma function was studied by . At the end of this paper he mentioned the existence of multiple gamma functions generalizing it, and studied these further in .
Double gamma functions are closely related to the q-gamma function, and triple gamma functions are related to the elliptic gamma function.
Definition
For , let
where is the Barnes zeta function. (This differs by a constant from Barnes's original definition.)
Properties
Considered as a meromorphic function of , has no zeros. It has poles at for non-negative integers . These poles are simple unless some of them coincide. Up to multiplication by the exponential of a polynomial, is the unique meromorphic function of finite order with these zeros and poles.
In the case of the double Gamma function, the asymptotic behaviour for is known, and the leading factor is
Infinite product representation
The multiple gamma function has an infinite product representation that makes it manifest that it is meromorphic, and that also makes the positions of its poles manifest. In the case of the double gamma function, this representation is
where we define the -independent coefficients
where is an -th order residue at .
Another representation as a product over leads to an algorithm for numerically computing the double Gamma function.
Reduction to the Barnes G-function
The double gamma function with parameters obeys the relations
It is related to the Barnes G-function by
The double gamma function and conformal field theory
For and , the function
is invariant under , and obeys the relations
For , it has the integral representation
From the function , we define the double Sine function and the Upsilon function by
These functions obey the relations
plus the relations that are obtained by . For they have the integral representations
The functions and appear in cor
|
https://en.wikipedia.org/wiki/Betavoltaic%20device
|
A betavoltaic device (betavoltaic cell or betavoltaic battery) is a type of nuclear battery which generates electric current from beta particles (electrons) emitted from a radioactive source, using semiconductor junctions. A common source used is the hydrogen isotope tritium.
Unlike most nuclear power sources which use nuclear radiation to generate heat which then is used to generate electricity, betavoltaic devices use a non-thermal conversion process, converting the electron-hole pairs produced by the ionization trail of beta particles traversing a semiconductor.
Betavoltaic power sources (and the related technology of alphavoltaic power sources) are particularly well-suited to low-power electrical applications where long life of the energy source is needed, such as implantable medical devices or military and space applications.
History
Betavoltaics were invented in the 1970s. Some pacemakers in the 1970s used betavoltaics based on promethium, but were phased out as cheaper lithium batteries were developed.
Early semiconducting materials weren't efficient at converting electrons from beta decay into usable current, so higher energy, more expensive—and potentially hazardous—isotopes were used. The more efficient semiconducting materials used can be paired with relatively benign isotopes such as tritium, which produce less radiation.
The Betacel was considered the first successfully commercialized betavoltaic battery.
Proposals
The primary use for betavoltaics is for remote and long-term use, such as spacecraft requiring electrical power for a decade or two. Recent progress has prompted some to suggest using betavoltaics to trickle-charge conventional batteries in consumer devices, such as cell phones and laptop computers. As early as 1973, betavoltaics were suggested for use in long-term medical devices such as pacemakers.
In 2018 a Russian design based on 2-micron thick nickel-63 slabs sandwiched between 10 micron diamond layers was introduced. It produc
|
https://en.wikipedia.org/wiki/Joel%20Spencer
|
Joel Spencer (born April 20, 1946) is an American mathematician. He is a combinatorialist who has worked on probabilistic methods in combinatorics and on Ramsey theory. He received his doctorate from Harvard University in 1970, under the supervision of Andrew Gleason. He is currently () a professor at the Courant Institute of Mathematical Sciences of New York University. Spencer's work was heavily influenced by Paul Erdős, with whom he coauthored many papers (giving him an Erdős number of 1).
In 1963, while studying at the Massachusetts Institute of Technology, Spencer became a Putnam Fellow. In 1984 Spencer received a Lester R. Ford Award. He was an Erdős Lecturer at Hebrew University of Jerusalem in 2001. In 2012 he became a fellow of the American Mathematical Society.
He was elected as a fellow of the Society for Industrial and Applied Mathematics in 2017, "for contributions to discrete mathematics and theory of computing, particularly random graphs and networks, Ramsey theory, logic, and randomized algorithms". In 2021 he received the Leroy P. Steele Prize for Mathematical Exposition with his coauthor Noga Alon for their book The Probabilistic Method.
Selected publications
Probabilistic methods in combinatorics, with Paul Erdős, New York: Academic Press, 1974.
Ramsey theory, with Bruce L. Rothschild and Ronald L. Graham, New York: Wiley, 1980; 2nd ed., 1990.
Ten lectures on the probabilistic method, Philadelphia: Society for Industrial and Applied Mathematics, 1987; 2nd ed., 1994.
The strange logic of random graphs, Berlin: Springer-Verlag, 2001.
The probabilistic method, with Noga Alon, New York: Wiley, 1992; 2nd ed., 2000; 3rd ed., 2008.
Deterministic random walks on regular trees, American Mathematical Society, New York, 2008.
Asymptopia, with Laura Florescu, American Mathematical Society, 2014.
See also
Packing in a hypergraph
|
https://en.wikipedia.org/wiki/Mark%20Jeffrey%20Rosker
|
Mark Jeffrey Rosker is a physicist at the Defense Advanced Research Projects Agency (DARPA) in Arlington, Virginia. He was named a Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2012 for his work on microwave and millimeter-wave phased arrays, gallium nitride semiconductors, and terahertz electronics.
|
https://en.wikipedia.org/wiki/Nodal%20homolog
|
Nodal homolog is a secretory protein that in humans is encoded by the NODAL gene which is located on chromosome 10q22.1. It belongs to the transforming growth factor beta superfamily (TGF-β superfamily). Like many other members of this superfamily it is involved in cell differentiation in early embryogenesis, playing a key role in signal transfer from the primitive node, in the anterior primitive streak, to lateral plate mesoderm (LPM).
Nodal signaling is important very early in development for mesoderm and endoderm formation and subsequent organization of left-right axial structures. In addition, Nodal seems to have important functions in neural patterning, stem cell maintenance and many other developmental processes, including left/right handedness.
Signaling
Nodal can bind type I and type II serine/threonine kinase receptors, with Cripto-1 acting as its co-receptor. Signaling through SMAD 2/3 and subsequent translocation of SMAD 4 to the nucleus promotes the expression of genes involved in proliferation and differentiation. Nodal also further activates its own expression via a positive feedback loop. It is tightly regulated by inhibitors Lefty A, Lefty B, Cerberus, and Tomoregulin-1, which can interfere with Nodal receptor binding.
Species specific Nodal ligands
Nodal is a widely distributed cytokine. The presence of Nodal is not limited to vertebrates, it is also known to be conserved in other deuterostomes (cephalochordates, tunicates and echinoderms) and protostomes such as snails, but neither the nematode C. elegans (another protosome) nor the fruit fly Drosophila (an arthropod) have a copy of nodal. Although mouse and human only have one nodal gene, the zebrafish contain three nodal paralogs: squint , cyclops and southpaw, and the frog five (xnr1,2,3,5 and 6). Even though the zebrafish Nodal homologs are very similar, they have specialized to perform different roles; for instance, Squint and Cyclops are important for mesoendoderm formation, whereas th
|
https://en.wikipedia.org/wiki/Bilirubin%20glucuronide
|
Bilirubin glucuronide is a water-soluble reaction intermediate over the process of conjugation of indirect bilirubin. Bilirubin glucuronide itself belongs to the category of conjugated bilirubin along with bilirubin di-glucuronide. However, only the latter one is primarily excreted into the bile in the normal setting.
Upon macrophages spot and phagocytize the effete Red Blood Corpuscles containing hemoglobin, unconjugated bilirubin is discharged from macrophages into the blood plasma. Most often, the free and water-insoluble unconjugated bilirubin which has an internal hydrodren bonding will bind to albumin and, to a much lesser extent, high density lipoprotein in order to decrease its hydrophobicity and to limit the probability of unnecessary contact with other tissues and keep bilirubin in the vascular space from traversing to extravascular space including brain, and from ending up increasing glomerular filtration. Nevertheless, there is still a little portion of indirect bilirubins stays free-of-bound. Free unconjugated bilirubin can poison the cerebrum.
Finally, albumin leads the indirect bilirubin to the liver. In the liver sinusoid, albumin disassociates with the indirect bilirubin and returns to the circulation while the hepatocyte transfers the indirect bilirubin to ligandin and glucuronide conjugates the indirect bilirubin in the endoplasmic reticulum by disrupting unconjugated bilirubin's internal hydrogen bonding, which is the thing that makes indirect bilirubin having the property of eternal half-elimination life and insoluble in water, and by attaching two molecules of glucuronic acid to it in a two step process. The reaction is a transfer of two glucuronic acid groups including UDP glucuronic acid sequentially to the propionic acid groups of the bilirubin, primarily catalyzed by UGT1A1. In greater detail about this reaction, a glucuronosyl moiety is conjugated to one of the propionic acid side chains, located on the C8 and C12 carbons of the two cent
|
https://en.wikipedia.org/wiki/Digital%20enclosure
|
Digital enclosure is a model of interrogating the relationship that has arisen between surveillance and the new emerging interactive economy developed by Mark Andrejevic, an associate professor at the University of Iowa. The new interactivity-enhanced spaces born of the new digital era has created "digital enclosures" in both a physical and virtual sense. It is this enclosure that allows for individuals or companies to own, operate, and claim ownership over the information generated by the users who interact within their boundaries.
Andrejevic cites the example of Google and Earthlink to provide the city of San Francisco with free wireless access within the city in multiple articles. This project was to be funded through "contextual advertising", the use of custom-targeted ads developed by tracking the location of users throughout the day. This potentially opens the door for google to create custom algorithms to gain information on users from search engine inquiries, map requests, and even Gmail accounts.
Companies who invest in such information from enclosures due to their potentially profit contributes to what Andrejevic describes as "the work of being watched", individuals who unknowingly or willingly submit to giving up their information that generates such economic value in exchange for use of digital commodities.
Andrejevic connects this idea back to the forcible separation of workers from the means of production – a process defined by Marx as "primitive" accumulation. Marx developed this from the first agrarian enclosure of land, where works were separated from the land in order to force workers to regain access to be contractually regulated.
The model of digital enclosures is necessary to understanding new information commodities being valuable resources in the new era of digital capitalism. The private ownership of such enclosures aims to separate users from the means of interaction, transaction, communication, and expression, much like the farmer's se
|
https://en.wikipedia.org/wiki/Cellular%20memory%20modules
|
Cellular memory modules are a form of epigenetic inheritance that allow cells to maintain their original identity after a series of cell divisions and developmental processes. Cellular memory modules implement these preserved characteristics into transferred environments through transcriptional memory. Cellular memory modules are primarily found in Drosophila.
History
Cellular memory modules were discovered by François Jacob and Jaques Monod in 1961 at the Pasteur Institute in Paris. The discovery led to Jacob and Monod, along with André Lwoff, receiving The Nobel Prize in Physiology or Medicine in 1965 for their discoveries regarding genetic control of enzyme and virus synthesis. These experimental results mapped the complex processes in which self-regulating processes express or suppress genes. Monod and Jacob proved how genetic information conversion during the construction of proteins was done through a messenger which evinced RNA. Lwoff aided in the experiment that won the Nobel Prize but did not work on the series of experiments that led to the discovery of cellular memory modules, which is why he remains uncredited in its discovery.
Locations and mechanisms: experiment overviews
Cellular memory modules have the same general process of genes undergoing transcription, these genes being transferred to an unfamiliar environment, and then these genes reverting to their original characteristics preserved through transcriptional memory. Cellular memory modules preserve repressed and active chromatin states in the Polycomb group (PcG) and trithorax group (trxG) proteins by using Polycomb- and trithorax response elements, which are just DNA sequences. Transcription resets and alters epigenetic marks on chromosomal memory elements that are regulated by PcG and trxG proteins. PcG genes maintain silent expression states during the development of Hox genes while trxG proteins maintain Hox gene expression patterns. PcG proteins bind to Polycomb response elements (PREs
|
https://en.wikipedia.org/wiki/Ramanujan%20Math%20Park
|
The Ramanujan Math Park is an Indian museum and activity center dedicated to mathematics education inside the Agastya Campus Creativity Lab located in Kuppam, in Chittoor, Andhra Pradesh. It is named after the Indian mathematician Srinivasa Ramanujan (1887-1920) who was from nearby Madras State. It is a joint project of Agastya International Foundation and the non-profit organization Gyanome.
Agastya is known for its hands-on teaching methods and the Math Park follows this tradition. The park features both indoor and outdoor exhibits as well as interactive touch screen stations, all designed to enhance the mathematical experience. There are plans to replicate this math park experience at other Government run schools elsewhere in India.
History
Ramanujan Math Park was conceived, partially funded and executed by Sujatha Ramdorai and her husband Srinivasan Ramdorai along with V.S.S Sastry, an Indian mathematics communicator based in nearby Kolar. It was inaugurated in 2017, on 22 December, Ramanujan's birthday and the day celebrated in India as National Math Day.
|
https://en.wikipedia.org/wiki/Remote%20digital%20terminal
|
In telecommunications, a remote digital terminal (RDT) typically accepts E1, T1 or OC-3 digital lines to communicate with a telephone Access network (AN) or telephone exchange (Local Digital Switch, LDS) on one side, and forms a local exchange (LE) on the other, which is connected to "plain old telephone service" (POTS) lines.
See also
Distributed switching
Remote concentrator
Digital access carrier system
|
https://en.wikipedia.org/wiki/Vertex%20cover
|
In graph theory, a vertex cover (sometimes node cover) of a graph is a set of vertices that includes at least one endpoint of every edge of the graph.
In computer science, the problem of finding a minimum vertex cover is a classical optimization problem. It is NP-hard, so it cannot be solved by a polynomial-time algorithm if P ≠ NP. Moreover, it is hard to approximate – it cannot be approximated up to a factor smaller than 2 if the unique games conjecture is true. On the other hand, it has several simple 2-factor approximations. It is a typical example of an NP-hard optimization problem that has an approximation algorithm. Its decision version, the vertex cover problem, was one of Karp's 21 NP-complete problems and is therefore a classical NP-complete problem in computational complexity theory. Furthermore, the vertex cover problem is fixed-parameter tractable and a central problem in parameterized complexity theory.
The minimum vertex cover problem can be formulated as a half-integral, linear program whose dual linear program is the maximum matching problem.
Vertex cover problems have been generalized to hypergraphs, see Vertex cover in hypergraphs.
Definition
Formally, a vertex cover of an undirected graph is a subset of such that , that is to say it is a set of vertices where every edge has at least one endpoint in the vertex cover . Such a set is said to cover the edges of . The upper figure shows two examples of vertex covers, with some vertex cover marked in red.
A minimum vertex cover is a vertex cover of smallest possible size. The vertex cover number is the size of a minimum vertex cover, i.e. . The lower figure shows examples of minimum vertex covers in the previous graphs.
Examples
The set of all vertices is a vertex cover.
The endpoints of any maximal matching form a vertex cover.
The complete bipartite graph has a minimum vertex cover of size .
Properties
A set of vertices is a vertex cover if and only if its complement is an
|
https://en.wikipedia.org/wiki/Boltzmann%20sampler
|
A Boltzmann sampler is an algorithm intended for random sampling of combinatorial structures. If the object size is viewed as its energy, and the argument of the corresponding generating function is interpreted in terms of the temperature of the physical system, then a Boltzmann sampler returns an object from a classical Boltzmann distribution.
The concept of Boltzmann sampler was proposed by Philippe Duchon, Philippe Flajolet, Guy Louchard and Gilles Schaeffer in 2004.
Description
The concept of Boltzmann sampling is closely related to the symbolic method in combinatorics.
Let be a combinatorial class with an ordinary generating function which has a nonzero radius of convergence , i.e. is complex analytic. Formally speaking, if each object
is equipped with a non-negative integer size , then the generating function is defined as
where denotes the number of objects of size . The size function is typically used to denote the number of vertices in a tree or in a graph, the number of letters in a word, etc.
A Boltzmann sampler for the class with a parameter such that , denoted as
returns an object with probability
Construction
Finite sets
If is finite, then an element is drawn with probability proportional to .
Disjoint union
If the target class is a disjoint union of two other classes, , and the generating functions and of and are known, then the Boltzmann sampler for can be obtained as
where stands for "if the random variable is 1, then execute , else execute ". More generally, if the disjoint union is taken over a finite set, the resulting Boltzmann sampler can be represented using a random choice with probabilities proportional to the values of the generating functions.
Cartesian product
If is a class constructed of ordered pairs where and , then the corresponding Boltzmann sampler can be obtained as
i.e. by forming a pair with and drawn independently from and .
Sequence
If is composed of all the finite sequences of ele
|
https://en.wikipedia.org/wiki/Altered%20Beast
|
Altered Beast is a 1988 beat 'em up arcade video game developed and manufactured by Sega. The game is set in Ancient Greece and follows a player character chosen by Zeus to rescue his daughter Athena from the demonic ruler of the underworld, Neff. Through the use of power-ups, the player character can assume the form of different magical beasts (wolf, dragon, bear, tiger, and golden wolf). It was ported to several home video game consoles and home computers. It was the pack-in game for the Mega Drive when that system launched in 1988.
The game was developed by Makoto Uchida, who developed the game as his first project as a lead developer. Uchida and his team used the System 16 arcade system board for its graphical capabilities with sprites. Altered Beast was ported numerous times in addition to its Genesis conversion, including for the Master System by Sega and to several computer systems and video game consoles by various third parties.
Altered Beasts arcade release and its various ports have all received mixed reviews, mainly targeting the game's gameplay and graphics. The game has been re-released several times for various consoles and as part of video game compilations, and two sequels to the game have been developed.
Gameplay
Altered Beast is a side scrolling beat 'em up game with light platform elements. It has five levels and can be played by up to two players simultaneously. Combat takes place across five levels set in Ancient Greece and populated by aggressive undead creatures and monsters resembling those from Greek mythology. The demonic god Neff waits at the end of each level. Between each level are small animations giving the player glimpses of Athena's peril. Players can punch, kick and jump.
The game's premise is that Neff, ruler of the underworld, captures the goddess Athena. Angry, her father, the Olympian god Zeus, decides to choose a champion to save her. Respecting the bravery of Roman Centurions, Zeus resurrects one of them and empowers him
|
https://en.wikipedia.org/wiki/Rindler%20coordinates
|
Rindler coordinates are a coordinate system used in the context of special relativity to describe the hyperbolic acceleration of a uniformly accelerating reference frame in flat spacetime. In relativistic physics the coordinates of a hyperbolically accelerated reference frame constitute an important and useful coordinate chart representing part of flat Minkowski spacetime. In special relativity, a uniformly accelerating particle undergoes hyperbolic motion, for which a uniformly accelerating frame of reference in which it is at rest can be chosen as its proper reference frame. The phenomena in this hyperbolically accelerated frame can be compared to effects arising in a homogeneous gravitational field. For general overview of accelerations in flat spacetime, see Acceleration (special relativity) and Proper reference frame (flat spacetime).
In this article, the speed of light is defined by , the inertial coordinates are , and the hyperbolic coordinates are . These hyperbolic coordinates can be separated into two main variants depending on the accelerated observer's position: If the observer is located at time at position (with as the constant proper acceleration measured by a comoving accelerometer), then the hyperbolic coordinates are often called Rindler coordinates with the corresponding Rindler metric. If the observer is located at time at position , then the hyperbolic coordinates are sometimes called Møller coordinates or Kottler–Møller coordinates with the corresponding Kottler–Møller metric. An alternative chart often related to observers in hyperbolic motion is obtained using Radar coordinates which are sometimes called Lass coordinates. Both the Kottler–Møller coordinates as well as Lass coordinates are denoted as Rindler coordinates as well.
Regarding the history, such coordinates were introduced soon after the advent of special relativity, when they were studied (fully or partially) alongside the concept of hyperbolic motion: In relation to flat Mi
|
https://en.wikipedia.org/wiki/Neural%20top%E2%80%93down%20control%20of%20physiology
|
Neural top–down control of physiology concerns the direct regulation by the brain of physiological functions (in addition to smooth muscle and glandular ones). Cellular functions include the immune system’s production of T-lymphocytes and antibodies, and nonimmune related homeostatic functions such as liver gluconeogenesis, sodium reabsorption, osmoregulation, and brown adipose tissue nonshivering thermogenesis. This regulation occurs through the sympathetic and parasympathetic system (the autonomic nervous system), and their direct innervation of body organs and tissues that starts in the brainstem. There is also a noninnervation hormonal control through the hypothalamus and pituitary (HPA). These lower brain areas are under control of cerebral cortex ones. Such cortical regulation differs between its left and right sides. Pavlovian conditioning shows that brain control over basic cell level physiological function can be learned.
Higher brain
Cerebral cortex
Sympathetic and parasympathetic nervous systems and the hypothalamus are regulated by the higher brain. Through them, the higher cerebral cortex areas can control the immune system, and the body’s homeostatic and stress physiology. Areas doing this include the insular cortex, the orbital, and the medial prefrontal cortices. These cerebral areas also control smooth muscle and glandular physiological processes through the sympathetic and parasympathetic nervous system including blood circulation, urogenital, gastrointestinal functions, pancreatic gut secretions, respiration, coughing, vomiting, piloerection, pupil dilation, lacrimation and salivation.
Lateralization
The sympathetic nervous system is predominantly controlled by the right side of the brain (focused upon the insular cortex), while the left side predominantly controls the parasympathetic nervous system. The cerebral cortex in rodents shows lateral specialization in its regulation of immunity with immunosuppression being controlled by the righ
|
https://en.wikipedia.org/wiki/Laundry%20symbol
|
A laundry symbol, also called a care symbol, is a pictogram indicating the manufacturer's suggestions as to methods of washing, drying, dry-cleaning and ironing clothing. Such symbols are written on labels, known as care labels or care tags, attached to clothing to indicate how a particular item should best be cleaned. While there are internationally recognized standards for the care labels and pictograms, their exact use and form differ by region. In some standards, pictograms coexist with or are complemented by written instructions.
Standards
GINETEX, the France-based European association for textile care labelling, was formed in 1963 in part to define international standards for the care and labelling of textiles. By the early 1970s, GINETEX was working with ISO to develop international standards for textile labelling, eventually leading to the ISO 3758 standard, Textiles – Care labelling code using symbols. ISO 3758 was supplemented in 1993, revised in 2005 and again in 2012 with reviews of the standard held on a five-year cycle.
In March 1970, the Canadian Government Specifications Board published 86-GP-1, Standard for Care Labelling of Textiles, which promoted a symbol-based textile care labelling system in which symbols were colored: green indicated "no precautions are necessary", yellow indicated "some caution is necessary", and red indicated "prohibited". Publication 86-GP-1 was revised several times over the following three decades; the most noteworthy change was in 1979, when temperatures changed from Fahrenheit to Celsius, and any additional instructions were to be added in text, in both English and French. In 2003, the system was withdrawn in favor of a black-and-white symbol-based system harmonized with North American and international standards. The inclusion of care symbols on garments made or sold in Canada has always been voluntary; only fabric content labels are mandatory (since 1972).
In 1996, in the United States, ASTM International publishe
|
https://en.wikipedia.org/wiki/Step%20detection
|
In statistics and signal processing, step detection (also known as step smoothing, step filtering, shift detection, jump detection or edge detection) is the process of finding abrupt changes (steps, jumps, shifts) in the mean level of a time series or signal. It is usually considered as a special case of the statistical method known as change detection or change point detection. Often, the step is small and the time series is corrupted by some kind of noise, and this makes the problem challenging because the step may be hidden by the noise. Therefore, statistical and/or signal processing algorithms are often required.
The step detection problem occurs in multiple scientific and engineering contexts, for example in statistical process control (the control chart being the most directly related method), in exploration geophysics (where the problem is to segment a well-log recording into stratigraphic zones), in genetics (the problem of separating microarray data into similar copy-number regimes), and in biophysics (detecting state transitions in a molecular machine as recorded in time-position traces). For 2D signals, the related problem of edge detection has been studied intensively for image processing.
Algorithms
When the step detection must be performed as and when the data arrives, then online algorithms are usually used,
and it becomes a special case of sequential analysis.
Such algorithms include the classical CUSUM method applied to changes in mean.
By contrast, offline algorithms are applied to the data potentially long after it has been received. Most offline algorithms for step detection in digital data can be categorised as top-down, bottom-up, sliding window, or global methods.
Top-down
These algorithms start with the assumption that there are no steps and introduce possible candidate steps one at a time, testing each candidate to find the one that minimizes some criteria (such as the least-squares fit of the estimated, underlying piecewise constant
|
https://en.wikipedia.org/wiki/American%20Institute%20for%20Stuttering
|
The American Institute for Stuttering is an American nonprofit organization that provides universally affordable speech therapy to people who stutter. The organization, legally known as The American Institute for Stuttering Treatment and Professional Training (AIS), was founded in 1998 by speech-language pathologist Catherine Otto Montgomery in New York, New York. The current clinical director is speech-language pathologist Heather Grossman, PhD. AIS currently has offices in New York City, Atlanta, GA, and Minneapolis, MN.
The organization provides stuttering therapy to children and adults, financial assistance to help make therapy affordable, and clinical training to speech-language pathologists seeking specialized knowledge in the treatment of stuttering.
Board of directors
The members of the board of directors are:
Chairman Eric Dinallo – Debevoise & Plimpton LLP
Treasurer Nolan Russo, Jr. – Capital Printing Corporation
Vice Chair Sander Flaum – Flaum Partners, Inc.
Secretary Z. Logan Gould – Maynard, Cooper & Gale, LLP
Aaron Graff – Ferring Pharmaceuticals, Inc
Amish Shah – Goodwin Procter LLP
Anthony Hooper – Biotech
Austin Pendleton – actor, director, and writer
Barry Blaustein – writer, director, producer
Carly Simon – singer/Songwriter
Chris Coleridge – Hum Nutrition Inc.
Clarence Page – Journalist
David Esseks – Allen and Overy, LLP
David Resnicow – Resnicow and associates
Emily Blunt – actress
John Stossel – Fox Business
Julie M. Henson – Taft
Kenyatta Bolden – Prudential Financial
Kenyon Martin – NBA player
Kerri Chase – McDermott Will & Emery LLP
Laureen Coyne – Risk Management Consultant
Norbert Lewandowski – Lewandowski & Company
Rachel Cortese – Speech Language Pathologist, Behavior Therapist
Susan Reichardt – Surface Transportation Board
Will Blodgett – Fairstead
William D. Marsillo – Boies Schiller Flexner LLP
See also
National Stuttering Association
Stuttering Foundation of America
Stuttering therapy
|
https://en.wikipedia.org/wiki/Ripe%20Digital%20Entertainment
|
Ripe Digital Entertainment (RDE) was an on-demand digital entertainment company with video on demand components on several platforms. RDE was founded by CEO Ryan Magnussen.
Networks
Ripe's 'networks', which consisted of Ripe TV, Octane TV and Flow TV were delivered over multiple platform. Ripe TV, launched in October 2005, focused on male content such as sports, comedy and content usually seen in the lad mag format of men's magazines such as FHM. Octane TV, which launched in August 2006, mainly involved automotive content. Flow TV launched in April 2007 and focused mainly on hip-hop music. All three services ended in June 2009 with the company's shutdown.
Distribution
The services were mainly distributed through cable video on demand services such as Comcast, Time Warner Cable and FiOS, along with syndication through the AOL, Google and Yahoo ad networks. Sprint and MVNO Helio provided RDE's content via mobile.
|
https://en.wikipedia.org/wiki/DIVX
|
DIVX (Digital Video Express) is a discontinued digital video format. Created in part by Circuit City, it was an unsuccessful attempt to create an alternative to video rental in the United States. The format's poor reception from consumers resulted in major financial losses for Circuit City and is credited with being part of the company's downfall.
Format
DIVX was a rental format variation on the DVD player in which a customer would buy a DIVX disc (similar to a DVD) for approximately US$4.50, which was watchable for up to 48 hours from its initial viewing. After this period, the disc could be viewed by paying a continuation fee to play it for two more days. Viewers who wanted to watch a disc an unlimited number of times could convert the disc to a "DIVX silver" disc for an additional fee. "DIVX gold" discs that could be played an unlimited number of times on any DIVX player were announced at the time of DIVX's introduction, but no DIVX gold titles were ever released.
Each DIVX disc was marked with a unique barcode in the burst cutting area that could be read by the player, and used to track the discs. The status of the discs was monitored through an account over a phone line. DIVX player owners had to set up an account with DIVX to which additional viewing fees could be charged. The player would call an account server over the phone line to charge for viewing fees similar to the way DirecTV and Dish Network satellite systems handle pay-per-view.
In addition to the normal Content Scramble System (CSS) encryption, DIVX discs used Triple DES encryption and an alternative channel modulation coding scheme, which prevented them from being read in standard DVD players. Most of the discs would be manufactured by United Kingdom-based Nimbus CD International.
DIVX players manufactured by Zenith Electronics (who would go bankrupt shortly before the launch of the format), Thomson Consumer Electronics (RCA and ProScan), and Matsushita Electric (Panasonic) started to become a
|
https://en.wikipedia.org/wiki/Synchronous%20Serial%20Interface
|
Synchronous Serial Interface (SSI) is a widely used serial interface standard for industrial applications between a master (e.g. controller) and a slave (e.g. sensor). SSI is based on RS-422 standards and has a high protocol efficiency in addition to its implementation over various hardware platforms, making it very popular among sensor manufacturers. SSI was originally developed by Max Stegmann GmbH in 1984 for transmitting the position data of absolute encoders – for this reason, some servo/drive equipment manufacturers refer to their SSI port as a "Stegmann Interface".
It was formerly covered by the German patent DE 34 45 617 which expired in 1990. It is very suitable for applications demanding reliability and robustness in measurements under varying industrial environments.
It is different from the Serial Peripheral Interface Bus (SPI): An SSI bus is differential, simplex, non-multiplexed, and relies on a time-out to frame the data. An SPI bus is single-ended, duplex, and uses a chip select signal to frame the data. However, SPI peripherals on microcontrollers can implement SSI with external differential driver-ICs and program-controlled timing.
Introduction
SSI is a synchronous, point-to-point, serial communication channel for digital data transmission. Synchronous data transmission is one in which the data is transmitted by synchronizing the transmission at the receiving and sending ends using a common clock signal. Since the start and stop bits are not present, this allows better use of data transmission bandwidth for more message bits and makes the whole transmission process simpler and easier. The clock needs its own bandwidth and should be included when determining the total bandwidth required for communication between the two devices.
In general, as mentioned earlier, it is a point-to-point connection from a master (e.g., PLC, Microcontroller) to a slave (e.g. rotary encoders). The master controls the clock sequence, and the slave transmits the curr
|
https://en.wikipedia.org/wiki/Decimal%20separator
|
A decimal separator is a symbol used to separate the integer part from the fractional part of a number written in decimal form (e.g., "." in 12.45). Different countries officially designate different symbols for use as the separator. The choice of symbol also affects the choice of symbol for the thousands separator used in digit grouping.
Any such symbol can be called a decimal mark, decimal marker, or decimal sign. Symbol-specific names are also used; decimal point and decimal comma refer to an (either baseline or middle) dot and comma respectively, when it is used as a decimal separator; these are the usual terms used in English, with the aforementioned generic terms reserved for abstract usage.
In many contexts, when a number is spoken, the function of the separator is assumed by the spoken name of the symbol: comma or point in most cases. In some specialized contexts, the word decimal is instead used for this purpose (such as in International Civil Aviation Organization-regulated air traffic control communications). In mathematics, the decimal separator is a type of radix point, a term that also applies to number systems with bases other than ten.
History
Hellenistic–Renaissance eras
In the Middle Ages, from the original Indian decimal writing, before printing, a bar ( ¯ ) over the units digit was used to separate the integral part of a number from its fractional part, as in 995 (meaning 99.95 in decimal point format). A similar notation remains in common use as an underbar to superscript digits, especially for monetary values without a decimal separator, as in 99. Later, a "separatrix" (i.e., a short, roughly vertical ink stroke) between the units and tenths position became the norm among Arab mathematicians (e.g. 99ˌ95), while an L-shaped or vertical bar () served as the separatrix in England. When this character was typeset, it was convenient to use the existing comma (99,95) or full stop (99.95) instead.
Positional decimal fractions appear for the firs
|
https://en.wikipedia.org/wiki/Bose%E2%80%93Hubbard%20model
|
The Bose–Hubbard model gives a description of the physics of interacting spinless bosons on a lattice. It is closely related to the Hubbard model that originated in solid-state physics as an approximate description of superconducting systems and the motion of electrons between the atoms of a crystalline solid. The model was introduced by Gersch and Knollman in 1963 in the context of granular superconductors. (The term 'Bose' in its name refers to the fact that the particles in the system are bosonic.) The model rose to prominence in the 1980s after it was found to capture the essence of the superfluid-insulator transition in a way that was much more mathematically tractable than fermionic metal-insulator models.
The Bose–Hubbard model can be used to describe physical systems such as bosonic atoms in an optical lattice, as well as certain magnetic insulators. Furthermore, it can be generalized and applied to Bose–Fermi mixtures, in which case the corresponding Hamiltonian is called the Bose–Fermi–Hubbard Hamiltonian.
Hamiltonian
The physics of this model is given by the Bose–Hubbard Hamiltonian:
.
Here, denotes summation over all neighboring lattice sites and , while and are bosonic creation and annihilation operators such that gives the number of particles on site . The model is parametrized by the hopping amplitude that describes boson mobility in the lattice, the on-site interaction which can be attractive () or repulsive (), and the chemical potential , which essentially sets the number of particles. If unspecified, typically the phrase 'Bose–Hubbard model' refers to the case where the on-site interaction is repulsive.
This Hamiltonian has a global symmetry, which means that it is invariant (its physical properties are unchanged) by the transformation . In a superfluid phase, this symmetry is spontaneously broken.
Hilbert space
The dimension of the Hilbert space of the Bose–Hubbard model is given by , where is the total number of particles, whil
|
https://en.wikipedia.org/wiki/Polignac%27s%20conjecture
|
In number theory, Polignac's conjecture was made by Alphonse de Polignac in 1849 and states:
For any positive even number n, there are infinitely many prime gaps of size n. In other words: There are infinitely many cases of two consecutive prime numbers with difference n.
Although the conjecture has not yet been proven or disproven for any given value of n, in 2013 an important breakthrough was made by Zhang Yitang who proved that there are infinitely many prime gaps of size n for some value of n < 70,000,000. Later that year, James Maynard announced a related breakthrough which proved that there are infinitely many prime gaps of some size less than or equal to 600. As of April 14, 2014, one year after Zhang's announcement, according to the Polymath project wiki, n has been reduced to 246. Further, assuming the Elliott–Halberstam conjecture and its generalized form, the Polymath project wiki states that n has been reduced to 12 and 6, respectively.
For n = 2, it is the twin prime conjecture. For n = 4, it says there are infinitely many cousin primes (p, p + 4). For n = 6, it says there are infinitely many sexy primes (p, p + 6) with no prime between p and p + 6.
Dickson's conjecture generalizes Polignac's conjecture to cover all prime constellations.
Conjectured density
Let for even n be the number of prime gaps of size n below x.
The first Hardy–Littlewood conjecture says the asymptotic density is of form
where Cn is a function of n, and means that the quotient of two expressions tends to 1 as x approaches infinity.
C2 is the twin prime constant
where the product extends over all prime numbers p ≥ 3.
Cn is C2 multiplied by a number which depends on the odd prime factors q of n:
For example, C4 = C2 and C6 = 2C2. Twin primes have the same conjectured density as cousin primes, and half that of sexy primes.
Note that each odd prime factor q of n increases the conjectured density compared to twin primes by a factor of . A heuristic argument follows. It r
|
https://en.wikipedia.org/wiki/Remote%20pickup%20unit
|
A remote pickup unit or RPU is a radio system using special radio frequencies set aside for electronic news-gathering (ENG) and remote broadcasting. It can also be used for other types of point-to-point radio links.
An RPU is used to send program material from a remote location to the broadcast station or network. Usually these systems use specialized high audio fidelity radio equipment. One manufacturer, Marti, was best known for manufacturing remote pickup equipment, so much so that the name is usually used to refer to a remote pickup unit regardless of who the actual equipment manufacturer actually is.
Today much of the remote broadcast use digital audio system fed over ISDN telephone lines. This method is favored because of reliability of telephone lines versus a radio link back to the station. The radio RPU remains much more favored for ENG however, because of its flexibility.
Footnotes
Broadcast engineering
|
https://en.wikipedia.org/wiki/RNA%20helicase%20database
|
The RNA helicase database stores data (sequence, structures...) about RNA helicases.
See also
Helicase
|
https://en.wikipedia.org/wiki/Distributed%20Sender%20Blackhole%20List
|
The Distributed Sender Blackhole List was a Domain Name System-based Blackhole List that listed IP addresses of insecure e-mail hosts. DSBL could be used by server administrators to tag or block e-mail messages that came from insecure servers, which is often spam.
The DSBL published its lists as domain name system (DNS) zones that could be queried by anyone on the Internet.
DSBL is a dead RBL as of May 2008. Its administrators continued to run their authoritative nameservers for several months after their decommissioning announcement; as of March 9, 2009, even those servers are offline. At this point, using any *.dsbl.org lookups in an RBL check results in DNS failures and can even prevent an SMTP server from starting a conversation.
Blocking
It is not possible for DSBL to block or intercept mail. E-mail is sometimes blocked or bounced with a message referencing DSBL. These messages were not blocked by DSBL; they were blocked by the administrator of the receiving mail server, who chose to reject messages coming from a potentially-insecure IP address listed by DSBL. See DNSBL for a description of how mail transfer agents interact with these lists.
Methodology
DSBL lists IP addresses of hosts that are demonstrated to be insecure. DSBL defines an insecure host as one that allows e-mail to be sent from anyone to anyone else. Normal servers only send mail from their own users to anyone else. Insecure servers are commonly abused by spammers, although DSBL does not claim that the hosts have sent spam or have been abused by spammers; only that they could be.
DSBL builds its lists by receiving specially-formatted "listme" e-mails triggered by testers. DSBL itself does not test hosts for security vulnerabilities. The testers use software that causes insecure servers to send a message to an e-mail address monitored by DSBL. The message includes a time-sensitive cryptographically secure cookie to prevent servers from being listed by mistake. When a valid listme message i
|
https://en.wikipedia.org/wiki/Intramolecular%20force
|
An intramolecular force (or primary forces) is any force that binds together the atoms making up a molecule or compound, not to be confused with intermolecular forces, which are the forces present between molecules. The subtle difference in the name comes from the Latin roots of English with inter meaning between or among and intra meaning inside. Chemical bonds are considered to be intramolecular forces which are often stronger than intermolecular forces present between non-bonding atoms or molecules.
Types
The classical model identifies three main types of chemical bonds — ionic, covalent, and metallic — distinguished by the degree of charge separation between participating atoms. The characteristics of the bond formed can be predicted by the properties of constituent atoms, namely electronegativity. They differ in the magnitude of their bond enthalpies, a measure of bond strength, and thus affect the physical and chemical properties of compounds in different ways. % of ionic character is directly proportional difference in electronegitivity of bonded atom.
Ionic bond
An ionic bond can be approximated as complete transfer of one or more valence electrons of atoms participating in bond formation, resulting in a positive ion and a negative ion bound together by electrostatic forces. Electrons in an ionic bond tend to be mostly found around one of the two constituent atoms due to the large electronegativity difference between the two atoms, generally more than 1.9, (greater difference in electronegativity results in a stronger bond); this is often described as one atom giving electrons to the other. This type of bond is generally formed between a metal and nonmetal, such as sodium and chlorine in NaCl. Sodium would give an electron to chlorine, forming a positively charged sodium ion and a negatively charged chloride ion.
Covalent bond
In a true covalent bond, the electrons are shared evenly between the two atoms of the bond; there is little or no charge separa
|
https://en.wikipedia.org/wiki/CDMA%20subscriber%20identity%20module
|
A CDMA subscriber identity module (CSIM) is an application to support CDMA2000 phones that runs on a UICC, with a file structure derived from the R-UIM card. By porting the application to the UICC (Universal Integrated Circuit Card), a card with CSIM, SIM, and USIM can operate with all major cellular technologies worldwide. The CSIM application allows users to change phones by simply removing the smart card from one mobile phone and inserting it into another mobile phone or broadband telephony device supporting the CDMA2000 radio interface.
CSIM application file system
The CSIM application contains a file system with a number of parameters needed to operate on cdmaOne/CDMA2000 ("CDMA") networks. Each parameter, or a group of related parameters, is specified with a unique identifier with an implicit or explicit length, and is considered a separate Elementary File (EF). The following examples are taken from the 3GPP2 specification.
Encryption keys, provisioned by the operator
Phone number
Call counts
Short Message Service parameters
Received Short Messages (255 bytes maximum per message)
IMSI (international mobile subscriber identifier)
TMSI (temporary mobile subscriber identifier, for position security)
UIMID (hardware identifier). It will be a pseudo (hashed) value if EUIMID is in use.
EUIMID, either short form (based on MEID) or long form (based on ICCID)
ICCID. Present even if it is not used as EUIMID.
MEID (hardware identifier)
Analog (AMPS) operational parameters
CDMA2000 home identifiers, such as SID and NID
CDMA2000 zone-based registration parameters, telling the handset to register when it changes to a new operator-defined zone (SID/NID)
CDMA2000 distance-based registration parameters, telling the handset to register when it travels a certain distance (each CDMA2000 transmitter emits its GPS location)
Random parameters (slot cycle index) to use during CSMA access probes
List of services available. If a service is not indicated as availa
|
https://en.wikipedia.org/wiki/Goldberg%E2%80%93Shprintzen%20syndrome
|
Goldberg–Shprintzen is a very rare connective tissue condition associated with mutations in KIAA1279 gene which encodes KIF-binding protein (KBP), a protein that may interact with microtubules and actin filaments. KBP may play a key role in cytoskeleton formation and neurite growth.
Hirschsprung's disease may be part of the presentation. Developmental abnormalities shown by people with Goldberg–Shprintzen syndrome may be ocular, cardiac, urogenital, and skeletal .
It was first observed by Sugarman and Vogel in 1981 and later established as a separate clinical identity by R B Goldberg and R J Shprintzen in 1982. The sibling pair had microcephaly, hypertelorism, short stature, submucous cleft palate, learning problems, and Hirschsprung aganglionic megacolon (HAM). This combination of symptoms had not been previously identified with a diagnosis.
Three more cases were observed in 1988 by Hurst et al., although one of these cases could have been Mowat-Wilson syndrome instead of Goldberg–Shprintzen syndrome. Additional cases have since been documented sporadically across the world, although it remains extremely uncommon.
Signs and Symptoms
Goldberg–Shprintzen syndrome presents a wide variety of phenotypic symptoms. It is characterized by a combination of any of the following symptoms:
Central nervous system: mental retardation, microcephaly, gyral anomalies of the brain indicative of defects in migration of neural crest cells and neurons, abnormal EEG, hypoplasia of the corpus callosum and cerebella, and enlargement of the subarachnoid space
Cardiovascular: progressive aortic root dilation, mitral valve prolapse, and mitral regurgitation
Urogenital: Hirschsprung disease and constipation
Optical: megalocornea and pale optic discs with poor fixation
Dysmorphic facial features: hypertelorism, submucous cleft palate, sloping forehead, sparse eyebrows, depressed nasal bridge, bulbous nose, full lips, sparse hair, arched eyebrows, long eyelashes, ptosis, downslanting
|
https://en.wikipedia.org/wiki/Boole%27s%20rule
|
In mathematics, Boole's rule, named after George Boole, is a method of numerical integration.
Formula
Simple Boole's Rule
It approximates an integral:
by using the values of at five equally spaced points:
It is expressed thus in Abramowitz and Stegun:
where the error term is
for some number between and where .
It is often known as Bode's rule, due to a typographical error that propagated from Abramowitz and Stegun.
The following constitutes a very simple implementation of the method in Common Lisp which ignores the error term:
(defun integrate-booles-rule (f x1 x5)
"Calculates the Boole's rule numerical integral of the function F in
the closed interval extending from inclusive X1 to inclusive X5
without error term inclusion."
(declare (type (function (real) real) f))
(declare (type real x1 x5))
(let ((h (/ (- x5 x1) 4)))
(declare (type real h))
(let* ((x2 (+ x1 h))
(x3 (+ x2 h))
(x4 (+ x3 h)))
(declare (type real x2 x3 x4))
(* (/ (* 2 h) 45)
(+ (* 7 (funcall f x1))
(* 32 (funcall f x2))
(* 12 (funcall f x3))
(* 32 (funcall f x4))
(* 7 (funcall f x5)))))))
Composite Boole's Rule
In cases where the integration is permitted to extend over equidistant sections of the interval , the composite Boole's rule might be applied. Given divisions, the integrated value amounts to:
where the error term is similar to above. The following Common Lisp code implements the aforementioned formula:
(defun integrate-composite-booles-rule (f a b n)
"Calculates the composite Boole's rule numerical integral of the
function F in the closed interval extending from inclusive A to
inclusive B across N subintervals."
(declare (type (function (real) real) f))
(declare (type real a b))
(declare (type (integer 1 *) n))
(let ((h (/ (- b a) n)))
(declare (type real h))
(flet ((f[i] (i)
(decl
|
https://en.wikipedia.org/wiki/Royal%20Society%20Africa%20Prize
|
The Royal Society Africa Prize (formerly known as the Royal Society Pfizer Prize) has been awarded by the Royal Society since 2006 to African-based researchers at the start of their career who are making innovative contributions to the biological sciences in Africa. £60,000 is awarded as a grant for the recipient to carry out a research project that is linked to an African centre of scientific excellence, normally a University or equivalent research centre, and a further £5,000 is given directly to the prizewinner.
The final award under the Pfizer name was made in 2016, after which the award was renamed the Royal Society Africa Prize, and consists of a grant of £11,000 and a gift of £1000.
Recipients
Source: Royal Society
Royal Society Pfizer Prize
2006: Alexis Nzila
2007: Hiba Mohamed, for her pioneering research into genetic susceptibility to leishmaniasis, a parasitic disease transmitted by sand fly bites
2008: Enock Matovu
2009: Linda-Gail Bekker, director of the Desmond Tutu HIV Centre, for her outstanding research into tuberculosis and HIV co-infections in Africa
2010: Collins Ouma of Maseno University, for his outstanding research into the effect of genetic variation in severe malarial anaemia (SMA) in children.
2011: Julie Makani of the Nuffield Department of Clinical Medicine and Muhimbili University, Tanzania, for her outstanding research into using anaemia in sickle cell disease as a model for translating genetic research into health benefits
2012: Martin Ota, for his outstanding research into the relationship of pneumococcal protein antibody levels to nasopharyngeal carriage of pneumococci in early infancy.
2013: Abdoulaye Diabate, for his important work on the identification of mosquito swarming cues
2014: Faith Osier, for her research on understanding the mechanisms of immunity to malaria infection in man
2015: Gordon A Awandare, for his achievements in molecular and cellular studies of malaria, including how malaria parasites invade red
|
https://en.wikipedia.org/wiki/Mir-154%20microRNA%20precursor%20family
|
In molecular biology mir-154 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms.
See also
MicroRNA
|
https://en.wikipedia.org/wiki/Fleet%20Street%20Publisher
|
Fleet Street Publisher was an Atari ST desktop publishing program produced by Mirrorsoft in the United Kingdom and released in November 1986. A IBM PC compatible version produced by Rowan Software was planned for 1987 but never released.
Running under GEM the program offered features such as multi-column text, the ability to design flow charts and graphics and multiple document sizes (flyers, menus, cards, etc.). Possible font sizes ranged from 4 to 216 points with support for accented characters. Character and line spacing where fully controllable by the user. The software came with a 150 image clipart gallery.
The software was superseded by Timeworks Publisher (Publish-It in the United States), which the market regarded as a much better product. This new version was produced by GST Software Products, and upgrades for the PC versions were available into the late 2000s.
Versions
Fleet Street Publisher (1986, published by Spectrum Holobyte and France Image Logiciel)
Fleet Street Publisher 1.1 (1987, published by Mirrorsoft)
Fleet Street Publisher 2.0 (1989, published by MichTron)
Fleet Street Publisher 3.0 (1989, published by MichTron)
|
https://en.wikipedia.org/wiki/Complex%20conjugate%20root%20theorem
|
In mathematics, the complex conjugate root theorem states that if P is a polynomial in one variable with real coefficients, and a + bi is a root of P with a and b real numbers, then its complex conjugate a − bi is also a root of P.
It follows from this (and the fundamental theorem of algebra) that, if the degree of a real polynomial is odd, it must have at least one real root. That fact can also be proved by using the intermediate value theorem.
Examples and consequences
The polynomial x2 + 1 = 0 has roots ± i.
Any real square matrix of odd degree has at least one real eigenvalue. For example, if the matrix is orthogonal, then 1 or −1 is an eigenvalue.
The polynomial
has roots
and thus can be factored as
In computing the product of the last two factors, the imaginary parts cancel, and we get
The non-real factors come in pairs which when multiplied give quadratic polynomials with real coefficients. Since every polynomial with complex coefficients can be factored into 1st-degree factors (that is one way of stating the fundamental theorem of algebra), it follows that every polynomial with real coefficients can be factored into factors of degree no higher than 2: just 1st-degree and quadratic factors.
If the roots are and , they form a quadratic
.
If the third root is , this becomes
.
Corollary on odd-degree polynomials
It follows from the present theorem and the fundamental theorem of algebra that if the degree of a real polynomial is odd, it must have at least one real root.
This can be proved as follows.
Since non-real complex roots come in conjugate pairs, there are an even number of them;
But a polynomial of odd degree has an odd number of roots;
Therefore some of them must be real.
This requires some care in the presence of multiple roots; but a complex root and its conjugate do have the same multiplicity (and this lemma is not hard to prove). It can also be worked around by considering only irreducible polynomials; any real polynomial o
|
https://en.wikipedia.org/wiki/Insulin%20aspart
|
Insulin aspart, sold under the brand name NovoLog, among others, is a modified type of medical insulin used to treat type 1 and type 2 diabetes. It is generally used by injection under the skin but may also be used by injection into a vein. Maximum effect occurs after about 1–3 hours and lasts for 3–5 hours. Generally a longer-acting insulin like insulin NPH is also needed.
Common side effects include low blood sugar, allergic reactions, itchiness, and pain at the site of injection. Other serious side effects may include low blood potassium. Use in pregnancy and breastfeeding is generally safe. It works the same as human insulin by increasing the amount of glucose that tissues take in and decreasing the amount of glucose made by the liver. It is a manufactured form of human insulin; where a single amino acid has been changed, specifically a proline with an aspartic acid at the B28 position.
Insulin aspart was approved for medical use in the United States in 2000. In 2020, it was the 80th most commonly prescribed medication in the United States, with more than 9million prescriptions. Manufacturing involves yeast, which have had the gene for insulin aspart put into their genome. This yeast then makes the insulin, which is harvested from the bioreactor.
Medical uses
Insulin aspart is indicated to improve glycemic control in people with diabetes.
Onset of action
The onset of action is approximately fifteen minutes, the peak action is reached in 45–90 minutes, and the duration is 3–5 hours.
Side effects
The safety of insulin aspart in people with diabetes is no different from for regular insulin. The side effects that are commonly associated with insulin therapy include: allergic reactions, injection site irritation, rashes, and hypoglycemia. The most common side effect is hypoglycemia. Long-term use of insulin, including insulin aspart, can cause lipodystrophy at the site of repeated injections or infusion. To reduce the risk of lipodystrophy, rotate the injecti
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.