source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Pulasan
Nephelium ramboutan-ake, the pulasan, is a tropical fruit in the soapberry family Sapindaceae. It is closely related to the rambutan and sometimes confused with it. Other related soapberry family fruits include lychee and longan. Usually eaten fresh, it is sweeter than the rambutan and lychee, but very rare outside Southeast Asia. Description The name pulasan comes from the Malay word pulas (twist), related to the Tagalog pilas (rip, tear). The fruit is opened through the act of twisting the fruit with both hands, thus the name pulasan. The pulasan tree is an ornamental. It attains a height of 10–15 m and has a short trunk to 30–40 cm thick. The branchlets are brown and hairy when young. The alternate leaves, pinnate or odd-pinnate, are 17–45 cm long, have 2 to 5 pairs of opposite or nearly opposite leaflets, are oblong or elliptic-lanceolate, 6.25-17.5 cm long and up to 5 cm wide; slightly wavy, dark-green and barely glossy on the upper surface; pale, and somewhat bluish, with a few short, silky hairs on the underside. Very small, greenish, petalless flowers with 4-5 hairy sepals are borne singly or in clusters on the branches of the erect, axillary or terminal, panicles clothed with fine yellowish or brownish hairs. The pulasan is ultra-tropical and thrives only in humid regions between 360 and 1,150 ft (110–350 m) of altitude. In Malaysia, it is said that the tree bears best after a long, dry season. The fruit is ovoid, 5-7.5 cm long, dark red, with its thick, leathery rind closely set with conical, blunt-tipped tubercles or thick, fleshy, straight spines, which are up to 1 cm long. There may be one or two small, undeveloped fruits nestled close to the stem. Within is the glistening, white or yellowish-white flesh (aril) to 1 cm thick, more or less clinging to the thin, grayish-brown seedcoat (testa) which separates from the seed. The flavor is generally much sweeter than that of the rambutan. The seed is ovoid, oblong or ellipsoid, light brown, somewhat flat
https://en.wikipedia.org/wiki/Linear%20response%20function
A linear response function describes the input-output relationship of a signal transducer, such as a radio turning electromagnetic waves into music or a neuron turning synaptic input into a response. Because of its many applications in information theory, physics and engineering there exist alternative names for specific linear response functions such as susceptibility, impulse response or impedance; see also transfer function. The concept of a Green's function or fundamental solution of an ordinary differential equation is closely related. Mathematical definition Denote the input of a system by (e.g. a force), and the response of the system by (e.g. a position). Generally, the value of will depend not only on the present value of , but also on past values. Approximately is a weighted sum of the previous values of , with the weights given by the linear response function : The explicit term on the right-hand side is the leading order term of a Volterra expansion for the full nonlinear response. If the system in question is highly non-linear, higher order terms in the expansion, denoted by the dots, become important and the signal transducer cannot adequately be described just by its linear response function. The complex-valued Fourier transform of the linear response function is very useful as it describes the output of the system if the input is a sine wave with frequency . The output reads with amplitude gain and phase shift . Example Consider a damped harmonic oscillator with input given by an external driving force , The complex-valued Fourier transform of the linear response function is given by The amplitude gain is given by the magnitude of the complex number and the phase shift by the arctan of the imaginary part of the function divided by the real one. From this representation, we see that for small the Fourier transform of the linear response function yields a pronounced maximum ("Resonance") at the frequency . The linear response func
https://en.wikipedia.org/wiki/Biohydrometallurgy
Biohydrometallurgy is a technique in the world of metallurgy that utilizes biological agents (bacteria) to recover and treat metals such as copper. Modern biohydrometallurgy advances started with the bioleaching of copper more efficiently in the 1950's Important Definitions Bio: Shortened form of Biology; refers to usage of bacteria. Hydro: Term referring to the usage of water; process occurs in aqueous environments Metallurgy: A process involving the separating and refining of metals from other substances; Bioleaching: Using biological agents (bacteria) to extract metals or soils; general term used to encompass all forms biotechnological forms of extraction (hydrometallurgy, biohydrometallurgy, biomining, etc) General Information Interdisciplinary field involving processes that make use of microbes, usually bacteria and archaea mainly take place in aqueous environment deal with metal production and treatment of metal containing materials and solutions "Biohydrometallurgy may generally referred to as the branch of biotechnology dealing with the study and application of the economic potential of the interactions between microbes and minerals. It concerns, thus, all those engaged, directly or indirectly, in the exploitation of mineral resources and in environmental protection: geologists, economic geologists, mining engineers, metallurgists, hydrometallurgists, chemists and chemical engineers. In addition to these specialists, there are the microbiologists whose work is indispensable in the design, implementation and running of biohydrometallurgical processes." Biohydrometallurgy was first used more than 300 years ago to recover copper. The uses have evolved to extracting gold, uranium, and other metals. Hydrometallurgy Hydrometallurgy refers to a specific process involving the chemical properties of water to create an aqueous solution for metal extraction through a series of chemical reactions Biohydrometallurgy as a Science Biohydrometallurgy repre
https://en.wikipedia.org/wiki/Confirm%20Project
CONFIRM was an ambitious IT project supposed to create a single computer reservations system/global distribution system used by airline, rental car, and hotel companies. It is often used as a case study as an example of a major failure in project management. History The system was pursued to develop in order to create synergies between AMR, Marriott, Hilton Hotels Corporation and Budget Rent-A-Car and fully integrate and unify the reservation systems of the companies involved. In 1988 the four large corporations made contracts to complete the system by June 1992 project at a cost of $55 million. Unfortunately, the project turned out to be much more complex to finish than the partners had anticipated. They all had high hopes based that AMR would expand on the extremely successful SABRE computer reservation system that had helped American Airlines create a sustainable competitive advantage after the deregulation of the airline industry. In April 1992, just three months before the system was intended to go live, Confirm failed tests at Los Angeles-based Hilton. AMR also told its partners that it needed another 15 to 18 months to complete the system. The project was never completed. In the process more than 500 technical personnel worked on the project; when the partners disbanded the project in July 1992 they had spent three and a half years and $125 million on the project. The technical complexity of this project was extreme. CONFIRM runs on two IBM 3090 mainframes. One houses the central reservations system, which runs under Transaction Processing Facility. The other mainframe houses a DB2 relational database in an MVS (an IBM mainframe operating system) environment. The database contains decision-support information such as customer histories and pricing data. The system required application-to-application bridging between the two mainframes (CPUs/IBM 3090) for some 60 applications. The main problem was to tie CONFIRM's transaction-processing facility-based cent
https://en.wikipedia.org/wiki/Elasticity%20of%20cell%20membranes
A cell membrane defines a boundary between a cell and its environment. The primary constituent of a membrane is a phospholipid bilayer that forms in a water-based environment due to the hydrophilic nature of the lipid head and the hydrophobic nature of the two tails. In addition there are other lipids and proteins in the membrane, the latter typically in the form of isolated rafts. Of the numerous models that have been developed to describe the deformation of cell membranes, a widely accepted model is the fluid mosaic model proposed by Singer and Nicolson in 1972. In this model, the cell membrane surface is modeled as a two-dimensional fluid-like lipid bilayer where the lipid molecules can move freely. The proteins are partially or fully embedded in the lipid bilayer. Fully embedded proteins are called integral membrane proteins because they traverse the entire thickness of the lipid bilayer. These communicate information and matter between the interior and the exterior of the cell. Proteins that are only partially embedded in the bilayer are called peripheral membrane proteins. The membrane skeleton is a network of proteins below the bilayer that links with the proteins in the lipid membrane. Elasticity of closed lipid vesicles The simplest component of a membrane is the lipid bilayer which has a thickness that is much smaller than the length scale of the cell. Therefore, the lipid bilayer can be represented by a two-dimensional mathematical surface. In 1973, based on similarities between lipid bilayers and nematic liquid crystals, Helfrich proposed the following expression for the curvature energy per unit area of the closed lipid bilayer where are bending rigidities, is the spontaneous curvature of the membrane, and and are the mean and Gaussian curvature of the membrane surface, respectively. The free energy of a closed bilayer under the osmotic pressure (the outer pressure minus the inner one) as: where dA and dV are the area element of the
https://en.wikipedia.org/wiki/Hypersensitive%20site
In genetics a hypersensitive site is a short region of chromatin and is detected by its super sensitivity to cleavage by DNase I and other various nucleases (DNase II and micrococcal nucleases). In a hypersensitive site, the nucleosomal structure is less compacted, increasing the availability of the DNA to binding by proteins, such as transcription factors and DNase I. These sites account for many inherited tendencies. Location Hypersensitive sites are found on every active gene, and many of these genes often have more than one hypersensitive site. Most often, hypersensitive sites are found only in chromatin of cells in which the associated gene is being expressed, and do not occur when the gene is inactive. In DNA being transcribed, 5'hypersensitive sites appear before transcription begins, and the DNA sequences within the hypersensitive sites are required for gene expression. Note: hypersensitive sites precede active promoters. Hypersensitive sites are generated as a result of the binding of transcription factors that displace histone octamers. They can also be located by indirect end labelling. A fragment of DNA is cut once at the hypersensitive site with DNase and at another site with a restriction enzyme. The distance from the known restriction site to the DNase cut is then measured to give the location.
https://en.wikipedia.org/wiki/Trans-acting
In the field of molecular biology, trans-acting (trans-regulatory, trans-regulation), in general, means "acting from a different molecule" (i.e., intermolecular). It may be considered the opposite of cis-acting (cis-regulatory, cis-regulation), which, in general, means "acting from the same molecule" (i.e., intramolecular). In the context of transcription regulation, a trans-acting factor is usually a regulatory protein that binds to DNA. The binding of a trans-acting factor to a cis-regulatory element in DNA can cause changes in transcriptional expression levels. microRNAs or other diffusible molecules are also examples of trans-acting factors that can regulate target sequences. The trans-acting gene may be on a different chromosome to the target gene, but the activity is via the intermediary protein or RNA that it encodes. Cis-acting elements, on the other hand, do not code for protein or RNA. Both the trans-acting gene and the protein/RNA that it encodes are said to "act in trans" on the target gene. See also Trans-regulatory element Transactivation Transrepression
https://en.wikipedia.org/wiki/Journal%20of%20Number%20Theory
The Journal of Number Theory (JNT) is a bimonthly peer-reviewed scientific journal covering all aspects of number theory. The journal was established in 1969 by R.P. Bambah, P. Roquette, A. Ross, A. Woods, and H. Zassenhaus (Ohio State University). It is currently published monthly by Elsevier and the editor-in-chief is Dorian Goldfeld (Columbia University). According to the Journal Citation Reports, the journal has a 2020 impact factor of 0.72.
https://en.wikipedia.org/wiki/Camillo%20Olivetti
Samuel David Camillo Olivetti (August 13, 1868 – December 1943) was an Italian electrical engineer and founder of Olivetti & Co., SpA., the Italian manufacturer of computers, printers and other business machines. The company was later run by his son Adriano. Biography Samuel David Camillo Olivetti was born in 1868 in a bourgeois Jewish family in Ivrea, Piedmont. His father, Salvador Benedetto, was a textile trader and his mother, Elvira Sacerdoti, that was from Modena, was bankers' daughter. From his father, Camillo Olivetti received the entrepreneurial style and the love for the progress, while from his mother the love for languages (Elvira spoke four languages). His cousin was the painter Raffaele Pontremoli. when Camillo was one years old, his father died. His mother looked after him, who was sent to the boarding school of «Calchi Taeggi» in Milan. At the end of high school, he enrolled at the Royal Italian Industrial Museum (later Politecnico di Torino from 1906) and at the Technical Application School, where he attended electrotechnics courses held by Galileo Ferraris. He graduated on the 31st December 1891 in industrial engineering, Camillo needed to improve his English and, on the other, to gain useful work experience. He stayed over a year in London where he worked in an industry that produced measuring instruments for electrical quantities, also doing the mechanic. Upon his return to Turin, he became Ferraris's assistant. In 1893 he accompanied his teacher to United States of America, who had been invited to lecture at the International Congress of Electrotechnics in Chicago. Olivetti acted as his interpreter. Together they visited the Thomas A. Edison laboratories at Llewellyn Park, New Jersey, where they met the brilliant American inventor in person. After this meeting, in 1893, Camillo wrote to his brother-in-law Carlo from Chicago: Camillo continued the journey from Chicago to San Francisco alone, carefully writing down the things he was discovering
https://en.wikipedia.org/wiki/Cynometra%20cauliflora
Cynometra cauliflora is a small, cauliflorous tree with a thick, heavily branched stem, and rather small flowers, about across, that appear on the stem in clusters. It is known in Indonesian as namu-namu, after the flattened, crescent shaped pods, which look similar to the Indonesian pastry, namu-namu. In the city of Ternate, it is called namo-namo, and ñam-ñam in the Maldives. It is a species of the genus Cynometra, native to Malaysia, and found mainly in northern peninsular Malaysia. Foliage Each leaf is made up of 2 leaflets, which if not examined closely may appear as two leaves. The leaflet is asymmetrical in shape measuring long and wide. They are glossy and smooth, dark green with pale green undersides, and are a rosy pink when young. Fruits The fruits are kidney-shaped, long and . The pod does not split open readily, but a line is visible along the fruit and divides it into two. The texture of the skin is coriaceous and uneven, colored a pale greenish/yellow to brown. The pod is edible, with aromatic and juicy yellow pulp, and relatively large seeds. Propagation The tree is usually grown from seed, although it can be propagated by budding, approach grafting and other methods. External links Singapore National Parks Board
https://en.wikipedia.org/wiki/Microbial%20food%20web
The microbial food web refers to the combined trophic interactions among microbes in aquatic environments. These microbes include viruses, bacteria, algae, heterotrophic protists (such as ciliates and flagellates). In aquatic environments, microbes constitute the base of the food web. Single celled photosynthetic organisms such as diatoms and cyanobacteria are generally the most important primary producers in the open ocean. Many of these cells, especially cyanobacteria, are too small to be captured and consumed by small crustaceans and planktonic larvae. Instead, these cells are consumed by phagotrophic protists which are readily consumed by larger organisms. Viruses can infect and break open bacterial cells and (to a lesser extent), planktonic algae (a.k.a. phytoplankton). Therefore, viruses in the microbial food web act to reduce the population of bacteria and, by lysing bacterial cells, release particulate and dissolved organic carbon (DOC). DOC may also be released into the environment by algal cells. One of the reasons phytoplankton release DOC termed "unbalanced growth" is when essential nutrients (e.g. nitrogen and phosphorus) are limiting. Therefore, carbon produced during photosynthesis is not used for the synthesis of proteins (and subsequent cell growth), but is limited due of a lack of the nutrients necessary for macromolecules. Excess photosynthate, or DOC is then released, or exuded. The microbial loop describes a pathway in the microbial food web where DOC is returned to higher trophic levels via the incorporation into bacterial biomass. See also Microbial cooperation Microbial intelligence Microbial population biology
https://en.wikipedia.org/wiki/Charge-induced%20voltage%20alteration
Charge-induced voltage alteration (CIVA) is a technique which uses a scanning electron microscope to locate open conductors on CMOS integrated circuits. This technique is used in semiconductor failure analysis. Theory of operation The scanning of an electron beam across the surface of the device may result in additional charge buildup on conductors which are disconnected from the remainder of the circuit (floating conductors). If a CMOS device is under active bias, the presence of open conductors may not prevent the circuit from operating at low clock frequencies as the result of quantum tunneling effects. By injecting charge into floating conductors that are operating in this tunneling mode, it is possible to produce additional loading which can be detected by monitoring the power supply current. These changes in supply current may be associated with the visual image of the device at the coordinates at which the change was detected. The result is a scanning electron microscope image which has an overlay of the floating conductors superimposed on it.
https://en.wikipedia.org/wiki/Monell%20Chemical%20Senses%20Center
The Monell Chemical Senses Center is a non-profit independent scientific institute located at the University City Science Center campus in Philadelphia, Pennsylvania. Monell conducts and publishes interdisciplinary basic research on taste, smell, and chemesthesis. History Monell was founded in 1968. The center's mission is to advance knowledge of the mechanisms and functions of the chemical senses. Knowledge gained from Monell’s research is relevant to issues related to public health, national health policy, and quality of life, including studies of obesity, diabetes, hypertension, pediatric health, occupational safety, environmental interactions, and national defense. Monell has a staff of more than 50 scientists and provides research opportunities for local high school and undergraduate students. Situated in Philadelphia’s University City Science Center, the center occupies two buildings with a total of . Monell is operated as a non-profit organization and receives funding from government grants, primarily from the National Institutes of Health through the National Institute on Deafness and Other Communication Disorders and the National Institute of Diabetes and Digestive and Kidney Diseases, as well as from private foundations and unrestricted corporate gifts. Research Selected achievements Characterized the first sweet-tasting protein, “Monellin,” broadening the concept of sweet taste Demonstrated that body odors can signal disease even before appearance of overt symptoms Revealed critical role of perinatal experience in establishing flavor preferences of infants, children, and adults Described role of liver chemosensors in control of appetite and satiety. Established that genetically-determined odortypes provide signals of individual identity. Developed the labeled magnitude scale to reliably measure human sensory perception. Pioneered use of living human tissue to characterize human olfactory and taste cell function. Identified the Sac locus co
https://en.wikipedia.org/wiki/Gordo%20%28monkey%29
Gordo was one of the first monkeys to travel into space. As part of the NASA space program, Gordo, also known as Old Reliable, was launched from Cape Canaveral on December 13, 1958, in the U.S. PGM-19 Jupiter rocket on its AM-13 mission. The rocket would travel over 1,500 miles and reach a height of 310 miles (500 km) before returning to Earth and landing in the South Atlantic. A technical malfunction prevented the capsule's parachute from opening and, despite a short search, neither his body nor the vessel were ever recovered. Background Gordo was a South American species of squirrel monkey, about one foot tall and weighing between 1 and 1.5 kg. He was chosen for space travel because of his species' similar anatomical makeup to man and sensitivity to changes in temperature. Missions using monkeys in space were not new: four monkeys called Albert, as well as monkeys called Patricia and Mike had all flown by 1962. The flight The flight took place at 0353 hours EST when the Jupiter AM-13 rocket containing Gordo was launched from the Atlantic Missile Range at Cape Canaveral. The rocket ascended to a height of 290 miles while travelling 1300 lateral miles. Scientists monitoring the flight found that, aside from a slight slowing of the pulse, Gordo suffered no ill effects from the entry into space or resultant weightlessness. Gordo wore a specially customized space suit and was fitted with instruments, including a thermometer and microphone, to monitor his health. In total the flight lasted 15 minutes, of which Gordo was weightless for 8.3 minutes, as the craft exited and re-entered Earth's atmosphere at a speed of over 10,000 miles per hour. The failure of Gordo's parachute meant that both he and the capsule were lost, despite efforts to recover them. It is believed that Gordo was alive at the time of impact. Legacy Despite the loss of Gordo, the mission was considered a success by NASA. It had gone some way towards alleviating the concerns over how the human bod
https://en.wikipedia.org/wiki/InBev-Baillet%20Latour%20Fund
The Artois-Baillet Latour Foundation is a Belgian non-profit organization which was founded on 1 March 1974. History The foundation was born by the initiative of Count Alfred de Baillet Latour, who was the Director of the Artois Breweries in Leuven, Belgium. He was the last male heir of the old House of Baillet, and was by motherside related to the House of Spoelberch, who own the Artois Company. In 1995 the name was changed in Interbrew-Baillet Latour Foundation and in 2005, the name was changed to InBev-Baillet Latour Fund. The Foundation wants to encourage and reward achievements of outstanding human value in the Arts and Sciences. It can do this by means of Prizes, or by any other means that the Foundation might deem appropriate. InBev-Baillet Latour Health Prize The Artois-Baillet Latour Health Prize was established in 1977 to be awarded periodically to recognize the merits of a person whose work has contributed prominently to the improvement of human health in the fields of metabolic disorders, infectious diseases, neurological diseases, cancer and cardiovascular disease. Since 2000 the prize has been awarded annually. Worth 250,000 euros, it is Belgium's major scientific prize. Since 2005, it has been known as the InBev-Baillet Latour Health Prize. Health Prize Laureates Source: 2018 Professor Laurence Zitvogel, University of Paris-Sud and Professor Guido Kroemer, Paris Descartes University (cancer) 2017 Professor , Zürich University (neurological disorders) 2016 Professor Charles M. Rice, Rockefeller University (infectious diseases) 2015 Professor Bruce M. Spiegelman, Harvard Medical School (metabolic disorders) 2014 Professor , Johns Hopkins University (cardiovascular disease) 2013 Professor Carlo M. Croce, Ohio State University (cancer) 2012 Professor Gero Miesenböck, University of Oxford (neurological disorders) 2011 Professor , Rockefeller University (infectious diseases) 2010 Professor Stephen O'Rahilly, University of Cambridge 2009 Professor K
https://en.wikipedia.org/wiki/List%20of%20network%20protocol%20stacks
This is a list of protocol stack architectures. A protocol stack is a suite of complementary communications protocols in a computer network or a computer bus system. See also Lists of network protocols IEEE 802 Network protocols Communications protocols Network protocol stacks stacks
https://en.wikipedia.org/wiki/Belgian%20Society%20of%20Biochemistry%20and%20Molecular%20Biology
The Belgian Society of Biochemistry and Molecular Biology (BMB) is a Belgian non-profit organization, concerned with biochemistry and molecular biology. The BMB was created, based on an initiative of Marcel Florkin, so a Belgian society could join the new International Union of Biochemistry. The first charter of the society was drafted by Edouard J. Bigwood, Jean Brachet, Christian de Duve, Marcel Florkin, Lucien Massart, Paul Putzeys, Laurent Vandendriessche and Claude Lièbecq. The first general assembly was held on 12 January 1952, and the first President of the society was Marcel Florkin, with Claude Lièbecq as secretary and treasurer. See also BIO.be National Fund for Scientific Research
https://en.wikipedia.org/wiki/Geographic%20routing
Geographic routing (also called georouting or position-based routing) is a routing principle that relies on geographic position information. It is mainly proposed for wireless networks and based on the idea that the source sends a message to the geographic location of the destination instead of using the network address. In the area of packet radio networks, the idea of using position information for routing was first proposed in the 1980s for interconnection networks. Geographic routing requires that each node can determine its own location and that the source is aware of the location of the destination. With this information, a message can be routed to the destination without knowledge of the network topology or a prior route discovery. Approaches There are various approaches, such as single-path, multi-path and flooding-based strategies (see for a survey). Most single-path strategies rely on two techniques: greedy forwarding and face routing. Greedy forwarding tries to bring the message closer to the destination in each step using only local information. Thus, each node forwards the message to the neighbor that is most suitable from a local point of view. The most suitable neighbor can be the one who minimizes the distance to the destination in each step (Greedy). Alternatively, one can consider another notion of progress, namely the projected distance on the source-destination-line (MFR, NFP), or the minimum angle between neighbor and destination (Compass Routing). Not all of these strategies are loop-free, i.e. a message can circulate among nodes in a certain constellation. It is known that the basic greedy strategy and MFR are loop free, while NFP and Compass Routing are not. Greedy forwarding can lead into a dead end, where there is no neighbor closer to the destination. Then, face routing helps to recover from that situation and find a path to another node, where greedy forwarding can be resumed. A recovery strategy such as face routing is necessary to ass
https://en.wikipedia.org/wiki/Paper%20popper
A paper popper is a party prank that is commonly used in schools. There are many variations of a paper popper, but they all involve a folded sheet of paper being gripped and right down. This causes air to be forced into the paper's holes, making the paper flaps pop out in the opposite direction, then making a loud popping noise. According to popular misconception, the sound is supposed to be caused by the paper breaking the sound barrier. In actuality, what happens is that air rushes in a blast wave to fill the vacuum created in the pocket that opens. The sound is described as the crack of a whip, which is an example of breaking the sound barrier, although the causes are different. Most papers work for poppers, although some are louder and some are quieter than others. Using bigger paper makes it louder and the thin paper with the red line and holes (Book paper) also makes it louder. Rectangle paper popper: First, fold the top/bottom of the paper down to the bottom/top holding it landscape. Second, still holding it landscape, take the right/left side and fold it in half It is now done. To use it, take the 2 outer pieces of paper and move them down while holding onto the 2 middle pieces of paper. And then, like every other paper popper, hold it high and then move it quickly downward. Another method is to hold it down low at your side so nobody sees it, then hold it above your head, whip it down and drop it so nobody knows it was you (move away quick) To reset it after using it, unfold it, since it should be folded in half to redo the previous folds. See also Paper plane External links Party Popper Triangle Popper Rectangle paper popper Section 3 Popper, paper Popper, paper Toy instruments and noisemakers
https://en.wikipedia.org/wiki/Negligible%20function
For a similar term, please see negligible set. (disambiguation) In mathematics, a negligible function is a function such that for every positive integer c there exists an integer Nc such that for all x > Nc, Equivalently, we may also use the following definition. A function is negligible, if for every positive polynomial poly(·) there exists an integer Npoly > 0 such that for all x > Npoly History The concept of negligibility can find its trace back to sound models of analysis. Though the concepts of "continuity" and "infinitesimal" became important in mathematics during Newton and Leibniz's time (1680s), they were not well-defined until the late 1810s. The first reasonably rigorous definition of continuity in mathematical analysis was due to Bernard Bolzano, who wrote in 1817 the modern definition of continuity. Later Cauchy, Weierstrass and Heine also defined as follows (with all numbers in the real number domain ): (Continuous function) A function is continuous at if for every , there exists a positive number such that implies This classic definition of continuity can be transformed into the definition of negligibility in a few steps by changing parameters used in the definition. First, in the case with , we must define the concept of "infinitesimal function": (Infinitesimal) A continuous function is infinitesimal (as goes to infinity) if for every there exists such that for all Next, we replace by the functions where or by where is a positive polynomial. This leads to the definitions of negligible functions given at the top of this article. Since the constants can be expressed as with a constant polynomial, this shows that infinitesimal functions are a subset of negligible functions. Use in cryptography In complexity-based modern cryptography, a security scheme is provably secure if the probability of security failure (e.g., inverting a one-way function, distinguishing cryptographically strong pseudorandom bits from truly random
https://en.wikipedia.org/wiki/Distance%20from%20a%20point%20to%20a%20plane
In Euclidean space, the distance from a point to a plane is the distance between a given point and its orthogonal projection on the plane, the perpendicular distance to the nearest point on the plane. It can be found starting with a change of variables that moves the origin to coincide with the given point then finding the point on the shifted plane that is closest to the origin. The resulting point has Cartesian coordinates : . The distance between the origin and the point is . Converting general problem to distance-from-origin problem Suppose we wish to find the nearest point on a plane to the point (), where the plane is given by . We define , , , and , to obtain as the plane expressed in terms of the transformed variables. Now the problem has become one of finding the nearest point on this plane to the origin, and its distance from the origin. The point on the plane in terms of the original coordinates can be found from this point using the above relationships between and , between and , and between and ; the distance in terms of the original coordinates is the same as the distance in terms of the revised coordinates. Restatement using linear algebra The formula for the closest point to the origin may be expressed more succinctly using notation from linear algebra. The expression in the definition of a plane is a dot product , and the expression appearing in the solution is the squared norm . Thus, if is a given vector, the plane may be described as the set of vectors for which and the closest point on this plane to the origin is the vector . The Euclidean distance from the origin to the plane is the norm of this point, . Why this is the closest point In either the coordinate or vector formulations, one may verify that the given point lies on the given plane by plugging the point into the equation of the plane. To see that it is the closest point to the origin on the plane, observe that is a scalar multiple of the vector defining the plane, a
https://en.wikipedia.org/wiki/List%20of%20software%20for%20the%20TRS-80
The TRS-80 series of computers were sold via Radio Shack & Tandy dealers in North America and Europe in the early 1980s. Much software was developed for these computers, particularly the relatively successful Color Computer I, II & III models, which were designed for both home office and entertainment (gaming) uses. A list of software for the TRS-80 computer series appears below. This list includes software that was sold labelled as a Radio Shack or Tandy product.# Note: This List is by no means complete, especially with regards to the earlier non-color computer models. Model I Model II VideoTex Color Computer TRS-80 Color Computer Color Computer 1 & 2 Color Computer 3 Model III Many of these titles also ran on the Model I, as the Model III was designed to be backward-compatible with the Model I. Model 16 & 16B Model 4, 4D & 4P Model 12 MC-10 Model 100 & 102 TRS-80 software
https://en.wikipedia.org/wiki/Compaq%20Evo
The Compaq Evo is a series of business PCs (desktop and laptop) and thin clients made by Compaq and Hewlett-Packard following the 2002 merger. The Evo brand was introduced by Compaq in May 2001 as a business-oriented brand. It replaced the Deskpro brand of desktops and the Armada brand of notebooks. Evo was rebranded as HP Compaq which was used until 2013. It is not to be confused with the later Intel Evo branding for performant laptops. Design The Desktops were small and made to be positioned horizontally instead of vertically so that the monitor could be placed on top to save space. Most featured a sleek silver and black compact design. The early models were shipped with CD-ROM drives but Compaq shipped Evos with CD-RW drives and DVD-ROM drives. The design of some models were only allowed for one CD or DVD drive, but some models had bigger designs for 2 CD or DVD drives. Some models also shipped with a 3½ floppy drive, positioned below the CD or DVD drive. Most models also had 2 USB 2.0 ports in the front for convenience as well as having two in the back for human interface devices and external volumes. Most also had a headphone and microphone jack in the front with a line in and line out in the back. The laptops were a conservative design, described by one reviewer as "The old-school black, squared-off-corner business notebook". Most models had a tough black case reminiscent of IBM's ThinkPad, a midsize 14" or 15" screen and good multimedia capability. Most desktops and some laptops were shipped with Pentium 4 processors and some Laptops were shipped with Centrino platforms. The thin clients were based on the Geode processor family. Distribution The Compaq Evo Desktop computer was a great option for businesses and schools because of its compact and cheap design. It was affordable and had the needed specs for businesses and schools. All Compaq Evo computers were shipped standard with Windows XP or Windows 2000 running on them. The last Evo-branded models
https://en.wikipedia.org/wiki/PAC-1
PAC-1 (first procaspase activating compound) is a synthesized chemical compound that selectively induces apoptosis, in cancerous cells. It was granted orphan drug status by the FDA in 2016. History PAC-1 was discovered in Professor Paul Hergenrother's laboratory at the University of Illinois at Urbana–Champaign during a process that screened many chemicals for anti-tumor potential. This molecule, when delivered to cancer cells, signals the cells to self-destruct by activating an "executioner" protein, procaspase-3. Then, the activated executioner protein begins a cascade of events that destroys the machinery of the cell. In 2011, Vanquish Oncology Inc. was founded to move PAC-1 forward to a human clinical trial. In 2013, Vanquish announced a multimillion-dollar angel investment into the company. In 2015, a phase I clinical trial of PAC-1 opened for enrollment of cancer patients, and in 2016, it was announced that PAC-1 had been granted Orphan Drug Designation for treatment of glioblastoma by the FDA, and in late 2017 a Phase 1b trial began of PAC-1 plus temozolomide for treatment of patients with recurrent glioblastoma or anaplastic astrocytoma. Mechanism of action In cells, the executioner protein, caspase-3, is stored in its inactive form, procaspase-3. This way, the cell can quickly undergo apoptosis by activating the protein that is already there. This inactive form is called a zymogen. Procaspase-3 is known to be inhibited by low levels of zinc. PAC-1 activates procaspase-3 by chelating zinc, thus relieving the zinc-mediated inhibition. This allows procaspase-3 to be an active enzyme, and it can then cleave another molecule of procaspase-3 to active caspase-3. Caspase-3 can further activate other molecules of procaspase-3 in the cell, causing an exponential increase in caspase-3 concentration. PAC-1 facilitates this process and causes the cell to undergo apoptosis quickly. This direct procaspase-3 activation mode-of-action for PAC-1 has been confirmed b
https://en.wikipedia.org/wiki/List%20of%20artificial%20objects%20leaving%20the%20Solar%20System
The artificial objects leaving the Solar System are all space probes and the upper stages of their launch vehicles, all launched by NASA. Three of the probes, Voyager 1, Voyager 2, and New Horizons are still functioning and are regularly contacted by radio communication, while Pioneer 10 and Pioneer 11 are now derelict. In addition to these spacecraft, some upper stages and de-spin weights are leaving the Solar System, assuming they continue on their trajectories. These objects are leaving the Solar System because their velocity and direction are taking them away from the Sun, and at their distance from the Sun, its gravitational pull is not sufficient to pull these objects back or into orbit. They are not impervious to the gravitational pull of the Sun and are being slowed, but are still traveling in excess of escape velocity to leave the Solar System and coast into interstellar space. Planetary exploration probes Pioneer 10 – launched in 1972, flew past Jupiter in 1973 and is heading in the direction of Aldebaran (65 light years away) in the constellation of Taurus. Contact was lost in January 2003, and it is estimated to have passed 134 astronomical units (AU; one AU is roughly the average distance between Earth and the Sun: 150 million kilometers (93 million miles)). Pioneer 11 – launched in 1973, flew past Jupiter in 1974 and Saturn in 1979. Contact was lost in November 1995, and it is estimated to be at 111 AU. The spacecraft is heading toward the constellation of Aquila, northwest of the constellation of Sagittarius. Barring an incident, Pioneer 11 will pass near one of the stars in the constellation in about 4 million years. Voyager 2 – launched in August 1977, flew past Jupiter in 1979, Saturn in 1981, Uranus in 1986, and Neptune in 1989. The probe left the heliosphere for interstellar space at 119 AU on 5 November 2018. Voyager 2 is still active. It is not heading toward any particular star, although in roughly 40,000 years it should pass 1.7 light-yea
https://en.wikipedia.org/wiki/DIGIC
Digital Imaging Integrated Circuit (often styled as "DiG!C") is Canon Inc.'s name for a family of signal processing and control units for digital cameras and camcorders. DIGIC units are used as image processors by Canon in its own digital imaging products. Several generations of DIGICs exist, and are distinguished by a version number suffix. Currently, DIGIC is implemented as an application-specific integrated circuit (ASIC) designed to perform high speed signal processing as well as the control operations in the product in which it has been incorporated. Over its numerous generations, DIGIC has evolved from a system involving a number of discrete integrated circuits to a single chip system, many of which are based around the ARM instruction set. Custom firmware for these units has been developed to add features to the cameras. DIGIC in Cameras Original DIGIC The original DIGIC was used on the PowerShot G3 (Sep 2002), Canon S1 IS (Mar 2004), A520 (Mar 2005), and other cameras. It consists of three separate chips: a video processing IC, an image processing IC and a camera control IC. DIGIC II DIGIC II is a single chip system introduced in 2004, unlike the first DIGIC, that allowed for more compact designs. DIGIC II also improved upon the original by adding a larger buffer and increasing processing speed. It has been used in some advanced consumer-level cameras and many digital SLRs such as Canon EOS 5D and Canon EOS 30D. DIGIC II uses high-speed DDR-SDRAM, which improves startup time and AF performance. It can write to memory card at speeds up to 5.8 MB/sec. Additionally, Canon claims DIGIC II improves color, sharpness, and automatic white balance with its CMOS sensor in its digital SLR camera line. DIGIC III The DIGIC III Image Processor, introduced in 2006, was advertised to deliver improved image quality, faster operation and extended battery life compared to its predecessor. DIGIC III provides a faster interface to the SD memory card for the Canon Powe
https://en.wikipedia.org/wiki/Craig%20plot
The Craig plot, named after Paul N. Craig, is a plot of two substituent parameters (e.g. Hansch-Fujita π constant and sigma constant) used in rational drug design. Two most used forms of a Craig plot are plotting the sigma constants of the Hammett equation versus hydrophobicity plotting the steric terms of the Taft equation against hydrophobicity See also Quantitative structure-activity relationship pKa
https://en.wikipedia.org/wiki/Fault%20injection
In computer science, fault injection is a testing technique for understanding how computing systems behave when stressed in unusual ways. This can be achieved using physical- or software-based means, or using a hybrid approach. Widely studied physical fault injections include the application of high voltages, extreme temperatures and electromagnetic pulses on electronic components, such as computer memory and central processing units. By exposing components to conditions beyond their intended operating limits, computing systems can be coerced into mis-executing instructions and corrupting critical data. In software testing, fault injection is a technique for improving the coverage of a test by introducing faults to test code paths; in particular error handling code paths, that might otherwise rarely be followed. It is often used with stress testing and is widely considered to be an important part of developing robust software. Robustness testing (also known as syntax testing, fuzzing or fuzz testing) is a type of fault injection commonly used to test for vulnerabilities in communication interfaces such as protocols, command line parameters, or APIs. The propagation of a fault through to an observable failure follows a well-defined cycle. When executed, a fault may cause an error, which is an invalid state within a system boundary. An error may cause further errors within the system boundary, therefore each new error acts as a fault, or it may propagate to the system boundary and be observable. When error states are observed at the system boundary they are termed failures. This mechanism is termed the fault-error-failure cycle and is a key mechanism in dependability. History The technique of fault injection dates back to the 1970s when it was first used to induce faults at a hardware level. This type of fault injection is called Hardware Implemented Fault Injection (HWIFI) and attempts to simulate hardware failures within a system. The first experiments in hardw
https://en.wikipedia.org/wiki/Humster
A humster is a hybrid cell line made from a hamster oocyte fertilized with human sperm. This is possible due to the unique promiscuity of hamster ova, which allows them to fuse with non-hamster sperm. It always consists of single cells, and cannot form a multi-cellular being. Humsters are usually destroyed before they divide into two cells; if isolated and left alone to divide, they would still be unviable. Humsters are routinely created mainly for two reasons: To avoid legal issues with working with pure human embryonic stem cell lines. To assess the viability of human sperm for in vitro fertilization Somatic cell hybrids between humans and hamsters or mice have been used for the mapping of various traits since at least the 1970s. See also Hamster zona-free ovum test Human–animal hybrid Recombinant DNA
https://en.wikipedia.org/wiki/Online%20health%20communities
Online health communities are online social networks related to health. They primarily provide a means for patients and their families to learn about illnesses, to seek and offer social support, and to connect with others in similar circumstances. These online groups can be composed of individuals with illnesses, groups of medical professionals with shared interests, non-professional caregivers and family of patients, or a combination. The term "online health community" is primarily academic jargon. Changes in the health care system coupled with increased infiltration and use of the Internet have resulted in heavier reliance on the Internet for disease and health education. Eighty percent of American adult Internet users currently go online to find health care information for themselves and their loved ones, with health searches one of the most popular uses of the Internet. Furthermore, 1 in 4 of people with an illness have gone online to find other people who share similar experiences. Individuals access these communities searching for a variety of resources. Existing medical and health websites, such as WebMD and Health Cloud, have recognized that they will have more visitors if they offer interactive community features such as discussion or Internet forums. Depending on the specific community, patients and medical professionals are able to engage in behaviors such as sharing their illness experiences, exchange knowledge, and increase disease-specific expertise. Even more, these online communities provide users with a breadth of social resources that may be directly beneficial to their health. These communities have been particularly useful in providing emotional and informational support to those with various illnesses such as cancer, HIV/Aids, infertility, diabetes, and other rare illnesses. These communities are distinct from general online communities in that they focus exclusively on health-related topics for those currently navigating the world of dise
https://en.wikipedia.org/wiki/CCL7
Chemokine (C-C motif) ligand 7 (CCL7) is a small cytokine that was previously called monocyte-chemotactic protein 3 (MCP3). CCL7 is a small protein that belongs to the CC chemokine family and is most closely related to CCL2 (previously called MCP1). Genomics In the human genome, CCL7 is encoded by the CCL7 gene which is one of the several chemokine genes clustered on chromosome 17q11.2-q12. This region contains the gene for the MCP subset of CC chemokines. The CCL7 gene has been given the locus symbol SCYA7. The gene consists of three exons and two introns. The first exon contains a 5′-untranslated region (5′-UTR), the information for the signal sequence (23 amino acids), and the mature protein's first two amino acids. The second exon encodes amino acids 3–42 of the mature proteins. The third exon is composed of the C-terminal region of the protein, a 3′-UTR containing one or more destabilizing AU-rich sequences and a polyadenylation signal. Molecular biology CCL7 was first characterized from osteosarcoma supernatant. CCL7 consists of 99 amino acids, which contains 23-amino acid signal peptide. The mature protein about 76 amino acids is secreted after cleavage of the signal peptide. In contrast to most chemokines, CCL7 exists in a general monomeric form, differing from the dimer formed in a highly concentrated solution. CCL7 can exist in four different glycotypes with a molecular weight 11, 13, 17 and 18 kDa in COS cells. CCL7 mediates effects on the immune cell types through binding to numerous receptors, including CCR1, CCR2, CCR3, CCR5, and CCR10. These receptors belongs to the G protein-coupled seven-transmembrane receptors. CCL7 can also interact with cell surface glycosaminoglycans (GAGs) present on all animal cell surfaces. Function CCL7 is expressed in many types of cells, including stromal cells, keratinocytes, airway smooth muscle cells, parenchymal cells, fibroblasts and leukocytes and also in tumor cells. CCL7 mainly acts as a chemoattract
https://en.wikipedia.org/wiki/Disseminated%20disease
Disseminated disease refers to a diffuse disease-process, generally either infectious or neoplastic. The term may sometimes also characterize connective tissue disease. A disseminated infection, for example, has extended beyond its origin or nidus and involved the bloodstream to "seed" other areas of the body. Similarly, one can view metastatic cancer as a disseminated infection in that it has extended into the bloodstream or into the lymphatic system and thus "seeded" distant sites (a process known as metastasis). Disseminated disease is often contrasted with localized disease.
https://en.wikipedia.org/wiki/Localized%20disease
A localized disease is an infectious or neoplastic process that originates in and is confined to one organ system or general area in the body, such as a sprained ankle, a boil on the hand, an abscess of finger. A localized cancer that has not extended beyond the margins of the organ involved can also be described as localized disease, while cancers that extend into other tissues are described as invasive. Tumors that are non-hematologic in origin but extend into the bloodstream or lymphatic system are known as metastatic. Localized diseases are contrasted with disseminated diseases and systemic diseases. Some diseases are capable of changing from local to disseminated diseases. Pneumonia, for example, is generally confined to one or both lungs but can become disseminated through sepsis, in which the microorganism responsible for the pneumonia "seeds" the bloodstream or lymphatic system and is transported to distant sites in the body. When that occurs, the process is no longer described as a localized disease, but rather as a disseminated disease. See also Disease Nosology
https://en.wikipedia.org/wiki/Globalization%20and%20disease
Globalization, the flow of information, goods, capital, and people across political and geographic boundaries, allows infectious diseases to rapidly spread around the world, while also allowing the alleviation of factors such as hunger and poverty, which are key determinants of global health. The spread of diseases across wide geographic scales has increased through history. Early diseases that spread from Asia to Europe were bubonic plague, influenza of various types, and similar infectious diseases. In the current era of globalization, the world is more interdependent than at any other time. Efficient and inexpensive transportation has left few places inaccessible, and increased global trade in agricultural products has brought more and more people into contact with animal diseases that have subsequently jumped species barriers (see zoonosis). Globalization intensified during the Age of Exploration, but trading routes had long been established between Asia and Europe, along which diseases were also transmitted. An increase in travel has helped spread diseases to natives of lands who had not previously been exposed. When a native population is infected with a new disease, where they have not developed antibodies through generations of previous exposure, the new disease tends to run rampant within the population. Etiology, the modern branch of science that deals with the causes of infectious disease, recognizes five major modes of disease transmission: airborne, waterborne, bloodborne, by direct contact, and through vector (insects or other creatures that carry germs from one species to another). As humans began traveling over seas and across lands which were previously isolated, research suggests that diseases have been spread by all five transmission modes. Travel patterns and globalization The Age of Exploration generally refers to the period between the 15th and 17th centuries. During this time, technological advances in shipbuilding and navigation made it e
https://en.wikipedia.org/wiki/Otis%20Boykin
Otis Frank Boykin (August 29, 1920March 26, 1982) was an American inventor and engineer. His inventions include electrical resistors used in computing, missile guidance, and pacemakers. Early life and education Otis Boykin was born on August 29, 1920, in Dallas, Texas. His father, Walter B. Boykin, was a carpenter, and later became a preacher. His mother, Sarah, was a maid, who died of heart failure when Otis was a year old. This inspired him to help improve the pacemaker. Boykin attended Booker T. Washington High School in Dallas, where he was the valedictorian, graduating in 1938. He attended Fisk University on a scholarship, worked as a laboratory assistant at the university's nearby aerospace laboratory, and left in 1941. Boykin then moved to Chicago, where he found work as a clerk at Electro Manufacturing Company. He was subsequently hired as a laboratory assistant for the Majestic Radio and Television Corporation; at that company, he rose to become foreman of their factory. By 1944, he was working for the P.J. Nilsen Research Labs. In 1946–1947, he studied at Illinois Institute of Technology, but dropped out after two years; some sources say it was because he could not afford his tuition, but he later stated he left for an employment opportunity and did not have time to return to finish his degree. One of his mentors was Dr. Denton Deere, an engineer and inventor with his own laboratory. Another mentor was Dr. Hal F. Fruth, with whom he collaborated on several experiments, including a more effective way to test automatic pilot control units in airplanes. The two men later went into business, opening an electronics research lab in the late 1940s. In the 1950s, Boykin and Fruth worked together at the Monson Manufacturing Corporation; Boykin was the company's chief engineer. In the early 1960s, Boykin was a senior project engineer at the Chicago Telephone Supply Corporation, later known as CTS Labs. It was here that he did much of his pacemaker research. But
https://en.wikipedia.org/wiki/Pulse%20compression
Pulse compression is a signal processing technique commonly used by radar, sonar and echography to either increase the range resolution when pulse length is constrained or increase the signal to noise ratio when the peak power and the bandwidth (or equivalently range resolution) of the transmitted signal are constrained. This is achieved by modulating the transmitted pulse and then correlating the received signal with the transmitted pulse. Simple pulse Signal description The ideal model for the simplest, and historically first type of signals a pulse radar or sonar can transmit is a truncated sinusoidal pulse (also called a CW --carrier wave-- pulse), of amplitude and carrier frequency, , truncated by a rectangular function of width, . The pulse is transmitted periodically, but that is not the main topic of this article; we will consider only a single pulse, . If we assume the pulse to start at time , the signal can be written the following way, using the complex notation: Range resolution Let us determine the range resolution which can be obtained with such a signal. The return signal, written , is an attenuated and time-shifted copy of the original transmitted signal (in reality, Doppler effect can play a role too, but this is not important here.) There is also noise in the incoming signal, both on the imaginary and the real channel. The noise is assumed to be band-limited, that is to have frequencies only in (this generally holds in reality, where a bandpass filter is generally used as one of the first stages in the reception chain); we write to denote that noise. To detect the incoming signal, a matched filter is commonly used. This method is optimal when a known signal is to be detected among additive noise having a normal distribution. In other words, the cross-correlation of the received signal with the transmitted signal is computed. This is achieved by convolving the incoming signal with a conjugated and time-reversed version of the transmitted
https://en.wikipedia.org/wiki/Table%20of%20mathematical%20symbols%20by%20introduction%20date
The following table lists many specialized symbols commonly used in modern mathematics, ordered by their introduction date. The table can also be ordered alphabetically by clicking on the relevant header title. See also History of mathematical notation History of the Hindu–Arabic numeral system Glossary of mathematical symbols List of mathematical symbols by subject Mathematical notation Mathematical operators and symbols in Unicode Sources External links RapidTables: Math Symbols List Jeff Miller: Earliest Uses of Various Mathematical Symbols Symbols by introduction date Symbols
https://en.wikipedia.org/wiki/John%20Fairclough
Sir John Whitaker Fairclough (23 August 1930 – 5 June 2003) was a British computer designer, and later government policy advisor. Education John Fairclough was educated at Thirsk Grammar School and then studied electrical engineering at Manchester University, before undertaking national service with the RAF. Career In 1954, he joined the Ferranti computer department and in 1957 he moved to IBM, including working in Raleigh, North Carolina, USA. He returned to the UK to be the Managing Director of IBM Hursley near Winchester in 1974. During 1986–90, Fairclough was Chief Scientific Adviser for the UK Conservative government led by Margaret Thatcher. He left the Cabinet Office and was knighted in 1990. That year, he joined the Board of NM Rothschild and Sons in 1990, becoming Chairman of its venture capital section. He was also involved with a number of start-up companies. He was President of the British Computer Society (1997–98). Personal life He married his first wife, Margaret Harvey, in 1954. After her death in 1996, he married his second wife, Karen, in 2000. He had two sons and a daughter from his first marriage.
https://en.wikipedia.org/wiki/Automoblox
Automoblox is a brand name of a wooden car construction toy designed by Patrick Calello and produced by Automoblox Company, LLC of Cranford, New Jersey. The toy consists of wooden car body sections with patented plastic interconnects, polycarbonate wheels and rubber tires, plastic passengers and polycarbonate screens. Each car can be disassembled into its component parts and re-assembled, parts from different cars can be combined to let children design their own models. It is very commonly used to teach CAD They have also been noted for their extensive use in the PLTW program.
https://en.wikipedia.org/wiki/Undescribed%20taxon
In taxonomy, an undescribed taxon is a taxon (for example, a species) that has been discovered, but not yet formally described and named. The various Nomenclature Codes specify the requirements for a new taxon to be validly described and named. Until such a description has been published, the taxon has no formal or official name, although a temporary, informal name is often used. A published scientific name may not fulfil the requirements of the Codes for various reasons. For example, if the taxon was not adequately described, its name is called a nomen nudum. It is possible for a taxon to be "undescribed" for an extensive period of time, even if unofficial descriptions are published. An undescribed species may be referred to with the genus name, followed by "sp"., but this abbreviation is also used to label specimens or images that are too incomplete to be identified at the species level. In some cases, there is more than one undescribed species in a genus. In this case, these are often referred to by a number or letter. In the shark genus Pristiophorus, for example, there were, for some time, four undescribed species, informally named Pristiophorus sp. A, B, C and D. (In 2008, sp. A was described as Pristiophorus peroniensis and sp. B as P. delicatus.) When a formal description for species C or D is published, its temporary name will be replaced with a proper binomial name. Provisional names in bacteriology In bacteriology, a valid publication of a name requires the deposition of the bacteria in a Bacteriology Culture Collection. Species for which this is impossible cannot receive a valid binomial name; these species are classified as Candidatus. Provisional names in botany A provisional name for a species may consist of the number or of some other designation of a specimen in a herbarium or other collection. It may also consist of the genus name followed by such a specimen identifier or by a provisional specific epithet which is enclosed by quotation marks. In
https://en.wikipedia.org/wiki/Directive%202003/30/EC
Directive 2003/30/EC was a European Union directive for promoting the use of biofuels for EU transport. The directive entered into force in May 2003, and stipulated that national measures must be taken by countries across the EU aiming at replacing 5.75% of all transport fossil fuels (petrol and diesel) with biofuels by 2010. The directive also called for an intermediate target of 2% by 31 December 2005. The target of 5.75% was to be met by 31 December 2010. These percentages were to be calculated on the basis of energy content of the fuel and were to apply to petrol and diesel fuel for transport purposes placed on the markets of member states. Member states were encouraged to take on national "indicative" targets in conformity with the overall target. Directive 2003/30/EC was repealed by Directive 2009/28/EC. Legislation COM(2001) 547, Communication of the European Commission of 7 November 2001 on an Action Plan and two Proposals for Directives to foster the use of Alternative Fuels for Transport, starting with the regulatory and fiscal promotion of biofuels COM(2006) 845, Communication of the European Commission to the Council and the European Parliament: Biofuels Progress Report, that proposes to raise the biofuel target to 10% by 2020. Related legislation: Council Directive 2003/96/EC of 27 October 2003 restructuring the Community framework for the taxation of energy products and electricity COM(2007)18: Proposal for a Directive amending Fuel Quality Directive 98/70/EC (also called Directive relating to the quality of petrol and diesel fuels ), as amended by Directive 2003/17/EC: fuel suppliers should reduce the lifecycle greenhouse gas emissions from transport fuels from 2011 onward, by 1% year. Results A 2007 progress report found that biofuel only held a 1% share, not reaching the target of 2% or the combined goals of the member nations, which was 1.4%. The 2010 target was officially replaced when the Directive 2003/30/EC was repealed by Directive
https://en.wikipedia.org/wiki/MPEG%20Surround
MPEG Surround (ISO/IEC 23003-1 or MPEG-D Part 1), also known as Spatial Audio Coding (SAC) is a lossy compression format for surround sound that provides a method for extending mono or stereo audio services to multi-channel audio in a backwards compatible fashion. The total bit rates used for the (mono or stereo) core and the MPEG Surround data are typically only slightly higher than the bit rates used for coding of the (mono or stereo) core. MPEG Surround adds a side-information stream to the (mono or stereo) core bit stream, containing spatial image data. Legacy stereo playback systems will ignore this side-information while players supporting MPEG Surround decoding will output the reconstructed multi-channel audio. Moving Picture Experts Group (MPEG) issued a call for proposals on MPEG Spatial Audio Coding in March 2004. The group decided that the technology that would be the starting point in standardization process, would be a combination of the submissions from two proponents - Fraunhofer IIS / Agere Systems and Coding Technologies / Philips. The MPEG Surround standard was developed by the Moving Picture Experts Group (ISO/IEC JTC 1/SC29/WG11) and published as ISO/IEC 23003 in 2007. It was the first standard of MPEG-D standards group, formally known as ISO/IEC 23003 - MPEG audio technologies. MPEG Surround was also defined as one of the MPEG-4 Audio Object Types in 2007. There is also the MPEG-4 No Delay MPEG Surround object type (LD MPEG Surround), which was published in 2010. The Spatial Audio Object Coding (SAOC) was published as MPEG-D Part 2 - ISO/IEC 23003–2 in 2010 and it extends MPEG Surround standard by re-using its spatial rendering capabilities while retaining full compatibility with existing receivers. MPEG SAOC system allows users on the decoding side to interactively control the rendering of each individual audio object (e.g. individual instruments, vocals, human voices). There is also the Unified Speech and Audio Coding (USAC) which will be de
https://en.wikipedia.org/wiki/Killian%27s%20dehiscence
Killian's dehiscence (also known as Killian's triangle) is a triangular area in the wall of the pharynx between the cricopharyngeus and thyropharyngeus which are the two parts of the inferior constrictors(also see Pharyngeal pouch). It can be seen as a locus minoris resistentiae. Similar triangular area between circular fibres of cricopharyngeus and longitudinal fibres of esophagus is Lamier's triangle or Lamier-hackermann's area. Clinical significance It represents a potentially weak spot where a pharyngoesophageal diverticulum (Zenker's diverticulum) is more likely to occur. Eponym It is named after the German ENT surgeon Gustav Killian.
https://en.wikipedia.org/wiki/DVB-SH
DVB-SH ("Digital Video Broadcasting - Satellite services to Handhelds") is a physical layer standard for delivering IP based media content and data to handheld terminals such as mobile phones or PDAs, based on a hybrid satellite/terrestrial downlink and for example a GPRS uplink. The DVB Project published the DVB-SH standard in February 2007. The DVB-SH system was designed for frequencies below 3 GHz, supporting UHF band, L Band or S-band. It complements and improves the existing DVB-H physical layer standard. Like its sister specification (DVB-H), it is based on DVB IP Datacast (IPDC) delivery, electronic service guides and service purchase and protection standards. Standard description Architecture There are two physical layers (terrestrial and satellite), what increases de system configuration options. Depending on the transmitted signal modulation we can find two kind of architectures: SH-A and SH-B. SH-A Architecture Both terrestrial and satellite layers use Orthogonal frequency-division multiplexing (OFDM) this modulation solves the multipath problem, where the same token can be received multiple times with a delay due to bounces. Use OFDM in both transmitters enables creating single-frequency networks (SFN), that increases spectral efficiency but forces the terrestrial signal being identical to the transmitted by the satellite. For this reason, is permitted to implement Multi-frequency network (MFN) where terrestrial and satellite components can broadcast through different channels with the same modulation. SH-B Architecture Terrestrial component uses OFDM but satellite uses Time-division multiplexing (TDM). Both components have to broadcast in different frequencies to avoid interference. It is not possible creating SFN but increases satellite´s signal transmission performance because TDM works better with these kinds of transmissions. Comparison with DVB-H The DVB-SH incorporates a number of enhancements when compared to DVB-H: More alternative co
https://en.wikipedia.org/wiki/Noncommutative%20standard%20model
In theoretical particle physics, the non-commutative Standard Model (best known as Spectral Standard Model ), is a model based on noncommutative geometry that unifies a modified form of general relativity with the Standard Model (extended with right-handed neutrinos). The model postulates that space-time is the product of a 4-dimensional compact spin manifold by a finite space . The full Lagrangian (in Euclidean signature) of the Standard model minimally coupled to gravity is obtained as pure gravity over that product space. It is therefore close in spirit to Kaluza–Klein theory but without the problem of massive tower of states. The parameters of the model live at unification scale and physical predictions are obtained by running the parameters down through renormalization. It is worth stressing that it is more than a simple reformation of the Standard Model. For example, the scalar sector and the fermions representations are more constrained than in effective field theory. Motivation Following ideas from Kaluza–Klein and Albert Einstein, the spectral approach seeks unification by expressing all forces as pure gravity on a space . The group of invariance of such a space should combine the group of invariance of general relativity with , the group of maps from to the standard model gauge group . acts on by permutations and the full group of symmetries of is the semi-direct product: Note that the group of invariance of is not a simple group as it always contains the normal subgroup . It was proved by Mather and Thurston that for ordinary (commutative) manifolds, the connected component of the identity in is always a simple group, therefore no ordinary manifold can have this semi-direct product structure. It is nevertheless possible to find such a space by enlarging the notion of space. In noncommutative geometry, spaces are specified in algebraic terms. The algebraic object corresponding to a diffeomorphism is the automorphism of the algebra of co
https://en.wikipedia.org/wiki/Stereocilia
Stereocilia (or stereovilli or villi) are non-motile apical cell modifications. They are distinct from cilia and microvilli, but are closely related to microvilli. They form single "finger-like" projections that may be branched, with normal cell membrane characteristics. They contain actin. Stereocilia are found in the vas deferens, the epididymis, and the sensory cells of the inner ear. Structure Stereocilia are cylindrical and non-motile. They are much longer and thicker than microvilli, form single "finger-like" projections that may be branched, and have more of the characteristics of the cellular membrane proper. Like microvilli, they contain actin and lack an axoneme. This distinguishes them from cilia. They do not have a Basal body at their base since they do not contain microtubules. They may or may not be covered by a glycocalyx coating. They have no fixed arrangement, different to the structure present in kinocilium. Organs containing stereocilia Stereocilia are found in: the vas deferens. the epididymis. Some consider epididymal stereocilia to be a variant of microvilli, rather than their own distinct type of structure. the sensory (hair) cells of the inner ear.
https://en.wikipedia.org/wiki/Non-peptidic%20antigen
Non-peptidic antigens are low-molecular-weight compounds that stimulate human Vγ9/Vδ2 T cells. The most potent activator for Vγ9/Vδ2 T cells is (E)-4-hydroxy-3-methyl-but-2-enyl pyrophosphate (HMB-PP), a natural intermediate of the non-mevalonate pathway of isopentenyl pyrophosphate (IPP) biosynthesis. HMB-PP is an essential metabolite in most pathogenic bacteria including Mycobacterium tuberculosis as well as in malaria parasites, but is absent from the human host. IPP itself is structurally closely related to HMB-PP and ubiquitously present in all living cells (i.e., also in human cells), yet its potency in vitro is 10,000-fold reduced; whether IPP represents a physiological 'danger' signal of stressed or transformed cells is still unclear. Of pharmacological interest and with bioactivities comparable to that of IPP are synthetic aminobisphosphonates such as zoledronate (Zometa) that are widely used to treat osteoporosis and bone metastases, and act as Vγ9/Vδ2 T cell receptor agonists. Certain alkylated amines have been described to activate Vγ9/Vδ2 T cells in vitro, however only at millimolar concentrations, i.e., with potencies 106-fold to 108-fold lower than those of HMB-PP, thereby questioning their physiological relevance.
https://en.wikipedia.org/wiki/Hilbert%27s%20arithmetic%20of%20ends
In mathematics, specifically in the area of hyperbolic geometry, Hilbert's arithmetic of ends is a method for endowing a geometric set, the set of ideal points or "ends" of a hyperbolic plane, with an algebraic structure as a field. It was introduced by German mathematician David Hilbert. Definitions Ends In a hyperbolic plane, one can define an ideal point or end to be an equivalence class of limiting parallel rays. The set of ends can then be topologized in a natural way and forms a circle. This usage of end is not canonical; in particular the concept it indicates is different from that of a topological end (see End (topology) and End (graph theory)). In the Poincaré disk model or Klein model of hyperbolic geometry, every ray intersects the boundary circle (also called the circle at infinity or line at infinity) in a unique point, and the ends may be identified with these points. However, the points of the boundary circle are not considered to be points of the hyperbolic plane itself. Every hyperbolic line has exactly two distinct ends, and every two distinct ends are the ends of a unique line. For the purpose of Hilbert's arithmetic, it is expedient to denote a line by the ordered pair (a, b) of its ends. Hilbert's arithmetic fixes arbitrarily three distinct ends, and labels them as 0, 1, and ∞ ;. The set H on which Hilbert defines a field structure is the set of all ends other than ∞, while H denotes the set of all ends including ∞. Addition Hilbert defines the addition of ends using hyperbolic reflections. For every end x in H, its negation −x is defined by constructing the hyperbolic reflection of line (x,∞) across the line (0,∞), and choosing −x to be the end of the reflected line. The composition of any three hyperbolic reflections whose axes of symmetry all share a common end is itself another reflection, across another line with the same end. Based on this "three reflections theorem", given any two ends x and y in H, Hilbert defines the sum x + y
https://en.wikipedia.org/wiki/Serial%20manipulator
Serial manipulators are the most common industrial robots and they are designed as a series of links connected by motor-actuated joints that extend from a base to an end-effector. Often they have an anthropomorphic arm structure described as having a "shoulder", an "elbow", and a "wrist". Serial robots usually have six joints, because it requires at least six degrees of freedom to place a manipulated object in an arbitrary position and orientation in the workspace of the robot. A popular application for serial robots in today's industry is the pick-and-place assembly robot, called a SCARA robot, which has four degrees of freedom. Structure In its most general form, a serial robot consists of a number of rigid links connected to joints. Simplicity considerations in manufacturing and control have led to robots with only revolute or prismatic joints and orthogonal, parallel and/or intersecting joint axes (instead of arbitrarily placed joint axes). Donald L. Pieper derived the first practically relevant result in this context, referred to as 321 kinematic structure: The inverse kinematics of serial manipulators with six revolute joints, and with three consecutive joints intersecting, can be solved in closed-form, i.e. analytically This result had a tremendous influence on the design of industrial robots. The main advantage of a serial manipulator is a large workspace with respect to the size of the robot and the floor space it occupies. The main disadvantages of these robots are: the low stiffness inherent to an open kinematic structure, errors are accumulated and amplified from link to link, the fact that they have to carry and move the large weight of most of the actuators, and the relatively low effective load that they can manipulate. Kinematics The position and orientation of a robot's end effector are derived from the joint positions by means of a geometric model of the robot arm. For serial robots, the mapping from joint positions to end-effector p
https://en.wikipedia.org/wiki/MaverickCrunch
The MaverickCrunch is a floating point math coprocessor core intended for digital audio. It was first presented by Cirrus Logic in June 2000 together with an ARM920T integer processor in their 200 MHz EP9302 EP9307 EP9312 and EP9315 System-on-Chip integrated circuits. Plagued with hardware bugs and poor compiler support, it was seldom used in any of the devices based on those chips and the product line was discontinued on April 1, 2008. Features The coprocessor has 16 64-bit registers which can be used for 32- or 64-bit integer and floating point operations and its floating point format is based on the IEEE-754 standard. It has its own instruction set which performs floating point addition, subtraction, multiplication, negation, absolute value, and comparisons as well as addition, multiplication and bit shifts on integers. It also has four 72-bit registers on which can perform a 32-bit multiply-and-accumulate instruction and a status register, as well as conversions between integer and floating point values and instructions to move data between itself and the ARM registers or memory. It operates in parallel with the main processor, both processors receiving their instructions from a single 32-bit instruction stream. Thus, to use it efficiently, integer and floating point instructions must be interleaved so as to keep both processors busy. Hardware bugs Five versions of the EP93xx silicon were issued: "D0" and "D1"/"E0"/"E1" and "E2", with major revisions to the MaverickCrunch core between D0 and D1 to fix its worst bugs. All have a dozen or more hardware bugs which either give imprecise or garbage results or clobber registers or memory when certain sequences of instructions are executed in a certain order. Compiler support A set of patches was submitted to the GNU Compiler Collection by Red Hat/Cygnus Solutions in 2003 to include a code generator for it, complete with flags to work around its defects. Unfortunately these never worked well enough for it to be
https://en.wikipedia.org/wiki/Homestead%20Technologies
Homestead Technologies is a web hosting company based in Burlington, Massachusetts. Homestead offers its members WYSIWIG tools to build and publish their own websites. Since its founding in 1997 as a free service provider, Homestead has expanded the scope of its services to include online marketing, paid search ads, SEO tools and e-commerce services. Homestead.com, which launched in June 1998, enables Internet users to build a website, with focus on small businesses, e-commerce, and their related services, such as consulting, listings and online business directories. History Founded in October 1997 by Justin S. Kitch and Thai Bui, in Menlo Park, California, Homestead quickly became popular due to its accessibility and ease of use. Its proprietary drag-and-drop SiteBuilder platform enabled users with no prior knowledge of coding or web programming to create fully functioning websites for personal, educational or business use. Homestead is geared mostly towards small business owners, non-profit administrators, retailers, and hobbyists, unlike competitors such as Angelfire and Yahoo! GeoCities, who target more casual users. By October 1999 member registrations reached two million. In October 2003, Homestead Technologies launched PhotoSite, a photo sharing and web hosting service tailored to amateur digital photography. In March 2005, PhotoSite was sold to United Online. In December 2007, Homestead was acquired by financial and tax preparation software giant Intuit for $170 million to bolster its web hosting offerings. On August 16, 2012, Intuit announced that Homestead would be purchased by web hosting company Endurance International Group of Burlington, Massachusetts. See also Tripod (web hosting) WordPress
https://en.wikipedia.org/wiki/Van%20der%20Waerden%20number
Van der Waerden's theorem states that for any positive integers r and k there exists a positive integer N such that if the integers {1, 2, ..., N} are colored, each with one of r different colors, then there are at least k integers in arithmetic progression all of the same color. The smallest such N is the van der Waerden number W(r, k). Tables of Van der Waerden numbers There are two cases in which the van der Waerden number W(r, k) is easy to compute: first, when the number of colors r is equal to 1, one has W(1, k) = k for any integer k, since one color produces only trivial colorings RRRRR...RRR (for the single color denoted R). Second, when the length k of the forced arithmetic progression is 2, one has W(r, 2) = r + 1, since one may construct a coloring that avoids arithmetic progressions of length 2 by using each color at most once, but using any color twice creates a length-2 arithmetic progression. (For example, for r = 3, the longest coloring that avoids an arithmetic progression of length 2 is RGB.) There are only seven other van der Waerden numbers that are known exactly. The table below gives exact values and bounds for values of W(r, k); values are taken from Rabung and Lotts except where otherwise noted. {| class="wikitable" ! k\r ! 2 colors ! 3 colors ! 4 colors ! 5 colors ! 6 colors |- | 3 | style="text-align:right;"| 9 | style="text-align:right;"| 27   | style="text-align:right;"| 76   | style="text-align:right;"| >170   | style="text-align:right;"| >223   |- | 4 | style="text-align:right;" | 35 | style="text-align:right;"| 293   | style="text-align:right;"| >1,048   | style="text-align:right;"| >2,254   | style="text-align:right;"| >9,778   |- | 5 | style="text-align:right;"| 178 | style="text-align:right;"| >2,173   | style="text-align:right;"| >17,705   | style="text-align:right;"| >98,740   | style="text-align:right;"| >98,748   |- | 6 | style="text-align:right;"| 1,132 | style="text-align:right;"| >11,191   | style="text-align:right;"|
https://en.wikipedia.org/wiki/Sudoku%20solving%20algorithms
A standard Sudoku contains 81 cells, in a 9×9 grid, and has 9 boxes, each box being the intersection of the first, middle, or last 3 rows, and the first, middle, or last 3 columns. Each cell may contain a number from one to nine, and each number can only occur once in each row, column, and box. A Sudoku starts with some cells containing numbers (clues), and the goal is to solve the remaining cells. Proper Sudokus have one solution. Players and investigators use a wide range of computer algorithms to solve Sudokus, study their properties, and make new puzzles, including Sudokus with interesting symmetries and other properties. There are several computer algorithms that will solve 9×9 puzzles (=9) in fractions of a second, but combinatorial explosion occurs as increases, creating limits to the properties of Sudokus that can be constructed, analyzed, and solved as increases. Techniques Backtracking Some hobbyists have developed computer programs that will solve Sudoku puzzles using a backtracking algorithm, which is a type of brute force search. Backtracking is a depth-first search (in contrast to a breadth-first search), because it will completely explore one branch to a possible solution before moving to another branch. Although it has been established that approximately 5.96 x 1126 final grids exist, a brute force algorithm can be a practical method to solve Sudoku puzzles. A brute force algorithm visits the empty cells in some order, filling in digits sequentially, or backtracking when the number is found to be not valid. Briefly, a program would solve a puzzle by placing the digit "1" in the first cell and checking if it is allowed to be there. If there are no violations (checking row, column, and box constraints) then the algorithm advances to the next cell and places a "1" in that cell. When checking for violations, if it is discovered that the "1" is not allowed, the value is advanced to "2". If a cell is discovered where none of the 9 digits is allowed
https://en.wikipedia.org/wiki/Etherloop
Etherloop is a kind of DSL technology that combines the features of Ethernet and DSL. It allows the combination of voice and data transmission on standard phone lines. Under the right conditions it will allow speeds of up to 6 megabits per second over a distance of up to 6.4 km (21,000 feet). Etherloop uses half-duplex transmission, and as such, is less susceptible to interference caused by poor line quality, bridge taps, etc. Also, etherloop modems can train up through line filters (although it is not recommended to do this). Etherloop has been deployed by various internet service providers in areas where the loop length is very long or line quality is poor. Some Etherloop modems (those made by Elastic Networks) offer a "Central Office mode", in which two modems are connected back to back over a phone line and used as a LAN extension. An example of a situation where this would be done is to extend Ethernet to a building that is too far to reach with straight Ethernet. See also Ethernet in the first mile (especially 2BASE-TL)
https://en.wikipedia.org/wiki/Canned%20tea
Canned tea is a relatively recent method of marketing tea which has been sold traditionally as leaf tea and also, for the last 100 years, in tea bag form. It utilises the canning process to produce a ready made drink. Perceived advantages includes ease of use (minimal or no preparation time) and the possibility of additives (such as flavours or sugar); the disadvantages are the cost of shipment (and therefore the price of the product) and a lack of freshness. Tea Tea is a beverage made by steeping processed leaves, buds, or twigs of the plant Camellia sinensis in hot water for a few minutes. The processing can include oxidation (called "fermentation" in the tea industry), heating, drying and the addition of herbs, flowers, spices and fruits. There are four main types of tea: black, oolong, green, and white. Tea is a natural source of caffeine, theophylline, theanine, and antioxidants; but it has almost no fat, carbohydrates, or protein Tea has been consumed in China for around 5,000 years. Canning Canning is a method of preserving food by heating it to a temperature that destroys contaminating microorganisms, and then sealing it in air-tight jars, cans, or pouches. Patented in the UK in 1810, this method of preservation was not used extensively for soft drinks until the pull tab version was patented in the USA in 1963 History In 1981, the first canned tea product, unsweetened oolong, was introduced in Japan by Ito En. The introduction very quickly led to more than one hundred variations, offered both in cans and in bottles. The oolong variety was well suited as the canned tea proved stable even at temperatures up to 50° - 60 °C in vending machines. The tea also appealed to consumers inclined to use vending machines for convenient access "on the street." Further, there are perceived health benefits to canned tea, as opposed to other canned soft drinks, due to tea's portrayal as an aid to maintaining a slim body build. The introduction of oolong was quickly fol
https://en.wikipedia.org/wiki/Stress%20testing%20%28software%29
Stress testing is a software testing activity that determines the robustness of software by testing beyond the limits of normal operation. Stress testing is particularly important for "mission critical" software, but is used for all types of software. Stress tests commonly put a greater emphasis on robustness, availability, and error handling under a heavy load, than on what would be considered correct behavior under normal circumstances. A system stress test refers to tests that put a greater emphasis on robustness, availability, and error handling under a heavy load, rather than on what would be considered correct behavior under normal circumstances. In particular, the goals of such tests may be to ensure the software does not crash in conditions of insufficient computational resources (such as memory or disk space), unusually high concurrency, or denial of service attacks. Examples: A web server may be stress tested using scripts, bots, and various denial of service tools to observe the performance of a web site during peak loads. These attacks generally are under an hour long, or until a limit in the amount of data that the web server can tolerate is found. Stress testing may be contrasted with load testing: Load testing examines the entire environment and database, while measuring the response time, whereas stress testing focuses on identified transactions, pushing to a level so as to break transactions or systems. During stress testing, if transactions are selectively stressed, the database may not experience much load, but the transactions are heavily stressed. On the other hand, during load testing the database experiences a heavy load, while some transactions may not be stressed. System stress testing, also known as stress testing, is loading the concurrent users over and beyond the level that the system can handle, so it breaks at the weakest link within the entire system. Field experience Failures may be related to: characteristics of non-production
https://en.wikipedia.org/wiki/IEEE%20802.1AE
IEEE 802.1AE (also known as MACsec) is a network security standard that operates at the medium access control layer and defines connectionless data confidentiality and integrity for media access independent protocols. It is standardized by the IEEE 802.1 working group. Details Key management and the establishment of secure associations is outside the scope of 802.1AE, but is specified by 802.1X-2010. The 802.1AE standard specifies the implementation of a MAC Security Entities (SecY) that can be thought of as part of the stations attached to the same LAN, providing secure MAC service to the client. The standard defines MACsec frame format, which is similar to the Ethernet frame, but includes additional fields: Security Tag, which is an extension of the EtherType Message authentication code (ICV) Secure Connectivity Associations that represent groups of stations connected via unidirectional Secure Channels Security Associations within each secure channel. Each association uses its own key (SAK). More than one association is permitted within the channel for the purpose of key change without traffic interruption (standard requires devices to support at least two) A default cipher suite of GCM-AES-128 (Galois/Counter Mode of Advanced Encryption Standard cipher with 128-bit key) GCM-AES-256 using a 256 bit key was added to the standard 5 years later. Security tag inside each frame in addition to EtherType includes: association number within the channel packet number to provide unique initialization vector for encryption and authentication algorithms as well as protection against replay attack optional LAN-wide secure channel identifier (not required on point-to-point links). The IEEE 802.1AE (MACsec) standard specifies a set of protocols to meet the security requirements for protecting data traversing Ethernet LANs. MACsec allows unauthorized LAN connections to be identified and excluded from communication within the network. In common with IPsec and TLS, M
https://en.wikipedia.org/wiki/Evolution%20as%20fact%20and%20theory
Many scientists and philosophers of science have described evolution as fact and theory, a phrase which was used as the title of an article by paleontologist Stephen Jay Gould in 1981. He describes fact in science as meaning data, not known with absolute certainty but "confirmed to such a degree that it would be perverse to withhold provisional assent". A scientific theory is a well-substantiated explanation of such facts. The facts of evolution come from observational evidence of current processes, from imperfections in organisms recording historical common descent, and from transitions in the fossil record. Theories of evolution provide a provisional explanation for these facts. Each of the words evolution, fact and theory has several meanings in different contexts. In biology, evolution refers to observed changes in organisms over successive generations, to their descent from a common ancestor, and at a technical level to a change in gene frequency over time; it can also refer to explanatory theories (such as Charles Darwin's theory of natural selection) which explain the mechanisms of evolution. To a scientist, fact can describe a repeatable observation capable of great consensus; it can refer to something that is so well established that nobody in a community disagrees with it; and it can also refer to the truth or falsity of a proposition. To the public, theory can mean an opinion or conjecture (e.g., "it's only a theory"), but among scientists it has a much stronger connotation of "well-substantiated explanation". With this number of choices, people can often talk past each other, and meanings become the subject of linguistic analysis. Evidence for evolution continues to be accumulated and tested. The scientific literature includes statements by evolutionary biologists and philosophers of science demonstrating some of the different perspectives on evolution as fact and theory. Evolution, fact and theory Evolution has been described as "fact and theory"; "
https://en.wikipedia.org/wiki/Ligand%20cone%20angle
In coordination chemistry, the ligand cone angle (θ) is a measure of the steric bulk of a ligand in a transition metal coordination complex. It is defined as the solid angle formed with the metal at the vertex of a cone and the outermost edge of the van der Waals spheres of the ligand atoms at the perimeter of the base of the cone. Tertiary phosphine ligands are commonly classified using this parameter, but the method can be applied to any ligand. The term cone angle was first introduced by Chadwick A. Tolman, a research chemist at DuPont. Tolman originally developed the method for phosphine ligands in nickel complexes, determining them from measurements of accurate physical models. Asymmetric cases The concept of cone angle is most easily visualized with symmetrical ligands, e.g. PR3. But the approach has been refined to include less symmetrical ligands of the type PRR′R″ as well as diphosphines. In such asymmetric cases, the substituent angles' half angles, , are averaged and then doubled to find the total cone angle, θ. In the case of diphosphines, the of the backbone is approximated as half the chelate bite angle, assuming a bite angle of 74°, 85°, and 90° for diphosphines with methylene, ethylene, and propylene backbones, respectively. The Manz cone angle is often easier to compute than the Tolman cone angle: Variations The Tolman cone angle method assumes empirical bond data and defines the perimeter as the maximum possible circumscription of an idealized free-spinning substituent. The metal-ligand bond length in the Tolman model was determined empirically from crystal structures of tetrahedral nickel complexes. In contrast, the solid-angle concept derives both bond length and the perimeter from empirical solid state crystal structures. There are advantages to each system. If the geometry of a ligand is known, either through crystallography or computations, an exact cone angle (θ) can be calculated. No assumptions about the geometry are made, unlike the To
https://en.wikipedia.org/wiki/Complementarity-determining%20region
Complementarity-determining regions (CDRs) are part of the variable chains in immunoglobulins (antibodies) and T cell receptors, generated by B-cells and T-cells respectively, where these molecules bind to their specific antigen. A set of CDRs constitutes a paratope. As the most variable parts of the molecules, CDRs are crucial to the diversity of antigen specificities generated by lymphocytes. Location and structure There are three CDRs (CDR1, CDR2 and CDR3), arranged non-consecutively, on the amino acid sequence of a variable domain of an antigen receptor. Since the antigen receptors are typically composed of two variable domains (on two different polypeptide chains, heavy and light chain), there are six CDRs for each antigen receptor that can collectively come into contact with the antigen. A single antibody molecule has two antigen receptors and therefore contains twelve CDRs total. There are three CDR loops per variable domain in antibodies. Sixty CDRs can be found on a pentameric IgM molecule. Since most sequence variation associated with immunoglobulins and T cell receptors are found in the CDRs, these regions are sometimes referred to as hypervariable regions. Within the variable domain, CDR1 and CDR2 are found in the variable (V) region of a polypeptide chain, and CDR3 includes some of V, all of diversity (D, heavy chains only) and joining (J) regions. CDR3 is the most variable. The tertiary structure of an antibody is important to analyze and design new antibodies. The three-dimensional structures of the non-H3 CDRs of antibodies have been clustered and classified by Chothia et al. and more recently by North et al. Homology modeling is a computational method to build tertiary structures from amino-acid sequences. The so-called H3-rules are empirical rules to build models of CDR3. See also Framework region Hypervariable region
https://en.wikipedia.org/wiki/Reciprocity%20%28network%20science%29
In network science, reciprocity is a measure of the likelihood of vertices in a directed network to be mutually linked. Like the clustering coefficient, scale-free degree distribution, or community structure, reciprocity is a quantitative measure used to study complex networks. Motivation In real network problems, people are interested in determining the likelihood of occurring double links (with opposite directions) between vertex pairs. This problem is fundamental for several reasons. First, in the networks that transport information or material (such as email networks, World Wide Web (WWW), World Trade Web, or Wikipedia ), mutual links facilitate the transportation process. Second, when analyzing directed networks, people often treat them as undirected ones for simplicity; therefore, the information obtained from reciprocity studies helps to estimate the error introduced when a directed network is treated as undirected (for example, when measuring the clustering coefficient). Finally, detecting nontrivial patterns of reciprocity can reveal possible mechanisms and organizing principles that shape the observed network's topology. Definitions Traditional definition A traditional way to define the reciprocity r is using the ratio of the number of links pointing in both directions to the total number of links L With this definition, is for a purely bidirectional network while for a purely unidirectional one. Real networks have an intermediate value between 0 and 1. However, this definition of reciprocity has some defects. It cannot tell the relative difference of reciprocity compared with purely random network with the same number of vertices and edges. The useful information from reciprocity is not the value itself, but whether mutual links occur more or less often than expected by chance. Besides, in those networks containing self-linking loops (links starting and ending at the same vertex), the self-linking loops should be excluded when calculating L. Ga
https://en.wikipedia.org/wiki/Abundance%20%28ecology%29
In ecology, local abundance is the relative representation of a species in a particular ecosystem. It is usually measured as the number of individuals found per sample. The ratio of abundance of one species to one or multiple other species living in an ecosystem is referred to as relative species abundances. Both indicators are relevant for computing biodiversity. A variety of sampling methods are used to measure abundance. For larger animals, these may include spotlight counts, track counts and roadkill counts, as well as presence at monitoring stations. In many plant communities the abundances of plant species are measured by plant cover, i.e. the relative area covered by different plant species in a small plot. Abundance is in simplest terms usually measured by identifying and counting every individual of every species in a given sector. It is common for the distribution of species to be skewed so that a few species take up the bulk of individuals collected. Relative species abundance is calculated by dividing the number of species from one group by the total number of species from all groups. Community ecology These measures are all a part of community ecology. Understanding patterns within a community is easy when the community has a relatively low number of species. However most communities do not have a low number of species. Measuring species abundance allows for understanding of how species are distributed within an ecosystem. For example, salt water marshes have an influx of sea water, causing only a few species which are adapted to be able to survive in both salt and fresh water to be abundant. Inversely in land locked wetlands, the species abundance is more evenly distributed among the species who live within the wetland. In most ecosystems in which abundance has been calculated, most often only a small number of species are abundant, while a large number are pretty rare. These abundant species are often generalists, with many rare species being spe
https://en.wikipedia.org/wiki/Adaptive%20step%20size
In mathematics and numerical analysis, an adaptive step size is used in some methods for the numerical solution of ordinary differential equations (including the special case of numerical integration) in order to control the errors of the method and to ensure stability properties such as A-stability. Using an adaptive stepsize is of particular importance when there is a large variation in the size of the derivative. For example, when modeling the motion of a satellite about the earth as a standard Kepler orbit, a fixed time-stepping method such as the Euler method may be sufficient. However things are more difficult if one wishes to model the motion of a spacecraft taking into account both the Earth and the Moon as in the Three-body problem. There, scenarios emerge where one can take large time steps when the spacecraft is far from the Earth and Moon, but if the spacecraft gets close to colliding with one of the planetary bodies, then small time steps are needed. Romberg's method and Runge–Kutta–Fehlberg are examples of a numerical integration methods which use an adaptive stepsize. Example For simplicity, the following example uses the simplest integration method, the Euler method; in practice, higher-order methods such as Runge–Kutta methods are preferred due to their superior convergence and stability properties. Consider the initial value problem where y and f may denote vectors (in which case this equation represents a system of coupled ODEs in several variables). We are given the function f(t,y) and the initial conditions (a, ya), and we are interested in finding the solution at t = b. Let y(b) denote the exact solution at b, and let yb denote the solution that we compute. We write , where is the error in the numerical solution. For a sequence (tn) of values of t, with tn = a + nh, the Euler method gives approximations to the corresponding values of y(tn) as The local truncation error of this approximation is defined by and by Taylor's theor
https://en.wikipedia.org/wiki/Klung%20Wilhelmy%20Science%20Award
The Klung Wilhelmy Science Award is an annual German award in the field of science, alternating annually between the categories of chemistry and physics. It is bestowed upon outstanding younger German scientists under the age of 40. Previous award names 1973 to 2001 – Otto-Klung-Award 2001 to 2007 – Otto-Klung-Weberbank-Award 2007 to 2013 – Klung-Wilhelmy-Weberbank-Award Selection process The prizewinners are selected by permanent committees at the Institutes of Chemistry and Biochemistry and the Department of Physics at the Free University of Berlin, with additional input from professors at other universities. Proposals and nominations by nationally and internationally renowned scientists are also taken into consideration. Self-nominations will not be accepted. The final decision on the selection recommendations is made by the following foundations: the Otto Klung Foundation at the Free University of Berlin and the Dr. Wilhelmy Foundation. The stated aim of these foundations is to strengthen the promotion of outstanding scientific achievements and to reward internationally accredited innovative approaches. Five of the previously chosen prizewinners later received the Nobel Prize. The prize was first awarded in 1973 by the Otto Klung Foundation. Since 2007, the prize has become one of the highest privately funded scientific endowments in Germany. The annual award ceremony, which has been held in November, is open to the public. Recipients From 1973 to 1978, the Otto Klung Foundation acting alone and trying to foster young academics presented the Otto-Klung-Award as a Junior Researcher Prize for outstanding scientific achievement to graduate students and postdoctoral students of the Free University of Berlin, Departments of Chemistry and Physics: Klaus-Peter Dinse (Physics 1973), Wolf-Dietrich Hunnius and Rolf Minkwitz (Chemistry 1974), Michael Grunze (Chemistry 1975), Günther Kerker (Physics 1976), Wolfgang Lubitz (Chemistry 1977), Andreas Gaupp (Physics 1
https://en.wikipedia.org/wiki/Sequential%20access%20memory
In computing, sequential access memory (SAM) is a class of data storage devices that read stored data in a sequence. This is in contrast to random access memory (RAM) where data can be accessed in any order. Sequential access devices are usually a form of magnetic storage or optical storage. While sequential access memory is read in sequence, arbitrary locations can still be accessed by "seeking" to the requested location. This operation, however, is often relatively inefficient (see seek time, rotational latency). Magnetic sequential access memory is typically used for secondary storage in general-purpose computers due to their higher density at lower cost compared to RAM, as well as resistance to wear and non-volatility. Magnetic tape is a type of sequential access memory still in use; historically, drum memory has also been used. See also Sequential access Basic sequential access method (BSAM) Queued sequential access method (QSAM) Secondary storage Hard disk drive Solid state drive Magnetic storage Magnetic tape Drum memory Locality of reference Streaming media
https://en.wikipedia.org/wiki/Podocin
Podocin is a protein component of the filtration slits of podocytes. Glomerular capillary endothelial cells, the glomerular basement membrane and the filtration slits function as the filtration barrier of the kidney glomerulus. Mutations in the podocin gene NPHS2 can cause nephrotic syndrome, such as focal segmental glomerulosclerosis (FSGS) or minimal change disease (MCD). Symptoms may develop in the first few months of life (congenital nephrotic syndrome) or later in childhood. Structure Podocin is a membrane protein of the band-7-stomatin family, consisting of 383 amino acids. It has a transmembrane domain forming a hairpin structure, with two cytoplasmic ends at the N- and C-terminus, the latter of which interacts with the cytosolic tail of nephrin, with CD2AP serving as an adaptor. Function Podocin is localized on the membranes of podocyte pedicels (foot-like long processes), where it oligomerizes in lipid rafts together with nephrin to form the filtration slits.
https://en.wikipedia.org/wiki/SmartQVT
SmartQVT is a unmaintained (since 2013) full Java open-source implementation of the QTV-Operational language which is dedicated to express model-to-model transformations. This tool compiles QVT transformations into Java programs to be able to run QVT transformations. The compiled Java programs are EMF-based applications. It is provided as Eclipse plug-ins running on top of the EMF metamodeling framework and is licensed under EPL. Components SmartQVT contains 3 main components: a code editor: this component helps the user to write QVT code by highlighting key words. a parser: this component converts QVT code files into model representations of the QVT programs (abstract syntax). a compiler: this component converts model representations of the QVT program into executable Java programs. See also Meta-Object Facility (MOF): a language to write metamodels Model transformation language Model-driven architecture (MDA) Programming tools Free software programmed in Java (programming language)
https://en.wikipedia.org/wiki/Slow-reacting%20substance%20of%20anaphylaxis
The slow-reacting substance of anaphylaxis or SRS-A is a mixture of the leukotrienes LTC4, LTD4 and LTE4. Mast cells secrete it during the anaphylactic reaction, inducing inflammation. It can be found in basophils. It induces prolonged, slow contraction of smooth muscle and has a major bronchoconstrictor role in asthma. Compared to histamine, it is approximately 1000 times more potent and has a slower onset but longer duration of action.
https://en.wikipedia.org/wiki/Dormand%E2%80%93Prince%20method
In numerical analysis, the Dormand–Prince (RKDP) method or DOPRI method, is an embedded method for solving ordinary differential equations (ODE). The method is a member of the Runge–Kutta family of ODE solvers. More specifically, it uses six function evaluations to calculate fourth- and fifth-order accurate solutions. The difference between these solutions is then taken to be the error of the (fourth-order) solution. This error estimate is very convenient for adaptive stepsize integration algorithms. Other similar integration methods are Fehlberg (RKF) and Cash–Karp (RKCK). The Dormand–Prince method has seven stages, but it uses only six function evaluations per step because it has the "First Same As Last" (FSAL) property: the last stage is evaluated at the same point as the first stage of the next step. Dormand and Prince chose the coefficients of their method to minimize the error of the fifth-order solution. This is the main difference with the Fehlberg method, which was constructed so that the fourth-order solution has a small error. For this reason, the Dormand–Prince method is more suitable when the higher-order solution is used to continue the integration, a practice known as local extrapolation. Butcher tableau The Butcher tableau is: The first row of b coefficients gives the fifth-order accurate solution, and the second row gives the fourth-order accurate solution. Applications , Dormand–Prince is the default method in the ode45 solver for MATLAB and GNU Octave and is the default choice for the Simulink's model explorer solver. It is an option in Python's SciPy ODE integration library and in Julia (programming language)'s ODE solvers library. Implementations for the languages Fortran, Java, and C++ are also available. Notes
https://en.wikipedia.org/wiki/Open%20Regulatory%20Annotation%20Database
The Open Regulatory Annotation Database (also known as ORegAnno) is designed to promote community-based curation of regulatory information. Specifically, the database contains information about regulatory regions, transcription factor binding sites, regulatory variants, and haplotypes. Overview Data Management For each entry, cross-references are maintained to EnsEMBL, dbSNP, Entrez Gene, the NCBI Taxonomy database and PubMed. The information within ORegAnno is regularly mapped and provided as a UCSC Genome Browser track. Furthermore, each entry is associated with its experimental evidence, embedded as an Evidence Ontology within ORegAnno. This allows the researcher to analyze regulatory data using their own conditions as to the suitability of the supporting evidence. Software and data access The project is open source - all data and all software that is produced in the project can be freely accessed and used. Database contents As of December 20, 2006, ORegAnno contained 4220 regulatory sequences (excluding deprecated records) for 2190 transcription factor binding sites, 1853 regulatory regions (enhancers, promoters, etc.), 170 regulatory polymorphisms, and 7 regulatory haplotypes for 17 different organisms (predominantly Drosophila melanogaster, Homo sapiens, Mus musculus, Caenorhabditis elegans, and Rattus norvegicus in that order). These records were obtained by manual curation of 828 publications by 45 ORegAnno users from the gene regulation community. The ORegAnno publication queue contained 4215 publications of which 858 were closed, 34 were in progress (open status), and 3321 were awaiting annotation (pending status). ORegAnno is continually updated and therefore current database contents should be obtained from www.oreganno.org. RegCreative Jamboree 2006 The RegCreative jamboree was stimulated by a community initiative to curate in perpetuity the genomic sequences which have been experimentally determined to control gene expression. This object
https://en.wikipedia.org/wiki/Subtractor
In electronics, a subtractor – a digital circuit that performs subtraction of numbers – can be designed using the same approach as that of an adder. The binary subtraction process is summarized below. As with an adder, in the general case of calculations on multi-bit numbers, three bits are involved in performing the subtraction for each bit of the difference: the minuend (), subtrahend (), and a borrow in from the previous (less significant) bit order position (). The outputs are the difference bit () and borrow bit . The subtractor is best understood by considering that the subtrahend and both borrow bits have negative weights, whereas the X and D bits are positive. The operation performed by the subtractor is to rewrite (which can take the values -2, -1, 0, or 1) as the sum . , where ⊕ represents exclusive or. Subtractors are usually implemented within a binary adder for only a small cost when using the standard two's complement notation, by providing an addition/subtraction selector to the carry-in and to invert the second operand. (definition of two's complement notation) Half subtractor The half subtractors can be designed through the combinational Boolean logic circuits [2] as shown in Figure 1 and 2.The half subtractor is a combinational circuit which is used to perform subtraction of two bits. It has two inputs, the minuend and subtrahend and two outputs the difference and borrow out . The borrow out signal is set when the subtractor needs to borrow from the next digit in a multi-digit subtraction. That is, when . Since and are bits, if and only if and . An important point worth mentioning is that the half subtractor diagram aside implements and not since on the diagram is given by . This is an important distinction to make since subtraction itself is not commutative, but the difference bit is calculated using an XOR gate which is commutative. The truth table for the half subtractor is: Using the table above and a Karnaugh map, we find
https://en.wikipedia.org/wiki/Visual%20sensor%20network
A visual sensor network or smart camera network or intelligent camera network is a network of spatially distributed smart camera devices capable of processing, exchanging data and fusing images of a scene from a variety of viewpoints into some form more useful than the individual images. A visual sensor network may be a type of wireless sensor network, and much of the theory and application of the latter applies to the former. The network generally consists of the cameras themselves, which have some local image processing, communication and storage capabilities, and possibly one or more central computers, where image data from multiple cameras is further processed and fused (this processing may, however, simply take place in a distributed fashion across the cameras and their local controllers). Visual sensor networks also provide some high-level services to the user so that the large amount of data can be distilled into information of interest using specific queries. The primary difference between visual sensor networks and other types of sensor networks is the nature and volume of information the individual sensors acquire: unlike most sensors, cameras are directional in their field of view, and they capture a large amount of visual information which may be partially processed independently of data from other cameras in the network. Alternatively, one may say that while most sensors measure some value such as temperature or pressure, visual sensors measure patterns. In light of this, communication in visual sensor networks differs substantially from traditional sensor networks. Applications Visual sensor networks are most useful in applications involving area surveillance, tracking, and environmental monitoring. Of particular use in surveillance applications is the ability to perform a dense 3D reconstruction of a scene and storing data over a period of time, so that operators can view events as they unfold over any period of time (including the current moment) f
https://en.wikipedia.org/wiki/Rayleigh%20length
In optics and especially laser science, the Rayleigh length or Rayleigh range, , is the distance along the propagation direction of a beam from the waist to the place where the area of the cross section is doubled. A related parameter is the confocal parameter, b, which is twice the Rayleigh length. The Rayleigh length is particularly important when beams are modeled as Gaussian beams. Explanation For a Gaussian beam propagating in free space along the axis with wave number , the Rayleigh length is given by where is the wavelength (the vacuum wavelength divided by , the index of refraction) and is the beam waist, the radial size of the beam at its narrowest point. This equation and those that follow assume that the waist is not extraordinarily small; . The radius of the beam at a distance from the waist is The minimum value of occurs at , by definition. At distance from the beam waist, the beam radius is increased by a factor and the cross sectional area by 2. Related quantities The total angular spread of a Gaussian beam in radians is related to the Rayleigh length by The diameter of the beam at its waist (focus spot size) is given by . These equations are valid within the limits of the paraxial approximation. For beams with much larger divergence the Gaussian beam model is no longer accurate and a physical optics analysis is required. See also Beam divergence Beam parameter product Gaussian function Electromagnetic wave equation John Strutt, 3rd Baron Rayleigh Robert Strutt, 4th Baron Rayleigh Depth of field
https://en.wikipedia.org/wiki/Istv%C3%A1n%20Orosz
István Orosz (born 24 October 1951) is a Hungarian painter, printmaker, graphic designer and animated film director. He is known for his mathematically inspired works, impossible objects, optical illusions, double-meaning images and anamorphoses. The geometric art of István Orosz, with forced perspectives and optical illusions, has been compared to works by M. C. Escher. Biography He was born in Kecskemét. He studied at the Hungarian University of Arts and Design (now Moholy-Nagy University of Art and Design) in Budapest as a pupil of István Balogh and Ernő Rubik. After graduating in 1975 he began to deal with theatre as a stage designer and animated film as animator and film director. He is known as painter, printmaker, poster designer, and illustrator as well. He likes to use visual paradox, double meaning images, and illusionistic approaches while following traditional printing techniques such as woodcutting and etching. He also tries to renew the technique of anamorphosis. He is a regular participant in the major international biennials of posters and graphic art and his works have been shown in individual and group exhibitions in Hungary and abroad. Film director at the PannóniaFilm Studio in Budapest, Habil. professor at University of West Hungary in Sopron, co-founder of Hungarian Poster Association, member of Alliance Graphique International (AGI) and Hungarian Art Academie. He often uses ΟΥΤΙΣ, or Utisz, (pronounced: outis) (No one) as artist's pseudonym. Quotes "Utisz - It was the Homeric hero Odysseus, who fought the Cyclops, had used this name, and had put out the monster's eye. I imagine that poster is nothing else but an Odysseus' gesture: some kind of attack upon the eye." "If you want to create a poster try to explain your idea in a sentence. Then try to reduce it, leave out phrases, attributes until you just have the bare essentials. When you do not need any letter at all you are ready with the poster." "... When I have drawn these impossible obje
https://en.wikipedia.org/wiki/Sidney%20Redner
Sidney Redner (born 1951) is a Canadian-born physicist, professor, and a resident faculty member at the Santa Fe Institute. He was formerly department chair of physics at Boston University. Redner has published over 200 journal articles, authored a book titled A Guide to First-Passage Processes (2001, ), and coauthored a book titled A Kinetic View of Statistical Physics (2010, ) with Pavel L. Krapivsky and Eli Ben-Naim. His research focuses mainly on non-equilibrium statistical mechanics and network structure. He received his Ph.D. in physics from the Massachusetts Institute of Technology in 1977 under Gene Stanley, also on faculty at Boston University. He was awarded the American Physical Society's Leo P. Kadanoff Prize for 2021.
https://en.wikipedia.org/wiki/Light-induced%20voltage%20alteration
Light-induced voltage alteration (LIVA) is a semiconductor analysis technique that uses a laser or infrared light source to induce voltage changes in a device while scanning the beam of light across its surface. The technique relies upon the generation of electron-hole pairs in the semiconductor material when exposed to photons. Theory of operation The device to be analyzed is biased using a constant current power supply. As the light source is scanned across the surface of the silicon, electron-hole pairs are generated. This causes subtle alterations of the operating characteristics of the device, which may result in slight changes to the power supply voltage. Any changes that are detected in the power supply voltage are noted and correlated with the position of the light source on the device. This allows the physical locations corresponding to power supply fluctuations to be mapped onto an image of the device. This provides the device analyst with specific locations at which the device may be examined for defects. Notes
https://en.wikipedia.org/wiki/Germaben
Germaben II is a clear, viscous liquid used to inhibit microbial, yeast and mould growth in cosmetics and other personal products. It contains propylene glycol, propylparaben, methylparaben, and diazolidinyl urea. It is a Registered Trademark of International Specialty Products. Germaben II is used primarily by small businesses or home hobbyists who make soap, lotions, shampoos and other body care products.
https://en.wikipedia.org/wiki/Dispose%20pattern
In object-oriented programming, the dispose pattern is a design pattern for resource management. In this pattern, a resource is held by an object, and released by calling a conventional method – usually called close, dispose, free, release depending on the language – which releases any resources the object is holding onto. Many programming languages offer language constructs to avoid having to call the dispose method explicitly in common situations. The dispose pattern is primarily used in languages whose runtime environment have automatic garbage collection (see motivation below). Motivation Wrapping resources in objects Wrapping resources in objects is the object-oriented form of encapsulation, and underlies the dispose pattern. Resources are typically represented by handles (abstract references), concretely usually integers, which are used to communicate with an external system that provides the resource. For example, files are provided by the operating system (specifically the file system), which in many systems represents open files with a file descriptor (an integer representing the file). These handles can be used directly, by storing the value in a variable and passing it as an argument to functions that use the resource. However, it is frequently useful to abstract from the handle itself (for example, if different operating systems represent files differently), and to store additional auxiliary data with the handle, so handles can be stored as a field in a record, along with other data; if this in an opaque data type, then this provides information hiding and the user is abstracted from the actual representation. For example, in C file input/output, files are represented by objects of the FILE type (confusingly called "file handles": these are a language-level abstraction), which stores an (operating system) handle to the file (such as a file descriptor), together with auxiliary information like I/O mode (reading, writing) and position in the stream.
https://en.wikipedia.org/wiki/Four%20corners%20of%20the%20world
Several cosmological and mythological systems portray four corners of the world or four quarters of the world corresponding approximately to the four points of the compass (or the two solstices and two equinoxes). At the center may lie a sacred mountain, garden, world tree, or other beginning-point of creation. Often four rivers run to the four corners of the world, and water or irrigate the four quadrants of Earth. Semitic religions In Christianity and Judaism, the Old Testament (Book of Genesis, ) identifies the Garden of Eden, and the four rivers as the Tigris, Euphrates, Pishon, and Gihon. The Tigris runs to Assyria, the Euphrates to Armenia, the Pishon to Havilah or Elam, and the Gihon to Ethiopia. The four corners of the earth are also spoken of in the book of Revelation 7:1. Mesopotamian traditions In Mesopotamian cosmology, four rivers flowing out of the garden of creation, which is the center of the world, define the four corners of the world. From the point of view of the Akkadians, the northern geographical horizon was marked by Subartu, the west by , the east by Elam and the south by Sumer; later rulers of all of Mesopotamia, such as Cyrus, claimed among their titles , "King of the Four Corners". Indian traditions In Hinduism, the sacred mountain Kailash has four sides, from which four rivers flow to the four quarters of the world (the Ganges, Indus, Oxus (Amu Darya), and Śita (Tarim)), dividing the world into four quadrants. Another account portrays a celestial mountain, Mount Meru, buttressed by four terrestrial mountain ranges which extend in four directions. Between them lie four sacred lakes, through which the celestial river divides into four earthly rivers, which flow to the four corners and irrigate the four quadrants of the Earth. Buddhism and the Bon religion of Tibet have similar accounts. See also Four continents Biblical cosmology Flat Earth Time Cube
https://en.wikipedia.org/wiki/Thermal%20laser%20stimulation
Thermal laser stimulation represents a class of defect imaging techniques which employ a laser to produce a thermal variation in a semiconductor device. This technique may be used for semiconductor failure analysis. There are four techniques associated with thermal laser stimulation: optical beam induced resistance change (OBIRCH), thermally induced voltage alteration (TIVA)), external induced voltage alteration (XIVA) and Seebeck effect imaging (SEI) Optical beam induced resistance change Optical beam induced resistance change (OBIRCH) is an imaging technique which uses a laser beam to induce a thermal change in the device. Laser stimulation highlights differences in thermal characteristics between areas containing defects and areas which are defect-free. As the laser locally heats a defective area on a metal line which is carrying a current, the resulting resistance changes can be detected by monitoring the input current to the device. OBIRCH is useful for detecting electromigration effects resulting in open metal lines. A constant voltage is applied to the device-under-test (DUT). An area of interest is selected on the device, and a laser beam is used to scan the area. The input current being drawn by the device is monitored for changes during this process. When a change in current is noted, the position of the laser at the time that the change occurred is marked on the image of the device. When the laser beam strikes a location which does not contain a void, good thermal transmission exists and the change in electrical resistance is small. In areas containing voids, however, thermal transmission is impeded, resulting in a larger change in resistance. The degree of resistance change is displayed visually on an image of the device, with areas of higher resistance being displayed as bright spots. Thermally induced voltage alteration Thermally induced voltage alteration (TIVA) is an imaging technique which uses a laser beam to pinpoint the location of electrical
https://en.wikipedia.org/wiki/Miron%20Nicolescu
Miron Nicolescu (; August 27, 1903 – June 30, 1975) was a Romanian mathematician, best known for his work in real analysis and differential equations. He was President of the Romanian Academy and Vice-President of the International Mathematical Union. Born in Giurgiu, the son of a teacher, he attended the Matei Basarab High School in Bucharest. After completing his undergraduate studies at the Faculty of Mathematics of the University of Bucharest in 1924, he went to Paris, where he enrolled at the École Normale Supérieure and the Sorbonne. In 1928, he completed his doctoral dissertation, Fonctions complexes dans le plan et dans l'espace, under the direction of Paul Montel. Upon returning to Romania, he taught at the University of Cernăuți until 1940, when he was named professor at the University of Bucharest. In 1936, he was elected an associate member of the Romanian Academy, and, in 1953, full member. After King Michael's Coup of August 23, 1944, Nicolescu joined the Social Democratic Party, and later became a member of the Romanian Communist Party. In 1963, he became director of the Institute of Mathematics of the Romanian Academy, a position he held until 1973. From 1966 until his death, he served as President of the Romanian Academy. Peter Freund (who met Nicolescu when he gave a lecture in Timișoara), described him as an "affable, debonair man, and a very handsome ladies' man." Nicolescu was awarded the Legion of Honour, Commander rank, and was elected in 1972 member of the German National Academy of Sciences Leopoldina. At the International Congress of Mathematicians held in Vancouver, British Columbia, Canada in 1974, he was elected Vice-President of the International Mathematical Union, a position he held from 1975 until his death (his term was completed by Gheorghe Vrănceanu). A technical high school in Sector 4 of Bucharest bears his name, and so does a boulevard in Giurgiu. Publications See also Heat equation
https://en.wikipedia.org/wiki/International%20Conference%20on%20Differential%20Geometric%20Methods%20in%20Theoretical%20Physics
International Conference on Differential Geometric Methods in Theoretical Physics are congresses held every few years on the subject of Differential geometric methods in Theoretical physics. Lectures, seminars, and discussions are held in different universities throughout the world, every few years, and a book compilation is published thereafter consisting of the papers submitted and discussed at the conference. Works 1981 Conference on Differential Geometric Methods in Theoretical Physics International Centre for Theoretical Physics, Trieste, 30 June-3 July 1981 by G. Denardo; H. D. Doebner; International Centre for Theoretical Physics. Type: Book Language: English Publisher: Singapore : World Scientific, ©1983. OCLC: 9738857 13th International Conference on Differential Geometric Methods in Theoretical Physics Proceedings of the XIII International Conference on Differential Geometric Methods in Theoretical Physics [eds.] Heinz Dietrich Doebner, Tchavdar D. Palev (13th: 1984: Shumen, Bulgaria) (Singapore: World Scientific, 1986) 14th International Conference on Differential Geometric Methods in Mathematical Physics Proceedings of the 14th International Conference on Differential Geometric Methods in Theoretical Physics held in Salamanca, Spain, June 24–29, 1985 edited by Pedro Luis García Pérez and A. Pérez-Rendón (14th: 1985: Salamanca, Spain) (Berlin ; New York: Springer-Verlag, 1987) Notes: Series: Lecture Notes in Mathematics (Springer-Verlag); 1251 15th International Conference on Differential Geometric Methods in Theoretical Physics Proceedings of the XV International Conference on Differential Geometric Methods in Theoretical Physics edited by H.D. Doebner, J.D. Hennig (15th: 1986: Clausthal-Zellerfeld, Germany) (Singapore: World Scientific, 1987) 16th International Conference on Differential Geometrical Methods in Theoretical Physics Proceedings of the NATO Advanced Research Workshop and the 16th International Conference on Differential Geometrical
https://en.wikipedia.org/wiki/Steviol%20glycoside
Steviol glycosides are the chemical compounds responsible for the sweet taste of the leaves of the South American plant Stevia rebaudiana (Asteraceae) and the main ingredients (or precursors) of many sweeteners marketed under the generic name stevia and several trade names. They also occur in the related species S. phlebophylla (but in no other species of Stevia) and in the plant Rubus chingii (Rosaceae). Steviol glycosides from Stevia rebaudiana have been reported to be between 30 and 320 times sweeter than sucrose, although there is some disagreement in the technical literature about these numbers. They are heat-stable, pH-stable, and do not ferment. Steviol glycosides do not induce a glycemic response when ingested, because humans cannot metabolize stevia. The acceptable daily intake (ADI) for steviol glycosides, expressed as steviol equivalents, has been established to be 4 mg/kg body weight/day, and is based on no observed effects of a 100 fold higher dose in a rat study. Structure These compounds are glycosides of steviol. Specifically, their molecules can be viewed as a steviol molecule, with its carboxyl hydrogen atom replaced by a glucose molecule to form an ester, and a hydroxyl hydrogen with combinations of glucose and rhamnose to form an acetal. The steviol glycosides found in S. rebaudiana leaves, and their dry weight percentage, include: Stevioside (5–10%) Dulcoside A (0.5–1%) Rebaudioside A (2–4%) Rebaudioside B Rebaudioside C (1–2%) Rebaudioside D Rebaudioside E Rebaudioside F Rubusoside Steviolbioside The last three are present only in minute quantities, and rebaudioside B has been claimed to be a byproduct of the isolation technique. A commercial steviol glycoside mixture extracted from the plant was found to have about 80% stevioside, 8% rebaudioside A, and 0.6% rebaudioside C. The Chinese plant Rubus chingii produces rubusoside, a steviol glycoside not found in Stevia. According to the EU Stevia Regulation of 13 July 2021,
https://en.wikipedia.org/wiki/Functional%20extinction
Functional extinction is the extinction of a species or other taxon such that: It disappears from the fossil record, or historic reports of its existence cease; The reduced population no longer plays a significant role in ecosystem function; or The population is no longer viable. There are no individuals able to reproduce, or the small population of breeding individuals will not be able to sustain itself due to inbreeding depression and genetic drift, which leads to a loss of fitness. In plant populations, self-incompatibility mechanisms may cause related plant specimens to be incompatible, which may lead to functional extinction if an entire population becomes self-incompatible. This does not occur in larger populations. In polygynous populations, where only a few males leave offspring, there is a much smaller reproducing population than if all viable males were considered. Furthermore, the successful males act as a genetic bottleneck, leading to more rapid genetic drift or inbreeding problems in small populations. Functionally extinct species in modern times Baiji dolphin Northern white rhinoceros Ivory-billed woodpecker Christmas Island shrew Yangtze giant softshell turtle South China Tiger Bornean rhinoceros North Atlantic right whale Vaquita Fernandina Island tortoise On May 10, 2019, the Australian Koala Foundation issued a press release that opened with the sentence "The Australian Koala Foundation (AKF) believes Koalas may be functionally extinct in the entire landscape of Australia." The press release was reported on by multiple news agencies around the world, with most repeating the AKF's statement. Despite this, koalas are not currently considered functionally extinct; while their population has decreased, the IUCN Red List lists them only as "Vulnerable". The AKF's press release was released on the eve of the 2019 elections in Australia, where topics such as climate change were major issues. Distinct animal populations can also become functionally
https://en.wikipedia.org/wiki/Eotaxin
The eotaxins are a CC chemokine subfamily of eosinophil chemotactic proteins. Eotaxin is a special CC chemokine because it primarily attracts eosinophils. By being a chemoattractant for eosinophils, eotaxin has a direct relationship with inflammation. This is because eosinophils are known to promote inflammation. In order to induce stimulation, eotaxin binds with the CCR-3 receptor. The binding of eotaxin with the CCR-3 Receptor recruits eosinophils, which ultimately induces inflammation. According to early studies, the production of eotaxin can be linked to Th2 lymphocytes. Eotaxin appears to be T-cell dependent because of evidence that suggests that eosinophil recruitment is regulated by Th2 lymphocytes. The regulation occurs because of the presence of the CCR-3 Receptor on the Th2 lymphocyte. Some examples of the types of cells that have the ability of synthesizing eotaxin are lung cells, vascular endothelial cells, and macrophages. In humans, there are three family members: CCL11 (eotaxin-1) CCL24 (eotaxin-2) CCL26 (eotaxin-3) Structure and Role After analyzing the cDNA from guinea pig cDNA, it is determined that eotaxin is a 96 amino acid sequence. The beginning 23 amino acids are hydrophobic, while the remaining 73 are what make up the active eotaxin protein. Within the 73 amino acids present in the active eotaxin protein, there is similarity with other C-C chemokines that are categorized as “eosinophil- activating.” Some of these are human MPC-3 with a 51% similarity, human MPC-1 with 53% similarity and  MPC-2 with a 54% similarity. According to another source, Human MPC- 3 shares 57% of the same amino acid sequence and has 65% of the same amino acids.  As for the nucleotide base pairs, eotaxin shares 71% of its nucleotide sequence with MPC-3. Both Human MPC-3 and eotaxin are active on eosinophils, which means that they both will bind to the receptors found on eosinophils.  In humans, the gene that encodes for the eotaxin protein is found on chromosom
https://en.wikipedia.org/wiki/Ecosystem%20approach
The ecosystem approach is a conceptual framework for resolving ecosystem issues. The idea is to protect and manage the environment through the use of scientific reasoning. Another point of the ecosystem approach is preserving the Earth and its inhabitants from potential harm or permanent damage to the planet itself. With the preservation and management of the planet through an ecosystem approach, the future monetary and planetary gain are the by-product of sustaining and/or increasing the capacity of that particular environment. This is possible as the ecosystem approach incorporates humans, the economy, and ecology to the solution of any given problem. The initial idea for an ecosystem approach would come to light during the second meeting (November 1995) at the Conference of the Parties (COP) it was the central topic in implementation and framework for the Convention on Biological Diversity (CBD), it would further elaborate on the ecosystem approach as using varies methodologies for solving complex issues. Throughout, the use and incorporation of ecosystem approaches, two similar terms have been created in that time: ecosystem-based management and ecosystem management. The Convention on Biological Diversity has seen ecosystem-based management as a supporting topic/concept for the ecosystem approach. Similarly, ecosystem management has a minor difference with the two terms. Conceptual the differences between the three terms come from a framework structure and the different methods used in solving complex issues. The key component and definition between the three terms refer to the concept of conservation and protection of the ecosystem. The use of the ecosystem approach has been incorporated with managing water, land, and living organisms ecosystems and advocating the nourishment and sustainment of those ecological space. Since the ecosystem approach is a conceptual model for solving problems, the key idea could combat various problems. History On December 29, 1
https://en.wikipedia.org/wiki/William%20Whitaker%27s%20Words
William Whitaker's Words is a computer program that parses the inflection or conjugation of a Latin word, and also translates the root into English. Given an English word, the program outputs Latin translations. The software, written in Ada, is free for download but can be used online through several different hosts as well. This program has gained popularity among Latinists because of its simple interface, high coverage of the Latin lexicon and mostly accurate results. Nevertheless, the user has to check the results, since WORDS uses a set of rules based on natural prefixation, suffixation, declension and conjugation to determine the possibility of an entry. As a consequence of this approach of analysing the structure of words, there is no guarantee that these words were ever used in Latin literature or speech, even if the program finds a possible meaning to a given word. A few years after the original author's death, the software became the subject of digital preservation efforts. Coverage The dictionary consists of about 39,000 entries, which would result in hundreds of thousands of variations, counting declensions and conjugations. Additionally, the dictionary contains prefixes and suffixes. In comparison, the Oxford Latin Dictionary, considered to be the most complete Latin lexicon published in the English language, has about 34,000 entries, excluding proper names. The Oxford Latin Dictionary has fewer entries because it only contains entries from Classical Latin, whereas WORDS contains words from many time periods. Parsing process For instance, given the Latin verb form amābantur, WORDS analyzes it as: amābantur = am + (ā + ba + nt + ur), where am = amo, amare, amavi, amatus (English to love) ā = theme vowel for indicative mood ba = marker for the imperfect nt = marker for third person plural number ur = marker for passive voice So amābantur is the passive, 3rd person, plural, imperfect, indicative form of the verb "to love", which would be translated
https://en.wikipedia.org/wiki/Differential%20backup
A differential backup is a type of data backup that preserves data, saving only the difference in the data since the last full backup. The rationale in this is that, since changes to data are generally few compared to the entire amount of data in the data repository, the amount of time required to complete the backup will be smaller than if a full backup was performed every time that the organization or data owner wishes to back up changes since the last full backup. Another advantage, at least as compared to the incremental backup method of data backup, is that at data restoration time, at most two backup media are ever needed to restore all the data. This simplifies data restores as well as increases the likelihood of shortening data restoration time. Meaning A differential backup is a cumulative backup of all changes made since the last full backup, i.e., the differences since the last full backup. The advantage to this is the quicker recovery time, requiring only a full backup and the last differential backup to restore the entire data repository. The disadvantage is that for each day elapsed since the last full backup, more data needs to be backed up, especially if a significant proportion of the data has changed, thus increasing backup time as compared to the incremental backup method. It is important to use the terms "differential backup" and "incremental backup" correctly. The two terms are widely used in the industry, and their use is universally standard. A differential backup refers to a backup made to include the differences since the last full backup, while an incremental backup contains only the changes since the last incremental backup. (Or, of course, since the last full backup if the incremental backup in questions is the first incremental backup immediately after the last full backup.) All the major data backup vendors have standardized on these definitions. Illustration The difference between incremental and differential backups can be illustra
https://en.wikipedia.org/wiki/Line%20notation
Line notation is a typographical notation system using ASCII characters, most often used for chemical nomenclature. Chemistry Cell notation for representation of an electrochemical cell Dyson / IUPAC (1944) Hayward (1961) International Chemical Identifier (InChI) Wiswesser Line Notation (WLN) (1952) Simplified molecular input line entry specification (SMILES) Smiles arbitrary target specification (SMARTS) SYBYL Line Notation (SLN) Mathematics Mathematical markup language Music GUIDO music notation Chess Forsyth–Edwards Notation Notation Chemical nomenclature Musical notation
https://en.wikipedia.org/wiki/Wiswesser%20line%20notation
Wiswesser line notation (WLN), invented by William J. Wiswesser in 1949, was the first line notation capable of precisely describing complex molecules. It was the basis of ICI Ltd's CROSSBOW database system developed in the late 1960s. WLN allowed for indexing the Chemical Structure Index (CSI) at the Institute for Scientific Information (ISI). It was also the tool used to develop the CAOCI (Commercially Available Organic Chemical Intermediates) database, the datafile from which Accelrys' (successor to MDL) ACD file was developed. WLN is still being extensively used by BARK Information Services. Descriptions of how to encode molecules as WLN have been published in several books. Examples 1H : methane 2H : ethane 3H : propane 1Y : isobutane 1X : neopentane Q1 : methanol 1R : toluene 1V1 : acetone 2O2 : diethyl ether 1VR : acetophenone ZR CVQ : 3-aminobenzoic acid QVYZ1R : phenylalanine QX2&2&2 : 3-ethylpentan-3-ol QVY3&1VQ : 2-propylbutanedioic acid L66J BMR& DSWQ IN1&1 : 6-dimethylamino-4-phenylamino-naphthalene-2-sulfonic acid QVR-/G 5 : pentachlorobenzoic acid
https://en.wikipedia.org/wiki/Range%20of%20a%20projectile
In physics, a projectile launched with specific initial conditions will have a range. It may be more predictable assuming a flat Earth with a uniform gravity field, and no air resistance. The horizontal ranges of a projectile are equal for two complementary angles of projection with the same velocity. The following applies for ranges which are small compared to the size of the Earth. For longer ranges see sub-orbital spaceflight. The maximum horizontal distance traveled by the projectile, neglecting air resistance, can be calculated as follows: where d is the total horizontal distance travelled by the projectile. v is the velocity at which the projectile is launched g is the gravitational acceleration—usually taken to be 9.81 m/s2 (32 f/s2) near the Earth's surface θ is the angle at which the projectile is launched y0 is the initial height of the projectile If y0 is taken to be zero, meaning that the object is being launched on flat ground, the range of the projectile will simplify to: Ideal projectile motion Ideal projectile motion states that there is no air resistance and no change in gravitational acceleration. This assumption simplifies the mathematics greatly, and is a close approximation of actual projectile motion in cases where the distances travelled are small. Ideal projectile motion is also a good introduction to the topic before adding the complications of air resistance. Derivations A launch angle of 45 degrees displaces the projectile the farthest horizontally. This is due to the nature of right triangles. Additionally, from the equation for the range : We can see that the range will be maximum when the value of is the highest (i.e. when it is equal to 1). Clearly, has to be 90 degrees. That is to say, is 45 degrees. Flat ground First we examine the case where (y0) is zero. The horizontal position of the projectile is In the vertical direction We are interested in the time when the projectile returns to the same height
https://en.wikipedia.org/wiki/Cockpit%20display%20system
The Cockpit display systems (or CDS) provides the visible (and audible) portion of the Human Machine Interface (HMI) by which aircrew manage the modern Glass cockpit and thus interface with the aircraft avionics. History Prior to the 1970s, cockpits did not typically use any electronic instruments or displays (see Glass cockpit history). Improvements in computer technology, the need for enhancement of situational awareness in more complex environments, and the rapid growth of commercial air transportation, together with continued military competitiveness, led to increased levels of integration in the cockpit. The average transport aircraft in the mid-1970s had more than one hundred cockpit instruments and controls, and the primary flight instruments were already crowded with indicators, crossbars, and symbols, and the growing number of cockpit elements were competing for cockpit space and pilot attention. Architecture Glass cockpits routinely include high-resolution multi-color displays (often LCD displays) that present information relating to the various aircraft systems (such as flight management) in an integrated way. Integrated Modular Avionics (IMA) architecture allows for the integration of the cockpit instruments and displays at the hardware and software level to be maximized. CDS software typically uses API code to integrate with the platform (such as OpenGL to access the graphics drivers for example). This software may be written manually or with the help of COTS tools such as GL Studio, VAPS, VAPS XT or SCADE Display. Standards such as ARINC 661 specify the integration of the CDS at the software level with the aircraft system applications (called User Applications or UA). See also Acronyms and abbreviations in avionics Avionics software Integrated Modular Avionics
https://en.wikipedia.org/wiki/CA/EZTEST
CA-EZTEST was a CICS interactive test/debug software package distributed by Computer Associates and originally called EZTEST/CICS, produced by Capex Corporation of Phoenix, Arizona with assistance from Ken Dakin from England. The product provided source level test and debugging features for computer programs written in COBOL, PL/I and Assembler (BAL) languages to complement their own existing COBOL optimizer product. Competition CA-EZTEST initially competed with three rival products: "Intertest" originally from On-line Software International, based in the United States. In 1991, Computer Associates International, Inc. acquired On-line Software and renamed the product CA-INTERTEST, then stopped selling CA-EZTEST. OLIVER (CICS interactive test/debug) from Advanced Programming Techniques in the UK. XPEDITER from Compuware Corporation who in 1994 acquired the OLIVER product. Early critical role Between them, these three products provided much needed third-party system software support for IBM's "flagship" teleprocessing product CICS, which survived for more than 20 years as a strategic product without any memory protection of its own. A single "rogue" application program (frequently by a buffer overflow) could accidentally overwrite data almost anywhere in the address space causing "down-time" for the entire teleprocessing system, possibly supporting thousands of remote terminals. This was despite the fact that much of the world's banking and other commerce relied heavily on CICS for secure transaction processing between 1970 and early 1990s. The difficulty in deciding which application program caused the problem was often insurmountable and frequently the system would be restarted without spending many hours investigated very large (and initially unformatted) "core dump"s requiring expert system programming support and knowledge. Early integrated testing environment Additionally, the product (and its competitors) provided an integrated testing environment wh
https://en.wikipedia.org/wiki/Capex%20Corporation
Capex Corporation was an American computer software company in existence from 1969 through 1982 and based in Phoenix, Arizona. It made a variety of software products, mostly system utilities for the IBM mainframe platform, and was known for its Optimizer add-on to the IBM COBOL compiler. Capex was acquired by Computer Associates in 1982. Origins The company was begun as a start-up in August 1969. In all there were eight original employees, with funding coming from venture capital. Some of the founders had been working for General Electric's computer division in Phoenix, on systems such as the GE-600 series. There, as a 1990 profile in the Phoenix Business Journal of one of them stated, they "encountered the types of limitations that would encourage him and countless other engineers and technical people to venture off and start their own businesses". Among these founders were A. LeRoy Ellison (1936–2017), who became president of the new company, and Harry N. Cantrell (1924–2004). Other original employees had been working on Univac systems. In putting out mainframe computer systems, General Electric, like other hardware vendors of the time, was providing software without cost to its customers. The Capex founders thought that hardware companies failed to have a sufficient understanding of the software world and that, as a consequence, there was a viable market to be found for an independent software company. This was especially the case following IBM's decision to unbundle software from its mainframes in 1969, which happened but two months after the founding of Capex. Capex became one of the first companies to capitalize on that change in the marketplace. Initially, they intended to offer products for multiple vendors' platforms but soon found that economically it made sense to focus only on the dominant IBM mainframe. Early executives of Capex Corporation included Russell E. Edwards and John J. Anderson. The office for the company was located on 3rd Street, in t
https://en.wikipedia.org/wiki/Craig%27s%20theorem
In mathematical logic, Craig's theorem (also known as Craig's trick) states that any recursively enumerable set of well-formed formulas of a first-order language is (primitively) recursively axiomatizable. This result is not related to the well-known Craig interpolation theorem, although both results are named after the same logician, William Craig. Recursive axiomatization Let be an enumeration of the axioms of a recursively enumerable set T of first-order formulas. Construct another set T* consisting of for each positive integer i. The deductive closures of T* and T are thus equivalent; the proof will show that T* is a recursive set. A decision procedure for T* lends itself according to the following informal reasoning. Each member of T* is of the form Since each formula has finite length, it is checkable whether or not it is of the said form. If it is of the said form and consists of j conjuncts, it is in T* if the (reoccurring) conjunct is ; otherwise it is not in T*. Again, it is checkable whether the conjunct is in fact by going through the enumeration of the axioms of T and then checking symbol-for-symbol whether the expressions are identical. Primitive recursive axiomatizations The proof above shows that for each recursively enumerable set of axioms there is a recursive set of axioms with the same deductive closure. A set of axioms is primitive recursive if there is a primitive recursive function that decides membership in the set. To obtain a primitive recursive axiomatization, instead of replacing a formula with one instead replaces it with (*) where f(x) is a function that, given i, returns a computation history showing that is in the original recursively enumerable set of axioms. It is possible for a primitive recursive function to parse an expression of form (*) to obtain and j. Then, because Kleene's T predicate is primitive recursive, it is possible for a primitive recursive function to verify that j is indeed a computation history
https://en.wikipedia.org/wiki/Data%20deficient
A data deficient (DD) species is one which has been categorized by the International Union for Conservation of Nature (IUCN) as offering insufficient information for a proper assessment of conservation status to be made. This does not necessarily indicate that the species has not been extensively studied; but it does indicate that little or no information is available on the abundance and distribution of the species. The IUCN recommends that care be taken to avoid classing species as "data deficient" when the absence of records may indicate dangerously low abundance: "If the range of a taxon is suspected to be relatively circumscribed, if a considerable period of time has elapsed since the last record of the taxon, threatened status may well be justified" (see also precautionary principle). See also IUCN Red List data deficient species List of data deficient amphibians IUCN Red List data deficient species (Annelida) List of data deficient arthropods List of data deficient birds IUCN Red List data deficient species (Cnidaria) List of data deficient fishes List of data deficient insects List of data deficient invertebrates List of data deficient mammals List of data deficient molluscs List of data deficient plants List of data deficient reptiles
https://en.wikipedia.org/wiki/Abacus%20Harmonicus
Abacus Harmonicus, or Abacum Arithmetico-Harmonicum is a table and tabular method described in Athanasius Kircher's comprehensive 1650 work on music, the Musurgia Vniversalis. The purpose is to generate counterpoint combinations. Also mentioned in early editions of the Encyclopædia Britannica, it is best described by the author's caption: "wonderful table that reveals all the secret art of counterpoint".
https://en.wikipedia.org/wiki/Merosity
Merosity (from the greek "méros," which means "having parts") refers to the number of component parts in a distinct whorl of a plant structure. The term is most commonly used in the context of a flower where it refers to the number of sepals in a whorl of the calyx, the number of petals in a whorl of the corolla, the number of stamens in a whorl of the androecium, or the number of carpels in a whorl of the gynoecium. The term may also be used to refer to the number of leaves in a leaf whorl. The adjective n-merous refers to a whorl of n parts, where n is any integer greater than one. In nature, five or three parts per whorl have the highest frequency of occurrence, but four or two parts per whorl are not uncommon. Be aware that two consecutive whorls of dimerous petals are often mistaken for tetramerous petals. If all of the whorls in a given floral arrangement have the same merosity, the flower is said to be isomerous, otherwise the flower is anisomerous. For example, Trillium is isomerous since all whorls are trimerous (one whorl of three sepals, zero or one whorl of three petals, two whorls of three stamens each, and one whorl of three carpels). Trillium also has one whorl of three leaves. Gallery See also Cyclic flower Floral diagram Floral formula
https://en.wikipedia.org/wiki/Circumzenithal%20arc
The circumzenithal arc, also called the circumzenith arc (CZA), upside-down rainbow, and the Bravais arc, is an optical phenomenon similar in appearance to a rainbow, but belonging to the family of halos arising from refraction of sunlight through ice crystals, generally in cirrus or cirrostratus clouds, rather than from raindrops. The arc is located at a considerable distance (approximately 46°) above the observed Sun and at most forms a quarter of a circle centered on the zenith. It has been called "a smile in the sky", its first impression being that of an upside-down rainbow. The CZA is one of the brightest and most colorful members of the halo family. Its colors, ranging from violet on top to red at the bottom, are purer than those of a rainbow because there is much less overlap in their formation. The intensity distribution along the circumzenithal arc requires consideration of several effects: Fresnel's reflection and transmission amplitudes, atmospheric attenuation, chromatic dispersion (i.e. the width of the arc), azimuthal angular dispersion (ray bundling), and geometrical constraints. In effect, the CZA is brightest when the Sun is observed at about 20°. Contrary to public awareness, the CZA is not a rare phenomenon, but it tends to be overlooked since it occurs so far overhead. It is worthwhile to look out for it when sun dogs are visible, since the same type of ice crystals that cause them (plate-shaped hexagonal prisms in horizontal orientation) are responsible for the CZA. Formation The light that forms the CZA enters an ice crystal through its flat top face, and exits through a side prism face. The refraction of almost parallel sunlight through what is essentially a 90-degree prism accounts for the wide color separation and the purity of color. The CZA can only form when the sun is at an altitude lower than 32.2°. The CZA is brightest when the sun is at 22° above the horizon, which causes sunlight to enter and exit the crystals at the minimum d
https://en.wikipedia.org/wiki/IEEE%20Registration%20Authority
The IEEE Registration Authority is the administrative body that is responsible for registering and administering organizationally unique identifiers (OUI) and other types of identifiers which are used in the computer and electronics industries (Individual Address Blocks (IAB), Manufacturer IDs, Standard Group MAC Addresses, Unique Registration Numbers (URN), EtherType values, etc.) The IEEE Registration Authority was formed in 1986 in response to a need for this service that was recognized by the P802 (LAN/MAN) standards group. The IEEE Registration Authority is currently recognized by ISO/IEC as the authorized registration authority to provide the service of globally assigning, administering, and registering OUIs. Note: The term 'Registration' as used in this context is "the assignment of unambiguous names to objects in a way which makes the assignment available to interested parties".