source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Cytidine%20diphosphate | Cytidine diphosphate, abbreviated CDP, is a nucleoside diphosphate. It is an ester of pyrophosphoric acid with the nucleoside cytidine. CDP consists of the pyrophosphate group, the pentose sugar ribose, and the nucleobase cytosine.
In Bacillus subtilis and Staphylococcus aureus, CDP-activated glycerol and ribitol are necessary to build wall teichoic acid.
See also
Nucleoside
Nucleotide
DNA
RNA
Oligonucleotide |
https://en.wikipedia.org/wiki/Splendid%20China%20%28Florida%29 | Splendid China was a theme park in Four Corners, Florida. It was opened in 1993 but closed on December 31, 2003. It was a sister park to the Splendid China in Shenzhen, China, and cost $100 million to build.
It was a miniature park with more than 60 replicas at a one-tenth scale at its height of popularity. Each piece was handcrafted to maintain authenticity. Initially, Chinese artists were hired to perform in the park. After a number of them tried to seek political asylum in the United States, they were replaced by local performers.
Citizens Against Communist Chinese Propaganda criticized the Chinese government's ownership of Splendid China; the state-owned corporation China Travel Service owned and operated the theme park.
History
Origins
The original idea for the Florida Splendid China theme park was that of Josephine Chen, a former educator from Taiwan. In 1988, Chen toured a prototype park, "Splendid China Miniature Scenic Spot", in Shenzhen, China operated by China Travel Service (CTS). This park is close to Hong Kong. 3.5 million people visited the park during the first year. CTS recouped its US$100 million investment during this first year.
She negotiated an agreement with CTS, controlled by Overseas Chinese Affairs Office, under the China National Tourism Administration. Under the agreement, Chen supplied the land and management services, while CTS would construct the building, and supply non-management personnel.
Construction
On December 19, 1989, groundbreaking ceremonies were held for the start of construction for Florida Splendid China.
The park's replica of the Great Wall took nearly seven million long bricks and stretched about . The replica of the Leshan Buddha was four stories tall.
Buyout
In December 1993, the American partners were bought out by the Chinese government.
Changes in 1996
In May 1996, the Orlando Sentinel reported that the President, General Manager, and Senior Vice-President for Entertainment were replaced in what was desc |
https://en.wikipedia.org/wiki/P680 | P680, or photosystem II primary donor, is the reaction-center chlorophyll a molecular dimer associated with photosystem II in plants, algae, and cyanobacteria, and central to oxygenic photosynthesis.
Etymology
Its name is derived from the word “pigment” (P) and the presence of a major bleaching band centered around 680-685 nm in the flash-induced absorbance difference spectra of P680/ P680+•.
Components
The structure of P680 consists of a heterodimer of two distinct chlorophyll molecules,
referred to as P and P. This “special pair” forms an excitonic dimer that functions as a single unit, excited by light energy as if they were a single molecule.
Action and function
Excitation
P680 receives excitation energy either by directly absorbing a photon of suitable frequency or indirectly from other chlorophylls within photosystem II, thereby exciting an electron to a higher energy level. The resulting P680 with a loosened electron is designated as P680*, which is a strong reducing agent.
Charge separation
Following excitation, the loosened electron of P680* is taken up by the primary electron acceptor, a pheophytin molecule located within photosystem II near P680. During this transfer, P680* is ionized and oxidized, producing cationic P680.
Recovery of P680
P680+ is the strongest biological oxidizing agent known, with an estimated redox potential of ~1.3 V. This makes it possible to oxidize water during oxygenic photosynthesis. P680+ recovers its lost electron by oxidizing water via the oxygen-evolving complex, which regenerates P680.
See also
P700
Photosystem I
Photosystem II |
https://en.wikipedia.org/wiki/Restructured%20steak | Restructured steak is a catch-all term to describe a class of imitation beef steaks made from smaller pieces of beef fused together by a binding agent. Its development started from the 1970s. Restructured steak is sometimes made using cheaper cuts of beef such as the hind quarter or fore quarter of beef.
Allowed food-grade agents include:
Sodium chloride (table salt) and phosphate salts. Salt can prevent microbiological growth and make myosin-type proteins more soluble. The allowed amount of phosphate in end products is 0.5% in the United States. It increases the emulsification of fat.
Animal blood plasma
Alginate: Sodium alginate forms an adhesive gel in the presence of Ca2+ ion.
Transglutaminase: an enzyme that helps the forming of cross-binding proteins.
Problems
Oxidation and food poisoning are the two most serious issues generally associated with restructured steak. To reduce the risk of food poisoning, restructured steaks should always be cooked until well-done. |
https://en.wikipedia.org/wiki/Oxygen-evolving%20complex | The oxygen-evolving complex (OEC), also known as the water-splitting complex, is a water-oxidizing enzyme involved in the photo-oxidation of water during the light reactions of photosynthesis. OEC is surrounded by 4 core proteins of photosystem II at the membrane-lumen interface. The mechanism for splitting water involves absorption of three photons before the fourth provides sufficient energy for water oxidation. Based on a widely accepted theory from 1970 by Kok, the complex can exist in 5 states, denoted S0 to S4, with S0 the most reduced and S4 the most oxidized. Photons trapped by photosystem II move the system from state S0 to S4. S4 is unstable and reacts with water producing free oxygen. For the complex to reset to the lowest state, S0, it uses 2 water molecules to pull out 4 electrons.
The OEC active site contains a cluster of manganese and calcium, with the formula Mn4Ca1OxCl1–2(HCO3)y. This cluster is coordinated by D1 and CP43 subunits and stabilized by peripheral membrane proteins. Other characteristics of it have been reviewed.
Currently, the mechanism of the complex is not completely understood. Along with the role of Ca+2, Cl−1, and the membrane proteins surrounding the metal cluster not being well understood. Much of what is known has been collected from flash photolysis experiments, electron paramagnetic resonance (EPR), and X-ray spectroscopy. |
https://en.wikipedia.org/wiki/Multigrid%20method | In numerical analysis, a multigrid method (MG method) is an algorithm for solving differential equations using a hierarchy of discretizations. They are an example of a class of techniques called multiresolution methods, very useful in problems exhibiting multiple scales of behavior. For example, many basic relaxation methods exhibit different rates of convergence for short- and long-wavelength components, suggesting these different scales be treated differently, as in a Fourier analysis approach to multigrid. MG methods can be used as solvers as well as preconditioners.
The main idea of multigrid is to accelerate the convergence of a basic iterative method (known as relaxation, which generally reduces short-wavelength error) by a global correction of the fine grid solution approximation from time to time, accomplished by solving a coarse problem. The coarse problem, while cheaper to solve, is similar to the fine grid problem in that it also has short- and long-wavelength errors. It can also be solved by a combination of relaxation and appeal to still coarser grids. This recursive process is repeated until a grid is reached where the cost of direct solution there is negligible compared to the cost of one relaxation sweep on the fine grid. This multigrid cycle typically reduces all error components by a fixed amount bounded well below one, independent of the fine grid mesh size. The typical application for multigrid is in the numerical solution of elliptic partial differential equations in two or more dimensions.
Multigrid methods can be applied in combination with any of the common discretization techniques. For example, the finite element method may be recast as a multigrid method. In these cases, multigrid methods are among the fastest solution techniques known today. In contrast to other methods, multigrid methods are general in that they can treat arbitrary regions and boundary conditions. They do not depend on the separability of the equations or other special |
https://en.wikipedia.org/wiki/Fermented%20fish | Fermented fish is a traditional preservation of fish. Before refrigeration, canning and other modern preservation techniques became available, fermenting was an important preservation method. Fish rapidly spoils, or goes rotten, unless some method is applied to stop the bacteria that produce the spoilage. Fermentation is a method which attacks the ability of microbials to spoil fish. It does this by making the fish muscle more acidic; bacteria usually cease multiplying when the pH drops below 4.5.
A modern approach, biopreservation, adds lactic acid bacteria to the fish to be fermented. This produces active antimicrobials such as lactic and acetic acid, hydrogen peroxide, and peptide bacteriocins. It can also produce the antimicrobial nisin, a particularly effective preservative.
Fermented fish preparations can be notable for their putrid smell. These days there are many other techniques of preserving fish, but fish is still fermented because some people enjoy the taste.
Risks
Alaska has witnessed a steady increase of cases of botulism since 1985. It has more cases of foodborne botulism than any other state in the United States of America. This is caused by the traditional Inuit/Yupik practice of allowing animal products such as whole fish, fish heads, walrus, sea lion, and whale flippers, beaver tails, seal oil, birds, etc., to ferment for an extended period of time before being consumed. The risk is exacerbated when a plastic container is used for this purpose instead of the old-fashioned, traditional method, a grass-lined hole, as the botulinum bacteria thrive in the anaerobic conditions created by the air-tight enclosure in plastic.
Preparations
See also
Food fermentation
Preserved fish
Notes |
https://en.wikipedia.org/wiki/Adjuncts | In brewing, adjuncts are unmalted grains (such as corn, rice, rye, oats, barley, and wheat) or grain products used in brewing beer which supplement the main mash ingredient (such as malted barley). This is often done with the intention of cutting costs, but sometimes also to create an additional feature, such as better foam retention, flavours or nutritional value or additives. Both solid and liquid adjuncts are commonly used.
Definition
Ingredients which are standard for certain beers, such as wheat in a wheat beer, may be termed adjuncts when used in beers which could be made without them — such as adding wheat to a pale ale for the purpose of creating a lasting head. The sense here is that the ingredient is additional and strictly unnecessary, though it may be beneficial and attractive. Under the Bavarian Reinheitsgebot purity law it would be considered that an adjunct is any beer ingredient other than water, barley, hops, and yeast; this, however, is an antiquated view. This purity law originated in 16th-century Germany and did not initially include yeast due to the fact that it had not been discovered yet.
The term adjunct is often used to refer to corn, rice, oats, unmalted barley and rye. The use of ingredients as substitutes for the main starch source is where the term adjunct is most often used.
Types of adjuncts and adjunct products
Adjuncts can be broadly classified according to the physical form in which they are used into solids and liquid syrups.
Solid adjuncts are either starchy adjuncts which need to be converted to simpler sugars, or solid sugar adjuncts which can be added after conversion. Solid starchy adjuncts are normally produced from cereals and are used in the form of flakes, grits, flour or purified starch and must be added before the mash tun to convert the starch into simple sugars which the yeast can use during fermentation. Cereals with a higher gelatinisation temperature than the standard mashing temperatures must be cooked in a |
https://en.wikipedia.org/wiki/Mom%27s%20Cancer | Mom's Cancer is an autobiographical graphic medicine webcomic by Brian Fies which describes his mother's fight against metastatic lung cancer, as well as his family's reactions to it. Mom's Cancer was the first webcomic to win an Eisner Award, winning in 2005. Its print collection, published in 2006, won a Harvey Award and a Deutscher Jugendliteraturpreis.
Characters and story
There are four main characters in the story, all of which are part of the family. None of the characters are ever referred by name, and the author uses his relationships to identify them.
Mom: the author's mother; she is in her 60s, and suffers from stage IV metastatic lung cancer.
Me: The author, a "self-employed writer" in his 40s, and the eldest child.
Nurse Sis: One of the author's sisters; slightly younger than him, and a registered nurse.
Kid Sis: The author's youngest sister; an actress and writer who lives with her mother.
Aside from the main characters, there are several secondary characters who appear several times, most notably the author's stepfather and his mother's head doctor.
A reviewer described the story as "not a discussion about the science of cancer, [and] not a guidebook to the troubles one may face with this disease... through this graphic novel, Fies simply tries to make sense of and document the reality of the course of his mother’s illness and how it affected his family... he simply tells the reader how it is, how he felt, and how it happened... Mixed messages from different doctors contradict and confound one another. As Fies’ mother struggles (and to an extent, fails) to understand her illness she is made to feel less than by providers who belittle her concerns or simply put on a big smile. The pain, fear, and anguish that patients face once they leave the hospital is brought to the forefront throughout the book."
Format
The story is divided into chapters, each one containing several multipanel pages. Although the webcomic is mostly presented in black and whi |
https://en.wikipedia.org/wiki/Dominating%20set | In graph theory, a dominating set for a graph is a subset of its vertices, such that any vertex of is either in , or has a neighbor in . The domination number is the number of vertices in a smallest dominating set for .
The dominating set problem concerns testing whether for a given graph and input ; it is a classical NP-complete decision problem in computational complexity theory. Therefore it is believed that there may be no efficient algorithm that can compute for all graphs . However, there are efficient approximation algorithms, as well as efficient exact algorithms for certain graph classes.
Dominating sets are of practical interest in several areas. In wireless networking, dominating sets are used to find efficient routes within ad-hoc mobile networks. They have also been used in document summarization, and in designing secure systems for electrical grids.
Formal definition
Given an undirected graph , a subset of vertices is called a dominating set if for every vertex , there is a vertex such that .
Every graph has at least one dominating set: if the set of all vertices, then by definition D is a dominating set, since there is no vertex . A more interesting challenge is to find small dominating sets. The domination number of is defined as: .
Variants
A connected dominating set is a dominating set that is also connected. If S is a connected dominating set, one can form a spanning tree of G in which S forms the set of non-leaf vertices of the tree; conversely, if T is any spanning tree in a graph with more than two vertices, the non-leaf vertices of T form a connected dominating set. Therefore, finding minimum connected dominating sets is equivalent to finding spanning trees with the maximum possible number of leaves.
A total dominating set (or strongly-dominating set) is a set of vertices such that all vertices in the graph, including the vertices in the dominating set themselves, have a neighbor in the dominating set. That is: for every |
https://en.wikipedia.org/wiki/Bucket%20toilet | A bucket toilet is a basic form of a dry toilet whereby a bucket (pail) is used to collect excreta. Usually, feces and urine are collected together in the same bucket, leading to odor issues. The bucket may be situated inside a dwelling, or in a nearby small structure (an outhouse).
Where people do not have access to improved sanitation – particularly in low-income urban areas of developing countries – an unimproved bucket toilet might be better than open defecation. They can play a temporary role in emergency sanitation, e.g. after earthquakes. However, the unimproved bucket toilet may carry significant health risks compared to an improved sanitation system. The bucket toilet system, with collection organised by the municipality, used to be widespread in wealthy countries; in Australia it persisted into the second half of the 20th century.
Once the basic bucket toilet has been "improved", it evolves into a number of different systems, which are more correctly referred to as either container-based sanitation systems, composting toilets, or urine-diverting dry toilets.
Applications
Unimproved toilets
Bucket toilets are used in households and even in health care facilities in some low- and middle- income countries where people do not have access to improved sanitation.
In those settings, bucket toilets are more likely to be used without a liner, or the liner is not removed each time the bucket is emptied. This is because the users cannot afford to regularly discard suitably sized, sturdy liners. Instead, the users may place some dry material in the base of the bucket (newspaper, sawdust, leaves, straw, or similar) in order to facilitate easier emptying.
Cold climates
Bucket toilets have been historically common in cold climates where installing running water can be difficult and expensive and subject to freezing-related pipe breakage, for example in Alaska and rural areas of Canada and Russia.
Emergencies
In natural disasters and other emergencies, the po |
https://en.wikipedia.org/wiki/Gene%20amplification | Gene amplification refers to a number of natural and artificial processes by which the number of copies of a gene is increased "without a proportional increase in other genes".
Artificial DNA amplification
In research or diagnosis DNA amplification can be conducted through methods such as:
Polymerase chain reaction, an easy, cheap, and reliable way to repeatedly replicate a focused segment of DNA by polymerizing nucleotides, a concept which is applicable to numerous fields in modern biology and related sciences.
Ligase chain reaction, a method that amplifies the nucleic acid used as the probe. For each of the two DNA strands, two partial probes are ligated to form the actual one; thus, LCR uses two enzymes: a DNA polymerase (used for initial template amplification and then inactivated) and a thermostable DNA ligase.
Transcription-mediated amplification, an isothermal, single-tube nucleic acid amplification system utilizing two enzymes, RNA polymerase and reverse transcriptase, to rapidly amplify the target RNA/DNA, enabling the simultaneous detection of multiple pathogenic organisms in a single tube.
Natural DNA amplification
DNA replication is a natural form of copying DNA with the amount of genes remaining constant. However, the amount of DNA or the number of genes can also increase within an organism through gene duplication, a major mechanism through which new genetic material is generated during molecular evolution. Common sources of gene duplications include ectopic recombination, retrotransposition event, aneuploidy, polyploidy, and replication slippage.
A piece of DNA or RNA that is the source and/or product of either natural or artificial amplification or replication events is called an amplicon. |
https://en.wikipedia.org/wiki/Cell%20junction | Cell junctions or junctional complexes, are a class of cellular structures consisting of multiprotein complexes that provide contact or adhesion between neighboring cells or between a cell and the extracellular matrix in animals. They also maintain the paracellular barrier of epithelia and control paracellular transport. Cell junctions are especially abundant in epithelial tissues. Combined with cell adhesion molecules and extracellular matrix, cell junctions help hold animal cells together.
Cell junctions are also especially important in enabling communication between neighboring cells via specialized protein complexes called communicating (gap) junctions. Cell junctions are also important in reducing stress placed upon cells.
In plants, similar communication channels are known as plasmodesmata, and in fungi they are called septal pores.
Types
In vertebrates, there are three major types of cell junction:
Adherens junctions, desmosomes and hemidesmosomes (anchoring junctions)
Gap junctions (communicating junction)
Tight junctions (occluding junctions)
Invertebrates have several other types of specific junctions, for example septate junctions or the C. elegans apical junction. In multicellular plants, the structural functions of cell junctions are instead provided for by cell walls. The analogues of communicative cell junctions in plants are called plasmodesmata.
Anchoring junctions
Cells within tissues and organs must be anchored to one another and attached to components of the extracellular matrix. Cells have developed several types of junctional complexes to serve these functions, and in each case, anchoring proteins extend through the plasma membrane to link cytoskeletal proteins in one cell to cytoskeletal proteins in neighboring cells as well as to proteins in the extracellular matrix.
Three types of anchoring junctions are observed, and differ from one another in the cytoskeletal protein anchor as well as the transmembrane linker protein that extends t |
https://en.wikipedia.org/wiki/Solar%20irradiance | Solar irradiance is the power per unit area (surface power density) received from the Sun in the form of electromagnetic radiation in the wavelength range of the measuring instrument.
Solar irradiance is measured in watts per square metre (W/m2) in SI units.
Solar irradiance is often integrated over a given time period in order to report the radiant energy emitted into the surrounding environment (joule per square metre, J/m2) during that time period. This integrated solar irradiance is called solar irradiation, solar exposure, solar insolation, or insolation.
Irradiance may be measured in space or at the Earth's surface after atmospheric absorption and scattering. Irradiance in space is a function of distance from the Sun, the solar cycle, and cross-cycle changes.
Irradiance on the Earth's surface additionally depends on the tilt of the measuring surface, the height of the Sun above the horizon, and atmospheric conditions.
Solar irradiance affects plant metabolism and animal behavior.
The study and measurement of solar irradiance have several important applications, including the prediction of energy generation from solar power plants, the heating and cooling loads of buildings, climate modeling and weather forecasting, passive daytime radiative cooling applications, and space travel.
Types
There are several measured types of solar irradiance.
Total solar irradiance (TSI) is a measure of the solar power over all wavelengths per unit area incident on the Earth's upper atmosphere. It is measured perpendicular to the incoming sunlight. The solar constant is a conventional measure of mean TSI at a distance of one astronomical unit (AU).
Direct normal irradiance (DNI), or beam radiation, is measured at the surface of the Earth at a given location with a surface element perpendicular to the Sun direction. It excludes diffuse solar radiation (radiation that is scattered or reflected by atmospheric components). Direct irradiance is equal to the extraterrestria |
https://en.wikipedia.org/wiki/Pharming%20%28genetics%29 | Pharming, a portmanteau of "farming" and "pharmaceutical", refers to the use of genetic engineering to insert genes that code for useful pharmaceuticals into host animals or plants that would otherwise not express those genes, thus creating a genetically modified organism (GMO). Pharming is also known as molecular farming, molecular pharming or biopharming.
The products of pharming are recombinant proteins or their metabolic products. Recombinant proteins are most commonly produced using bacteria or yeast in a bioreactor, but pharming offers the advantage to the producer that it does not require expensive infrastructure, and production capacity can be quickly scaled to meet demand, at greatly reduced cost.
History
The first recombinant plant-derived protein (PDP) was human serum albumin, initially produced in 1990 in transgenic tobacco and potato plants. Open field growing trials of these crops began in the United States in 1992 and have taken place every year since. While the United States Department of Agriculture has approved planting of pharma crops in every state, most testing has taken place in Hawaii, Nebraska, Iowa, and Wisconsin.
In the early 2000s, the pharming industry was robust. Proof of concept has been established for the production of many therapeutic proteins, including antibodies, blood products, cytokines, growth factors, hormones, recombinant enzymes and human and veterinary vaccines. By 2003 several PDP products for the treatment of human diseases were under development by nearly 200 biotech companies, including recombinant gastric lipase for the treatment of cystic fibrosis, and antibodies for the prevention of dental caries and the treatment of non-Hodgkin's lymphoma.
However, in late 2002, just as ProdiGene was ramping up production of trypsin for commercial launch it was discovered that volunteer plants (left over from the prior harvest) of one of their GM corn products were harvested with the conventional soybean crop later plante |
https://en.wikipedia.org/wiki/Lobularia%20maritima | Lobularia maritima (syn. Alyssum maritimum) is a species of low-growing flowering plant in the family Brassicaceae. Its common name is sweet alyssum or , also commonly referred to as just alyssum (from the genus Alyssum in which it was formerly classified).
Etymology
The genus name Lobularia comes from a Greek word meaning 'small pod', referring to the shape of the fruits. The name of the species maritima refers to its preferred coastal habitat.
Description
Lobularia maritima is an annual plant (rarely a short-lived perennial plant) growing to tall by broad. The stem is very branched, with dense clusters of small flowers. The leaves are 1–4 cm long and 3–5 mm, broad, alternate, sessile, quite hairy, oval to lanceolate, with an entire margin.
The flowers are about in diameter, sweet-smelling, with an aroma similar to that of honey, with four white rounded petals (or pink, rose-red, violet. yellow and lilac) and four sepals. The six stamens have yellow anthers. The flowers are produced throughout the growing season, or year-round in areas free of frost. They are pollinated by insects (entomophily) as its sweet honey-like fragrance is attractive to bees, flower flies, stingless wasps and butterflies. The fruits are numerous elongated seedpods rather hairy, oval to rounded, each containing two seeds. The dispersal of seed is affected by the wind (anemochory).
Distribution
This plant is native to the Mediterranean Basin and the Macaronesia region: (Canary Islands, Madeira, Cape Verde). It is widely naturalized elsewhere in the temperate world, including the United States. There is an endemic subspecies in the local flora of the Columbretes Islands of the western Mediterranean.
Habitat
It is common on sandy beaches and dunes, but can also grow on cultivated fields, walls, slopes and waste ground, preferably on calcareous soil, at an altitude of above sea level.
Cultivation
Lobularia maritima is cultivated in gardens, with many horticultural varieties with purple |
https://en.wikipedia.org/wiki/Parental%20investment | Parental investment, in evolutionary biology and evolutionary psychology, is any parental expenditure (e.g. time, energy, resources) that benefits offspring. Parental investment may be performed by both males and females (biparental care), females alone (exclusive maternal care) or males alone (exclusive paternal care). Care can be provided at any stage of the offspring's life, from pre-natal (e.g. egg guarding and incubation in birds, and placental nourishment in mammals) to post-natal (e.g. food provisioning and protection of offspring).
Parental investment theory, a term coined by Robert Trivers in 1972, predicts that the sex that invests more in its offspring will be more selective when choosing a mate, and the less-investing sex will have intra-sexual competition for access to mates. This theory has been influential in explaining sex differences in sexual selection and mate preferences, throughout the animal kingdom and in humans.
History
In 1859, Charles Darwin published On the Origin of Species. This introduced the concept of natural selection to the world, as well as related theories such as sexual selection. For the first time, evolutionary theory was used to explain why females are "coy" and males are "ardent" and compete with each other for females' attention. In 1930, Ronald Fisher wrote The Genetical Theory of Natural Selection, in which he introduced the modern concept of parental investment, introduced the sexy son hypothesis, and introduced Fisher's principle. In 1948, Angus John Bateman published an influential study of fruit flies in which he concluded that because female gametes are more costly to produce than male gametes, the reproductive success of females was limited by the ability to produce ovum, and the reproductive success of males was limited by access to females. In 1972, Trivers continued this line of thinking with his proposal of parental investment theory, which describes how parental investment affects sexual behavior. He conclude |
https://en.wikipedia.org/wiki/PDP-4 | The PDP-4 was the successor to the Digital Equipment Corporation's PDP-1.
History
This 18-bit machine, first shipped in 1962, was a compromise: "with slower memory and different packaging" than the PDP-1, but priced at $65,000 - less than half the price of its predecessor. All later 18-bit PDP machines (7, 9 and 15) are based on a similar, but enlarged instruction set, more powerful than, but based on the same concepts as, the 12-bit PDP-5/PDP-8 series.
Approximately 54 were sold.
Hardware
The system's memory cycle is 8 microseconds, compared to 5 microseconds for the PDP-1.
The PDP-4 weighs about .
Mass storage
Both the PDP-1 and the PDP-4 were introduced as paper tape-based systems. The only use, if any, for IBM-compatible 200 BPI or 556 BPI magnetic tape was for data. The use of "mass storage" drums - not even a megabyte and non-removable - were an available option, but were not in the spirit of the “personal” or serially shared systems that DEC offered.
It was in this setting that DEC introduced DECtape, initially called "MicroTape", for both the PDP-1 and PDP-4.
Software
DEC provided an editor, an assembler, and a FORTRAN II compiler. The assembler was different from that of the PDP-1 in two ways:
Unlike the PDP-1, macros were not supported.
It was a one-pass assembler; paper-tape input did not have to be read twice.
Photos
PDP-4
See also
Programmed Data Processor |
https://en.wikipedia.org/wiki/PDP-9 | The PDP-9, the fourth of the five 18-bit minicomputers produced by Digital Equipment Corporation, was introduced in 1966. A total of 445 PDP-9 systems were produced, of which 40 were the compact, low-cost PDP-9/L units.
History
The 18-bit PDP systems preceding the PDP-9 were the PDP-1, PDP-4 and PDP-7. Its successor was the PDP-15.
Hardware
The PDP-9, which is "two metres wide and about 75cm deep," is approximately twice the speed of the PDP-7. It was built using discrete transistors, and has an optional integrated vector graphics terminal. The PDP-9 has a memory cycle time of 1 microsecond, and weighs about . The PDP-9/L has a memory cycle time of 1.5 microseconds, and weighs about .
It was DEC's first microprogrammed machine.
A typical configuration included:
300 cps Paper Tape Reader
50 cps Paper Tape Punch
10 cps Console Teleprinter, Model 33 KSR
Among the improvements of the PDP-9 over its PDP-7 predecessor were:
the addition of Status flags for reader and punch errors, thus providing added flexibility and for error detection
an entirely new design for multi-level interrupts, called the Automatic Priority Interrupt (API) option
a more advanced form of memory management
User/university-based research projects for extending the PDP-9 included:
a hardware capability for floating-point arithmetic, at a time when machines in this price range used software
a PDP-9 controlled parallel computer
Software
The system came with a single-user keyboard monitor. DECsys provided an interactive, single-user, program development environment for Fortran and assembly language programs.
Both FORTRAN II and FORTRAN IV were implemented for the PDP-9.
MUMPS was originally developed on the PDP-7, and ran on several PDP-9s at the Massachusetts General Hospital.
Sales
The PDP-7, of which 120 were sold, was described as "highly successful". The PDP-9 sold 445 units. Both have submodels, the PDP-7A and the PDP-9/L, neither of which accounted for a substantial percent |
https://en.wikipedia.org/wiki/PDP-5 | The PDP-5 was Digital Equipment Corporation's first 12-bit computer, introduced in 1963.
History
An earlier 12-bit computer, named LINC has been described as the first minicomputer and also "the first modern personal computer." It had 2,048 12-bit words, and the first LINC was built in 1962.
DEC's founder, Ken Olsen, had worked with both it and a still earlier computer, the 18-bit 64,000-word TX-0, at MIT's Lincoln Laboratory.
Neither of these machines was mass-produced.
Applicability
Although the LINC computer was intended primarily for laboratory use, the PDP-5's 12-bit system had a far wider range of use. An example of DEC's "The success of the PDP-5 ... proved that a market for minicomputers did exist"
is:
"Data-processing computers have accomplished for mathematicians what the wheel did for transportation"
"Very reliable data was obtained with ..."
"A PDP-5 computer was used very successfully aboard Evergreen for ..."
all of which described the same PDP-5 used by the United States Coast Guard.
The architecture of the PDP-5 was specified by Alan Kotok and Gordon Bell; the principal logic designer was the young engineer Edson de Castro who went on later to found Data General.
Hardware
By contrast with the 4-cabinet PDP-1, the minimum configuration of the PDP-5 was a single 19-inch cabinet with "150 printed circuit board modules holding over 900 transistors." Additional cabinets were required to house many peripheral devices.
The minimum configuration weighed about .
The machine was offered with from 1,024 to 32,768 12-bit words of core memory. Addressing more than 4,096 words of memory required the addition of a Type 154 Memory Extension Control unit (in modern terms, a memory management unit); this allowed adding additional Type 155 4,096 word core memory modules.
Instruction set
Of the 12 bits in each word, exactly 3 were used for instruction op-codes.
The PDP-5's instruction set was later expanded in its successor, the PDP-8. The biggest c |
https://en.wikipedia.org/wiki/PDP-15 | The PDP-15 was the fifth and last of the 18-bit minicomputers produced by Digital Equipment Corporation. The PDP-1 was first delivered in December 1959 and the first PDP-15 was delivered in February 1970. More than 400 of these successors to the PDP-9 (and 9/L) were ordered within the first eight months.
In addition to operating systems, the PDP-15 has compilers for Fortran and ALGOL.
History
The 18-bit PDP systems preceding the PDP-15 were named PDP-1, PDP-4, PDP-7 and PDP-9.
The last PDP-15 was produced in 1979.
Hardware
The PDP-15 was DEC's only 18-bit machine constructed from TTL integrated circuits rather than discrete transistors, and, like every DEC 18-bit system could be equipped with:
an optional X-Y (point-plot or vector graphics) display.
a hardware floating-point option, with a 10x speedup, was offered.
up to 128Kwords of core main memory
Models
The PDP-15 models offered by DEC were:
PDP-15/10: a 4K-word paper-tape based system
PDP-15/20: 8K, added DECtape
PDP-15/30: 16K word, added memory protection and a foreground/background Monitor
PDP-15/35: Added a 524K-word fixed-head disk drive
PDP-15/40: 24K memory
PDP-15/50:
PDP-15/76
PDP-15/76: 15/40 plus PDP-11 frontend. The PDP-15/76 was a dual-processor system that shared memory with an attached PDP-11/05. The PDP-11 served as a peripheral processor and enabled use of Unibus peripherals.
Software
DECsys, RSX-15, and XVM/RSX were the operating systems supplied by DEC for the PDP-15. A batch processing monitor (BOSS-15: Batch Operating Software System) was also available.
DECsys
The first DEC-supplied mass-storage operating system available for the PDP-15 was DECsys, an interactive single-user system. This software was provided on a DECtape reel, of which copies were made for each user. This copied DECtape was then added to by the user, and thus was storage
for personal programs and data. A second DECtape was used as a scratch tape by the assembler and the Fortran compiler.
RSX-15
RSX- |
https://en.wikipedia.org/wiki/Encyclopedia%20of%20World%20Problems%20and%20Human%20Potential | The Encyclopedia of World Problems and Human Potential (EWPHP) is a research work published by the Union of International Associations (UIA). It is available online since 2000, and was previously available as a CD-ROM and as a three-volume book.
The EWPHP began under the direction of Anthony Judge in 1972 and eventually came to comprise more than 100,000 entries and 700,000 links, as well as hundreds of pages of introductory notes and commentaries on problems, strategies, values, concepts of human development, and various intellectual resources.
Contributors and history
The project was originally conceived in 1972 by James Wellesley-Wesley, who provided financial support through the foundation Mankind 2000, and Anthony Judge, by whom the work was orchestrated.
Work on the first edition started with funds from Mankind 2000, matching those of the UIA. The publisher Klaus Saur, of Munich, provided funds, in conjunction with those from the UIA, for work on the 2nd, 3rd, and 4th editions. Seed funding for the third volume of the 4th edition was also provided on behalf of Mankind 2000. In the nineties, seed funding was provided, again on behalf of Mankind 2000, for computer equipment which subsequently allowed the UIA to develop a website and make available for free the 1994–1995 edition of the EWPHP databases. The UIA, on the initiative of Nadia McLaren, a consultant ecologist who has been a primary editor for the EWPHP, instigated two multi-partner projects funded by the European Union, with matching funds from the UIA. The work done through those two projects, Ecolynx: Information Context for Biodiversity Conservation (mainly) and Interactive Health Ecology Access Links, resulted in a fifth, web-based edition of the EWPHP in 2000. Two other individuals supported the project: Robert Jungk of Mankind 2000, and Christian de Laet of the UIA.
EWPHP began as a processing of documents gathered from entities profiled in the Yearbook of International Organizations. The Uni |
https://en.wikipedia.org/wiki/List%20of%20finite%20simple%20groups | In mathematics, the classification of finite simple groups states that every finite simple group is cyclic, or alternating, or in one of 16 families of groups of Lie type, or one of 26 sporadic groups.
The list below gives all finite simple groups, together with their order, the size of the Schur multiplier, the size of the outer automorphism group, usually some small representations, and lists of all duplicates.
Summary
The following table is a complete list of the 18 families of finite simple groups and the 26 sporadic simple groups, along with their orders. Any non-simple members of each family are listed, as well as any members duplicated within a family or between families. (In removing duplicates it is useful to note that no two finite simple groups have the same order, except that the group A8 = A3(2) and A2(4) both have order 20160, and that the group Bn(q) has the same order as Cn(q) for q odd, n > 2. The smallest of the latter pairs of groups are B3(3) and C3(3) which both have order 4585351680.)
There is an unfortunate conflict between the notations for the alternating groups An and the groups of Lie type An(q). Some authors use various different fonts for An to distinguish them. In particular,
in this article we make the distinction by setting the alternating groups An in Roman font and the Lie-type groups An(q) in italic.
In what follows, n is a positive integer, and q is a positive power of a prime number p, with the restrictions noted. The notation (a,b) represents the greatest common divisor of the integers a and b.
Cyclic groups, Zp
Simplicity: Simple for p a prime number.
Order: p
Schur multiplier: Trivial.
Outer automorphism group: Cyclic of order p − 1.
Other names: Z/pZ, Cp
Remarks: These are the only simple groups that are not perfect.
Alternating groups, An, n > 4
Simplicity: Solvable for n < 5, otherwise simple.
Order: n!/2 when n > 1.
Schur multiplier: 2 for n = 5 or n > 7, 6 for n = 6 or 7; see Covering groups of the alt |
https://en.wikipedia.org/wiki/Bone%20canaliculus | Bone canaliculi are microscopic canals between the lacunae of ossified bone. The radiating processes of the osteocytes (called filopodia) project into these canals. These cytoplasmic processes are joined together by gap junctions. Osteocytes do not entirely fill up the canaliculi. The remaining space is known as the periosteocytic space, which is filled with periosteocytic fluid. This fluid contains substances too large to be transported through the gap junctions that connect the osteocytes.
In cartilage, the lacunae and hence, the chondrocytes, are isolated from each other. Materials picked up by osteocytes adjacent to blood vessels are distributed throughout the bone matrix via the canaliculi.
Dental canaliculi
The dental canaliculi (sometimes called dentinal tubules) are the blood supply of a tooth. Odontoblast process run in the canaliculi that transverse the dentin layer and are referred as dentinal tubules. The number and size of the canaliculi decrease as the tubules move away from the pulp and toward the enamel or cementum.
See also
Lacrimal canaliculi |
https://en.wikipedia.org/wiki/Double-chance%20function | In software engineering, a double-chance function is a software design pattern with a strong application in cross-platform and scalable development.
Examples
Computer graphics
Consider a graphics API with functions to DrawPoint, DrawLine, and DrawSquare. It is easy to see that DrawLine can be implemented solely in terms of DrawPoint, and DrawSquare can in turn be implemented through four calls to DrawLine. If you were porting this API to a new architecture you would have a choice: implement three different functions natively (taking more time to implement, but likely resulting in faster code), or write DrawPoint natively, and implement the others as described above using common, cross-platform, code. An important example of this approach is the X11 graphics system, which can be ported to new graphics hardware by providing a very small number of device-dependent primitives, leaving higher level functions to a hardware-independent layer.
The double-chance function is an optimal method of creating such an implementation, whereby the first draft of the port can use the "fast to market, slow to run" version with a common DrawPoint function, while later versions can be modified as "slow to market, fast to run". Where the double-chance pattern scores high is that the base API includes the self-supporting implementation given here as part of the null driver, and all other implementations are extensions of this. Consequently, the first port is, in fact, the first usable implementation.
One typical implementation in C++ could be:
class CBaseGfxAPI {
virtual void DrawPoint(int x, int y) = 0; /* Abstract concept for the null driver */
virtual void DrawLine(int x1, int y1, int x2, int y2) { /* DrawPoint() repeated */}
virtual void DrawSquare(int x1, int y1, int x2, int y2) { /* DrawLine() repeated */}
};
class COriginalGfxAPI : public CBaseGfxAPI {
virtual void DrawPoint(int x, int y) { /* The only necessary native calls */ }
virtual void DrawL |
https://en.wikipedia.org/wiki/Microsoft%20Compiled%20HTML%20Help | Microsoft Compiled HTML Help is a Microsoft proprietary online help format, consisting of a collection of HTML pages, an index and other navigation tools. The files are compressed and deployed in a binary format with the extension .CHM, for Compiled HTML. The format is often used for software documentation.
It was introduced as the successor to Microsoft WinHelp with the release of Windows 95 OSR 2.5 and consequently, Windows 98. Within the Windows NT family, the CHM file support is introduced in Windows NT 4.0 and is still supported in Windows 11. Although the format was designed by Microsoft, it has been successfully reverse-engineered and is now supported in many document viewer applications.
History
Microsoft has announced that they do not intend to add any new features to HTML Help.
File format
Help is delivered as a binary file with the .chm extension. It contains a set of HTML files, a hyperlinked table of contents, and an index file. The file format has been reverse-engineered and documentation of it is freely available.
The file starts with bytes "ITSF" (in ASCII), for "Info-Tech Storage Format", which is the internal name given by Microsoft to the generic storage file format used in with CHM files.
CHM files support the following features:
Data compression (using LZX)
Built-in search engine
Ability to merge multiple .chm help files
Extended character support, although it does not fully support Unicode.
Use in Windows applications
The Microsoft Reader's .lit file format is a modification of the HTML Help CHM format. CHM files are sometimes used for e-books.
Sumatra PDF supports viewing CHM documents since version 1.9.
Various applications, such as HTML Help Workshop and 7-Zip can decompile CHM files. The hh.exe utility on Windows and the extract_chmLib utility (a component of chmlib) on Linux can also decompile CHM files.
Microsoft's HTML Help Workshop and Compiler generate CHM files by instructions stored in a HTML Help project. The file name |
https://en.wikipedia.org/wiki/WinHelp | Microsoft WinHelp is a proprietary format for online help files that can be displayed by the Microsoft Help browser winhelp.exe or winhlp32.exe. The file format is based on Rich Text Format (RTF). It remained a popular Help platform from Windows 3.0 through Windows XP. WinHelp was removed in Windows Vista purportedly to discourage software developers from using the obsolete format and encourage use of newer help formats. Support for WinHelp files would eventually be removed entirely in Windows 10.
History
1990 – WinHelp 1.0 shipped with Windows 3.0.
1995 – WinHelp 4.0 shipped with Windows 95 / Windows NT.
2006 – Microsoft announced its intentions to phase out WinHelp as a supported platform. WinHelp is not part of Windows Vista out of the box. WinHelp files come in 16 bit and 32 bit types. Vista treats these files types differently. When starting an application that uses the 32 bit .hlp format, Windows warns that the format is no longer supported. A downloadable viewer for 32 bit .hlp files is available from the Microsoft Download Center. The 16 bit WinHelp files continue to display in Windows Vista (32 bit only) without the viewer download.
January 9, 2009 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows Server 2008 at the Microsoft Download Center.
October 14, 2009 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows 7 and Windows Server 2008 R2 at the Microsoft Download Center.
October 26, 2012 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows 8 at the Microsoft Download Center.
November 5, 2013 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows 8.1 at the Microsoft Download Center.
July 15, 2015 - Microsoft completely removed Windows Help from Windows 10. Attempting to open a .hlp file just brings users to a help page detailing that it was removed.
File format
A WinHelp file has a ".hlp" suffix. It |
https://en.wikipedia.org/wiki/Flag%20of%20Palestine | The flag of Palestine () is a tricolor of three equal horizontal stripes (black, white, and green from top to bottom) overlaid by a red triangle issuing from the hoist. This flag is derived from the Pan-Arab colors and is used to represent the State of Palestine and the Palestinian people. It was first adopted on 28 May 1964 by the Palestinian Liberation Organization. The flag day is celebrated on 30 September.
The flag is similar to that of Syria's Ba'ath Party, which uses the same shapes and colours but a 2:3 ratio as opposed to Palestine's 1:2, as well as the short-lived Arab Federation of Iraq and Jordan (which had an equilateral triangle at the hoist). It is also very similar to the Flag of Jordan and to the Flag of Western Sahara, all of which draw their inspiration from the Arab Revolt against Ottoman rule (1916–1918). The flag of the Arab Revolt had the same graphic form, but the colours were arranged differently (white on the bottom, rather than in the middle).
In 2021, President Mahmoud Abbas approved the annual lowering of the flag to lament the Balfour Declaration.
Origin
The flag used by the Arab Palestinian nationalists in the first half of the 20th century is the flag of the 1916 Arab Revolt. The origins of the flag are the subject of dispute and mythology. In one version, the colours were chosen by the Arab nationalist 'Literary Club' in Istanbul in 1909, based on the words of the 13th-century Arab poet Safi al-Din al-Hili:
Ask the high rising spears, of our aspirations
Bring witness the swords, did we lose hope
We are a band, honor halts our souls
Of beginning with harm, those who won't harm us
White are our deeds, black are our battles,
Green are our fields, red are our swords.
Another version credits the Young Arab Society, which was formed in Paris in 1911. Yet another version is that the flag was designed by Sir Mark Sykes of the British Foreign Office. Whatever the correct story, the flag was used by Sharif Hussein by 1917 at the latest a |
https://en.wikipedia.org/wiki/Silver%20staining | In pathology, silver staining is the use of silver to selectively alter the appearance of a target in microscopy of histological sections; in temperature gradient gel electrophoresis; and in polyacrylamide gels.
In traditional stained glass, silver stain is a technique to produce yellow to orange or brown shades (or green on a blue glass base), by adding a mixture containing silver compounds (notably silver nitrate), and firing lightly. It was introduced soon after 1800, and is the "stain" in the term "stained glass". Silver compounds are mixed with binding substances, applied to the surface of glass, and then fired in a furnace or kiln.
History
Camillo Golgi perfected silver staining for the study of the nervous system. Although the exact chemical mechanism by which this occurs is unknown, Golgi's method stains a limited number of cells at random in their entirety.
Silver staining was introduced by Kerenyi and Gallyas as a sensitive procedure to detect trace amounts of proteins in gels. The technique has been extended to the study of other biological macromolecules that have been separated in a variety of supports.
Classical Coomassie brilliant blue staining can usually detect a 50 ng protein band; silver staining increases the sensitivity typically 50 times.
Many variables can influence the color intensity and every protein has its own staining characteristics; clean glassware, pure reagents, and water of highest purity are the key points to successful staining.
Chemistry
Some cells are argentaffin. These reduce silver solution to metallic silver after formalin fixation. Other cells are argyrophilic. These reduce silver solution to metallic silver after being exposed to the stain that contains a reductant, for example hydroquinone or formalin.
Silver nitrate forms insoluble silver phosphate with phosphate ions; this method is known as the Von Kossa Stain. When subjected to a reducing agent, usually hydroquinone, it forms black elementary silver. This is use |
https://en.wikipedia.org/wiki/Vital%20stain | A vital stain in a casual usage may mean a stain that can be applied on living cells without killing them. Vital stains have been useful for diagnostic and surgical techniques in a variety of medical specialties. In supravital staining, living cells have been removed from an organism, whereas intravital staining is done by injecting or otherwise introducing the stain into the body. The term vital stain is used by some authors to refer to an intravital stain, and by others interchangeably with a supravital stain, the core concept being that the cell being examined is still alive. In a more strict sense, the term vital staining has a meaning contrasting with supravital staining. While in supravital staining the living cells take up the stain, in "vital staining" – the most accepted but apparently paradoxical meaning of this term, the living cells exclude the stain i.e. stain negatively and only the dead cells stain positively and thus viability can be assessed by counting the percentage of total cells that stain negatively. Very bulky or highly charged stains that don't cross live plasma membrane are used as vital stains and supravital stains are those that are either small or are pumped actively into live cells. Since supravital and intravital nature of the staining depends on the dye, a combination of supravital and vital dyes can also be used in a sophisticated way to better classify cells into distinct subsets (e.g. viable, dead, dying etc.).
List of common vital stains
Eosin dye exclusion
Propidium iodide, DNA stain that can differentiate necrotic, apoptotic and normal cells.
Trypan Blue, a living-cell exclusion dye
Triphenyl tetrazolium chloride
Erythrosine, which is Red No. 3 in food coloring, can be used as an exclusion dye.
7-Aminoactinomycin D used e.g. in flowcytometric studies of hematopoietic stem cell viability.
Other Vital Stains :
Janus Green B is a basic dye and vital stain used in histology.
See also
Cytopathology
Histology
Staining
Viability ass |
https://en.wikipedia.org/wiki/Sublinear%20function | In linear algebra, a sublinear function (or functional as is more often used in functional analysis), also called a quasi-seminorm or a Banach functional, on a vector space is a real-valued function with only some of the properties of a seminorm. Unlike seminorms, a sublinear function does not have to be nonnegative-valued and also does not have to be absolutely homogeneous. Seminorms are themselves abstractions of the more well known notion of norms, where a seminorm has all the defining properties of a norm that it is not required to map non-zero vectors to non-zero values.
In functional analysis the name Banach functional is sometimes used, reflecting that they are most commonly used when applying a general formulation of the Hahn–Banach theorem.
The notion of a sublinear function was introduced by Stefan Banach when he proved his version of the Hahn-Banach theorem.
There is also a different notion in computer science, described below, that also goes by the name "sublinear function."
Definitions
Let be a vector space over a field where is either the real numbers or complex numbers
A real-valued function on is called a (or a if ), and also sometimes called a or a , if it has these two properties:
Positive homogeneity/Nonnegative homogeneity: for all real and all
This condition holds if and only if for all positive real and all
Subadditivity/Triangle inequality: for all
This subadditivity condition requires to be real-valued.
A function is called or if for all although some authors define to instead mean that whenever these definitions are not equivalent.
It is a if for all
Every subadditive symmetric function is necessarily nonnegative.
A sublinear function on a real vector space is symmetric if and only if it is a seminorm.
A sublinear function on a real or complex vector space is a seminorm if and only if it is a balanced function or equivalently, if and only if for every unit length scalar (satisfying ) and ever |
https://en.wikipedia.org/wiki/Managed%20code | Managed code is computer program code that requires and will execute only under the management of a Common Language Infrastructure (CLI); Virtual Execution System (VES); virtual machine, e.g. .NET, CoreFX, or .NET Framework; Common Language Runtime (CLR); or Mono. The term was coined by Microsoft.
Managed code is the compiler output of source code written in one of over twenty high-level programming languages, including C#, J# and Visual Basic .NET.
Terminology
The distinction between managed and unmanaged code is prevalent and only relevant when developing applications that interact with CLR implementations. Since many older programming languages have been ported to the CLR, the differentiation is needed to identify managed code, especially in a mixed setup. In this context, code that does not rely on the CLR is termed "unmanaged".
A source of confusion was created when Microsoft started connecting the .NET Framework with C++, and the choice of how to name the Managed Extensions for C++. It was first named Managed C++ and then renamed to C++/CLI. The creator of the C++ programming language and member of the C++ standards committee, Bjarne Stroustrup, even commented on this issue, "On the difficult and controversial question of what the CLI binding/extensions to C++ is to be called, I prefer C++/CLI as a shorthand for "The CLI extensions to ISO C++". Keeping C++ as part of the name reminds people what is the base language and will help keep C++ a proper subset of C++ with the C++/CLI extensions."
Uses
The Microsoft Visual C++ compiler can produce both managed code, running under CLR, or unmanaged binaries, running directly on Windows.
Benefits of using managed code include programmer convenience (by increasing the level of abstraction, creating smaller models) and enhanced security guarantees, depending on the platform (including the VM implementation). There are many historical examples of code running on virtual machines, such as the language UCSD Pascal usin |
https://en.wikipedia.org/wiki/Zoogeography | Zoogeography is the branch of the science of biogeography that is concerned with geographic distribution (present and past) of animal species.
As a multifaceted field of study, zoogeography incorporates methods of molecular biology, genetics, morphology, phylogenetics, and Geographic Information Systems (GIS) to delineate evolutionary events within defined regions of study around the globe. Once proposed by Alfred Russel Wallace, known to be the father of zoogeography, phylogenetic affinities can be quantified among zoogeographic regions, further elucidating the phenomena surrounding geographic distributions of organisms and explaining evolutionary relationships of taxa.
Advancements in molecular biology and theory of evolution within zoological research has unraveled questions concerning speciation events and has expanded phylogenic relationships amongst taxa. Integration of phylogenetics with GIS provides a means for communicating evolutionary origins through cartographic design. Related research linking phylogenetics and GIS has been conducted in areas of the southern Atlantic, Mediterranean, and Pacific Oceans. Recent innovations in DNA bar-coding, for example, have allowed for explanations of phylogenetic relationships within two families of marine venomous fishes, scorpaenidae and tetraodontidae, residing in the Andaman Sea. Continued efforts to understand species evolutionary divergence articulated in the geologic time scale based on fossil records for killifish (Aphanius and Aphanolebias) in locales of the Mediterranean and Paratethys areas revealed climatological influences during the Miocene Further development of research within zoogeography has expanded upon knowledge of the productivity of South Atlantic ocean regions and distribution of organisms in analogous regions, providing both ecological and geographic data to supply a framework for the taxonomic relationships and evolutionary branching of benthic polychaetes.
Modern-day zoogeography also plac |
https://en.wikipedia.org/wiki/Carvone | Carvone is a member of a family of chemicals called terpenoids. Carvone is found naturally in many essential oils, but is most abundant in the oils from seeds of caraway (Carum carvi), spearmint (Mentha spicata), and dill.
Uses
Both carvones are used in the food and flavor industry. R-(−)-Carvone is also used for air freshening products and, like many essential oils, oils containing carvones are used in aromatherapy and alternative medicine. S-(+)-Carvone has shown a suppressant effect against high-fat diet induced weight gain in mice.
Food applications
As the compound most responsible for the flavor of caraway, dill and spearmint, carvone has been used for millennia in food. Wrigley's Spearmint Gum and spearmint flavored Life Savers are major users of natural spearmint oil from Mentha spicata. Caraway seed is extracted with alcohol to make the European drink Kümmel.
Agriculture
S-(+)-Carvone is also used to prevent premature sprouting of potatoes during storage, being marketed in the Netherlands for this purpose under the name Talent.
Insect control
R-(−)-Carvone has been approved by the U.S. Environmental Protection Agency for use as a mosquito repellent.
Organic synthesis
Carvone is available inexpensively in both enantiomerically pure forms, making it an attractive starting material for the asymmetric total synthesis of natural products. For example, (S)-(+)-carvone was used to begin a 1998 synthesis of the terpenoid quassin:<ref>(a) Shing, T. K. M.; Jiang, Q; Mak, T. C. W. J. Org. Chem. 1998, 63, 2056-2057. (b) Shing, T. K. M.; Tang, Y. J. Chem. Soc. Perkin Trans. 1 1994, 1625.</ref>
Stereoisomerism and odor
Carvone forms two mirror image forms or enantiomers: R-(−)-carvone, has a sweetish minty smell, like spearmint leaves. Its mirror image, S-(+)-carvone, has a spicy aroma with notes of rye, like caraway seeds. The fact that the two enantiomers are perceived as smelling different is evidence that olfactory receptors must respond more strongly to on |
https://en.wikipedia.org/wiki/Momentum%20theory | In fluid dynamics, momentum theory or disk actuator theory is a theory describing a mathematical model of an ideal actuator disk, such as a propeller or helicopter rotor, by W.J.M. Rankine (1865), Alfred George Greenhill (1888) and Robert Edmund Froude (1889).
The rotor is modeled as an infinitely thin disc, inducing a constant velocity along the axis of rotation. The basic state of a helicopter is hovering. This disc creates a flow around the rotor. Under certain mathematical premises of the fluid, there can be extracted a mathematical connection between power, radius of the rotor, torque and induced velocity. Friction is not included.
For a stationary open rotor with no outer duct, such as a helicopter in hover, the power required to produce a given thrust is:
where:
T is the thrust
is the density of air (or other medium)
A is the area of the rotor disc
P is power
A device which converts the translational energy of the fluid into rotational energy of the axis or vice versa is called a Rankine disk actuator. The real life implementations of such devices include marine and aviation propellers, windmills, helicopter rotors, centrifugal pumps, wind turbines, turbochargers and chemical agitators.
See also
Blade element theory
Circulation (fluid dynamics)
Disk loading
Kutta–Joukowski theorem |
https://en.wikipedia.org/wiki/S3%20Savage | Savage was a product-line of PC graphics chipsets designed by S3.
Graphics Processors
Savage 3D
At the 1998 E3 Expo S3 introduced the first Savage product, Savage3D. Compared to its ViRGE-derived predecessor (Trio3D), Savage3D was a technological leap forward. Its innovative feature-set included the following:
"free" (single-cycle) trilinear-filtering
hardware motion-compensation and subpicture alpha-blending (MPEG-2 video)
integrated NTSC/PAL TV-encoder, (optional) Macrovision
S3 Texture Compression (S3TC)
multi-tap X/Y interpolating front-end (BITBLT) and back-end (overlay) video-scaler
Unfortunately for S3, deliveries of the Savage3D were hampered by poor manufacturing yields. Only one major board-vendor, Hercules, made any real effort to ship a Savage3D product. S3's yield problems forced Hercules to hand pick usable chips from the silicon wafers. Combined with poor drivers and the chip's lack of multitexturing support, the Savage3D failed in the market.
Savage 3D also dropped support for the S3D API from the S3 ViRGE predecessor.
In early 1999, S3 retired the Savage3D and released the Savage4 family. Many of the Savage3D's limitations were addressed by the Savage 4 chipset.
Savage4
Savage4 was an evolution of Savage 3D technology in many ways. S3 refined the chip, fixing hardware bugs and streamlining the chip for both cost reduction and performance. They added single-pass multi-texturing, meaning the board could sample 2 textures per pixel in one pass (not one clock cycle) through the rendering engine instead of halving its texture fillrate in dual-textured games like Savage 3D. Savage4 supported the then-new AGP 4X although at the older 3.3 voltage specification. It was manufactured on a 250 nm process, like Savage 3D. The graphics core was clocked at 125 MHz, with the board's SDRAM clocked at either 125 MHz or 143 MHz (Savage4 Pro). They could be equipped with 8-32 MiB memory. And while an integrated TV encoder was dropped, the DVD accelerat |
https://en.wikipedia.org/wiki/Semiquinone | Semiquinones (or ubisemiquinones, if their origin is ubiquinone) are free radicals resulting from the removal of one hydrogen atom with its electron during the process of dehydrogenation of a hydroquinone, such as hydroquinone itself or catechol, to a quinone or alternatively the addition of a single hydrogen atom with its electron to a quinone. Semiquinones are highly unstable.
E.g. ubisemiquinone is the first of two stages in reducing the supplementary form of CoQ10 (ubiquinone) to its active form ubiquinol. |
https://en.wikipedia.org/wiki/Videocipher | VideoCipher is a brand name of analog scrambling and de-scrambling equipment for cable and satellite television invented primarily to enforce Television receive-only (TVRO) satellite equipment to only receive TV programming on a subscription basis.
The second version of Videocipher, Videocipher II, was the primary encryption scheme used by major cable TV programmers to prevent TVRO owners from receiving free terrestrial television programming. It was especially notable due to the widespread compromise of its encryption scheme.
Background
Satellite providers
Though the first half of the 1980s, HBO, Cinemax and other premium television providers with analog satellite transponders faced a fast growing market of TVRO equipment owners. Satellite television consumers could watch these services simply by pointing their dish at a satellite, and tuning into the provider's transponder. Two open questions existed about this practice: whether the Communications Act of 1934 applied as a case of "unauthorized reception" by TVRO consumers; and to what it extent it was legal for a service provider to encrypt their signals in an effort to prevent its reception.
The Cable Communications Policy Act of 1984 clarified all of these matters, making the following legal:
Reception of unencrypted satellite signals by a consumer
Reception of encrypted satellite signals by a consumer, when they have received authorization to legally decrypt it
This created a framework for the wide deployment of encryption on analog satellite signals. It further created a framework (and implicit mandate to provide) subscription services to TVRO consumers to allow legal decryption of those signals. HBO and Cinemax became the first two services to announce intent to encrypt their satellite feeds late in 1984.
Videocipher technology
Videocipher was invented in 1983 by Linkabit Corporation (later bought out by M/A-COM in 1985, operated as M/A-COM Linkabit). In the mid 1980s, M/A-COM began divesting divis |
https://en.wikipedia.org/wiki/Human%20ecosystem | Human ecosystems are human-dominated ecosystems of the anthropocene era that are viewed as complex cybernetic systems by conceptual models that are increasingly used by ecological anthropologists and other scholars to examine the ecological aspects of human communities in a way that integrates multiple factors as economics, sociopolitical organization, psychological factors, and physical factors related to the environment.
A human ecosystem has three central organizing concepts: human environed unit (an individual or group of individuals), environment, interactions and transactions between and within the components. The total environment includes three conceptually distinct, but interrelated environments: the natural, human constructed, and human behavioral. These environments furnish the resources and conditions necessary for life and constitute a life-support system.
Further reading
Basso, Keith 1996 “Wisdom Sits in Places: Landscape and Language among the Western Apache.” Albuquerque: University of New Mexico Press.
Douglas, Mary 1999 “Implicit Meanings: Selected Essays in Anthropology.” London and New York: Routledge, Taylor & Francis Group.
Nadasdy, Paul 2003 “Hunters and Bureaucrats: Power, Knowledge, and Aboriginal-State Relations in the Southwest Yukon.” Vancouver and Toronto: UBC Press. |
https://en.wikipedia.org/wiki/Rugae | In anatomy, rugae are a series of ridges produced by folding of the wall of an organ. Most commonly rugae refers to the gastric rugae of the internal surface of the stomach.
Function
A purpose of the gastric rugae is to allow for expansion of the stomach after the consumption of foods and liquids. This expansion increases the volume of the stomach to hold larger amounts of food. The folds also result in greater surface area, allowing the stomach to absorb nutrients more quickly.
Location
Rugae can appear in the following locations in humans:
Wrinkles of the labia and scrotum
Hard palate immediately behind the upper anterior teeth
Inside the urinary bladder
Vagina
Gallbladder
Inside the stomach
Inside the rectum
Difference between rugae and plicae
With few exceptions (e.g. the scrotum), rugae are only evident when an organ or tissue is deflated or relaxed. For example, rugae are evident within the stomach when it is deflated. However, when the stomach distends, the rugae unfold to allow for the increase in volume. On the other hand, plicae remain folded regardless of distension as is evident within the plicae of the small intestine walls. |
https://en.wikipedia.org/wiki/Yakov%20Sinai | Yakov Grigorevich Sinai (; born September 21, 1935) is a Russian–American mathematician known for his work on dynamical systems. He contributed to the modern metric theory of dynamical systems and connected the world of deterministic (dynamical) systems with the world of probabilistic (stochastic) systems. He has also worked on mathematical physics and probability theory. His efforts have provided the groundwork for advances in the physical sciences.
Sinai has won several awards, including the Nemmers Prize, the Wolf Prize in Mathematics and the Abel Prize. He serves as the professor of mathematics at Princeton University since 1993 and holds the position of Senior Researcher at the Landau Institute for Theoretical Physics in Moscow, Russia.
Biography
Yakov Grigorevich Sinai was born into a Russian Jewish academic family on September 21, 1935, in Moscow, Soviet Union (now Russia). His parents, Nadezda Kagan and Gregory Sinai, were both microbiologists. His grandfather, Veniamin Kagan, headed the Department of
Differential Geometry at Moscow State University and was a major influence on Sinai's life.
Sinai received his bachelor's and master's degrees from Moscow State University. In 1960, he earned his Ph.D., also from Moscow State; his adviser was Andrey Kolmogorov. Together with Kolmogorov, he showed that even for "unpredictable" dynamic systems, the level of unpredictability of motion can be described mathematically. In their idea, which became known as Kolmogorov–Sinai entropy, a system with zero entropy is entirely predictable, while a system with non-zero entropy has an unpredictability factor directly related to the amount of entropy.
In 1963, Sinai introduced the idea of dynamical billiards, also known as "Sinai Billiards". In this idealized system, a particle bounces around inside a square boundary without loss of energy. Inside the square is a circular wall, of which the particle also bounces off. He then proved that for most initial trajector |
https://en.wikipedia.org/wiki/Christian%20views%20on%20cloning | Christians take multiple positions in the debate on the morality of human cloning. Since Dolly the sheep was successfully cloned on 5 July 1996, and the possibility of cloning humans became a reality, Christian leaders have been pressed to take an ethical stance on its morality. While many Christians tend to disagree with the practice, such as Roman Catholics and a majority of fundamentalist pastors, including Southern Baptists, the views taken by various other Christian denominations are diverse and often conflicting. It is hard to pinpoint any one, definite stance of the Christian religion, since there are so many Christian denominations and so few official statements from each of them concerning the morality of human cloning.
There are certain Protestant denominations that do not disagree with the acceptability of human cloning. Mary Seller, for example, a member of the Church of England's Board of Social Responsibility and a professor of developmental genetics, states, "Cloning, like all science, must be used responsibly. Cloning humans is not desirable. But cloning sheep has its uses." On the other hand, according to a survey of Christian fundamentalist pastors, responses indicated a "common account of human cloning as primarily reproductive in nature, proscribed by its violation of God's will and role." Many of these pastors acknowledged the reason for this violation being rooted in the religiously motivated view that human cloning is an example of scientists 'playing God.'" It is not only this that many Christians are concerned about, however; other concerns include whether the dignity of the human person is overlooked, as well as the role of the parents as co-creators. All of these things may contribute to why many fundamentalist Christian pastors see human reproductive cloning as simply "forbidden territory."
Understanding cloning
Some scientists do argue that the plurality of views comes from the differing understandings of what exactly human cloning is |
https://en.wikipedia.org/wiki/Cue%20card | Cue cards, also known as note cards, are cards with words written on them that help actors and speakers remember what they have to say. They are typically used in television productions where they can be held off-camera and are unseen by the audience. Cue cards are still currently being used on many late night talk shows including The Tonight Show Starring Jimmy Fallon, Late Night with Seth Meyers and formerly on Conan as well as variety and sketch comedy shows like Saturday Night Live due to the practice of last minute script changes. Many other TV shows, including game and reality shows, still use cue cards due to their mobility, as a teleprompter only allows the actor or broadcaster to look directly into the camera.
History
Cue cards were originally used to aid aging actors. One early use was by John Barrymore in the late 1930s.
Cue cards however did not become widespread until 1949 when Barney McNulty a CBS page and former military pilot, was asked to write ailing actor Ed Wynn's script lines on large sheets of paper to help him remember his script. McNulty volunteered for this duty because his training as a pilot taught him to write very quickly and clearly. McNulty soon saw the necessity of this concept and formed the company "Ad-Libs". McNulty continued to be Bob Hope's personal cue card man until he stopped performing. McNulty who died in 2000 at the age of 77 was known in Hollywood as the "Cue-Card King".
Marlon Brando was also a frequent user of cue cards, feeling that this helped bring realism and spontaneity to his performances, instead of giving the impression that he was merely reciting a writer's speech. During production of the film Last Tango in Paris, he had cue cards posted about the set, although director Bernardo Bertolucci declined his request to have lines written on actress Maria Schneider's rear end. Tony Mendez became a minor celebrity for his cue card work on the Late Show with David Letterman.
See also
Wally Feresten - cue card h |
https://en.wikipedia.org/wiki/143%20%28number%29 | 143 (one hundred [and] forty-three) is the natural number following 142 and preceding 144.
In mathematics
143 is the sum of seven consecutive primes (11 + 13 + 17 + 19 + 23 + 29 + 31). But this number is never the sum of an integer and its base 10 digits, making it a self number. It is also the product of a twin prime pair (11 × 13).
Every positive integer is the sum of at most 143 seventh powers (see Waring's problem).
143 is the difference in the first exception to the pattern shown below:
.
In the military
Vickers Type 143 was a British single-seat fighter biplane in 1929
United States Air Force 143d Airlift Wing airlift unit at Quonset Point, Rhode Island
was a United States Navy during World War II
was a United States Navy during World War II
was a United States Navy during World War II
was a United States Navy in World War II
was a United States Navy patrol boat
was a United States Navy during the Cuban Missile Crisis
was a United States Navy during World War I
In transportation
London Buses route 143 is a Transport for London contracted bus route in London
Air Canada Flight 143, landed at Gimli, Manitoba Air Force Base after gliding after running out of fuel on July 23, 1983
Philippine Airlines Flight 143 exploded prior to takeoff on May 11, 1990, at Manila Airport
Bristol Type 143 was a British twin-engined monoplane aircraft of the Bristol Aeroplane Company
British Rail Class 143 diesel multiple unit, part of the Pacer family of trains introduced in 1985
East 143rd Street–St. Mary's Street station on the IRT Pelham Line of the New York City Subway
143rd Street station on Metra's SouthWest Service in Orland Park, Illinois
In media
Musicians Ray J and Bobby Brackins wrote the song "143"
On Mister Rogers' Neighborhood: "Transformations", 143 is used to mean "I love you". 1 meaning I for 1 letter, 4 meaning Love for the 4 letters, and 3 meaning You for the 3 letters. Reportedly, Fred Rogers maintained his weight at exactly |
https://en.wikipedia.org/wiki/Universality%20%28dynamical%20systems%29 | In statistical mechanics, universality is the observation that there are properties for a large class of systems that are independent of the dynamical details of the system. Systems display universality in a scaling limit, when a large number of interacting parts come together. The modern meaning of the term was introduced by Leo Kadanoff in the 1960s, but a simpler version of the concept was already implicit in the van der Waals equation and in the earlier Landau theory of phase transitions, which did not incorporate scaling correctly.
The term is slowly gaining a broader usage in several fields of mathematics, including combinatorics and probability theory, whenever the quantitative features of a structure (such as asymptotic behaviour) can be deduced from a few global parameters appearing in the definition, without requiring knowledge of the details of the system.
The renormalization group provides an intuitively appealing, albeit mathematically non-rigorous, explanation of universality. It classifies operators in a statistical field theory into relevant and irrelevant. Relevant operators are those responsible for perturbations to the free energy, the imaginary time Lagrangian, that will affect the continuum limit, and can be seen at long distances. Irrelevant operators are those that only change the short-distance details. The collection of scale-invariant statistical theories define the universality classes, and the finite-dimensional list of coefficients of relevant operators parametrize the near-critical behavior.
Universality in statistical mechanics
The notion of universality originated in the study of phase transitions in statistical mechanics. A phase transition occurs when a material changes its properties in a dramatic way: water, as it is heated boils and turns into vapor; or a magnet, when heated, loses its magnetism. Phase transitions are characterized by an order parameter, such as the density or the magnetization, that changes as a function of |
https://en.wikipedia.org/wiki/J%C3%BCrgen%20Moser | Jürgen Kurt Moser (July 4, 1928 – December 17, 1999) was a German-American mathematician, honored for work spanning over four decades, including Hamiltonian dynamical systems and partial differential equations.
Life
Moser's mother Ilse Strehlke was a niece of the violinist and composer Louis Spohr. His father was the neurologist Kurt E. Moser (July 21, 1895 – June 25, 1982), who was born to the merchant Max Maync (1870–1911) and Clara Moser (1860–1934). The latter descended from 17th century French Huguenot immigrants to Prussia. Jürgen Moser's parents lived in Königsberg, German empire and resettled in Stralsund, East Germany as a result of the second world war. Moser attended the Wilhelmsgymnasium (Königsberg) in his hometown, a high school specializing in mathematics and natural sciences education, from which David Hilbert had graduated in 1880. His older brother Friedrich Robert Ernst (Friedel) Moser (August 31, 1925 – January 14, 1945) served in the German Army and died in Schloßberg during the East Prussian offensive.
Moser married the biologist Dr. Gertrude C. Courant (Richard Courant's daughter, Carl Runge's granddaughter and great-granddaughter of Emil DuBois-Reymond) on September 10, 1955 and took up permanent residence in New Rochelle, New York in 1960, commuting to work in New York City. In 1980 he moved to Switzerland, where he lived in Schwerzenbach near Zürich. He was a member of the Akademisches Orchester Zürich. He was survived by his younger brother, the photographic printer and processor Klaus T. Moser-Maync from Northport, New York, his wife, Gertrude Moser from Seattle, their daughters, the theater designer Nina Moser from Seattle and the mathematician Lucy I. Moser-Jauslin from Dijon, and his stepson, the lawyer Richard D. Emery from New York City. Moser played the piano and the cello, performing chamber music since his childhood in the tradition of a musical family, where his father played the violin and his mother the piano. He was a lifelo |
https://en.wikipedia.org/wiki/Contractible%20space | In mathematics, a topological space X is contractible if the identity map on X is null-homotopic, i.e. if it is homotopic to some constant map. Intuitively, a contractible space is one that can be continuously shrunk to a point within that space.
Properties
A contractible space is precisely one with the homotopy type of a point. It follows that all the homotopy groups of a contractible space are trivial. Therefore any space with a nontrivial homotopy group cannot be contractible. Similarly, since singular homology is a homotopy invariant, the reduced homology groups of a contractible space are all trivial.
For a topological space X the following are all equivalent:
X is contractible (i.e. the identity map is null-homotopic).
X is homotopy equivalent to a one-point space.
X deformation retracts onto a point. (However, there exist contractible spaces which do not strongly deformation retract to a point.)
For any path-connected space Y, any two maps f,g: Y → X are homotopic.
For any space Y, any map f: Y → X is null-homotopic.
The cone on a space X is always contractible. Therefore any space can be embedded in a contractible one (which also illustrates that subspaces of contractible spaces need not be contractible).
Furthermore, X is contractible if and only if there exists a retraction from the cone of X to X.
Every contractible space is path connected and simply connected. Moreover, since all the higher homotopy groups vanish, every contractible space is n-connected for all n ≥ 0.
Locally contractible spaces
A topological space X is locally contractible at a point x if for every neighborhood U of x there is a neighborhood V of x contained in U such that the inclusion of V is nulhomotopic in U. A space is locally contractible if it is locally contractible at every point. This definition is occasionally referred to as the "geometric topologist's locally contractible," though is the most common usage of the term. In Hatcher's standard Algebraic Topology text, |
https://en.wikipedia.org/wiki/Hans%20Lewy | Hans Lewy (20 October 1904 – 23 August 1988) was a Jewish American mathematician, known for his work on partial differential equations and on the theory of functions of several complex variables.
Life
Lewy was born in Breslau, Silesia, on October 20, 1904. He began his studies at the University of Göttingen in 1922, after being advised to avoid the more local University of Breslau because it was too old-fashioned, supporting himself during the Weimar hyperinflation by a side job doing railroad track maintenance. At Göttingen, he studied both mathematics and physics; his teachers there included Max Born, Richard Courant, James Franck, David Hilbert, Edmund Landau, Emmy Noether, and Alexander Ostrowski. He earned his doctorate in 1926, at which time he and his friend Kurt Otto Friedrichs both became assistants to Courant and privatdozents at Göttingen. The famous Courant-Friedrichs-Lewy condition originated from that time in 1928.
At the recommendation of Courant, Lewy was granted a Rockefeller Fellowship, which he used in 1929 to travel to Rome and study algebraic geometry with Tullio Levi-Civita and Federigo Enriques, and then in 1930 to travel to Paris, where he attended the seminar of Jacques Hadamard. After Hitler's election as chancellor in 1933, Lewy was advised by Herbert Busemann to leave Germany again. He was offered a position in Madrid, but declined it, fearing for the future there under Francisco Franco. He revisited Italy and France, but then at the invitation of the Emergency Committee in Aid of Displaced Foreign Scholars and with the assistance of Hadamard found a two-year position in America at Brown University. At the end of that term, in 1935, he moved to the University of California, Berkeley.
During World War II, Lewy obtained a pilot's license, but then worked at the Aberdeen Proving Ground. He married Helen Crosby in 1947.
In 1950, Lewy was fired from Berkeley for refusing to sign a loyalty oath. He taught at Harvard University and Stanford |
https://en.wikipedia.org/wiki/Current-mode%20logic | Current mode logic (CML), or source-coupled logic (SCL), is a digital design style used both for logic gates and for board-level digital signaling of digital data.
The basic principle of CML is that current from a constant current generator is steered between two alternate paths depending on whether a logic zero or logic one is being represented. Typically, the generator is connected to the two sources of a pair of differential FETs, with the two paths being their two drains. The bipolar equivalent emitter-coupled logic (ECL) operates similarly, with the output being taken from the collectors of the BJT transistors.
As a differential PCB-level interconnect, it is intended to transmit data at speeds between 312.5 Mbit/s and 3.125 Gbit/s across standard printed circuit boards.
The transmission is point-to-point, unidirectional, and is usually terminated at the destination with 50 Ω resistors to Vcc on both differential lines. CML is frequently used in interfaces to fiber optic components. The principle difference between CML and ECL as a link technology is the output impedance of the driver stage: the emitter follower of ECL has a low resistance of around 5 Ω whereas CML connects to the drains of the driving transistors, that have a high impedance, and so the impedance of the pull up/down network (typically 50 Ω resistive) is the effective output impedance. Matching this drive impedance close to the driven transmission line's characteristic impedance greatly reduces undesirable ringing.
CML signals have also been found useful for connections between modules. CML is the physical layer used in DVI, HDMI and FPD-Link III video links, the interfaces between a display controller and a monitor.
In addition, CML has been widely used in high-speed integrated systems, such as for serial data transceivers and frequency synthesizers in telecommunication systems.
Operation
The fast operation of CML circuits is mainly due to their lower output voltage swing compared to the s |
https://en.wikipedia.org/wiki/Integrated%20device%20manufacturer | An integrated device manufacturer (IDM) is a semiconductor company which designs, manufactures, and sells integrated circuit (IC) products.
IDM is often used to refer to a company which handles semiconductor manufacturing in-house, compared to a fabless semiconductor company, which outsources production to a third-party semiconductor fabrication plant.
Examples of IDMs are Intel, Samsung, and Texas Instruments, examples of fabless companies are AMD, Nvidia, and Qualcomm, and examples of pure play foundries are GlobalFoundries, TSMC, and UMC.
Due to the dynamic nature of the semiconductor industry, the term IDM has become less accurate than when it was coined.
OSATs
The term OSATs means "outsourced semiconductor assembly and test providers". OSATs have dominated IC packaging and testing.
Fabless operations
The terms fabless (fabrication-less), foundry, and IDM are now used to describe the role a company has in a business relationship. For example, Freescale owns and operates fabrication facilities (fab) where it manufactures many chip product lines, as a traditional IDM would. Yet it is known to contract with merchant foundries for other products, as would fabless companies.
Manufacturers
Many electronic manufacturing companies engage in business that would qualify them as an IDM:
Analog Devices
ams AG
Belling
Changxin Memory Technologies
Cypress Semiconductor
CR Microelectronics
Fujitsu
Good-Ark Electronics
Hitachi
IBM
IM Flash Technologies
Infineon
Intersil
Intel
LSI Corporation
Matsushita
Maxim Integrated Products
Micron Technology
Mitsubishi
National Semiconductor
Nexperia
NXP (formerly Philips + Freescale Semiconductors)
OMMIC
ON Semiconductor
Pericom
Qorvo
Renesas (formerly NEC semiconductor)
Samsung
SK Hynix
STMicroelectronics
Sony
Texas Instruments
Tsinghua Unigroup
Toshiba
Reading
Understanding fabless IC technology By Jeorge S. Hurtarte, Evert A. Wolsheimer, Lisa M. Tafoya 1.4.1 Integrated device manufacturer Pag |
https://en.wikipedia.org/wiki/Biodiversity%20of%20New%20Zealand | The biodiversity of New Zealand, a large island country located in the south-western Pacific Ocean, is varied and distinctive. The species of New Zealand accumulated over many millions of years as lineages evolved in the local circumstances. New Zealand's pre-human biodiversity exhibited high levels of species endemism, but has experienced episodes of biological turnover. Global extinction approximately 65 Ma (million years ago) resulted in the loss of fauna such as non-avian dinosaurs, pterosaurs and marine reptiles e.g. mosasaurs, elasmosaurs and plesiosaurs. The ancient fauna is not well known, but at least one species of terrestrial mammal existed in New Zealand around 19 Ma. For at least several million years before the arrival of human and commensal species, the islands had no terrestrial mammals except for bats and seals, the main component of the terrestrial fauna being insects and birds. As recently as the 14th century a component has been introduced by humans, including many terrestrial mammals.
New Zealand has developed a national Biodiversity Action Plan to address conservation of considerable numbers of threatened flora and fauna within New Zealand.
Evolution
The break-up of the supercontinent of Gondwana left the resulting continents and microcontinents with shared biological affinities. Zealandia (the continental crust from which New Zealand and New Caledonia later developed) began to move away from Antarctic Gondwana 85 Ma ago, the break being complete by 66 Ma ago. It has been moving northwards since then, changing both in relief and climate. About 23 million years ago New Zealand was mostly underwater. One estimate suggests just 18% of the present surface area remained above the water. However geological evidence does not rule out the possibility that it was entirely submerged, or at least restricted to small islands. Today about 93% of the Zealandian continent remains below the sea. Several elements of Gondwana biota are present in New Zealand |
https://en.wikipedia.org/wiki/Moore%27s%20second%20law | Rock's law or Moore's second law, named for Arthur Rock or Gordon Moore, says that the cost of a semiconductor chip fabrication plant doubles every four years. As of 2015, the price had already reached about 14 billion US dollars.
Rock's law can be seen as the economic flip side to Moore's (first) law – that the number of transistors in a dense integrated circuit doubles every two years. The latter is a direct consequence of the ongoing growth of the capital-intensive semiconductor industry— innovative and popular products mean more profits, meaning more capital available to invest in ever higher levels of large-scale integration, which in turn leads to the creation of even more innovative products.
The semiconductor industry has always been extremely capital-intensive, with ever-dropping manufacturing unit costs. Thus, the ultimate limits to growth of the industry will constrain the maximum amount of capital that can be invested in new products; at some point, Rock's Law will collide with Moore's Law.
It has been suggested that fabrication plant costs have not increased as quickly as predicted by Rock's law – indeed plateauing in the late 1990s – and also that the fabrication plant cost per transistor (which has shown a pronounced downward trend) may be more relevant as a constraint on Moore's Law.
See also
Semiconductor device fabrication
Fabless manufacturing
Wirth's law, an analogous law about software complicating over time
Semiconductor consolidation |
https://en.wikipedia.org/wiki/Nagle%27s%20algorithm | Nagle's algorithm is a means of improving the efficiency of TCP/IP networks by reducing the number of packets that need to be sent over the network. It was defined by John Nagle while working for Ford Aerospace. It was published in 1984 as a Request for Comments (RFC) with title Congestion Control in IP/TCP Internetworks in .
The RFC describes what Nagle calls the "small-packet problem", where an application repeatedly emits data in small chunks, frequently only 1 byte in size. Since TCP packets have a 40-byte header (20 bytes for TCP, 20 bytes for IPv4), this results in a 41-byte packet for 1 byte of useful information, a huge overhead. This situation often occurs in Telnet sessions, where most keypresses generate a single byte of data that is transmitted immediately. Worse, over slow links, many such packets can be in transit at the same time, potentially leading to congestion collapse.
Nagle's algorithm works by combining a number of small outgoing messages and sending them all at once. Specifically, as long as there is a sent packet for which the sender has received no acknowledgment, the sender should keep buffering its output until it has a full packet's worth of output, thus allowing output to be sent all at once.
Algorithm
The RFC defines the algorithm as
inhibit the sending of new TCP segments when new outgoing data arrives from the user if any previously transmitted data on the connection remains unacknowledged.
Where MSS is the maximum segment size, the largest segment that can be sent on this connection, and the window size is the currently acceptable window of unacknowledged data, this can be written in pseudocode as
if there is new data to send then
if the window size ≥ MSS and available data is ≥ MSS then
send complete MSS segment now
else
if there is unconfirmed data still in the pipe then
enqueue data in the buffer until an acknowledge is received
else
send data immediately
|
https://en.wikipedia.org/wiki/Four-frequency | The four-frequency of a massless particle, such as a photon, is a four-vector defined by
where is the photon's frequency and is a unit vector in the direction of the photon's motion. The four-frequency of a photon is always a future-pointing and null vector. An observer moving with four-velocity will observe a frequency
Where is the Minkowski inner-product (+−−−)
Closely related to the four-frequency is the four-wavevector defined by
where , is the speed of light and and is the wavelength of the photon. The four-wavevector is more often used in practice than the four-frequency, but the two vectors are related (using ) by
See also
Four-vector
Wave vector |
https://en.wikipedia.org/wiki/Rubberhose%20%28file%20system%29 | In computing, rubberhose (also known by its development codename Marutukku) is a deniable encryption archive containing multiple file systems whose existence can only be verified using the appropriate cryptographic key.
Name and history
The project was originally named Rubberhose, as it was designed to be resistant to attacks by people willing to use torture on those who knew the encryption keys. This is a reference to the rubber-hose cryptanalysis euphemism.
It was written in 1997–2000 by Julian Assange, Suelette Dreyfus, and Ralf Weinmann.
Technical
The following paragraphs are extracts from the project's documentation:
Status
Rubberhose is not actively maintained, although it is available for Linux kernel 2.2, NetBSD and FreeBSD. The latest version available, still in alpha stage, is v0.8.3.
See also
Rubber-hose cryptanalysis
Key disclosure law
StegFS
VeraCrypt hidden volumes |
https://en.wikipedia.org/wiki/Supercow%20%28dairy%29 | Supercow (also super cow or super-cow) is a term used in the dairy industry to denote lines or individual animals that have superior milk production: that is, which produce more milk per day, or in some cases produce more fat per gallon of milk.
Until recently, super-cows have been developed through selective breeding – either traditional breeding or, since the 1960s, artificial insemination. Now the term tends to be applied to cows that have been genetically altered or whose genome has been studied in order to improve breeding. By 2023, according to a news report "[i]n many countries, including the United States, farmers breed clones with conventional animals to add desirable traits, such as high milk production or disease resistance, into the gene pool". In that year, Chinese scientists reported the cloning of three so called "super-cows" with a milk productivity "nearly 1.7 times the amount of milk an average cow in the United States produced in 2021" and a plan for 1,000 such super-cows in the near-term.
See also
Environmental impacts of animal agriculture
Genetically modified animal |
https://en.wikipedia.org/wiki/Bandhavgarh%20National%20Park | Bandhavgarh National Park is a national park of India, located in the Umaria district of Madhya Pradesh. Bandhavgarh, with an area of , was declared a national park in 1968 and then became Tiger Reserve in 1993. The current core area is spread over .
This park has a large biodiversity. The park has a large breeding population of leopards, and various species of deer. Maharaja Martand Singh of Rewa captured the first white tiger in this region in 1951. This white tiger, Mohan, is now stuffed and on display in the palace of the Maharajas of Rewa. Historically villagers and their cattle have been at a threat from the tiger. Rising mining activities around the park are putting the tigers at risk. The park derives its name from the most prominent hill of the area, which is said to have been given by Lord Rama to his brother Lakshmana to keep a watch on Lanka (Bandhav = Brother, Garh = Fort). The fort was built by a Gond Dynasty king.
Structure
The three main zones of the national park are Tala, Magdhi and Khitauli. Tala is the richest zone in terms of biodiversity, mainly tigers. Altogether, these three ranges comprise the 'core' of the Bandhavgarh Tiger Reserve, constituting a total area of 716 km2.
Fauna
With the tiger at the apex of the food chain, it contains at least 37 species of mammals. According to forest officials, there are more than 250 species of birds, about 80 species of butterflies, a number of reptiles. But many people have the species' list of about 350 birds along with photographs. The richness and tranquity of grasslands invites pairs of sarus cranes to breed in the rainy season.
One of the biggest attractions of this national park is the Bengal tiger (Panthera tigris tigris). Bandhavgarh has a very high density of tigers within its jungles. The 105 km2 of park area open to tourists was reported to have 22 tigers, a density of one tiger for every 4.77 km2. (Population estimation exercise 2001). The population of tigers in the park in 2012 was abo |
https://en.wikipedia.org/wiki/Holoplankton | Holoplankton are organisms that are planktic (they live in the water column and cannot swim against a current) for their entire life cycle. Holoplankton can be contrasted with meroplankton, which are planktic organisms that spend part of their life cycle in the benthic zone. Examples of holoplankton include some diatoms, radiolarians, some dinoflagellates, foraminifera, amphipods, krill, copepods, and salps, as well as some gastropod mollusk species. Holoplankton dwell in the pelagic zone as opposed to the benthic zone. Holoplankton include both phytoplankton and zooplankton and vary in size. The most common plankton are protists.
Reproduction
Holoplankton have unique traits that make reproduction in the water column possible. Both sexual and asexual reproduction are used depending on the type of plankton. Some invertebrate holoplankton release sperm into the water column which are then taken up by the females for fertilization. Other species release both sperm and egg to increase the likelihood of fertilization. Environmental, mechanical, or chemical cues can all trigger this release.
Diatoms are single celled phytoplankton that can occur as individuals or as long chains. They can reproduce sexually and asexually.
Diatoms are important oxygen producers and are usually the first step in the food chain.
Copepods are small holoplanktonic crustaceans that swim using their hind legs and antennae.
Defenses
Because of their small size and sluggish swimming abilities, holoplanktonic species have made certain specialized adaptations and in some cases are equipped with special defenses. Adaptations include flat bodies, lateral spines, oil droplets, floats filled with gases, sheaths made of gel-like substances, and ion replacement.
Zooplankton have adapted by developing transparent bodies, bright colors, bad tastes and cyclomorphosis (seasonal changes in body shape). When predators release a chemical in the water to signal zooplankton; cyclomorphosis allows holoplankto |
https://en.wikipedia.org/wiki/Meroplankton | Meroplankton are a wide variety of aquatic organisms which have both planktonic and benthic stages in their life cycles. Much of the meroplankton consists of larval stages of larger organism. Meroplankton can be contrasted with holoplankton, which are planktonic organisms that stay in the pelagic zone as plankton throughout their entire life cycle.
After a period of time in the plankton, many meroplankton graduate to the nekton or adopt a benthic (often sessile) lifestyle on the seafloor. The larval stages of benthic invertebrates make up a significant proportion of planktonic communities. The planktonic larval stage is particularly crucial to many benthic invertebrate in order to disperse their young. Depending on the particular species and the environmental conditions, larval or juvenile-stage meroplankton may remain in the pelagic zone for durations ranging from hour to months.
Not all meroplankton are larvae or juvenile stages of larger organisms. Many dinoflagellates are meroplanktonic, undergoing a seasonal cycle of encystment and dormancy in the benthic zone followed by excystment and reproduction in the pelagic zone before returning to the benthic zone once more. There also exist meroplanktonic diatoms; these have a seasonal resting phase below the photic zone and can be found commonly amongst the benthos of lakes and coastal zones.
Spatial distribution
Meroplankton species composition depends on spatial distribution and reproductive habits of adults in a given area. Biotic and abiotic factors such as tidal and lunar cycles and availability of food determine adult spawning schedules, in turn, determining subsequent meroplankton populations. Behavioural factors, such as predator avoidance are also important. Freshwater inputs play a key role in meroplankton species composition in estuarine environments. Effects of tides contribute greatly to meroplankton species distribution. One study conducted in a Patagonian Fjord found that species composition of the |
https://en.wikipedia.org/wiki/California%20Botanic%20Garden | The California Botanic Garden (formerly the Rancho Santa Ana Botanic Garden) is a botanical garden in Claremont, California, in the United States, just south of the San Gabriel foothills. The garden, at , is the largest botanic garden in the state dedicated to California native plants. It contains some 70,000 native Californian plants, representing 2,000 native species, hybrids and cultivars. The seed bank has embryos for the thousands of rare plants.
The garden has an active research department, specializing in systematic botany and floristics. The herbarium of Pomona College, housed at the garden, was transferred from the college in 1996. It holds over 1,200,000 specimens. The journal Aliso is published by the organization semiannually. The garden offers graduate degrees in botany through Claremont Graduate University.
The garden originated in 1927 when Susanna Bixby Bryant established a native garden on her rancho in Orange County. The garden relocated to Claremont in 1951. The facility, run by a non-profit organization, was open to the public with free admission for 58 years; in 2009 an admission fee was implemented. In 2019, the garden was renamed "California Botanic Garden" to better represent the contents of its collections.
See also
Bernard Field Station
John R. Rodman Arboretum in Claremont, California
List of California native plants
List of botanical gardens and arboretums in California
List of botanical gardens in the United States |
https://en.wikipedia.org/wiki/Ecology%20%28journal%29 | Ecology is a scientific journal that publishes research and synthesizes papers in the field of ecology. It was founded in 1920 as the continuation of Plant World, and is published by the Ecological Society of America. According to the Journal Citation Reports, it is currently ranked 15th out of 136 journals in the Ecology category. |
https://en.wikipedia.org/wiki/British%20Ecological%20Society | The British Ecological Society is a learned society in the field of ecology that was founded in 1913. It is the oldest ecological society in the world. The Society's original objective was "to promote and foster the study of Ecology in its widest sense" and this remains the central theme guiding its activities today. The Society had, circa 2013 around 4,000 members of which 14% are students. Of its members, 42% are outside the United Kingdom, in a total of 92 countries. The head office is located in London.
History
The Society evolved out of the British Vegetation Committee, which was founded in 1904 to promote the survey and study of vegetation in the British Isles. This initiative was in turn the outcome of what many historians perceive to have been the emergence of modern ecology in the 1890s. The British Ecological Society's inaugural meeting was held at University College London on 12 April 1913 and was attended by 47 members. Sir Arthur Tansley became the first President and the first issue of Journal of Ecology was printed in time for the meeting.
In its early days the society shared the London offices of The Linnean Society.
Publications
Publication of scientific journals is a principal activity. The Journal of Ecology was first published in 1913 in time for the inaugural meeting of the Society, followed by the Journal of Animal Ecology (1932), Journal of Applied Ecology (1962), Functional Ecology (1987), and Methods in Ecology and Evolution (2010). Members can subscribe to these journals at a low cost. The Society also partners with Wiley-Blackwell on the open access journal Ecology and Evolution.
Meetings
The Society also runs several major scientific meetings for ecologists each year. The Annual Meeting currently attracts 1,200 delegates each year and provides the opportunity for ecologists to present papers and posters on a wide variety of topics; an important element has always been the active participation of research students. There is an |
https://en.wikipedia.org/wiki/Ecological%20Society%20of%20America | The Ecological Society of America (ESA) is a professional organization of ecological scientists. Based in the United States and founded in 1915, ESA publications include peer-reviewed journals, newsletters, fact sheets, and teaching resources. It holds an annual meeting at different locations in the USA and Canada. In addition to its publications and annual meeting, ESA is engaged in public policy, science, education and diversity issues.
ESA's 9,000 members are researchers, educators, natural resource managers, and students in over 90 countries. Members work on a wide range of topics, from agroecology to marine diversity and explore the relationships between organisms and their past, present, and future environments. the society has 32 topical sections, six regional chapters, and ten committees.
History
The first discussions on the formation of the society took place in 1914 in the lobby of the Hotel Walton in Philadelphia, Pennsylvania, at a meeting of animal and plant ecologists organized by Henry Chandler Cowles. On December 28, 1915, in Columbus, Ohio, at the meeting of the American Association for the Advancement of Science, a group of about 50 people voted to form the Ecological Society of America, adopted a constitution, and set the next meeting. Dr. Victor E. Shelford of the University of Illinois served as the first president.
The society was founded to unify the science of ecology, stimulate research in all aspects of the discipline, encourage communication among ecologists, and promote the responsible application of ecological data and principles to the solution of environmental problems. The society has grown to 10,000 members worldwide.
In the 1940s, the society decided that it should focus on research and not pursue an activist or political focus on ecological preservation, leading Shelford to found what became The Nature Conservancy.
Divisions
The Public Affairs Office works to engage in environmental and science policy, share ecological sci |
https://en.wikipedia.org/wiki/Remote%20backup%20service | A remote, online, or managed backup service, sometimes marketed as cloud backup or backup-as-a-service, is a service that provides users with a system for the backup, storage, and recovery of computer files. Online backup providers are companies that provide this type of service to end users (or clients). Such backup services are considered a form of cloud computing.
Online backup systems are typically built for a client software program that runs on a given schedule. Some systems run once a day, usually at night while computers aren't in use. Other newer cloud backup services run continuously to capture changes to user systems nearly in real-time. The online backup system typically collects, compresses, encrypts, and transfers the data to the remote backup service provider's servers or off-site hardware.
There are many products on the market – all offering different feature sets, service levels, and types of encryption. Providers of this type of service frequently target specific market segments. High-end LAN-based backup systems may offer services such as Active Directory, client remote control, or open file backups. Consumer online backup companies frequently have beta software offerings and/or free-trial backup services with fewer live support options.
History
In the mid-1980s, the computer industry was in a great state of change with modems at speeds of 1200 to 2400 baud, making transfers of large amounts of data slow (1 MB in 72 minutes). While faster modems and more secure network protocols were in development, tape backup systems gained in popularity. During that same period the need for an affordable, reliable online backup system was becoming clear, especially for businesses with critical data.
More online/remote backup services came into existence during the heyday of the dot-com boom in the late 1990s. The initial years of these large industry service providers were about capturing market share and understanding the importance and the role that thes |
https://en.wikipedia.org/wiki/EDonkey%20network | The eDonkey Network (also known as the eDonkey2000 network or eD2k) is a decentralized, mostly server-based, peer-to-peer file sharing network created in 2000 by US developers Jed McCaleb and Sam Yagan that is best suited to share big files among users, and to provide long term availability of files. Like most sharing networks, it is decentralized, as there is no central hub for the network; also, files are not stored on a central server but are exchanged directly between users based on the peer-to-peer principle.
The server part of the network is proprietary freeware. There are two families of server software for the eD2k network: the original one from MetaMachine, written in C++, closed-source and proprietary, and no longer maintained; and eserver, written in C, also closed-source and proprietary, although available free of charge and for several operating systems and computer architectures. The eserver family is currently in active development and support, and almost all eD2k servers as of 2008 run this server software.
There are many programs that act as the client part of the network. Most notably, eDonkey2000, the original client by MetaMachine, closed-source but freeware, and no longer maintained but very popular in its day; and eMule, a free program for Windows written in Visual C++ and licensed under the GNU GPL.
The original eD2k protocol has been extended by subsequent releases of both eserver and eMule programs, generally working together to decide what new features the eD2k protocol should support. However, the eD2k protocol is not formally documented (especially in its current extended state), and it can be said that in practice the eD2k protocol is what eMule and eserver do together when running, and also how eMule clients communicate among themselves. As eMule is open source, its code is freely available for peer-review of the workings of the protocol. Examples of eD2k protocol extensions are "peer exchange among clients", "protocol obfuscation" |
https://en.wikipedia.org/wiki/Comparison%20of%20display%20technology | This is a comparison of various properties of different display technologies.
General characteristics
Major technologies are CRT, LCD and its derivatives (Quantum dot display, LED backlit LCD, WLCD, OLCD), Plasma, and OLED and its derivatives (Transparent OLED, PMOLED, AMOLED). An emerging technology is Micro LED and cancelled and now obsolete technologies are SED and FED.
Temporal characteristics
Different display technologies have vastly different temporal characteristics, leading to perceptual differences for motion, flicker, etc.
The figure shows a sketch of how different technologies present a single white/grey frame. Time and intensity is not to scale. Notice that some have a fixed intensity, while the illuminated period is variable. This is a kind of pulse-width modulation. Others can vary the actual intensity in response to the input signal.
Single-chip DLPs use a kind of "chromatic multiplexing" in which each color is presented serially. The intensity is varied by modulating the "on" time of each pixel within the time-span of one color. Multi-chip DLPs are not represented in this sketch, but would have a curve identical to the plasma display.
LCDs have a constant (backlit) image, where the intensity is varied by blocking the light shining through the panel.
CRTs use an electron beam, scanning the display, flashing a lit image. If interlacing is used, a single full-resolution image results in two "flashes". The physical properties of the phosphor are responsible for the rise and decay curves.
Plasma displays modulate the "on" time of each sub-pixel, similar to DLP.
Movie theaters use a mechanical shutter to illuminate the same frame 2 or 3 times, increasing the flicker frequency to make it less perceptible to the human eye.
Research
Researchers announced a display that uses silicon metasurface pixels that do not require polarized light and require half the energy. It employs a transparent conductive oxide as a heater that can quickly change the pixe |
https://en.wikipedia.org/wiki/Alpine%20plant | Alpine plants are plants that grow in an alpine climate, which occurs at high elevation and above the tree line. There are many different plant species and taxa that grow as a plant community in these alpine tundra. These include perennial grasses, sedges, forbs, cushion plants, mosses, and lichens. Alpine plants are adapted to the harsh conditions of the alpine environment, which include low temperatures, dryness, ultraviolet radiation, wind, drought, poor nutritional soil, and a short growing season.
Some alpine plants serve as medicinal plants.
Ecology
Alpine plants occur in a tundra: a type of natural region or biome that does not contain trees. Alpine tundra occurs in mountains worldwide. It transitions to subalpine forests below the tree line; stunted forests occurring at the forest-tundra ecotone are known as Krummholz. With increasing elevation, it ends at the snow line where snow and ice persist through summer, also known as the Nival Zone.
Alpine plants are not limited to higher elevations. However, high-elevation areas have different ecology than those growing at higher latitudes. One of the biggest distinctions is that the lower bound of a tropical alpine area is difficult to define due to a mixture of human disturbances, dry climates, and a naturally lacking tree line. The other major difference between tropical and arctic-alpine ecology is the temperature differences. The tropics have a summer/winter cycle every day, whereas the higher latitudes stay cold both day and night.
In the northern latitudes, the main factor to overcome is the cold. Intense frost action processes have a profound effect on what little soil there is and the vegetation of arctic-alpine regions. Tropical alpine regions are subject to these conditions as well, but they seldom happen. Because northern alpine areas cover a massive area it can be difficult to generalize the characteristics that define the ecology. One factor in alpine ecology is wind in an area. Wind pruning is a |
https://en.wikipedia.org/wiki/66%20block | A 66 block is a type of punch-down block used to connect sets of wires in a telephone system. They have been manufactured in four common configurations, A, B, E and M. A and B styles have the clip rows on 0.25" centers while E and M have the clip rows on 0.20" centers. The A blocks have 25 slotted holes on the left side for position the incoming building cable with a 50 slot fanning strip on the right side for distribution cables. They have been obsolete for many years. The B & M styles have 50 slot fanning strip on both sides. The B style is used mainly in distribution panels where several destinations (often 1A2 key telephones) need to connect to the same source. The M blocks are often used to connect a single instrument to such a distribution block. The E style has 5 columns of 10 2 clips rows and are used for transitioning from the 25 pair distribution cable to a 25 pair RJ21 style female ribbon connector.
66 blocks are designed to terminate 20 through 26 AWG insulated solid copper wire or 18 & 19 gauge skinned solid copper wire. The 66 series connecting block, introduced in the Bell System in 1962, was the first terminating device with insulation displacement connector technology. The term 66 block reflects its Western Electric model number.
The 25-pair standard non-split 66 block contains 50 rows; each row has two (E) or four (M) or six (A) & (B) columns of clips that are electrically bonded. The 25-pair split 50 66 block is the industry standard for easy termination of voice cabling, and is a standard network termination by telephone companies—generally on commercial properties. Each row contains four (M) or six (B) clips, but the left-side clips are electrically isolated from the right-side clips. Smaller versions also exist with fewer rows for smaller-scale use, such as residential.
66 E blocks are available pre-assembled with an RJ-21 female connector that accepts a quick connection to a 25-pair cable with a male end. These connections are typica |
https://en.wikipedia.org/wiki/110%20block | A 110 block is a type of punch-down block used to terminate runs of on-premises wiring in a structured cabling system. The designation 110 is also used to describe a type of insulation displacement contact (IDC) connector used to terminate twisted pair cables, which uses a punch-down tool similar to the type used for the older 66 block.
Usage
Telephone distribution
Early residential telephone systems used simple screw terminals to join cables to sockets in a tree topology. These screw-terminal blocks have been slowly replaced by 110 blocks and connectors. Modern homes usually have phone service entering the house to a single 110 block, whence it is distributed by on-premises wiring to outlet boxes throughout the home in star topology. At the outlet box, cables are attached to ports with insulation-displacement contacts (IDCs), and those ports fit into special faceplates.
In commercial settings, this style of home run or star topology wiring was already in use on 66 blocks in telecom closets and switchrooms. The 110 block has been slowly replacing the 66 block, especially for data communications usage.
Computer networks
The 110 style insulation-displacement connection is often used at both ends of networking cables such as Category 5, Category 5e, Category 6, and Category 6A cables which run through buildings, as shown in the image. In switch rooms, 110 blocks are often built into the backs of patch panels to terminate cable runs. At the other end, 110 connections may be used with keystone modules that are attached to wall plates. 110 blocks are preferred over 66 blocks in high-speed networks because they introduce less crosstalk and many are certified for use in Category 5, Category 6 and Category 6A wiring systems.
Individual Category 5e and better-rated 8P8C jacks (keystone and patch panel) with IDC connectors commonly use the same punchdown 'teeth' dimensions and tools as a full size 110 block.
Advantages
110 style blocks allow a much higher density of te |
https://en.wikipedia.org/wiki/Ex%20vivo | Ex vivo (Latin: "out of the living") literally means that which takes place outside an organism. In science, ex vivo refers to experimentation or measurements done in or on tissue from an organism in an external environment with minimal alteration of natural conditions.
A primary advantage of using ex vivo tissues is the ability to perform tests or measurements that would otherwise not be possible or ethical in living subjects. Tissues may be removed in many ways, including in part, as whole organs, or as larger organ systems.
Examples of ex vivo specimen use include:
bioassays;
using cancerous cell lines, like DU145 for prostate cancer, in drug testing of anticancer agents;
measurements of physical, thermal, electrical, mechanical, optical and other tissue properties, especially in various environments that may not be life-sustaining (for example, at extreme pressures or temperatures);
realistic models for surgical procedure development;
investigations into the interaction of different energy types with tissues; or
as phantoms in imaging technique development.
The term ex vivo means that the samples to be tested have been extracted from the organism. The term in vitro (lit. "within the glass") means the samples to be tested are obtained from a repository. In the case of cancer cells, a strain that would produce favorable results, then grown to produce a control sample and the number of samples required for the number of tests. These two terms are not synonymous even though the testing in both cases is "within the glass".
In cell biology, ex vivo procedures often involve living cells or tissues taken from an organism and cultured in a laboratory apparatus, usually under sterile conditions with no alterations, for up to 24 hours to obtain sufficient cells for the experiments. Experiments generally start after 24 hours of incubation. Using living cells or tissue from the same organism are still considered to be ex vivo. One widely performed ex vivo study is |
https://en.wikipedia.org/wiki/Eventual%20consistency | Eventual consistency is a consistency model used in distributed computing to achieve high availability that informally guarantees that, if no new updates are made to a given data item, eventually all accesses to that item will return the last updated value. Eventual consistency, also called optimistic replication, is widely deployed in distributed systems and has origins in early mobile computing projects. A system that has achieved eventual consistency is often said to have converged, or achieved replica convergence. Eventual consistency is a weak guarantee – most stronger models, like linearizability, are trivially eventually consistent.
Eventually-consistent services are often classified as providing BASE semantics (basically-available, soft-state, eventual consistency), in contrast to traditional ACID (atomicity, consistency, isolation, durability). In chemistry, a base is the opposite of an acid, which helps in remembering the acronym. According to the same resource, these are the rough definitions of each term in BASE:
Basically available: reading and writing operations are available as much as possible (using all nodes of a database cluster), but might not be consistent (the write might not persist after conflicts are reconciled, and the read might not get the latest write)
Soft-state: without consistency guarantees, after some amount of time, we only have some probability of knowing the state, since it might not yet have converged
Eventually consistent: If we execute some writes and then the system functions long enough, we can know the state of the data; any further reads of that data item will return the same value
Eventual consistency is sometimes criticized as increasing the complexity of distributed software applications. This is partly because eventual consistency is purely a liveness guarantee (reads eventually return the same value) and does not guarantee safety: an eventually consistent system can return any value before it converges.
Conflic |
https://en.wikipedia.org/wiki/Index%20Kewensis | The 1893 Index Kewensis (IK), maintained by the Royal Botanic Gardens, Kew, is a publication that aims to register all botanical names for seed plants at the rank of species and genera. It later came to include names of taxonomic families and ranks below that of species.
The Index is currently maintained as part of the International Plant Names Index in combination with the Gray Herbarium and Australian Plant Name indexes. This database is anticipated to complete the task of creating a complete list of plant names, although it does not determine which are accepted species names.
History
The preparation for this venture was made by Benjamin Daydon Jackson of the Linnaean Society, directed by Joseph Dalton Hooker at Kew. Charles Darwin provided the funding for the indexing project. When he died in 1882 his will stipulated that provision be made for £250 per annum over a 5-year period.
In providing citations of plant names, the starting point was taken from 1753 onward; the year of publication for the Species Plantarum of Linnaeus.
Darwin had found difficulties in applying these to the plants he studied, and Hooker's directive was to 'the compilation of an Index to the Names and Authorities of all known flowering plants and their countries'.
While the Index has never fulfilled this original charter, it was the most comprehensive for over 100 years.
Previous attempts at a comprehensive index had relied on secondary sources, this was the first attempt to provide the original publication details of the names.
A note on the country of origin was also included.
The publications of De Candolle, Pfeiffer, and Bentham provided models for the acceptance of names.
However, the editor admitted that not all earlier sources were included; this sometimes led to subsequent errors in botanical nomenclature.
The scope of the project was also changed in early editions, the editor noting that to include a full synonymy was too ambitious. The work originally indicated acceptanc |
https://en.wikipedia.org/wiki/Thermal%20hydraulics | Thermal hydraulics (also called thermohydraulics) is the study of hydraulic flow in thermal fluids. The area can be mainly divided into three parts: thermodynamics, fluid mechanics, and heat transfer, but they are often closely linked to each other. A common example is steam generation in power plants and the associated energy transfer to mechanical motion and the change of states of the water while undergoing this process. Thermal-hydraulic analysis can determine important parameters for reactor design such as plant efficiency and coolability of the system.
The common adjectives are "thermohydraulic", "thermal-hydraulic" and "thermalhydraulic".
Thermodynamic analysis
In the thermodynamic analysis, all states defined in the system are assumed to be in thermodynamic equilibrium; each state has mechanical, thermal, and phase equilibrium, and there is no macroscopic change with respect to time. For the analysis of the system, the first law and second law of thermodynamics can be applied.
In power plant analysis, a series of states can comprise a cycle. In this case, each state represents condition at the inlet/outlet of individual component. The example of components are pumpcompressor, turbine, reactor, and heat exchanger. By considering the constitutive equation for the given type of fluid, thermodynamic state of each point can be analyzed. As a result, the thermal efficiency of the cycle can be defined.
Examples of the cycle include the Carnot cycle, Brayton cycle, and Rankine cycle. Based on the simple cycle, modified or combined cycle also exists.
Thermo-Hydraulic Improvement Parameter (THIP)
Authors (Sahu et al.[6])observe that Thermo-hydraulic Parameter (THP) is less sensitive towards the Friction Factor Improvement Factor (FFER). The deviation between the terms (fR/fS) and (fR/fS)0.33 has been found 48 % to 64 % for the range of roughness and other parameters with (Re) 2900 – 14,000, which has been used for the present study.
Therefore, in order to eval |
https://en.wikipedia.org/wiki/Kad%20network | The Kad network is a peer-to-peer (P2P) network which implements the Kademlia P2P overlay protocol. The majority of users on the Kad Network are also connected to servers on the eDonkey network, and Kad Network clients typically query known nodes on the eDonkey network in order to find an initial node on the Kad network.
Usage
The Kad network uses a UDP-based protocol to:
Find sources for eD2k hashes.
Search for eD2k hashes based on keywords in the file name.
Find comments and ratings for files (hashes).
Provide buddy services for firewalled (Low ID) nodes.
Store locations, comments and (keywords out of) filenames.
Note that the Kad network is not used to actually transfer files across the P2P network. Instead, when a file transfer is initiated, clients connect directly to each other (using the standard public IP network). This traffic is susceptible to blocking/shaping/tracking by an ISP or any other opportunistic middle-man.
As with all decentralized networks, the Kad network requires no official or common servers. As such, it cannot be disabled by shutting down a given subset of key nodes. While the decentralization of the network prevents a simple shut-down, traffic analysis and deep packet inspection will more readily identify the traffic as P2P due to the high variable-destination packet throughput. The large packet volume typically causes a reduction in available CPU and/or network resources usually associated with P2P traffic.
Clients
Client search
The Kad network supports searching of files by name and a number of secondary characteristics such as size, extension, bit-rate, and more. Features vary based on client used.
Major clients
Only a few major clients currently support the Kad network implementation. However, they comprise over 80% of the user base and are probably closer to 95% of ed2k installations.
eMule: An open source Windows client which is the most popular, with 80% of network users. It also runs on Linux using the Wine librar |
https://en.wikipedia.org/wiki/Pelorus%20%28instrument%29 | In marine navigation, a pelorus is a reference tool for maintaining bearing of a vessel at sea. It is a "simplified compass" without a directive element, suitably mounted and provided with vanes to permit observation of relative bearings.
The instrument was named for one Pelorus, said to have been the pilot for Hannibal, circa 203 BC.
Ancient use
Harold Gatty described the use of a pelorus by Polynesians before the use of a compass. In equatorial waters the nightly course of stars overhead is nearly uniform during the year. This regularity simplified navigation for the Polynesians using a pelorus, or dummy compass:
Reading from North to South, in their rising and setting positions, these stars are:
{| class="wikitable"
|+ Stars
|- valign="top"
! Point
! Star
|- valign="top"
| N || Polaris
|- valign="top"
| NbE || "the Guards" (Ursa Minor)
|- valign="top"
| NNE || Alpha Ursa Major
|- valign="top"
| NEbN ||Alpha Cassiopeiae
|- valign="top"
| NE || Capella
|- valign="top"
| NEbE || Vega
|- valign="top"
| ENE || Arcturus
|- valign="top"
| EbN || the Pleiades
|- valign="top"
| E || Altair
|- valign="top"
| EbS || Orion's belt
|- valign="top"
| ESE || Sirius
|- valign="top"
| SEbE || Beta Scorpionis
|- valign="top"
| SE || Antares
|- valign="top"
| SEbS || Alpha Centauri
|- valign="top"
| SSE || Canopus
|- valign="top"
| SbE || Achernar
|- valign="top"
| S || Southern Cross
|}
The true position of these stars is only approximate to their theoretical equidistant rhumbs on the sidereal compass. Over time, the elaboration of the pelorus points led to the modern compass rose.
Modern use
In appearance and use, a pelorus resembles a compass or compass repeater, with sighting vanes or a sighting telescope attached, but it has no directive properties. That is, it remains at any relative direction to which it is set. It is generally used by setting 000° at the lubber's line. Relative bearings are then observed. They can be converted to bearings true, magnetic, gr |
https://en.wikipedia.org/wiki/Copycat%20%28software%29 | Copycat is a model of analogy making and human cognition based on the concept of the parallel terraced scan, developed in 1988 by Douglas Hofstadter, Melanie Mitchell, and others at the Center for Research on Concepts and Cognition, Indiana University Bloomington. The original Copycat was written in Common Lisp and is bitrotten (as it relies on now-outdated graphics libraries for Lucid Common Lisp); however, Java and Python ports exist. The latest version in 2018 is a Python3 port by Lucas Saldyt and J. Alan Brogan.
Description
Copycat produces answers to such problems as "abc is to abd as ijk is to what?" (abc:abd :: ijk:?). Hofstadter and Mitchell consider analogy making as the core of high-level cognition, or high-level perception, as Hofstadter calls it, basic to recognition and categorization.
High-level perception emerges from the spreading activity of many independent processes, called codelets, running in parallel, competing or cooperating. They create and destroy temporary perceptual constructs, probabilistically trying out variations to eventually produce an answer. The codelets rely on an associative network, slipnet, built on pre-programmed concepts and their associations (a long-term memory). The changing activation levels of the concepts make a conceptual overlap with neighboring concepts.
Copycat's architecture is tripartite, consisting of a slipnet, a working area (also called workspace, similar to blackboard systems), and the coderack (with the codelets). The slipnet is a network composed of nodes, which represent permanent concepts, and weighted links, which are relations, between them. It differs from traditional semantic networks, since the effective weight associated with a particular link may vary through time according to the activation level of specific concepts (nodes). The codelets build structures in the working area and modify activations in the slipnet accordingly (bottom-up processes), and the current state of slipnet determines prob |
https://en.wikipedia.org/wiki/Shirley%20Corriher | Shirley O. Corriher (born February 23, 1935) is an American biochemist and author of CookWise: The Hows and Whys of Successful Cooking, winner of a James Beard Foundation award, and BakeWise: The Hows and Whys of Successful Baking. CookWise shows how scientific insights can be applied to traditional cooking, while BakeWise applies the same idea to baking. Some compare Corriher's approach to that of Harold McGee (whom Corriher thanks as her "intellectual hero" in the "My Gratitude and Thanks" section of Cookwise) and Alton Brown. She has made a number of appearances as a food consultant on Brown's show Good Eats and has released a DVD, Shirley O. Corriher's Kitchen Secrets Revealed.
Personal life
After graduating from Vanderbilt University in 1959, she and her husband opened a boys' school in Atlanta, Georgia, where she was responsible for cooking three meals a day for 30 boys. By 1970 the school had grown to 140 students. That same year she divorced her husband and left the school; she took up cooking to support herself and her three children.
Corriher lives with her current husband, Arch, in Atlanta.
Books
See also
Harold McGee
Alton Brown |
https://en.wikipedia.org/wiki/Inverse%20semigroup | In group theory, an inverse semigroup (occasionally called an inversion semigroup) S is a semigroup in which every element x in S has a unique inverse y in S in the sense that and , i.e. a regular semigroup in which every element has a unique inverse. Inverse semigroups appear in a range of contexts; for example, they can be employed in the study of partial symmetries.
(The convention followed in this article will be that of writing a function on the right of its argument, e.g. x f rather than f(x), and
composing functions from left to right—a convention often observed in semigroup theory.)
Origins
Inverse semigroups were introduced independently by Viktor Vladimirovich Wagner in the Soviet Union in 1952, and by Gordon Preston in the United Kingdom in 1954. Both authors arrived at inverse semigroups via the study of partial bijections of a set: a partial transformation α of a set X is a function from A to B, where A and B are subsets of X. Let α and β be partial transformations of a set X; α and β can be composed (from left to right) on the largest domain upon which it "makes sense" to compose them:
where α−1 denotes the preimage under α. Partial transformations had already been studied in the context of pseudogroups. It was Wagner, however, who was the first to observe that the composition of partial transformations is a special case of the composition of binary relations. He recognised also that the domain of composition of two partial transformations may be the empty set, so he introduced an empty transformation to take account of this. With the addition of this empty transformation, the composition of partial transformations of a set becomes an everywhere-defined associative binary operation. Under this composition, the collection of all partial one-one transformations of a set X forms an inverse semigroup, called the symmetric inverse semigroup (or monoid) on X, with inverse the functional inverse defined from image to domain (equivalently, the convers |
https://en.wikipedia.org/wiki/Laser%20trimming | Laser trimming is the manufacturing process of using a laser to adjust the operating parameters of an electronic circuit.
One of the most common applications uses a laser to burn away small portions of resistors, raising their resistance value. The burning operation can be conducted while the circuit is being tested by automatic test equipment, leading to optimum final values for the resistor(s) in the circuit.
The resistance value of a film resistor is defined by its geometric dimensions (length, width, height) and the resistor material. A lateral cut in the resistor material by the laser narrows or lengthens the current flow path and increases the resistance value. The same effect is obtained whether the laser changes a thick-film or a thin-film resistor on a ceramic substrate or an SMD-resistor on a SMD circuit. The SMD-resistor is produced with the same technology and may be laser trimmed as well.
Trimmable chip capacitors are built up as multilayer plate capacitors. Vaporizing the top layer with a laser decreases the capacitance by reducing the area of the top electrode.
Passive trim is the adjustment of a resistor to a given value. If the trimming adjusts the whole circuit output such as output voltage, frequency, or switching threshold, this is called active trim. During the trim process, the corresponding parameter is measured continuously and compared to the programmed nominal value. The laser stops automatically when the value reaches the nominal value.
Trimming LTCC resistances in a pressure chamber
One type of passive trimmer uses a pressure chamber to enable resistor trimming in a single run. The LTCC boards are contacted by test probes on the assembly side and trimmed with a laser beam from the resistor side. This trimming method requires no contact points between the resistances, because the fine pitch adapter contacts the component on the opposite side of where the trimming occurs. Therefore, the LTCC can be arranged more compactly and less exp |
https://en.wikipedia.org/wiki/Daptomycin | Daptomycin, sold under the brand name Cubicin among others, is a lipopeptide antibiotic used in the treatment of systemic and life-threatening infections caused by Gram-positive organisms.
Daptomycin was removed from the World Health Organization's List of Essential Medicines in 2019. The World Health Organization classifies daptomycin as critically important for human medicine.
Medical uses
In the United States, daptomycin is indicated for use in adults for skin and skin structure infections caused by Gram-positive infections, S. aureus bacteraemia, and right-sided S. aureus endocarditis. It binds avidly to pulmonary surfactant, so cannot be used in the treatment of pneumonia. There seems to be a difference in working daptomycin on hematogenous pneumonia.
Adverse effects
Common adverse drug reactions associated with daptomycin therapy include:
Cardiovascular: low blood pressure, high blood pressure, swelling
Central nervous system: insomnia
Dermatological: rash
Gastrointestinal: diarrhea, abdominal pain
Hematological: eosinophilia
Respiratory: dyspnea
Other: injection site reactions, fever, hypersensitivity
Less common, but serious adverse events reported in the literature include
Hepatotoxicity: elevated transaminases
Nephrotoxicity: acute kidney injury from rhabdomyolysis
Also, myopathy and rhabdomyolysis have been reported in patients simultaneously taking statins, but whether this is due entirely to the statin or whether daptomycin potentiates this effect is unknown. Due to the limited data available, the manufacturer recommends that statins be temporarily discontinued while the patient is receiving daptomycin therapy. Creatine kinase levels are usually checked regularly while individuals undergo daptomycin therapy.
In July 2010, the FDA issued a warning that daptomycin could cause life-threatening eosinophilic pneumonia. The FDA said it had identified seven confirmed cases of eosinophilic pneumonia between 2004 and 2010 and an additional 36 p |
https://en.wikipedia.org/wiki/Semantics%20encoding | A semantics encoding is a translation between formal languages. For programmers, the most familiar form of encoding is the compilation of a programming language into machine code or byte-code. Conversion between document formats are also forms of encoding. Compilation of TeX or LaTeX documents to PostScript are also commonly encountered encoding processes. Some high-level preprocessors such as OCaml's Camlp4 also involve encoding of a programming language into another.
Formally, an encoding of a language A into language B is a mapping of all terms of A into B. If there is a satisfactory encoding of A into B, B is considered at least as powerful (or at least as expressive) as A.
Properties
An informal notion of translation is not sufficient to help determine expressivity of languages, as it permits trivial encodings such as mapping all elements of A to the same element of B. Therefore, it is necessary to determine the definition of a "good enough" encoding. This notion varies with the application.
Commonly, an encoding is expected to preserve a number of properties.
Preservation of compositions
soundness For every n-ary operator of A, there exists an n-ary operator of B such that
completeness For every n-ary operator of A, there exists an n-ary operator of B such that
(Note: as far as the author is aware of, this criterion of completeness is never used.)
Preservation of compositions is useful insofar as it guarantees that components can be examined either separately or together without "breaking" any interesting property. In particular, in the case of compilations, this soundness guarantees the possibility of proceeding with separate compilation of components, while completeness guarantees the possibility of de-compilation.
Preservation of reductions
This assumes the existence of a notion of reduction on both language A and language B. Typically, in the case of a programming language, reduction is the relation which models the execution of a |
https://en.wikipedia.org/wiki/The%20Hermit%20%28tarot%20card%29 | The Hermit (IX) is the ninth trump or Major Arcana card in most traditional tarot decks. It is used in game playing as well as in divination.
Description
The Rider–Waite version of the card shows an old man, standing on a mountain peak, carrying a staff in one hand and a lit lantern containing a six-pointed star in the other. In the background is a mountain range.
According to Eden Gray, his lantern is the Lamp of Truth, used to guide the unknowing, his patriarch's staff helps him navigate narrow paths as he seeks enlightenment and his cloak is a form of discretion.
Interpretation
According to A.E. Waite's 1910 book Pictorial Key to the Tarot, the Hermit card carries several divinatory associations:
9. THE HERMIT. Prudence, circumspection; also and especially treason, dissimulation, roguery, corruption. Reversed: Concealment, disguise, policy fear, unreasoned caution.
The card is usually thought to connote aspects of healing/recovery, particularly the kind that happens over time. In that regard, The Hermit is sometimes considered the mature and wiser version of The Magician. As such, both cards represent the astrological sign of Virgo. The Hermit is the "withdrawal from events and relationship to introspect and gather strength". Seeking the inner voice or calling upon vision from within. A need of understanding and advice, or a wise person who will offer knowing guidance. A card of personal experience and thoughtful temperance.
Examples
In other media
A version of Pamela Colman Smith's Hermit designed by Barrington Colby is depicted on the inner jacket sleeve of Led Zeppelin IV.
In the X/1999 tarot version made by CLAMP, The Hermit is Satsuki Yatouji.
JoJo's Bizarre Adventure: Stardust Crusaders uses tarot cards to name character's Stands. Joseph Joestar has the Stand Hermit Purple, with abilities inspired by thoughtography.
The Binding of Isaac utilizes tarot cards as pickups that can be used to gain/activate in-game abilities, the Hermit warping the pla |
https://en.wikipedia.org/wiki/Superatom | In chemistry, a superatom is any cluster of atoms that seem to exhibit some of the properties of elemental atoms.
Sodium atoms, when cooled from vapor, naturally condense into clusters, preferentially containing a magic number of atoms (2, 8, 20, 40, 58, etc.), with the outermost electron of each atom entering an orbital encompassing all the atoms in the cluster. Superatoms tend to behave chemically in a way that will allow them to have a closed shell of electrons, in this new counting scheme.
Aluminum clusters
Certain aluminum clusters have superatom properties. These aluminium clusters are generated as anions ( with n = 1, 2, 3, … ) in helium gas and reacted with a gas containing iodine. When analyzed by mass spectrometry one main reaction product turns out to be . These clusters of 13 aluminium atoms with an extra electron added do not appear to react with oxygen when it is introduced in the same gas stream, indicating a halide-like character and a magic number of 40 free electrons. Such a cluster is known as a superhalogen. The cluster component in ion is similar to an iodide ion or better still a bromide ion. The related cluster is expected to behave chemically like the triiodide ion.
Similarly it has been noted that clusters with 42 electrons (2 more than the magic numbers) appear to exhibit the properties of an alkaline earth metal which typically adopt +2 valence states. This is only known to occur when there are at least 3 iodine atoms attached to an cluster, . The anionic cluster has a total of 43 itinerant electrons, but the three iodine atoms each remove one of the itinerant electrons to leave 40 electrons in the jellium shell.
It is particularly easy and reliable to study atomic clusters of inert gas atoms by computer simulation because interaction between two atoms can be approximated very well by the Lennard-Jones potential. Other methods are readily available and it has been established that the magic numbers are 13, 19, 23, 26, 29, 32, 34, 4 |
https://en.wikipedia.org/wiki/ATM%20Adaptation%20Layer%205 | ATM Adaptation Layer 5 (AAL5) is an ATM adaptation layer used to send variable-length packets up to 65,535 octets in size across an Asynchronous Transfer Mode (ATM) network.
Unlike most network frames, which place control information in the header, AAL5 places control information in an 8-octet trailer at the end of the packet. The AAL5 trailer contains a 16-bit length field, a 32-bit cyclic redundancy check (CRC) and two 8-bit fields labeled UU and CPI that are currently unused.
Each AAL5 packet is divided into an integral number of ATM cells and reassembled into a packet before delivery to the receiving host. This process is known as Segmentation and Reassembly (see below). The last cell contains padding to ensure that the entire packet is a multiple of 48 octets long. The final cell contains up to 40 octets of data, followed by padding bytes and the 8-octet trailer. In other words, AAL5 places the trailer in the last 8 octets of the final cell where it can be found without knowing the length of the packet; the final cell is identified by a bit in the ATM header (see below), and the trailer is always in the last 8 octets of that cell.
Convergence, segmentation, and reassembly
When an application sends data over an ATM connection using AAL5, the host delivers a block of data to the AAL5 interface. AAL5 generates a trailer, divides the information into 48-octet pieces, and transfers each piece across the ATM network in a single cell. On the receiving end of the connection, AAL5 reassembles incoming cells into a packet, checks the CRC to ensure that all pieces arrived correctly, and passes the resulting block of data to the host software. The process of dividing a block of data into cells and regrouping them is known as ATM segmentation and reassembly (SAR).
By separating the functions of segmentation and reassembly from cell transport, AAL5 follows the layering principle. The ATM cell transfer layer is classified as "machine-to-machine" because the layering pri |
https://en.wikipedia.org/wiki/Superabundant%20number | In mathematics, a superabundant number is a certain kind of natural number. A natural number is called superabundant precisely when, for all :
where denotes the sum-of-divisors function (i.e., the sum of all positive divisors of , including itself). The first few superabundant numbers are . For example, the number 5 is not a superabundant number because for , and 5, the sigma is , and .
Superabundant numbers were defined by . Unknown to Alaoglu and Erdős, about 30 pages of Ramanujan's 1915 paper "Highly Composite Numbers" were suppressed. Those pages were finally published in The Ramanujan Journal 1 (1997), 119–153. In section 59 of that paper, Ramanujan defines generalized highly composite numbers, which include the superabundant numbers.
Properties
proved that if n is superabundant, then there exist a k and a1, a2, ..., ak such that
where pi is the i-th prime number, and
That is, they proved that if n is superabundant, the prime decomposition of n has non-increasing exponents (the exponent of a larger prime is never more than that a smaller prime) and that all primes up to are factors of n. Then in particular any superabundant number is an even integer, and it is a multiple of the k-th primorial
In fact, the last exponent ak is equal to 1 except when n is 4 or 36.
Superabundant numbers are closely related to highly composite numbers. Not all superabundant numbers are highly composite numbers. In fact, only 449 superabundant and highly composite numbers are the same . For instance, 7560 is highly composite but not superabundant. Conversely, 1163962800 is superabundant but not highly composite.
Alaoglu and Erdős observed that all superabundant numbers are highly abundant.
Not all superabundant numbers are Harshad numbers. The first exception is the 105th superabundant number, 149602080797769600. The digit sum is 81, but 81 does not divide evenly into this superabundant number.
Superabundant numbers are also of interest in connection with the |
https://en.wikipedia.org/wiki/Colossally%20abundant%20number | In number theory, a colossally abundant number (sometimes abbreviated as CA) is a natural number that, in a particular, rigorous sense, has many divisors. Particularly, it is defined by a ratio between the sum of an integer's divisors and that integer raised to a power higher than one. For any such exponent, whichever integer has the highest ratio is a colossally abundant number. It is a stronger restriction than that of a superabundant number, but not strictly stronger than that of an abundant number.
Formally, a number is said to be colossally abundant if there is an such that for all ,
where denotes the sum-of-divisors function.
The first 15 colossally abundant numbers, 2, 6, 12, 60, 120, 360, 2520, 5040, 55440, 720720, 1441440, 4324320, 21621600, 367567200, 6983776800 are also the first 15 superior highly composite numbers, but neither set is a subset of the other.
History
Colossally abundant numbers were first studied by Ramanujan and his findings were intended to be included in his 1915 paper on highly composite numbers. Unfortunately, the publisher of the journal to which Ramanujan submitted his work, the London Mathematical Society, was in financial difficulties at the time and Ramanujan agreed to remove aspects of the work to reduce the cost of printing. His findings were mostly conditional on the Riemann hypothesis and with this assumption he found upper and lower bounds for the size of colossally abundant numbers and proved that what would come to be known as Robin's inequality (see below) holds for all sufficiently large values of n.
The class of numbers was reconsidered in a slightly stronger form in a 1944 paper of Leonidas Alaoglu and Paul Erdős in which they tried to extend Ramanujan's results.
Properties
Colossally abundant numbers are one of several classes of integers that try to capture the notion of having many divisors. For a positive integer n, the sum-of-divisors function σ(n) gives the sum of all those numbers that divide n, i |
https://en.wikipedia.org/wiki/Highly%20abundant%20number | In number theory, a highly abundant number is a natural number with the property that the sum of its divisors (including itself) is greater than the sum of the divisors of any smaller natural number.
Highly abundant numbers and several similar classes of numbers were first introduced by , and early work on the subject was done by . Alaoglu and Erdős tabulated all highly abundant numbers up to 104, and showed that the number of highly abundant numbers less than any is at least proportional to .
Formal definition and examples
Formally, a natural number n is called highly abundant if and only if for all natural numbers m < n,
where σ denotes the sum-of-divisors function. The first few highly abundant numbers are
1, 2, 3, 4, 6, 8, 10, 12, 16, 18, 20, 24, 30, 36, 42, 48, 60, ... .
For instance, 5 is not highly abundant because σ(5) = 5+1 = 6 is smaller than σ(4) = 4 + 2 + 1 = 7, while 8 is highly abundant because σ(8) = 8 + 4 + 2 + 1 = 15 is larger than all previous values of σ.
The only odd highly abundant numbers are 1 and 3.
Relations with other sets of numbers
Although the first eight factorials are highly abundant, not all factorials are highly abundant. For example,
σ(9!) = σ(362880) = 1481040,
but there is a smaller number with larger sum of divisors,
σ(360360) = 1572480,
so 9! is not highly abundant.
Alaoglu and Erdős noted that all superabundant numbers are highly abundant, and asked whether there are infinitely many highly abundant numbers that are not superabundant. This question was answered affirmatively by .
Despite the terminology, not all highly abundant numbers are abundant numbers. In particular, none of the first seven highly abundant numbers (1, 2, 3, 4, 6, 8, and 10) is abundant. Along with 16, the ninth highly abundant number, these are the only highly abundant numbers that are not abundant.
7200 is the largest powerful number that is also highly abundant: all larger highly abundant numbers have a prime factor that divides them only o |
https://en.wikipedia.org/wiki/Superior%20highly%20composite%20number | In number theory, a superior highly composite number is a natural number which, in a particular rigorous sense, has many divisors. Particularly, it is defined by a ratio between the number of divisors an integer has and that integer raised to some positive power. For any possible exponent, whichever integer has the highest ratio is a superior highly composite number. It is a stronger restriction than that of a highly composite number, which is defined as having more divisors than any smaller positive integer.
The first 10 superior highly composite numbers and their factorization are listed.
For a superior highly composite number there exists a positive real number such that for all natural numbers smaller than we have
and for all natural numbers larger than we have
where , the divisor function, denotes the number of divisors of . The term was coined by Ramanujan (1915).
For example, the number with the most divisors per square root of the number itself is 12; this can be demonstrated using some highly composites near 12.
120 is another superior highly composite number because it has the highest ratio of divisors to itself raised to the .4 power.
The first 15 superior highly composite numbers, 2, 6, 12, 60, 120, 360, 2520, 5040, 55440, 720720, 1441440, 4324320, 21621600, 367567200, 6983776800 are also the first 15 colossally abundant numbers, which meet a similar condition based on the sum-of-divisors function rather than the number of divisors. Neither set, however, is a subset of the other.
Properties
All superior highly composite numbers are highly composite. This is easy to prove: if there is some number k that has the same number of divisors as n but is less than n itself (i.e. , but ), then for all positive ε, so if a number "n" is not highly composite, it cannot be superior highly composite.
An effective construction of the set of all superior highly composite numbers is given by the following monotonic mapping from the positive real number |
https://en.wikipedia.org/wiki/TippingPoint | TippingPoint Technologies was an American computer hardware and software company active between 1999 and 2015. Its focus was on network security products, particularly intrusion prevention systems for networks. In 2015, it was acquired by Trend Micro.
History
The company was founded in January 1999 under the name Shbang! in Texas.
Its co-founders were John F. McHale, Kent A. Savage (first chief executive), and Kenneth A. Kalinoski. Its business was to develop and sell Internet appliances.
In May 1999, the company changed its name to Netpliance and in November they released the i-Opener, a low-cost computer intended for browsing the World Wide Web. The hardware was sold at a loss, and costs were recouped through a subscription service plan. When the device was found to be easily modded to avoid the service plan, Netpliance changed the terms of sale to charge a termination fee. In 2001, the Federal Trade Commission fined the company $100,000 for inaccurate advertising and unfair billing of customers.
In 2002, the company discontinued operations of its internet appliance business and renamed itself TippingPoint. CEO Savage was replaced by chairman of the board McHale. McHale stepped down in 2004, but remained chairman of the board. The position was filled by Kip McClanahan, former CEO of BroadJump.
In January 2005, TippingPoint was acquired by the network equipment company 3Com for $442 million, operating as a division of 3Com led by James Hamilton (TippingPoint President), later replaced by Alan Kessler. 3Com itself was subsequently acquired by computer manufacturer Hewlett-Packard in April 2010 for approximately $2.7 billion.
On Oct 21, 2015, TippingPoint was acquired by Trend Micro for approximately $300 million.
Technology
The TippingPoint NGIPS is a network Intrusion Prevention System (IPS) deals with IT threat protection. It combines application-level security with user awareness and inbound/outbound messaging inspection capabilities, to protect the user' |
https://en.wikipedia.org/wiki/Object-based%20language | The term object-based language may be used in a technical sense to describe any programming language that uses the idea of encapsulating state and operations inside objects. Object-based languages need not support inheritance or subtyping, but those that do are also termed object-oriented. Object-based languages that do not support inheritance or subtyping are usually not considered to be true object-oriented languages.
Examples of object-oriented languages, in rough chronological order, include Simula, Smalltalk, C++ (which object model is based on Simula's), Objective-C (which object model is based on Smalltalk's), Eiffel, Xojo (formerly REALbasic), Python, Ruby, Java, Visual Basic .NET, C#, and Fortran 2003. Examples of a language that is object-based, but not object-oriented are early versions of Ada, Visual Basic (VB), and Fortran 90. These languages all support the definition of an object as a data structure, but lack polymorphism and inheritance.
In practice, the term object-based is usually applied to those object-based languages that are not also object-oriented, although all object-oriented languages are also object-based, by definition. Instead, the terms object-based and object-oriented are normally used as mutually exclusive alternatives, rather than as categories that overlap.
Sometimes, the term object-based is applied to prototype-based programming languages, true object-oriented languages that lack classes, but in which objects instead inherit their code and data directly from other template objects. An example of a commonly used prototype-based scripting language is JavaScript.
Both object-based and object-oriented languages (whether class-based or prototype-based) may be statically type-checked. Statically checking prototype-based languages can be difficult, because these languages often allow objects to be dynamically extended with new behavior, and even to have their parent object (from which they inherit) changed, at runtime. |
https://en.wikipedia.org/wiki/List%20of%20natural%20phenomena | A natural phenomenon is an observable event which is not man-made. Examples include: sunrise, weather, fog, thunder, tornadoes; biological processes, decomposition, germination; physical processes, wave propagation, erosion; tidal flow, and natural disasters such as electromagnetic pulses, volcanic eruptions, hurricanes and earthquakes.
History
Over many intervals of time, natural phenomena have been observed by a series of countless events as a feature created by nature.
Physical phenomena
The act of:
Freezing
Boiling
Gravity
Magnetism
Gallery
Chemical phenomena
Oxidation
Fire
Rusting
Biological phenomena
Metabolism
Catabolism
Anabolism
Decomposition – by which organic substances are broken down into a much simpler form of matter
Fermentation – converts sugar to acids, gases and/or alcohol.
Growth
Birth
Death
Population decrease
Gallery
Astronomical phenomena
Supernova
Gamma ray bursts
Quasars
Blazars
Pulsars
Cosmic microwave background radiation.
Geological phenomena
Mineralogic phenomena
Lithologic phenomena
Rock types
Igneous rock
Igneous formation processes
Sedimentary rock
Sedimentary formation processes (sedimentation)
Quicksand
Metamorphic rock
Endogenic phenomena
Plate tectonics
Continental drift
Earthquake
Oceanic trench
Phenomena associated with igneous activity
Geysers and hot springs
Bradyseism
Volcanic eruption
Earth's magnetic field
Exogenic phenomena
Slope phenomena
Slump
Landslide
Weathering phenomena
Erosion
Glacial and peri-glacial phenomena
Glaciation
Moraines
Hanging valleys
Atmospheric phenomena
Impact phenomena
Impact crater
Coupled endogenic-exogenic phenomena
Orogeny
Drainage development
Stream capture
Gallery
Meteorological phenomena
Violent meteorological phenomena are called storms. Regular, cyclical phenomena include seasons and atmospheric circulation. climate change is often semi-regular.
Atmospheric optical phenomena
Oceanographic
Oceanographic phenomena inc |
https://en.wikipedia.org/wiki/The%20Fat%20Duck | The Fat Duck is a fine dining restaurant in Bray, Berkshire, England, owned by the chef Heston Blumenthal. Housed in a 16th-century building, the Fat Duck opened on 16 August 1995. Although it originally served food similar to a French bistro, it soon acquired a reputation for precision and invention, and has been at the forefront of many modern culinary developments, such as food pairing, flavour encapsulation and multi-sensory cooking.
The number of staff in the kitchen increased from four when the Fat Duck first opened to 42, resulting in a ratio of one kitchen staff member per customer. The Fat Duck gained its first Michelin star in 1999, its second in 2002 and its third in 2004, making it one of eight restaurants in the United Kingdom to earn three Michelin stars. It lost its three stars in 2016, as it was under renovation, preventing it from being open for assessment. It regained all three stars in the following year.
The Fat Duck is known for its fourteen-course tasting menu featuring dishes such as nitro-scrambled egg and bacon ice cream, an Alice in Wonderland-inspired mock turtle soup involving a bouillon packet made up to look like a fob watch dissolved in tea, and a dish called Sound of the Sea which includes an audio element. It has an associated laboratory where Blumenthal and his team develop new dish concepts. In 2009, the Fat Duck suffered from the largest norovirus outbreak ever documented at a restaurant, with more than 400 diners falling ill.
Description
The Fat Duck is located on the high street of Bray, Berkshire. The owner, Heston Blumenthal, has owned the premises since it opened at the location in 1995. It is not the only Michelin three-star restaurant in Bray, the other being Michel Roux's restaurant the Waterside Inn. As of 2023, it is one of eight restaurants in the United Kingdom with three Michelin stars.
The Fat Duck has fourteen tables, and can seat 42 diners. It has a high proportion of chefs working, 42, equating to one chef p |
https://en.wikipedia.org/wiki/Minimum%20mean%20square%20error | In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. In the Bayesian setting, the term MMSE more specifically refers to estimation with quadratic loss function. In such case, the MMSE estimator is given by the posterior mean of the parameter to be estimated. Since the posterior mean is cumbersome to calculate, the form of the MMSE estimator is usually constrained to be within a certain class of functions. Linear MMSE estimators are a popular choice since they are easy to use, easy to calculate, and very versatile. It has given rise to many popular estimators such as the Wiener–Kolmogorov filter and Kalman filter.
Motivation
The term MMSE more specifically refers to estimation in a Bayesian setting with quadratic cost function. The basic idea behind the Bayesian approach to estimation stems from practical situations where we often have some prior information about the parameter to be estimated. For instance, we may have prior information about the range that the parameter can assume; or we may have an old estimate of the parameter that we want to modify when a new observation is made available; or the statistics of an actual random signal such as speech. This is in contrast to the non-Bayesian approach like minimum-variance unbiased estimator (MVUE) where absolutely nothing is assumed to be known about the parameter in advance and which does not account for such situations. In the Bayesian approach, such prior information is captured by the prior probability density function of the parameters; and based directly on Bayes theorem, it allows us to make better posterior estimates as more observations become available. Thus unlike non-Bayesian approach where parameters of interest are assumed to be deterministic, but unknown constants, the Bayesian estimator seeks to estimate a par |
https://en.wikipedia.org/wiki/Complex%20convexity | Complex convexity is a general term in complex geometry.
Definition
A set in is called if its intersection with any complex line is contractible.
Background
In complex geometry and analysis, the notion of convexity and its generalizations play an important role in understanding function behavior. Examples of classes of functions with a rich structure are, in addition to the convex functions, the subharmonic functions and the plurisubharmonic functions.
Geometrically, these classes of functions correspond to convex domains and pseudoconvex domains, but there are also other types of domains, for instance lineally convex domains which can be generalized using convex analysis.
A great deal is already known about these domains, but there remain some fascinating, unsolved problems. This theme is mainly theoretical, but there are computational aspects of the domains studied, and these computational aspects are certainly worthy of further study. |
https://en.wikipedia.org/wiki/Parabolic%20cylinder%20function | In mathematics, the parabolic cylinder functions are special functions defined as solutions to the differential equation
This equation is found when the technique of separation of variables is used on Laplace's equation when expressed in parabolic cylindrical coordinates.
The above equation may be brought into two distinct forms (A) and (B) by completing the square and rescaling , called H. F. Weber's equations:
and
If is a solution, then so are
If is a solution of equation (), then is a solution of (), and, by symmetry,
are also solutions of ().
Solutions
There are independent even and odd solutions of the form (). These are given by (following the notation of Abramowitz and Stegun (1965)):
and
where is the confluent hypergeometric function.
Other pairs of independent solutions may be formed from linear combinations of the above solutions. One such pair is based upon their behavior at infinity:
where
The function approaches zero for large values of and , while diverges for large values of positive real .
and
For half-integer values of a, these (that is, U and V) can be re-expressed in terms of Hermite polynomials; alternatively, they can also be expressed in terms of Bessel functions.
The functions U and V can also be related to the functions (a notation dating back to Whittaker (1902)) that are themselves sometimes called parabolic cylinder functions:
Function was introduced by Whittaker and Watson as a solution of eq.~() with bounded at . It can be expressed in terms of confluent hypergeometric functions as
Power series for this function have been obtained by Abadir (1993). |
https://en.wikipedia.org/wiki/Potential%20energy%20surface | A potential energy surface (PES) describes the energy of a system, especially a collection of atoms, in terms of certain parameters, normally the positions of the atoms. The surface might define the energy as a function of one or more coordinates; if there is only one coordinate, the surface is called a potential energy curve or energy profile. An example is the Morse/Long-range potential.
It is helpful to use the analogy of a landscape: for a system with two degrees of freedom (e.g. two bond lengths), the value of the energy (analogy: the height of the land) is a function of two bond lengths (analogy: the coordinates of the position on the ground).
The PES concept finds application in fields such as chemistry and physics, especially in the theoretical sub-branches of these subjects. It can be used to theoretically explore properties of structures composed of atoms, for example, finding the minimum energy shape of a molecule or computing the rates of a chemical reaction.
Mathematical definition and computation
The geometry of a set of atoms can be described by a vector, , whose elements represent the atom positions. The vector could be the set of the Cartesian coordinates of the atoms, or could also be a set of inter-atomic distances and angles.
Given , the energy as a function of the positions, , is the value of for all of interest. Using the landscape analogy from the introduction, E gives the height on the "energy landscape" so that the concept of a potential energy surface arises.
To study a chemical reaction using the PES as a function of atomic positions, it is necessary to calculate the energy for every atomic arrangement of interest. Methods of calculating the energy of a particular atomic arrangement of atoms are well described in the computational chemistry article, and the emphasis here will be on finding approximations of to yield fine-grained energy-position information.
For very simple chemical systems or when simplifying approximations are |
https://en.wikipedia.org/wiki/Atrazine | Atrazine is a chlorinated herbicide of the triazine class. It is used to prevent pre-emergence broadleaf weeds in crops such as maize (corn), soybean and sugarcane and on turf, such as golf courses and residential lawns. Atrazine's primary manufacturer is Syngenta and it is one of the most widely used herbicides in the United States, Canadian, and Australian agriculture. Its use was banned in the European Union in 2004, when the EU found groundwater levels exceeding the limits set by regulators, and Syngenta could not show that this could be prevented nor that these levels were safe.
At least two significant Canadian farm well studies showed that atrazine was the most common contaminant found. , atrazine was the most commonly detected pesticide contaminating drinking water in the U.S. Studies suggest it is an endocrine disruptor, an agent that can alter the natural hormonal system. However, in 2006 the U.S. Environmental Protection Agency (EPA) had stated that under the Food Quality Protection Act "the risks associated with the pesticide residues pose a reasonable certainty of no harm", and in 2007, the EPA said that atrazine does not adversely affect amphibian sexual development and that no additional testing was warranted. EPA's 2009 review concluded that "the agency's scientific bases for its regulation of atrazine are robust and ensure prevention of exposure levels that could lead to reproductive effects in humans". However, in their 2016 Refined Ecological Risk Assessment for Atrazine, it was stated that "it is difficult to make definitive conclusions about the impact of atrazine at a given concentration but multiple studies have reported effects to various endpoints at environmentally-relevant concentrations." EPA started a registration review in 2013.
The EPA's review has been criticized, and the safety of atrazine remains controversial. EPA has however stated that "If at any time EPA determines there are urgent human or environmental risks from atrazine ex |
https://en.wikipedia.org/wiki/Biostimulation | Biostimulation involves the modification of the environment to stimulate existing bacteria capable of bioremediation. This can be done by addition of various forms of rate limiting nutrients and electron acceptors, such as phosphorus, nitrogen, oxygen, or carbon (e.g. in the form of molasses). Alternatively, remediation of halogenated contaminants in anaerobic environments may be stimulated by adding electron donors (organic substrates), thus allowing indigenous microorganisms to use the halogenated contaminants as electron acceptors. EPA Anaerobic Bioremediation Technologies Additives are usually added to the subsurface through injection wells, although injection well technology for biostimulation purposes is still emerging. Removal of the contaminated material is also an option, albeit an expensive one. Biostimulation can be enhanced by bioaugmentation. This process, overall, is referred to as bioremediation and is an EPA-approved method for reversing the presence of oil or gas spills. While biostimulation is usually associated with remediation of hydrocarbon or high production volume chemical spills, it is also potentially useful for treatment of less frequently encountered contaminant spills, such as pesticides, particularly herbicides.
The primary advantage of biostimulation is that bioremediation will be undertaken by already present native microorganisms that are well-suited to the subsurface environment, and are well distributed spatially within the subsurface. The primary disadvantage is that the delivery of additives in a manner that allows the additives to be readily available to subsurface microorganisms is based on the local geology of the subsurface. Tight, impermeable subsurface lithology (tight clays or other fine-grained material) make it difficult to spread additives throughout the affected area. Fractures in the subsurface create preferential pathways in the subsurface which additives preferentially follow, preventing even distribution o |
https://en.wikipedia.org/wiki/Transcend%20Information | Transcend Information, Inc. () is a Taiwanese company headquartered in Taipei, Taiwan that manufactures and distributes memory products. Transcend deals in over 2,000 products including memory modules, flash memory cards, USB flash drives, portable hard drives, multimedia products, solid-state drives, dashcams, body cameras, personal cloud storage, card readers and accessories.
It has offices in the United States, Germany, the Netherlands, United Kingdom, Japan, Hong Kong, China and South Korea. It was the first Taiwanese memory module manufacturer to receive ISO 9001 certification.
History
Transcend Information. Inc. was founded in 1988 by Mr. Chung-Won Shu, Peter, with its headquarters in Taipei, Taiwan. Today, Transcend has become a global brand of digital storage, multimedia and industrial products with 13 offices worldwide. The design, development, and manufacture all their products is done in-house, as well as the marketing and sales.
Product lines
Transcend manufactures memory modules, flash memory cards, USB flash drives, card readers, external hard drives, solid state drives and industrial-grade products.
Memory cards – SD / CF / micro SD
USB flash drives – USB 3.0 / USB 2.0
External hard drives – Rugged series / Classic series / 3.5" Desktop storage
Solid-state drives – 2.5" SSDs / M.2 SSDs / mSATA SSDs / Portable SSDs
Card readers – USB 3.0 / USB 2.0 / USB Type - C
Dashcams
Body cameras
Personal cloud storage
Apple solutions - Lightning / USB 3.1 flash drives / SSD upgrade kit / Expansion cards for MacBook Air and MacBook Pro / Portable hard drive and portable SSD for Mac
Multimedia products - Digital media players / DVD writers
Memory modules - For Desktop / Notebook / Server
Embedded solutions
Flash: SSD / CF / SD / MMC / eMMC / SATA Module / PATA Module / USB Module
DRAM: Industrial temperature / Workstation / Server / Standard / Low voltage / Low profile
JetFlash
JetFlash are a series of flash based USB drives designed and manufac |
https://en.wikipedia.org/wiki/Megaminx | The Megaminx or Mégaminx (, ) is a dodecahedron-shaped puzzle similar to the Rubik's Cube. It has a total of 50 movable pieces to rearrange, compared to the 20 movable pieces of the Rubik's Cube.
History
The Megaminx, or Magic Dodecahedron, was invented by several people independently and produced by several different manufacturers with slightly different designs. Uwe Mèffert eventually bought the rights to some of the patents and continues to sell it in his puzzle shop under the Megaminx moniker. It is also known by the name Hungarian Supernova, invented by Dr. Christoph Bandelow. His version came out first, shortly followed by Meffert's Megaminx. The proportions of the two puzzles are slightly different.
Speed-solving the Megaminx became an official World Cube Association event in 2003, with the first official single-solve record set by American Grant Tregay with a time of 2 minutes 12.82 seconds during the World Rubik's Games Championship in Canada. The first sub-minute solve in official competition was achieved by Japanese solver Takumi Yoshida with a time of 59.33s at the January 2009 Amagasaki Open, and the first sub-30-second single solve was achieved by Peruvian solver Juan Pablo Huanqui at a 2017 Santiago event with a time of 29.93 seconds. The current world record time for a Megaminx solve is 24.44 seconds, set by Leandro Martín López of Argentina on 23 July 2023
Description
The Megaminx is made in the shape of a dodecahedron, and has 12 faces and center pieces, 20 corner pieces, and 30 edge pieces. The face centers each have a single color, which identifies the color of that face in the solved state. The edge pieces have two colors, and the corner pieces have three. Each face contains a center piece, 5 corner pieces and 5 edge pieces. The corner and edge pieces are shared with adjacent faces. The face centers can only rotate in place, but the other pieces can be permuted by twisting the face layer around the face center.
There are two main versions |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.