source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Chief%20Choc
Chief Choc is the former mascot of Mississippi College, a private Christian university located in Clinton, Mississippi. In August 2005, the NCAA announced that schools with "hostile and abusive racial/ethnic/national origin mascots, nicknames or imagery" would be banned from championship events. Overview The original announcement named 18 colleges and universities with "hostile and abusive" mascots, including the Mississippi College Choctaws. According to Mississippi College President Lee Royce, however, the university shares a "mutual relationship of respect and cooperation" with the Mississippi Band of Choctaw Indians and the tribe approved of the school's use of Choctaw names and images. NCAA spokesman Bob Williams said the appeals were granted because individual, sovereign tribes announced their approval of the use of names and images. Most of the other schools use generic terms like "Indians" or "Braves" as mascots, so Williams said "the tribal issue isn't in play at all in their particular cases." Four schools - Chowan College, Midwestern State University, Southeastern Oklahoma State University and the University of Louisiana-Monroe - decided not to fight this ruling, making immediate changes to their mascots and logos. Alcorn State University (the "Braves") and Arkansas State University (the Indians) have not filed appeals or started changing their mascots and logos. In a letter dated February 17, 2006, Mississippi College received word that the NCAA has removed its policy restrictions in the use of the name Choctaw for MC athletics. In a statement, President Lee Royce stated: “We are pleased with the ruling from the NCAA giving their approval of our request to remove Mississippi College from the list of institutions subject to the policy’s restrictions. We are very appreciative of the Mississippi Band of Choctaw Indians’ support of our use of the Choctaw name, and look forward to continuing our mutual relationship of respect and cooperation.” "Alt
https://en.wikipedia.org/wiki/Bernhard%20Nocht%20Institute%20for%20Tropical%20Medicine
Bernhard Nocht Institute for Tropical Medicine (Bernhard-Nocht-Institut für Tropenmedizin) (BNITM) in Hamburg is Germany's largest institution for tropical medicine, with a workforce of about 250 people in Hamburg. It is member of the Leibniz-Association. History The cholera epidemic of the year 1892 claimed thousands of lives and prompted the Senate and Parliament of the City of Hamburg to reform the health care system. The Tropical Medicine Institute was founded with the support of the Imperial Government to research ship and tropical diseases and to train ship and colonial physicians. In 1893, the naval physician was introduced to the newly created position of port physician. For the medical care of seamen suffering from internal diseases, he was also given a department in the St. Georg General Hospital. Contrary to the plans of the bacteriologist Robert Koch, Nocht established Hamburg in 1899 as the location for an institute for the research of tropical diseases, since "due to overseas traffic there are many people with treatment needs at this point". On 1 October 1900 the "Institute for Maritime and Tropical Diseases" with 24 employees was opened in the former administration building of the naval hospital at Hamburg's Landungsbrücken. Since 2006, the inpatient care has taken place at the University Medical Center Hamburg-Eppendorf. Building Between 1910 and 1914, the three-part clinker building with laboratory wing, hospital and animal house was built according to plans by Fritz Schumacher. The building wing is located in the St. Pauli district between Bernhard Nocht Street on the high north side and the slope of leading down to the harbor shore. After 1945 the building, damaged by bombs, was re-built. From 2003 a new wing was built on the site of the former animal house, which was put into operation at the end of January 2008. In particular, the high-security laboratories were completely redesigned and have since then been among the safest in the world (b
https://en.wikipedia.org/wiki/Amanita%20farinosa
Amanita farinosa, commonly known as the Eastern American floury amanita or the American floury amanita, is a North American poisonous mushroom of the genus Amanita, a genus of fungi including some of the most deadly mushrooms. Taxonomy Two recent molecular studies show that Amanita farinosa is part of a subgroup within Amanita with its close relatives Amanita muscaria, A. gemmata and A. roseotincta. Description The cap is 2.5–7 cm (1–2.8 inches) in diameter, domed in young and flat in older specimens, with a striate margin. It is whitish grey and covered with brownish grey volval or mealy material. The gills are white. They may be attached to the stem or free. They are close and crowded and not bruising. The stem, or the stipe, lacks a ring and at its base a smallish bulb or volva. It measures up to 6.5 cm high, 1–3 centimeters thick. The stem is white to tan in color. The spores are white. They are 5.5–8 x 6–8 µm in measurement and inamyloid. The spores are smooth and round to broadly elliptical. The flesh is white in color. It doesn't stain on exposure. In old specimens, the smell can be strong and that of mink. Distribution and habitat An uncommon mushroom, it is found across North America in late summer to late autumn in coniferous or deciduous woodlands. See also List of Amanita species Footnotes farinosa Poisonous fungi Fungi of North America Fungi of Central America Fungi of Mexico Fungi described in 1822 Fungi without expected TNC conservation status
https://en.wikipedia.org/wiki/Hemoglobin%20variants
Hemoglobin variants are different types of hemoglobin molecules, by different combinations of its subunits and/or mutations thereof. Hemoglobin variants are a part of the normal embryonic and fetal development. They may also be pathologic mutant forms of hemoglobin in a population, caused by variations in genetics. Some well-known hemoglobin variants, such as sickle-cell anemia, are responsible for diseases and are considered hemoglobinopathies. Other variants cause no detectable pathology, and are thus considered non-pathological variants. Some normal hemoglobin types are; Hemoglobin A (Hb A), which is 95–98% of hemoglobin found in adults, Hemoglobin A2 (Hb A2), which is 2–3% of hemoglobin found in adults, and Hemoglobin F (Hb F), which is found in adults up to 2.5% and is the primary hemoglobin that is produced by the fetus during pregnancy. Hemoglobin variants occur when there are genetic changes in specific genes, or globins, that cause changes or alterations in the amino acid. They could affect the structure, behavior, the production rate, and/or the stability of that specific gene. Usually there are four genes that code for alpha globin and two genes that code for beta globin. If the genes for alpha chains is mutated, the most common condition that occurs is alpha thalassemia, which causes a decrease in production of that gene. The level of severity of alpha thalassemia is determined by the number of genes that are affected. Hemoglobin variants are most often inherited characteristics. First, abnormal beta gene can be inherited in an autosomal recessive fashion. This means that the person who inherits this will have two copies of the altered gene. Both of these genes can be passed to offspring. The next way they can be inherited is in a heterozygous fashion. This means that the person has one normal beta gene and one abnormal beta gene. This person is considered to be a carrier of whichever hemoglobin variant is inherited. Only the abnormal gene can be pas
https://en.wikipedia.org/wiki/Entropy%20and%20life
Research concerning the relationship between the thermodynamic quantity entropy and both the origin and evolution of life began around the turn of the 20th century. In 1910, American historian Henry Adams printed and distributed to university libraries and history professors the small volume A Letter to American Teachers of History proposing a theory of history based on the second law of thermodynamics and on the principle of entropy. The 1944 book What is Life? by Nobel-laureate physicist Erwin Schrödinger stimulated further research in the field. In his book, Schrödinger originally stated that life feeds on negative entropy, or negentropy as it is sometimes called, but in a later edition corrected himself in response to complaints and stated that the true source is free energy. More recent work has restricted the discussion to Gibbs free energy because biological processes on Earth normally occur at a constant temperature and pressure, such as in the atmosphere or at the bottom of the ocean, but not across both over short periods of time for individual organisms. Ideas about the relationship between entropy and living organisms have inspired hypotheses and speculations in many contexts, including psychology, information theory, the origin of life, and the possibility of extraterrestrial life. Early views In 1863, Rudolf Clausius published his noted memoir On the Concentration of Rays of Heat and Light, and on the Limits of Its Action, wherein he outlined a preliminary relationship, based on his own work and that of William Thomson (Lord Kelvin), between living processes and his newly developed concept of entropy. Building on this, one of the first to speculate on a possible thermodynamic perspective of organic evolution was the Austrian physicist Ludwig Boltzmann. In 1875, building on the works of Clausius and Kelvin, Boltzmann reasoned: In 1876, American civil engineer Richard Sears McCulloh, in his Treatise on the Mechanical Theory of Heat and its Applicatio
https://en.wikipedia.org/wiki/Hostway
Hostway is a global web hosting and technology infrastructure company headquartered in Austin, Texas, United States. It provides hosting services to individuals, small to medium-sized businesses, and large corporations with web sites, databases, business applications, and managed web hosting. Hostway Services Inc. merged with Hosting.com in January 2019. The merged company rebranded to Ntirety in September 2019. History Founding In 1998, University of Chicago alumni Lucas Roh, John Lee, Arnold Choi and two other business partners acquired Spectronet Inc., a Web hosting company in the Chicago suburbs with annual revenues of $30,000. They renamed the company Hostway Services Inc. and began preparations to expand its reach in the US and international markets. Roh previously served as a computer scientist at Argonne National Laboratory, conducting pioneering research in the emerging field of "automatic differentiation". He also worked for six years as a software and hardware engineer at Tektronics and Hewlett-Packard prior to becoming a founding partner, president and CEO of Hostway. Globalization Initially, company growth was focused on sales throughout the United States, primarily in Chicago. In 2000, the founders expanded Hostway to international markets with the launch of Hostway Korea. South Korea was chosen as the starting point for Hostway's expansion, due to Roh being raised there. This would serve as the beginning of a much larger push towards its globalization of web hosting services. From 2000-2007, Hostway grew via acquisition, increasing its client base and building its experience with the dedicated server/managed hosting & domain name registration markets. The most significant of these was the 2003 merger with Canada's NetNation Communications with approximately US$10 million paid in cash to NetNation stockholders. In June 2006, Hostway announced the opening of its new data center located at Boeing International's worldwide headquarters, in the he
https://en.wikipedia.org/wiki/Cluster%20root
Cluster roots, also known as proteoid roots, are plant roots that form clusters of closely spaced short lateral rootlets. They may form a two- to five-centimetre-thick mat just beneath the leaf litter. They enhance nutrient uptake, possibly by chemically modifying the soil environment to improve nutrient solubilisation. As a result, plants with proteoid roots can grow in soil that is very low in nutrients, such as the phosphorus-deficient native soils of Australia. They were first described by Adolf Engler in 1894, after he discovered them on plants of the family Proteaceae growing in Leipzig Botanic Gardens. In 1960, Helen Purnell examined 44 species from ten Proteaceae genera, finding proteoid roots in every genus except Persoonia; she then coined the name "proteoid roots" in reference to the plant family in which it was known to occur. Proteoid roots are now known to occur in 27 different Proteaceae genera, plus around 30 species from other families, including Betulaceae, Casuarinaceae, Eleagnaceae, Leguminosae, Moraceae and Myricaceae. Similar structures also occur in species of Cyperaceae and Restionaceae, but their physiology is yet to be studied. Two forms are recognised: simple cluster roots form rootlets only along a root; compound cluster roots form the primary rootlets, and also form secondary rootlets on the primary rootlets. Some Proteaceae, such as Banksia and Grevillea, are valued by the horticulture and floriculture industries. In cultivation, only slow-release low-phosphorus fertilizers should be used, as higher levels cause phosphorus toxicity and sometimes iron deficiency, leading to plant death. Crop management should minimise root disturbance, and weed control should be via slashing or contact herbicides. Many plants with proteoid roots have economic value. Cultivated crops with proteoid roots include Lupinus and Macadamia.
https://en.wikipedia.org/wiki/Automated%20optical%20inspection
Automated optical inspection (AOI) is an automated visual inspection of printed circuit board (PCB) (or LCD, transistor) manufacture where a camera autonomously scans the device under test for both catastrophic failure (e.g. missing component) and quality defects (e.g. fillet size or shape or component skew). It is commonly used in the manufacturing process because it is a non-contact test method. It is implemented at many stages through the manufacturing process including bare board inspection, solder paste inspection (SPI), pre-reflow and post-re-flow as well as other stages. Historically, the primary place for AOI systems has been after solder re-flow or "post-production." Mainly because, post-re-flow AOI systems can inspect for most types of defects (component placement, solder shorts, missing solder, etc.) at one place in the line with one single system. In this way the faulty boards are reworked and the other boards are sent to the next process stage. SMT inspection AOIs for a PCB board with components may inspect the following features: Area defects Billboarding Component offset Component polarity Component presence or absence Component Skew Excessive Solder Joints Flipped component Height Defects Insufficient Paste around Leads Insufficient Solder Joints Lifted Leads No Population tests Paste Registration Severely Damaged Components Tombstoning Volume Defects Wrong Part Solder Bridging Presence of Foreign Material on the board AOI can be used in the following locations in the SMT lines: post paste, pre-reflow, post-reflow, or wave areas. Bare PCB inspection AOI for a bare PCB board inspection may detect these features: Line width violations Spacing violation Excess copper Missing pad – a feature that should be on the board is missing Short circuits Gold Finger damage Cuts Hole breakage – a drilled hole (via) is outside of its landing pad Wrong mounting components identified The triggering of a defects report may be ei
https://en.wikipedia.org/wiki/Package-merge%20algorithm
The package-merge algorithm is an O(nL)-time algorithm for finding an optimal length-limited Huffman code for a given distribution on a given alphabet of size n, where no code word is longer than L. It is a greedy algorithm, and a generalization of Huffman's original algorithm. Package-merge works by reducing the code construction problem to the binary coin collector's problem. The coin collector's problem Suppose a coin collector has a number of coins of various denominations, each of which has a numismatic value unrelated to its denomination. The coin collector has run out of money and needs to use some of his coin collection to buy something of cost N. He wishes to select a subset of coins from his collection of minimum numismatic value whose denominations total N. The binary version of this problem is that all denominations are powers of 2, that is, 1, 1/2, 1/4, etc. dollars. Description of the package-merge algorithm Assume that the largest denomination is 1 dollar, and that N is an integer. (The algorithm works even if these assumptions do not hold, by trivial modifications.) The coin collector first separates his coins into lists, one for each denomination, sorted by numismatic value. He then packages the smallest denomination coins in pairs, starting from the pair of smallest total numismatic value. If there is one coin left over, it will be the coin of highest numismatic value of that denomination, and it is set aside and ignored henceforth. These packages are then merged into the list of coins of the next smallest denomination, again in order of numismatic value. The items in that list are then packaged in pairs, and merged into the next smallest list, and so forth. Finally, there is a list of items, each of which is a 1 dollar coin or a package consisting of two or more smaller coins whose denominations total 1 dollar. They are also sorted in order of numismatic value. The coin collector then selects the least value N of them. Note that
https://en.wikipedia.org/wiki/Interleukin%2029
Interleukin-29 (IL-29) is a cytokine and it belongs to type III interferons group, also termed interferons λ (IFN-λ). IL-29 (alternative name IFNλ1) plays an important role in the immune response against pathogenes and especially against viruses by mechanisms similar to type I interferons, but targeting primarily cells of epithelial origin and hepatocytes. IL-29 is encoded by the IFNL1 gene located on chromosome 19 in humans. It is a pseudogene in mice meaning the IL-29 protein is not produced in them. Structure IL-29 is, with the rest of IFN-λ, structurally related to the IL-10 family, but its primary amino acid sequence (and also function) is more similar to type I interferons. It binds to a heterodimeric receptor composed of one subunit IFNL1R specific for IFN-λ and a second subunit IL10RB shared among the IL-10 family cytokines. Function Effects on immune response to pathogens   IL-29 exhibits antiviral effects by inducing similar signaling pathways as type I interferons. IL-29 receptor signals through JAK-STAT pathways leading to activated expression of interferon-stimulated genes and production of antiviral proteins. Further consequences of IL-29 signalization comprise the upregulated expression of MHC class I molecules, or enhanced expression of the costimulatory molecules and chemokine receptors on pDC, which are the main producers of IFN-α. IL-29 expression is dominant in virus-infected epithelial cells of the respiratory, gastrointestinal and urogenital tracts, also in other mucosal tissues and skin. Hepatocytes infected by HCV or HBV viruses stimulate the immune response by producing IL-29 (IFN-λ in general) rather than type I interferons. It is also produced by maturing macrophages, dendritic cells or mastocytes. It plays a role in defense against pathogens apart from viruses. It affects the function of both innate and adaptive immune system. Besides described antiviral effects, IL-29 modulates cytokine production of other cells, for example, it
https://en.wikipedia.org/wiki/Hirschberg%E2%80%93Sinclair%20algorithm
The Hirschberg–Sinclair algorithm is a distributed algorithm designed for leader election problem in a synchronous ring network. It is named after its inventors, Dan Hirschberg and J. B. Sinclair. The algorithm requires the use of unique IDs (UID) for each process. The algorithm works in phases and sends its UID out in both directions. The message goes out a distance of 2Phase Number hops and then the message heads back to the originating process. While the messages are heading "out" each receiving process will compare the incoming UID to its own. If the UID is greater than its own UID then it will continue the message on. Otherwise if the UID is less than its own UID, it will not pass the information on. At the end of a phase, a process can determine if it will send out messages in the next round by if it received both of its incoming messages. Phases continue until a process receives both of its out messages, from both of its neighbors. At this time the process knows it is the largest UID in the ring and declares itself the leader.
https://en.wikipedia.org/wiki/Pre-charge
Pre-charge of the powerline voltages in a high voltage DC application is a preliminary mode which limits the inrush current during the power up procedure. A high-voltage system with a large capacitive load can be exposed to high electric current during initial turn-on. This current, if not limited, can cause considerable stress or damage to the system components. In some applications, the occasion to activate the system is a rare occurrence, such as in commercial utility power distribution. In other systems such as vehicle applications, pre-charge will occur with each use of the system, multiple times per day. Precharging is implemented to increase the lifespan of electronic components and increase reliability of the high voltage system. Background: inrush currents into capacitors Inrush currents into capacitive components are a key concern in power-up stress to components. When DC input power is applied to a capacitive load, the step response of the voltage input will cause the input capacitor to charge. The capacitor charging starts with an inrush current and ends with an exponential decay down to the steady state condition. When the magnitude of the inrush peak is very large compared to the maximum rating of the components, then component stress is to be expected. The current into a capacitor is known to be : the peak inrush current will depend upon the capacitance C and the rate of change of the voltage (dV/dT). The inrush current will increase as the capacitance value increases, and the inrush current will increase as the voltage of the power source increases. This second parameter is of primary concern in high voltage power distribution systems. By their nature, high voltage power sources will deliver high voltage into the distribution system. Capacitive loads will then be subject to high inrush currents upon power-up. The stress to the components must be understood and minimized. The objective of a pre-charge function is to limit the magnitude of the inru
https://en.wikipedia.org/wiki/Vapor%E2%80%93liquid%20equilibrium
In thermodynamics and chemical engineering, the vapor–liquid equilibrium (VLE) describes the distribution of a chemical species between the vapor phase and a liquid phase. The concentration of a vapor in contact with its liquid, especially at equilibrium, is often expressed in terms of vapor pressure, which will be a partial pressure (a part of the total gas pressure) if any other gas(es) are present with the vapor. The equilibrium vapor pressure of a liquid is in general strongly dependent on temperature. At vapor–liquid equilibrium, a liquid with individual components in certain concentrations will have an equilibrium vapor in which the concentrations or partial pressures of the vapor components have certain values depending on all of the liquid component concentrations and the temperature. The converse is also true: if a vapor with components at certain concentrations or partial pressures is in vapor–liquid equilibrium with its liquid, then the component concentrations in the liquid will be determined dependent on the vapor concentrations and on the temperature. The equilibrium concentration of each component in the liquid phase is often different from its concentration (or vapor pressure) in the vapor phase, but there is a relationship. The VLE concentration data can be determined experimentally or approximated with the help of theories such as Raoult's law, Dalton's law, and Henry's law. Such vapor–liquid equilibrium information is useful in designing columns for distillation, especially fractional distillation, which is a particular specialty of chemical engineers. Distillation is a process used to separate or partially separate components in a mixture by boiling (vaporization) followed by condensation. Distillation takes advantage of differences in concentrations of components in the liquid and vapor phases. In mixtures containing two or more components, the concentrations of each component are often expressed as mole fractions. The mole fraction of
https://en.wikipedia.org/wiki/Tight%20span
In metric geometry, the metric envelope or tight span of a metric space M is an injective metric space into which M can be embedded. In some sense it consists of all points "between" the points of M, analogous to the convex hull of a point set in a Euclidean space. The tight span is also sometimes known as the injective envelope or hyperconvex hull of M. It has also been called the injective hull, but should not be confused with the injective hull of a module in algebra, a concept with a similar description relative to the category of R-modules rather than metric spaces. The tight span was first described by , and it was studied and applied by Holsztyński in the 1960s. It was later independently rediscovered by and ; see for this history. The tight span is one of the central constructions of T-theory. Definition The tight span of a metric space can be defined as follows. Let (X,d) be a metric space, and let T(X) be the set of extremal functions on X, where we say an extremal function on X to mean a function f from X to R such that For any x, y in X, d(x,y) ≤ f(x) + f(y), and For each x in X, f(x) = sup{d(x,y) - f(y):y in X}. In particular (taking x = y in property 1 above) f(x) ≥ 0 for all x. One way to interpret the first requirement above is that f defines a set of possible distances from some new point to the points in X that must satisfy the triangle inequality together with the distances in (X,d). The second requirement states that none of these distances can be reduced without violating the triangle inequality. The tight span of (X,d) is the metric space (T(X),δ), where is analogous to the metric induced by the norm. (If d is bounded, then δ is the subspace metric induced by the metric induced by the norm. If d is not bounded, then every extremal function on X is unbounded and so Regardless, it will be true that for any f,g in T(X), the difference belongs to , i.e., is bounded.) Equivalent definitions of extremal functions For a function f from X
https://en.wikipedia.org/wiki/Nice%20name
In set theory, a nice name is used in forcing to impose an upper bound on the number of subsets in the generic model. It is used in the context of forcing to prove independence results in set theory such as Easton's theorem. Formal definition Let ZFC be transitive, a forcing notion in , and suppose is generic over . Then for any -name in , we say that is a nice name for a subset of if is a -name satisfying the following properties: (1) (2) For all -names , forms an antichain. (3) (Natural addition): If , then there exists in such that .
https://en.wikipedia.org/wiki/Maximal%20pair
In computer science, a maximal pair within a string is a pair of matching substrings that are maximal, where "maximal" means that it is not possible to make a longer matching pair by extending the range of both substrings to the left or right. Example For example, in this table, the substrings at indices 2 to 4 (in red) and indices 6 to 8 (in blue) are a maximal pair, because they contain identical characters (abc), and they have different characters to the left (x at index 1 and y at index 5) and different characters to the right (y at index 5 and w at index 9). Similarly, the substrings at indices 6 to 8 (in blue) and indices 10 to 12 (in green) are a maximal pair. However, the substrings at indices 2 to 4 (in red) and indices 10 to 12 (in green) are not a maximal pair, as the character y follows both substrings, and so they can be extended to the right to make a longer pair. Formal definition Formally, a maximal pair of substrings with starting positions and respectively, and both of length , is specified by a triple , such that, given a string of length , (meaning that the substrings have identical contents), but (they have different characters to their left) and (they also have different characters to their right; together, these two inequalities are the condition for being maximal). Thus, in the example above, the maximal pairs are (the red and blue substrings) and (the green and blue substrings), and is not a maximal pair. Related concepts and time complexity A maximal repeat is the string represented by a maximal pair. A supermaximal repeat is a maximal repeat never occurring as a proper substring of another maximal repeat. In the above example, abc and abcy are both maximal repeats, but only abcy is a supermaximal repeat. Maximal pairs, maximal repeats and supermaximal repeats can each be found in time using a suffix tree, if there are such structures.
https://en.wikipedia.org/wiki/Intermittent%20fault
An intermittent fault, often called simply an "intermittent", (or anecdotally "interfailing") is a malfunction of a device or system that occurs at intervals, usually irregular, in a device or system that functions normally at other times. Intermittent faults are common to all branches of technology, including computer software. An intermittent fault is caused by several contributing factors, some of which may be effectively random, which occur simultaneously. The more complex the system or mechanism involved, the greater the likelihood of an intermittent fault. Intermittent faults are not easily repeatable because of their complicated behavioral patterns. These are also sometimes referred to as “soft” failures, since they do not manifest themselves all the time and disappear in an unpredictable manner. In contrast, “hard” failures are permanent failures that occur over a period of time (or are sometimes instantaneous). They have a specific failure site (location of failure), mode (how the failure manifests itself), and mechanism, and there is no unpredictable recovery for the failed system. Since intermittent faults are not easily repeatable, it is more difficult to conduct a failure analysis for them, understand their root causes, or isolate their failure site than it is for permanent failures. Intermittent failures can be a cause of no-fault-found (NFF) occurrences in electronic products and systems. NFF implies that a failure (fault) occurred or was reported to have occurred during a product’s use. The product was analyzed or tested to confirm the failure, but “a failure or fault” could be not found. A common example of the NFF phenomenon occurs when your computer “hangs up”. Clearly, a “failure” has occurred. However, if the computer is rebooted, it often works again. The impact of NFF and intermittent failures can be profound. Due to their characteristics, manufacturers may assume a cause(s) rather than spend the time and cost to determine a root cause. For
https://en.wikipedia.org/wiki/Frequency%20separation
Frequency separation within astrophysics, is a term used in both Helioseismology and Asteroseismology. It refers to the spacing in frequency between adjacent modes of oscillation, having the same angular degree (l) but different radial order (n). For a Sun-like star, the frequency can be further described using the 'large frequency spacing' between modes of different radial order (136 μHz in the Sun), and the 'small frequency spacing' between modes of even and odd angular degree within the same radial order (9.0 μHz in the Sun). The period corresponding to the large frequency spacing can be shown to be approximately the same as the time required for a sound wave to travel to the centre of the Sun and return, confirming the global nature of the oscillations seen. A further frequency separation, the rotational splitting can be seen in high-resolution solar data between modes of the same angular degree, but different azimuthal order (m). This gives information
https://en.wikipedia.org/wiki/QF-Test
QF-Test from Quality First Software is a cross-platform software tool for the GUI test automation specialized on Java/Swing, SWT, Eclipse plug-ins and RCP applications, Java applets, Java Web Start, ULC and cross-browser test automation of static and dynamic web-based applications (HTML and AJAX frameworks like ExtJS, GWT, GXT, RAP, Qooxdoo, RichFaces, Vaadin, PrimeFaces, ICEfaces and ZK). Version 4.0 added Windows support for the Web browser Chrome, support for JavaFX and the AJAX frameworks jQuery UI and jQueryEasyUI were added. Testing of PDF documents via record and replay. RESTful web service testing. Overview QF-Test (the successor of qftestJUI, available since 2001) enables regression and load testing and runs on Windows and all major Unix systems. Its commercial application is primarily performed by testers and developers in the field of Quality Assurance. Since December 2008 a webtest add-on is available which allows for test automation of browser-based GUIs (like Internet Explorer and Firefox) in addition to its Java GUI testing functionality which was extended to JavaFX in July 2014. Features QF-Test's capture/replay function enables recording of tests for beginners, while modularization allows for creating large test suites in a concise arrangement. For the advanced user who requires even more control over his application, the tool offers access to internal program structures through the standard scripting languages Jython, the Java implementation of the popular Python language, JavaScript and Groovy. The tool also offers a batch mode, allowing to run tests unattended and then generate XML, HTML and JUnit reports. Thus the tool can be integrated into existing build/test frameworks like Jenkins, Ant or Maven. Another mode is the so-called Daemon mode for distributed test execution. A specific integration with many test management tools like HP Quality Center / HP ALM, QMetry, TestLink, SQS-TEST/Professional Suite, Rational Quality Manager, Scapa
https://en.wikipedia.org/wiki/Distributive%20law%20between%20monads
In category theory, an abstract branch of mathematics, distributive laws between monads are a way to express abstractly that two algebraic structures distribute one over the other one. Suppose that and are two monads on a category C. In general, there is no natural monad structure on the composite functor ST. However, there is a natural monad structure on the functor ST if there is a distributive law of the monad S over the monad T. Formally, a distributive law of the monad S over the monad T is a natural transformation such that the diagrams                   commute. This law induces a composite monad ST with as multiplication: , as unit: . See also distributive law
https://en.wikipedia.org/wiki/Lipschitz%20domain
In mathematics, a Lipschitz domain (or domain with Lipschitz boundary) is a domain in Euclidean space whose boundary is "sufficiently regular" in the sense that it can be thought of as locally being the graph of a Lipschitz continuous function. The term is named after the German mathematician Rudolf Lipschitz. Definition Let . Let be a domain of and let denote the boundary of . Then is called a Lipschitz domain if for every point there exists a hyperplane of dimension through , a Lipschitz-continuous function over that hyperplane, and reals and such that where is a unit vector that is normal to is the open ball of radius , In other words, at each point of its boundary, is locally the set of points located above the graph of some Lipschitz function. Generalization A more general notion is that of weakly Lipschitz domains, which are domains whose boundary is locally flattable by a bilipschitz mapping. Lipschitz domains in the sense above are sometimes called strongly Lipschitz by contrast with weakly Lipschitz domains. A domain is weakly Lipschitz if for every point there exists a radius and a map such that is a bijection; and are both Lipschitz continuous functions; where denotes the unit ball in and A (strongly) Lipschitz domain is always a weakly Lipschitz domain but the converse is not true. An example of weakly Lipschitz domains that fails to be a strongly Lipschitz domain is given by the two-bricks domain Applications of Lipschitz domains Many of the Sobolev embedding theorems require that the domain of study be a Lipschitz domain. Consequently, many partial differential equations and variational problems are defined on Lipschitz domains.
https://en.wikipedia.org/wiki/Freelance%20model
Freelance, in aerial, railway, naval, or bus model building, refers to companies that produce models that are not based on existing livery. Such models are sometimes frowned upon in the model-building community because they do not represent existing items, but are original designs. Because they require no licensing fees for trademark and design owners, and can thus be produced less expensively, freelance models are quite popular in the United States. They have not become popular in Europe, although a few European companies produce them. Freelance companies Railway FTL - Ferrovie e Tranvie Locali - Local Railroads and Tramways Beetrains SAFF - Società Anonima Ferrovie Federate - Joint-stock company Federate Railroads So.Ge.R.I.T. SITAV Società Intermodale Trasporti Alta Valle - High Valley Intermodal Society Transport FRA Ferrovie Regionali dell'Appennino - Regional Railways of Apennine Bus SAFF - Società Anonima Ferrovie Federate - Joint-stock company Federate Railroads Scale modeling
https://en.wikipedia.org/wiki/Consequentia%20mirabilis
Consequentia mirabilis (Latin for "admirable consequence"), also known as Clavius's Law, is used in traditional and classical logic to establish the truth of a proposition from the inconsistency of its negation. It is thus related to reductio ad absurdum, but it can prove a proposition using just its own negation and the concept of consistency. For a more concrete formulation, it states that if a proposition is a consequence of its negation, then it is true, for consistency. In formal notation: . Equivalent forms Given being equivalent to , the principle is equivalent to . History Consequentia mirabilis was a pattern of argument popular in 17th-century Europe that first appeared in a fragment of Aristotle's Protrepticus: "If we ought to philosophise, then we ought to philosophise; and if we ought not to philosophise, then we ought to philosophise (i.e. in order to justify this view); in any case, therefore, we ought to philosophise." Barnes claims in passing that the term consequentia mirabilis refers only to the inference of the proposition from the inconsistency of its negation, and that the term Lex Clavia (or Clavius' Law) refers to the inference of the proposition's negation from the inconsistency of the proposition. See also Ex falso quodlibet Tertium non datur Peirce's law
https://en.wikipedia.org/wiki/Interface%20conditions%20for%20electromagnetic%20fields
Interface conditions describe the behaviour of electromagnetic fields; electric field, electric displacement field, and the magnetic field at the interface of two materials. The differential forms of these equations require that there is always an open neighbourhood around the point to which they are applied, otherwise the vector fields and H are not differentiable. In other words, the medium must be continuous. On the interface of two different media with different values for electrical permittivity and magnetic permeability, that condition does not apply. However, the interface conditions for the electromagnetic field vectors can be derived from the integral forms of Maxwell's equations. Interface conditions for electric field vectors Electric field strength where: is normal vector from medium 1 to medium 2. Therefore, the tangential component of E is continuous across the interface. {| class="toccolours collapsible collapsed" width="80%" style="text-align:left" !Outline of proof from Faraday's law |- |We begin with the integral form of Faraday's law: Choose as a small square across the interface. Then, have the sides perpendicular to the interface shrink to infinitesimal length. The area of integration now looks like a line, which has zero area. In other words: Since remains finite in this limit, the whole right hand side goes to zero. All that is left is: Two of our sides are infinitesimally small, leaving only Assuming we made our square small enough that E is roughly constant, its magnitude can be pulled out of the integral. As the remaining sides to our original loop, the in each region run in opposite directions, so we define one of them as the tangent unit vector and the other as After dividing by l, and rearranging, This argument works for any tangential direction. The difference in electric field dotted into any tangential vector is zero, meaning only the components of parallel to the normal vector can change between mediums. Thus, th
https://en.wikipedia.org/wiki/BKM%20algorithm
The BKM algorithm is a shift-and-add algorithm for computing elementary functions, first published in 1994 by Jean-Claude Bajard, Sylvanus Kla, and Jean-Michel Muller. BKM is based on computing complex logarithms (L-mode) and exponentials (E-mode) using a method similar to the algorithm Henry Briggs used to compute logarithms. By using a precomputed table of logarithms of negative powers of two, the BKM algorithm computes elementary functions using only integer add, shift, and compare operations. BKM is similar to CORDIC, but uses a table of logarithms rather than a table of arctangents. On each iteration, a choice of coefficient is made from a set of nine complex numbers, 1, 0, −1, i, −i, 1+i, 1−i, −1+i, −1−i, rather than only −1 or +1 as used by CORDIC. BKM provides a simpler method of computing some elementary functions, and unlike CORDIC, BKM needs no result scaling factor. The convergence rate of BKM is approximately one bit per iteration, like CORDIC, but BKM requires more precomputed table elements for the same precision because the table stores logarithms of complex operands. As with other algorithms in the shift-and-add class, BKM is particularly well-suited to hardware implementation. The relative performance of software BKM implementation in comparison to other methods such as polynomial or rational approximations will depend on the availability of fast multi-bit shifts (i.e. a barrel shifter) or hardware floating point arithmetic. Overview In order to solve the equation the BKM algorithm takes advantage of a basic property of logarithms Using Pi notation, this identity generalizes to Because any number can be represented by a product, this allows us to choose any set of values which multiply to give the value we started with. In computer systems, it's much faster to multiply and divide by multiples of 2, but because not every number is a multiple of 2, using is a better option than a more simple choice of . Since we want to start with large cha
https://en.wikipedia.org/wiki/MTV-1
The MTV-1 Micro TV was the second model of a near pocket-sized television. The first was the Panasonic IC model TR-001 introduced in 1970. The MTV-1 was developed by Clive Sinclair (Sinclair Radionics Ltd). It was shown to the public at trade shows in London and Chicago in January, 1977, and released for sale in 1978. Development spanned 10 years and included a cash infusion of (about ) from the UK government in 1976. The MTV-1 used a German AEG Telefunken black-and-white, electrostatic deflection cathode ray tube (CRT) and included a rechargeable 4-AA-cell NiCad battery pack. It measured and weighed . It was able to receive either PAL or NTSC transmissions on VHF or UHF. A Welsh company, Wolsey Electronics, manufactured it for Sinclair. Custom ICs made by Texas Instruments and Sinclair contributed to its small size and low power consumption. The original (about ) price tag proved to be too high to sell many of them, and Sinclair lost over in 1978, eventually selling its remaining inventory to liquidators at greatly reduced prices. The MTV-1B, released later in 1978 at the much lower price of , was able to receive only British and South African UHF PAL signals. The original MTV-1 was the world's first multistandard TV, being capable of switching between PAL and NTSC.
https://en.wikipedia.org/wiki/S%C3%ADlvio%20Meira
Sílvio Romero de Lemos Meira (Taperoá, Paraíba, 2 February 1955) is a Brazilian computer scientist, professor and entrepreneur in the area of software engineering and innovation. Currently, he is an emeritus professor at the Centre of Informatics of the Federal University of Pernambuco. Additionally, he also is the chairman of the Porto Digital - the biggest Brazilian tech park - and is member of the board of directors in several companies. Meira is also founder and chief scientist at tds.company. He is co-founder of the Recife Center for Advanced Studies and Systems, where he served as chief scientist for 12 years.
https://en.wikipedia.org/wiki/Liquid%20state%20machine
A liquid state machine (LSM) is a type of reservoir computer that uses a spiking neural network. An LSM consists of a large collection of units (called nodes, or neurons). Each node receives time varying input from external sources (the inputs) as well as from other nodes. Nodes are randomly connected to each other. The recurrent nature of the connections turns the time varying input into a spatio-temporal pattern of activations in the network nodes. The spatio-temporal patterns of activation are read out by linear discriminant units. The soup of recurrently connected nodes will end up computing a large variety of nonlinear functions on the input. Given a large enough variety of such nonlinear functions, it is theoretically possible to obtain linear combinations (using the read out units) to perform whatever mathematical operation is needed to perform a certain task, such as speech recognition or computer vision. The word liquid in the name comes from the analogy drawn to dropping a stone into a still body of water or other liquid. The falling stone will generate ripples in the liquid. The input (motion of the falling stone) has been converted into a spatio-temporal pattern of liquid displacement (ripples). LSMs have been put forward as a way to explain the operation of brains. LSMs are argued to be an improvement over the theory of artificial neural networks because: Circuits are not hard coded to perform a specific task. Continuous time inputs are handled "naturally". Computations on various time scales can be done using the same network. The same network can perform multiple computations. Criticisms of LSMs as used in computational neuroscience are that LSMs don't actually explain how the brain functions. At best they can replicate some parts of brain functionality. There is no guaranteed way to dissect a working network and figure out how or what computations are being performed. There is very little control over the process. Universal function approximati
https://en.wikipedia.org/wiki/Ross%E2%80%93Littlewood%20paradox
The Ross–Littlewood paradox (also known as the balls and vase problem or the ping pong ball problem) is a hypothetical problem in abstract mathematics and logic designed to illustrate the paradoxical, or at least non-intuitive, nature of infinity. More specifically, like the Thomson's lamp paradox, the Ross–Littlewood paradox tries to illustrate the conceptual difficulties with the notion of a supertask, in which an infinite number of tasks are completed sequentially. The problem was originally described by mathematician John E. Littlewood in his 1953 book Littlewood's Miscellany, and was later expanded upon by Sheldon Ross in his 1988 book A First Course in Probability. The problem starts with an empty vase and an infinite supply of balls. An infinite number of steps are then performed, such that at each step 10 balls are added to the vase and 1 ball removed from it. The question is then posed: How many balls are in the vase when the task is finished? To complete an infinite number of steps, it is assumed that the vase is empty at one minute before noon, and that the following steps are performed: The first step is performed at 30 seconds before noon. The second step is performed at 15 seconds before noon. Each subsequent step is performed in half the time of the previous step, i.e., step n is performed at 2 minutes before noon. This guarantees that a countably infinite number of steps is performed by noon. Since each subsequent step takes half as much time as the previous step, an infinite number of steps is performed by the time one minute has passed. The question is then: How many balls are in the vase at noon? Solutions Answers to the puzzle fall into several categories. Vase contains infinitely many balls The most intuitive answer seems to be that the vase contains an infinite number of balls by noon, since at every step along the way more balls are being added than removed. By definition, at each step, there will be a greater number of balls than at
https://en.wikipedia.org/wiki/Hardware-in-the-loop%20simulation
Hardware-in-the-loop (HIL) simulation, HWIL, or HITL, is a technique that is used in the development and testing of complex real-time embedded systems. HIL simulation provides an effective testing platform by adding the complexity of the process-actuator system, known as a plant, to the test platform. The complexity of the plant under control is included in testing and development by adding a mathematical representation of all related dynamic systems. These mathematical representations are referred to as the "plant simulation". The embedded system to be tested interacts with this plant simulation. How HIL works HIL simulation must include electrical emulation of sensors and actuators. These electrical emulations act as the interface between the plant simulation and the embedded system under test. The value of each electrically emulated sensor is controlled by the plant simulation and is read by the embedded system under test (feedback). Likewise, the embedded system under test implements its control algorithms by outputting actuator control signals. Changes in the control signals result in changes to variable values in the plant simulation. For example, a HIL simulation platform for the development of automotive anti-lock braking systems may have mathematical representations for each of the following subsystems in the plant simulation: Vehicle dynamics, such as suspension, wheels, tires, roll, pitch and yaw; Dynamics of the brake system's hydraulic components; Road characteristics. Uses In many cases, the most effective way to develop an embedded system is to connect the embedded system to the real plant. In other cases, HIL simulation is more efficient. The metric of development and testing efficiency is typically a formula that includes the following factors: 1. Cost 2. Duration 3. Safety 4. Feasibility The cost of the approach should be a measure of the cost of all tools and effort. The duration of development and testing affects the time-to-market for
https://en.wikipedia.org/wiki/Sensitive%20skin%20%28electronic%20device%29
Sensitive skin, also known as sensate skin or electronic skin (e-skin), is an electronic sensing skin placed on the surface of a machine such as a robotic arm. The goal of the skin is to sense important environmental parameters—such as proximity to objects, heat, moisture, and direct touch sensations. Examples of a sensitive skin have been made by a group at the University of Tokyo led by Prof. Takao Someya. The group's work has been published in Nature, Science and Advanced Materials journals, as well as Proceedings of the (United States) National Academy of Sciences. They have demonstrated pressure and temperature measurements at a resolution of a few millimeters, as well as organic transistors, organic solar cells, and organic light emitting diodes about 1 μm thick. The sensor arrays are based on many organic thin film transistors that are turned on one at a time and the current that flows through the sensor (electrical resistance-based) tells you how much pressure is being sensed. They have simultaneously measured pressure and temperature by using two separate arrays and laying one on top of another. See also Electronic nose Electronic skin Tactile sensor
https://en.wikipedia.org/wiki/Dual%20bundle
In mathematics, the dual bundle is an operation on vector bundles extending the operation of duality for vector spaces. Definition The dual bundle of a vector bundle is the vector bundle whose fibers are the dual spaces to the fibers of . Equivalently, can be defined as the Hom bundle that is, the vector bundle of morphisms from to the trivial line bundle Constructions and examples Given a local trivialization of with transition functions a local trivialization of is given by the same open cover of with transition functions (the inverse of the transpose). The dual bundle is then constructed using the fiber bundle construction theorem. As particular cases: The dual bundle of an associated bundle is the bundle associated to the dual representation of the structure group. The dual bundle of the tangent bundle of a differentiable manifold is its cotangent bundle. Properties If the base space is paracompact and Hausdorff then a real, finite-rank vector bundle and its dual are isomorphic as vector bundles. However, just as for vector spaces, there is no natural choice of isomorphism unless is equipped with an inner product. This is not true in the case of complex vector bundles: for example, the tautological line bundle over the Riemann sphere is not isomorphic to its dual. The dual of a complex vector bundle is indeed isomorphic to the conjugate bundle but the choice of isomorphism is non-canonical unless is equipped with a hermitian product. The Hom bundle of two vector bundles is canonically isomorphic to the tensor product bundle Given a morphism of vector bundles over the same space, there is a morphism between their dual bundles (in the converse order), defined fibrewise as the transpose of each linear map Accordingly, the dual bundle operation defines a contravariant functor from the category of vector bundles and their morphisms to itself.
https://en.wikipedia.org/wiki/Computer-based%20mathematics%20education
Computer-based mathematics education (CBME) is an approach to teaching mathematics that emphasizes the use of computers. Computers in math education Computers are used in education in a number of ways, such as interactive tutorials, hypermedia, simulations and educational games. Tutorials are types of software that present information, check learning by question/answer method, judge responses, and provide feedback. Educational games are more like simulations and are used from the elementary to college level. E learning systems can deliver math lessons and exercises and manage homework assignments. See also ALEKS, a computer-based education system that includes mathematics among its curricula Computer-Based Math, a project aimed at using computers for computational tasks and spending more classroom time on applications Mathletics (educational software), a popular K-12 mathematics learning program from 3P Learning Mathspace, a similar program for students aged 7-18, founded in Australia in 2010 Sokikom, a team-based math learning game
https://en.wikipedia.org/wiki/Fibrin%20ring%20granuloma
A fibrin ring granuloma, also known as doughnut granuloma, is a histopathological finding that is characteristic of Q fever. On hematoxylin-eosin staining, the fibrin ring granuloma consists of a central lipid vacuole (usually washed-out during fixing and staining, leaving only an empty hole) surrounded by a dense red fibrin ring and epithelioid macrophages. Fibrin ring granulomas may also be seen in Hodgkin's disease and infectious mononucleosis. See also Granuloma
https://en.wikipedia.org/wiki/Vigilante%20%28video%20game%29
is a 1988 beat 'em up arcade video game developed and published by Irem in Japan and Europe, and published in North America by Data East. It is considered as a spiritual sequel to Irem's earlier Kung-Fu Master (1984). Plot The game takes place in downtown New York City. The game's plot involves a lone, professional martial artist who became a vigilante to fight an evil gang called the Skinheads ruled by a man known as the Giant Devil, in order to protect his "turf" and save a female hostage named Madonna, who was kidnapped by them. Gameplay Players control the titular character using punches and kicks to defeat the Skinheads in a 2D platform manner, while sometimes picking up and using nunchaku against them. If players get hurt while holding nunchuku, they become unarmed. There are five stages in order of appearance: a street, a junkyard, the Brooklyn Bridge, a back street scene and on top of a building that is under construction. Skinheads with Mohawk or spiked hairdo attack the vigilante with knives, chains, motorbikes, guns and other kinds of weapons. They will also choke him if he lets them get too close. Development An arcade sequel to Kung-Fu Master called Beyond Kung-Fu: Return of the Master was developed by Irem and underwent location testing in 1987, but was shelved after it underperformed. The Kung-Fu sequel was then revamped into Vigilante, after Irem decided to give the game a more Americanized setting, which was released in 1988. Ports The arcade game was later ported to several different home computers and consoles. The Master System version was published exclusively in North America and Europe by Sega, and is one of several games in the console to include an FM sound switch for enhanced music quality. In the Master System version, Madonna was renamed "Maria" and the Skinheads were called the "Rogues". The ones for Commodore 64, ZX Spectrum, Atari ST, Amiga and the Amstrad CPC were reprogrammed by Emerald Software and published by U.S. Gold mostl
https://en.wikipedia.org/wiki/List%20of%20open-source%20health%20software
The following is a list of notable software packages and applications licensed under an open-source license or in the public domain for use in the health care industry. Public health and biosurveillance Epi Info is public domain statistical software for epidemiology developed by Centers for Disease Control and Prevention. Spatiotemporal Epidemiological Modeler is a tool, originally developed at IBM Research, for modelings and visualizing the spread of infectious diseases. It is maintained by the Eclipse Foundation and available under terms of the Eclipse Public License. Electronic records and medical practice management Health system management iHRIS is an integrated Human Resource Information System developed by IntraHealth International under USAID-funded CapacityPlus project and deployed in more than 20 countries in the world. iHRIS is distributed under the GNU GPL. DHIS is a district health management information system and data warehouse. DHIS2 is released under the BSD license. HRHIS is a human resource for health information system for management of human resources for health developed by University of Dar es Salaam college of information and communication technology, Department of Computer Science and Engineering, for Ministry of Health and Social Welfare (Tanzania) and funded by the Japan International Cooperation Agency (JICA). Disease management Breathing Games is a series of research-backed, co-created games to prevent, diagnose and treat chronic respiratory diseases. They are released under the Peer Production licence. Nightscout is a collection of software tools, including mobile clients, to enable DIY cloud-based continuous glucose monitoring "…for informational and educational purposes." Individual components are available under various open-source licenses, including the GNU GPL, GNU AGPL, MIT License, and BSD licenses. Imaging/visualization CamBA is a collection of neuroimaging pipelines distributed under the GNU GPL. Drishti is a volu
https://en.wikipedia.org/wiki/Gurdon%20Saltonstall%20Hubbard
Gurdon Saltonstall Hubbard (August 22, 1802 – September 14, 1886) was an American fur trader, insurance underwriter, and land speculator. He was influential in the development of the city of Chicago and responsible for its growth during the 19th century. First arriving in Chicago in 1818, he settled in the area in the late 1820s. He became one of the most prominent residents of the town and was one of its first trustees in 1833. He went on to build Chicago's first stockyard and help foment a land boom for Chicago in the East. In addition to his work in developing and promoting Chicago, Hubbard was known for his athletic prowess. Hubbard Street in Chicago is named for him, as is Hubbard High School. Early life Gurdon Saltonstall Hubbard was born in Windsor, Vermont, on August 22, 1802. His parents were Abigail Sage and Elizur Hubbard. Sage was a daughter of the general Comfort Sage and his wife Sarah Hamlin, and was from Middletown, Connecticut. The elder Hubbard was a son of the Revolutionary War officer George Hubbard and his wife Thankful Hatch. Hubbard was descended from and named for the Connecticut governor Gurdon Saltonstall. Hubbard disliked going to school at an early age and was often truant. When his father, a lawyer, lost his money in 1812 in speculative ventures, his aunt took him in at her house in Bridgewater, Massachusetts, where he was educated. The situation of the Hubbards continued to worsen, and the elder Hubbard decided to take the rest of the family north to settle in Montreal and continue his practice. Hubbard left Bridgewater to join them and they departed in early May 1815. The family discovered upon their arrival that Elizur could not practice law as he was an American, but made money by renting a house to lodgers. Hubbard started working at a local hardware store in April 1816. He befriended John Dyde, the son of a boarding house proprietor. Dyde's father housed William Matthews, an agent of the American Fur Company. In 1818 an expedit
https://en.wikipedia.org/wiki/Digital%20program%20insertion
Digital program insertion (DPI) allows cable headends and broadcast affiliates to insert locally generated commercials and short programs into remotely distributed regional programs before they are delivered to home viewers. Digital program insertion also refers to a specific technology which allows an MPEG transport stream to be spliced into a currently flowing MPEG transport stream seamlessly and with little or no artifacts. The controlling signaling used to initiate an MPEG is referred to as an SCTE-35 message. The communication API between MPEG splicers and content delivery servers or ad insertion servers is referred to as SCTE30 messages.
https://en.wikipedia.org/wiki/Mark%20Nemenman
Mark Nemenman (, ) (6 November 1936, Minsk, Belarus - 20 September 2022, San Mateo, California) was a Soviet computer scientist, notable as a pioneer in systems programming and programming language research. He was one of the main developers of the AKI language (in Russian АКИ - АвтоКод ИНЖЕНЕР - Engineer's Autocode) in 1964, before BASIC became known. He led the development of system software for Minsk-32, the most popular of Minsk family of computers. He was awarded Lenin Komsomol Prize in 1970, received his Ph.D. in 1975 (scientific advisor - Andrey Ershov), Professor since 1984. Nemenman authored more than 70 papers and 5 books. He is the father of two sons, one of whom is theoretical physicist Ilya Nemenman. Books Nemenman, Tsagelsky, Matyushevskaya Autocode for engineering problems solving on Minsk 2 Minsk, 1965 Nemenman Programming in AKI Minsk, 1972 Kushnerev, Nemenman, Tsagelsky Programming for Computer Minsk-32 Moscow, 1973 Belokurskaya, Kushnerev, Nemenman Minsk 32 Dispatcher Moscow, 1973 Lopato, Nemenman, Pykhtin, Tikmenov Personal-professional Computers Moscow, 1988 Belokurskaya, Emelyanchik, Nemenman Personal Computers ES. Abacus Package Moscow, 19 External links Biography (in Russian) Documents from Andrey Ershov Archive Minsk family of computers Nemenman, Mark oral history (transcript) / (video recording) by Computer History Museum 1936 births 20th-century Belarusian Jews 21st-century Belarusian Jews Living people Scientists from Minsk Belarusian Jews Belarusian computer scientists Computer programmers American people of Belarusian-Jewish descent Soviet computer scientists
https://en.wikipedia.org/wiki/Grandi%27s%20series
In mathematics, the infinite series , also written is sometimes called Grandi's series, after Italian mathematician, philosopher, and priest Guido Grandi, who gave a memorable treatment of the series in 1703. It is a divergent series, meaning that it does not have a sum. However, it can be manipulated to yield a number of mathematically interesting results. For example, many summation methods are used in mathematics to assign numerical values even to a divergent series. For example, the Cesàro summation and the Ramanujan summation of this series is 1/2. Unrigorous methods One obvious method to find the sum of the series 1 − 1 + 1 − 1 + 1 − 1 + 1 − 1 + ... is to treat it like a telescoping series and perform the subtractions in place: (1 − 1) + (1 − 1) + (1 − 1) + ... = 0 + 0 + 0 + ... = 0. On the other hand, a similar bracketing procedure leads to the apparently contradictory result 1 + (−1 + 1) + (−1 + 1) + (−1 + 1) + ... = 1 + 0 + 0 + 0 + ... = 1. Thus, by applying parentheses to Grandi's series in different ways, one can obtain either 0 or 1 as a "value". (Variations of this idea, called the Eilenberg–Mazur swindle, are sometimes used in knot theory and algebra.). By taking the average of these two "values", one can justify that the series converges to . Treating Grandi's series as a divergent geometric series and using the same algebraic methods that evaluate convergent geometric series to obtain a third value: S = 1 − 1 + 1 − 1 + ..., so 1 − S = 1 − (1 − 1 + 1 − 1 + ...) = 1 − 1 + 1 − 1 + ... = S' 1 − S = S1 = 2S, resulting in S = . The same conclusion results from calculating −S, subtracting the result from S, and solving 2S = 1. The above manipulations do not consider what the sum of a series actually means and how said algebraic methods can be applied to divergent geometric series. Still, to the extent that it is important to be able to bracket series at will, and that it is more important to be able to perform arithmetic with them, one can arrive a
https://en.wikipedia.org/wiki/The%20Cobra%20Event
The Cobra Event is a 1998 thriller novel by Richard Preston describing an attempted bioterrorism attack on the United States. The perpetrator of the attack has genetically engineered a virus, called "Cobra", that fuses the incurable and highly contagious common cold with one of the world's most virulent diseases, smallpox. The disease that results from the virus, called brainpox in the novel, has symptoms that mimic those of Lesch–Nyhan syndrome and the common cold. The book is divided between descriptions of the virus and the government's attempt to stop the imminent threat posed by it. Plot summary The book is divided into six sections. The first section, called "Trial", starts with a teenaged girl named Kate Moran who dies violently one day in school. The next section, titled "1969", describes tests done in the 1960s by the U.S. government involving weaponized viruses. The third section, "Diagnosis", describes the autopsy of Kate Moran and introduces the key characters of Dr. Alice Austen, Mark Littleberry, and Will Hopkins. The last three sections—"Decision", "Reachdeep", and "The Operation"—describe these three characters' journeys to discover the source of the lethal Cobra virus. "Cobra" and its effects The virus described in the novel is a fictional chimera that attacks the human brain. The infective agent, code-named "Cobra" by the protagonists, is a recombinant virus made from modified variants of the nuclear polyhedrosis virus (which normally infects moths and butterflies), rhinovirus, and smallpox. The infection initially presents common cold-like symptoms and a characteristic blistering process in the nose and mouth, before invading the nervous system. Although not as contagious as the influenza virus, it spreads rapidly through the same vectors as the common cold, mainly via airborne particulate matter coming in contact with the mucous membranes of the respiratory system. Although tussis is the primary source of these particles, the inclusion of nucl
https://en.wikipedia.org/wiki/Linux%20Trace%20Toolkit
The Linux Trace Toolkit (LTT) is a set of tools that is designed to log program execution details from a patched Linux kernel and then perform various analyses on them, using console-based and graphical tools. LTT has been mostly superseded by its successor LTTng (Linux Trace Toolkit Next Generation). LTT allows the user to see in-depth information about the processes that were running during the trace period, including when context switches occurred, how long the processes were blocked for, and how much time the processes spent executing vs. how much time the processes were blocked. The data is logged to a text file and various console-based and graphical (GTK+) tools are provided for interpreting that data. In order to do data collection, LTT requires a patched Linux kernel. The authors of LTT claim that the performance hit for a patched kernel compared to a regular kernel is minimal; Their testing has reportedly shown that this is less than 2.5% on a "normal use" system (measured using batches of kernel makes) and less than 5% on a file I/O intensive system (measured using batches of tar). Usage Collecting trace data Data collection is Started by: trace 15 foo This command will cause the LTT tracedaemon to do a trace that lasts for 15 seconds, writing trace data to foo.trace and process information from the /proc filesystem to foo.proc. The trace command is actually a script which runs the program tracedaemon with some common options. It is possible to run tracedaemon directly and in that case, the user can use a number of command-line options to control the data which is collected. For the complete list of options supported by tracedaemon, see the online manual page for tracedaemon. Viewing the results Viewing the results of a trace can be accomplished with: traceview foo This command will launch a graphical (GTK+) traceview tool that will read from foo.trace and foo.proc. This tool can show information in various interesting ways, including Eve
https://en.wikipedia.org/wiki/Pascal%20matrix
In mathematics, particularly matrix theory and combinatorics, a Pascal matrix is a matrix (possibly infinite) containing the binomial coefficients as its elements. It is thus an encoding of Pascal's triangle in matrix form. There are three natural ways to achieve this: as a lower-triangular matrix, an upper-triangular matrix, or a symmetric matrix. For example, the 5 × 5 matrices are: There are other ways in which Pascal's triangle can be put into matrix form, but these are not easily extended to infinity. Definition The non-zero elements of a Pascal matrix are given by the binomial coefficients: such that the indices i, j start at 0, and ! denotes the factorial. Properties The matrices have the pleasing relationship Sn = LnUn. From this it is easily seen that all three matrices have determinant 1, as the determinant of a triangular matrix is simply the product of its diagonal elements, which are all 1 for both Ln and Un. In other words, matrices Sn, Ln, and Un are unimodular, with Ln and Un having trace n. The trace of Sn is given by with the first few terms given by the sequence 1, 3, 9, 29, 99, 351, 1275, … . Construction A Pascal matrix can actually be constructed by taking the matrix exponential of a special subdiagonal or superdiagonal matrix. The example below constructs a 7 × 7 Pascal matrix, but the method works for any desired n × n Pascal matrices. The dots in the following matrices represent zero elements. It is important to note that one cannot simply assume exp(A) exp(B) = exp(A + B), for n × n matrices A and B; this equality is only true when AB = BA (i.e. when the matrices A and B commute). In the construction of symmetric Pascal matrices like that above, the sub- and superdiagonal matrices do not commute, so the (perhaps) tempting simplification involving the addition of the matrices cannot be made. A useful property of the sub- and superdiagonal matrices used for the construction is that both are nilpotent; that is, when raised to a suf
https://en.wikipedia.org/wiki/Measuring%20the%20World
Measuring the World () is a novel by Austrian author Daniel Kehlmann, published in 2005 by Rowohlt Verlag, Reinbek. The novel re-imagines the lives of German mathematician Carl Friedrich Gauss and German geographer Alexander von Humboldt—who was accompanied on his journeys by French explorer Aimé Bonpland—and their many groundbreaking ways of taking the world's measure, as well as Humboldt's and Bonpland's travels in America and their meeting in 1828. One subplot fictionalises the conflict between Gauss and his son Eugene; while Eugene wanted to become a linguist, his father decreed that he study law. The book was a bestseller; by 2012 it had sold more than 2.3 million copies in Germany alone. A film version directed by Detlev Buck was released in 2012. Translations The English translation is by Carol Brown Janeway (November 2006). Rosa Pilar Blanco translated the book into Spanish.
https://en.wikipedia.org/wiki/Program%20of%20All-Inclusive%20Care%20for%20the%20Elderly
Program of All-inclusive Care for the Elderly (PACE) are programs within the United States that provide comprehensive health services for individuals age 55 and over who are sufficiently frail to be categorized as "nursing home eligible" by their state's Medicaid program. The ultimate goal of PACE programs is to keep eligible older adults out of nursing homes and within their communities for as long as possible. Services include primary and specialty medical care, nursing, nutrition, social services, therapies (occupational, physical, speech, recreation, etc.), pharmaceuticals, day health center services, home care, health-related transportation, minor modification to the home to accommodate disabilities, and anything else the program determines is medically necessary to maximize a member's health. If you or a loved one are eligible for nursing home level care but prefer to continue living at home, a PACE program can provide expansive health care and social opportunities during the day while you retain the comfort and familiarity of your home outside of day hours. History PACE was developed by On Lok Senior Health Services, a not-for-profit community-based organization launched in the early 1970s in the Chinatown-North Beach area of San Francisco, California. 1971–1978 On Lok Senior Health Services was created in 1971 to address the long-term care needs of older immigrants in San Francisco's Chinatown-North Beach neighborhood. After its founding, between 1973 and 1975, On Lok expanded to include day centers, in-home care, home-delivered meals, and housing assistance. In 1974, On Lok started being reimbursed by Medicaid for its provision of adult day health services. Later, in 1978, these health services were broadened to include comprehensive medical care for older adults certified to be nursing home-eligible. 1979 The Department of Health and Human Services provided a four-year grant to On Lok to develop a model of care delivery for individuals with long-ter
https://en.wikipedia.org/wiki/Combinatorial%20explosion
In mathematics, a combinatorial explosion is the rapid growth of the complexity of a problem due to how the combinatorics of the problem is affected by the input, constraints, and bounds of the problem. Combinatorial explosion is sometimes used to justify the intractability of certain problems. Examples of such problems include certain mathematical functions, the analysis of some puzzles and games, and some pathological examples which can be modelled as the Ackermann function. Examples Latin squares A Latin square of order is an array with entries from a set of elements with the property that each element of the set occurs exactly once in each row and each column of the array. An example of a Latin square of order three is given by, {| class="wikitable" style="margin-left:auto;margin-right:auto;text-align:center;width:6em;height:6em;table-layout:fixed;" |- | 1|| 2 || 3 |- | 2 || 3 || 1 |- | 3 || 1 || 2 |} A common example of a Latin square would be a completed Sudoku puzzle. A Latin square is a combinatorial object (as opposed to an algebraic object) since only the arrangement of entries matters and not what the entries actually are. The number of Latin squares as a function of the order (independent of the set from which the entries are drawn) provides an example of combinatorial explosion as illustrated by the following table. Sudoku A combinatorial explosion can also occur in some puzzles played on a grid, such as Sudoku. A Sudoku is a type of Latin square with the additional property that each element occurs exactly once in sub-sections of size (called boxes). Combinatorial explosion occurs as increases, creating limits to the properties of Sudokus that can be constructed, analyzed, and solved, as illustrated in the following table. Games One example in a game where combinatorial complexity leads to a solvability limit is in solving chess (a game with 64 squares and 32 pieces). Chess is not a solved game. In 2005 all chess game endings with six piec
https://en.wikipedia.org/wiki/Genetic%20algorithms%20in%20economics
Genetic algorithms have increasingly been applied to economics since the pioneering work by John H. Miller in 1986. It has been used to characterize a variety of models including the cobweb model, the overlapping generations model, game theory, schedule optimization and asset pricing. Specifically, it has been used as a model to represent learning, rather than as a means for fitting a model. Genetic algorithm in the cobweb model The cobweb model is a simple supply and demand model for a good over t periods. Firms (agents) make a production quantity decision in a given period, however their output is not produced until the following period. Thus, the firms are going to have to use some sort of method to forecast what the future price will be. The GA is used as a sort of learning behaviour for the firms. Initially their quantity production decisions are random, however each period they learn a little more. The result is the agents converge within the area of the rational expectations (RATEX) equilibrium for the stable and unstable case. If the election operator is used, the GA converges exactly to the RATEX equilibrium. There are two types of learning methods these agents can be deployed with: social learning and individual learning. In social learning, each firm is endowed with a single string which is used as its quantity production decision. It then compares this string against other firms' strings. In the individual learning case, agents are endowed with a pool of strings. These strings are then compared against other strings within the agent's population pool. This can be thought of as mutual competing ideas within a firm whereas in the social case, it can be thought of as a firm learning from more successful firms. Note that in the social case and in the individual learning case with identical cost functions, that this is a homogeneous solution, that is all agents' production decisions are identical. However, if the cost functions are not identical, this wil
https://en.wikipedia.org/wiki/Model%20maker
A Model maker is a professional Craftsperson who creates a three-dimensional representation of a design or concept. Most products in use and in development today first take form as a model. This "model" may be an exacting duplicate (prototype) of the future design or a simple mock-up of the general shape or concept. Many prototype models are used for testing physical properties of the design, others for usability and marketing studies. Mock-ups are generally used as part of the design process to help convey each new iteration. Some model makers specialize in "scale models" that allow an easier grasp of the whole design or for portability of the model to a trade show or an architect or client's office. Other scale models are used in museum displays and in the movie special effects industry. Model makers work in many environments from private studio/shops to corporate design and engineering facilities to research laboratories. The model maker must be highly skilled in the use of many machines, such as manual lathes, manual mills, Computer Numeric Control (CNC) machines, lasers, wire EDM, water jet saws, tig welders, sheet metal fabrication tools and wood working tools. Fabrication processes model makers take part in are powder coating, shearing, punching, plating, folding, forming and anodizing. Some model makers also use increasingly automated processes, for example cutting parts directly with digital data from computer-aided design plans on a CNC mill or creating the parts through rapid prototyping. Hand tools used by a model maker are an exacto knife, tweezers, sprue cutter, tape, glue, paint, and paint brushes. There are two basic processes used by the model maker to create models: additive and subtractive. Additive can be as simple as adding clay to create a form, sculpting and smoothing to the final shape. Body fillers, foam and resins are also used in the same manner. Most rapid prototyping technologies are based on the additive process, solidifying thin la
https://en.wikipedia.org/wiki/Bilaminar%20embryonic%20disc
The bilaminar embryonic disc, bilaminar blastoderm or embryonic disc is the distinct two-layered structure of cells formed in an embryo. In the development of the human embryo this takes place by day eight. It is formed when the inner cell mass, also known as the embryoblast, forms a bilaminar disc of two layers, an upper layer called the epiblast (primitive ectoderm) and a lower layer called the hypoblast (primitive endoderm), which will eventually form into fetus. These two layers of cells are stretched between two fluid-filled cavities at either end: the primitive yolk sac and the amniotic sac. The epiblast is adjacent to the trophoblast and made of columnar epithelial cells; the hypoblast is closest to the blastocoel (blastocystic cavity) and made of cuboidal cells. As the two layers become evident, a basement membrane forms between the layers. This distinction of layers of the bilaminar disc defines the primitive dorso ventral axis and polarity in embryogenesis. The epiblast migrates away from the trophoblast downwards, forming the amniotic cavity in between, the lining of which is formed from amnioblasts developed from the epiblast. The hypoblast is pushed down and forms the yolk sac (exocoelomic cavity) lining. Some hypoblast cells migrate along the inner cytotrophoblast lining of the blastocoel, secreting an extracellular matrix along the way. These hypoblast cells and extracellular matrix are called Heuser's membrane (or the exocoelomic membrane), and they cover the blastocoel to form the yolk sac (or exocoelomic cavity). Cells of the hypoblast migrate along the outer edges of this reticulum and form the extraembryonic mesoderm; this disrupts the extraembryonic reticulum. Soon pockets form in the reticulum, which ultimately coalesce to form the chorionic cavity (extraembryonic coelom). Initial formation The one-celled zygote, a eukaryotic cell formed by a fertilization event between two gametes at the start of embryonic development, undergoes cleavage b
https://en.wikipedia.org/wiki/Supplementum%20Plantarum
, commonly abbreviated to or just , and further abbreviated by botanists to Suppl. Pl., is a 1782 book by Carolus Linnaeus the Younger. Written entirely in Latin, it was intended as a supplement to the 1737 and the 1753 , both written by the author's father, the "father of modern taxonomy", Carl Linnaeus. Its full title means: “Supplement of Plants, the 13th edition of A System of Vegetables, the 6th edition of The Genera of Plants and the 2nd edition of The Species of Plants”, listing the components of the book in order of presentation. The Systematis Vegetabilium (13th edition) in the title refers to as published in 1774 by Johan Andreas Murray, a student of Linnaeus, Sr. The cover page indicates that it was published in 1781, and it was long believed to have been published in October of that year. In 1976, however, Hermann Manitz used a letter written by Jakob Friedrich Ehrhart to show that it had in fact been published in April 1782. Furthermore, the cover page states that the book was originally printed in Brunswick (), Lower Saxony, northwestern Germany by the printshop Orphanotropheum ( means ‘At the expense of the Orphanotropheum’). The book has 467 pages. The work was translated by Erasmus Darwin's Lichfield Botanical Society as A System of Vegetables (1785). It leaves the binomial nomenclature untranslated in the original Latin, but uses English in the keys and descriptions.
https://en.wikipedia.org/wiki/John%20Scales%20Avery
John Scales Avery (born in 1933) is a theoretical chemist noted for his research publications in quantum chemistry, thermodynamics, evolution, and history of science. Since the early 1990s, Avery has been an active world peace activist. During these years, he was part of a group associated with the Pugwash Conferences on Science and World Affairs. In 1995, this group received the Nobel Peace Prize for their efforts. Presently, he is an Associate Professor in quantum chemistry at the University of Copenhagen. His 2003 book Information Theory and Evolution set forth the view that the phenomenon of life, including its origin and evolution, that including human cultural evolution, has it background situated over thermodynamics, statistical mechanics, and information theory. Early life Avery was born in Lebanon to American parents. Avery's parents were both born in the United States, in the state of Michigan, where they studied at the University of Michigan. His father studied medicine while his mother studied bacteriology. After graduation, his parents did research together at the Marine Biological Laboratory in Woods Hole, Massachusetts. Later, his father did research in a borderline area between physics and medicine with Arthur Holly Compton, discoverer of the "Compton effect", at the University of Chicago. In 1926, his father moved the family to Beirut, where his father worked as a professor of anatomy at the American University of Beirut. The family stayed in Beirut until the start of World War II. It was during these tumultuous years that John Scales Avery was born. Education 1950 - graduated, Phillips Academy, Andover, Mass 1954 - B.Sc., physics, M.I.T 1955 - M.Sc., physics, University of Chicago 1965 - Ph.D., theoretical chemistry, Imperial College, London Thermodynamics In his recent 2003 book Information Theory and Evolution (2nd Edition, 2012), Avery combines information theory with thermodynamics to account for the phenomenon of life, including its o
https://en.wikipedia.org/wiki/Z-matrix%20%28mathematics%29
In mathematics, the class of Z-matrices are those matrices whose off-diagonal entries are less than or equal to zero; that is, the matrices of the form: Note that this definition coincides precisely with that of a negated Metzler matrix or quasipositive matrix, thus the term quasinegative matrix appears from time to time in the literature, though this is rare and usually only in contexts where references to quasipositive matrices are made. The Jacobian of a competitive dynamical system is a Z-matrix by definition. Likewise, if the Jacobian of a cooperative dynamical system is J, then (−J) is a Z-matrix. Related classes are L-matrices, M-matrices, P-matrices, Hurwitz matrices and Metzler matrices. L-matrices have the additional property that all diagonal entries are greater than zero. M-matrices have several equivalent definitions, one of which is as follows: a Z-matrix is an M-matrix if it is nonsingular and its inverse is nonnegative. All matrices that are both Z-matrices and P-matrices are nonsingular M-matrices. In the context of quantum complexity theory, these are referred to as stoquastic operators. See also Hurwitz matrix M-matrix Metzler matrix P-matrix
https://en.wikipedia.org/wiki/Matrix%20equivalence
In linear algebra, two rectangular m-by-n matrices A and B are called equivalent if for some invertible n-by-n matrix P and some invertible m-by-m matrix Q. Equivalent matrices represent the same linear transformation V → W under two different choices of a pair of bases of V and W, with P and Q being the change of basis matrices in V and W respectively. The notion of equivalence should not be confused with that of similarity, which is only defined for square matrices, and is much more restrictive (similar matrices are certainly equivalent, but equivalent square matrices need not be similar). That notion corresponds to matrices representing the same endomorphism V → V under two different choices of a single basis of V, used both for initial vectors and their images. Properties Matrix equivalence is an equivalence relation on the space of rectangular matrices. For two rectangular matrices of the same size, their equivalence can also be characterized by the following conditions The matrices can be transformed into one another by a combination of elementary row and column operations. Two matrices are equivalent if and only if they have the same rank. Canonical form The rank property yields an intuitive canonical form for matrices of the equivalence class of rank as , where the number of s on the diagonal is equal to . This is a special case of the Smith normal form, which generalizes this concept on vector spaces to free modules over principal ideal domains. See also Row equivalence Matrix congruence Matrices Equivalence (mathematics)
https://en.wikipedia.org/wiki/Bonferroni%20correction
In statistics, the Bonferroni correction is a method to counteract the multiple comparisons problem. Background The method is named for its use of the Bonferroni inequalities. An extension of the method to confidence intervals was proposed by Olive Jean Dunn. Statistical hypothesis testing is based on rejecting the null hypothesis if the likelihood of the observed data under the null hypotheses is low. If multiple hypotheses are tested, the probability of observing a rare event increases, and therefore, the likelihood of incorrectly rejecting a null hypothesis (i.e., making a Type I error) increases. The Bonferroni correction compensates for that increase by testing each individual hypothesis at a significance level of , where is the desired overall alpha level and is the number of hypotheses. For example, if a trial is testing hypotheses with a desired , then the Bonferroni correction would test each individual hypothesis at . Likewise, when constructing multiple confidence intervals the same phenomenon appears. Definition Let be a family of hypotheses and their corresponding p-values. Let be the total number of null hypotheses, and let be the number of true null hypotheses (which is presumably unknown to the researcher). The family-wise error rate (FWER) is the probability of rejecting at least one true , that is, of making at least one type I error. The Bonferroni correction rejects the null hypothesis for each , thereby controlling the FWER at . Proof of this control follows from Boole's inequality, as follows: This control does not require any assumptions about dependence among the p-values or about how many of the null hypotheses are true. Extensions Generalization Rather than testing each hypothesis at the level, the hypotheses may be tested at any other combination of levels that add up to , provided that the level of each test is decided before looking at the data. For example, for two hypothesis tests, an overall of 0.05 could be maintai
https://en.wikipedia.org/wiki/Polarized%20membrane
A polarized membrane is a lipid membrane that has a positive electrical charge on one side and a negative charge on another side, which produces the resting potential in living cells. Whether or not a membrane is polarized is determined by the distribution of dissociable protons and permeant ions inside and outside the membrane that travel passively through ion channel or actively via ion pump, creating an action potential. See also Membrane transporter
https://en.wikipedia.org/wiki/Neddylation
Neddylation (also NEDDylation) is the process by which the ubiquitin-like protein NEDD8 is conjugated to its target proteins. This process is analogous to ubiquitination, although it relies on its own E1 and E2 enzymes. No NEDD8-specific E3 has yet been identified and it is possible that the Neddylation system relies on E3 ligases with dual specificity. NEDD8 NEDD8 (neural-precursor-cell-expressed developmentally down-regulated 8) is a protein involved in the regulation of cell growth, viability and development. Neddylation process NEDD8 links itself to a protein through an isopeptide linkage between its carboxy-terminal glycine and the lysine of the substrate. The neddylation of the substrate causes in a structural change, and there are three main biochemical effects that result. First, neddylation can cause a conformational change in the substrate which may restrict molecular movement and the positioning of different binding partners. Second, it can cause the target protein to become incompatible with other proteins that it usually binds with. For example, CAND1 does not bind to neddylated proteins. In addition, neddylation can recruit NEDD8-interacting proteins. When NEDD8 binds to the ubiquitin E2 Ubc4, the interaction stimulates cullin-based ubiquitin ligases, although the exact mechanism is unclear. Disease association Neddylation is involved in the pathogenesis of Alzheimer's disease where its activation appears to drive neurons into apoptosis by initiating cell cycle reentry. Also, evidence shows that increased NEDD8 conjugation in human oral carcinoma cells led to abnormal higher degrees of proliferation. Because NEDD8 conjugation to cullin proteins plays an important role in the regulation of the cell cycle, an upregulation in conjugation causes this proliferation.
https://en.wikipedia.org/wiki/WSMO
WSMO or Web Service Modeling Ontology is a conceptual model for relevant aspects related to Semantic Web Services. It provides an ontology based framework, which supports the deployment and interoperability of Semantic Web Services. The WSMO has four main components: Goals - The client's objectives when consulting a Web Service. Ontologies - A formal Semantic description of the information used by all other components. Mediators - Connectors between componentes with mediation facilities. Provides interoperability between different ontologies. WebServices - Semantic description of Web Services. May include functional (Capability) and usage (Interface) descriptions. The WSMO working group, part of the ESSI Cluster aligns the research and development efforts in the areas of Semantic Web Services between several European FP6 research projects. WSMO working group includes the WSML working group, which aims at developing a language called Web Services Modeling Language (WSML) that formalizes the Web Services Modeling Ontology (WSMO). External links WSMO Working Group Web Site WSMO tools Web services Semantic Web
https://en.wikipedia.org/wiki/Global%20Ocean%20Ecosystem%20Dynamics
Global Ocean Ecosystem Dynamics (GLOBEC) is the International Geosphere-Biosphere Programme (IGBP) core project responsible for understanding how global change will affect the abundance, diversity and productivity of marine populations. The programme was initiated by SCOR and the IOC of UNESCO in 1991, to understand how global change will affect the abundance, diversity and productivity of marine populations comprising a major component of oceanic ecosystems. The aim of GLOBEC is to advance our understanding of the structure and functioning of the global ocean ecosystem, its major subsystems, and its response to physical forcing so that a capability can be developed to forecast the responses of the marine ecosystem to global change. Structure GLOBEC encompasses an integrated suite of research activities consisting of Regional Programmes, National Activities and cross-cutting research focal activities. The GLOBEC programme has been developed by the Scientific Steering Committee (SSC) and is co-ordinated through the GLOBEC International Project Office (IPO). Regional Programmes: Ecosystem Structure of Subarctic Seas (ESSAS) CLimate Impacts on Oceanic TOp Predators (CLIOTOP) ICES Cod and Climate Change (CCC) PICES Climate Change and Carrying Capacity (CCCC) Southern Ocean GLOBEC (SO GLOBEC) Small Pelagic Fish and Climate Change (SPACC) National Programmes: GLOBEC has several active national programmes and scientists from nearly 30 countries participate in GLOBEC activities on a national or regional level. Focus Working Groups: There are four GLOBEC cross-cutting research focal activities: Focus 1. Retrospective analysis Focus 2. Process studies Focus 3. Prediction and modelling Focus 4. Feedback from ecosystem changes Publications GLOBEC produces a report series, special contributions series and a biannual newsletter, all of which can be downloaded from the GLOBEC website. GLOBEC science has contributed to over 2000 refereed scientific publications which can
https://en.wikipedia.org/wiki/Epiploic%20appendagitis
Epiploic appendagitis (EA) is an uncommon, benign, self-limiting inflammatory process of the epiploic appendices. Other, older terms for the process include appendicitis epiploica and appendagitis, but these terms are used less now in order to avoid confusion with acute appendicitis. Epiploic appendices are small, fat-filled sacs or finger-like projections along the surface of the upper and lower colon and rectum. They may become acutely inflamed as a result of torsion (twisting) or venous thrombosis. The inflammation causes pain, often described as sharp or stabbing, located on the left, right, or central regions of the abdomen. There is sometimes nausea and vomiting. The symptoms may mimic those of acute appendicitis, diverticulitis, or cholecystitis. The pain is characteristically intense during/after defecation or micturition (espec. in the sigmoid type) due to the effect of traction on the pedicle of the lesion caused by straining and emptying of the bowel and bladder. Initial lab studies are usually normal. EA is usually diagnosed incidentally on CT scan which is performed to exclude more serious conditions. Although it is self-limiting, epiploic appendagitis can cause severe pain and discomfort. It is usually thought to be best treated with an anti-inflammatory and a moderate to severe pain medication (depending on the case) as needed. Surgery is not recommended in nearly all cases. Sand and colleagues, however, recommend laparoscopic surgery to excise the inflamed appendage in most cases in order to prevent recurrence. Signs and symptoms The condition commonly occurs in patients in their 40s and 50s predominantly in men. Epiploic appendagitis is normally misdiagnosed in most patients. Epiploic appendagitis presents with an acute onset of pain, commonly in the left lower quadrant the symptoms often lead to a misdiagnosis for diverticulitis. Diverticulitis manifests with evenly distributed lower abdominal pain accompanied with nausea, fever, and leukocy
https://en.wikipedia.org/wiki/Aminoallyl%20nucleotide
Aminoallyl nucleotide is a nucleotide with a modified base containing an allylamine. They are used in post-labeling of nucleic acids by fluorescence detection in microarray. They are reactive with N-Hydroxysuccinimide ester group which helps attach a fluorescent dye to the primary amino group on the nucleotide. These nucleotides are known as 5-(3-aminoallyl)-nucleotides since the aminoallyl group is usually attached to carbon 5 of the pyrimidine ring of uracil or cytosine. The primary amine group in the aminoallyl moiety is aliphatic and thus more reactive compared to the amine groups that are directly attached to the rings (aromatic) of the bases. Common names of aminoallyl nucleosides are initially abbreviated with aa- or AA- to indicate aminoallyl. The 5-carbon sugar is indicated with or without the lowercase "d" indicating deoxyribose if included or ribose if not. Finally the nitrogenous base and number of phosphates are indicated (i.e. aa-UTP = aminoallyl uridine triphosphate). History The goal of combining fluorescence and nucleic acids has been to provide a non-isotopic tag that is detectable to study DNA or RNA. This type of labeling allows scientists to study DNA or RNA in their structure, function, or formation with other nucleic acids. The first base modification for fluorescent labeling occurred in 1971 with a 4-thiouridine and 4-thiouracil. This research along with others, which included various types of direct and non-direct labeling via: analogs, addition via enzymes, or other methods made labeling of nucleotides much safer for scientist to study DNA. As instrumentation and technologies become more advanced in the field of DNA microarray, better reagents and techniques will be needed to further scientific studies. Fluorescent labeling with Cy3 was shown to be more insufficient and skew results; the method of aminoallyl nucleotide incorporation was opted instead. Using aminoallyl nucleotides as indirect fluorescent labeling seemed to nullify the sen
https://en.wikipedia.org/wiki/Kugelblitz%20%28astrophysics%29
A kugelblitz is a theoretical astrophysical object predicted by general relativity. It is a concentration of heat, light or radiation so intense that its energy forms an event horizon and becomes self-trapped. In other words, if enough radiation is aimed into a region of space, the concentration of energy can warp spacetime so much that it creates a black hole. This would be a black hole whose original mass–energy was in the form of radiant energy rather than matter, however as soon as it forms, it is indistinguishable from an ordinary black hole. John Archibald Wheeler's 1955 Physical Review paper entitled "Geons" refers to the kugelblitz phenomenon and explores the idea of creating such particles (or toy models of particles) from spacetime curvature. The kugelblitz phenomenon has been considered a possible basis for interstellar engines (drives) for future black hole starships. See also Bekenstein bound Micro black hole
https://en.wikipedia.org/wiki/Bodies%3A%20The%20Exhibition
Bodies: The Exhibition is an exhibition showcasing human bodies that have been preserved through a process called plastination and dissected to display bodily systems. It opened in Tampa, Florida on August 20, 2005. It is similar to, though not affiliated with, the exhibition Body Worlds (which opened in 1995). The exhibit displays internal organs and organic systems, bodies staged in active poses, and fetuses in various stages of development. The show is operated by Premier Exhibitions which presents and promotes similar exhibits including "Bodies Revealed", and "Our Body: The Universe Within", and other entertainment exhibits. The Exhibition showcases 13 whole-body specimens and more than 260 organs and partial body specimens. These real human bodies have been meticulously dissected, preserved through an innovative process. Concerns have been raised by human rights advocates that the bodies are those of executed Chinese prisoners, and that the families of the victims have not consented. The exhibition has claimed that the presumed origin of the bodies and fetuses "relies solely on the representations of its Chinese partners" and that they "cannot independently verify" that the bodies do not belong to executed prisoners. Catholics also had high levels of concern about allowing children to view displays containing human remains. Exhibit organization The exhibit is set up so that one starts at the skeletal system, and more layers (muscular, nervous, circulatory, digestive, respiratory, urinary, and reproductive systems; as well as fetal development and the treated body) are added in successive rooms. Containing about twenty bodies in total, each exhibition uses real human bodies that have been preserved permanently by a process called "polymer preservation" (commonly referred to as "plastination") so that they will not decay. This exhibition is organized by the publicly traded corporation, Premier Exhibitions Incorporated, which also staged Bodies Revealed first i
https://en.wikipedia.org/wiki/Digital%20television%20adapter
A digital television adapter (DTA), commonly known as a converter box or decoder box, is a television tuner that receives a digital television (DTV) transmission, and converts the digital signal into an analog signal that can be received and displayed on an analog television set. Some also have an HDMI output since some TVs with HDMI do not have a digital tuner. The input digital signal may be over-the-air terrestrial television signals received by a television antenna, or signals from a digital cable system. It normally does not refer to satellite TV, which has always required a set-top box either to operate the big satellite dish, or to be the integrated receiver/decoder (IRD) in the case of direct-broadcast satellites (DBS). In North America and South Korea, these ATSC tuner boxes convert from ATSC to NTSC, while in most of Europe and other places such as Australia and most Asian countries, they convert from Digital Video Broadcasting (DVB) to PAL, and in Japan, the Philippines and almost all countries in Latin America, they convert from ISDB-T to either NTSC or PAL. Because the DTV transition did nothing to reduce the number of broadcast television system standards (and in fact further complicated them), and due to varying frequency allocations and bandplans, there are many other combinations specific to other countries. United States On June 12, 2009, all full-power analog television transmissions ended in the United States. Viewers who watch broadcast television on older analog TV sets must use a digital converter box. Since many of the low-power TV stations continued to broadcast in analog for a while, consumers who watch low-power stations needed an adapter with an analog passthrough feature that allows the viewer to watch both digital and analog signals. Viewers who receive their television signals through cable or satellite were not affected by this change and did not need a digital television adapter (however, see the cable TV exception below). A
https://en.wikipedia.org/wiki/Symmetric%20level-index%20arithmetic
The level-index (LI) representation of numbers, and its algorithms for arithmetic operations, were introduced by Charles Clenshaw and Frank Olver in 1984. The symmetric form of the LI system and its arithmetic operations were presented by Clenshaw and Peter Turner in 1987. Michael Anuta, Daniel Lozier, Nicolas Schabanel and Turner developed the algorithm for symmetric level-index (SLI) arithmetic, and a parallel implementation of it. There has been extensive work on developing the SLI arithmetic algorithms and extending them to complex and vector arithmetic operations. Definition The idea of the level-index system is to represent a non-negative real number as where and the process of exponentiation is performed times, with . and are the level and index of respectively. is the LI image of . For example, so its LI image is The symmetric form is used to allow negative exponents, if the magnitude of is less than 1. One takes or and stores it (after substituting +1 for 0 for the reciprocal sign since for the LI image is and uniquely defines and we can do away without a third state and use only one bit for the two states −1 and +1) as the reciprocal sign . Mathematically, this is equivalent to taking the reciprocal (multiplicative inverse) of a small magnitude number, and then finding the SLI image for the reciprocal. Using one bit for the reciprocal sign enables the representation of extremely small numbers. A sign bit may also be used to allow negative numbers. One takes sgn(X) and stores it (after substituting +1 for 0 for the sign since for the LI image is and uniquely defines and we can do away without a third state and use only one bit for the two states −1 and +1) as the sign . Mathematically, this is equivalent to taking the inverse (additive inverse) of a negative number, and then finding the SLI image for the inverse. Using one bit for the sign enables the representation of negative numbers. The mapping function is called the
https://en.wikipedia.org/wiki/Apartness%20relation
In constructive mathematics, an apartness relation is a constructive form of inequality, and is often taken to be more basic than equality. It is often written as (⧣ in unicode) to distinguish from the negation of equality (the denial inequality) which is weaker. Description An apartness relation is a symmetric irreflexive binary relation with the additional condition that if two elements are apart, then any other element is apart from at least one of them (this last property is often called co-transitivity or comparison). That is, a binary relation is an apartness relation if it satisfies: The complement of an apartness relation is an equivalence relation, as the above three conditions become reflexivity, symmetry, and transitivity. If this equivalence relation is in fact equality, then the apartness relation is called tight. That is, is a if it additionally satisfies: 4. In classical mathematics, it also follows that every apartness relation is the complement of an equivalence relation, and the only tight apartness relation on a given set is the complement of equality. So in that domain, the concept is not useful. In constructive mathematics, however, this is not the case. The prototypical apartness relation is that of the real numbers: two real numbers are said to be apart if there exists (one can construct) a rational number between them. In other words, real numbers and are apart if there exists a rational number such that or The natural apartness relation of the real numbers is then the disjunction of its natural pseudo-order. The complex numbers, real vector spaces, and indeed any metric space then naturally inherit the apartness relation of the real numbers, even though they do not come equipped with any natural ordering. If there is no rational number between two real numbers, then the two real numbers are equal. Classically, then, if two real numbers are not equal, one would conclude that there exists a rational number between them. H
https://en.wikipedia.org/wiki/Center-of-momentum%20frame
In physics, the center-of-momentum frame (COM frame), also known as zero-momentum frame, is the inertial frame in which the total momentum of the system vanishes. It is unique up to velocity, but not origin. The center of momentum of a system is not a location, but a collection of relative momenta/velocities: a reference frame. Thus "center of momentum" is a short for "center-of-momentum ". A special case of the center-of-momentum frame is the center-of-mass frame: an inertial frame in which the center of mass (which is a single point) remains at the origin. In all center-of-momentum frames, the center of mass is at rest, but it is not necessarily at the origin of the coordinate system. In special relativity, the COM frame is necessarily unique only when the system is isolated. Properties General The center of momentum frame is defined as the inertial frame in which the sum of the linear momenta of all particles is equal to 0. Let S denote the laboratory reference system and S′ denote the center-of-momentum reference frame. Using a Galilean transformation, the particle velocity in S′ is where is the velocity of the mass center. The total momentum in the center-of-momentum system then vanishes: Also, the total energy of the system is the minimal energy as seen from all inertial reference frames. Special relativity In relativity, the COM frame exists for an isolated massive system. This is a consequence of Noether's theorem. In the COM frame the total energy of the system is the rest energy, and this quantity (when divided by the factor c2, where c is the speed of light) gives the rest mass (invariant mass) of the system: The invariant mass of the system is given in any inertial frame by the relativistic invariant relation but for zero momentum the momentum term (p/c)2 vanishes and thus the total energy coincides with the rest energy. Systems that have nonzero energy but zero rest mass (such as photons moving in a single direction, or, equivalently, pla
https://en.wikipedia.org/wiki/Hurst%20exponent
The Hurst exponent is used as a measure of long-term memory of time series. It relates to the autocorrelations of the time series, and the rate at which these decrease as the lag between pairs of values increases. Studies involving the Hurst exponent were originally developed in hydrology for the practical matter of determining optimum dam sizing for the Nile river's volatile rain and drought conditions that had been observed over a long period of time. The name "Hurst exponent", or "Hurst coefficient", derives from Harold Edwin Hurst (1880–1978), who was the lead researcher in these studies; the use of the standard notation H for the coefficient also relates to his name. In fractal geometry, the generalized Hurst exponent has been denoted by H or Hq in honor of both Harold Edwin Hurst and Ludwig Otto Hölder (1859–1937) by Benoît Mandelbrot (1924–2010). H is directly related to fractal dimension, D, and is a measure of a data series' "mild" or "wild" randomness. The Hurst exponent is referred to as the "index of dependence" or "index of long-range dependence". It quantifies the relative tendency of a time series either to regress strongly to the mean or to cluster in a direction. A value H in the range 0.5–1 indicates a time series with long-term positive autocorrelation, meaning that the decay in autocorrelation is slower than exponential, following a power law; for the series it means that a high value tends to be followed by another high value and that future excursions to more high values do occur. A value in the range 0 – 0.5 indicates a time series with long-term switching between high and low values in adjacent pairs, meaning that a single high value will probably be followed by a low value and that the value after that will tend to be high, with this tendency to switch between high and low values lasting a long time into the future, also following a power law. A value of H=0.5 indicates short-memory, with (absolute) autocorrelations decaying exponentially
https://en.wikipedia.org/wiki/Developmental%20coordination%20disorder
Developmental coordination disorder (DCD), also known as developmental motor coordination disorder, developmental dyspraxia or simply dyspraxia (from Ancient Greek praxis 'activity'), is a neurodevelopmental disorder characterized by impaired coordination of physical movements as a result of brain messages not being accurately transmitted to the body. Deficits in fine or gross motor skills movements interfere with activities of daily living. It is often described as disorder in skill acquisition, where the learning and execution of coordinated motor skills is substantially below that expected given the individual's chronological age. Difficulties may present as clumsiness, slowness and inaccuracy of performance of motor skills (e.g., catching objects, using cutlery, handwriting, riding a bike, use of tools or participating in team sports or swimming). It is also often accompanied by difficulty with organisation and/or problems with attention, working memory and time management. A diagnosis of DCD is reached only in the absence of other neurological impairments such as cerebral palsy, multiple sclerosis, or Parkinson's disease. The condition is lifelong and its onset is in early childhood. It is thought to affect about 5% of the population. Signs and symptoms The World Health Organisation (WHO) recognise DCD as a condition, and have published their definition in the International Classification of Diseases. This describes DCD as: The American Psychiatric Association (APA)'s Diagnostic and Statistical Manual, DSM-5 classifies Developmental Coordination Disorder (DCD) as a discrete motor disorder under the broader heading of neurodevelopmental disorders. It is often described as a disorder in skill acquisition or motor learning, where the learning and execution of coordinated motor skills is substantially below that expected given the individual's chronological age. Various areas of development can be affected by DCD and these may persist into adulthood. In chil
https://en.wikipedia.org/wiki/Controlled-environment%20agriculture
Controlled-environment agriculture (CEA) -- which includes indoor agriculture (IA) and vertical farming—is a technology-based approach toward food production. The aim of CEA is to provide protection from the outdoor elements and maintain optimal growing conditions throughout the development of the crop. Production takes place within an enclosed growing structure such as a greenhouse or plant factory. Plants are often grown in a soilless medium in order to supply the proper amounts of water and nutrients to the root zone as well as supplemental lighting to ensure a sufficient daily light integral. CEA optimizes the use of resources such as water, energy, space, capital and labor. CEA technologies include hydroponics, aeroponics, aquaculture, and aquaponics. Different techniques are available for growing food in controlled environment agriculture. Currently, the greenhouse industry is the largest component of the CEA industry but another quickly growing segment is the vertical farming industry. Controlled Environment Agriculture has the ability to produce crops all year round, with the possibility of increased yield by adjusting the amount of carbon and nutrients the plants receive (Benke et al). In consideration to urban agriculture, CEA can exist inside repurposed structures, built to purpose structures or in basements and subterranean spaces. The trend is increasingly growing into alternative food networks, as entrepreneurs and households seek to meet the growing demand for fresh food products. Technical implementation Controllable variables: Environmental: Temperature (air, nutrient solution, root-zone, leaf) Relative Humidity (%RH) Carbon dioxide (CO2) Light (intensity, spectrum, duration and intervals) Cultural: Water Quality Nutrient concentration (PPM of Nitrogen, Potassium, Phosphorus, etc) Nutrient pH (acidity) Cropping duration and density Cultivar Pest controls CEA facilities can range from fully 100% environmentally controlled enclosed c
https://en.wikipedia.org/wiki/Quantum%20mind
The quantum mind or quantum consciousness is a group of hypotheses proposing that classical mechanics alone cannot explain consciousness, positing instead that quantum-mechanical phenomena, such as entanglement and superposition, may play an important part in the brain's function and could explain critical aspects of consciousness. These scientific hypotheses are as yet untested, and can overlap with quantum mysticism. History Eugene Wigner developed the idea that quantum mechanics has something to do with the workings of the mind. He proposed that the wave function collapses due to its interaction with consciousness. Freeman Dyson argued that "mind, as manifested by the capacity to make choices, is to some extent inherent in every electron". Other contemporary physicists and philosophers considered these arguments unconvincing. Victor Stenger characterized quantum consciousness as a "myth" having "no scientific basis" that "should take its place along with gods, unicorns and dragons". David Chalmers argues against quantum consciousness. He instead discusses how quantum mechanics may relate to dualistic consciousness. Chalmers is skeptical that any new physics can resolve the hard problem of consciousness. He argues that quantum theories of consciousness suffer from the same weakness as more conventional theories. Just as he argues that there is no particular reason why particular macroscopic physical features in the brain should give rise to consciousness, he also thinks that there is no particular reason why a particular quantum feature, such as the EM field in the brain, should give rise to consciousness either. Approaches Bohm David Bohm viewed quantum theory and relativity as contradictory, which implied a more fundamental level in the universe. He claimed that both quantum theory and relativity pointed to this deeper theory, which he formulated as a quantum field theory. This more fundamental level was proposed to represent an undivided wholeness and a
https://en.wikipedia.org/wiki/Blocking%20antibody
A blocking antibody is an antibody that does not have a reaction when combined with an antigen, but prevents other antibodies from combining with that antigen. This function of blocking antibodies has had a variety of clinical and experimental uses. The term can also be used for inhibiting antibody, prozone phenomenon and, agglutination reaction. Blocking antibodies have been described as a mechanism for HSV-1 to evade the immune system. Uses Blocking antibodies can be used in a variety of medical and scientific manners, thus far been to treat cancer, Graves' disease, and prevent the growth of malaria in mosquitoes. Cancer treatment Blocking antibodies have been used in clinical trials of cancer treatments. The blocking antibody ipilimumab has been effectively used in the clinical treatment of melanoma, RCC, and NSCLC with some degree of success. This is accomplished through the blocking of the coinhibitory molecule CTLA-4. The blocking antibody does not directly target tumor cells, but rather blocks the regulatory functions of CTLA-4, resulting in enhanced T-cell function. Some new treatments hypothesize the blocking of PD-1, a programmed cell-death protein, which will result in longer-lived T-cells. The blocking antibody BMS-936559 has been shown to bind to PD-L1 and prevent its binding to PD-1. These new treatments are not without side-effects and immune-related adverse events have been observed in a variety of patients. The tolerance that immune cells normally have to host tissues can be lost, resulting in permanent damage to host cells. Graves' disease Studies have shown that blocking antibodies can bind to and prevent thyrotropin binding, resulting in reduced cAMP levels in human thyroid cells. This interaction has been used primarily as a method of indicating that Graves' disease immunoglobulins are pluritopic, meaning that they have multiple effects, rather than indicating a possible treatment for this disease. Malaria Blocking antibodies have a
https://en.wikipedia.org/wiki/Single-strand%20conformation%20polymorphism
Single-strand conformation polymorphism (SSCP), or single-strand chain polymorphism, is defined as a conformational difference of single-stranded nucleotide sequences of identical length as induced by differences in the sequences under certain experimental conditions. This property allows sequences to be distinguished by means of gel electrophoresis, which separates fragments according to their different conformations. Physical background A single nucleotide change in a particular sequence, as seen in a double-stranded DNA, cannot be distinguished by gel electrophoresis techniques, which can be attributed to the fact that; the physical properties of the double strands are almost identical for both alleles. After denaturation, single-stranded DNA undergoes a characteristic 3-dimensional folding and may assume a unique conformational state based on its DNA sequence. The difference in shape between two single-stranded DNA strands with different sequences can cause them to migrate differently through an electrophoresis gel, even though the number of nucleotides is the same, which is, in fact, an application of SSCP. Applications in molecular biology SSCP used to be a way to discover new DNA polymorphisms apart from DNA sequencing but is now being supplanted by sequencing techniques on account of efficiency and accuracy. These days, SSCP is most applicable as a diagnostic tool in molecular biology. It can be used in genotyping to detect homozygous individuals of different allelic states, as well as heterozygous individuals that should each demonstrate distinct patterns in an electrophoresis experiment. SSCP is also widely used in virology to detect variations in different strains of a virus, the idea being that a particular virus particle present in both strains will have undergone changes due to mutation, and that these changes will cause the two particles to assume different conformations and, thus, be differentiable on an SSCP gel.
https://en.wikipedia.org/wiki/One%20thousand%20origami%20cranes
The crane is considered a mystical or holy creature (others include the dragon and the tortoise) in Japan and is said to live for a thousand years. That is why are made, one for each year. In some stories, it is believed that the cranes must be completed within one year and they must all be made by the person (or group of people) who will make the wish at the end. Cultural significance In Japan, cranes have been thought a symbol of long life. An old phrase says "cranes live a thousand years". Here "a thousand" is not necessarily to designate the exact number, but a poetic expression of huge amounts. Historically well-wishers offered a picture of a crane to shrines and temples as well as paper cranes. Origami, specially crafted and patterned paper, was invented in Edo period. In the late 17th century books referring not only to "paper cranes" but also to "one thousand cranes" were published. In modern times, cranes are often given to a person who is seriously ill, to wish for their recovery. They are usually created by friends, classmates, or colleagues as a collective effort, offered to a shrine on the person's behalf or directly gifted to. Another common use is for sports teams or athletes, wishing them victories. Cranes are also a symbol of peace, and are thus often seen at war memorials along with its original meaning for wishing good health. Several temples, including some in Tokyo and Hiroshima, have eternal flames for world peace. At these temples, school groups or individuals often donate senbazuru to add to the prayer for peace. The cranes are left exposed to the elements, slowly becoming tattered and dissolving as symbolically, the wish is released. In this way, they are related to the prayer flags of India and Tibet. The Japanese space agency JAXA used the folding of one thousand cranes as one of the tests for candidates of its astronaut program. Sadako Sasaki The one thousand origami cranes were globally popularized through the story of Sadako Sas
https://en.wikipedia.org/wiki/Weyl%27s%20lemma%20%28Laplace%20equation%29
In mathematics, Weyl's lemma, named after Hermann Weyl, states that every weak solution of Laplace's equation is a smooth solution. This contrasts with the wave equation, for example, which has weak solutions that are not smooth solutions. Weyl's lemma is a special case of elliptic or hypoelliptic regularity. Statement of the lemma Let be an open subset of -dimensional Euclidean space , and let denote the usual Laplace operator. Weyl's lemma states that if a locally integrable function is a weak solution of Laplace's equation, in the sense that for every smooth test function with compact support, then (up to redefinition on a set of measure zero) is smooth and satisfies pointwise in . This result implies the interior regularity of harmonic functions in , but it does not say anything about their regularity on the boundary . Idea of the proof To prove Weyl's lemma, one convolves the function with an appropriate mollifier and shows that the mollification satisfies Laplace's equation, which implies that has the mean value property. Taking the limit as and using the properties of mollifiers, one finds that also has the mean value property, which implies that it is a smooth solution of Laplace's equation. Alternative proofs use the smoothness of the fundamental solution of the Laplacian or suitable a priori elliptic estimates. Generalization to distributions More generally, the same result holds for every distributional solution of Laplace's equation: If satisfies for every , then is a regular distribution associated with a smooth solution of Laplace's equation. Connection with hypoellipticity Weyl's lemma follows from more general results concerning the regularity properties of elliptic or hypoelliptic operators. A linear partial differential operator with smooth coefficients is hypoelliptic if the singular support of is equal to the singular support of for every distribution . The Laplace operator is hypoelliptic, so if , then the sing
https://en.wikipedia.org/wiki/Polynomial%20matrix
In mathematics, a polynomial matrix or matrix of polynomials is a matrix whose elements are univariate or multivariate polynomials. Equivalently, a polynomial matrix is a polynomial whose coefficients are matrices. A univariate polynomial matrix P of degree p is defined as: where denotes a matrix of constant coefficients, and is non-zero. An example 3×3 polynomial matrix, degree 2: We can express this by saying that for a ring R, the rings and are isomorphic. Properties A polynomial matrix over a field with determinant equal to a non-zero element of that field is called unimodular, and has an inverse that is also a polynomial matrix. Note that the only scalar unimodular polynomials are polynomials of degree 0 – nonzero constants, because an inverse of an arbitrary polynomial of higher degree is a rational function. The roots of a polynomial matrix over the complex numbers are the points in the complex plane where the matrix loses rank. The determinant of a matrix polynomial with Hermitian positive-definite (semidefinite) coefficients is a polynomial with positive (nonnegative) coefficients. Note that polynomial matrices are not to be confused with monomial matrices, which are simply matrices with exactly one non-zero entry in each row and column. If by λ we denote any element of the field over which we constructed the matrix, by I the identity matrix, and we let A be a polynomial matrix, then the matrix λI − A is the characteristic matrix of the matrix A. Its determinant, |λI − A| is the characteristic polynomial of the matrix A.
https://en.wikipedia.org/wiki/Hamiltonian%20matrix
In mathematics, a Hamiltonian matrix is a -by- matrix such that is symmetric, where is the skew-symmetric matrix and is the -by- identity matrix. In other words, is Hamiltonian if and only if where denotes the transpose. Properties Suppose that the -by- matrix is written as the block matrix where , , , and are -by- matrices. Then the condition that be Hamiltonian is equivalent to requiring that the matrices and are symmetric, and that . Another equivalent condition is that is of the form with symmetric. It follows easily from the definition that the transpose of a Hamiltonian matrix is Hamiltonian. Furthermore, the sum (and any linear combination) of two Hamiltonian matrices is again Hamiltonian, as is their commutator. It follows that the space of all Hamiltonian matrices is a Lie algebra, denoted . The dimension of is . The corresponding Lie group is the symplectic group . This group consists of the symplectic matrices, those matrices which satisfy . Thus, the matrix exponential of a Hamiltonian matrix is symplectic. However the logarithm of a symplectic matrix is not necessarily Hamiltonian because the exponential map from the Lie algebra to the group is not surjective. The characteristic polynomial of a real Hamiltonian matrix is even. Thus, if a Hamiltonian matrix has as an eigenvalue, then , and are also eigenvalues. It follows that the trace of a Hamiltonian matrix is zero. The square of a Hamiltonian matrix is skew-Hamiltonian (a matrix is skew-Hamiltonian if ). Conversely, every skew-Hamiltonian matrix arises as the square of a Hamiltonian matrix. Extension to complex matrices As for symplectic matrices, the definition for Hamiltonian matrices can be extended to complex matrices in two ways. One possibility is to say that a matrix is Hamiltonian if , as above. Another possibility is to use the condition where the superscript asterisk () denotes the conjugate transpose. Hamiltonian operators Let be a vector space, equipped wi
https://en.wikipedia.org/wiki/Cylindric%20algebra
In mathematics, the notion of cylindric algebra, invented by Alfred Tarski, arises naturally in the algebraization of first-order logic with equality. This is comparable to the role Boolean algebras play for propositional logic. Cylindric algebras are Boolean algebras equipped with additional cylindrification operations that model quantification and equality. They differ from polyadic algebras in that the latter do not model equality. Definition of a cylindric algebra A cylindric algebra of dimension (where is any ordinal number) is an algebraic structure such that is a Boolean algebra, a unary operator on for every (called a cylindrification), and a distinguished element of for every and (called a diagonal), such that the following hold: (C1) (C2) (C3) (C4) (C5) (C6) If , then (C7) If , then Assuming a presentation of first-order logic without function symbols, the operator models existential quantification over variable in formula while the operator models the equality of variables and . Hence, reformulated using standard logical notations, the axioms read as (C1) (C2) (C3) (C4) (C5) (C6) If is a variable different from both and , then (C7) If and are different variables, then Cylindric set algebras A cylindric set algebra of dimension is an algebraic structure such that is a field of sets, is given by , and is given by . It necessarily validates the axioms C1–C7 of a cylindric algebra, with instead of , instead of , set complement for complement, empty set as 0, as the unit, and instead of . The set X is called the base. A representation of a cylindric algebra is an isomorphism from that algebra to a cylindric set algebra. Not every cylindric algebra has a representation as a cylindric set algebra. It is easier to connect the semantics of first-order predicate logic with cylindric set algebra. (For more details, see .) Generalizations Cylindric algebras have been genera
https://en.wikipedia.org/wiki/WinBuilder
WinBuilder is a free application designed to build and customize boot disks (Live CDs) based on Microsoft Windows (WinPE). Scripts Scripts are the building blocks of Live CD compiled with Winbuilder, and are used to build the core components of the Live CD for adding applications, and for configuration of a specific project. Scripts consist of a human-readable text file with the extension, and can contain Winbuilder scripting commands, the scripts graphical user interface, and sometimes encoded applications to be extracted when needed. Projects A Project is a collection of Winbuilder scripts maintained for building a complete Live CD. Each project contains multiple scripts, each responsible for adding features or applications to the build. Users may build their own projects from scratch or use one of many projects actively developed by the boot-land community. Projects can be downloaded directly using Winbuilder's built-in download manager and used as-is or further customized to meet the individual's needs. Actively developed projects Gena - The Gena project provides a basis for building a PE environment using Windows XP sources and is the successor to LiveXP nativeEx_multiPE - Live CD/DVD / Live USB that can be built from either Windows XP, Windows 2003, Windows Vista, or Windows 7 x86 sources. Win7PE SE - The Win7PE SE project provides a basis for building a PE environment using Windows 7 or Windows Vista sources. A fork of Multi 7PEs. Win8PE SE - The Win8PE SE project provides a basis for building a PE environment using a Windows 8 source and supports both x86 and x64 architectures. Win8.1PE SE - The Win8.1PE SE project provides a basis for building a PE environment using a Windows 8.1 source and supports both x86 and x64 architectures. Win10PE SE - Live CD/DVD that can be built from Windows 10 sources and supports both x86 and x64 architectures. Inactive Projects BartPEcore - running bartPE inside a Winbuilder environment BB-7PE - A Windows 7 bas
https://en.wikipedia.org/wiki/Mathieu%20transformation
The Mathieu transformations make up a subgroup of canonical transformations preserving the differential form The transformation is named after the French mathematician Émile Léonard Mathieu. Details In order to have this invariance, there should exist at least one relation between and only (without any involved). where . When a Mathieu transformation becomes a Lagrange point transformation. See also Canonical transformation
https://en.wikipedia.org/wiki/Social%20trap
In psychology, a social trap is a conflict of interest or perverse incentive where individuals or a group of people act to obtain short-term individual gains, which in the long run leads to a loss for the group as a whole. Social traps are the cause of countless environmental issues, including overfishing, energy "brownout" and "blackout" power outages during periods of extreme temperatures, the overgrazing of cattle on the Sahelian Desert, the destruction of the rainforest by logging interests and agriculture, and, most importantly, climate change. Origin of the concept The term social trap was first introduced to the scientific community by John Platt's 1973 paper in American Psychologist, and in a book developed in an interdisciplinary symposium held at the University of Michigan. Building upon the concept of the "tragedy of the commons" in Garrett Hardin's pivotal article in Science (1968), Platt and others in the seminar applied behavioral psychology concepts to actions of people operating in social traps. By applying the findings of basic research on "schedules of operant reinforcement" (B.F. Skinner 1938, 1948, 1953, 1957; Keller and Schoenfeld, 1950), Platt recognized that individuals operating for short-term positive gain ("reinforcement") had a tendency to over-exploit a resource, which led to a long-term overall loss to society. The application of behavioral psychology terms to behaviors in the tragedy of the commons led to the realization that the same short-term versus long-term cause-and-effect relationship also applied to other human traps, in addition to the exploitation of commonly held resources. Types of social traps Besides defining this psychological phenomenon, Platt also distinguishes between social traps and social fences (countertraps). Social traps represent a behavior or action which prioritizes individual gains at the expense of collective gains. Social fence refers to a short-term avoidance behavior by individuals that leads to a
https://en.wikipedia.org/wiki/Vexillography
Vexillography ( ) is the art and practice of designing flags; a person who designs flags is a vexillographer. Vexillography is allied with vexillology, the scholarly study of flags, but is not synonymous with that discipline. Background of flag design Flag designs exhibit a number of regularities, arising from a variety of practical concerns, historical circumstances, and cultural prescriptions that have shaped and continue to shape their evolution. Vexillographers face the necessity for the design to be manufactured (and often mass-produced) into or onto a piece of cloth, which will subsequently be hoisted aloft in the outdoors to represent an organization, individual, idea, or group. In this respect, flag design departs considerably from logo design: logos are predominantly still images suitable for reading off a page, screen, or billboard; while flags are alternately draped and fluttering images - visible from a variety of distances and angles (including the reverse). The prevalence of simple bold colors and shapes in flag design attests to these practical issues. Flag design has a history, and new designs often refer back to previous designs, effectively quoting, elaborating, or commenting upon them. Families of current flags may derive from a few common ancestors - as in the cases of the Pan-African colours, the Pan-Arab colors, the Pan-Slavic colors, the Nordic Cross flag and the Ottoman flag. Certain cultures prescribe the proper design of their own flags, through heraldic or other authoritative systems. Prescription may be based on religious principles: see, for example, Islamic flags. Vexillographers have begun to articulate design principles, such as those jointly published by the North American Vexillological Association and the Flag Institute in their Guiding Principles of Flag Design. Principles of design In 2006, the North American Vexillological Association published a booklet titled “Good” Flag, “Bad” Flag to aid those wishing to design or re-de
https://en.wikipedia.org/wiki/Universal%20Powerline%20Association
The Universal Powerline Association (UPA) was a trade association that covered power line communication (PLC) markets and applications. The UPA promoted and certified power line communication technology from 2004 to 2010. History An interest group for the UPA was established by the founding members in May 2004. A memorandum of understanding was signed by the founding members in September 2004, and the UPA was officially announced in January 2005. It was a non-profit trade group registered in Belgium. Design of Systems on Silicon (DS2), a supplier of integrated circuits and software for PLC, was among the founding members; UPA systems were based on its technology. Other members included: AcBel, Ambient Corporation, bpl, Corinex Communications, IBEC, Netgear, PCN Technology, Toshiba, Toyo Network Systems, and Junaid. The Universal Powerline Association released specifications related to three aspects of powerline technology. The UPA coexistence specification was published in June 2005. The UPA Access specification (the European OPERA project endorsed specification for Internet access over power lines) was published in February 2006. In-home systems and solutions, called "triple play" scenarios, were published in February 2006 as the UPA Digital Home Specification v1.0. The UPA worked with international standardisation bodies such as IEEE and ETSI. Certification events (known as Plugtests) were held in January 2006 and 2007, with products demonstrated at the Consumer Electronics Show. DS2 had financial problems in 2010. Marvell Technology Group acquired the intellectual property of DS2 in August 2011, although it continued to provide UPA specifications compliant devices. With the exit of DS2 from the market, the UPA suspended activities in November 2010 and discontinued its website, although the UPA market continued through Marvell and its partners.
https://en.wikipedia.org/wiki/Black%20cocaine
Black cocaine () is a mixture of regular cocaine base or cocaine hydrochloride with various other substances. These other substances are added to camouflage the typical appearance (pigments and dyes, e.g. charcoal), to interfere with color-based drug tests (mixing thiocyanates and iron salts or cobalt salts forms deep red complexes in solution), to make the mixture undetectable by drug sniffing dogs (activated carbon may sufficiently absorb trace odors). Since the result is usually black, it is generally smuggled as toner, fingerprint powder, fertilizer, pigment, metal moldings, or charcoal. The pure cocaine base can be recovered from the mixture by extraction (freebase) or acid-base extraction (hydrochloride) using common organic solvents such as methylene chloride or acetone. A second process is required to convert cocaine base into powdered cocaine hydrochloride. It was reported that in the mid-1980s Chilean dictator Augusto Pinochet ordered his army to build a clandestine cocaine laboratory in Chile where chemists mixed cocaine with other chemicals to produce what Pinochet's former top aide for intelligence Manuel Contreras described as a "black cocaine" capable of being smuggled past drug agents in the US and Europe. Black cocaine was detected in Bogota, Colombia in May 1998. In 2008, a new type of black cocaine was discovered by police in Spain. It had been manufactured into rubber-like sheets and made into luggage. In 2021, of black cocaine disguised as charcoal, in 30 sacks among 1,364 sacks of charcoal, were seized in Spain, one of the biggest cocaine seizures recorded in Castilla y León.
https://en.wikipedia.org/wiki/Tissue%20bank
A tissue bank is an establishment that collects and recovers human cadaver tissue for the purposes of medical research, education and allograft transplantation. A tissue bank may also refer to a location where biomedical tissue is stored under cryogenic conditions and is generally used in a more clinical sense. The United States Navy Tissue Bank is generally accepted as the first full tissue banking service of its kind in the world although it is not the largest or only tissue bank today. Donating a body Most medical schools need donated bodies for teaching students about the anatomy and physiology of the body, as well as how to perform medical procedures. There are no upper age limits for donating your body to science. Each school has different policies and procedures for donated bodies. Act 368 of the Public Acts of Michigan 1978, Article 10, Part 101, authorizes an individual to will their body to a medical institution. Medical institutions will only accept a full body, meaning the body cannot be used for organ donation prior to body donation because it would leave the body incomplete. At the time of death, the institution where the body is being donated to should be contacted as soon as possible. The decision to accept or reject the donation will be made at that time. A body could be turned away if it has already begun decomposing, the person was extremely obese, recent surgery was conducted, thoughts of possible contagious disease or severe trauma to the body occurred. There is no an age restriction as long as the donor or the donor’s legal representative gives consent. It is very uncommon for a school to turn down a body due to having an adequate supply; they are always in high demand. At Michigan State University people can choose whether to donate their body for three years or without any restriction as to the length of time. A body donated for an indefinite period can be used for educational purposes as long as it remains an effective teaching tool. At t
https://en.wikipedia.org/wiki/Phosphodiesterase%203
PDE3 is a phosphodiesterase. The PDEs belong to at least eleven related gene families, which are different in their primary structure, substrate affinity, responses to effectors, and regulation mechanism. Most of the PDE families are composed of more than one gene. PDE3 is clinically significant because of its role in regulating heart muscle, vascular smooth muscle and platelet aggregation. PDE3 inhibitors have been developed as pharmaceuticals, but their use is limited by arrhythmic effects and they can increase mortality in some applications. Function PDE3 enzymes are involved in regulation of cardiac and vascular smooth muscle contractility. Molecules that inhibit PDE3 were originally investigated for the treatment of heart failure, but, because of unwanted arrhythmic side-effects, they are not studied for that indication any longer. Nonetheless, the PDE3 inhibitor milrinone is approved for use in heart failure in intravenous form. Both PDE3A and PDE3B are expressed in vascular smooth muscle cells and are likely to modulate contraction. Their expression in vascular smooth muscle is altered under specific conditions such as elevated cAMP and hypoxia. Isoforms and genes The PDE3 family in mammals consists of two members, PDE3A and PDE3B. The PDE3 isoforms are structurally similar, containing an N-terminal domain important for the localization and a C-terminus end. The 44-amino acid insertion in the catalytic domain differs in the PDE3 isoforms, and the N-terminal portions of the isoforms are quite divergent. PDE3A and PDE3B have strikingly similar pharmacological and kinetic properties, but the distinction is in expression profiles and affinity for cGMP. The PDE3 family is composed of two genes, PDE3A and PDE3B. In cells expressing both genes, PDE3A is usually dominant. Three different variants of PDE3A (PDE3A1-3) are products of alternate startcodon usage of the PDE3A gene. The PDE3B encodes a single isoform only. In their full-length both PDE3A and PDE3B
https://en.wikipedia.org/wiki/BH3%20interacting-domain%20death%20agonist
The BH3 interacting-domain death agonist, or BID, gene is a pro-apoptotic member of the Bcl-2 protein family. Bcl-2 family members share one or more of the four characteristic domains of homology entitled the Bcl-2 homology (BH) domains (named BH1, BH2, BH3 and BH4), and can form hetero- or homodimers. Bcl-2 proteins act as anti- or pro-apoptotic regulators that are involved in a wide variety of cellular activities. Interactions BID is a pro-apoptotic Bcl-2 protein containing only the BH3 domain. In response to apoptotic signaling, BID interacts with another Bcl-2 family protein, Bax, leading to the insertion of Bax into organelle membranes, primarily the outer mitochondrial membrane. Bax is believed to interact with, and induce the opening of the mitochondrial voltage-dependent anion channel, VDAC. Alternatively, growing evidence suggest that activated Bax and/or Bak form an oligomeric pore, MAC in the outer membrane. This results in the release of cytochrome c and other pro-apoptotic factors (such as SMAC/DIABLO) from the mitochondria, often referred to as mitochondrial outer membrane permeabilization, leading to activation of caspases. This defines BID as a direct activator of Bax, a role common to some of the pro-apoptotic Bcl-2 proteins containing only the BH3 domain. The anti-apoptotic Bcl-2 proteins, including Bcl-2 itself, can bind BID and inhibit BID's ability to activate Bax. As a result, the anti-apoptotic Bcl-2 proteins may inhibit apoptosis by sequestering BID, leading to reduced Bax activation. The expression of BID is upregulated by the tumor suppressor p53, and BID has been shown to be involved in p53-mediated apoptosis. The p53 protein is a transcription factor that, when activated as part of the cell's response to stress, regulates many downstream target genes, including BID. However, p53 also has a transcription-independent role in apoptosis. In particular, p53 interacts with Bax, promoting Bax activation and the insertion of Bax into the mi
https://en.wikipedia.org/wiki/Info%20%28Unix%29
Info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command-line interface. Info reads info files generated by the texinfo program and presents the documentation as a tree with simple commands to traverse the tree and to follow cross references. For instance, pressing the space bar scrolls down within the current tree node or goes to the next node in the current document if already at the bottom of the current node, allowing to read the contents of an info file sequentially. Pressing the backspace key moves in the opposite direction. Furthermore: goes to the next node in the current document. goes to the previous node in the current document. goes to the next node on the same level as the current node. goes to the previous node on the same level as the current node. ("up") goes to the parent of the current node. goes to the last visited node. Moving the cursor over a link (a word preceded by an asterisk) and pressing the enter key follows the link. Pressing the tab key will move the cursor to the next nearest link. The C implementation of info was designed as the main documentation system of GNU based operating systems and was then ported to other Unix-like operating systems. However, info files had already been in use on ITS emacs. On the TOPS-20 operating system INFO was called XINFO. List of Info readers GNU info, distributed with Texinfo pinfo tkman tkinfo khelpcenter emacs info.vim (Vim plugin) vinfo (Vim plugin) GNOME Yelp See also Manual page (Unix) List of Unix commands
https://en.wikipedia.org/wiki/P53%20upregulated%20modulator%20of%20apoptosis
The p53 upregulated modulator of apoptosis (PUMA) also known as Bcl-2-binding component 3 (BBC3), is a pro-apoptotic protein, member of the Bcl-2 protein family. In humans, the Bcl-2-binding component 3 protein is encoded by the BBC3 gene. The expression of PUMA is regulated by the tumor suppressor p53. PUMA is involved in p53-dependent and -independent apoptosis induced by a variety of signals, and is regulated by transcription factors, not by post-translational modifications. After activation, PUMA interacts with antiapoptotic Bcl-2 family members, thus freeing Bax and/or Bak which are then able to signal apoptosis to the mitochondria. Following mitochondrial dysfunction, the caspase cascade is activated ultimately leading to cell death. Structure The PUMA protein is part of the BH3-only subgroup of Bcl-2 family proteins. This group of proteins only share sequence similarity in the BH3 domain, which is required for interactions with Bcl-2-like proteins, such as Bcl-2 and Bcl-xL. Structural analysis has shown that PUMA directly binds to antiapoptotic Bcl-2 family proteins via an amphiphatic α-helical structure which is formed by the BH3 domain. The mitochondrial localization of PUMA is dictated by a hydrophobic domain on its C-terminal portion. PUMA protein degradation is regulated by phosphorylation at a conserved serine residue at position 10.[31] Mechanism of action Biochemical studies have shown that PUMA interacts with antiapoptotic Bcl-2 family members such as Bcl-xL, Bcl-2, Mcl-1, Bcl-w, and A1, inhibiting their interaction with the proapoptotic molecules, Bax and Bak. When the inhibition of these is lifted, they result in the translocation of Bax and activation of mitochondrial dysfunction resulting in release of mitochondrial apoptogenic proteins cytochrome c, SMAC, and apoptosis-inducing factor (AIF) leading to caspase activation and cell death. Because PUMA has high affinity for binding to Bcl-2 family members, another hypothesis is that PUMA direc
https://en.wikipedia.org/wiki/Phorbol-12-myristate-13-acetate-induced%20protein%201
Phorbol-12-myristate-13-acetate-induced protein 1 is a protein that in humans is encoded by the PMAIP1 gene, and is also known as Noxa. Noxa (Latin for damage) is a pro-apoptotic member of the Bcl-2 protein family. Bcl-2 family members can form hetero- or homodimers, and they act as anti- or pro-apoptotic regulators that are involved in a wide variety of cellular activities. The expression of Noxa is regulated by the tumor suppressor p53, and Noxa has been shown to be involved in p53-mediated apoptosis. Interactions Noxa has been shown to interact with: BCL2-like 1, Bcl-2, and MCL1. See also Apoptosis Apoptosome Bcl-2 Bcl-2-associated X protein (BAX) BH3 interacting domain death agonist (BID) Caspases Cytochrome c Mitochondrion p53 upregulated modulator of apoptosis (PUMA) 12-O-Tetradecanoylphorbol-13-acetate (Phorbol-12-myristate-13-acetate)
https://en.wikipedia.org/wiki/Radio%20resource%20management
Radio resource management (RRM) is the system level management of co-channel interference, radio resources, and other radio transmission characteristics in wireless communication systems, for example cellular networks, wireless local area networks, wireless sensor systems, and radio broadcasting networks. RRM involves strategies and algorithms for controlling parameters such as transmit power, user allocation, beamforming, data rates, handover criteria, modulation scheme, error coding scheme, etc. The objective is to utilize the limited radio-frequency spectrum resources and radio network infrastructure as efficiently as possible. RRM concerns multi-user and multi-cell network capacity issues, rather than the point-to-point channel capacity. Traditional telecommunications research and education often dwell on channel coding and source coding with a single user in mind, but when several users and adjacent base stations share the same frequency channel it may not be possible to achieve the maximum channel capacity. Efficient dynamic RRM schemes may increase the system spectral efficiency by an order of magnitude, which often is considerably more than what is possible by introducing advanced channel coding and source coding schemes. RRM is especially important in systems limited by co-channel interference rather than by noise, for example cellular systems and broadcast networks homogeneously covering large areas, and wireless networks consisting of many adjacent access points that may reuse the same channel frequencies. The cost for deploying a wireless network is normally dominated by base station sites (real estate costs, planning, maintenance, distribution network, energy, etc.) and sometimes also by frequency license fees. So, the objective of radio resource management is typically to maximize the system spectral efficiency in bit/s/Hz/area unit or Erlang/MHz/site, under some kind of user fairness constraint, for example, that the grade of service should be above
https://en.wikipedia.org/wiki/Newick%20format
In mathematics, Newick tree format (or Newick notation or New Hampshire tree format) is a way of representing graph-theoretical trees with edge lengths using parentheses and commas. It was adopted by James Archie, William H. E. Day, Joseph Felsenstein, Wayne Maddison, Christopher Meacham, F. James Rohlf, and David Swofford, at two meetings in 1986, the second of which was at Newick's restaurant in Dover, New Hampshire, US. The adopted format is a generalization of the format developed by Meacham in 1984 for the first tree-drawing programs in Felsenstein's PHYLIP package. Examples The following tree: could be represented in Newick format in several ways ((,)); no nodes are named (A,B,(C,D)); leaf nodes are named (A,B,(C,D)E)F; all nodes are named (:0.1,:0.2,(:0.3,:0.4):0.5); all but root node have a distance to parent (:0.1,:0.2,(:0.3,:0.4):0.5):0.0; all have a distance to parent (A:0.1,B:0.2,(C:0.3,D:0.4):0.5); distances and leaf names (popular) (A:0.1,B:0.2,(C:0.3,D:0.4)E:0.5)F; distances and all names ((B:0.2,(C:0.3,D:0.4)E:0.5)F:0.1)A; a tree rooted on a leaf node (rare) Newick format is typically used for tools like PHYLIP and is a minimal definition for a phylogenetic tree. Rooted, unrooted, and binary trees When an unrooted tree is represented in Newick notation, an arbitrary node is chosen as its root. Whether rooted or unrooted, typically a tree's representation is rooted on an internal node and it is rare (but legal) to root a tree on a leaf node. A rooted binary tree that is rooted on an internal node has exactly two immediate descendant nodes for each internal node. An unrooted binary tree that is rooted on an arbitrary internal node has exactly three immediate descendant nodes for the root node, and each other internal node has exactly two immediate descendant nodes. A binary tree rooted from a leaf has at most one immediate de
https://en.wikipedia.org/wiki/Bare%20area%20of%20the%20liver
The bare area of the liver (nonperitoneal area) is a large triangular area on the diaphragmatic surface of the liver. It is the only part of the liver with no peritoneal covering, although it is still covered by Glisson’s capsule. It is attached directly to the diaphragm by loose connective tissue. The bare area of the liver is relevant to the portacaval anastomosis, encloses the right extraperitoneal subphrenic space, and can be a site of spread of infection from the abdominal cavity to the thoracic cavity Structure The bare area of the liver is found on the posterosuperior surface of the right lobe of the liver. This lies close to the thoracic diaphragm. It is the only part of the liver that has no peritoneal covering. It lies between the two layers of the coronary ligament, as well as the right triangular ligament. The coronary ligament represents reflections of the visceral peritoneum covering the liver onto the diaphragm. The bare area of the liver is attached to the thoracic diaphragm by loose connective tissue. It touches the bottom surface of the diaphragm. It is also not covered in capsule. Clinical significance The bare area of the liver is clinically important because of the portacaval anastomosis. It is a site where infection can spread from the abdominal cavity to the thoracic cavity. It encloses the right extraperitoneal subphrenic space. History The bare area of the liver may also be known as the nonperitoneal area.
https://en.wikipedia.org/wiki/Walking%20fish
A walking fish, or ambulatory fish, is a fish that is able to travel over land for extended periods of time. Some other modes of non-standard fish locomotion include "walking" along the sea floor, for example, in handfish or frogfish. Types Most commonly, walking fish are amphibious fish. Able to spend longer times out of water, these fish may use a number of means of locomotion, including springing, snake-like lateral undulation, and tripod-like walking. The mudskippers are probably the best land-adapted of contemporary fish and are able to spend days moving about out of water and can even climb mangroves, although to only modest heights. The climbing gourami is often specifically referred to as a "walking fish", although it does not actually "walk", but rather moves in a jerky way by supporting itself on the extended edges of its gill plates and pushing itself by its fins and tail. Some reports indicate that it can also climb trees. The epaulette shark (Hemiscyllium ocellatum) tends to live in shallow waters where swimming is difficult, and can often be seen walking over rocks and sand by using its muscular pectoral fins. It lives in areas of great variation in water depth, usually where the tide falls below its location. If it finds itself out of water, it can survive for several hours, and is capable of walking over land to get to water. This means that it is easily observed by beachgoers in its natural range. There are a number of fish that are less adept at actual walking, such as the walking catfish. Despite being known for "walking on land", this fish usually wriggles and may use its pectoral fins to aid in its movement. Walking catfish have a respiratory system that allows them to live out of water for several days. Some are invasive species, for example, the northern snakehead in the U.S. Polypterids have rudimentary lungs and can also move about on land, though rather clumsily. The mangrove rivulus can survive for months out of water and can move to
https://en.wikipedia.org/wiki/Weakly%20o-minimal%20structure
In model theory, a weakly o-minimal structure is a model-theoretic structure whose definable sets in the domain are just finite unions of convex sets. Definition A linearly ordered structure, M, with language L including an ordering relation <, is called weakly o-minimal if every parametrically definable subset of M is a finite union of convex (definable) subsets. A theory is weakly o-minimal if all its models are weakly o-minimal. Note that, in contrast to o-minimality, it is possible for a theory to have models that are weakly o-minimal and to have other models that are not weakly o-minimal. Difference from o-minimality In an o-minimal structure the definable sets in are finite unions of points and intervals, where interval stands for a sets of the form , for some a and b in . For weakly o-minimal structures this is relaxed so that the definable sets in M are finite unions of convex definable sets. A set is convex if whenever a and b are in , a < b and c ∈  satisfies that a < c < b, then c is in C. Points and intervals are of course convex sets, but there are convex sets that are not either points or intervals, as explained below. If we have a weakly o-minimal structure expanding (R,<), the real ordered field, then the structure will be o-minimal. The two notions are different in other settings though. For example, let R be the ordered field of real algebraic numbers with the usual ordering < inherited from R. Take a transcendental number, say π, and add a unary relation S to the structure given by the subset (−π,π) ∩ R. Now consider the subset A of R defined by the formula so that the set consists of all strictly positive real algebraic numbers that are less than π. The set is clearly convex, but cannot be written as a finite union of points and intervals whose endpoints are in R. To write it as an interval one would either have to include the endpoint π, which isn't in R, or one would require infinitely many intervals, such as the union Sinc
https://en.wikipedia.org/wiki/Meningohydroencephalocoele
Meningohydroencephalocoele (AmE: meningohydroencephalocele) is a form of meningocele (AmE) - a developmental abnormality of the central nervous system. Like meningocoele, meningohydroencephalocoele is caused by defects in bone ossification; in particular, the intramembranous ossification related to the closure of infantile fontanelles. It refers to the protrusion of the meninges between the un-fused bones, to lie subcutaneously. Meningocoele - refers to herniation of meninges. Meningoencephalocoele refers to the condition if brain tissue is included with the meninges in the herniation. Meningohydroencephalocoele refers to the condition including meninges, brain tissue and part of the ventricular system in the herniation. Encephalocoele defects occur in approximately 1 in 2000 live births.
https://en.wikipedia.org/wiki/Sexual%20differentiation%20in%20humans
Sexual differentiation in humans is the process of development of sex differences in humans. It is defined as the development of phenotypic structures consequent to the action of hormones produced following gonadal determination. Sexual differentiation includes development of different genitalia and the internal genital tracts and body hair plays a role in sex identification. The development of sexual differences begins with the XY sex-determination system that is present in humans, and complex mechanisms are responsible for the development of the phenotypic differences between male and female humans from an undifferentiated zygote. Females typically have two X chromosomes, and males typically have a Y chromosome and an X chromosome. At an early stage in embryonic development, both sexes possess equivalent internal structures. These are the mesonephric ducts and paramesonephric ducts. The presence of the SRY gene on the Y chromosome causes the development of the testes in males, and the subsequent release of hormones which cause the paramesonephric ducts to regress. In females, the mesonephric ducts regress. Divergent sexual development, known as intersex, can be a result of genetic and hormonal factors. Sex determination Most mammals, including humans, have an XY sex-determination system: the Y chromosome carries factors responsible for triggering male development. In the absence of a Y chromosome, the fetus will undergo female development. This is because of the presence of the sex-determining region of the Y chromosome, also known as the SRY gene. Thus, male mammals typically have an X and a Y chromosome (XY), while female mammals typically have two X chromosomes (XX). Chromosomal sex is determined at the time of fertilization; a chromosome from the sperm cell, either X or Y, fuses with the X chromosome in the egg cell. Gonadal sex refers to the gonads, that is the testis or ovaries, depending on which genes are expressed. Phenotypic sex refers to the struct
https://en.wikipedia.org/wiki/Constant%20air%20volume
Constant air volume (CAV) is a type of heating, ventilating, and air-conditioning (HVAC) system. In a simple CAV system, the supply air flow rate is constant, but the supply air temperature is varied to meet the thermal loads of a space. Most CAV systems are small, and serve a single thermal zone. However, variations such as CAV with reheat, CAV multizone, and CAV primary-secondary systems can serve multiple zones and larger buildings. In mid- to large-size buildings, new central CAV systems are somewhat rare. Due to fan energy savings potential, variable air volume (VAV) systems are more common. However, in small buildings and residences, CAV systems are often the system of choice due to their simplicity, low cost, and reliability. Such small CAV systems often have on/off control, rather than supply air temperature modulation, to vary their heating or cooling capacities. There are two types of CAV systems that are commonly in use to modify the supply air temperature: the terminal reheat system and the mixed air system. The terminal reheat system cools the air in the air handling unit down to the lowest possible needed temperature within its zone of spaces. This supplies a comfortable quality to the space, but wastes energy. The mixed air system has two air streams, typically one for the coldest and one for the hottest needed air temperature in the zone. The two air streams are strategically combined to offset the space's load. The mixed air system option is not as proficient at controlling the humidity, yet it does do well at controlling the temperature.
https://en.wikipedia.org/wiki/Falconer%27s%20formula
Heritability is the proportion of variance caused by genetic factors of a specific trait in a population. Falconer's formula is a mathematical formula that is used in twin studies to estimate the relative contribution of genetic vs. environmental factors to variation in a particular trait (that is, the heritability of the trait) based on the difference between twin correlations. Statistical models for heritability commonly include an error that will absorb phenotypic variation that cannot be described by genetics when analyzed. These are unique subject-specific influences on a trait. Falconer's formula was first proposed by the Scottish geneticist Douglas Falconer. The formula is where is the broad sense heritability, is the (monozygotic, MZ) identical twin correlation, and is the (dizygotic, DZ) fraternal twin correlation. Falconer's formula assumes the equal contribution of environmental factors in MZ pairs and DZ pairs. Therefore, additional phenotypic correlation between the two pairs is due to genetic factors. Subtracting the correlation of the DZ pairs from MZ pairs yields the variance in phenotypes contributed by genetic factors. The correlation of same sex MZ twins is always higher than the DZ twin correlation with various sexes and thus all gender differences are evaluated as heritable. To avoid this error, only genetic studies comparing MZ twins with the same sex DZ twins are valid. Correlations between (additive genetics) and (common environment) must be included in the derivation shown below. See also Quantitative genetics
https://en.wikipedia.org/wiki/Internal%20urethral%20orifice
The internal urethral orifice is the opening of the urinary bladder into the urethra. It is placed at the apex of the trigonum vesicae, in the most dependent part of the bladder. It is usually somewhat crescent-shaped; the mucous membrane immediately behind it presents a slight elevation in males, the uvula vesicae, caused by the middle lobe of the prostate. See also Internal sphincter muscle of urethra
https://en.wikipedia.org/wiki/Stool%20guaiac%20test
The stool guaiac test or guaiac fecal occult blood test (gFOBT) is one of several methods that detects the presence of fecal occult blood (blood invisible in the feces). The test involves placing a fecal sample on guaiac paper (containing a phenolic compound, alpha-guaiaconic acid, extracted from the wood resin of Guaiacum trees) and applying hydrogen peroxide which, in the presence of blood, yields a blue reaction product within seconds. The American College of Gastroenterology has recommended the abandoning of gFOBT testing as a colorectal cancer screening tool, in favor of the fecal immunochemical test (FIT). Though the FIT is preferred, even the guaiac FOB testing of average risk populations may have been sufficient to reduce the mortality associated with colon cancer by about 25%. With this lower efficacy, it was not always cost effective to screen a large population with gFOBT. Methodology The stool guaiac test involves fasting from iron supplements, red meat (the blood it contains can turn the test positive), certain vegetables (which contain a chemical with peroxidase properties that can turn the test positive), and vitamin C and citrus fruits (which can turn the test falsely negative) for a period of time before the test. It has been suggested that cucumber, cauliflower and horseradish, and often other vegetables, should be avoided for three days before the test. In testing, feces are applied to a thick piece of paper attached to a thin film coated with guaiac. Either the patient or medical professional smears a small fecal sample on to the film. The fecal sample is obtained by catching the stool and transferring a sample with an applicator. Digital rectal examination specimens are also used but this method is discouraged for colorectal cancer screening due to very poor performance characteristics. Both sides of the test card can be peeled open, to access the inner guaiac paper. One side of the card is marked for application of the stool and the o