source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Lander%E2%80%93Green%20algorithm
The Lander–Green algorithm is an algorithm, due to Eric Lander and Philip Green for computing the likelihood of observed genotype data given a pedigree. It is appropriate for relatively small pedigrees and a large number of markers. It is used in the analysis of genetic linkage.
https://en.wikipedia.org/wiki/IDRO%20Group
The Industrial Development & Renovation Organization of Iran (IDRO) known as IDRO Group was established in 1967 in Iran. IDRO Group is one of the largest companies in Iran. It is also one of the largest conglomerates in Asia. IDRO's objective is to develop Iran's industry sector and to accelerate the industrialization process of the country and to export Iranian products worldwide. Today, IDRO owns 117 subsidiaries and affiliated companies both domestically as well as internationally. Businesses In the course of its 40 years of activity, IDRO has gradually become a major shareholder of some key industries in Iran. In recent years and in accordance with the country's privatization policy, IDRO has made great efforts to privatize its affiliated companies. While carrying on its privatization policies and lessening its role as a holding company, IDRO intends to concentrate on its prime missions and to turn into an industrial development agency. IDRO has focused its activities on the following areas in order to materialize such strategy and to expedite the industrial development of Iran: Promotion of local and foreign investments with minority holdings owned by IDRO (less than 50% of the shares) with particular emphasis on new, hi-tech and export-oriented industries. Restructuring the existing industries through participation of reputable foreign companies in order to transfer new technologies and to enhance the non-oil exports of Iran. Development of general contracting activities with the participation of the Iranian private sector and credible foreign companies. Rendering consultancy and support services to foreign investors. Privatization of the existing subsidiaries. Industrial Investment Management Development Automotive Industry Industrial Equipment Machinery Marine Industry Railway Industry Hi-Tech Industries Development General Contracting Health care Banking Privatization IDRO had privatized 140 of its companies worth about 2,000 billion rials ($200 milli
https://en.wikipedia.org/wiki/Choline%20transporter
The high-affinity choline transporter (ChT) also known as solute carrier family 5 member 7 is a protein in humans that is encoded by the SLC5A7 gene. It is a cell membrane transporter and carries choline into acetylcholine-synthesizing neurons. Hemicholinium-3 is an inhibitor of the ChT and can be used to deplete acetylcholine stores, while coluracetam is an enhancer of the ChT and can increase cholinergic neurotransmission by enhancing acetylcholine synthesis. Function Choline is a direct precursor of acetylcholine (ACh), a neurotransmitter of the central and peripheral nervous system that regulates a variety of autonomic, cognitive, and motor functions. SLC5A7 is a Na(+)- and Cl(-)- dependent high-affinity transporter that mediates the uptake of choline for acetylcholine synthesis in cholinergic neurons. Mutations in the SLC5A7 gene have been associated with Distal spinal muscular atrophy with vocal cord paralysis (distal hereditary motor neuropathy type 7A). The ChT seems to be a site of action of some β-neurotoxins found in snake venoms, which disrupt peripheral cholinergic transmission by interfering with presynaptic acetylcholine synthesis. It is hypothesized that these toxins irreversibly block the ChT. Model organisms Model organisms have been used in the study of SLC5A7 function. A conditional knockout mouse line called Slc5a7tm1a(KOMP)Wtsi was generated at the Wellcome Trust Sanger Institute. Male and female animals underwent a standardized phenotypic screen to determine the effects of deletion. Additional screens performed: - In-depth immunological phenotyping Choline transporter in the human brain microvascular endothelial cells Choline is a necessary reagent for the synthesis of acetylcholine in the central nervous system. Neurons get their choline by specific protein transporters known as choline transporters. In the human brain microvascular endothelial cells, two systems initiate the choline absorption. The first system is known as the Ch
https://en.wikipedia.org/wiki/Body%20reactivity
Body reactivity is usually understood as the body's ability to react in a proper way to influence the environment. Resistance of an organism is its stability under the influence of pathogenic factors. The body reactivity can range from homeostasis to a fight or flight response. Ultimately, they are all governed by the nervous system. Nervous system divisions The central nervous system (CNS) consists of parts that are encased by the bones of the skull and spinal column: the brain and spinal cord. The peripheral nervous system (PNS) is found outside those bones and consists of the nerves and most of the sensory organs. Central nervous system The CNS can be divided into the brain and spinal cord. The CNS processes many different kinds of incoming sensory information. It is also the source of thoughts, emotions, and memories. Most signals that stimulate muscles to contract and glands to secrete originate in the CNS. The spinal cord and spinal nerves contribute to homeostasis by providing quick reflexive responses to many stimuli. The spinal cord is the pathway for sensory input to the brain and motor output from the brain. The brain is responsible for integrating most sensory information and coordinating body function, both consciously and unconsciously. Peripheral nervous system The PNS can be divided into the autonomic and somatic nervous system. The autonomic nervous system can be divided into the parasympathetic, sympathetic, and enteric nervous system. The sympathetic nervous system regulates the “fight or flight” responses. The parasympathetic nervous system regulates the “rest and digest” responses. The enteric nervous system innervates the viscera (gastrointestinal tract, pancreas, and gall bladder). The somatic nervous system consists of peripheral nerve fibers that send sensory information to the central nervous system and motor nerve fibers that project to skeletal muscle. The somatic nervous system engages in voluntary reactions, and the autonomic nervous
https://en.wikipedia.org/wiki/Flux%20tube
A flux tube is a generally tube-like (cylindrical) region of space containing a magnetic field, B, such that the cylindrical sides of the tube are everywhere parallel to the magnetic field lines. It is a graphical visual aid for visualizing a magnetic field. Since no magnetic flux passes through the sides of the tube, the flux through any cross section of the tube is equal, and the flux entering the tube at one end is equal to the flux leaving the tube at the other. Both the cross-sectional area of the tube and the magnetic field strength may vary along the length of the tube, but the magnetic flux inside is always constant. As used in astrophysics, a flux tube generally means an area of space through which a strong magnetic field passes, in which the behavior of matter (usually ionized gas or plasma) is strongly influenced by the field. They are commonly found around stars, including the Sun, which has many flux tubes from tens to hundreds of kilometers in diameter. Sunspots are also associated with larger flux tubes of 2500 km diameter. Some planets also have flux tubes. A well-known example is the flux tube between Jupiter and its moon Io. Definition The flux of a vector field passing through any closed orientable surface is the surface integral of the field over the surface. For example, for a vector field consisting of the velocity of a volume of liquid in motion, and an imaginary surface within the liquid, the flux is the volume of liquid passing through the surface per unit time. A flux tube can be defined passing through any closed, orientable surface in a vector field , as the set of all points on the field lines passing through the boundary of . This set forms a hollow tube. The tube follows the field lines, possibly turning, twisting, and changing its cross sectional size and shape as the field lines converge or diverge. Since no field lines pass through the tube walls there is no flux through the walls of the tube, so all the field lines ent
https://en.wikipedia.org/wiki/Cryogenic%20electron%20microscopy
Cryogenic electron microscopy (cryo-EM) is a cryomicroscopy technique applied on samples cooled to cryogenic temperatures. For biological specimens, the structure is preserved by embedding in an environment of vitreous ice. An aqueous sample solution is applied to a grid-mesh and plunge-frozen in liquid ethane or a mixture of liquid ethane and propane. While development of the technique began in the 1970s, recent advances in detector technology and software algorithms have allowed for the determination of biomolecular structures at near-atomic resolution. This has attracted wide attention to the approach as an alternative to X-ray crystallography or NMR spectroscopy for macromolecular structure determination without the need for crystallization. In 2017, the Nobel Prize in Chemistry was awarded to Jacques Dubochet, Joachim Frank, and Richard Henderson "for developing cryo-electron microscopy for the high-resolution structure determination of biomolecules in solution." Nature Methods also named cryo-EM as the "Method of the Year" in 2015. History Early development In the 1960s, the use of transmission electron microscopy for structure determination methods was limited because of the radiation damage due to high energy electron beams. Scientists hypothesized that examining specimens at low temperatures would reduce beam-induced radiation damage. Both liquid helium (−269 °C or 4 K or −452.2 °F) and liquid nitrogen (−195.79 °C or 77 K or −320 °F) were considered as cryogens. In 1980, Erwin Knapek and Jacques Dubochet published comments on beam damage at cryogenic temperatures sharing observations that: Thin crystals mounted on carbon film were found to be from 30 to 300 times more beam-resistant at 4 K than at room temperature... Most of our results can be explained by assuming that cryoprotection in the region of 4 K is strongly dependent on the temperature. However, these results were not reproducible and amendments were published in Nature just two years later
https://en.wikipedia.org/wiki/Polynomial-time%20reduction
In computational complexity theory, a polynomial-time reduction is a method for solving one problem using another. One shows that if a hypothetical subroutine solving the second problem exists, then the first problem can be solved by transforming or reducing it to inputs for the second problem and calling the subroutine one or more times. If both the time required to transform the first problem to the second, and the number of times the subroutine is called is polynomial, then the first problem is polynomial-time reducible to the second. A polynomial-time reduction proves that the first problem is no more difficult than the second one, because whenever an efficient algorithm exists for the second problem, one exists for the first problem as well. By contraposition, if no efficient algorithm exists for the first problem, none exists for the second either. Polynomial-time reductions are frequently used in complexity theory for defining both complexity classes and complete problems for those classes. Types of reductions The three most common types of polynomial-time reduction, from the most to the least restrictive, are polynomial-time many-one reductions, truth-table reductions, and Turing reductions. The most frequently used of these are the many-one reductions, and in some cases the phrase "polynomial-time reduction" may be used to mean a polynomial-time many-one reduction. The most general reductions are the Turing reductions and the most restrictive are the many-one reductions with truth-table reductions occupying the space in between. Many-one reductions A polynomial-time many-one reduction from a problem A to a problem B (both of which are usually required to be decision problems) is a polynomial-time algorithm for transforming inputs to problem A into inputs to problem B, such that the transformed problem has the same output as the original problem. An instance x of problem A can be solved by applying this transformation to produce an instance y of problem B
https://en.wikipedia.org/wiki/Geodemography
Geodemography is the study of people based on where they live; it links the sciences of demography, the study of human population dynamics, and geography, the study of the locational and spatial variation of both physical and human phenomena on Earth, along with sociology. It includes the application of geodemographic classifications for business, social research and public policy but has a parallel history in academic research seeking to understand the processes by which settlements (notably, cities) evolve and neighborhoods are formed. Geodemographic systems estimate the most probable characteristics of people based on the pooled profile of all people living in a small area near a particular address. Origins The origins of geodemographics are often identified as Charles Booth and his studies of deprivation and poverty in early twentieth century London, and the Chicago School of sociology. Booth developed the idea of 'classifying neighborhoods', exemplified by his multivariate classification of the 1891 UK Census data to create a generalized social index of London's (then) registration districts. Research at the Chicago School – though generally qualitative in nature – strengthened the idea that such classifications could be meaningful by developing the idea of 'natural areas' within cities: conceived as geographical units with populations of broadly homogenous social-economic and cultural characteristics. The idea that census outputs could serve to identify and to characterize the geographies of cities gathered momentum with the increased availability of national census data and the computational ability to look for patterns in such data. Of particular importance to the emerging geodemographic industry was the development of clustering techniques to group statistically similar neighborhoods into classes on a 'like with like' basis. More recently, data have become available at finer geographical resolutions (such as postal units), often originating from private
https://en.wikipedia.org/wiki/Calmar%20ratio
Calmar ratio (or Drawdown ratio) is a performance measurement used to evaluate Commodity Trading Advisors and hedge funds. It was created by Terry W. Young and first published in 1991 in the trade journal Futures. Young owned California Managed Accounts, a firm in Santa Ynez, California, which managed client funds and published the newsletter CMA Reports. The name of his ratio "Calmar" is an acronym of his company's name and its newsletter: CALifornia Managed Accounts Reports. Young defined it thus: Young believed the Calmar ratio was superior because It should be mentioned that a competitor newsletter, Managed Account Reports (founded in 1979 by publisher Leon Rose), had previously defined and popularized another performance measurement, the MAR Ratio, equal to the compound annual return from inception, divided by the maximum drawdown from inception. Although the Calmar ratio and MAR ratio are sometimes assumed to be identical, they are in fact different: Calmar ratio uses 36 months of performance data, whereas MAR ratio uses all performance data from inception onwards. Later versions of the Calmar ratio introduce the risk free rate into the numerator to create a Sharpe type ratio. See also Modigliani risk-adjusted performance Omega ratio Risk return ratio Sharpe ratio Sterling ratio Sortino ratio
https://en.wikipedia.org/wiki/List%20of%20fast%20rotators%20%28minor%20planets%29
This is a list of fast rotators—"minor planets" (which includes asteroids) that have an exceptionally short rotation period, also called "rotation rate" or "spin rate". In some cases the rotation period is not constant because the object tumbles (see List of tumblers). In this list the periods are sourced from the Light Curve Data Base (LCDB), and are given in both seconds and hours. Most minor planets have rotation periods between 2 and 20 hours. , a group of 887 bodies – most of them are stony near-Earth asteroids with small diameters of barely 1 kilometre – have an estimated period of less than 2.2 hours. According to the Minor Planet Center, most small bodies are thought to be rubble piles – conglomerations of smaller pieces, loosely coalesced under the influence of gravity. Bodies with a period below 2.2 hours – known as the "cohesionless spin-barrier" – cannot be merely held together by self-gravity, but must be formed of a contiguous solid, as they would fly apart otherwise. Via the deduction of strength boundary limits, rotation periods give an insight into the body's internal composition, and, from its degree of fracture, its collisional history can be inferred. Bodies with an uncertain period are displayed in dark-grey. They have an Asteroid Lightcurve Database (LCDB) quality code, U, of less than 2, which corresponds to an estimated error margin of larger than 30%. A trailing plus sign (+) or minus sign (−) indicate slightly better or worse quality, respectively, than the unsigned value. This list also includes a small group of bodies which have no provisional designation in the LCDB. Fastest rotators Periods 0.02–0.05 hour Periods 0.05–0.1 hour Periods 0.1–0.5 hour Periods 0.5–1 hour Periods 1–2 hours Periods 2–2.2 hours
https://en.wikipedia.org/wiki/Protochlorophyllide
Protochlorophyllide, or monovinyl protochlorophyllide, is an intermediate in the biosynthesis of chlorophyll a. It lacks the phytol side-chain of chlorophyll and the reduced pyrrole in ring D. Protochlorophyllide is highly fluorescent; mutants that accumulate it glow red if irradiated with blue light. In angiosperms, the later steps which convert protochlorophyllide to chlorophyll are light-dependent, and such plants are pale (chlorotic) if grown in the darkness. Gymnosperms, algae, and photosynthetic bacteria have another, light-independent enzyme and grow green in the darkness as well. Conversion to chlorophyll The enzyme that converts protochlorophyllide to chlorophyllide a, the next intermediate on the biosynthetic pathway, is protochlorophyllide reductase, EC 1.3.1.33. There are two structurally unrelated proteins with this activity: the light-dependent and the dark-operative. The light-dependent reductase needs light to operate. The dark-operative version is a completely different protein, consisting of three subunits that exhibit significant sequence similarity to the three subunits of nitrogenase, which catalyzes the formation of ammonia from dinitrogen. This enzyme might be evolutionary older but (being similar to nitrogenase) is highly sensitive to free oxygen and does not work if its concentration exceeds about 3%. Hence, the alternative, light-dependent version needed to evolve. Most of the photosynthetic bacteria have both light-dependent and light-independent reductases. Angiosperms have lost the dark-operative form and rely on 3 slightly different copies of light-dependent version, frequently abbreviated as POR A, B, and C. Gymnosperms have much more copies of the similar gene (Loblolly pine has about 11 Loblolly Pine (Pinus taeda L.) Contains Multiple Expressed Genes Encoding Light-Dependent NADPH:Protochlorophyllide Oxidoreductase (POR)). In plants, POR is encoded in the cell nucleus and only later transported to its place of work, chloroplast.
https://en.wikipedia.org/wiki/Ultraviolet%20fixed%20point
In a quantum field theory, one may calculate an effective or running coupling constant that defines the coupling of the theory measured at a given momentum scale. One example of such a coupling constant is the electric charge. In approximate calculations in several quantum field theories, notably quantum electrodynamics and theories of the Higgs particle, the running coupling appears to become infinite at a finite momentum scale. This is sometimes called the Landau pole problem. It is not known whether the appearance of these inconsistencies is an artifact of the approximation, or a real fundamental problem in the theory. However, the problem can be avoided if an ultraviolet or UV fixed point appears in the theory. A quantum field theory has a UV fixed point if its renormalization group flow approaches a fixed point in the ultraviolet (i.e. short length scale/large energy) limit. This is related to zeroes of the beta-function appearing in the Callan–Symanzik equation. The large length scale/small energy limit counterpart is the infrared fixed point. Specific cases and details Among other things, it means that a theory possessing a UV fixed point may not be an effective field theory, because it is well-defined at arbitrarily small distance scales. At the UV fixed point itself, the theory can behave as a conformal field theory. The converse statement, that any QFT which is valid at all distance scales (i.e. isn't an effective field theory) has a UV fixed point is false. See, for example, cascading gauge theory. Noncommutative quantum field theories have a UV cutoff even though they are not effective field theories. Physicists distinguish between trivial and nontrivial fixed points. If a UV fixed point is trivial (generally known as Gaussian fixed point), the theory is said to be asymptotically free. On the other hand, a scenario, where a non-Gaussian (i.e. nontrivial) fixed point is approached in the UV limit, is referred to as asymptotic safety. Asymptotically
https://en.wikipedia.org/wiki/Survival%20rate
Survival rate is a part of survival analysis. It is the proportion of people in a study or treatment group still alive at a given period of time after diagnosis. It is a method of describing prognosis in certain disease conditions, and can be used for the assessment of standards of therapy. The survival period is usually reckoned from date of diagnosis or start of treatment. Survival rates are based on the population as a whole and cannot be applied directly to an individual. There are various types of survival rates (discussed below). They often serve as endpoints of clinical trials and should not be confused with mortality rates, a population metric. Overall survival Patients with a certain disease (for example, colorectal cancer) can die directly from that disease or from an unrelated cause (for example, a car accident). When the precise cause of death is not specified, this is called the overall survival rate or observed survival rate. Doctors often use mean overall survival rates to estimate the patient's prognosis. This is often expressed over standard time periods, like one, five, and ten years. For example, prostate cancer has a much higher one-year overall survival rate than pancreatic cancer, and thus has a better prognosis. Sometimes the overall survival is reported as a death rate (%) without specifying the period the % applies to (possibly one year) or the period it is averaged over (possibly five years), e.g. Obinutuzumab: A Novel Anti-CD20 Monoclonal Antibody for Chronic Lymphocytic Leukemia. Net survival rate When someone is interested in how survival is affected by the disease, there is also the net survival rate, which filters out the effect of mortality from other causes than the disease. The two main ways to calculate net survival are relative survival and cause-specific survival or disease-specific survival. Relative survival has the advantage that it does not depend on accuracy of the reported cause of death; cause specific survival has the
https://en.wikipedia.org/wiki/Microhematuria
Microhematuria, also called microscopic hematuria (both usually abbreviated as MH), is a medical condition in which urine contains small amounts of blood; the blood quantity is too low to change the color of the urine (otherwise, it is known as gross hematuria). While not dangerous in itself, it may be a symptom of kidney disease, such as IgA nephropathy or Sickle cell trait, which should be monitored by a doctor. The American Urological Association (AUA) recommends a definition of microscopic hematuria as three or more red blood cells per high-power microscopic field in urinary sediment from two of three properly collected urinalysis specimens. Microhematuria is usually asymptomatic, and as of 2001 there were medical guidelines on how to handle asymptomatic microhematuria (AMH) so as to avoid problems such as overtreatment or misdiagnosis. In 2020 American Urological Association guidelines were updated. See also Proteinuria Hematuria Myoglobinuria Hemoglobinuria
https://en.wikipedia.org/wiki/Nonlocal%20Lagrangian
In field theory, a nonlocal Lagrangian is a Lagrangian, a type of functional containing terms that are nonlocal in the fields , i.e. not polynomials or functions of the fields or their derivatives evaluated at a single point in the space of dynamical parameters (e.g. space-time). Examples of such nonlocal Lagrangians might be: The Wess–Zumino–Witten action. Actions obtained from nonlocal Lagrangians are called nonlocal actions. The actions appearing in the fundamental theories of physics, such as the Standard Model, are local actions; nonlocal actions play a part in theories that attempt to go beyond the Standard Model and also in some effective field theories. Nonlocalization of a local action is also an essential aspect of some regularization procedures. Noncommutative quantum field theory also gives rise to nonlocal actions. Quantum measurement Quantum field theory Theoretical physics
https://en.wikipedia.org/wiki/Encyclopedia%20of%20Earth
The Encyclopedia of Earth (abbreviated EoE) is an electronic reference about the Earth, its natural environments, and their interaction with society. The Encyclopedia is described as a free, fully searchable collection of articles written by scholars, professionals, educators, and other approved experts, who collaborate and review each other's work. The articles are written in non-technical language and are intended to be useful to students, educators, scholars, and professionals, as well as to the general public. The authors, editors, and even copy editors are attributed on the articles with links to biographical pages on those individuals. The Encyclopedia of Earth is a component of the larger Earth Portal (part of the Digital Universe project), which is a constellation of subject-specific information portals that contain news services, structured metadata, a federated environmental search engine, and other information resources. The technology platform for the Encyclopedia of Earth is a modified version of MediaWiki, which is closed to all but approved users. Once an article is reviewed and approved it is published to a public site. The EoE was launched in September 2006 with about 360 articles, and as of November 30, 2010 had 7,678 articles. Authoring and publishing process Contributors to the Encyclopedia of Earth are made up of scientists, educators, and professionals within the environmental field. Contributors are vetted by the Environmental Information Coalition (EIC) Stewardship Committee, the governing body of the Encyclopedia of Earth, before they are given access to the author's wiki. Within the wiki, where they operate under their real names and are given attribution for the published articles. Articles are written, edited, and published in a two-step process: Content for the Encyclopedia is created, maintained, and governed by group of experts via a restricted-access wiki that uses a modified version of MediaWiki. Upon completion, content is rev
https://en.wikipedia.org/wiki/Health%20Services%20Laboratories
Health Services Laboratories (HSL) is an independent provider of pathology and diagnostic services to the NHS. Formed in 2015, it is a partnership between The Doctors Laboratory, University College London Hospitals NHS Foundation Trust and the Royal Free London NHS Foundation Trust. The Doctors Laboratory is the longest-established specialist provider of clinically-led pathology services in the UK and is based in central London. It is owned by Sonic Healthcare, an Australian company providing pathology and radiology services. HSL's flagship laboratory, the Halo, is based at 1 Mabledon Place, in the heart of London's 'med-city', and is one of the largest pathology laboratories in Europe. Currently in development, it is spread over 11 floors with five split-level basements. The facility will be home to more than 1,000 staff working within a connected suite of laboratories spanning more than 100,000 square feet. The Halo will also have dedicated clinical and non-clinical cores for vertical connectivity. Its other laboratories include rapid response laboratories (RRLs) at Mortimer Market, 60 Whitfield Street, the Royal Free London and North Middlesex Hospital. HSL provides services to a number of NHS trusts. It does not provide services to individuals on a commercial basis. Lord Carter of Coles chairs the HSL board. The Doctors Laboratory was an early adopter of noninvasive prenatal testing. It started using Roche Diagnostics's Ariosa Diagnostics Harmony test in late 2012. In May 2017 a report to the North Middlesex University Hospital NHS Trust, which uses the company's services, outlined a rise in incidents reported during the transition to the brand new laboratory. HSL and North Middlesex worked in close partnership to resolve these early teething issues. In the COVID-19 pandemic in England the firm had £60 million in contracts mostly for PCR testing services in the new Surge Capacity Laboratory in MedCity. Profits were more than £10 million, up from £1.9 mill
https://en.wikipedia.org/wiki/Multifactorial%20disease
Multifactorial diseases are not confined to any specific pattern of single gene inheritance and are likely to be caused when multiple genes come together along with the effects of environmental factors. In fact, the terms ‘multifactorial’ and ‘polygenic’ are used as synonyms and these terms are commonly used to describe the architecture of disease causing genetic component. Multifactorial diseases are often found gathered in families yet, they do not show any distinct pattern of inheritance. It is difficult to study and treat multifactorial diseases because specific factors associated with these diseases have not yet been identified. Some common multifactorial disorders include schizophrenia, diabetes, asthma, depression, high blood pressure, Alzheimer’s, obesity, epilepsy, heart diseases, Hypothyroidism, club foot, cancer, birth defects and even dandruff. The Multifactorial threshold model assumes that gene defects for multifactorial traits are usually distributed within populations.  Firstly, different populations might have different thresholds. This is the case in which occurrences of a particular disease is different in males and females (e.g. Pyloric stenosis). The distribution of susceptibility is the same but threshold is different. Secondly, threshold may be same but the distributions of susceptibility may be different. It explains the underlying risks present in first degree relatives of affected individuals. Characteristics Multifactorial disorders exhibit a combination of distinct characteristics which are clearly differentiated from Mendelian inheritance. The risk of multifactorial diseases may get increased due to environmental influences. The disease is not sex-limited but it occurs more frequently in one gender than the other; females are more likely to have neural tube defects compared to males. The disease occurs more commonly in a distinct ethnic group (i.e., Africans, Asians, Caucasians etc.) The diseases may have more in common than g
https://en.wikipedia.org/wiki/Spatial-numerical%20association%20of%20response%20codes
The spatial-numerical association of response codes (SNARC) is an example of the spatial organisation of magnitude information. Put simply, when presented with smaller numbers (0 to 4), people tend to respond faster if those stimuli are associated with the left extrapersonal hemiside of their perceived surroundings; when presented with larger numbers (6 to 9), people respond faster if those stimuli are instead associated with the right extrapersonal hemiside of their perceived surroundings. The SNARC effect is this automatic association that occurs between the location of the response hand and the semantic magnitude of a modality-independent number. Even for tasks in which magnitude is irrelevant, like parity judgement or phoneme detection, larger numbers are faster responded to with the right response key while smaller numbers are faster responded to with the left. This also occurs when the hands are crossed, with the right hand activating the left response key and vice versa. The explanation given by Dehaene and colleagues is that the magnitude of a number on an oriented mental number line is automatically activated. The mental number line is assumed to be oriented from left to right in populations with a left-to-right writing system (e.g. English), and oriented from right to left in populations with a right-to-left writing system (e.g. Iranian) Effects The SNARC has been observed primarily in two scenarios: attentional and oculomotor. The first of these involves people being faster to detect left probes after smaller numbers are shown and right probes after large numbers, whereas the oculomotor effects are seen when participants look at greater speeds towards the left after detecting small numbers and to the right after detecting large ones. Newer research shows a motor bias to also be associated with the SNARC effect. In an experiment conducted into random number generation, participants tended to generate numbers of a larger magnitude when turning their head
https://en.wikipedia.org/wiki/Sphingomyelin
Sphingomyelin (SPH, ) is a type of sphingolipid found in animal cell membranes, especially in the membranous myelin sheath that surrounds some nerve cell axons. It usually consists of phosphocholine and ceramide, or a phosphoethanolamine head group; therefore, sphingomyelins can also be classified as sphingophospholipids. In humans, SPH represents ~85% of all sphingolipids, and typically make up 10–20 mol % of plasma membrane lipids. Sphingomyelin was first isolated by German chemist Johann L.W. Thudicum in the 1880s. The structure of sphingomyelin was first reported in 1927 as N-acyl-sphingosine-1-phosphorylcholine. Sphingomyelin content in mammals ranges from 2 to 15% in most tissues, with higher concentrations found in nerve tissues, red blood cells, and the ocular lenses. Sphingomyelin has significant structural and functional roles in the cell. It is a plasma membrane component and participates in many signaling pathways. The metabolism of sphingomyelin creates many products that play significant roles in the cell. Physical characteristics Composition Sphingomyelin consists of a phosphocholine head group, a sphingosine, and a fatty acid. It is one of the few membrane phospholipids not synthesized from glycerol. The sphingosine and fatty acid can collectively be categorized as a ceramide. This composition allows sphingomyelin to play significant roles in signaling pathways: the degradation and synthesis of sphingomyelin produce important second messengers for signal transduction. Sphingomyelin obtained from natural sources, such as eggs or bovine brain, contains fatty acids of various chain length. Sphingomyelin with set chain length, such as palmitoylsphingomyelin with a saturated 16 acyl chain, is available commercially. Properties Ideally, sphingomyelin molecules are shaped like a cylinder, however many molecules of sphingomyelin have a significant chain mismatch (the lengths of the two hydrophobic chains are significantly different). The hydrophob
https://en.wikipedia.org/wiki/Bacteriophage%20PBC1
Bacteriophage PBC1 is a bacteriophage that infects the spore-forming bacterium Bacillus cereus.
https://en.wikipedia.org/wiki/I/O%20virtualization
In virtualization, input/output virtualization (I/O virtualization) is a methodology to simplify management, lower costs and improve performance of servers in enterprise environments. I/O virtualization environments are created by abstracting the upper layer protocols from the physical connections. The technology enables one physical adapter card to appear as multiple virtual network interface cards (vNICs) and virtual host bus adapters (vHBAs). Virtual NICs and HBAs function as conventional NICs and HBAs, and are designed to be compatible with existing operating systems, hypervisors, and applications. To networking resources (LANs and SANs), they appear as normal cards. In the physical view, virtual I/O replaces a server’s multiple I/O cables with a single cable that provides a shared transport for all network and storage connections. That cable (or commonly two cables for redundancy) connects to an external device, which then provides connections to the data center networks. Background Server I/O is a critical component to successful and effective server deployments, particularly with virtualized servers. To accommodate multiple applications, virtualized servers demand more network bandwidth and connections to more networks and storage. According to a survey, 75% of virtualized servers require 7 or more I/O connections per device, and are likely to require more frequent I/O reconfigurations. In virtualized data centers, I/O performance problems are caused by running numerous virtual machines (VMs) on one server. In early server virtualization implementations, the number of virtual machines per server was typically limited to six or less. But it was found that it could safely run seven or more applications per server, often using 80 percent of total server capacity, an improvement over the average 5 to 15 percent utilized with non-virtualized servers . However, increased utilization created by virtualization placed a significant strain on the server’s I/O cap
https://en.wikipedia.org/wiki/CONQUEST
CONQUEST is a linear scaling, or O(N), density functional theory (DFT) electronic structure open-source code. The code is designed to perform DFT calculations on very large systems containing many thousands of atoms. It can be run at different levels of precision ranging from ab initio tight binding up to full DFT with plane wave accuracy. It has been applied to the study of three-dimensional reconstructions formed by Ge on Si(001), containing over 20,000 atoms. Tests on the UK's national supercomputer HECToR in 2009 demonstrated the capability of the code to perform ground-state calculations on systems of over 1,000,000 atoms. Methodology Instead of solving for the Kohn-Sham eigenstates as normal DFT codes do, CONQUEST solves for the one particle density matrix, . To make the problem computationally tractable, the density matrix is written in separable form: , where is a support function centred on atom i (with support functions on the same atom notated by ) and is the density matrix in the basis of the support functions. The ground state is found as a series of nested loops: • Minimise the energy with respect to the density matrix for fixed charge density and support functions • Find self-consistency between charge density and potential • Minimise the energy with respect to the support functions The support functions are confined within spheres of given cutoff radius and the density matrix is forced to zero beyond a given range: . These approximations give linear scaling behaviour, and as the radii are increased tend to the exact result. Developers CONQUEST is jointly developed at the Department of Physics and Astronomy and London Centre for Nanotechnology, University College London in the UK and at the Computational Materials Science Centre, National Institute for Materials Science, Tsukuba, Japan. In the UK, the development team includes Dr. David Bowler, Dr. Veronika Brazdova, Prof. Mike Gillan, Dr. Andrew Horsfield, Mr. Alex Sena, Mr. Lianheng Tong,
https://en.wikipedia.org/wiki/Cisco%20PIX
Cisco PIX (Private Internet eXchange) was a popular IP firewall and network address translation (NAT) appliance. It was one of the first products in this market segment. In 2005, Cisco introduced the newer Cisco Adaptive Security Appliance (Cisco ASA), that inherited many of the PIX features, and in 2008 announced PIX end-of-sale. The PIX technology was sold in a blade, the FireWall Services Module (FWSM), for the Cisco Catalyst 6500 switch series and the 7600 Router series, but has reached end of support status as of September 26, 2007. PIX History PIX was originally conceived in early 1994 by John Mayes of Redwood City, California and designed and coded by Brantley Coile of Athens, Georgia. The PIX name is derived from its creators' aim of creating the functional equivalent of an IP PBX to solve the then-emerging registered IP address shortage. At a time when NAT was just being investigated as a viable approach, they wanted to conceal a block or blocks of IP addresses behind a single or multiple registered IP addresses, much as PBXs do for internal phone extensions. When they began, RFC 1597 and RFC 1631 were being discussed, but the now-familiar RFC 1918 had not yet been submitted. The design, and testing were carried out in 1994 by John Mayes, Brantley Coile and Johnson Wu of Network Translation, Inc., with Brantley Coile being the sole software developer. Beta testing of PIX serial number 000000 was completed and first customer acceptance was on December 21, 1994 at KLA Instruments in San Jose, California. The PIX quickly became one of the leading enterprise firewall products and was awarded the Data Communications Magazine "Hot Product of the Year" award in January 1995. Shortly before Cisco acquired Network Translation in November 1995, Mayes and Coile hired two longtime associates, Richard (Chip) Howes and Pete Tenereillo, and shortly after acquisition 2 more longtime associates, Jim Jordan and Tom Bohannon. Together they continued development on Fine
https://en.wikipedia.org/wiki/Three-taxon%20analysis
Three-taxon analysis (or TTS, three-item analysis, 3ia) is a cladistic based method of phylogenetic reconstruction. Introduced by Nelson and Platnick in 1991 to reconstruct organisms' phylogeny, this method can also be applied to biogeographic areas. It attempts to reconstruct complex phylogenetic trees by breaking the problem down into simpler chunks. Rather than try to resolve the relationships of all X taxa at once, it considers taxa 3 at a time. It is relatively easy to generate three-taxon statements (3is); that is, statements of the form "A and B are more closely related to one another than to C". Once each group of three taxa has been considered, the method constructs a tree that is consistent with as many three-item statements as possible. From a theoretical point of view, the method has three main problems: (1) character evolution is a priori assumed to be irreversible; (2) 3is that are not logically independent are treated as if they are; (3) 3is that are considered as independent support for a given tree may be mutually exclusive on that tree. A computer program that implement three-taxon analysis is LisBeth (for systematic and biogeographic studies). LisBeth have been freely released. A recent simulation-based study found that Three-taxon analysis yields good power and an error rate intermediate between parsimony with ordered states and parsimony with unordered states. See also Cladistics
https://en.wikipedia.org/wiki/Grotesque%20%28architecture%29
In architecture, a grotesque () or chimera () is a fantastic or mythical figure carved from stone and fixed to the walls or roof of a building. A chimera is a kind of grotesque in which the figure is a combination of animals (including humans). Grotesque are often called gargoyles, although the term gargoyle refers to figures carved specifically to drain water away from the sides of buildings. In the Middle Ages, the term babewyn was used to refer to both gargoyles and chimerae. This word is derived from the Italian word , which means "baboon". Grotesques often depict whimsical, mythical creatures in dramatic or humorous ways. They have historically been a key element of architecture in many periods including the Renaissance and Medieval periods and have stylistically developed in conjunction with these times. Although grotesques typically depict a wide range of subjects, they are often hybrids of different mythical, human, and animalistic features. Many scholars describe grotesques as being used to ward off evil and as reminders of the separation of the earth and the divine. Grotesques are predominantly carved into buildings of religious significance, in particular churches and cathedrals. Despite their presence in religious spaces, their anthropomorphic designs are largely not directly religious and instead are often more whimsical without religious connotations. They commonly exist on high ledges and rooftops and are frequently positioned out of view from common areas. Prominent examples of preserved grotesques exist on buildings such as the Florence Cathedral and Notre-Dame de Paris. Historically, grotesques have also had significant design influence from sculptural trends and often their architects were originally sculptors or artists. This meant that the widespread emergence of grotesques also often converged with popular art styles that existed at the time, especially the combined rise of the gothic style and the addition of grotesques in architecture. Key
https://en.wikipedia.org/wiki/Tidal%20volume
Tidal volume (symbol VT or TV) is the volume of air moved into or out of the lungs during a normal breath. In a healthy, young human adult, tidal volume is approximately 500 ml per inspiration or 7 ml/kg of body mass. Mechanical ventilation Tidal volume plays a significant role during mechanical ventilation to ensure adequate ventilation without causing trauma to the lungs. Tidal volume is measured in milliliters and ventilation volumes are estimated based on a patient's ideal body mass. Measurement of tidal volume can be affected (usually overestimated) by leaks in the breathing circuit or the introduction of additional gas, for example during the introduction of nebulized drugs. Ventilator-induced lung injury such as Acute lung injury (ALI) /Acute Respiratory Distress Syndrome (ARDS) can be caused by ventilation with very large tidal volumes in normal lungs, as well as ventilation with moderate or small volumes in previously injured lungs, and research shows that the incidence of ALI increases with higher tidal volume settings in nonneurologically impaired patients. . Similarly A 2018 systematic review by The Cochrane Collaboration provided evidence that low tidal volume ventilation reduced post operative pneumonia and reduced the requirement for both invasive and non invasive ventilation after surgery Initial settings of mechanical ventilation: Patients without pre-existing lung disease Protective lung ventilation strategies should be applied with VT 6ml/kg to 8ml/kg with RR = 12 to 20 and an average starting target minute ventilation of 7 L/min. Patients with chronic obstructive lung disease Protective lung volumes apply 6ml/kg to 8ml/kg with a rate high enough for proper alveolar ventilation but does not create or aggravate intrinsic Positive End-Expiry Pressure (PEEP). Acute respiratory distress syndrome Protective lung ventilation strategies apply. VT 6 to 8 ml/kg or as low as 5 ml/kg in severe cases. Permissive hypercapnia can be employed in an attem
https://en.wikipedia.org/wiki/Linopristin
Linopristin is an antibiotic of the streptogramin B class. It is one of the components of the antibiotic combination NXL103.
https://en.wikipedia.org/wiki/Urn%20problem
In probability and statistics, an urn problem is an idealized mental exercise in which some objects of real interest (such as atoms, people, cars, etc.) are represented as colored balls in an urn or other container. One pretends to remove one or more balls from the urn; the goal is to determine the probability of drawing one color or another, or some other properties. A number of important variations are described below. An urn model is either a set of probabilities that describe events within an urn problem, or it is a probability distribution, or a family of such distributions, of random variables associated with urn problems. History In Ars Conjectandi (1713), Jacob Bernoulli considered the problem of determining, given a number of pebbles drawn from an urn, the proportions of different colored pebbles within the urn. This problem was known as the inverse probability problem, and was a topic of research in the eighteenth century, attracting the attention of Abraham de Moivre and Thomas Bayes. Bernoulli used the Latin word urna, which primarily means a clay vessel, but is also the term used in ancient Rome for a vessel of any kind for collecting ballots or lots; the present-day Italian word for ballot box is still urna. Bernoulli's inspiration may have been lotteries, elections, or games of chance which involved drawing balls from a container, and it has been asserted that elections in medieval and renaissance Venice, including that of the doge, often included the choice of electors by lot, using balls of different colors drawn from an urn. Basic urn model In this basic urn model in probability theory, the urn contains x white and y black balls, well-mixed together. One ball is drawn randomly from the urn and its color observed; it is then placed back in the urn (or not), and the selection process is repeated. Possible questions that can be answered in this model are: Can I infer the proportion of white and black balls from n observations? With what deg
https://en.wikipedia.org/wiki/M-325
In the history of cryptography, M-325, also known as SIGFOY, was an American rotor machine designed by William F. Friedman and built in 1944. Between 1944 and 1946, more than 1,100 machines were deployed within the United States Foreign Service. Its use was discontinued in 1946 because of faults in operation. Friedman applied for a patent on the M-325 on 11 August 1944; it was and was granted on 17 March 1959 (US patent #2,877,565). Like the Enigma, the M-325 contains three intermediate rotors and a reflecting rotor. See also Hebern rotor machine SIGABA
https://en.wikipedia.org/wiki/Cross-covariance%20matrix
In probability theory and statistics, a cross-covariance matrix is a matrix whose element in the i, j position is the covariance between the i-th element of a random vector and j-th element of another random vector. A random vector is a random variable with multiple dimensions. Each element of the vector is a scalar random variable. Each element has either a finite number of observed empirical values or a finite or infinite number of potential values. The potential values are specified by a theoretical joint probability distribution. Intuitively, the cross-covariance matrix generalizes the notion of covariance to multiple dimensions. The cross-covariance matrix of two random vectors and is typically denoted by or . Definition For random vectors and , each containing random elements whose expected value and variance exist, the cross-covariance matrix of and is defined by where and are vectors containing the expected values of and . The vectors and need not have the same dimension, and either might be a scalar value. The cross-covariance matrix is the matrix whose entry is the covariance between the i-th element of and the j-th element of . This gives the following component-wise definition of the cross-covariance matrix. Example For example, if and are random vectors, then is a matrix whose -th entry is . Properties For the cross-covariance matrix, the following basic properties apply: If and are independent (or somewhat less restrictedly, if every random variable in is uncorrelated with every random variable in ), then where , and are random vectors, is a random vector, is a vector, is a vector, and are matrices of constants, and is a matrix of zeroes. Definition for complex random vectors If and are complex random vectors, the definition of the cross-covariance matrix is slightly changed. Transposition is replaced by Hermitian transposition: For complex random vectors, another matrix called the pseudo-cross-
https://en.wikipedia.org/wiki/Privacy%20in%20Australian%20law
There is no absolute right to privacy in Australian law and there is no clearly recognised tort of invasion of privacy or similar remedy available to people who feel their privacy has been violated. Privacy is, however, affected and protected in limited ways by common law in Australia and a range of federal, state and territorial laws, as well as administrative arrangements. What is privacy? There is no statutory definition of privacy in Australia. The Australian Law Reform Commission (ALRC) was given a reference to review Australian privacy law in 2006. During that review it considered the definition of privacy in 2007 in its Discussion paper 72. In it, the ALRC found there is no "precise definition of universal application" of privacy; instead it conducted the inquiry considering the contextual use of the term "privacy". In reaching that conclusion, the ALRC began by considering the concept of privacy: "It has been suggested that privacy can be divided into some separate, but related concepts: Information privacy, which involves the establishment of rules governing the collection and handling of personal data such as credit information, and medical and government records. It is also known as "data protection"; Bodily privacy, which concerns the protection of people’s physical selves against invasive procedures such as genetic tests, drug testing and cavity searches; Privacy of communications, which covers the security and privacy of mail, telephones, e-mail and other forms of communication; and Territorial privacy, which concerns the setting of limits on intrusion into the domestic and other environments such as the workplace or public space. This includes searches, video surveillance and ID checks. Privacy at common law It is unclear if a tort of invasion of privacy exists under Australian law. The ALRC summarised the position in 2007: "In Australia, no jurisdiction has enshrined in legislation a cause of action for invasion of privacy; however, the door t
https://en.wikipedia.org/wiki/Orbiting%20body
In astrodynamics, an orbiting body is any physical body that orbits a more massive one, called the primary body. The orbiting body is properly referred to as the secondary body (), which is less massive than the primary body (). Thus, or . Under standard assumptions in astrodynamics, the barycenter of the two bodies is a focus of both orbits. An orbiting body may be a spacecraft (i.e. an artificial satellite) or a natural satellite, such as a planet, dwarf planet, moon, moonlet, asteroid, or comet. A system of two orbiting bodies is modeled by the Two-Body Problem and a system of three orbiting bodies is modeled by the Three-Body Problem. These problems can be generalized to an N-body problem. While there are a few analytical solutions to the n-body problem, it can be reduced to a 2-body system if the secondary body stays out of other bodies' Sphere of Influence and remains in the primary body's sphere of influence. See also Barycenter Double planet Primary (astronomy) Satellite Two-body problem Three-body problem N-body problem
https://en.wikipedia.org/wiki/Aplastic%20anemia
Aplastic anemia (AA) is a severe hematologic condition in which the body fails to make blood cells in sufficient numbers. Aplastic anemia is associated with cancer and various cancer syndromes. Blood cells are produced in the bone marrow by stem cells that reside there. Aplastic anemia causes a deficiency of all blood cell types: red blood cells, white blood cells, and platelets. It occurs most frequently in people in their teens and twenties but is also common among the elderly. It can be caused by heredity, immune disease, or exposure to chemicals, drugs, or radiation. However, in about half of cases, the cause is unknown. Aplastic anemia can be definitively diagnosed by bone marrow biopsy. Normal bone marrow has 30–70% blood stem cells, but in aplastic anemia, these cells are mostly gone and are replaced by fat. First-line treatment for aplastic anemia consists of immunosuppressive drugs—typically either anti-lymphocyte globulin or anti-thymocyte globulin—combined with corticosteroids, chemotherapy, and ciclosporin. Hematopoietic stem cell transplantation is also used, especially for patients under 30 years of age with a related, matched marrow donor. Aplastic anemia is known to have caused the deaths of Eleanor Roosevelt, Luana Reyes, Molly Holzschlag, and Marie Curie. Signs and symptoms Anemia may lead to fatigue, pale skin, severe bruising, and a fast heart rate. Low platelets are associated with an increased risk of bleeding, bruising, and petechiae, with lower blood counts that impact the ability of the blood to clot. Low white blood cells increase the risk of infections. Causes Aplastic anemia can be caused by immune disease or exposure to certain chemicals, drugs, radiation, or infection; in about half the cases, a definitive cause is unknown. It is not a hereditary condition, nor is it contagious. Aplastic anemia is also sometimes associated with exposure to toxins such as benzene or with the use of certain drugs, including chloramphenicol, carbam
https://en.wikipedia.org/wiki/English%20terms%20with%20diacritical%20marks
English rarely uses diacritics, which are symbols indicating the modification of a letter's sound when spoken. Most of the affected words are in terms imported from other languages The two dots accent (diaeresis or umlaut), the grave accent and the acute accent are the only diacritics native to Modern English, and their usage has tended to fall off except in certain publications and particular cases. Proper nouns are not generally counted as English terms except when accepted into the language as an eponym – such as Geiger–Müller tube, or the English terms roentgen after Wilhelm Röntgen, and biro after László Bíró, in which case any diacritical mark is often lost. Unlike continental European languages, English orthography tends to use digraphs (like "sh", "oo", and "ea") rather than diacritics to indicate more sounds than can be accommodated by the letters of the Latin alphabet. Unlike other systems (such as Spanish orthography) where the spelling indicates the pronunciation, English spelling is highly varied, and diacritics alone would be insufficient to make it reliably phonetic. (See .) Types of diacritical marks Though limited, the following diacritical marks in English may be encountered, particularly for marking in poetry: the acute accent (née) and grave accent (English poetry marking, changèd), modifying vowels or marking stresses the circumflex (entrepôt), borrowed from French the diaeresis (Zoë), indicating a second syllable in two consecutive vowels the tittle, the dot found on the regular small i and small j, is removed when another diacritic is required the macron (English poetry marking, lēad pronounced , not ), lengthening vowels, as in Māori; or indicating omitted n or m (in pre-Modern English, both in print and in handwriting). the breve (English poetry marking, drŏll pronounced , not ), shortening vowels the umlaut (über), altering Germanic vowels the cedilla (soupçon, façade), in French, Portuguese and in Catalan it is a softening c
https://en.wikipedia.org/wiki/Lankford%20coefficient
The Lankford coefficient (also called Lankford value, R-value, or plastic strain ratio) is a measure of the plastic anisotropy of a rolled sheet metal. This scalar quantity is used extensively as an indicator of the formability of recrystallized low-carbon steel sheets. Definition If and are the coordinate directions in the plane of rolling and is the thickness direction, then the R-value is given by where is the in-plane plastic strain, transverse to the loading direction, and is the plastic strain through-the-thickness. More recent studies have shown that the R-value of a material can depend strongly on the strain even at small strains . In practice, the value is usually measured at 20% elongation in a tensile test. For sheet metals, the values are usually determined for three different directions of loading in-plane ( to the rolling direction) and the normal R-value is taken to be the average The planar anisotropy coefficient or planar R-value is a measure of the variation of with angle from the rolling direction. This quantity is defined as Anisotropy of steel sheets Generally, the Lankford value of cold rolled steel sheet acting for deep-drawability shows heavy orientation, and such deep-drawability is characterized by . However, in the actual press-working, the deep-drawability of steel sheets cannot be determined only by the value of and the measure of planar anisotropy, is more appropriate. In an ordinary cold rolled steel, is the highest, and is the lowest. Experience shows that even if is close to 1, and can be quite high leading to a high average value of . In such cases, any press-forming process design on the basis of does not lead to an improvement in deep-drawability. See also Yield surface
https://en.wikipedia.org/wiki/76th%20meridian%20west
The meridian 76° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, North America, the Atlantic Ocean, the Caribbean Sea, South America, the Pacific Ocean, the Southern Ocean, and Antarctica to the South Pole. The 76th meridian west forms a great circle with the 104th meridian east. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 76th meridian west passes through: {| class="wikitable plainrowheaders" ! scope="col" width="120" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Ellesmere Island |-valign="top" | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Baffin Bay | style="background:#b0e0e6;" | Passing just east of Bylot Island, Nunavut, (at ) |- | ! scope="row" | | Nunavut — Baffin Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Foxe Basin | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Baird Peninsula, Baffin Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Foxe Basin | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Prince Charles Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Foxe Basin | style="background:#b0e0e6;" | |- | ! scope="row" | | Nunavut — Foxe Peninsula, Baffin Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Hudson Strait | style="background:#b0e0e6;" | |-valign="top" | ! scope="row" | | Quebec Ontario — from , passing just west of Ottawa (at ) |-valign="top" | ! scope="row" | | New York Pennsylvania — from Maryland — from |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Chesapeake Bay | style="background:#b0e0e6;" | |
https://en.wikipedia.org/wiki/Category%20of%20relations
In mathematics, the category Rel has the class of sets as objects and binary relations as morphisms. A morphism (or arrow) R : A → B in this category is a relation between the sets A and B, so . The composition of two relations R: A → B and S: B → C is given by (a, c) ∈ S o R ⇔ for some b ∈ B, (a, b) ∈ R and (b, c) ∈ S. Rel has also been called the "category of correspondences of sets". Properties The category Rel has the category of sets Set as a (wide) subcategory, where the arrow in Set corresponds to the relation defined by . A morphism in Rel is a relation, and the corresponding morphism in the opposite category to Rel has arrows reversed, so it is the converse relation. Thus Rel contains its opposite and is self-dual. The involution represented by taking the converse relation provides the dagger to make Rel a dagger category. The category has two functors into itself given by the hom functor: A binary relation R ⊆ A × B and its transpose RT ⊆ B × A may be composed either as R RT or as RT R. The first composition results in a homogeneous relation on A and the second is on B. Since the images of these hom functors are in Rel itself, in this case hom is an internal hom functor. With its internal hom functor, Rel is a closed category, and furthermore a dagger compact category. The category Rel can be obtained from the category Set as the Kleisli category for the monad whose functor corresponds to power set, interpreted as a covariant functor. Perhaps a bit surprising at first sight is the fact that product in Rel is given by the disjoint union (rather than the cartesian product as it is in Set), and so is the coproduct. Rel is monoidal closed, if one defines both the monoidal product A ⊗ B and the internal hom A ⇒ B by the cartesian product of sets. It is also a monoidal category if one defines the monoidal product by the disjoint union of sets. The category Rel was the prototype for the algebraic structure called an allegory by Peter J. Freyd and An
https://en.wikipedia.org/wiki/History%20of%20trauma%20and%20orthopaedics
The treatment of broken bones and dislocated joints can be traced as far back as the Ancient Greeks. Hippocrates is credited with a method of reduction of a dislocated shoulder. 16th century Spanish texts talk about the Aztecs use of reduction of fractures using fir branches. The modern discipline of orthopaedics in trauma care developed during the course of World War I, but it was not until after World War II that orthopaedics became the dominant field treating fractures in much of the world. Today, the discipline encompasses conditions such as bone fractures and bone loss, as well as spinal pathology and joint disease. Greek and Roman Several volumes of the Hippocratic Corpus, Articulations or On Joints, On Fractures, On the Instruments of Reduction discuss Ancient Greek medicine relating to orthopaedics, and Hippocrates is credited with a method of reduction of a dislocated shoulder. Renaissance Period 16th century Spanish texts talk about the Aztecs use of reduction of fractures, as well intramedullary fixation using fir branches. Peter Lowe was the first surgeon to use the term amputation in his 1597 book A discourse of the Whole Art of Chirurgerie. Nicolas Andry has been credited with the term 'orthopaedics', taken from the title of his 1741 book Orthopédie on childhood deformity correction. The frontispiece of the book bore an engraving of a sapling being splinted with a stake, a symbol now referred to as the Tree of Andry and adopted by many orthopaedic associations internationally. In 1768, Percivall Pott published his book Some Few Remarks upon Fractures and Dislocations following his compound femoral fracture on the use of splinting to avoid amputation. Pott's student, John Hunter, expanded on the knowledge of bone healing. Around the same time, Jean-André Venel published his work Orthopaedia, or the Art of Preventing and Correcting Deformities in Children, one of the first surgeons to discuss the practical application for treating congenital de
https://en.wikipedia.org/wiki/Fish
A fish (: fish or fishes) is an aquatic, craniate, gill-bearing animal that lacks limbs with digits. Included in this definition are the living hagfish, lampreys, and cartilaginous and bony fish as well as various extinct related groups. Approximately 95% of living fish species are ray-finned fish, belonging to the class Actinopterygii, with around 99% of those being teleosts. The earliest organisms that can be classified as fish were soft-bodied chordates that first appeared during the Cambrian period. Although they lacked a true spine, they possessed notochords which allowed them to be more agile than their invertebrate counterparts. Fish would continue to evolve through the Paleozoic era, diversifying into a wide variety of forms. Many fish of the Paleozoic developed external armor that protected them from predators. The first fish with jaws appeared in the Silurian period, after which many (such as sharks) became formidable marine predators rather than just the prey of arthropods. Most fish are ectothermic ("cold-blooded"), allowing their body temperatures to vary as ambient temperatures change, though some of the large active swimmers like white shark and tuna can hold a higher core temperature. Fish can acoustically communicate with each other, most often in the context of feeding, aggression or courtship. Fish are abundant in most bodies of water. They can be found in nearly all aquatic environments, from high mountain streams (e.g., char and gudgeon) to the abyssal and even hadal depths of the deepest oceans (e.g., cusk-eels and snailfish), although no species has yet been documented in the deepest 25% of the ocean. With 34,300 described species, fish exhibit greater species diversity than any other group of vertebrates. Fish are an important resource for humans worldwide, especially as food. Commercial and subsistence fishers hunt fish in wild fisheries or farm them in ponds or in cages in the ocean (in aquaculture). They are also caught by recreational
https://en.wikipedia.org/wiki/Thymidylate%20kinase
Thymidylate kinase (; dTMP kinase) catalyzes the phosphorylation of thymidine 5'-monophosphate (dTMP) to form thymidine 5'-diphosphate (dTDP) in the presence of ATP and magnesium: ATP + thymidine 5'-phosphate ADP + thymidine 5'-diphosphate Thymidylate kinase is a ubiquitous enzyme of about 25 Kd and is important in the dTTP synthesis pathway for DNA synthesis. The function of dTMP kinase in eukaryotes comes from the study of a cell cycle mutant, cdc8, in Saccharomyces cerevisiae. Structural and functional analyses suggest that the cDNA codes for authentic human dTMP kinase. The mRNA levels and enzyme activities corresponded to cell cycle progression and cell growth stages. Thymidylate kinase's subfamily is predicted thymidylate kinase, TKRP1. Human protein DTYMK contains this domain. Structural studies As of late 2007, 40 structures have been solved for this class of enzymes, with PDB accession codes , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , and . See also Thymidine kinase Thymidylate synthase Thymidine kinase in clinical chemistry
https://en.wikipedia.org/wiki/Magic%20bullet%20%28medicine%29
The magic bullet is a scientific concept developed by a German Nobel laureate Paul Ehrlich in 1907. While working at the Institute of Experimental Therapy (Institut für experimentelle Therapie), Ehrlich formed an idea that it could be possible to kill specific microbes (such as bacteria), which cause diseases in the body, without harming the body itself. He named the hypothetical agent as Zauberkugel, and used the English translation "magic bullet" in The Harben Lectures at London. The name itself is a reference to an old German myth about a bullet that cannot miss its target. Ehrlich had in mind Carl Maria von Weber's popular 1821 opera Der Freischütz, in which a young hunter is required to hit an impossible target in order to marry his bride. Ehrlich envisioned that just like a bullet fired from a gun to hit a specific target, there could be a way to specifically target invading microbes. His continued research to discover the magic bullet resulted in further knowledge of the functions of the body's immune system, and in the development of Salvarsan, the first effective drug for syphilis, in 1909. His works were the foundation of immunology, and for his contributions he shared the 1908 Nobel Prize in Physiology or Medicine with Élie Metchnikoff. Ehrlich's discovery of Salvarsan in 1909 for the treatment of syphilis is termed as the first magic bullet. This led to the foundation of the concept of chemotherapy. Background Research on antibody In the early 1890s, Paul Ehrlich started to work with Emil Behring, professor of medicine at the University of Marburg. Behring had been investigating antibacterial agents and discovered a diphtheria antitoxin. (For that discovery, Bering was the first recipient of the Nobel Prize in Physiology or Medicine in 1901. Ehrlich was also nominated for that year.) From Behring's work, Ehrlich understood that antibodies produced in the blood could attack invading pathogens without any harmful effect on the body. He speculated that
https://en.wikipedia.org/wiki/Colloidal%20gold%20protein%20assay
The colloidal gold protein assay is a highly sensitive biochemical assay for determining the total concentration of protein in a solution (~0.1 ng/µL to 200 ng/µL). It was first described in 1987 by two groups who used commercially available "Aurodye" colloidal gold solutions. Notably, the formulation of Aurodye changed between 1987 and 1990 such that it became incompatible with protein assays, however vendors such as Bio-Rad & Diversified Biotech starting offering colloidal gold formulations that were suitable for protein assays. These products have since been discontinued and there are no vendors that currently explicitly sell colloidal gold for the assay, however detailed synthetic procedures were published to produce the ~17-40 nm gold nanoparticles that are suitable for the assay, along with modifications to increase the shelf stability of the colloidal gold & adapt the assay to microplate format & increase it's sensitivity. Gold nanoparticles in the ~17-40 nm size range that are presumably compatible with the assay are currently commercially available. Mechanism The total protein concentration is readout by an increase in absorbance at 565 nm, which can then be measured using colorimetric techniques, including using microplate readers. Most common reagents, except thiols and SDS, are compatible with the assay. An optimized formulation for the assay to maximize sensitivity in microplate format was described. Comparison to other assays While the colloidal gold assay is the most sensitive in-solution colorimetric protein assay, it may be equally sensitive or surpassed in sensitivity by fluorescent protein assays such as the CBQCA, FQ, NanoOrange, Quant-iT, and EZQ assays. See also Colloidal gold Bradford assay BCA assay
https://en.wikipedia.org/wiki/PHI-base
https://canto.phi-base.org/ The Pathogen-Host Interactions database (PHI-base) is a biological database that contains curated information on genes experimentally proven to affect the outcome of pathogen-host interactions. The database is maintained by researchers at Rothamsted Research, together with external collaborators since 2005. Since April 2017 PHI-base is part of ELIXIR, the European life-science infrastructure for biological information via its ELIXIR-UK node. Background The Pathogen-Host Interactions database was developed to utilise effectively the growing number of verified genes that mediate an organism's ability to cause disease and / or to trigger host responses. The web-accessible database catalogues experimentally verified pathogenicity, virulence and effector genes from bacterial, fungal and oomycete pathogens which infect animal, plant and fungal hosts. PHI-base is the first on-line resource devoted to the identification and presentation of information on fungal and oomycete pathogenicity genes and their host interactions. As such, PHI-base aims to be a resource for the discovery of candidate targets in medically and agronomically important fungal and oomycete pathogens for intervention with synthetic chemistries and natural products (fungicides). Each entry in PHI-base is curated by domain experts and supported by strong experimental evidence (gene disruption experiments) as well as literature references in which the experiments are described. Each gene in PHI-base is presented with its nucleotide and deduced amino acid sequence as well as a detailed structured description of the predicted protein's function during the host infection process. To facilitate data interoperability, genes are annotated using controlled vocabularies (Gene Ontology terms, EC Numbers, etc.), and links to other external data sources such as UniProt, EMBL and the NCBI taxonomy services. Current developments Version 4.15 (May 2, 2023) of PHI-base provides informa
https://en.wikipedia.org/wiki/International%20Conference%20on%20Cold%20Fusion
The International Conference on Cold Fusion (ICCF) (also referred to as Annual Conference on Cold Fusion in 1990-1991 and mostly as International Conference on Condensed Matter Nuclear Science since 2007) is an annual or biennial conference on the topic of cold fusion. An international conference on cold fusion was held in Santa Fe, New Mexico US in 1989. However, the first ICCF conference (ICCF1) took place in 1990 in Salt Lake City, Utah, under the title "First Annual Conference on Cold Fusion". Its location has since rotated between Russia, US, Europe, and Asia. It was held in India for the first time in 2011. The conferences have been criticized as events which attract "crackpots" and "pseudo-scientists". Reception The First Annual Conference on Cold Fusion was held in March 1990 in Salt Lake City, Utah, United States. Robert L. Park of the American Physical Society derisively referred to it as a "seance of true believers." The conference was attended by more than 200 researchers from the United States, Italy, Japan, India and Taiwan and dozens of reporters from all over the U.S. and abroad. The Third International Conference on Cold Fusion was held in 1992 in Nagoya, Japan. It was described by The New York Times, "depending on one's point of view" as "either a turning point in which evidence was presented that will convince the skeptics that cold fusion exists or a religious revival where claims of miracles were lapped up by ardent believers." The conference was sponsored by seven Japanese scientific societies, it was attended by 200 Japanese scientists and more than 100 from abroad. Tomohiro Taniguchi, then director of the Electric Power Technology Division at Japan's Ministry of International Trade and Industry, reportedly said that the Ministry of International Trade and Industry was willing to finance research in the field in view of "encouraging evidence, especially after the conference." The conference was also covered by the Associated Press. A jour
https://en.wikipedia.org/wiki/Windows%20NT%203.5
Windows NT 3.5 is a major release of the Windows NT operating system developed by Microsoft and oriented towards businesses. It was released on September 21, 1994, as the successor to Windows NT 3.1 and the predecessor to Windows NT 3.51. One of the primary goals during Windows NT 3.5 development was to improve the operating system's performance. As a result, the project was codenamed "Daytona", after the Daytona International Speedway in Daytona Beach, Florida. On December 31, 2001, Microsoft declared Windows NT 3.5 obsolete and stopped providing support and updates for the system. Features Windows NT 3.5 comes in two editions: NT Workstation and NT Server. They respectively replace the NT and NT Advanced Server editions of Windows NT 3.1. The Workstation edition allows only 10 concurrent clients to access the file server and does not support Mac clients. Windows NT 3.5 includes integrated Winsock and TCP/IP support. (Its predecessor, Windows NT 3.1, only includes an incomplete implementation of TCP/IP based on the AT&T UNIX System V "STREAMS" API.) TCP/IP and IPX/SPX stacks in Windows NT 3.5 are rewritten. NetBIOS over TCP/IP (NetBT) support as a compatibility layer for TCP/IP was introduced as also the Microsoft DHCP and WINS clients and DHCP and WINS servers. Windows NT 3.5 can share files via the File Transfer Protocol, and printers through the Line Printer Daemon protocol. It can act as a Gopher, HTTP, or WAIS server, and includes Remote Access Service for remote dial-up modem access to LAN services using either SLIP or PPP protocols. Windows NT 3.5 Resource Kit includes the first implementation of Microsoft DNS. Other new features in Windows NT 3.5 include support for the VFAT file system, Object Linking and Embedding (OLE) version 2.0 and support for input/output completion ports. Microsoft updated the graphical user interface to be consistent with that of Windows for Workgroups 3.11. NT 3.5 shows performance improvements over NT 3.1, and requires less
https://en.wikipedia.org/wiki/Heishansaurus
Heishansaurus, meaning "Heishan lizard" after the area in China where it was discovered, is the name given to a dubious genus of herbivorous ornithischian dinosaur, probably belonging to the Ankylosauridae. In 1930, Swedish palaeontologist Anders Birger Bohlin discovered dinosaur fossils, in the context of the Swedish-Chinese expeditions headed by Sven Hedin, near Jiayuguan ("Chia-Yu-Kuan"), in the west of Gansu Province. In 1953, Bohlin named these as the type species Heishansaurus pachycephalus. The generic name refers to the Heishan, the "Black Mountains". The specific name pachycephalus, meaning "thick-headed", was inspired by Bohlin's identification of the taxon as a pachycephalosaur. Today this dinosaur is more probably considered an ankylosaur. The fossils, from the Minhe Formation dating from the Late Cretaceous (Campanian or Maastrichtian stage), were fragmentary. The type is the only known specimen. The material consisted of poorly preserved cranial and postcranial fragments plus some dermal scutes. It contained skull fragments including a maxilla, teeth, vertebrae from the neck, back and tail, osteoderms and spikes. Today, the specimen is lost. Of one dorsal vertebra a cast remains, preserved in the American Museum of Natural History with the inventory number AMNH 2062. Bohlin considered the species to be a member of the pachycephalosaurians because he mistook an osteoderm for the thick skull roof typical of this group. The material is probably ankylosaurid. It has been seen as a junior synonym of Pinacosaurus but the genus is more generally considered a nomen dubium, especially since Bohlin's description can only be checked by comparison with his published drawings. Notes
https://en.wikipedia.org/wiki/Computer-aided%20lean%20management
Computer-aided lean management, in business management, is a methodology of developing and using software-controlled, lean systems integration. Its goal is to drive innovation towards cost and cycle-time savings. It attempts to create an efficient use of capital and resources through the development and use of one integrated system model to run a business's planning, engineering, design, maintenance, and operations. Overview Computer-Aided Lean Management (CALM) is a management philosophy that uses computational software to reduce risk and inefficiencies. CALM acts on uncertainties and business inefficiencies to increase profitability through the use of computational decision-making tools that enable opportunities for additional value creation. It is based on the application of software to enable continuous improvement through an Integrated System Model (ISM) of the business’s physical assets, business processes, and machine learning. This unique integration of software applications using lean principles was developed in the aerospace industry and has migrated to the energy industry. The creation of an integrated system model removes the barriers posed by the silos or stovepipes inherent in the departmentalization of most companies. Integration enables lean uses of information for the creation of actionable knowledge. CALM strives to create such a lean management approach to running the company through the rigors of software enforcement. From this software enforcement comes clear policy and procedures that are adhered to, activity-based costing, measurement of effectiveness, and the capability of using advanced algorithms for dramatic improvements in optimization of resources. CALM creates business capabilities through software to enable technology application, streamlining of processes, and a lean organizational structure. The methodology is based on a commonsense approach for running a business, by measuring of actions taken and using those measurem
https://en.wikipedia.org/wiki/Pitfall%20trap
A pitfall trap is a trapping pit for small animals, such as insects, amphibians and reptiles. Pitfall traps are a sampling technique, mainly used for ecology studies and ecologic pest control. Animals that enter a pitfall trap are unable to escape. This is a form of passive collection, as opposed to active collection where the collector catches each animal (by hand or with a device such as a butterfly net). Active collection may be difficult or time-consuming, especially in habitats where it is hard to see the animals such as thick grass. Structure and composition Pitfall traps come in a variety of sizes and designs. They come in two main forms: dry and wet pitfall traps. Dry pitfall traps consist of a container (tin, jar or drum) buried in the ground with its rim at surface level used to trap mobile animals that fall into it. Wet pitfall traps are basically the same, but contain a solution designed to kill and preserve the trapped animals. The fluids that can be used in these traps include formalin (10% formaldehyde), methylated spirits, alcohol, ethylene glycol, trisodium phosphate, picric acid or even (with daily checked traps) plain water. A little detergent is usually added to break the surface tension of the liquid to promote quick drowning. The opening is usually covered by a sloped stone or lid or some other object. This is done to reduce the amount of rain and debris entering the trap, and to prevent animals in dry traps from drowning (when it rains) or overheating (during the day) as well as to keep out predators. One or more fence-lines of some sort may be added to channel targets into the trap. Traps may also be baited. Lures or baits of varying specificity can be used to increase the capture rate of a certain target species or group by placing them in, above or near the trap. Examples of baits include meat, dung, fruit and pheromones. Uses Pitfall traps can be used for various purposes: During the mating seasons of toads, frogs and salamanders in t
https://en.wikipedia.org/wiki/Agaricus%20macrosporus
Agaricus macrosporus is a rare, edible mushroom found from June at wood fringes and in meadows. Description The white cap is hemispherical and white when young, but later flattens out up to 25 cm in diameter and becomes yellowish or tan. Its flesh is very thick. The gills are pinkish grey when young, and become brown with age. The spores measure 12 by 6 μm and are purplish-brown and almond-shaped. The stem is strong and thick, with a broad ring. It may measure 8 to 12 cm in height and up to 3 cm in diameter. The flesh is white with a mild taste and a smell of aniseed, turning slowly orange when cut. Similar species There is a danger of confusing this mushroom with dangerous amanitas such as Amanita phalloides and Amanita pantherina. Agaricus excellens is different by its taller and slimmer stipe which is striped lengthwise. Agaricus augustus does not have the pure white cap in young specimens. See also List of Agaricus species
https://en.wikipedia.org/wiki/OutGuess
OutGuess is a steganographic software for hiding data in the most redundant content data bits of existing (media) files. It has handlers for image files in the common Netpbm and JPEG formats, so it can, for example, specifically alter the frequency coefficients of JPEG files. It is written in C and published as Free Software under the terms of the old BSD license. It has been tested on a variety of Unix-like operating systems and is included in the standard software repositories of the popular Linux distributions Debian and Arch Linux (via user repository) and their derivatives. Method of operation An algorithm estimates the capacity for hidden data without the distortions of the decoy data becoming apparent. OutGuess determines bits in the decoy data that it considers most expendable and then distributes secret bits based on a shared secret in a pseudorandom pattern across these redundant bits, flipping some of them according to the secret data. For JPEG images, OutGuess recompresses the image to a user-selected quality level and then embeds secret bits into the least significant bits (LSB) of the quantized coefficients while skipping zeros and ones. Subsequently, corrections are made to the coefficients to make the global histogram of discrete cosine transform (DCT) coefficients match that of the decoy image, counteracting detection by the chi-square attack that is based on the analysis of first-order statistics. This technique is criticized because it actually facilitates detection by further disturbing other statistics. Also, data embedded in JPEG frequency coefficients has poor robustness and does not withstand JPEG reencoding. History OutGuess was originally developed in Germany in 1999 by Niels Provos. In 1999, Andreas Westfeld published the statistical chi-square attack, which can detect common methods for steganographically hiding messages in LSBs of quantized JPEG coefficients. In response, Provos implemented a method that exactly preserves the DCT histo
https://en.wikipedia.org/wiki/Ovorubin
Ovorubin (PcOvo or PcPV1) is the most abundant perivitellin (>60 % total protein) of the perivitelline fluid from Pomacea canaliculata snail eggs. This glyco-lipo-caroteno protein complex is a approx. 300 kDa multimer of a combination of multiple copies of six different ~30 kDa subunits. Together with the other perivitellins from Pomacea canaliculata eggs, ovorubin serves a nutrient source for developing embryos, notably to the intermediate and late stages. Moreover, after hatching, the protein is still detected in the lumen of the digestive gland ready to be endocytosed, therefore, acting as a nutrient source for the newly hatched snail. Ovorubin contains carbohydrates and carotenoid pigments as main prosthetic groups, which are related to many physiological roles on Pomacea aerial egg-laying strategy. Given that carbohydrates tend to retain water, the high glycosylation of ovorubin (~17 % w/w) was proposed as an embryo defense against water loss. The carotenoid pigments stabilized by ovorubin also provide the eggs of antioxidant and photoprotective capacities, crucial roles to cope with the harsh conditions of the aerial environment. The presence of carotenoid pigments is also responsible for the brightly reddish coloration of Ovorubin, and therefore snail eggs, which was related to a warning coloration (aposematism) advertising predators about the presence of deterrents. In fact, field evidence of egg unpalatability is provided by the fact that most animals foraging in habitats where the apple snails live ignore these eggs. Like most other studied perivitellins from Pomacea snails, ovorubin is highly stable in a wide range of pH values and withstands gastrointestinal digestion, characteristics associated with an antinutritive defense system that deters predation by lowering the nutritional value of the eggs.
https://en.wikipedia.org/wiki/SETIcon
SETIcon was a public convention organized by the SETI Institute and held twice in Santa Clara, California. It was an interdisciplinary conference, with talks focusing on a wide range of issues related to the search for extraterrestrial intelligence (SETI). The programs included panels with prominent scientists from NASA and SETI; talks by science fiction writers, artists, and actors; and panels to explore controversial issues and compare perspectives. The sessions were non-technical and meant to increase awareness, interest, and funding for the SETI Institute. The convention attracted professional scientists, educators, space enthusiasts, and science fiction fans. Astronomical tattoos were not uncommon among attendees. The first SETIcon took place on 13–15 August 2010 and drew about 1,000 people. It commemorated the 25th anniversary of the SETI Institute, the 50th anniversary of SETI, and the 80th birthday of Frank Drake, the radio astronomer who started Project Ozma and pioneered the SETI program. Speakers included scientists Frank Drake, Alex Filippenko, Seth Shostak, Andrew Fraknoi, David Morrison and Jill Tarter; author Robert J. Sawyer; professional skeptic Phil Plait; graphic artist Paul Duffield; actor Tim Russ; and musician Mickey Hart. During one of the panels, Seth Shostak said he expected proof of extraterrestrial life to be found within 25 years. "Young people in the audience, I think there's a really good chance you're going to see this happen." SETIcon II was held on 22–24 June 2012 and drew about 600 people and 60 speakers, including Bill Nye. Recent findings from NASA's Kepler mission fueled much of the discussion—between 2009 and 2012 the experiment detected about 2,300 new exoplanets, making the discovery of intelligent extraterrestrial life increasingly plausible. Among the returning speakers was Alex Filippenko, who during a panel called "Did the Big Bang Require a Divine Spark?" argued against the necessity of God as the first cause of the Big
https://en.wikipedia.org/wiki/Milne-Thomson%20circle%20theorem
In fluid dynamics the Milne-Thomson circle theorem or the circle theorem is a statement giving a new stream function for a fluid flow when a cylinder is placed into that flow. It was named after the English mathematician L. M. Milne-Thomson. Let be the complex potential for a fluid flow, where all singularities of lie in . If a circle is placed into that flow, the complex potential for the new flow is given by with same singularities as in and is a streamline. On the circle , , therefore Example Consider a uniform irrotational flow with velocity flowing in the positive direction and place an infinitely long cylinder of radius in the flow with the center of the cylinder at the origin. Then , hence using circle theorem, represents the complex potential of uniform flow over a cylinder. See also Potential flow Conformal mapping Velocity potential Milne-Thomson method for finding a holomorphic function
https://en.wikipedia.org/wiki/Symbolic%20trajectory%20evaluation
Symbolic trajectory evaluation (STE) is a lattice-based model checking technology that uses a form of symbolic simulation. STE is essentially used for computer hardware, that is circuit verification. The technique uses abstraction, meaning that details of the circuit behaviour are removed from the circuit model. It was first developed by Carl Seger and Randy Bryant in 1995 as an alternative to "classical" symbolic model checking.
https://en.wikipedia.org/wiki/Wet%20wipe
A wet wipe, also known as a wet towel, moist towelette, disposable wipe, disinfecting wipe, or a baby wipe (in specific circumstances) is a small to medium-sized moistened piece of plastic or cloth that either comes folded and individually wrapped for convenience or, in the case of dispensers, as a large roll with individual wipes that can be torn off. Wet wipes are used for cleaning purposes like personal hygiene and household cleaning; each is a separate product depending on the chemicals added and medical or office cleaning wipes are not intended for skin hygiene. In 2013, owing to increasing sales of the product in affluent countries, Consumer Reports reported that efforts to make the wipes "flushable" down the toilet had not entirely succeeded, according to their test. Invention American Arthur Julius is seen as the inventor of wet wipes. Julius worked in the cosmetics industry and in 1957, adjusted a soap portioning machine, putting it in a loft in Manhattan. Julius trademarked the name Wet-Nap in 1958, a name for the product that is still being used. After fine tuning his new hand-cleaning aid together with a mechanic, he unveiled his invention at the 1960 National Restaurant Show in Chicago and in 1963 started selling Wet-Nap products to Colonel Harland Sanders to be distributed to customers of Kentucky Fried Chicken. Production Ninety percent of wet wipes on the market are produced from nonwoven fabrics made of polyester or polypropylene. The material is moistened with water or other liquids (e.g., isopropyl alcohol) depending on the applications. The material may be treated with softeners, lotions, or perfume to adjust the tactile and olfactory properties. Preservatives such as methylisothiazolinone are used to prevent bacterial or fungal growth in the package. The finished wet wipes are folded and put in pocket size package or a box dispenser. Uses Wet wipes can serve a number of personal and household purposes. Although marketed primarily for wipi
https://en.wikipedia.org/wiki/Ginseng
Ginseng () is the root of plants in the genus Panax, such as Korean ginseng (P. ginseng), South China ginseng (P. notoginseng), and American ginseng (P. quinquefolius), characterized by the presence of ginsenosides and gintonin. Ginseng is common in the cuisines and medicines of China and Korea. Ginseng has been used in traditional medicine over centuries, though modern clinical research is inconclusive about its medical effectiveness. There is no substantial evidence that ginseng is effective for treating any medical condition and it has not been approved by the US Food and Drug Administration (FDA) as a prescription drug. Although ginseng is sold as a dietary supplement, inconsistent manufacturing practices for supplements have led to analyses of some ginseng products contaminated with toxic metals or unrelated filler compounds, and its excessive use may have adverse effects or untoward interactions with prescription drugs. History One of the first written texts covering the use of ginseng as a medicinal herb was the Shen Nong Pharmacopoeia, written in China in 196 AD. In his Compendium of Materia Medica herbal of 1596, Li Shizhen described ginseng as a "superior tonic". However, the herb was not used as a "cure-all" medicine, but more specifically as a tonic for patients with chronic illnesses and those who were convalescing. Control over ginseng fields in China and Korea became an issue in the 16th century. Ginseng species Ginseng plants belong only to the genus Panax. Cultivated species include Panax ginseng (Korean ginseng), Panax notoginseng (South China ginseng), Panax pseudoginseng (Himalayan ginseng), Panax quinquefolius (American ginseng), Panax trifolius (Dwarf ginseng), and Panax vietnamensis (Vietnamese ginseng). Ginseng is found in cooler climates – Korean Peninsula, Northeast China, Russian Far East, Canada and the United States, although some species grow in warm regions – South China ginseng being native to Southwest China and Vietnam. Panax
https://en.wikipedia.org/wiki/Internet%20rush%20hour
Internet rush hour is the time period when the majority of Internet users are online at the same time. Typically, in the UK the peak hours are between 7 and 11 pm. During this time frame, users commonly experience slowness while browsing or downloading content. The congestion experienced during the rush hour is similar to transportation rush hour, where demand for resources outweighs capacity. In contrast to the hours cited above from a source dated 2011, a Google Analytics report dated 2017 indicates very strongly that daily web use peaks between 9am and midday, falling off steadily throughout the day with a modest levelling off between 7pm and 10pm, and then collapses to a base at 4am. Reasons Growth Global Internet usage has increased significantly across the world from 12,881,000 hosts in 1998 to 908,585,739 hosts in 2012. Internet use has surged with the introduction of mobile devices and tablet computing. Internet access has also changed during this time frame from 56kbit/s dial-up to high-speed bandwidth access at 100Mbit/s or higher. The increase in Internet users and increase in access bandwidth is a contributing factor to the Internet Rush Hour. The table below shows the big picture of world internet usage versus the population. Infrastructure End users connect to the Internet through Internet service providers (ISPs). The Tier 1 ISPs own the infrastructure, which includes routers, switches and fiber optic footprints. The back bone of the Internet is connected through Tier 1 ISPs that peer with other Tier 1 ISPs in a transit-free network. These peering agreements between Tier 1 ISPs have no overt settlements, meaning there is no money exchanged for the right to pass traffic between Tier 1 peers. Tier 2 and Tier 3 ISPs are customers of the Tier 1 ISPs and rely on the Tier 1 ISPs to route their traffic across the Internet. This is a disadvantage for the lower tier ISPs due to the amount of traffic hops and shared common gateways to Tier 1 ISPs.
https://en.wikipedia.org/wiki/L%C3%AA%20V%C4%83n%20%C4%90%E1%BB%87
Celso-Léon Lê Văn Đệ (24 August 1906 – 16 March 1966) was a South Vietnamese painter who designed the South Vietnamese flag. Early life and education A Roman Catholic, he was born in Mỏ Cày, Bến Tre, and was in charge of Asian Arts at the International Exhibition of Catholic Press at Vatican in 1936. He established the École supérieure nationale des Beaux Arts in Saigon. Career Some of his works have been shown at the Vietnam National Museum of Fine Arts, Hanoi. Others are exhibited in Europe. A wall panel, Mater Amabilis and Saint Magdalene at the food of the cross was on show at the Missionaria Arte museum at the Vatican. In 1930, he received a scholarship from the “Society for the Intellectual and Physical Improvement of the Indigenous People of Cochinchina” to study at the Beaux-Arts in Paris as a student of Jean-Pierre Laurens (1875-1932). The latter had welcomed two years earlier Fang Ganmin (1906-1984) who returned to China to become an influential teacher (notably of Zao Wou-Ki, Chu Teh Chun and Wu Guanzhong). He was in Paris when Fang fell in love with Sue Ailan (1905-1985) whom he met at the Beaux-Arts in the Humbert studio (the only one authorized to receive female students) and whom he would marry on his return to China. Lé Van Dé was noticed in 1932 and 1933 during the Salon Officiel des Artistes Français (Paris) and obtained an “honorable mention”. In 1934, the French government bought from him a large canvas (179 X 225 cm) “The family interior in Tonkin” which he exhibited at the International Exhibition of Colonial Art in Naples. He died in Saigon.
https://en.wikipedia.org/wiki/List%20of%20parties%20to%20the%20Biological%20Weapons%20Convention
The list of parties to the Biological Weapons Convention encompasses the states which have signed and ratified or acceded to the Biological Weapons Convention (BWC), a multilateral treaty outlawing biological weapons. On 10 April 1972, the BWC was opened for signature. The Netherlands became the first state to deposit their signature of the treaty that same day. The treaty closed for signature upon coming into force on 26 March 1975 with the deposit of ratification by 22 states. Since then, states that did not sign the BWC can only accede to it. A total of 197 states may become members of the BWC, including all 193 United Nations member states, the Cook Islands, the Holy See, the State of Palestine and Niue. As of February 2023, 185 states have ratified or acceded to the treaty, most recently South Sudan in February 2023. As well, the Republic of China (Taiwan), which is currently only recognized by , deposited their instruments of ratification of the BWC with the United States government prior to the US's decision to switch their recognition of the sole legitimate government of China from the Republic of China (ROC) to the People's Republic of China (PRC). A further five states have signed but not ratified the treaty. Several countries made reservations when ratifying the agreement declaring that it did not imply their complete satisfaction that the BWC allows the stockpiling of biological agents and toxins for "prophylactic, protective or other peaceful purposes", nor should it imply recognition of other countries they do not recognise. States Parties According to the treaties database maintained by the United Nations Office for Disarmament Affairs, as of February 2023, 185 states have ratified or acceded to the BWC. However, the status of the succession of a number of additional states to the BWC is unclear. For further details, see the Succession of colonies to the BWC section below. Multiple dates indicate the different days in which states submitted thei
https://en.wikipedia.org/wiki/List%20of%20desktop%20publishing%20software
The following is a list of major desktop publishing software. A wide range of related software tools exist in this field including many plug-ins and tools related to the applications listed below. Several software directories provide more comprehensive listings of desktop publishing software, including VersionTracker and Tucows. Free software This section lists free software which does desktop publishing. All of these are required to be open-source. While not required, the software listed in this section is available free of charge. (In principle, in rare cases, free software is sold without being distributed over the Internet.) Desktop publishing software for Windows, macOS, Linux and other operating systems Collabora Online Draw and Collabora Online Writer. The applications for Windows, macOS, Linux and ChromeOS are also known as Collabora Office. LibreOffice Draw and LibreOffice Writer for Windows, macOS, Linux, BSDs and others LyX for Windows, MacOS, Linux, UNIX, OS/2 and Haiku, based on the LaTeX typesetting system, initial release in 1995 Scribus for Windows, macOS, Linux, BSD, Unix, Haiku, OS/2, based on the free Qt toolkit, initial release in 2003 Online desktop publishing software Collabora Online Draw and Collabora Online Writer Scenari, open source single-source publishing tool with support for chain publication Proprietary Desktop publishing software for Windows XEditpro Automated Publishing Tool - DiacriTech, 1997 Adobe InDesign Adobe FrameMaker Adobe PageMaker, discontinued in 2004 Affinity Publisher CatBase Calamus CorelDRAW Corel Ventura, previously Ventura Publisher, originally developed by Xerox, now owned by Corel FrameMaker, now owned by Adobe InPage - DTP which works with English + Urdu, Arabic, Persian, Pashto etc. MadCap Flare Microsoft Publisher PageStream, formerly known as Publishing Partner Prince XML, by YesLogic QuarkXPress RagTime Ready, Set, Go! Xara Designer Pro X Xara Page & Layout Designer Deskt
https://en.wikipedia.org/wiki/Set%20Theory%3A%20An%20Introduction%20to%20Independence%20Proofs
Set Theory: An Introduction to Independence Proofs is a textbook and reference work in set theory by Kenneth Kunen. It starts from basic notions, including the ZFC axioms, and quickly develops combinatorial notions such as trees, Suslin's problem, ◊, and Martin's axiom. It develops some basic model theory (rather specifically aimed at models of set theory) and the theory of Gödel's constructible universe L. The book then proceeds to describe the method of forcing. Kunen completely rewrote the book for the 2011 edition (under the title "Set Theory"), including more model theory.
https://en.wikipedia.org/wiki/Polytolypa
Polytolypa is a monotypic genus of fungus containing the single species Polytolypa hystricis. First classified in the Onygenaceae family, as of 2008 it is considered to be in the Ajellomycetaceae, although there is still uncertainty as to its phylogenetic relationships with other similar genera. This species is only known from a single specimen derived in the laboratory from a specimen of dung of the North American porcupine, Erethizon dorsatum, collected in Ontario, Canada. Polytolypa hystricis contains bioactive compounds that have antifungal activity. Taxonomy, phylogeny, and naming The genus was first described in 1993 by University of Toronto mycologists J.A. Scott and D.W. Malloch, who grew the fungus in moist chamber cultures of porcupine dung collected in Stoneleigh, Ontario, Canada. The generic name Polytolypa is from the Greek word poly (πολυ) meaning "many", and tolype (τολυπη), meaning "skein of yarn". The specific epithet hystricis comes from the Greek hystrix (υστριξ), or "porcupine". The genus has been classified in the Onygenaceae, a fungal family characterized by species capable of digesting human hair in vitro, and with spores that are punctate (with minute surface punctures) when viewed with scanning electron microscopy. However, as Scott and colleagues demonstrated using traditional laboratory tests to determine keratinolytic activity, P. hystricus is not able to digest hair. There is still uncertainty as to its phylogenetic relationships with other similar genera. Polytolypa is thought to be evolutionarily most closely related to the genera Malbranchea and Spiromastix. The grouping of Polytolypa and Spiromastix represent a sister clade to the Ajellomyces clade, based on analysis of partial nuclear LSU sequence data. However, the phylogenetics of Polytolypa are still unclear and await further study. The 10th edition of the Dictionary of the Fungi (2008) considers the genus to be in the Ajellomycetaceae family, although uncertainty with this cla
https://en.wikipedia.org/wiki/Neurotree
Academic Family Tree, which began as Neurotree, is an online database for academic genealogy, containing numerous "family trees" of academic disciplines. Neurotree was established in 2005 as a family tree of neuroscientists. Later that year Academic Family Tree incorporated Neurotree and academic genealogies of other scholarly disciplines. Unlike a conventional genealogy or family tree, in which connections among individuals are from kinship (e.g., parents to children), connections in Academic Family Tree are from mentoring relationships, usually among people working in academic settings (e.g., doctoral supervisors to students). Academic Family Tree has been used as sources of information for the history and prospects of academic fields such as psychology, meteorology, organizational communication, and neuroscience. Academic Family Tree has been used to address infometrics, to research issues of scientific methodology, and to examine mentor characteristics that predict mentee academic success. Functioning and scope Discipline-specific family trees of Academic Family Tree are volunteer-run; accuracy is maintained by a group of volunteer editors. Hierarchical connections between parents and children are defined as any meaningful mentoring relationship (research assistant, graduate student, postdoctoral fellow, or research scientist) between researchers. Continuous records extend well into the Middle Ages and earlier. As of 29 September 2023, Academic Family Tree contained 871,361 people with 882,278 connections among them. Academic Family Tree encompasses a broad range of discipline-specific trees. As of 29 September 2023, there were 73 trees spanning science (e.g., human genetics, microbiology, and psychology), mathematics and philosophy, engineering, the humanities (e.g., economics, law, theology, and music), and business (e.g., organizational communication and advertising). All trees within Academic Family Tree are closely linked. A search for a person in on
https://en.wikipedia.org/wiki/16%3A10%20aspect%20ratio
16:10 (1.6:1) is an aspect ratio commonly used for computer displays and tablet computers. It is equal to 8/5, close to the golden ratio (), which is approximately 1.618. History LCD computer displays with a 16:10 ratio first rose to mass market prominence in 2003. By 2008, the 16:10 aspect ratio had become the most common aspect ratio for LCD monitors and laptop displays. After 2010, however, 16:9 became the mainstream standard. This shift was driven by lower manufacturing costs and the 16:9 aspect ratio being used as a standard in modern televisions. Rise in popularity from 2003 Until about 2003, most computer monitors had a 4:3 aspect ratio, with some using a 5:4 ratio. Between 2003 and 2006, monitors with 16:10 aspect ratios became commonly available, first in laptops, and later in display monitors. Such displays were considered better suited for word processing and computer-aided design. From 2005 to 2008, 16:10 overtook 4:3 as the highest-selling aspect ratio for LCD monitors. At the time, 16:10 made up 90% of the notebook market, and was the most commonly used aspect ratio for laptops. However, 16:10 had a short reign as the most common aspect ratio. Decline from 2008 Around 2008–2010, computer display manufacturers began a rapid shift to the 16:9 aspect ratio. By 2011, 16:10 had almost disappeared from new mass-market products. By October 2012, the market share of 16:10 displays had dropped to less than 23%, according to Net Applications. The primary reason for this move was considered to be production efficiency: Since display panels for TVs use the 16:9 aspect ratio, it became more efficient for display manufacturers to produce computer display panels in the same aspect ratio. A 2008 report by DisplaySearch also cited several other reasons, including the ability for PC and monitor manufacturers to expand their product ranges by offering products with wider screens and higher resolutions. This helped consumers adopt such products more easily, "stimu
https://en.wikipedia.org/wiki/Construction%20of%20a%20complex%20null%20tetrad
Calculations in the Newman–Penrose (NP) formalism of general relativity normally begin with the construction of a complex null tetrad , where is a pair of real null vectors and is a pair of complex null vectors. These tetrad vectors respect the following normalization and metric conditions assuming the spacetime signature Only after the tetrad gets constructed can one move forward to compute the directional derivatives, spin coefficients, commutators, Weyl-NP scalars , Ricci-NP scalars and Maxwell-NP scalars and other quantities in NP formalism. There are three most commonly used methods to construct a complex null tetrad: All four tetrad vectors are nonholonomic combinations of orthonormal tetrads; (or ) are aligned with the outgoing (or ingoing) tangent vector field of null radial geodesics, while and are constructed via the nonholonomic method; A tetrad which is adapted to the spacetime structure from a 3+1 perspective, with its general form being assumed and tetrad functions therein to be solved. In the context below, it will be shown how these three methods work. Note: In addition to the convention employed in this article, the other one in use is . Nonholonomic tetrad The primary method to construct a complex null tetrad is via combinations of orthonormal bases. For a spacetime with an orthonormal tetrad , the covectors of the nonholonomic complex null tetrad can be constructed by and the tetrad vectors can be obtained by raising the indices of via the inverse metric . Remark: The nonholonomic construction is actually in accordance with the local light cone structure. Example: A nonholonomic tetrad Given a spacetime metric of the form (in signature(-,+,+,+)) the nonholonomic orthonormal covectors are therefore and the nonholonomic null covectors are therefore la (na) aligned with null radial geodesics In Minkowski spacetime, the nonholonomically constructed null vectors respectively match the outgoing and ingoing null radial ra
https://en.wikipedia.org/wiki/Half-metal
A half-metal is any substance that acts as a conductor to electrons of one spin orientation, but as an insulator or semiconductor to those of the opposite orientation. Although all half-metals are ferromagnetic (or ferrimagnetic), most ferromagnets are not half-metals. Many of the known examples of half-metals are oxides, sulfides, or Heusler alloys. Types of half-metallic compounds theoretically predicted so far include some Heusler alloys, such as Co2FeSi, NiMnSb, and PtMnSb; some Si-containing half–Heusler alloys with Curie temperatures over 600K, such as NiCrSi and PdCrSi; some transition-metal oxides, including rutile structured CrO2; some perovskites, such as LaMnO3 and SeMnO3; and a few more simply structured zincblende(ZB) compounds, including CrAs and superlattices. NiMnSb and CrO2 have been experimentally determined to be half-metals at very low temperatures. In half-metals, the valence band for one spin orientation is partially filled while there is a gap in the density of states for the other spin orientation. This results in conducting behavior for only electrons in the first spin orientation. In some half-metals, the majority spin channel is the conducting one while in others the minority channel is. Half-metals were first described in 1983, as an explanation for the electrical properties of manganese-based Heusler alloys. Some notable half-metals are chromium(IV) oxide, magnetite, and lanthanum strontium manganite (LSMO), as well as chromium arsenide. Half-metals have attracted some interest for their potential use in spintronics.
https://en.wikipedia.org/wiki/Transformation%20efficiency
Transformation efficiency refers to the ability of a cell to take up and incorporate exogenous DNA, such as plasmids, during a process called transformation. The efficiency of transformation is typically measured as the number of transformants (cells that have taken up the exogenous DNA) per microgram of DNA added to the cells. A higher transformation efficiency means that more cells are able to take up the DNA, and a lower efficiency means that fewer cells are able to do so. In molecular biology, transformation efficiency is a crucial parameter, it is used to evaluate the ability of different methods to introduce plasmid DNA into cells and to compare the efficiency of different plasmid, vectors and host cells. This efficiency can be affected by a number of factors, including the method used for introducing the DNA, the type of cell and plasmid used, and the conditions under which the transformation is performed. Therefore, measuring and optimizing transformation efficiency is an important step in many molecular biology applications, including genetic engineering, gene therapy and biotechnology. Measurement By measuring the transformation efficiency, we can utilize the information from our experiment to evaluate how effectively our transformation went. This is a quantification of how many cells were altered by 1 µg of plasmid DNA. In essence, it is a sign that the transformation experiment was successful. It should be determined under conditions of cell excess. Transformation efficiency is typically measured as the number of transformed cells per total number of cells. It can be represented as a percentage or as colony forming units (CFUs) per microgram of DNA. One of the most common ways to measure transformation efficiency is by performing a colony forming assay. Here is an example of how to calculate transformation efficiency using colony forming units (CFUs): Plate a known number of cells on agar plates containing the appropriate antibiotics. Incubate th
https://en.wikipedia.org/wiki/Archaic%20smile
The archaic smile was used by sculptors in Archaic Greece, especially in the second quarter of the 6th century BCE, possibly to suggest that their subject was alive and infused with a sense of well-being. One of the most famous examples of the archaic smile is the Kroisos Kouros, and the Peplos Kore is another. By the middle of the Archaic Period of ancient Greece (roughly 800 BCE to 480 BCE), the art that proliferated contained images of people who had the archaic smile, as evidenced by statues found in excavations all across the Greek mainland, Asia Minor, and on islands in the Aegean Sea. The significance of the convention is not known although it is often assumed that for the Greeks, that kind of smile reflected a state of ideal health and well-being. It has also been suggested that it is simply the result of a technical difficulty in fitting the curved shape of the mouth to the somewhat-blocklike head typical of Archaic sculpture. Richard Neer theorizes that the archaic smile may actually be a marker of status, since aristocrats of multiple cities throughout Greece were referred to as the Geleontes or "smiling ones". There are alternative views to the archaic smile being "flat and quite unnatural looking". John Fowles describes the archaic smile in his novel The Magus as "full of the purest metaphysical good humour [...] timelessly intelligent and timelessly amused. [...] Because a star explodes and a thousand worlds like ours die, we know this world is. That is the smile: that what might not be, is [...] When I die, I shall have this by my bedside. It is the last human face I want to see." The Greek archaic smile is also found on Etruscan artworks during the same time period nearby on the west side of the Italian peninsula, as consequence of the influence of Greek art on Etruscan art. An example of this commonly featured in art history texts is the Sarcophagus of the Spouses, a terracotta work found in the necropolis of Cerveteri. It features a smiling coup
https://en.wikipedia.org/wiki/Procedural%20reasoning%20system
In artificial intelligence, a procedural reasoning system (PRS) is a framework for constructing real-time reasoning systems that can perform complex tasks in dynamic environments. It is based on the notion of a rational agent or intelligent agent using the belief–desire–intention software model. A user application is predominately defined, and provided to a PRS system is a set of knowledge areas. Each knowledge area is a piece of procedural knowledge that specifies how to do something, e.g., how to navigate down a corridor, or how to plan a path (in contrast with robotic architectures where the programmer just provides a model of what the states of the world are and how the agent's primitive actions affect them). Such a program, together with a PRS interpreter, is used to control the agent. The interpreter is responsible for maintaining beliefs about the world state, choosing which goals to attempt to achieve next, and choosing which knowledge area to apply in the current situation. How exactly these operations are performed might depend on domain-specific meta-level knowledge areas. Unlike traditional AI planning systems that generate a complete plan at the beginning, and replan if unexpected things happen, PRS interleaves planning and doing actions in the world. At any point, the system might only have a partially specified plan for the future. PRS is based on the BDI or belief–desire–intention framework for intelligent agents. Beliefs consist of what the agent believes to be true about the current state of the world, desires consist of the agent's goals, and intentions consist of the agent's current plans for achieving those goals. Furthermore, each of these three components is typically explicitly represented somewhere within the memory of the PRS agent at runtime, which is in contrast to purely reactive systems, such as the subsumption architecture. History The PRS concept was developed by the Artificial Intelligence Center at SRI International duri
https://en.wikipedia.org/wiki/Anti-transglutaminase%20antibodies
Anti-transglutaminase antibodies (ATA) are autoantibodies against the transglutaminase protein. Antibodies serve an important role in the immune system by detecting cells and substances that the rest of the immune system then eliminates. These cells and substances can be foreign (for example, viruses) and also can be produced by the body (for example, cancer cells). Antibodies against the body's own products are called autoantibodies. Autoantibodies can sometimes errantly be directed against healthy portions of the organism, causing autoimmune diseases. ATA can be classified according to 2 different schemes: transglutaminase isoform and immunoglobulin reactivity subclass (IgA, IgG) toward transglutaminases. Transglutaminase isoform reactivity Anti-tissue transglutaminase Antibodies to tissue transglutaminase (abbreviated as anti-tTG or anti-TG2) are found in patients with several conditions, including celiac disease, juvenile diabetes, inflammatory bowel disease, and various forms of arthritis. In celiac disease, ATA are involved in the destruction of the villous extracellular matrix and target the destruction of intestinal villous epithelial cells by killer cells. Deposits of anti-tTG in the intestinal epithelium predict celiac disease. Anti-endomysial reactivity The endomysium is a layer of connective tissue that ensheaths a muscle fiber. The endomysium contains a form of transglutaminase called "tissue transglutaminase" or "tTG" for short, and antibodies that bind to this form of transglutaminase are called endomysial autoantibodies (EmA). The antiendomysial antibody test is a histological assay for patient serum binding to esophageal tissue from primate. EmA are present in celiac disease. They do not cause any direct symptoms to muscles, but detection of EmA is useful in the diagnosis of the disease. Anti-epidermal transglutaminase Antibodies to epidermal transglutaminase (eTG, also keratinocyte transglutaminase) are the autoantibodies believed to caus
https://en.wikipedia.org/wiki/George%20Lane%20%28mental%20calculator%29
George Lane (born 1964) is a British mental calculator and author. He is a five-time gold medalist in the Mental Calculations event at the annual Mind Sports Olympiad, and is one of only five Grandmasters of Mental Calculation, as recognised by the Mind Sports Organisation. George is a regular facilitator at the annual Junior Mental Calculation World Championship. He has written two books.
https://en.wikipedia.org/wiki/Algebraic%20fraction
In algebra, an algebraic fraction is a fraction whose numerator and denominator are algebraic expressions. Two examples of algebraic fractions are and . Algebraic fractions are subject to the same laws as arithmetic fractions. A rational fraction is an algebraic fraction whose numerator and denominator are both polynomials. Thus is a rational fraction, but not because the numerator contains a square root function. Terminology In the algebraic fraction , the dividend a is called the numerator and the divisor b is called the denominator. The numerator and denominator are called the terms of the algebraic fraction. A complex fraction is a fraction whose numerator or denominator, or both, contains a fraction. A simple fraction contains no fraction either in its numerator or its denominator. A fraction is in lowest terms if the only factor common to the numerator and the denominator is 1. An expression which is not in fractional form is an integral expression. An integral expression can always be written in fractional form by giving it the denominator 1. A mixed expression is the algebraic sum of one or more integral expressions and one or more fractional terms. Rational fractions If the expressions a and b are polynomials, the algebraic fraction is called a rational algebraic fraction or simply rational fraction. Rational fractions are also known as rational expressions. A rational fraction is called proper if , and improper otherwise. For example, the rational fraction is proper, and the rational fractions and are improper. Any improper rational fraction can be expressed as the sum of a polynomial (possibly constant) and a proper rational fraction. In the first example of an improper fraction one has where the second term is a proper rational fraction. The sum of two proper rational fractions is a proper rational fraction as well. The reverse process of expressing a proper rational fraction as the sum of two or more fractions is called resolving it into p
https://en.wikipedia.org/wiki/Rigid-band%20model
The Rigid-Band Model (or RBM) is one of the models used to describe the behavior of metal alloys. In some cases the model is even used for non-metal alloys such as Si alloys. According to the RBM the shape of the constant energy surfaces (hence the Fermi surface as well) and curve of density of states of the alloy are the same as those of the solvent metal under the following conditions: The excess charge of the solute atoms localizes around them. The mean free path of the electrons is much greater than the lattice spacing of the alloy. The electron states of interest in the pure solvent are all in one energy band, which is greatly separated in energy from the other bands. The only effect of the addition of the solute, given that its valence is greater than that of the solvent, is the addition of electrons to the valence band. This results to swelling the Fermi surface and filling the density of states curve to a higher energy. Theory In a pure metal, because of the periodicity of the lattice, the features of its electronic structure are well known. The single-particle states can be described in terms of Bloch states, the energy structure is characterized by Brillouin zone boundaries, energy gaps and energy bands. In reality though no metal is perfectly pure. When the amount of the foreign element is dilute, the added atoms may be treated as impurities. But when its concentration exceeds several atomic % , an alloy is formed and the interaction among the added atoms can no longer be neglected. Before giving a more mathematical outline of the RBM it is convenient to give somewhat of a visualization of what happens to a metal upon alloying it. In a pure metal, we'll take silver as an example, all lattice sites are occupied by silver atoms. When different kind of atoms are dissolved into it, for example 10% of copper, some random lattice sites become occupied by copper atoms. Since silver has a valence of 1 and copper has a valence of 2, the alloy will now have
https://en.wikipedia.org/wiki/WFXQ-CD
WFXQ-CD (channel 28) is a low-power, Class A television station in Springfield, Massachusetts, United States. It is a translator of dual NBC/CW+ affiliate WWLP (channel 22), owned by Nexstar Media Group. WFXQ-CD's transmitter is located at the old Mount Tom Ski Area summit in Holyoke. Its parent station maintains studios at Broadcast Center in the Sandy Hill section of Chicopee at the northwest corner of the I-391/MA 116/Chicopee Street interchange. History The station first went on-the-air May 6, 1987, on VHF channel 11. Using the calls W11BJ, it originally aired a low-powered analog signal from the Rattlesnake Mountain transmitter site of Connecticut's Fox affiliate WTIC-TV (channel 61). The station was an independent that aired local shows to a senior retirement community in Farmington, Connecticut. It used a live skycam weather forecast which consisted of a character generator and a home video camera with shots of the window from the transmitter building. The owner of the station was the Chase family (who also owned WTIC-TV). When LIN TV bought W11BJ in 2004, there was a construction permit approved to broadcast this station on UHF channel 28 from a new transmitter on Mount Tom in Holyoke. During the building of this transmitter, WWLP temporarily put on a simulcast of Connecticut's UPN affiliate WCTX (a sister station) through an off-air pickup. In early 2006, W28CT signed-on from the top of Mount Tom and the W11BJ transmitter on Rattlesnake Mountain was shut down. Right from the start, the station began to simulcast WWLP in a full-time manner. LIN TV had initially changed the call sign to WXCW-CA in reference to The CW in anticipation of it becoming an affiliate of that network. This affiliation eventually went to cable-only WB affiliate "WBQT". As a result, the channel's call sign was changed again to WFXQ-CA referring to a possible Fox affiliation. This caused rumors on several television industry message boards that it would become an affiliate of that ne
https://en.wikipedia.org/wiki/Steam%20turbine%20governing
Steam turbine governing is the procedure of controlling the flow rate of steam to a steam turbine so as to maintain its speed of rotation as constant. The variation in load during the operation of a steam turbine can have a significant impact on its performance. In a practical situation the load frequently varies from the designed or economic load and thus there always exists a considerable deviation from the desired performance of the turbine. The primary objective in the steam turbine operation is to maintain a constant speed of rotation irrespective of the varying load. This can be achieved by means of governing in a steam turbine. There are many types of governors. Overview Steam turbine governing is the procedure of monitoring and controlling the flow rate of steam into the turbine with the objective of maintaining its speed of rotation as constant. The flow rate of steam is monitored and controlled by interposing valves between the boiler and the turbine. Depending upon the particular method adopted for control of steam flow rate, different types of governing methods are being practiced. The principal methods used for governing are described below. Throttle governing In throttle governing the pressure of steam is reduced at the turbine entry thereby decreasing the availability of energy. In this method steam is passed through a restricted passage thereby reducing its pressure across the governing valve. The flow rate is controlled using a partially opened steam control valve. The reduction in pressure leads to a throttling process in which the enthalpy of steam remains constant. Throttle governing – Small turbines Low initial cost and simple mechanism makes throttle governing the most apt method for small steam turbines. The mechanism is illustrated in figure 1. The valve is actuated by using a centrifugal governor which consists of flying balls attached to the arm of the sleeve. A geared mechanism connects the turbine shaft to the rotating shaft on which
https://en.wikipedia.org/wiki/Dimension%20function
In mathematics, the notion of an (exact) dimension function (also known as a gauge function) is a tool in the study of fractals and other subsets of metric spaces. Dimension functions are a generalisation of the simple "diameter to the dimension" power law used in the construction of s-dimensional Hausdorff measure. Motivation: s-dimensional Hausdorff measure Consider a metric space (X, d) and a subset E of X. Given a number s ≥ 0, the s-dimensional Hausdorff measure of E, denoted μs(E), is defined by where μδs(E) can be thought of as an approximation to the "true" s-dimensional area/volume of E given by calculating the minimal s-dimensional area/volume of a covering of E by sets of diameter at most δ. As a function of increasing s, μs(E) is non-increasing. In fact, for all values of s, except possibly one, Hs(E) is either 0 or +∞; this exceptional value is called the Hausdorff dimension of E, here denoted dimH(E). Intuitively speaking, μs(E) = +∞ for s < dimH(E) for the same reason as the 1-dimensional linear length of a 2-dimensional disc in the Euclidean plane is +∞; likewise, μs(E) = 0 for s > dimH(E) for the same reason as the 3-dimensional volume of a disc in the Euclidean plane is zero. The idea of a dimension function is to use different functions of diameter than just diam(C)s for some s, and to look for the same property of the Hausdorff measure being finite and non-zero. Definition Let (X, d) be a metric space and E ⊆ X. Let h : [0, +∞) → [0, +∞] be a function. Define μh(E) by where Then h is called an (exact) dimension function (or gauge function) for E if μh(E) is finite and strictly positive. There are many conventions as to the properties that h should have: Rogers (1998), for example, requires that h should be monotonically increasing for t ≥ 0, strictly positive for t > 0, and continuous on the right for all t ≥ 0. Packing dimension Packing dimension is constructed in a very similar way to Hausdorff dimension, except that one "packs" E f
https://en.wikipedia.org/wiki/Flags%20depicting%20the%20Southern%20Cross
The Southern Cross or Crux, a constellation visible in the Southern Hemisphere, is depicted on flags and coats of arms of various countries and sub-national entities. This star constellation is visible mostly in the southern hemisphere and it therefore symbolises the southern location of its users. The term Southern Cross can also refer to the blue saltire as used in various flags of the Confederate States of America in the American Civil War. This list is an incomplete list and some of the flags in this list might not have official status. Flag proportions may vary between the different flags, and sometimes even vary between different versions of the same flag. National flags of countries in the Southern Hemisphere Other flags of the Commonwealth of Australia Other flags of the Federative Republic of Brazil Other flags of the Realm of New Zealand Other flags of Papua New Guinea Other flags in South America Other flags with the Southern Cross See also Nordic Cross Flag Union Flag
https://en.wikipedia.org/wiki/FROSTBURG
FROSTBURG was a Connection Machine 5 (CM-5) massively parallel supercomputer used by the US National Security Agency (NSA) to perform higher-level mathematical calculations. The CM-5 was built by the Thinking Machines Corporation, based in Cambridge, Massachusetts, at a cost of US$25 million. The system was installed at NSA in 1991, and operated until 1997. It was the first massively parallel processing computer bought by NSA, originally containing 256 processing nodes. The system was upgraded in 1993 with another 256 nodes, for a total of 512 nodes. The system had a total of 500 billion 32-bit words (≈ 2 terabytes) of storage, 2.5 billion words (≈ 10 gigabytes) of memory, and could perform at a theoretical maximum 65.5 gigaFLOPS. The operating system CMost was based on Unix, but optimized for parallel processing. FROSTBURG is now on display at the National Cryptologic Museum. See also HARVEST Cryptanalytic computer
https://en.wikipedia.org/wiki/Central%20Security%20Service
The Central Security Service (CSS) is a combat support agency of the United States Department of Defense which was established in 1972 to integrate the National Security Agency (NSA) and the Service Cryptologic Components (SCC) of the United States Armed Forces in the field of signals intelligence, cryptology, and information assurance at the tactical level. In 2002, the CSS had approximately 25,000 uniformed members. It is part of the United States Intelligence Community. History After World War II ended, the United States had two military organizations for the collection of signals intelligence (SIGINT): the Army Security Agency (ASA) and the Naval Communications Intelligence Organization (OP-20-G). The latter was deactivated and reorganized into the much smaller Communications Support Activities (CSA) in 1946, leaving ASA as the main SIGINT agency. Additionally, the United States Air Force established its own US Air Force Security Service (USAFSS) for the collection of communications intelligence in 1948. On May 20, 1949, the Secretary of Defense created the Armed Forces Security Agency (AFSA), which became responsible for the direction and control of all US communications intelligence (COMINT) and communications security (COMSEC) activities. However, at the tactical level these tasks continued to be performed by the respective army, navy, and air force agencies, which were not willing to accept the authority of the newly created AFSA. In trying to get control over the military SIGINT elements, AFSA was replaced by the new and more powerful National Security Agency (NSA) on October 24, 1952. Tactical military intelligence was traditionally collected by specialized soldiers, sailors, airmen, Marines, and coast guardsmen deployed around the world. For example, during the Vietnam War, each of the military services deployed its own cryptologic units, supported by the NSA, which set up a number of SIGINT Support Groups (SSGs) as merging points for signal intellige
https://en.wikipedia.org/wiki/Wedge%20%28symbol%29
Wedge (∧) is a symbol that looks similar to an in-line caret (^). It is used to represent various operations. In Unicode, the symbol is encoded and by \wedge and \land in TeX. The opposite symbol (∨) is called a vel, or sometimes a (descending) wedge. Some authors who call the descending wedge vel often call the ascending wedge ac (the corresponding Latin word for "and", also spelled "atque"), keeping their usage parallel Use Wedge is used to represent various operations: Logical conjunction in propositional logic and first-order logic Meet in lattice theory Exterior product or wedge product in differential geometry See also Turned v Vel (symbol) List of mathematical symbols List of logic symbols Wedge (disambiguation) /\ (disambiguation)
https://en.wikipedia.org/wiki/M2%20hol%20family
The Mycobacterial 2 TMS Phage Holin (M2 Hol) Family (TC# 1.E.36) is a group of transporters belonging to the Holin Superfamily VII. The Mycobactrerial 2 transmembrane segment (TMS) Holins have been identified and recognized by Catalao et al (2012). The Mycobacterium phage D29 gp11 protein (TC# 1.E.36.1.7) is a holin that, upon expression, rapidly kills both E. coli and Mycobacterium smegmatis. Shortening gp11 from its C-terminus resulted in diminished cytotoxicity and smaller holes. The two TMSs at the N-terminus alone do not integrate into the cytoplasmic membrane and do not show toxicity. Fusion of the two TMSs and a small C-terminal coiled-coil region resulted in restoration of cell killing. The second TMS is dispensable for toxicity. The gp11 C-terminal region is therefore necessary but not sufficient for toxicity. See also Holin Lysin Transporter Classification Database Further reading
https://en.wikipedia.org/wiki/National%20Institute%20of%20Metrology%20Standardization%20and%20Industrial%20Quality
The National Institute of Metrology, Standardization and Industrial Quality (INMETRO) is a Brazilian federal autarchy, linked to MDIC, the Ministry of Development, Industry and Foreign Trade. In Brazil, certification bodies must be accredited by INMETRO. Electrical and electronic products that meet Brazilian requirements and that are certified by an INMETRO accredited organization must carry the mandatory INMETRO mark along with the mark of the certification organization, such as UCIEE (União Certificadora para o Controle da Conformidade de Produtos, Processos ou Serviços). General information INMETRO acts as Executive Secretariat of the National Council of Metrology, Standardization and Industrial Quality (CONMETRO), an inter-ministerial collegiate entity which is the normative agency of the National System of Metrology, Standardization and Industrial Quality (SINMETRO). SINMETRO, CONMETRO and INMETRO were created by Brazilian Law 5966, of 11 of December 1973. In this occasion, the INMETRO substitutes the National Institute of Weights and Measurements (INPM), significantly extending its range of performance for the Brazilian society. In the scope of its ample institucional mission, INMETRO aims to fortify national companies, increasing their productivity by means of the adoption of mechanisms destined to the improvement of the quality of products and services. Being so, INMETRO mission incorporates the promotion of the quality of life of the Brazilian citizens and the competitiveness of the Brazilian economy, by means of Metrology and Evaluation of Conformity concepts and mechanisms. Some of the main attributions of INMETRO are: To execute the Brazilian Metrology and Evaluation of Conformity policies; To verify the observance of the technical and legal requirements related to measurement units and methods and instruments, as well as its application to material products; To keep the standards of the units of measurement, as well as to implement and keep the
https://en.wikipedia.org/wiki/Chemical%20defenses%20in%20Cannabis
Cannabis (/ˈkænəbɪs/) is commonly known as marijuana or hemp and has two known strains: Cannabis sativa and Cannabis indica, both of which produce chemicals to deter herbivory. The chemical composition includes specialized terpenes and cannabinoids, mainly tetrahydrocannabinol (THC), and cannabidiol (CBD). These substances play a role in defending the plant from pathogens including insects, fungi, viruses and bacteria. THC and CBD are stored mostly in the trichomes of the plant, and can cause psychological and physical impairment in the user, via the endocannabinoid system and unique receptors. THC increases dopamine levels in the brain, which attributes to the euphoric and relaxed feelings cannabis provides. As THC is a secondary metabolite, it poses no known effects towards plant development, growth, and reproduction. However, some studies show secondary metabolites such as cannabinoids, flavonoids, and terpenes are used as defense mechanisms against biotic and abiotic environmental stressors. Biosynthesis pathways Cannabinoids The production of the cannabinoids THC and CBD are a result of a series of chemical reactions, and are just two types of over a hundred that are known. Inside the transcriptomes of glandular trichomes in the cannabis plant, the pathway for cannabinoid production takes place. Beginning with the formation of 3,5,7-trioxododecaneoyl-COA by the condensation reaction between hexanoyl-CoA and malonyl-CoA, catalyzed by type III polyketide synthase (PKS), the product is then used to form olivetolic acid. After the geranylation of olivetolic acid, cannabigerolic acid (CBGA) or cannabigerivarinic acid (CBGVA) is formed. The decarboxylation of these acids yield what we recognize as THC and CBD. Terpenes Terpenes are a key component in chemotaxonomical classification of cannabis strains as terpene composition is a phenotypic trait. Majority of terpenes found in cannabis are hydrocarbons, which are a direct product of terpene synthase (TPS) enzym
https://en.wikipedia.org/wiki/Lindstr%C3%B6m%27s%20theorem
In mathematical logic, Lindström's theorem (named after Swedish logician Per Lindström, who published it in 1969) states that first-order logic is the strongest logic (satisfying certain conditions, e.g. closure under classical negation) having both the (countable) compactness property and the (downward) Löwenheim–Skolem property. Lindström's theorem is perhaps the best known result of what later became known as abstract model theory, the basic notion of which is an abstract logic; the more general notion of an institution was later introduced, which advances from a set-theoretical notion of model to a category-theoretical one. Lindström had previously obtained a similar result in studying first-order logics extended with Lindström quantifiers. Lindström's theorem has been extended to various other systems of logic, in particular modal logics by Johan van Benthem and Sebastian Enqvist. Notes
https://en.wikipedia.org/wiki/International%20Institute%20of%20Agriculture
The International Institute of Agriculture (IIA) was founded in Rome in 1905 by the King of Italy Victor Emmanuel III with the intent of creating a clearinghouse for collection of agricultural statistics. It was created primarily due to the efforts of David Lubin. In 1930, the IIA published the first world agricultural census. After World War II, both its assets and mandate were handed over to the Food and Agriculture Organization (FAO) of the United Nations. History In 1904 the idea of such an institute came to David Lubin of Sacramento, California, and his project found favor with the king of Italy. The latter gave a building in Rome and an annual income of $60,000. The king called the first congress in 1906, and delegates attended from 40 countries. At the congress, a treaty was formed making the institute a permanent organization and defining its scope and activities. Government The government of the IIA was vested in the general assembly of delegates from affiliated countries, meeting every two years, and in a permanent executive committee, on which there was one representative from each country. This permanent committee had direct charge of the IIA. The general officers were the president (also chairman of the permanent committee), the vice president and the secretary general. The work of the institute was divided among four bureaus: Bureau of the secretary general had charge of the personnel, financial and other routine business, the building and its equipment, the printing and distribution of publications, the library and general bibliographical work, and, as a more recent service, the preparation and publication of an annual compilation of agricultural legislation in the different countries of the world. Bureau of general statistics collected, collated and published statistics of production and commerce in agricultural products, both animal and vegetable, throughout the world. Bureau of agricultural intelligence and plant diseases collected and publis
https://en.wikipedia.org/wiki/RV%20Song%20of%20the%20Whale
RV Song of the Whale is a research vessel owned by Marine Conservation Research International and operated by Marine Conservation Research Ltd. The 70-foot vessel was designed specifically to carry out research on cetaceans (and other marine fauna) using benign research techniques such as passive acoustic monitoring. Description Song of the Whale is a cutter-rigged steel-hulled research vessel commissioned by the International Fund for Animal Welfare (IFAW) and built in 2004. The vessel was designed by Rogers Yacht Design of Lymington and built to Lloyd's Special Service Craft Rules for world-wide service - the first sailing vessel to meet those standards for 30 years. Ordered from Blondecell Ltd, the subcontracted steel hull was fabricated by Corus Steel and assembled by Riverside Fabrication at Falmouth, Cornwall. The addition of the composite superstructure and the full outfitting was carried out at Blondecell's facility at Cracknore Hard, Marchwood, Hampshire. The vessel cost £1.5 million. The design minimises acoustic emissions to facilitate the benign research techniques favoured by her former owners IFAW and the engine-room is encased in a Faraday cage to contain electrical fields. The outfit of Song of the Whale includes the latest computerised recording and tracking devices to ensure that best and most advanced acoustic research can be carried out. To assist physical observation, there is a two-person crow's nest. The new vessel was launched in St Katharine Docks, London, on 6 June 2004 by Pierce and Keely Brosnan. Operation Song of the Whale was formerly owned by IFAW until 14 March 2014 when the vessel was granted to Marine Conservation Research International of Kelvedon, UK. She is based at Ipswich and continues to carry out the research for which she designed, using benign techniques. She replaced a smaller vessel of the same name, a converted 46-foot luxury yacht, which had been in service for 17 years. Song of the Whale carries out most of i
https://en.wikipedia.org/wiki/Human%20viruses%20in%20water
Viruses are a major cause of human waterborne and water-related diseases. Waterborne diseases are caused by water that is contaminated by human and animal urine and feces that contain pathogenic microorganisms. A subject can get infected through contact with or consumption of the contaminated water. Viruses affect all living organisms from single cellular plants, bacteria and animal to the highest forms of plants and animals including human beings. Within a specific kingdom ( Plantae, Animalia, Fungi etc) the localization of viruses colonizing the host can vary: Some human viruses, for example, HIV, colonizes only the immune system, while influenza viruses on the other hand can colonize either the upper respiratory tract or the lower respiratory tract depending on the type (human Influenza virus or avian influenza viruses respectively). Different viruses can have different routes of transmission; for example, HIV is directly transferred by contaminated body fluids from an infected host into the tissue or bloodstream of a new host while influenza is airborne and transmitted through inhalation of contaminated air containing viral particles by a new host. Research has also suggested that solid surface plays a role in the transmission of water viruses. In a experiments that used E.coli phages, Qβ, fr, T4, and MS2 confirmed that viruses survive on a solid surface longer compared to when they are in water. Because of this adaptation to survive longer on solid surfaces, viruses now have a prolonged opportunities to infect humans. Enteric viruses primarily infect the intestinal tract through ingestion of food and water contaminated with viruses of fecal origin. Some viruses can be transmitted through all three routes of transmission. Water virology started about half a century ago when scientists attempted to detect the polio virus in water samples. Since then, other pathogenic viruses that are responsible for gastroenteritis, hepatitis, and many other virus strains hav
https://en.wikipedia.org/wiki/Vietnam%27s%20Red%20Data%20Book
Vietnam's Red Data Book (Vietnamese: Sách đỏ Việt Nam) is a list of rare and endangered species of fauna and flora native to Vietnam. This is the important legal basis for relating governmental regulations for the protection of biodiversity and wildlife in Vietnam. Criteria for this book are set forth on the basis of those specified by IUCN Red List. Nature conservation in Vietnam Ecology literature IUCN Red List
https://en.wikipedia.org/wiki/Architecture%20description%20language
Architecture description languages (ADLs) are used in several disciplines: system engineering, software engineering, and enterprise modelling and engineering. The system engineering community uses an architecture description language as a language and/or a conceptual model to describe and represent system architectures. The software engineering community uses an architecture description language as a computer language to create a description of a software architecture. In the case of a so-called technical architecture, the architecture must be communicated to software developers; a functional architecture is communicated to various stakeholders and users. Some ADLs that have been developed are: Acme (developed by CMU), AADL (standardized by the SAE), C2 (developed by UCI), SBC-ADL (developed by National Sun Yat-Sen University), Darwin (developed by Imperial College London), and Wright (developed by CMU). Overview The ISO/IEC/IEEE 42010 document, Systems and software engineering—Architecture description, defines an architecture description language as "any form of expression for use in architecture descriptions" and specifies minimum requirements on ADLs. The enterprise modelling and engineering community have also developed architecture description languages catered for at the enterprise level. Examples include ArchiMate (now a standard of The Open Group), DEMO, ABACUS (developed by the University of Technology, Sydney). These languages do not necessarily refer to software components, etc. Most of them, however, refer to an application architecture as the architecture that is communicated to the software engineers. Most of the writing below refers primarily to the perspective from the software engineering community. A standard notation (ADL) for representing architectures helps promote mutual communication, the embodiment of early design decisions, and the creation of a transferable abstraction of a system. Architectures in the past were largely represented by
https://en.wikipedia.org/wiki/Visual%20descriptor
In computer vision, visual descriptors or image descriptors are descriptions of the visual features of the contents in images, videos, or algorithms or applications that produce such descriptions. They describe elementary characteristics such as the shape, the color, the texture or the motion, among others. Introduction As a result of the new communication technologies and the massive use of Internet in our society, the amount of audio-visual information available in digital format is increasing considerably. Therefore, it has been necessary to design some systems that allow us to describe the content of several types of multimedia information in order to search and classify them. The audio-visual descriptors are in charge of the contents description. These descriptors have a good knowledge of the objects and events found in a video, image or audio and they allow the quick and efficient searches of the audio-visual content. This system can be compared to the search engines for textual contents. Although it is certain, that it is relatively easy to find text with a computer, is much more difficult to find concrete audio and video parts. For instance, imagine somebody searching a scene of a happy person. The happiness is a feeling and it is not evident its shape, color and texture description in images. The description of the audio-visual content is not a superficial task and it is essential for the effective use of this type of archives. The standardization system that deals with audio-visual descriptors is the MPEG-7 (Motion Picture Expert Group - 7). Types Descriptors are the first step to find out the connection between pixels contained in a digital image and what humans recall after having observed an image or a group of images after some minutes. Visual descriptors are divided in two main groups: General information descriptors: contain low level descriptors which give a description about color, shape, regions, textures and motion. Specific domain infor
https://en.wikipedia.org/wiki/Ebook
An ebook (short for electronic book), also known as an e-book or eBook, is a book publication made available in electronic form, consisting of text, images, or both readable on the flat-panel display of computers or other electronic devices. Although sometimes defined as "an electronic version of a printed book", some e-books exist without a printed equivalent. E-books can be read on dedicated e-reader devices, but also on any computer device that features a controllable viewing screen, including desktop computers, laptops, tablets and smartphones. In the 2000s, there was a trend of print and e-book sales moving to the Internet, where readers buy traditional paper books and e-books on websites using e-commerce systems. With print books, readers are increasingly browsing through images of the covers of books on publisher or bookstore websites and selecting and ordering titles online. The paper books are then delivered to the reader by mail or another delivery service. With e-books, users can browse through titles online, and then when they select and order titles, the e-book can be sent to them online or the user can download the e-book. By the early 2010s, e-books had begun to overtake hardcover by overall publication figures in the U.S. The main reasons for people buying e-books are possibly lower prices, increased comfort (as they can buy from home or on the go with mobile devices) and a larger selection of titles. With e-books, "electronic bookmarks make referencing easier, and e-book readers may allow the user to annotate pages." "Although fiction and non-fiction books come in e-book formats, technical material is especially suited for e-book delivery because it can be digitally searched" for keywords. In addition, for programming books, code examples can be copied. In the U.S., the amount of e-book reading is increasing. By 2014, 28% of adults had read an e-book, compared to 23% in 2013. By 2014, 50% of American adults had an e-reader or a tablet, compared to
https://en.wikipedia.org/wiki/Lilienfeld%20Prize
The Julius Edgar Lilienfeld Prize of the American Physical Society, to remember Julius Edgar Lilienfeld, has been awarded annually, since 1989. (It was not awarded in 2002). The purpose of the Prize is to recognize outstanding contributions to physics. Recipients Source: American Physical Society External links J. E. Lilienfeld Prize for Theoretical Particle Physics APS See also List of physics awards
https://en.wikipedia.org/wiki/Computational%20sustainability
Computational sustainability is an emerging field that attempts to balance societal, economic, and environmental resources for the future well-being of humanity using methods from mathematics, computer science, and information science fields. Sustainability in this context refers to the world's ability to sustain biological, social, and environmental systems in the long term. Using the power of computers to process large quantities of information, decision making algorithms allocate resources based on real-time information. Applications advanced by this field are widespread across various areas. For example, artificial intelligence and machine learning techniques are created to promote long-term biodiversity conservation and species protection. Smart grids implement renewable resources and storage capabilities to control the production and expenditure of energy. Intelligent transportation system technologies can analyze road conditions and relay information to drivers so they can make smarter, more environmentally-beneficial decisions based on real-time traffic information. Origins and motivations The field of computational sustainability has been motivated by Our Common Future, a 1987 report from the World Commission on Environment and Development about the future of humanity. More recently, computational sustainability research has also been driven by the United Nation's sustainable development goals, a set of 17 goals for the sustainability of human economic, social, and environmental well-being world-wide. Researchers in computational sustainability have primarily focused on addressing problems in areas related to the environment (e.g., biodiversity conservation), sustainable energy infrastructure and natural resources, and societal aspects (e.g., global hunger crises). The computational aspects of computational sustainability leverage techniques from mathematics and computer science, in the areas of artificial intelligence, machine learning, algorithms, gam
https://en.wikipedia.org/wiki/NetShow
NetShow was Microsoft's original framework for Internet network broadcasting, intended to compete with RealNetworks RealMedia & Vivo (acquired in 1998 by RealNetworks). It was later renamed and marketed under the Windows Media umbrella. NetShow 1.0 came out in 1996. A newer version, 2.0, was included in Windows NT 4.0 SP3 in 1997. Version 3.0 came out mid-1998. The whole product line was renamed Windows Media in October, 1999, four months before Windows 2000 appeared. The NetShow name is still carried on in the user-agent string in current versions of Windows Media Player, which reports as "NSPlayer". Components NetShow Player (version 2.0 was included with Internet Explorer 4 March 10 1997, now incorporated into Windows Media Player) NetShow Services (renamed Windows Media Services) It was eventually incorporated into the media server functionality of Microsoft Internet Information Services (IIS). Netshow server and encoder functionality was also integrated into PowerPoint as part of the Online Broadcast functionality in Microsoft Office 2000. NetShow Encoder (renamed Windows Media Encoder) NetShow Real-Time Encoder ASF Editor NetShow Presenter VIDTOASF WAVTOASF ASFCHOP NetShow Channel (renamed Windows Media Station) NetShow Server (also known as Theater Server): A high bitrate, full frame, on-demand streaming media solution for closed-circuit networks (such as airplanes or hotels). Related technologies Active Streaming Format (ASF) (later renamed Advanced Streaming Format, then to Advanced Systems Format) Microsoft Media Server (MMS)
https://en.wikipedia.org/wiki/Vaccine%20passports%20during%20the%20COVID-19%20pandemic
A vaccine passport or proof of vaccination is an immunity passport employed as a credential in countries and jurisdictions as part of efforts to control the COVID-19 pandemic via vaccination. A vaccine passport is typically issued by a government or health authority, and usually consists of a digital or printed record. Some credentials may include a scannable QR code, which can also be provisioned via mobile app. It may or may not use a COVID-19 vaccine card as a basis of authentication. The use of vaccine passports is based on the general presumption that a vaccinated individual would be less likely to transmit SARS-CoV-2 to others, and less likely to experience a severe outcome (hospitalization or death) if they were to be infected, thus making it relatively safer for them to congregate. A vaccine passport is typically coordinated with policies enforced by individual businesses, or enforceable public health orders, that require patrons to present proof of vaccination for COVID-19 as a condition of entry or service. Government-mandated use of vaccine passports typically applies to discretionary public spaces and events (such as indoor restaurants, bars, or large-scale in-person events, such as concerts and sports), and not essential businesses, such as retail stores or health care. In France, Italy, Ireland, and Canada, vaccine uptake increased after various levels of governments announced plans to introduce vaccine passports. An intention by some jurisdictions is to prevent future lockdowns and restrictions. Vaccine passports are controversial and have raised scientific, ethical and legal concerns. Critics have also argued that vaccine passports violate civil liberties via coercion. In the United States, there is no vaccine passport at a federal level, and some US states have preemptively banned vaccine passports in certain public and private sector contexts, citing discrimination and privacy concerns. England initially decided against mandating vaccine passpo
https://en.wikipedia.org/wiki/DIGIC
Digital Imaging Integrated Circuit (often styled as "DiG!C") is Canon Inc.'s name for a family of signal processing and control units for digital cameras and camcorders. DIGIC units are used as image processors by Canon in its own digital imaging products. Several generations of DIGICs exist, and are distinguished by a version number suffix. Currently, DIGIC is implemented as an application-specific integrated circuit (ASIC) designed to perform high speed signal processing as well as the control operations in the product in which it has been incorporated. Over its numerous generations, DIGIC has evolved from a system involving a number of discrete integrated circuits to a single chip system, many of which are based around the ARM instruction set. Custom firmware for these units has been developed to add features to the cameras. DIGIC in Cameras Original DIGIC The original DIGIC was used on the PowerShot G3 (Sep 2002), Canon S1 IS (Mar 2004), A520 (Mar 2005), and other cameras. It consists of three separate chips: a video processing IC, an image processing IC and a camera control IC. DIGIC II DIGIC II is a single chip system introduced in 2004, unlike the first DIGIC, that allowed for more compact designs. DIGIC II also improved upon the original by adding a larger buffer and increasing processing speed. It has been used in some advanced consumer-level cameras and many digital SLRs such as Canon EOS 5D and Canon EOS 30D. DIGIC II uses high-speed DDR-SDRAM, which improves startup time and AF performance. It can write to memory card at speeds up to 5.8 MB/sec. Additionally, Canon claims DIGIC II improves color, sharpness, and automatic white balance with its CMOS sensor in its digital SLR camera line. DIGIC III The DIGIC III Image Processor, introduced in 2006, was advertised to deliver improved image quality, faster operation and extended battery life compared to its predecessor. DIGIC III provides a faster interface to the SD memory card for the Canon Powe
https://en.wikipedia.org/wiki/Bit
The bit is the most basic unit of information in computing and digital communications. The name is a portmanteau of binary digit. The bit represents a logical state with one of two possible values. These values are most commonly represented as either , but other representations such as true/false, yes/no, on/off, or +/− are also widely used. The relation between these values and the physical states of the underlying storage or device is a matter of convention, and different assignments may be used even within the same device or program. It may be physically implemented with a two-state device. A contiguous group of binary digits is commonly called a bit string, a bit vector, or a single-dimensional (or multi-dimensional) bit array. A group of eight bits is called one byte, but historically the size of the byte is not strictly defined. Frequently, half, full, double and quadruple words consist of a number of bytes which is a low power of two. A string of four bits is a nibble. In information theory, one bit is the information entropy of a random binary variable that is 0 or 1 with equal probability, or the information that is gained when the value of such a variable becomes known. As a unit of information, the bit is also known as a shannon, named after Claude E. Shannon. The symbol for the binary digit is either "bit", per the IEC 80000-13:2008 standard, or the lowercase character "b", per the IEEE 1541-2002 standard. Use of the latter may create confusion with the capital "B" which is the international standard symbol for the byte. History The encoding of data by discrete bits was used in the punched cards invented by Basile Bouchon and Jean-Baptiste Falcon (1732), developed by Joseph Marie Jacquard (1804), and later adopted by Semyon Korsakov, Charles Babbage, Herman Hollerith, and early computer manufacturers like IBM. A variant of that idea was the perforated paper tape. In all those systems, the medium (card or tape) conceptually carried an array of hole p
https://en.wikipedia.org/wiki/Sum-of-squares%20optimization
A sum-of-squares optimization program is an optimization problem with a linear cost function and a particular type of constraint on the decision variables. These constraints are of the form that when the decision variables are used as coefficients in certain polynomials, those polynomials should have the polynomial SOS property. When fixing the maximum degree of the polynomials involved, sum-of-squares optimization is also known as the Lasserre hierarchy of relaxations in semidefinite programming. Sum-of-squares optimization techniques have been applied across a variety of areas, including control theory (in particular, for searching for polynomial Lyapunov functions for dynamical systems described by polynomial vector fields), statistics, finance and machine learning.<ref>Tan, W., Packard, A., 2004. "Searching for control Lyapunov functions using sums of squares programming". In: Allerton Conf. on Comm., Control and Computing. pp. 210–219.</ref>A. Chakraborty, P. Seiler, and G. Balas, "Susceptibility of F/A-18 Flight Controllers to the Falling-Leaf Mode: Nonlinear Analysis," AIAA Journal of Guidance, Control, and Dynamics, vol. 34 no. 1 (2011), pp. 73–85. Optimization problem Given a vector and polynomials for , , a sum-of-squares optimization problem is written as Here "SOS" represents the class of sum-of-squares (SOS) polynomials. The quantities are the decision variables. SOS programs can be converted to semidefinite programs (SDPs) using the duality of the SOS polynomial program and a relaxation for constrained polynomial optimization using positive-semidefinite matrices, see the following section. Dual problem: constrained polynomial optimization Suppose we have an -variate polynomial , and suppose that we would like to minimize this polynomial over a subset . Suppose furthermore that the constraints on the subset can be encoded using polynomial equalities of degree at most , each of the form where is a polynomial of degree at most . A natural,
https://en.wikipedia.org/wiki/AMG%20LASSO
AMG LASSO is a media recognition service launched by the All Media Guide in 2004. The LASSO service automatically recognizes CDs, DVDs, and digital audio files in formats such as MP3, WMA, and others. The service uses CD table of contents (ToC), DVD ToC, and acoustic fingerprint based recognition to recognize media. LASSO is available in versions for PCs and embedded devices. LASSO competes with user submitted services like freedb, Gracenote, MusicBrainz, and Discogs. See also List of online music databases External links Macrovision's LASSO Product Page (using AMG data) Online music and lyrics databases Acoustic fingerprinting