source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/FM%20broadcast%20band
The FM broadcast band is a range of radio frequencies used for FM broadcasting by radio stations. The range of frequencies used differs between different parts of the world. In Europe and Africa (defined as International Telecommunication Union (ITU) region 1) and in Australia and New Zealand, it spans from 87.5 to 108 megahertz (MHz) - also known as VHF Band II - while in the Americas (ITU region 2) it ranges from 88 to 108 MHz. The FM broadcast band in Japan uses 76 to 95 MHz, and in Brazil, 76 to 108 MHz. The International Radio and Television Organisation (OIRT) band in Eastern Europe is from 65.9 to 74.0 MHz, although these countries now primarily use the 87.5 to 108 MHz band, as in the case of Russia. Some other countries have already discontinued the OIRT band and have changed to the 87.5 to 108 MHz band. Narrow band Frequency Modulation was developed and demonstrated by Hanso Idzerda in 1919. Wide band Frequency modulation radio originated in the United States during the 1930s; the system was developed by the American electrical engineer Edwin Howard Armstrong. However, FM broadcasting did not become widespread, even in North America, until the 1960s. Frequency-modulated radio waves can be generated at any frequency. All the bands mentioned in this article are in the very high frequency (VHF) range, which extends from 30 to 300 MHz. CCIR bandplan Center frequencies While all countries use FM channel center frequencies ending in 0.1, 0.3, 0.5, 0.7, and 0.9 MHz, some countries also use center frequencies ending in 0.0, 0.2, 0.4, 0.6, and 0.8 MHz. A few others also use 0.05, 0.15, 0.25, 0.35, 0.45, 0.55, 0.65, 0.75, 0.85, and 0.95 MHz. An ITU conference in Geneva, Switzerland, on December 7, 1984, resolved to discontinue the use of 50 kHz channel spacings throughout Europe. Most countries have used 100 kHz or 200 kHz channel spacings for FM broadcasting since this ITU conference in 1984. Some digitally-tuned FM radios are unable to tune using 50 kHz or
https://en.wikipedia.org/wiki/Christopher%20Llewellyn%20Smith
Sir Christopher Hubert Llewellyn Smith (born 19 November 1942) is an Emeritus Professor of Physics at the University of Oxford. Education Llewellyn Smith was educated at the University of Oxford (BA) and completed his Doctor of Philosophy degree in theoretical physics at New College, Oxford in 1967. Career and research After his DPhil he worked at the Lebedev Physical Institute in Moscow, CERN and then the SLAC National Accelerator Laboratory before returning to Oxford in 1974. Llewellyn Smith was elected a Fellow of the Royal Society in 1984. While Chairman of Oxford Physics (1987–92), he led the merger of five different departments into a single Physics Department. Llewellyn Smith was Director General of CERN from 1994 to 1998. Thereafter he served as Provost and President of University College London (1999–2002). Awards and honours Llewellyn Smith received the James Clerk Maxwell Medal and Prize in 1979, and Glazebrook Medal and Prize of the Institute of Physics in 1999 and was knighted in 2001. In 2004, he became Chairman of the Consultative Committee for Euratom on Fusion (CCE-FU). Until 2009 he was Director of UKAEA Culham Division, which holds the responsibility for the United Kingdom's fusion programme and operation of the Joint European Torus (JET). He is a member of the Advisory Council for the Campaign for Science and Engineering. In 2013, he joined the National Institute of Science Education and Research (NISER), Bhubaneswar, India as a Distinguished Professor. In 2015, he was awarded the Royal Medal of the Royal Society. Personal life Llewellyn Smith married in 1966 and has one son and one daughter.
https://en.wikipedia.org/wiki/Hip%20protector
A hip protector is a specialized form of pants or underwear containing pads (either hard or soft) along the outside of each hip/leg, designed to prevent hip fractures following a fall. Recent developments include the use of double-sided adhesive films that are breathable and more comfortable to wear than specialised pants. The adhesive films are safer because they can be worn in the bath or shower, or when toileting. Hip protectors are most commonly used in elderly individuals who have a high risk of falls and hip fractures (for example, due to history of a previous fall and underlying osteoporosis). Most hip fractures follow an impact due to a lateral fall. The pads are located over the trochanters, the bony extrusions of the hip region. Effectiveness A 2014 Cochrane review found that hip protectors decrease the number of hip fractures among the elderly. A number of reviews have found that hip protectors are cost-effective, particularly among residents of care homes and for long-stay patients in hospital A previous review found that the effect for preventing hip fracture among nursing home residents was small and not effective among community dwelling elderly individuals. A 2007 review found a decreased risk of hip fractures in elderly nursing home residents. However, acceptance and long-term compliance towards them has historically been quite low, mainly because of discomfort, dislike of their appearance by the person wearing it, and disagreement about fracture risk. More modern hip protectors do not suffer from these disadvantages because they are slimmer with a low profile, so less noticeable, have ventilation holes and ducting to keep the skin cool under the pad and are soft and pliable conforming to the contours of the hip. Better independent testing procedures developed by Professor Julian Minns have established a reliable baseline for impact absorption performance. Research which has found hip protectors to be beneficial found that hard, energy-shunt
https://en.wikipedia.org/wiki/Mat%C3%A9rn%20covariance%20function
In statistics, the Matérn covariance, also called the Matérn kernel, is a covariance function used in spatial statistics, geostatistics, machine learning, image analysis, and other applications of multivariate statistical analysis on metric spaces. It is named after the Swedish forestry statistician Bertil Matérn. It specifies the covariance between two measurements as a function of the distance between the points at which they are taken. Since the covariance only depends on distances between points, it is stationary. If the distance is Euclidean distance, the Matérn covariance is also isotropic. Definition The Matérn covariance between measurements taken at two points separated by d distance units is given by where is the gamma function, is the modified Bessel function of the second kind, and ρ and are positive parameters of the covariance. A Gaussian process with Matérn covariance is times differentiable in the mean-square sense. Spectral density The power spectrum of a process with Matérn covariance defined on is the (n-dimensional) Fourier transform of the Matérn covariance function (see Wiener–Khinchin theorem). Explicitly, this is given by Simplification for specific values of ν Simplification for ν half integer When , the Matérn covariance can be written as a product of an exponential and a polynomial of order : which gives: for : for : for : The Gaussian case in the limit of infinite ν As , the Matérn covariance converges to the squared exponential covariance function Taylor series at zero and spectral moments The behavior for can be obtained by the following Taylor series (reference is needed, the formula below leads to division by zero in case ): When defined, the following spectral moments can be derived from the Taylor series: See also Radial basis function
https://en.wikipedia.org/wiki/Q-matrix
In mathematics, a Q-matrix is a square matrix whose associated linear complementarity problem LCP(M,q) has a solution for every vector q. Properties M is a Q-matrix if there exists d > 0 such that LCP(M,0) and LCP(M,d) have a unique solution. Any P-matrix is a Q-matrix. Conversely, if a matrix is a Z-matrix and a Q-matrix, then it is also a P-matrix. See also P-matrix Z-matrix
https://en.wikipedia.org/wiki/Generalized%20flag%20variety
In mathematics, a generalized flag variety (or simply flag variety) is a homogeneous space whose points are flags in a finite-dimensional vector space V over a field F. When F is the real or complex numbers, a generalized flag variety is a smooth or complex manifold, called a real or complex flag manifold. Flag varieties are naturally projective varieties. Flag varieties can be defined in various degrees of generality. A prototype is the variety of complete flags in a vector space V over a field F, which is a flag variety for the special linear group over F. Other flag varieties arise by considering partial flags, or by restriction from the special linear group to subgroups such as the symplectic group. For partial flags, one needs to specify the sequence of dimensions of the flags under consideration. For subgroups of the linear group, additional conditions must be imposed on the flags. In the most general sense, a generalized flag variety is defined to mean a projective homogeneous variety, that is, a smooth projective variety X over a field F with a transitive action of a reductive group G (and smooth stabilizer subgroup; that is no restriction for F of characteristic zero). If X has an F-rational point, then it is isomorphic to G/P for some parabolic subgroup P of G. A projective homogeneous variety may also be realised as the orbit of a highest weight vector in a projectivized representation of G. The complex projective homogeneous varieties are the compact flat model spaces for Cartan geometries of parabolic type. They are homogeneous Riemannian manifolds under any maximal compact subgroup of G, and they are precisely the coadjoint orbits of compact Lie groups. Flag manifolds can be symmetric spaces. Over the complex numbers, the corresponding flag manifolds are the Hermitian symmetric spaces. Over the real numbers, an R-space is a synonym for a real flag manifold and the corresponding symmetric spaces are called symmetric R-spaces. Flags in a vector space
https://en.wikipedia.org/wiki/Monkey%20gland%20sauce
Monkey gland sauce is a dark coloured, thick, sweet and tangy sauce from South Africa. It is typically served as a topping for grilled steaks or burgers, but is also used as a marinade, a dipping sauce for onion rings and chips, or on roasted potatoes. It has been featured as a restaurant item since the 1930s, becoming a South African restaurant and fast food staple condiment. Ingredients The main components of monkey gland sauce are chutney and tomato sauce – which result in a sweet mixture. Then an addition of onions, vinegar, garlic and Worcestershire sauce, give it a savoury-sweet flavour. Naming Despite its name, the sauce does not contain any monkey glands. There are various theories on the origins of the sauce but the most likely is that it originated with French chefs at the old Carlton Hotel in Johannesburg. South African diners added sauces such as chutney, tomato sauce, and Worcester sauce to the French dishes before eating it. Thus, the disgruntled chefs combined all the condiments to create a sauce which they named monkey gland sauce. There was speculation at the time that monkey glands could slow down ageing. A more outlandish theory is that it was named after Russian-born French scientist, Dr Serge Abrahamovitch Voronoff, who was a regular visitor at the Savoy Hotel in London. One of his medical experiments involved grafting monkey testicle tissue onto impotent men as a cure. The hotel renamed his favourite steak dish the "monkey gland steak" when he became famous. Then an ex-Savoy waiter brought it over to South Africa in the 1930s. See also Monkey Gland – cocktail List of dips List of sauces Notes
https://en.wikipedia.org/wiki/Organellar%20DNA
Organellar DNA (oDNA) is DNA contained in organelles (such as mitochondria and chloroplasts), outside the nucleus of eukaryotic cells. Mitochondria contain mitochondrial DNA Plastids (e.g., chloroplasts) contain plastid DNA Inheritance of organelle DNA The traits encoded by this type of DNA, in animals, generally pass from mother to offspring rather than from the father in a process called cytoplasmic inheritance. This is due to the ovum provided from the mother being larger than the male sperm cell, and therefore has more organelles, where the organellar DNA is found. Although maternal inheritance is most common, there are also paternal and biparental patterns of inheritance that take place. The latter two patterns of inheritance are found most often in plants. Recombination of organelle DNA is very limited, meaning that any traits that are encoded by the oDNA are likely to remain the same as they are passed from generation to generation. Structure Unlike nuclear DNA, which is present as linear molecules inside the chromosomes, the entire genomes of chloroplasts and mitochondria are present on a single molecule of double-stranded circular DNA molecule; this is very similar structure to a bacterial chromosome. Although the functionality and genetic structure vary significantly between different organelles and their host species, genetic characteristic patterns allow the differentiation between nucleolar and organellar DNA. A recently published machine-learning approach using only the genome sequences and multiple genome annotation tools can classify them. See also Nuclear DNA Non-Mendelian Inheritance
https://en.wikipedia.org/wiki/MASTAR%20MOSFET%20Model
The MASTAR (Model for Analog Simulation of subThreshold, saturation and weak Avalanche Regions) is an analytical model of Metal-Oxide Semiconductor Field-Effect Transistors, developed using the voltage-doping transformation (VDT) technique. MASTAR offers good accuracy and continuity in current and its derivatives in all operation regimes of the MOSFET devices. The model has been successfully used in CAD/EDA simulation tools. The official ITRS definition of the acronym MASTAR is Model for Assessment of CMOS Technologies And Roadmaps. This software is developed by STMicroelectronics and is freely distributed on ITRS organization web site. See also Transistor models SPICE
https://en.wikipedia.org/wiki/Guillotine%20cutting
Guillotine cutting is the process of producing small rectangular items of fixed dimensions from a given large rectangular sheet, using only guillotine-cuts. A guillotine-cut (also called an edge-to-edge cut) is a straight bisecting line going from one edge of an existing rectangle to the opposite edge, similarly to a paper guillotine. Guillotine cutting is particularly common in the glass industry. Glass sheets are scored along horizontal and vertical lines, and then broken along these lines to obtain smaller panels. It is also useful for cutting steel plates, cutting of wood sheets to make furniture, and cutting of cardboard into boxes. There are various optimization problems related to guillotine cutting, such as: maximize the total area of the produced pieces, or their total value; minimize the amount of waste (unused parts) of the large sheet, or the total number of sheets. They have been studied in combinatorial geometry, operations research and industrial engineering. A related but different problem is guillotine partition. In that problem, the dimensions of the small rectangles are not fixed in advance. The challenge comes from the fact that the original sheet might not be rectangular - it can be any rectilinear polygon. In particular, it might contain holes (representing defects in the raw material). The optimization goal is usually to minimize the number of small rectangles, or minimize the total length of the cuts. Terminology and assumptions The following terms and notations are often used in the literature on guillotine cutting. The large rectangle, also called the stock sheet, is the raw rectangular sheet which should be cut. It is characterized by its width W0 and height H0, which are the primary inputs to the problem The small rectangles, also called items, are the required outputs of the cutting. They are characterized by their width wi and height hi and for i in 1,...,m, where m is the number of rectangles. Often, it is allowed to have sever
https://en.wikipedia.org/wiki/Mycobacterium%20lepromatosis
Mycobacterium lepromatosis is an aerobic, acid-fast bacillus (AFB), and the second known causative agent of Hansen's disease (leprosy). It was discovered in 2008. Analysis of the 16S rRNA gene confirms that the species is distinct from Mycobacterium leprae. Characteristics Members of the Mycobacterium genus are characterized by being Gram-positive, non-motile, non-spore-forming, and possess a bacilliary cell shape. Bacteria in the Mycobacterium genus are characteristically known for possessing an outer membrane, a capsule, as well as a uniquely thick, waxy, hydrophobic cell wall abundant in mycolic acids. Many species of Mycobacterium are opportunistic pathogenic bacteria and can cause serious disease. The 16S ribosomal RNA (rRNA), a genetic marker of bacterial evolution is used in establishing phylogenetic relationships due to its highly conserved nature. 16S rRNA sequencing of M. lepromatosis displays a 2.1% genomic discrepancy from other Mycobacterium. The rpoB and hspa5 gene sequences were also analyzed, displaying speciation from its two most related ancestors of Mycobacterium, M. leprae and M. tuberculosis. M. lepromatosis exemplifies a 9.1% deviance from M. leprae in nucleotide structure, evidence of a novel bacterium. A phylogenetic analysis of 20 genes and pseudogenes, isolated to 5 conserved protein genes, was used to construct phylogenetic trees displaying long terminal branches from the nodes, suggesting speciation occurred long ago. Disease Mycobacterium lepromatosis can induce diffuse lepromatous leprosy (DLL), typically known to occur in Mexico and the Caribbean. DLL is a severe form of leprosy which manifests through nerve invasion and extensive skin ulcerations due to massive AFB burden in internal organs. M. lepromatosis, like M. leprae, has not been cultured in the laboratory because they both lack genes necessary to grow outside their hosts. These genes are believed to have been lost through reductive evolution. The novel bacterium was disc
https://en.wikipedia.org/wiki/Hippocrates%20of%20Chios
Hippocrates of Chios (; c. 470 – c. 410 BC) was an ancient Greek mathematician, geometer, and astronomer. He was born on the isle of Chios, where he was originally a merchant. After some misadventures (he was robbed by either pirates or fraudulent customs officials) he went to Athens, possibly for litigation, where he became a leading mathematician. On Chios, Hippocrates may have been a pupil of the mathematician and astronomer Oenopides of Chios. In his mathematical work there probably was some Pythagorean influence too, perhaps via contacts between Chios and the neighboring island of Samos, a center of Pythagorean thinking: Hippocrates has been described as a 'para-Pythagorean', a philosophical 'fellow traveler'. "Reduction" arguments such as reductio ad absurdum argument (or proof by contradiction) have been traced to him, as has the use of power to denote the square of a line. Mathematics The major accomplishment of Hippocrates is that he was the first to write a systematically organized geometry textbook, called Elements (Στοιχεῖα, Stoicheia), that is, basic theorems, or building blocks of mathematical theory. From then on, mathematicians from all over the ancient world could, at least in principle, build on a common framework of basic concepts, methods, and theorems, which stimulated the scientific progress of mathematics. Only a single, famous fragment of Hippocrates' Elements is existent, embedded in the work of Simplicius. In this fragment the area is calculated of some so-called Hippocratic lunes. This was part of a research program to square the circle, that is, to construct a square with the same area as a circle. The strategy, apparently, was to divide a circle into a number of crescent-shaped parts. If it were possible to calculate the area of each of those parts, then the area of the circle as a whole would be known too. Only much later was it proven (by Ferdinand von Lindemann, in 1882) that this approach had no chance of success, because the fa
https://en.wikipedia.org/wiki/Gerontophobia
Gerontophobia is the fear of age-related self-degeneration (similar to Gerascophobia), or a hatred or fear of the elderly due to memento mori. The term comes from the Greek γέρων – gerōn, "old man" and φόβος – phobos, "fear". Gerontophobia has been linked to Thanatophobia as fear of old age can be a precursor to fear of death. Gerontophobia can be caused by harmful stereotypes of elderly people displayed in the media Ageism Discriminatory aspects of ageism have been strongly linked to gerontophobia. This irrational fear or hatred of the elderly can be associated with the expectation that someday all young people including oneself will be old inevitably and suffer from the irreversible health decline that comes with old age, which is associated with disability, disease, and death. The sight of aged people could be a possible reminder of death (memento mori) and inevitable biological vulnerability. This unwillingness to accept these can manifest in feelings of hostility and discriminatory acts towards the elderly. History Old age was previously seen as a golden age in the Middle Ages Around the time of the Anglo-Saxons there was a shift towards more negative views of the elderly, which led to more and more literature developing a Gerontophobic view. Portrayal in Literature and the Media Gerontophobia is heavily portrayed in literature and the media starting as early as Anglo-saxon poetry but is also found in common literary classics such as Shakespeare's King Lear, Jonathan Swift's Gulliver's Travels, and Jane Austen's Persuasion Gerontophobia can also be found in many TV shows and movies. Treatments for Gerontophobia Treatment for Gerontophobia can include better education about the elderly and aging as well as an increase in exposure and insight therapy. See also Gerascophobia Gerontocracy Intergenerational equity List of phobias
https://en.wikipedia.org/wiki/ThreadX
Azure RTOS ThreadX is a highly deterministic, embedded real-time operating system (RTOS) programmed mostly in the language C. Overview ThreadX was originally developed and marketed by Express Logic of San Diego, California, United States. The author of ThreadX is William Lamie, who was also the original author of the Nucleus RTOS in 1990. William Lamie was President and CEO of Express Logic. Express Logic was purchased for an undisclosed sum by Microsoft on April 18, 2019. The name ThreadX is derived from the threads that are used as the executable elements, and the letter X represents context switching, i.e., it switches threads. ThreadX provides priority-based, preemptive scheduling, fast interrupt response, memory management, interthread communication, mutual exclusion, event notification, and thread synchronization features. Major distinguishing technology characteristics of ThreadX include preemption-threshold, priority inheritance, efficient timer management, fast software timers, picokernel design, event-chaining, and small size: minimal size on an ARM architecture processor is about 2 KB. ThreadX supports multi-core processor environments via either asymmetric multiprocessing (AMP) or symmetric multiprocessing (SMP). Application thread isolation with memory management unit (MMU) or memory protection unit (MPU) memory protection is available with ThreadX Modules. ThreadX has extensive safety certifications from Technischer Überwachungsverein (TÜV, English: Technical Inspection Association) and UL (formerly Underwriters Laboratories) and is Motor Industry Software Reliability Association MISRA C compliant. ThreadX is the foundation of Express Logic's X-Ware Internet of things (IoT) platform, which also includes embedded file system support (FileX), embedded UI support (GUIX), embedded Internet protocol suite (TCP/IP) and cloud connectivity (NetX/NetX Duo), and Universal Serial Bus (USB) support (USBX). ThreadX has won high appraisal from developers and i
https://en.wikipedia.org/wiki/Higher-order%20programming
Higher-order programming is a style of computer programming that uses software components, like functions, modules or objects, as values. It is usually instantiated with, or borrowed from, models of computation such as lambda calculus which make heavy use of higher-order functions. A programming language can be considered higher-order if components, such as procedures or labels, can be used just like data. For example, these elements could be used in the same way as arguments or values. For example, in higher-order programming, one can pass functions as arguments to other functions and functions can be the return value of other functions (such as in macros or for interpreting). This style of programming is mostly used in functional programming, but it can also be very useful in object-oriented programming. A slightly different interpretation of higher-order programming in the context of object-oriented programming are higher order messages, which let messages have other messages as arguments, rather than functions. Examples of languages supporting this are Wolfram Language, C#, Java, ECMAScript (ActionScript, JavaScript, JScript), F#, Haskell, Lisp (Common Lisp, Scheme, Clojure, others), Lua, Oz, Perl, PHP, Prolog, Python, Ruby, Smalltalk, Scala, ML, and Erlang. See also Prolog#Higher-order programming Higher-order logic programming
https://en.wikipedia.org/wiki/ISO%205427
ISO 5427 is an 8-bit extension to the KOI-7 N1 character set, which was standardised by the ISO. The first half was published in 1979, and the second half was published in 1981. It supports the Russian (also before 1918), Belarusian, Bulgarian (also before 1945), Ukrainian, Macedonian, and Serbian languages. In July and October 1983, there were some revisions. Code table
https://en.wikipedia.org/wiki/Bifidobacterium%20longum
Bifidobacterium longum is a Gram-positive, catalase-negative, rod-shaped bacterium present in the human gastrointestinal tract and one of the 32 species that belong to the genus Bifidobacterium. It is a microaerotolerant anaerobe and considered to be one of the earliest colonizers of the gastrointestinal tract of infants. When grown on general anaerobic medium, B. longum forms white, glossy colonies with a convex shape. B. longum is one of the most common bifidobacteria present in the gastrointestinal tracts of both children and adults. B. longum is non-pathogenic, is often added to food products, and its production of lactic acid is believed to prevent growth of pathogenic organisms. Classification In 2002, three previously distinct species of Bifidobacterium, B. infantis, B. longum, and B. suis, were unified into a single species named B. longum with the biotypes infantis, longum, and suis, respectively. This occurred as the three species had extensive DNA similarity including a 16S rRNA gene sequence similarity greater than 97%. In addition, the three original species were phenotypically difficult to distinguish due to different carbohydrate fermentation patterns among strains of the same species. As probiotic activity varies among strains of B. longum, interest exists in the exact classification of new strains, although this is made difficult by the high gene similarity between the three biotypes. Currently, strain identification is done through polymerase chain reaction (PCR) on the subtly different 16S rRNA gene sequences. Environment B. longum colonizes the human gastrointestinal tract, where it, along with other Bifidobacterium species, represents up to 90% of the bacteria of an infant's gastrointestinal tract. This number gradually drops to 3% in an adult's gastrointestinal tract as other enteric bacteria such as Bacteroides and Eubacterium begin to dominate. Some strains of B. longum were found to have high tolerance for gastric acid and bile, suggesti
https://en.wikipedia.org/wiki/Congenital%20afibrinogenemia
Congenital afibrinogenemia is a rare, genetically inherited blood fibrinogen disorder in which the blood does not clot normally due to the lack of fibrinogen, a blood protein necessary for coagulation. This disorder is autosomal recessive, meaning that two unaffected parents can have a child with the disorder. The lack of fibrinogen expresses itself with excessive and, at times, uncontrollable bleeding. Signs and symptoms As this is a disorder that is present in an individual from birth, there are no warning signs to look for. The first symptom usually seen is hemorrhage from the umbilical cord that is difficult to stop. Other symptoms include: Nasal and oral mucosa bleeds Gastrointestinal bleeding Excessive/spontaneous bleeding or bruising from minor injury Prolonged menstruation in women Spontaneous abortion during pregnancy CNS hemorrhaging Spontaneous bleeding of the mouth, nose, and gastrointestinal tract are common. Since blood clots can not be formed, minor injuries tend to lead to excessive bleeding or bruising. The biggest concern for individuals with afibrinogenemia is CNS hemorrhage, bleeding in the brain, which can be fatal. Many of these symptoms are chronic, and will continue to occur for the entirety of the affected individual's life. Causes A missense or nonsense mutation to the genes that code for the fibrinogen protein are affected. Usually the mutation leads to an early stop in the production of the protein. Due to the problem being genetically based, there is no way to prevent the disease. Individuals can get genetic testing done to see if they are a carrier of the trait, and if so may choose to complete genetic counseling to better understand the disorder and help manage family planning. Parents can choose to do prenatal genetic testing for the disorder to determine if their child will have the disease. The only risk factor is if both parents of a child carry the recessive allele linked to the disorder. Mechanism Individuals with th
https://en.wikipedia.org/wiki/Metacharacter
A metacharacter is a character that has a special meaning to a computer program, such as a shell interpreter or a regular expression (regex) engine. In POSIX extended regular expressions, there are 14 metacharacters that must be escaped (preceded by a backslash (\)) in order to drop their special meaning and be treated literally inside an expression: opening and closing square brackets ([ and ]); backslash (\); caret (^); dollar sign ($); period/full stop/dot (.); vertical bar/pipe symbol (|); question mark (?); asterisk (*); plus and minus signs (+ and -); opening and closing curly brackets/braces ({ and }); and opening and closing parentheses (( and )). For example, to match the arithmetic expression (1+1)*3=6 with a regex, the correct regex is \(1\+1\)\*3=6; otherwise, the parentheses, plus sign, and asterisk will have special meanings. Other examples Some other characters may have special meaning in some environments. In some Unix shells the semicolon (";") is a statement separator. In XML and HTML, the ampersand ("&") introduces an HTML entity. It also has special meaning in MS-DOS/Windows Command Prompt. In some Unix shells and MS-DOS/Windows Command Prompt, the less-than sign and greater-than sign ("<" and ">") are used for redirection and the backtick/grave accent ("`") is used for command substitution. In many programming languages, strings are delimited using quotes (" or '). In some cases, escape characters (and other methods) are used to avoid delimiter collision, e.g. "He said, \"Hello\"". In printf format strings, the percent sign ("%") is used to introduce format specifiers and must be escaped as "%%" to be interpreted literally. In SQL, the percent is used as a wildcard character. In SQL, the underscore ("_") is used to match any single character. Escaping The term "to escape a metacharacter" means to make the metacharacter ineffective (to strip it of its special meaning), causing it to have its literal meaning. For example, in PCRE, a do
https://en.wikipedia.org/wiki/FastICA
FastICA is an efficient and popular algorithm for independent component analysis invented by Aapo Hyvärinen at Helsinki University of Technology. Like most ICA algorithms, FastICA seeks an orthogonal rotation of prewhitened data, through a fixed-point iteration scheme, that maximizes a measure of non-Gaussianity of the rotated components. Non-gaussianity serves as a proxy for statistical independence, which is a very strong condition and requires infinite data to verify. FastICA can also be alternatively derived as an approximative Newton iteration. Algorithm Prewhitening the data Let the denote the input data matrix, the number of columns corresponding with the number of samples of mixed signals and the number of rows corresponding with the number of independent source signals. The input data matrix must be prewhitened, or centered and whitened, before applying the FastICA algorithm to it. Centering the data entails demeaning each component of the input data , that is, for each and . After centering, each row of has an expected value of . Whitening the data requires a linear transformation of the centered data so that the components of are uncorrelated and have variance one. More precisely, if is a centered data matrix, the covariance of is the -dimensional identity matrix, that is, A common method for whitening is by performing an eigenvalue decomposition on the covariance matrix of the centered data , , where is the matrix of eigenvectors and is the diagonal matrix of eigenvalues. The whitened data matrix is defined thus by Single component extraction The iterative algorithm finds the direction for the weight vector that maximizes a measure of non-Gaussianity of the projection , with denoting a prewhitened data matrix as described above. Note that is a column vector. To measure non-Gaussianity, FastICA relies on a nonquadratic nonlinear function , its first derivative , and its second derivative . Hyvärinen states that the functions
https://en.wikipedia.org/wiki/Superheating
In thermodynamics, superheating (sometimes referred to as boiling retardation, or boiling delay) is the phenomenon in which a liquid is heated to a temperature higher than its boiling point, without boiling. This is a so-called metastable state or metastate, where boiling might occur at any time, induced by external or internal effects. Superheating is achieved by heating a homogeneous substance in a clean container, free of nucleation sites, while taking care not to disturb the liquid. This may occur by microwaving water in a very smooth container. Disturbing the water may cause an unsafe eruption of hot water and result in burns. Cause Water is said to "boil" when bubbles of water vapor grow without bound, bursting at the surface. For a vapor bubble to expand, the temperature must be high enough that the vapor pressure exceeds the ambient pressure (the atmospheric pressure, primarily). Below that temperature, a water vapor bubble will shrink and vanish. Superheating is an exception to this simple rule; a liquid is sometimes observed not to boil even though its vapor pressure does exceed the ambient pressure. The cause is an additional force, the surface tension, which suppresses the growth of bubbles. Surface tension makes the bubble act like an elastic balloon. The pressure inside is raised slightly by the "skin" attempting to contract. For the bubble to expand, the temperature must be raised slightly above the boiling point to generate enough vapor pressure to overcome both surface tension and ambient pressure. What makes superheating so explosive is that a larger bubble is easier to inflate than a small one; just as when blowing up a balloon, the hardest part is getting started. It turns out the excess pressure due to surface tension is inversely proportional to the diameter of the bubble. That is, . This can be derived by imagining a plane cutting a bubble into two halves. Each half is pulled towards the middle with a surface tension force , which
https://en.wikipedia.org/wiki/Simulation%20hypothesis
The simulation hypothesis proposes that what humans experience as the world is actually a simulated reality, such as a computer simulation in which humans themselves are constructs. There has been much debate over this topic, ranging from philosophical discourse to practical applications in computing. The simulation hypothesis, as formulated by Nick Bostrom, is part of a long tradition of skeptical scenarios. It was presented by Bostrom as not merely a philosophical speculation but an empirical claim with quantifiable probabilities. The hypothesis has received criticism from some physicists, such as Sabine Hossenfelder who has called it pseudoscience and cosmologist George F. R. Ellis, who stated that "[the hypothesis] is totally impracticable from a technical viewpoint" and that "Late-night pub discussion is not a viable theory." Versions of the hypothesis have also been featured in science fiction, appearing as a central plot device in many stories and films, such as The Matrix. Origins Human history is full of thinkers who observed the difference between how things seem and how they might actually be, with dreams, illusions and hallucinations providing poetic and philosophical metaphors. For example, the "Butterfly Dream" of Zhuangzi from ancient China, or the Indian philosophy of Maya, or in Ancient Greek philosophy Anaxarchus and Monimus likened existing things to a scene-painting and supposed them to resemble the impressions experienced in sleep or madness. In the Western philosophical tradition, Plato's Allegory of the Cave stands out as an influential example. Aztec philosophical texts theorized that the world was a painting or book written by the Teotl. René Descartes' Evil Demon philosophically formalized these epistemic doubts, to be followed by a large literature with subsequent variations like Brain in a Vat. Simulation Argument Nick Bostrom's premise: Nick Bostrom's conclusion: Expanded Argument Bostrom attempted to assess the probability of
https://en.wikipedia.org/wiki/Caloron
In mathematical physics, a caloron is the finite temperature generalization of an instanton. Finite temperature and instantons At zero temperature, instantons are the name given to solutions of the classical equations of motion of the Euclidean version of the theory under consideration, and which are furthermore localized in Euclidean spacetime. They describe tunneling between different topological vacuum states of the Minkowski theory. One important example of an instanton is the BPST instanton, discovered in 1975 by Belavin, Polyakov, Schwartz and Tyupkin. This is a topologically stable solution to the four-dimensional SU(2) Yang–Mills field equations in Euclidean spacetime (i.e. after Wick rotation). Finite temperatures in quantum field theories are modeled by compactifying the imaginary (Euclidean) time (see thermal quantum field theory). This changes the overall structure of spacetime, and thus also changes the form of the instanton solutions. According to the Matsubara formalism, at finite temperature, the Euclidean time dimension is periodic, which means that instanton solutions have to be periodic as well. In SU(2) Yang–Mills theory In SU(2) Yang–Mills theory at zero temperature, the instantons have the form of the BPST instanton. The generalization thereof to finite temperature has been found by Harrington and Shepard: where is the anti-'t Hooft symbol, r is the distance from the point x to the center of the caloron, ρ is the size of the caloron, is the Euclidean time and T is the temperature. This solution was found based on a periodic multi-instanton solution first suggested by 't Hooft and published by Witten.
https://en.wikipedia.org/wiki/DDAIP
DDAIP is a pharmaceutical ingredient added to topical products to increase penetration through the skin. Chemically, DDAIP is an ester of N,N-dimethylalanine and dodecanol, although as of now the structural formula shows an ester with decanol (C10) instead. DDAIP is typically formulated as its hydrochloride salt (DDAIP.HCl). This salt is a white crystalline solid with a melting range of 88-93 °C and is an amphiphilic molecule with a pKa of 4.87 that is soluble in water up to about 40% w/v. DDAIP is proprietary to NexMed USA, a subsidiary of Apricus Biosciences. Mechanism of action DDAIP is a permeation enhancer that temporarily changes the permeation dynamics of the lipid bilayer and opens up the tight junctions between skin cells so active drug molecules can be rapidly absorbed through the skin into systemic circulation. It can also improve the solubility of compounds resulting in enhanced drug permeation. Clinical use DDAIP hydrochloride is a functional inactive excipient currently used in the topical drug Vitaros, an alprostadil vasodilator cream used to treat erectile dysfunction. It is also used in MycoVa, a terbinafine antifungal nail lacquer for onychomycosis currently in phase-III clinical trials. Safety Overall, about 5,000 patients have been exposed to this compound with no serious adverse events recorded. DDAIP is primarily metabolized by esterases on cell surfaces and plasma to N,N-dimethylalanine, which is further demethylated to alanine; and dodecanol which is oxidized to lauric acid, both naturally occurring compounds already present in the body.
https://en.wikipedia.org/wiki/4-Phenylthiosemicarbazide
4-Phenylthiosemicarbazide (4-PTSC) is a thiosemicarbazide used as an agricultural chemical and pesticide. It also possesses antibacterial properties attributed to electron delocalization in the thiosemicarbazide moiety.
https://en.wikipedia.org/wiki/C12orf40
C12orf40, also known as Chromosome 12 Open Reading Frame 40, HEL-206, and Epididymis Luminal Protein 206 is a protein that in humans is encoded by the C12orf40 gene. Gene Human gene In humans, the gene for C12orf40 is located on chromosome 12. There are 13 exons in the canonical isoform that is transcribed into an mRNA of 2797 base pairs. Three other isoforms have been isolated. Evolution Homologs exist as distant as the green sea turtle and chickens at approximately 60% sequence identity, suggesting that the gene may have arisen in the amniotes after their divergence from other tetrapods; the first 4 exons are conserved with 36% identity as distantly as the anemone. Protein Properties The human C12orf40 protein is 652 amino acids in length. Its molecular weight is predicted to be 74.52 kDa, and its isoelectric point 7.822. Amino acids 229-652 contain a domain of unknown function (DUF4552) which is conserved in vertebrates. C12orf40 is predicted to be a soluble protein with no transmembrane segments. Its secondary and tertiary structures are not currently known. Interactions Experimental evidence shows that C12orf40 has a physical interaction with dynein light chain 2 (DYNLL2). This protein is part of a complex that regulates the function of the motor protein dynein. Expression Within the cell, C12orf40 is predicted to be present in the nucleus based on signals within its sequence. An analysis of normal human tissues shows that C12orf40 expression occurs primarily in the testis, suggesting importance to the male reproductive system. Clinical significance The function of C12orf40 is not yet well understood. However, the three prime untranslated region (3' UTR) of C12orf40 is highly similar to the 3' UTR of the cystic fibrosis transmembrane conductance regulator (CFTR), which may mean that the two genes share certain expression patterns. In the fibroblasts of hypertrophic scars, exposure to the immunosuppressant Tacrolimus causes C12orf40 up-regulation. In
https://en.wikipedia.org/wiki/Rhinal%20sulcus
In the human brain, the entorhinal cortex appears as a longitudinal elevation anterior to the parahippocampal gyrus, with a corresponding internal furrow, the external rhinal sulcus (or rhinal fissure), separating it from the inferiolateral surface of the hemisphere close to the lamina terminalis. It is analogous to the collateral fissure found further caudally in the inferior part of the temporal lobe.
https://en.wikipedia.org/wiki/Sodium%20diacetate
Sodium diacetate is a compound with formula . It is a salt of acetic acid. It is a colorless solid that is used in seasonings and as an antimicrobial agent. Preparation and structure The salt forms upon half-neutralization of acetic acid followed by evaporation of the solution. It can be viewed as the result of homoassociation, an effect that enhances the acidity of acetic acid in concentrated solution: 2 CH3CO2H + NaOH → Na+[(CH3CO2)2H]− + H2O Also described as the sodium acid salt of acetic acid, it is best described as the sodium salt of the hydrogen-bonded anion (CH3CO2)2H−. The O···O distance is about 2.47 angstrom. The species has no significant existence in solution but forms stable crystals. Applications As a food additive, it has E number E262 and is used to impart a salt and vinegar flavor. See also Sodium acetate
https://en.wikipedia.org/wiki/Central%20dogma%20of%20molecular%20biology
The central dogma of molecular biology is an explanation of the flow of genetic information within a biological system. It is often stated as "DNA makes RNA, and RNA makes protein", although this is not its original meaning. It was first stated by Francis Crick in 1957, then published in 1958: He re-stated it in a Nature paper published in 1970: "The central dogma of molecular biology deals with the detailed residue-by-residue transfer of sequential information. It states that such information cannot be transferred back from protein to either protein or nucleic acid." A second version of the central dogma is popular but incorrect. This is the simplistic DNA → RNA → protein pathway published by James Watson in the first edition of The Molecular Biology of the Gene (1965). Watson's version differs from Crick's because Watson describes a two-step (DNA → RNA and RNA → protein) process as the central dogma. While the dogma as originally stated by Crick remains valid today, Watson's version does not. The dogma is a framework for understanding the transfer of sequence information between information-carrying biopolymers, in the most common or general case, in living organisms. There are 3 major classes of such biopolymers: DNA and RNA (both nucleic acids), and protein. There are conceivable direct transfers of information that can occur between these. The dogma classes these into 3 groups of 3: three general transfers (believed to occur normally in most cells), two special transfers (known to occur, but only under specific conditions in case of some viruses or in a laboratory), and four unknown transfers (believed never to occur). The general transfers describe the normal flow of biological information: DNA can be copied to DNA (DNA replication), DNA information can be copied into mRNA (transcription), and proteins can be synthesized using the information in mRNA as a template (translation). The special transfers describe: RNA being copied from RNA (RNA replication), D
https://en.wikipedia.org/wiki/Elongated%20pentagonal%20orthobicupola
In geometry, the elongated pentagonal orthobicupola or cantellated pentagonal prism is one of the Johnson solids (). As the name suggests, it can be constructed by elongating a pentagonal orthobicupola () by inserting a decagonal prism between its two congruent halves. Rotating one of the cupolae through 36 degrees before inserting the prism yields an elongated pentagonal gyrobicupola (). Formulae The following formulae for volume and surface area can be used if all faces are regular, with edge length a:
https://en.wikipedia.org/wiki/Cross-border%20flag%20for%20Ireland
There is no cross-border flag universally accepted as representing both jurisdictions on the island of Ireland. This can be a problem in contexts where a body organised on an all-island basis needs to be represented by a flag in an international context. The island is politically divided into the Republic of Ireland (a sovereign state comprising 26 of Ireland's 32 traditional counties) and Northern Ireland (comprising the remaining 6 counties in the north-east of Ireland, and a part of the United Kingdom), but all-island organisations are common. Examples include the Catholic Church in Ireland, the Presbyterian Church in Ireland, the Church of Ireland, the Gaelic Athletic Association, the Orange Order, Scouting Ireland and the Irish Rugby Football Union. Saint Patrick's Saltire The Saint Patrick's Saltire was incorporated into the Union Flag in 1801 by way of the Act of Union 1800 to represent Ireland within the new United Kingdom of Great Britain and Ireland. The Church of Ireland orders that, apart from the flag of the Anglican Communion, only this saltire may be flown on its church grounds—as opposed to the tricolour, the Union Flag or the former flag of Northern Ireland. This follows the practice of other Anglican churches in England, Scotland, and Wales, which fly the flags of their respective patron saints instead of the Union Flag. The saltire is also flown by the Catholic St. Patrick's College, Maynooth, on graduation days. Modified versions have been used formerly by the Irish Rugby Football Union, and currently by the Royal College of Surgeons in Ireland and the Commissioners of Irish Lights. Irish tricolour The tricolour of green, white and orange, the official flag of the Republic of Ireland, was originally intended by nationalists to represent the entire island. However, this symbolism is rejected by Ulster unionists, as illustrated by an exchange in the House of Commons of Northern Ireland in 1951: The Gaelic Athletic Association (GAA) flies
https://en.wikipedia.org/wiki/Bill%20English%20%28computer%20engineer%29
William Kirk English (January 27, 1929July 26, 2020) was an American computer engineer who contributed to the development of the computer mouse while working for Douglas Engelbart at SRI International's Augmentation Research Center. He would later work for Xerox PARC and Sun Microsystems. Early life English was born on January 27, 1929, in Lexington, Kentucky. The only son of Harry English and Caroline (Gray) English, he had two half-brothers from his father's previous marriage. Harry English was an electrical engineer who managed coal mines and Caroline was a homemaker. William, or Bill as he was known, attended a boarding school in Arizona and then studied electrical engineering at the University of Kentucky. Career English served in the US Navy until the late 1950s, including postings in northern California and Japan. He then joined the Stanford Research Institute in the 1960s to work on magnets, and built one of the first all-magnetic arithmetic units with Hewitt Crane. In 1964, he was the first person to join Douglas Engelbart's lab, the Augmentation Research Center. He and Douglas Engelbart share credit for creating the first computer mouse in 1963; English built the initial prototype, and was its first user, based on Engelbart's notes. English led a 1965 project, sponsored by NASA, which evaluated the best way to select a point on a computer display; the mouse was the winner. English was also instrumental at The Mother of All Demos in 1968, which showcased the mouse and other technologies developed as part of their NLS (oN-Line System). In particular, English figured out how to connect a terminal in the San Francisco Civic Auditorium to the host computer at SRI away, and also transmitted audio and video between the locations. He left SRI in 1971 and went to Xerox PARC, where he managed the Office Systems Research Group. While working at PARC, English developed a ball mouse, in which a ball replaced the original set of wheels. It worked similarly to a mov
https://en.wikipedia.org/wiki/Power%20transfer%20unit
In aviation, a power transfer unit (PTU) is a device that transfers hydraulic power from one of an aircraft's hydraulic systems to another in the event that the other system has failed or been turned off. The PTU is used when, for example, there is right hydraulic system pressure but no left hydraulic system pressure. In this example, the PTU transfers hydraulic power from the right hydraulic system to the left hydraulic system. A PTU consists of a hydraulic motor paired with a hydraulic pump via a shaft. As the connection is purely mechanical, there is no intermixing of hydraulic fluid between the left and right hydraulic systems during PTU operation. Design philosophy Large transport category aircraft with hydraulically powered flight controls and utilities typically have multiple, independent hydraulic systems powered by a combination of engine-driven and electrically driven hydraulic pumps. Multiple hydraulic systems are typically needed for redundancy, where for instance if one system fails or loses hydraulic fluid, a surviving system may still provide sufficient power for critical systems to continue safe flight and landing. On airliners or business jets with powered flight controls, it is typical to have at least two hydraulic power control units (actuators) for each critical flight control surface — these are the elevators, rudder and ailerons. Only two sources might be used if some form of mechanical reversion is present (i.e. the pilot can still fly the aeroplane manually, but with some difficulty, via mechanical linkages and cables if hydraulic power is lost). On fly-by-wire aircraft, at least three independent power sources are needed. Spoilers and flaps meanwhile are considered secondary flight controls, and may only have a single hydraulic power source, providing the flight control can be deployed symmetrically. Likewise, landing gear, brakes and nosewheel steering are systems which are not considered critical for flight, and are subsequently ty
https://en.wikipedia.org/wiki/Tree%20traversal
In computer science, tree traversal (also known as tree search and walking the tree) is a form of graph traversal and refers to the process of visiting (e.g. retrieving, updating, or deleting) each node in a tree data structure, exactly once. Such traversals are classified by the order in which the nodes are visited. The following algorithms are described for a binary tree, but they may be generalized to other trees as well. Types Unlike linked lists, one-dimensional arrays and other linear data structures, which are canonically traversed in linear order, trees may be traversed in multiple ways. They may be traversed in depth-first or breadth-first order. There are three common ways to traverse them in depth-first order: in-order, pre-order and post-order. Beyond these basic traversals, various more complex or hybrid schemes are possible, such as depth-limited searches like iterative deepening depth-first search. The latter, as well as breadth-first search, can also be used to traverse infinite trees, see below. Data structures for tree traversal Traversing a tree involves iterating over all nodes in some manner. Because from a given node there is more than one possible next node (it is not a linear data structure), then, assuming sequential computation (not parallel), some nodes must be deferred—stored in some way for later visiting. This is often done via a stack (LIFO) or queue (FIFO). As a tree is a self-referential (recursively defined) data structure, traversal can be defined by recursion or, more subtly, corecursion, in a natural and clear fashion; in these cases the deferred nodes are stored implicitly in the call stack. Depth-first search is easily implemented via a stack, including recursively (via the call stack), while breadth-first search is easily implemented via a queue, including corecursively. Depth-first search In depth-first search (DFS), the search tree is deepened as much as possible before going to the next sibling. To traverse binary tre
https://en.wikipedia.org/wiki/Retrospective%20diagnosis
A retrospective diagnosis (also retrodiagnosis or posthumous diagnosis) is the practice of identifying an illness after the death of the patient (sometimes a historical figure) using modern knowledge, methods and disease classifications. Alternatively, it can be the more general attempt to give a modern name to an ancient and ill-defined scourge or plague. Historical research Retrospective diagnosis is practised by medical historians, general historians and the media with varying degrees of scholarship. At its worst it may become "little more than a game, with ill-defined rules and little academic credibility". The process often requires "translating between linguistic and conceptual worlds separated by several centuries", and assumes our modern disease concepts and categories are privileged. Crude attempts at retrospective diagnosis fail to be sensitive to historical context, may treat historical and religious records as scientific evidence, or ascribe pathology to behaviours that require none. Darin Hayton, a historian of science at Haverford College, claims that retrodiagnosing famous individuals with autism in the media is pointless, as historical accounts often contain incomplete information. The understanding of the history of illness can benefit from modern science. For example, knowledge of the insect vectors of malaria and yellow fever can be used to explain the changes in extent of those diseases caused by drainage or urbanisation in historical times. The practice of retrospective diagnosis has been applied in parody, where characters from fiction are "diagnosed". Squirrel Nutkin may have had Tourette syndrome and Tiny Tim could have had distal renal tubular acidosis (type I). Postmortem diagnosis Post-mortem diagnosis is considered a research tool, and also a quality control practice and it allows to evaluate the performance of the clinical case definitions. The term retrospective diagnosis is also sometimes used by a clinical pathologist to descr
https://en.wikipedia.org/wiki/Osteoid%20osteoma
An osteoid osteoma is a benign (non-cancerous) bone tumor that arises from osteoblasts and some components of osteoclasts. It was originally thought to be a smaller version of an osteoblastoma. Osteoid osteomas tend to be less than 1.5 cm in size. The tumor can be in any bone in the body but are most common in long bones, such as the femur and tibia. They account for 10 to 12 percent of all benign bone tumors and 2 to 3 percent of all abnormal bone growths. Osteoid osteomas may occur at any age, and are most common in patients between the ages of 4 and 25 years old. Males are affected approximately three times more commonly than females. Signs and symptoms Common symptoms of an osteoid osteoma are: dull or sharp pain that worsens at night pain that is usually relieved by aspirin or other anti-inflammatory drugs limping painful scoliosis and muscle spasticity (when the tumor is located in the spine) growth disturbance (when the tumor is involved with a bone's growth plate) muscle wasting bowing deformity nerve symptoms like sciatica (when the tumor is located in the spine) The most common symptom is pain that can be relieved with over the counter pain medication in the beginning. After the benign tumor develops further the pain can not be alleviated with medication and minor to severe swelling starts to occur. Although, in some cases the pain level remains the same for years, and regular NSAIDs intake keeps the pain at bay. The tumor is often found through x-ray imaging. Conventional radiographs reveal a well-demarcated lytic lesion (nidus) surrounded by a distinct zone of sclerosis that allow doctors to identify the tumor. Characterized by being less than 1.5 cm in diameter, osteoid osteomas most frequently occur in young men (male:female ratio 3:1) and may occur in any bone of the body, most frequently around the knee but often also seen in the vertebrae, in the long bones and less commonly in the mandible or other craniofacial bones. Severe pain typica
https://en.wikipedia.org/wiki/AAL1gator
The AAL1gator is a semiconductor device that implements the Circuit Emulation Service. It was developed between 1994 and 1998 and became a run-away success. It also played a role in the acquisition of four companies. The name was based on the fact that the AAL1gator implements the ATM AAL-1 standard. Development of the AAL1gator The AAL1gator was developed by Network Synthesis, Inc. under contract from Integrated Telecom Technology (IgT). It was the first semiconductor solution to implement the Circuit Emulation Service standard from the ATM Forum. It implemented 8 DS1/E1 lines worth of CES and had 256 channels. It flexibly converted the PDH DS1 signal into Asynchronous Transfer Mode cells. The AAL1gator was principally designed by the Network Synthesis CEO, Brian Holden and a consultant, Ed Lennox. Brian Holden was also involved in the ATM Forum standardization effort for the Circuit Emulation Service. Additional design efforts came from Andy Annudurai, Ravi Sajwan, and Imran Chaudhri (who also came up with the name). Chee Hu did most of the work on getting the "C" version to work at speed and to be manufacturable. Denis Smetana did most of the work on the "D" version and on the later 32 DS1 version. Jim Jacobson of OnStream Networks was the Beta Customer. Patents on the AAL1gator Two U.S. patents were issued on the AAL1gator's calendar-based transmit scheduler, one on the original product and an even better one on the "D" version enhancements designed by Denis Smetana. The scheduler implemented several intricate methods of minimizing the jitter caused by the scheduling of the 256 channels. The AAL1gator also could have gotten another patent on its method of queuing the SRTS samples, but the designers were too busy to get the application in. Functions of the AAL1gator The AAL1gator could flexibly map individual DS0s or groups of DS0s into 256 ATM VCs. It also had a high speed mode which mapped a single DS-3 into ATM. Additionally, it had a high p
https://en.wikipedia.org/wiki/Gene%20product
A gene product is the biochemical material, either RNA or protein, resulting from expression of a gene. A measurement of the amount of gene product is sometimes used to infer how active a gene is. Abnormal amounts of gene product can be correlated with disease-causing alleles, such as the overactivity of oncogenes which can cause cancer. A gene is defined as "a hereditary unit of DNA that is required to produce a functional product". Regulatory elements include: Promoter region TATA box Polyadenylation sequences Enhancers These elements work in combination with the open reading frame to create a functional product. This product may be transcribed and be functional as RNA or is translated from mRNA to a protein to be functional in the cell. RNA products RNA molecules that do not code for any proteins still maintain a function in the cell. The function of the RNA depends on its classification. These roles include: aiding protein synthesis catalyzing reactions regulating various processes. Protein synthesis is aided by functional RNA molecules such as tRNA, which helps add the correct amino acid to a polypeptide chain during translation, rRNA, a major component of ribosomes (which guide protein synthesis), as well as mRNA which carry the instructions for creating the protein product. One type of functional RNA involved in regulation are microRNA (miRNA), which works by repressing translation. These miRNAs work by binding to a complementary target mRNA sequence to prevent translation from occurring. Short-interfering RNA (siRNA) also work by negative regulation of transcription. These siRNA molecules work in RNA-induced silencing complex (RISC) during RNA interference by binding to a target DNA sequence to prevent transcription of a specific mRNA. Protein products Proteins are the product of a gene that are formed from translation of a mature mRNA molecule. Proteins contain 4 elements in regards to their structure: primary, secondary, tertiary and quaternary.
https://en.wikipedia.org/wiki/Fibrous%20tunic%20of%20eyeball
The sclera and cornea form the fibrous tunic of the bulb of the eye; the sclera is opaque, and constitutes the posterior five-sixths of the tunic; the cornea is transparent, and forms the anterior sixth. The term "corneosclera" is also used to describe the sclera and cornea together.
https://en.wikipedia.org/wiki/Polynomial%20and%20rational%20function%20modeling
In statistical modeling (especially process modeling), polynomial functions and rational functions are sometimes used as an empirical technique for curve fitting. Polynomial function models A polynomial function is one that has the form where n is a non-negative integer that defines the degree of the polynomial. A polynomial with a degree of 0 is simply a constant function; with a degree of 1 is a line; with a degree of 2 is a quadratic; with a degree of 3 is a cubic, and so on. Historically, polynomial models are among the most frequently used empirical models for curve fitting. Advantages These models are popular for the following reasons. Polynomial models have a simple form. Polynomial models have well known and understood properties. Polynomial models have moderate flexibility of shapes. Polynomial models are a closed family. Changes of location and scale in the raw data result in a polynomial model being mapped to a polynomial model. That is, polynomial models are not dependent on the underlying metric. Polynomial models are computationally easy to use. Disadvantages However, polynomial models also have the following limitations. Polynomial models have poor interpolatory properties. High-degree polynomials are notorious for oscillations between exact-fit values. Polynomial models have poor extrapolatory properties. Polynomials may provide good fits within the range of data, but they will frequently deteriorate rapidly outside the range of the data. Polynomial models have poor asymptotic properties. By their nature, polynomials have a finite response for finite x values and have an infinite response if and only if the x value is infinite. Thus polynomials may not model asymptotic phenomena very well. While no procedure is immune to the bias-variance tradeoff, polynomial models exhibit a particularly poor tradeoff between shape and degree. In order to model data with a complicated structure, the degree of the model must be high, indicating that the associa
https://en.wikipedia.org/wiki/Actibind
Actibind is an actin-binding fungal T(2)-RNase protein that is produced by the black mold Aspergillus niger, a microorganism used in biotechnology and food technology. In plants, actibind binds actin, a major component of the cytoskeleton, interfering with the plants' pollen tubes and halting cell growth. Research published in the journal Cancer on 15 May 2006 reports evidence that actibind has antiangiogenic and anticarcinogenic characteristics. In human colon cancer, breast cancer and melanoma, increasing the level of actibind was found to reduce the ability of these cells to form tumorogenic colonies. In animal models, increased actibind inhibited the growth of colon cancer-derived tumors, metastases and blood vessel formation. During the completion of the Human Genome Project, the gene encoding for RNaseT2, the human actibind-like protein, was found on chromosome 6. Why ACTIBIND? The reason why ACTIBIND is an enzyme of interest in biochemical laboratories is due to the fact that researchers observed that ACTIBIND inhibits the elongation of pollen tubes by interfering with the intracellular actin network of the plant cell. The specific actin network ACTIBIND inhibits are actin rich pseudopods, which are important for a variety of cellular functions including elongation of plant pollen tubes, motility of mammalian cells, and most importantly cancer cell function. In cancer cells specifically, the actin rich pseudopods help the cancer's invasion and metastasis. Because ACTIBIND was able to interfere with actin networks in plant cells, researchers aimed to find out whether ACTIBIND could also inhibit mammalian cancer cell development. ACTIBIND as an Antiangiogenic ACTIBIND, an extracellular ribonuclease (T2-RNase), has been researched heavily in biochemical laboratories due the enzyme's ability to efficiently hydrolyze/degrade RNA molecules. The ability to cleave off RNA molecules it what makes ACTIBIND an effective antiangiogenic. Angiogenesis is the growth of
https://en.wikipedia.org/wiki/Centrifugal%20governor
A centrifugal governor is a specific type of governor with a feedback system that controls the speed of an engine by regulating the flow of fuel or working fluid, so as to maintain a near-constant speed. It uses the principle of proportional control. Centrifugal governors, also known as "centrifugal regulators" and "fly-ball governors", were invented by Christiaan Huygens and used to regulate the distance and pressure between millstones in windmills in the 17th century. In 1788, James Watt adapted one to control his steam engine where it regulates the admission of steam into the cylinder(s), a development that proved so important he is sometimes called the inventor. Centrifugal governors' widest use was on steam engines during the Steam Age in the 19th century. They are also found on stationary internal combustion engines and variously fueled turbines, and in some modern striking clocks. A simple governor does not maintain an exact speed but a speed range, since under increasing load the governor opens the throttle as the speed (RPM) decreases. Operation The devices shown are on steam engines. Power is supplied to the governor from the engine's output shaft by a belt or chain connected to the lower belt wheel. The governor is connected to a throttle valve that regulates the flow of working fluid (steam) supplying the prime mover. As the speed of the prime mover increases, the central spindle of the governor rotates at a faster rate, and the kinetic energy of the balls increases. This allows the two masses on lever arms to move outwards and upwards against gravity. If the motion goes far enough, this motion causes the lever arms to pull down on a thrust bearing, which moves a beam linkage, which reduces the aperture of a throttle valve. The rate of working-fluid entering the cylinder is thus reduced and the speed of the prime mover is controlled, preventing over-speeding. Mechanical stops may be used to limit the range of throttle motion, as seen near the mass
https://en.wikipedia.org/wiki/Abelian%20and%20Tauberian%20theorems
In mathematics, Abelian and Tauberian theorems are theorems giving conditions for two methods of summing divergent series to give the same result, named after Niels Henrik Abel and Alfred Tauber. The original examples are Abel's theorem showing that if a series converges to some limit then its Abel sum is the same limit, and Tauber's theorem showing that if the Abel sum of a series exists and the coefficients are sufficiently small (o(1/n)) then the series converges to the Abel sum. More general Abelian and Tauberian theorems give similar results for more general summation methods. There is not yet a clear distinction between Abelian and Tauberian theorems, and no generally accepted definition of what these terms mean. Often, a theorem is called "Abelian" if it shows that some summation method gives the usual sum for convergent series, and is called "Tauberian" if it gives conditions for a series summable by some method that allows it to be summable in the usual sense. In the theory of integral transforms, Abelian theorems give the asymptotic behaviour of the transform based on properties of the original function. Conversely, Tauberian theorems give the asymptotic behaviour of the original function based on properties of the transform but usually require some restrictions on the original function. Abelian theorems For any summation method L, its Abelian theorem is the result that if c = (cn) is a convergent sequence, with limit C, then L(c) = C. An example is given by the Cesàro method, in which L is defined as the limit of the arithmetic means of the first N terms of c, as N tends to infinity. One can prove that if c does converge to C, then so does the sequence (dN) where To see that, subtract C everywhere to reduce to the case C = 0. Then divide the sequence into an initial segment, and a tail of small terms: given any ε > 0 we can take N large enough to make the initial segment of terms up to cN average to at most ε/2, while each term in the tail is bo
https://en.wikipedia.org/wiki/Transmembrane%20protein%20217
Transmembrane Protein 217 is a protein encoded by the gene TMEM217. TMEM217 has been found to have expression correlated with the lymphatic system and endothelial tissues and has been predicted to have a function linked to the cytoskeleton. Gene TMEM217 is located on the chromosome 6 minus strand at 6p21.2. The gene consists of 46,857 base pairs and is flanked by TBC1D22B (TBC1 Domain Family Member 22B) and PIM1. It was previously known as C6orf128 (Chromosome 6 open reading frame 128). mRNA TMEM217 has three common isoforms formed from the alternative splicing of three exons. Isoform 1 translates for the longest polypeptide, consisting of 1590 nucleotides. The 5’ un-translated region of isoform 1 is relatively short and is predicted to fold into several stem loop domains within conserved areas of the un-translated region. Protein Primary Protein Sequence The longest polypeptide of transmembrane protein 217 consists of 229 amino acids. This protein isoform has a predicted weight of 26.6 kDa and isoelectric point at a pH of 9.3. It is notably rich in isoleucine and phenylalanine, and deficient in alanine, aspartate, and proline compared to other proteins. Transmembrane protein 217 contains the domain of unknown function, DUF4534, between amino acids 11-171. Secondary Structure Transmembrane protein 217 is predicted to have four transmembrane domains. These transmembrane domains consist primarily of uncharged amino acids in predicted alpha helices. The N-terminus and C-terminus of the protein are predicted to be facing the cytosol with the C-terminus containing a long predicted coiled tail extending from the final transmembrane domain. Post-Translational Modifications There are several predicted phosphorylation and glycosylation sites on transmembrane protein 217 in highly conserved parts of the protein, where the phosphorylation sites are located primarily on the C-terminal tail. There are also two highly conserved cysteine residues, which have the potenti
https://en.wikipedia.org/wiki/Personal%20Data%20Privacy%20and%20Security%20Act%20of%202009
The Personal Data Privacy and Security Act of 2009 ( Official title: A bill to prevent and mitigate identity theft, to ensure privacy, to provide notice of security breaches, and to enhance criminal penalties, law enforcement assistance, and other protections against security breaches, fraudulent access, and misuse of personally identifiable information), was a bill proposed in the United States Congress to increase protection of personally identifiable information by private companies and government agencies, set guidelines and restrictions on personal data sharing by data brokers, and to enhance criminal penalty for identity theft and other violations of data privacy and security. The bill was sponsored in the United States Senate by Patrick Leahy (Democrat-Vermont), where it is known as S.1490. Status Senator Patrick Leahy introduced the bill on July 22, 2009 and was referred to the Senate Judiciary Committee where it was approved. The last action was on December 17, 2009. This bill did not come up for debate during the 111th United States Congress and at the end of the 2009-2010 session and never became law. Summary The Congressional Research Service (CRS) of the Library of Congress summarizes the bill in its four main parts. Title I - Enhancing Punishment for Identity Theft and Other Violations of Data Privacy and Security Section 101 Amends the federal criminal code to add intentionally accessing a computer without authorization to the definition of racketeering activity. Section 102 Imposes a fine and/or prison term of up to five years for intentionally and willfully concealing a security breach involving sensitive personally identifiable information that causes economic damage to one or more persons. It defines "sensitive personally identifiable information" to include an individual's name in combination with other personal information, such as a social security number, home address, date of birth, biometrics data, or financial account information.
https://en.wikipedia.org/wiki/Handheld%20Isothermal%20Silver%20Standard%20Sensor
The Handheld Isothermal Silver Standard Sensor (HISSS) project was sponsored by DARPA in the 2000s to develop a hand-held sensor that is capable of identifying biological weapon threats across the entire spectrum including bacteria, viruses and toxins. The program began in early part of the 21st century with the following goals: DNA detection without polymerase chain reaction (PCR) RNA detection without PCR or reverse transcription antibody-based protein detection at sensitivities unachievable by traditional methods The final goal was to give field units the ability to detect threat agents across the complete spectrum of biological warfare weapons. The main contractor for this project was Northrop Grumman with subcontractors Ionian Technologies and Ribomed. External links HISSS Project Biological warfare
https://en.wikipedia.org/wiki/MathML
Mathematical Markup Language (MathML) is a mathematical markup language, an application of XML for describing mathematical notations and capturing both its structure and content, and is one of a number of mathematical markup languages. Its aim is to natively integrate mathematical formulae into World Wide Web pages and other documents. It is part of HTML5 and standardised by ISO/IEC since 2015. History Following some experiments in the Arena browser based on proposals for mathematical markup in HTML, MathML 1 was released as a W3C recommendation in April 1998 as the first XML language to be recommended by the W3C. Version 1.01 of the format was released in July 1999 and version 2.0 appeared in February 2001. Implementations of the specification appeared in Amaya 1.1, Mozilla 1.0 and Opera 9.5. In October 2003, the second edition of MathML Version 2.0 was published as the final release by the W3C Math Working Group. MathML was originally designed before the finalization of XML namespaces. However, it was assigned a namespace immediately after the Namespace Recommendation was completed, and for XML use, the elements should be in the namespace with namespace URL http://www.w3.org/1998/Math/MathML. When MathML is used in HTML (as opposed to XML) this namespace is automatically inferred by the HTML parser and need not be specified in the document. MathML version 3 Version 3 of the MathML specification was released as a W3C recommendation on 20 October 2010. A recommendation of A MathML for CSS Profile was later released on 7 June 2011; this is a subset of MathML suitable for CSS formatting. Another subset, Strict Content MathML, provides a subset of content MathML with a uniform structure and is designed to be compatible with OpenMath. Other content elements are defined in terms of a transformation to the strict subset. New content elements include which associates bound variables () to expressions, for example a summation index. The new element allows structure s
https://en.wikipedia.org/wiki/Unequal%20crossing%20over
Unequal crossing over is a type of gene duplication or deletion event that deletes a sequence in one strand and replaces it with a duplication from its sister chromatid in mitosis or from its homologous chromosome during meiosis. It is a type of chromosomal crossover between homologous sequences that are not paired precisely. Normally genes are responsible for occurrence of crossing over. It exchanges sequences of different links between chromosomes. Along with gene conversion, it is believed to be the main driver for the generation of gene duplications and is a source of mutation in the genome. Mechanisms During meiosis, the duplicated chromosomes (chromatids) in eukaryotic organisms are attached to each other in the centromere region and are thus paired. The maternal and paternal chromosomes then align alongside each other. During this time, recombination can take place via crossing over of sections of the paternal and maternal chromatids and leads to reciprocal recombination or non-reciprocal recombination. Unequal crossing over requires a measure of similarity between the sequences for misalignment to occur. The more similarity within the sequences, the more likely unequal crossing over will occur. One of the sequences is thus lost and replaced with the duplication of another sequence. When two sequences are misaligned, unequal crossing over may create a tandem repeat on one chromosome and a deletion on the other. The rate of unequal crossing over will increase with the number of repeated sequences around the duplication. This is because these repeated sequences will pair together, allowing for the mismatch in the cross over point to occur. Consequences for the organism Unequal crossing over is the process most responsible for creating regional gene duplications in the genome. Repeated rounds of unequal crossing over cause the homogenization of the two sequences. With the increase in the duplicates, unequal crossing over can lead to dosage imbalanc
https://en.wikipedia.org/wiki/ICT%201900%20series
ICT 1900 was a family of mainframe computers released by International Computers and Tabulators (ICT) and later International Computers Limited (ICL) during the 1960s and 1970s. The 1900 series was notable for being one of the few non-American competitors to the IBM System/360, enjoying significant success in the European and British Commonwealth markets. Origins In early 1963, ICT was engaged in negotiations to buy the computer business of Ferranti. In order to sweeten the deal, Ferranti demonstrated to ICT the Ferranti-Packard 6000 (FP6000) machine, which had been developed by its Canadian subsidiary Ferranti-Packard, to a design known as Harriac that had been initiated in Ferranti by Harry Johnson and fleshed out by Stanley Gill and John Iliffe. The FP6000 was an advanced design, notably including hardware support for multiprogramming. ICT considered using the FP6000 as their medium-sized processor in the 1965–1968 timeframe, replacing the ICT 1302. Another plan being considered was to license a new range of machines being developed by RCA, probably compatible with the expected IBM 8000. On 7 April 1964 IBM announced the System/360 series, a family of compatible machines spanning nearly the complete range of customer needs. It was immediately obvious that ICT would need a coherent response. Two paths were available: develop a range of machines based on the FP6000, using the flexibility of its design to produce smaller or larger machines, or cooperate with RCA who were re-targeting their development to a System/360 compatible range to be known as the RCA Spectra 70. One major consideration was that the FP6000 was already running, while the RCA Spectra range would take some years to become available. In the end, the decision was made to go with a range of machines based on the FP6000. The centrepiece of the new range was the ICT 1904, a version of the FP6000 with the ICT standard peripheral interface. For higher-end machines, a new larger processor,
https://en.wikipedia.org/wiki/Non-expanding%20horizon
A non-expanding horizon (NEH) is an enclosed null surface whose intrinsic structure is preserved. An NEH is the geometric prototype of an isolated horizon which describes a black hole in equilibrium with its exterior from the quasilocal perspective. It is based on the concept and geometry of NEHs that the two quasilocal definitions of black holes, weakly isolated horizons and isolated horizons, are developed. Definition of NEHs A three-dimensional submanifold ∆ is defined as a generic (rotating and distorted) NEH if it respects the following conditions: (i) ∆ is null and topologically ; (ii) Along any null normal field tangent to ∆, the outgoing expansion rate vanishes; (iii) All field equations hold on ∆, and the stress–energy tensor on ∆ is such that is a future-directed causal vector () for any future-directed null normal . Condition (i) is fairly trivial and just states the general fact that from a 3+1 perspective an NEH ∆ is foliated by spacelike 2-spheres ∆'=S2, where S2 emphasizes that ∆' is topologically compact with genus zero (). The signature of ∆ is (0,+,+) with a degenerate temporal coordinate, and the intrinsic geometry of a foliation leaf ∆'=S2 is nonevolutional. The property in condition (ii) plays a pivotal role in defining NEHs and the rich implications encoded therein will be extensively discussed below. Condition (iii) makes one feel free to apply the Newman–Penrose (NP) formalism of Einstein-Maxwell field equations to the horizon and its near-horizon vicinity; furthermore, the very energy inequality is motivated from the dominant energy condition and is a sufficient condition for deriving many boundary conditions of NEHs. Note: In this article, following the convention set up in refs., "hat" over the equality symbol means equality on the black-hole horizons (NEHs), and "hat" over quantities and operators (, , etc.) denotes those on a foliation leaf of the horizon. Also, ∆ is the standard symbol for both an NEH and the directional deri
https://en.wikipedia.org/wiki/Treviso%20Arithmetic
The Treviso Arithmetic, or Arte dell'Abbaco, is an anonymous textbook in commercial arithmetic written in vernacular Venetian and published in Treviso, Italy, in 1478. The author explains the motivation for writing this textbook: The Treviso Arithmetic is the earliest known printed mathematics book in the West, and one of the first printed European textbooks dealing with a science. The Arithmetic as an early printed book There appears to have been only one edition of the work. David Eugene Smith translated parts of the Treviso Arithmetic for educational purposes in 1907. Frank J. Swetz translated the complete work using Smith's notes in 1987 in his Capitalism & Arithmetic: The New Math of the 15th Century. Swetz used a copy of the Treviso housed in the Manuscript Library at Columbia University. The volume found its way to this collection via a curious route. Maffeo Pinelli (1785), an Italian bibliophile, is the first known owner. After his death his library was purchased by a London book-dealer and sold at auction on February 6, 1790. The book was obtained for three shillings by Mr. Wodhull. About 100 years later the Arithmetic appeared in the library of Brayton Ives, a New York lawyer. When Ives sold the collection of books at auction, George Arthur Plimpton, a New York publisher, acquired the Treviso and made it an acquisition to his extensive collection of early scientific texts. Plimpton donated his library to Columbia University in 1936. Original copies of the Treviso Arithmetic are extremely rare. There are 123 pages of text with 32 lines of print to a page. The pages are unnumbered, untrimmed and have wide margins. Some of the margins contain written notes. The size of the book is 14.5 cm by 20.6 cm. The book included information taken from the 1202 Liber Abaci, such as lattice multiplication. George G. Joseph in Crest of the Peacock suggests that John Napier read this book to create Napier's bones (or rods). Reasons for publication The Treviso Arithm
https://en.wikipedia.org/wiki/Elementary%20modes
Elementary modes may be considered minimal realizable flow patterns through a biochemical network that can sustain a steady state. This means that elementary modes cannot be decomposed further into simpler pathways. All possible flows through a network can be constructed from linear combinations of the elementary modes. The set of elementary modes for a given network is unique (up to an arbitrary scaling factor). Given the fundamental nature of elementary modes in relation to uniqueness and non-decomposability, the term `pathway' can be defined as an elementary mode. Note that the set of elementary modes will change as the set of expressed enzymes change during transitions from one cell state to another. Mathematically, the set of elementary modes is defined as the set of flux vectors, , that satisfy the steady state condition, where is the stoichiometry matrix, is the vector of rates, the vector of steady state floating (or internal) species and , the vector of system parameters. An important condition is that the rate of each irreversible reaction must be non-negative, . A more formal definition is given by: An elementary mode, , is defined as a vector of fluxes, , such that the three conditions listed in the following criteria are satisfied. The vector must satisfy: , that is: the steady state condition. For all irreversible reactions: . This means that all flow patterns must use reactions that proceed in their most natural direction. This makes the pathway described by the elementary mode a thermodynamically feasible pathway. The vector must be elementary. That is, it should not be possible to generate by combining two other vectors that satisfy the first and second requirements using the same set of enzymes that appear as non-zero entries in . In other words, it should not be possible to decompose into two other pathways that can themselves sustain a steady state. This is called elementarity. A more formal test is that the null space of the subm
https://en.wikipedia.org/wiki/N-Acetylserotonin
N-Acetylserotonin (NAS), also known as normelatonin, is a naturally occurring chemical intermediate in the endogenous production of melatonin from serotonin. It also has biological activity in its own right, including acting as a melatonin receptor agonist, an agonist of the TrkB, and having antioxidant effects. Biological function Like melatonin, NAS is an agonist at the melatonin receptors MT1, MT2, and MT3, and may be considered to be a neurotransmitter. In addition, NAS is distributed in some areas of the brain where serotonin and melatonin are not, suggesting that it may have unique central duties of its own instead of merely functioning as a precursor in the synthesis of melatonin. NAS is known to have anti-depressant, neurotrophic and cognition-enhancing effects and has been proposed to be a target for the treatment of aging-associated cognitive decline and depression TrkB receptor NAS has been shown to act as a potent TrkB receptor agonist, while serotonin and melatonin do not. Subchronic and chronic administration of NAS to adult mice induces proliferation of neural progenitor cells (NPC)s, blockage of TrkB abolished this effect suggesting that it is TrkB-dependent. NAS was also found to significantly enhance NPC proliferation in sleep-deprived mice. It is thought that the anti-depressant and neurotrophic effects of NAS are in part due to its role as a TrkB agonist. Antioxidant properties NAS acts as a potent antioxidant, NAS effectiveness as an anti-oxidant has been found to be different depending on the experimental model used, it has been described as being between 5 and 20 times more effect than melatonin at protecting against oxidant damage. NAS has been shown to protect against lipid peroxidation in microsomes and mitochondria. NAS has also been reported to lower resting levels of ROS in peripheral blood lymphocytes and to exhibit anti-oxidant effects against t-butylated hydroperoxide- and diamide-induced ROS. NAS has also been observed to inhibit
https://en.wikipedia.org/wiki/Lipid%20signaling
Lipid signaling, broadly defined, refers to any biological cell signaling event involving a lipid messenger that binds a protein target, such as a receptor, kinase or phosphatase, which in turn mediate the effects of these lipids on specific cellular responses. Lipid signaling is thought to be qualitatively different from other classical signaling paradigms (such as monoamine neurotransmission) because lipids can freely diffuse through membranes (see osmosis). One consequence of this is that lipid messengers cannot be stored in vesicles prior to release and so are often biosynthesized "on demand" at their intended site of action. As such, many lipid signaling molecules cannot circulate freely in solution but, rather, exist bound to special carrier proteins in serum. Sphingolipid second messengers Ceramide Ceramide (Cer) can be generated by the breakdown of sphingomyelin (SM) by sphingomyelinases (SMases), which are enzymes that hydrolyze the phosphocholine group from the sphingosine backbone. Alternatively, this sphingosine-derived lipid (sphingolipid) can be synthesized from scratch (de novo) by the enzymes serine palmitoyl transferase (SPT) and ceramide synthase in organelles such as the endoplasmic reticulum (ER) and possibly, in the mitochondria-associated membranes (MAMs) and the perinuclear membranes. Being located in the metabolic hub, ceramide leads to the formation of other sphingolipids, with the C1 hydroxyl (-OH) group as the major site of modification. A sugar can be attached to ceramide (glycosylation) through the action of the enzymes, glucosyl or galactosyl ceramide synthases. Ceramide can also be broken down by enzymes called ceramidases, leading to the formation of sphingosine, Moreover, a phosphate group can be attached to ceramide (phosphorylation) by the enzyme, ceramide kinase. It is also possible to regenerate sphingomyelin from ceramide by accepting a phosphocholine headgroup from phosphatidylcholine (PC) by the action of an enzyme called
https://en.wikipedia.org/wiki/The%20Lady%20Vanishes
The Lady Vanishes is a 1938 British mystery thriller film directed by Alfred Hitchcock, starring Margaret Lockwood and Michael Redgrave. Written by Sidney Gilliat and Frank Launder, based on the 1936 novel The Wheel Spins by Ethel Lina White, the film is about a beautiful English tourist travelling by train in continental Europe who discovers that her elderly travelling companion seems to have disappeared from the train. After her fellow passengers deny ever having seen the elderly lady, the young woman is helped by a young musicologist, the two proceeding to search the train for clues to the old lady's disappearance. The Lady Vanishes was filmed at the Gainsborough Studios in Islington, London. Hitchcock caught Hollywood's attention with the film and moved to Hollywood soon after its release. Although the director's three previous efforts had done poorly at the box office, The Lady Vanishes was widely successful, and confirmed American producer David O. Selznick's belief that Hitchcock indeed had a future in Hollywood cinema. The British Film Institute ranked The Lady Vanishes the 35th best British film of the 20th century. In 2017, a poll of 150 actors, directors, writers, producers and critics for Time Out magazine saw it ranked the 31st best British film ever. It is one of Hitchcock's most renowned British films, and the first of three screen versions of White's novel to date. Plot After visiting the fictional country of Bandrika, English tourist Iris Henderson is returning home to get married, but an avalanche blocks the railway line. The stranded passengers are forced to spend the night at a hotel. In the same predicament are Charters and Caldicott, cricket enthusiasts anxious to see the last days of a Test match in Manchester, and Miss Froy, a governess and music teacher. Miss Froy listens to a folk singer in the street, but he is strangled to death by an unseen murderer. That evening, Iris is bothered by a loud noise from the room above hers. It is cause
https://en.wikipedia.org/wiki/Shinobi%20%281987%20video%20game%29
is a side-scrolling hack and slash video game produced by Sega, originally released for arcades on the Sega System 16 board in 1987. The player controls ninja Joe Musashi, to stop the Zeed terrorist organization from kidnapping students of his clan. Shinobi was a commercial success in arcades; it topped the monthly Japanese table arcade charts in December 1987, and became a blockbuster arcade hit in the United States, where it was the highest-grossing conversion kit of 1988 and one of the top five conversion kits of 1989. It was adapted by Sega to its Master System game console, followed by conversions to the Nintendo Entertainment System, PC Engine, and home computers. It was re-released as downloadable emulated versions of the original arcade game for the Wii and Xbox 360. The arcade game joined the Nintendo Switch in January 2020 through the Sega Ages series. Shinobis success inspired various sequels and spin-offs of the Shinobi series. Gameplay The controls of Shinobi consist of an eight-way joystick and three action buttons for attacking, jumping, and using ninjutsu techniques called "ninja magic". The player can walk, or perform a crouching walk by pressing the joystick diagonally downward. The player can jump to higher or lower floors by pressing the jump button while holding the joystick up or down. The protagonist Joe Musashi's standard weapons are an unlimited supply of shurikens, and punches and kicks. Rescuing certain hostages in each stage will grant him an attack upgrade replacing throwing stars with a gun, and his close-range attack becomes a katana slash. Musashi's ninjutsu techniques can only be used once per stage and will clear the screen of all enemies, or greatly damage a boss. Depending on the stage, the three ninjutsu techniques are a thunderstorm, a tornado, and a doppelganger attack. Enemies include punks, mercenaries, ninjas, and the Mongolian swordsmen guarding each hostage. Musashi can bump into most enemies without harm and can only
https://en.wikipedia.org/wiki/Glossary%20of%20ecology
This glossary of ecology is a list of definitions of terms and concepts in ecology and related fields. For more specific definitions from other glossaries related to ecology, see Glossary of biology, Glossary of evolutionary biology, and Glossary of environmental science. A B C D E F G H I J K L M N O P Q R S T U V W X Y Z See also Outline of ecology History of ecology
https://en.wikipedia.org/wiki/Visual%20Prolog
Visual Prolog, previously known as PDC Prolog and Turbo Prolog, is a strongly typed object-oriented extension of Prolog. As Turbo Prolog, it was marketed by Borland but it is now developed and marketed by the Danish firm PDC that originally created it. Visual Prolog can build Microsoft Windows GUI-applications, console applications, DLLs (dynamic link libraries), and CGI-programs. It can also link to COM components and to databases by means of ODBC. Visual Prolog contains a compiler which generates x86 and x86-64 machine code. Unlike standard Prolog, programs written in Visual Prolog are statically typed. This allows some errors to be caught at compile-time instead of run-time. History Hanoi example In the Towers of Hanoi example, the Prolog inference engine figures out how to move a stack of any number of progressively smaller disks, one at a time, from the left pole to the right pole in the described way, by means of a center as transit, so that there's never a bigger disk on top of a smaller disk. The predicate hanoi takes an integer indicating the number of disks as an initial argument. class hanoi predicates hanoi : (unsigned N). end class hanoi implement hanoi domains pole = left; center; right. clauses hanoi(N) :- move(N, left, center, right). class predicates move : (unsigned N, pole A, pole B, pole C). clauses move(0, _, _, _) :- !. move(N, A, B, C) :- move(N-1, A, C, B), stdio::writef("move a disc from % pole to the % pole\n", A, C), move(N-1, B, A, C). end implement hanoi goal console::init(), hanoi::hanoi(4). Reception Bruce F. Webster of BYTE praised Turbo Prolog in September 1986, stating that it was the first Borland product to excite him as much as Turbo Pascal did. He liked the user interface and low price, and reported that two BYU professors stated that it was superior to the Prolog they used at the university. While q
https://en.wikipedia.org/wiki/PCell
PCell stands for parameterized cell, a concept used widely in the automated design of analog integrated circuits. A PCell represents a part or a component of the circuit whose structure is dependent on one or more parameters. Hence, it is a cell which is automatically generated by electronic design automation (EDA) software based on the values of these parameters. For example, one can create a transistor PCell and then use different instances of the same with different user defined lengths and widths. Vendors of EDA software sometimes use different names for the concept of parameterized cells, e.g. T-Cell and Magic Cell. Application In electronic circuit designs, cells are basic units of functionality. A given cell may be placed or instantiated many times. A P-Cell is more flexible than a non-parameterized cell because different instances may have different parameter values and, therefore, different structures. For example, rather than having many different cell definitions to represent the variously sized transistors in a given design, a single PCell may take a transistor's dimensions (width and length) as parameters. Different instances of a single PCell can then represent transistors of different sizes, but otherwise similar characteristics. The structures within an integrated circuit and the rules (design rules) governing their physical dimensions are often complex, thereby making the structures tedious to draw by hand. By using PCells a circuit designer can easily generate a large number of various structures that only differ in a few parameters, thus increasing design productivity and consistency. Most often, PCell implies a physical PCell, i.e., a physical representation of an electronic component describing its physical structure inside an integrated circuit (IC). Although most PCells are physical PCells, device symbols in circuit schematics may also be implemented as PCells. Underlying characteristics of all PCells are a dependence on (input) parame
https://en.wikipedia.org/wiki/Swiss%20annuity
A Swiss annuity simply refers to a fixed or variable annuity marketed from Switzerland or issued by a Swiss based life insurance company but has no legal definition. Insurance brokers promoting annuity contracts issued by insurance companies domiciled in jurisdictions outside of Switzerland, such as Liechtenstein, also market such contracts as Swiss annuities. The hallmarks of a Swiss annuity generally include the ability to invest in multiple currencies, the custody of assets within Switzerland, and the flexibility of withdrawals. Often touted benefits of a Swiss annuity include the safety of Switzerland plus some degree of asset protection. U.S. tax considerations For U.S. taxpayers owning a fixed annuity issued by a non-U.S. insurance company, including a Swiss annuity, the interest credited within the policy is subject to U.S. income tax on an annual basis under the original issue discount rules. Variable annuities issued by non-U.S. insurance companies may permit tax deferral in the same manner as an annuity issued by a U.S. domestic insurance company presuming the diversification requirements and investor control limitations of the code and as articulated by the IRS are respected. A Swiss annuity is considered a foreign financial account and reportable to the U.S. Department of Treasury each year by filing Form TD F 90-22.1 (the "FBAR"). Further, a Swiss annuity generally is considered a specified foreign financial asset, which must be reported annually using IRS Form 8938 assuming the value is above certain thresholds. It is unclear whether U.S. taxpayers can exchange an existing domestically-issued life insurance policy or annuity contract for a Swiss annuity on a tax-deferred basis as a result of the ambiguity of the wording of Section 1035(c) of the Internal Revenue Code of 1986, as amended. The U.S. Treasury Department and IRS have not issued regulations applying what was meant by Section 1035(c), which provides "To the extent provided in regulat
https://en.wikipedia.org/wiki/ICMP%20tunnel
An ICMP tunnel establishes a covert connection between two remote computers (a client and proxy), using ICMP echo requests and reply packets. An example of this technique is tunneling complete TCP traffic over ping requests and replies. Technical details ICMP tunneling works by injecting arbitrary data into an echo packet sent to a remote computer. The remote computer replies in the same manner, injecting an answer into another ICMP packet and sending it back. The client performs all communication using ICMP echo request packets, while the proxy uses echo reply packets. In theory, it is possible to have the proxy use echo request packets (which makes implementation much easier), but these packets are not necessarily forwarded to the client, as the client could be behind a translated address (NAT). This bidirectional data flow can be abstracted with an ordinary serial line. ICMP tunneling is possible because RFC 792, which defines the structure of ICMP packets, allows for an arbitrary data length for any type 0 (echo reply) or 8 (echo message) ICMP packets. Uses ICMP tunneling can be used to bypass firewalls rules through obfuscation of the actual traffic. Depending on the implementation of the ICMP tunneling software, this type of connection can also be categorized as an encrypted communication channel between two computers. Without proper deep packet inspection or log review, network administrators will not be able to detect this type of traffic through their network. Mitigation One way to prevent this type of tunneling is to block ICMP traffic, at the cost of losing some network functionality that people usually take for granted (e.g. it might take tens of seconds to determine that a peer is offline, rather than almost instantaneously). Another method for mitigating this type of attack is to only allow fixed sized ICMP packets through firewalls, which can impede or eliminate this type of behavior. ICMP-tunnels are sometimes used to circumvent firewalls th
https://en.wikipedia.org/wiki/AMD%20Instinct
AMD Instinct is AMD's brand of professional GPUs. It replaced AMD's FirePro S brand in 2016. Compared to the Radeon brand of mainstream consumer/gamer products, the Instinct product line is intended to accelerate deep learning, artificial neural network, and high-performance computing/GPGPU applications. The Radeon Instinct product line directly competes with Nvidia's Ampere and Intel Xeon Phi and incoming Intel Xe lines of machine learning and GPGPU cards. Before MI100 introduction in November 2020, the Instinct family was known as AMD Radeon Instinct, AMD dropped the Radeon brand from its name. Supercomputers based on (AMD CPUs and) AMD Instinct GPUs now take the lead on the Green500 supercomputer list with over 50% lead over any other, and top the first 4 spots, including the second, which is the current fastest in the world on the TOP500 list, Frontier. Products The three initial Radeon Instinct products were announced on December 12, 2016, and released on June 20, 2017, with each based on a different architecture. MI6 The MI6 is a passively cooled, Polaris 10 based card with 16 GB of GDDR5 memory and with a <150 W TDP. At 5.7 TFLOPS (FP16 and FP32), the MI6 is expected to be used primarily for inference, rather than neural network training. The MI6 has a peak double precision (FP64) compute performance of 358 GFLOPS. MI8 The MI8 is a Fiji based card, analogous to the R9 Nano, and expected to have a <175W TDP. The MI8 has 4 GB of High Bandwidth Memory. At 8.2 TFLOPS (FP16 and FP32), the MI8 is marked toward inference. The MI8 has a peak (FP64) double precision compute performance 512 GFLOPS. MI25 The MI25 is a Vega based card, utilizing HBM2 memory. The MI25 performance is expected to be 12.3 TFLOPS using FP32 numbers. In contrast to the MI6 and MI8, the MI25 is able to increase performance when using lower precision numbers, and accordingly is expected to reach 24.6 TFLOPS when using FP16 numbers. The MI25 is rated at <300W TDP with passive cooling.
https://en.wikipedia.org/wiki/Milk%20fiber
Milk fiber or milk wool is a type of Azlon, a regenerated protein fiber based on the casein protein found in milk. There are several trade names for milk-casein-based fibers, including Lanital, Fibrolane and Aralac. Invention and history First produced and patented in Italy in 1935 by Antonio Ferretti and sold under the name Lanital, milk fiber was created under an Italian national self-sufficiency drive and was intended to capitalize on previous successes with rayon. Milk fibers enjoyed a brief period of success in the 1930s and 1940s. The popularity of milk fibers declined rapidly once full-synthetic fibers were developed. Fully synthetic fibers such as acrylic were able to significantly undercut milk fiber on price while being more durable. During the 2010s several producers tried to reintroduce milk fibers to commercial production. Production process The production process of milk fiber was of some public interest and was documented on film by several contemporary sources. A simplified overview of the process is as follows: Acid is mixed with milk to extract the casein. Water is evaporated to form casein crystals. The casein is hydrated to a thick syrup and extruded through spinnerets. The resulting fiber is passed through a hardening bath. The continuous fiber is then cut to the desired length.
https://en.wikipedia.org/wiki/Residue%20%28complex%20analysis%29
In mathematics, more specifically complex analysis, the residue is a complex number proportional to the contour integral of a meromorphic function along a path enclosing one of its singularities. (More generally, residues can be calculated for any function that is holomorphic except at the discrete points {ak}k, even if some of them are essential singularities.) Residues can be computed quite easily and, once known, allow the determination of general contour integrals via the residue theorem. Definition The residue of a meromorphic function at an isolated singularity , often denoted , , or , is the unique value such that has an analytic antiderivative in a punctured disk . Alternatively, residues can be calculated by finding Laurent series expansions, and one can define the residue as the coefficient a−1 of a Laurent series. The definition of a residue can be generalized to arbitrary Riemann surfaces. Suppose is a 1-form on a Riemann surface. Let be meromorphic at some point , so that we may write in local coordinates as . Then, the residue of at is defined to be the residue of at the point corresponding to . Examples Residue of a monomial Computing the residue of a monomial makes most residue computations easy to do. Since path integral computations are homotopy invariant, we will let be the circle with radius . Then, using the change of coordinates we find that hence our integral now reads as Application of monomial residue As an example, consider the contour integral where C is some simple closed curve about 0. Let us evaluate this integral using a standard convergence result about integration by series. We can substitute the Taylor series for into the integrand. The integral then becomes Let us bring the 1/z5 factor into the series. The contour integral of the series then writes Since the series converges uniformly on the support of the integration path, we are allowed to exchange integration and summation. The series of the
https://en.wikipedia.org/wiki/Numbers%20%28TV%20series%29
Numbers (stylized as NUMB3RS) is an American crime drama television series that was broadcast on CBS from January 23, 2005, to March 12, 2010, for six seasons and 118 episodes. The series was created by Nicolas Falacci and Cheryl Heuton, and follows FBI Special Agent Don Eppes (Rob Morrow) and his brother Charlie Eppes (David Krumholtz), a college mathematics professor and prodigy, who helps Don solve crimes for the FBI. Brothers Ridley and Tony Scott produced Numbers; its production companies are the Scott brothers' Scott Free Productions and CBS Television Studios (originally Paramount Network Television, and later CBS Paramount Network Television). The show focuses equally on the relationships among Don Eppes, his brother Charlie Eppes, and their father, Alan Eppes (Judd Hirsch), and on the brothers' efforts to fight crime, usually in Los Angeles. A typical episode begins with a crime, which is subsequently investigated by a team of FBI agents led by Don and mathematically modeled by Charlie, with the help of Larry Fleinhardt (Peter MacNicol) and Amita Ramanujan (Navi Rawat). The insights provided by Charlie's mathematics were always in some way crucial to solving the crime. On May 18, 2010, CBS canceled the series after six seasons. Cast and characters The show revolved around three intersecting groups of characters: the FBI, scientists at the fictitious California Institute of Science (CalSci), and the Eppes family. Don Eppes (Rob Morrow), Charlie's older brother, is the lead FBI agent at the Los Angeles Violent Crimes Squad. Professor Charlie Eppes (David Krumholtz) is a mathematical genius, who in addition to teaching at CalSci, consults for the FBI and NSA. Alan Eppes (Judd Hirsch) is a former L.A. city planner, a widower, and the father of both Charlie and Don Eppes. Alan lives in a historic two-story California bungalow furnished with period Arts and Crafts furniture. David Sinclair (Alimi Ballard) is an FBI field agent and was later made Don's se
https://en.wikipedia.org/wiki/Unicore
Unicore is the name of a computer instruction set architecture designed by the Microprocessor Research and Development Center (MPRC) of Peking University in the PRC. The computer built on this architecture is called the Unity-863. The CPU is integrated into a fully functional SoC to make a PC-like system. The processor is very similar to the ARM architecture, but uses a different instruction set. It is supported by the Linux kernel as of version 2.6.39. Support will be removed in Linux kernel version 5.9 as nobody seems to maintain it and the code is falling behind the rest of the kernel code and compiler requirements. Instruction set The instructions are almost identical to the standard ARM formats, except that conditional execution has been removed, and the bits reassigned to expand all the register specifiers to 5 bits. Likewise, the immediate format is 9 bits rotated by a 5-bit amount (rather than 8 bit rotated by 4), the load/store offset sizes are 14 bits for byte/word and 10 bits for signed byte or half-word. Conditional moves are provided by encoding the condition in the (unused by ARM) second source register field Rn for MOV and MVN instructions. The meaning of various flag bits (such as S=1 enables setting the condition codes) is identical to the ARM instruction set. The load/store multiple instruction can only access half of the register set, depending on the H bit. If H=0, the 16 bits indicate R0–R15; if H=1, R16–R31.
https://en.wikipedia.org/wiki/Buddleja%20%C3%97%20weyeriana
Buddleja × weyeriana is one of the more remarkable Buddleja hybrids, the first crossing of an Asiatic species (B. davidii) with a South American (B. globosa). The hybrid was raised during the First World War by the eponymous Major William van de Weyer at his home, Smedmore House, at Corfe Castle, England. Van de Weyer was hoping to achieve an inflorescence the size of davidii with the colour of globosa, but met with only limited success. Description The F1 progeny of × weyeriana was a disappointment, the inflorescences small and the colours drab. Van de Weyer rightly deduced the colour was a function of a recessive trait, and reputedly crossed the F1 plants to raise the superior F2 cultivars 'Golden Glow' and 'Moonlight', although both retained vestiges of the lilac colouring of davidii. However, Moore considered it more likely the cultivars arose from backcrossing the F1 hybrids with B. globosa. The hybrid's nectar is relatively complex, comprising three constituents in almost equal proportions: sucrose, fructose, and glucose, whereas the nectar of the common Buddleja davidii is almost exclusively sucrose. It was not until half a century later that the colour objective was realized, when a Dutch nursery cloned a pure orange-yellow sport of 'Golden Glow' to produce 'Sungold'. An inflorescence the size of davidii with this desired colour remains (2014) elusive. Ploidy: 2n = 76. Cultivation Buddleja × weyeriana 'Golden Glow' is the original plant of van de Weyer exhibited in 1920. A specimen is grown as part of the NCCPG national collection held by the Longstock Park Nursery. together with the cultivars 'Sungold', 'Moonlight', and Buddleja × weyeriana 'Variegata'. The authenticity of the other cultivars remains a matter of contention.
https://en.wikipedia.org/wiki/Enron%20Corpus
The Enron Corpus is a database of over 600,000 emails generated by 158 employees of the Enron Corporation in the years leading up to the company's collapse in December 2001. The corpus was generated from Enron email servers by the Federal Energy Regulatory Commission (FERC) during its subsequent investigation. A copy of the email database was subsequently purchased for $10,000 by Andrew McCallum, a computer scientist at the University of Massachusetts Amherst. He released this copy to researchers, providing a trove of data that has been used for studies on social networking and computer-mediated communication. Creation In the legal investigation into Enron's collapse, the discovery process required collecting and preserving vast amounts of data, for which the FERC hired Aspen Systems (now part of Lockheed Martin). The emails were collected at Enron Corporation headquarters in Houston during two weeks in May 2002 by Joe Bartling, a litigation support and data analysis contractor for Aspen. In addition to the Enron employee emails, all of Enron's enterprise database systems, hosted in Oracle databases on Sun Microsystems servers, were captured and preserved, including its online energy trading platform, EnronOnline. Once collected, the Enron emails were processed and hosted in proprietary electronic discovery platforms (first Concordance, then iCONECT) for review by investigators from the FERC, Commodity Futures Trading Commission, and Department of Justice. At the conclusion of the investigation, and upon the issuance of the FERC staff report, the emails and information collected were deemed to be in the public domain, to be used for historical research and academic purposes. The email archive was made publicly available and searchable via the web using iCONECT 24/7, but the sheer volume of email of over 160GB made it impractical to use. Copies of the collected emails and databases were made available on hard drives. Jitesh Shetty and Jafar Adibi from the Univ
https://en.wikipedia.org/wiki/Trollius%20%C3%97%20cultorum
Trollius × cultorum is a group of hybrid flowering plants of garden origin, belonging to the buttercup family Ranunculaceae. There are several cultivars, derived from T. europaeus, T. asiaticus and T. chinensis. These are clump-forming herbaceous perennials whose preferred location is heavy, moist or even boggy ground, in full sun or partial shade. Typically growing to tall, they bear showy double flowers up to in diameter. Flowers appear in shades of cream, yellow and orange. The curved “petals” are actually sepals, surrounding the smaller, nectar-bearing petals. The spherical or cupped shape of the blooms gives rise to the common name globeflower, which they share with other Trollius species. The Latin specific epithet cultorum means “relating to gardens, cultivation”. Trollius × cultorum is an unresolved name, meaning that it has not yet been accepted as a correct botanical name or synonym. However, plants are widely offered in the horticultural trade. The cultivars ‘Superbus’ and ‘Orange Princess’ have gained the Royal Horticultural Society’s Award of Garden Merit. Other cultivars include ‘Alabaster’, ‘Lemon Queen’ and ‘New Moon’.
https://en.wikipedia.org/wiki/List%20of%20blindness%20effects
There are several psychological and physiological effects that cause blindness to some visual stimulus. Banner blindness or ad blindness, consciously or subconsciously ignoring banner-like advertisements at web pages. Change blindness, the inability to detect some changes in busy scenes. Choice blindness, a result in a perception experiment by Petter Johansson and colleagues. Color blindness, a color vision deficiency. Cortical blindness, a loss of vision caused by damage to the visual area in the brain. Flash blindness, a visual impairment following exposure to a light flash. Hysterical blindness (nowadays known as conversion disorder), the appearance of neurological symptoms without a neurological cause. Inattentional blindness or perceptual blindness, failing to notice some stimulus that is in plain sight. Motion blindness, a neuropsychological disorder causing an inability to perceive motion. See also Blindness (disambiguation) Cognitive psychology Physiology
https://en.wikipedia.org/wiki/Hilbert%27s%20arithmetic%20of%20ends
In mathematics, specifically in the area of hyperbolic geometry, Hilbert's arithmetic of ends is a method for endowing a geometric set, the set of ideal points or "ends" of a hyperbolic plane, with an algebraic structure as a field. It was introduced by German mathematician David Hilbert. Definitions Ends In a hyperbolic plane, one can define an ideal point or end to be an equivalence class of limiting parallel rays. The set of ends can then be topologized in a natural way and forms a circle. This usage of end is not canonical; in particular the concept it indicates is different from that of a topological end (see End (topology) and End (graph theory)). In the Poincaré disk model or Klein model of hyperbolic geometry, every ray intersects the boundary circle (also called the circle at infinity or line at infinity) in a unique point, and the ends may be identified with these points. However, the points of the boundary circle are not considered to be points of the hyperbolic plane itself. Every hyperbolic line has exactly two distinct ends, and every two distinct ends are the ends of a unique line. For the purpose of Hilbert's arithmetic, it is expedient to denote a line by the ordered pair (a, b) of its ends. Hilbert's arithmetic fixes arbitrarily three distinct ends, and labels them as 0, 1, and ∞ ;. The set H on which Hilbert defines a field structure is the set of all ends other than ∞, while H denotes the set of all ends including ∞. Addition Hilbert defines the addition of ends using hyperbolic reflections. For every end x in H, its negation −x is defined by constructing the hyperbolic reflection of line (x,∞) across the line (0,∞), and choosing −x to be the end of the reflected line. The composition of any three hyperbolic reflections whose axes of symmetry all share a common end is itself another reflection, across another line with the same end. Based on this "three reflections theorem", given any two ends x and y in H, Hilbert defines the sum x + y
https://en.wikipedia.org/wiki/Ivan%20Ku%C5%A1%C4%8Der
Ivan Kuščer (17 June 1918 – 31 January 2000) was a Slovenian physicist. He was one of the founders of the school of physics at the University of Ljubljana and a prominent member of the international research group in the field of transport theory, contributing to the theoretical aspects of the diffusion of light, neutron science and dynamics in rarefied gases. Teaching Kuščer was associated with the Department of Physics at the University of Ljubljana for over 50 years. He graduated from the University in 1941 with majors in mathematics and chemistry, and a minor in physics, becoming a docent in physics in 1945 and receiving his Ph.D. in 1951. He was elected full professor at the Faculty of Natural Sciences and Technology in 1963. He invested vital effort in developing courses to build up the physics program on a level of established programs abroad. Together with a colleague Anton Moljk, he organized postgraduate studies for physics teachers. The duo authored a series of three textbooks in Slovenian covering basic undergraduate physics concepts; the book series also served as a college textbook due to the comprehensive and precise treatment of the subject. Later Kuščer coauthored two university textbooks, on thermodynamics with S. Žumer, and on mathematical physics with Alojz Kodre, also published in German by Springer. Research His early research was on classical radiative transfer using the Boltzmann equation because of his interest in the optics of underwater light fields: this was also the theme of his doctoral dissertation. In later research, he turned to solving classical one-dimensional transport problems for radiative transfer, including polarized light and thermal neutrons, using the Boltzmann equation. In the early 1960s he was a major contributor to the development of the singular eigenfunction method for solving neutral particle transport problems. He moved on to investigate discrete relaxation times in neutron thermalization and relaxation constant
https://en.wikipedia.org/wiki/Biskit
Biskit is an open source software package that facilitates research in structural bioinformatics and molecular modelling. Written in Python, it consists of: An object-oriented programming library for manipulating and analyzing macromolecular structures, protein complexes and molecular dynamics trajectories A set of programs for solving specific tasks, such as automatic prediction of protein structures by homology modeling, and possible prediction of protein complex structures through flexible protein-protein docking The library delegates many calculations to more specialized third-party software. It currently utilizes 15 external applications, including X-PLOR, Hex, T-Coffee, DSSP and MODELLER. The latest Biskit version, 2.4.0, was released on 4 Mar 2012. It was originally developed at the Pasteur Institute. The name "Biskit" refers to the research group's name, Unité de BioInformatique Structurale. External links Structural bioinformatics software Molecular modelling Physics software Computational chemistry software Free science software Free software programmed in Python Molecular dynamics
https://en.wikipedia.org/wiki/Peripheral%20tolerance
In immunology, peripheral tolerance is the second branch of immunological tolerance, after central tolerance. It takes place in the immune periphery (after T and B cells egress from primary lymphoid organs). Its main purpose is to ensure that self-reactive T and B cells which escaped central tolerance do not cause autoimmune disease. Peripheral tolerance prevents immune response to harmless food antigens and allergens, too. Deletion of self-reactive T cells in the thymus is only 60-70% efficient, and naive T cell repertoire contains a significant portion of low-avidity self-reactive T cells. These cells can trigger an autoimmune response, and there are several mechanisms of peripheral tolerance to prevent their activation. Antigen-specific mechanisms of peripheral tolerance include persistent of T cell in quiescence, ignorance of antigen and direct inactivation of effector T cells by either clonal deletion, conversion to regulatory T cells (Tregs) or induction of anergy. Tregs, which are also generated during thymic T cell development, further suppress the effector functions of conventional lymphocytes in the periphery. Dendritic cells (DCs) participate in the negative selection of autoreactive T cells in the thymus, but they also mediate peripheral immune tolerance through several mechanisms. Dependence of a particular antigen on either central or peripheral tolerance is determined by its abundance in the organism. B cell peripheral tolerance is much less studied and is largely mediated by B cell dependence on T cell help. Cells mediating peripheral tolerance Regulatory T cells Tregs are the central mediators of immune suppression and they play a key role in maintaining peripheral tolerance. The master regulator of Treg phenotype and function is Foxp3. Natural Tregs (nTregs) are generated in the thymus during the negative selection. TCR of nTregs shows a high affinity for self-peptides, Induced Tregs (iTreg) develop from conventional naive helper T cells a
https://en.wikipedia.org/wiki/Gallium%20manganese%20arsenide
Gallium manganese arsenide, chemical formula is a magnetic semiconductor. It is based on the world's second most commonly used semiconductor, gallium arsenide, (chemical formula ), and readily compatible with existing semiconductor technologies. Differently from other dilute magnetic semiconductors, such as the majority of those based on II-VI semiconductors, it is not paramagnetic but ferromagnetic, and hence exhibits hysteretic magnetization behavior. This memory effect is of importance for the creation of persistent devices. In , the manganese atoms provide a magnetic moment, and each also acts as an acceptor, making it a p-type material. The presence of carriers allows the material to be used for spin-polarized currents. In contrast, many other ferromagnetic magnetic semiconductors are strongly insulating and so do not possess free carriers. is therefore a candidate as a spintronic material. Growth Like other magnetic semiconductors, is formed by doping a standard semiconductor with magnetic elements. This is done using the growth technique molecular beam epitaxy, whereby crystal structures can be grown with atom layer precision. In the manganese substitute into gallium sites in the GaAs crystal and provide a magnetic moment. Because manganese has a low solubility in GaAs, incorporating a sufficiently high concentration for ferromagnetism to be achieved proves challenging. In standard molecular beam epitaxy growth, to ensure that a good structural quality is obtained, the temperature the substrate is heated to, known as the growth temperature, is normally high, typically ~600 °C. However, if a large flux of manganese is used in these conditions, instead of being incorporated, segregation occurs where the manganese accumulate on the surface and form complexes with elemental arsenic atoms. This problem was overcome using the technique of low-temperature molecular beam epitaxy. It was found, first in and then later used for , that by utilising non-equilibriu
https://en.wikipedia.org/wiki/Vaughan%27s%20identity
In mathematics and analytic number theory, Vaughan's identity is an identity found by that can be used to simplify Vinogradov's work on trigonometric sums. It can be used to estimate summatory functions of the form where f is some arithmetic function of the natural integers n, whose values in applications are often roots of unity, and Λ is the von Mangoldt function. Procedure for applying the method The motivation for Vaughan's construction of his identity is briefly discussed at the beginning of Chapter 24 in Davenport. For now, we will skip over most of the technical details motivating the identity and its usage in applications, and instead focus on the setup of its construction by parts. Following from the reference, we construct four distinct sums based on the expansion of the logarithmic derivative of the Riemann zeta function in terms of functions which are partial Dirichlet series respectively truncated at the upper bounds of and , respectively. More precisely, we define and , which leads us to the exact identity that This last expansion implies that we can write where the component functions are defined to be We then define the corresponding summatory functions for to be so that we can write Finally, at the conclusion of a multi-page argument of technical and at times delicate estimations of these sums, we obtain the following form of Vaughan's identity when we assume that , , and : It is remarked that in some instances sharper estimates can be obtained from Vaughan's identity by treating the component sum more carefully by expanding it in the form of The optimality of the upper bound obtained by applying Vaughan's identity appears to be application-dependent with respect to the best functions and we can choose to input into equation (V1). See the applications cited in the next section for specific examples that arise in the different contexts respectively considered by multiple authors. Applications Vaughan's identity has been used to s
https://en.wikipedia.org/wiki/Gastric%20glands
The gastric glands are glands in the lining of the stomach that play an essential role in the process of digestion. All of the glands have mucus-secreting foveolar cells. Mucus lines the entire stomach, and protects the stomach lining from the effects of hydrochloric acid released from other cells in the glands. There are two types of gland in the stomach, the oxyntic gland, and the pyloric gland. The major type of gastric gland is the oxyntic gland that is present in 80 per cent of the stomach, and is often referred to simply as the gastric gland. The oxyntic gland is an exocrine gland and contains the parietal cells that produce hydrochloric acid, and intrinsic factor. Intrinsic factor is necessary for the absorption of vitamin B12. The other type of gland in the stomach is the pyloric gland found in the pyloric region taking up the remaining 20 per cent of the stomach area. The pyloric gland secretes gastrin from its G cells. Pyloric glands are similar in structure to the oxyntic glands but are endocrine glands with hardly any parietal cells. Types of gland The gastric glands are glands in the lining of the stomach that play an essential role in the process of digestion. All of the glands have mucus-secreting foveolar cells. Mucus lines the entire stomach of protects the stomach lining from the effects of hydrochloric acid released from other cells in the glands. Gastric glands are mostly exocrine glands and are all located beneath the gastric pits within the gastric mucosa—the mucous membrane of the stomach. The gastric mucosa is pitted with innumerable gastric pits which each house 3-5 gastric glands. The cells of the exocrine glands are foveolar (mucus), chief cells, and parietal cells. The other type of gastric gland is the pyloric gland which is an endocrine gland that secretes the hormone gastrin produced by its G cells. The cardiac glands are found in the cardia of the stomach which is the part nearest to the heart, enclosing the opening where the
https://en.wikipedia.org/wiki/Cantharellus%20minor
Cantharellus minor is a fungus native to eastern North America. It is one of the smallest of the genus Cantharellus, which includes other edible chanterelles. It is suspected of being mycorrhizal, found in association with oaks and moss. Recently, C. minor has been reported from semi-evergreen to evergreen forests in the Western Ghats, Kerala, India forming ectomycorrhizal associations with tree species like Vateria indica, Diospyros malabarica, Hopea parviflora, and Myristica species. The cap of C. minor ranges from wide and is convex and umbonate, often shallowly depressed, becoming funnel-shaped in some. The yellowish gills are decurrent, and fade to yellowish white in maturity. The stipe is less than tall. They fruit in the summer and fall. Although insubstantial, they are edible.
https://en.wikipedia.org/wiki/Calais%20%28Reuters%20product%29
Calais is a service created by Thomson Reuters that automatically extracts semantic information from web pages in a format that can be used on the semantic web. Calais was launched in January 2008, and is free to use. The technology is now available via the website of Refinitiv, a provider of financial market data and infrastructure founded in 2018, that is a subsidiary of London Stock Exchange Group. The Calais Web service reads unstructured text and returns Resource Description Framework formatted results identifying entities, facts and events within the text. The service appears to be based on technology acquired when Reuters purchased ClearForest in 2007. The technology has also been used to automatically tag blog articles, and organize museum collections. Calais uses natural language processing technologies delivered via a web service interface.
https://en.wikipedia.org/wiki/Deflate
In computing, Deflate (stylized as DEFLATE, and also called Flate) is a lossless data compression file format that uses a combination of LZ77 and Huffman coding. It was designed by Phil Katz, for version 2 of his PKZIP archiving tool. Deflate was later specified in RFC 1951 (1996). Katz also designed the original algorithm used to construct Deflate streams. This algorithm was patented as , and assigned to PKWARE, Inc. As stated in the RFC document, an algorithm producing Deflate files was widely thought to be implementable in a manner not covered by patents. This led to its widespread use – for example, in gzip compressed files and PNG image files, in addition to the ZIP file format for which Katz originally designed it. The patent has since expired. Stream format A Deflate stream consists of a series of blocks. Each block is preceded by a 3-bit header: First bit: Last-block-in-stream marker: 1: This is the last block in the stream. 0: There are more blocks to process after this one. Second and third bits: Encoding method used for this block type: 00: A stored (a.k.a. raw or literal) section, between 0 and 65,535 bytes in length 01: A static Huffman compressed block, using a pre-agreed Huffman tree defined in the RFC 10: A dynamic Huffman compressed block, complete with the Huffman table supplied 11: Reserved—don't use. The stored block option adds minimal overhead and is used for data that is incompressible. Most compressible data will end up being encoded using method 10, the dynamic Huffman encoding, which produces an optimized Huffman tree customized for each block of data individually. Instructions to generate the necessary Huffman tree immediately follow the block header. The static Huffman option is used for short messages, where the fixed saving gained by omitting the tree outweighs the percentage compression loss due to using a non-optimal (thus, not technically Huffman) code. Compression is achieved through two steps: The matching and
https://en.wikipedia.org/wiki/Cystidium
A cystidium (: cystidia) is a relatively large cell found on the sporocarp of a basidiomycete (for example, on the surface of a mushroom gill), often between clusters of basidia. Since cystidia have highly varied and distinct shapes that are often unique to a particular species or genus, they are a useful micromorphological characteristic in the identification of basidiomycetes. In general, the adaptive significance of cystidia is not well understood. Classification By position Cystidia may occur on the edge of a lamella (or analogous hymenophoral structure) (cheilocystidia), on the face of a lamella (pleurocystidia), on the surface of the cap (dermatocystidia or pileocystidia), on the margin of the cap (circumcystidia) or on the stipe (caulocystidia). Especially the pleurocystidia and cheilocystidia are important for identification within many genera. Sometimes the cheilocystidia give the gill edge a distinct colour which is visible to the naked eye or with a hand lens. By morphology Chrysocystidia are cystidia whose contents contain a distinct refractive yellow body, that becomes more deeply yellow when exposed to ammonia or other alkaline compounds. Chrysocystidia are characteristic of many (though not all) members of the agaric family Strophariaceae. Gloeocystidia have an oily or granular appearance under the microscope. Like gloeohyphae, they may be yellowish or clear (hyaline) and can sometimes selectively be coloured by sulphovanillin or other reagents. Metuloids are thick-walled cystidia with an apex having any of several distinct shapes.
https://en.wikipedia.org/wiki/Elongated%20pentagonal%20orthocupolarotunda
In geometry, the elongated pentagonal orthocupolarotunda is one of the Johnson solids (). As the name suggests, it can be constructed by elongating a pentagonal orthocupolarotunda () by inserting a decagonal prism between its halves. Rotating either the cupola or the rotunda through 36 degrees before inserting the prism yields an elongated pentagonal gyrocupolarotunda (). Formulae The following formulae for volume and surface area can be used if all faces are regular, with edge length a:
https://en.wikipedia.org/wiki/Tricholoma%20caligatum
Tricholoma caligatum is a mushroom of the agaric genus Tricholoma. It is a large species with a distinct sheathing ring on the stem, found in mycorrhizal association with various trees throughout the Mediterranean. It is sometimes referred to as the European Matsutake, though it is certainly gastronomically inferior to the true Matsutake (Tricholoma matsutake), a related species highly prized in Japan. Taxonomy and name Tricholoma caligatum was originally described in 1834 as "Agaricus caligatus" and was transferred to genus Tricholoma in 1914. Considerable controversy exists regarding the application of this name to Central European and North American collections, which likely represent different species. The name caligatum (Latin: 'boot') refers to the appearance of the mushroom, which looks like it is wearing a boot due to the presence of dark fibrils on the lower portion of the stem. Description The cap (pileus) is hemispherical at first, soon becoming convex to flat, reaching 12–15 cm in diameter, and it is covered in large, chestnut to dark-brown fibrous scales or patches. The inrolled cap edge tends to host cottony bits of partial veil. The gills (lamellae) are adnate to sinuate, crowded, whitish to cream (and reddish-brown-spotted in age). The stem (stipe) is 4–12 cm long, tapering and somewhat rooting at the base, and has a well-developed cottony ring covering the gills when young. Below the ring the stem is covered in dark bands of scales, which are the same colour as the cap. The flesh is thick and fibrous, and has a distinct, spicy, penetrating smell resembling nutmeg or allspice. The spore print is white. Ecology and distribution Tricholoma caligatum is a strictly Southern species, locally common in Mediterranean woods. It is associated with several conifers, such as Pinus pinea, Pinus halepensis, Pinus nigra, Pinus brutia, and Pinus pinaster, but also with evergreen oaks, strawberry trees and terebinth bushes. It appears in coastal and high altitud
https://en.wikipedia.org/wiki/No-arbitrage%20bounds
In financial mathematics, no-arbitrage bounds are mathematical relationships specifying limits on financial portfolio prices. These price bounds are a specific example of good–deal bounds, and are in fact the greatest extremes for good–deal bounds. The most frequent nontrivial example of no-arbitrage bounds is put–call parity for option prices. In incomplete markets, the bounds are given by the subhedging and superhedging prices. The essence of no-arbitrage in mathematical finance is excluding the possibility of "making money out of nothing" in the financial market. This is necessary because the existence of arbitrage is not only unrealistic, but also contradicts the possibility of an economic equilibrium. All mathematical models of financial markets have to satisfy a no-arbitrage condition to be realistic models. See also Box spread Indifference price
https://en.wikipedia.org/wiki/Timeline%20of%20crystallography
This is a timeline of crystallography. 18th Century 1723 – Moritz Anton Cappeller introduces the term ‘crystallography’. 1766 – Pierre-Joseph Macquer, in his Dictionnaire de Chymie, promotes mechanisms of crystallization based on the idea that crystals are composed of polyhedral molecules (primitive integrantes). 1772 – Jean-Baptiste L. Romé de l'Isle develops geometrical ideas on crystal structure in his Essai de Cristallographie. He also described the twinning phenomenon in crystals. 1781 – Abbé René Just Haüy (often termed the "Father of Modern Crystallography") discovers that crystals always cleave along crystallographic planes. Based on this observation, and the fact that the inter-facial angles in each crystal species always have the same value, Haüy concluded that crystals must be periodic and composed of regularly arranged rows of tiny polyhedra (molécules intégrantes). This theory explained why all crystal planes are related by small rational numbers (the law of rational indices). 1783 – Jean-Baptiste L. Romé de l'Isle in the second edition of his Cristallographie uses the contact goniometer to discover the law of constant interfacial angles: angles are constant and characteristic for crystals of the same chemical substance. 1784 – René Just Haüy publishes his Law of Decrements: a crystal is composed of molecules arranged periodically in three dimensions. 1795 – René Just Haüy lectures on his Law of Symmetry: “[…] the manner in which Nature creates crystals is always obeying [...] the law of the greatest possible symmetry, in the sense that oppositely situated but corresponding parts are always equal in number, arrangement, and form of their faces”. 19th Century 1801 – René Just Haüy publishes his multi-volume Traité de Minéralogie in Paris. A second edition under the title Traité de Cristallographie was published in 1822. 1815 – René Just Haüy publishes his Law of Symmetry. 1815 – Christian Samuel Weiss, founder of the dynamist school of crysta
https://en.wikipedia.org/wiki/A%E2%88%9E-operad
{{DISPLAYTITLE:A∞-operad}} In the theory of operads in algebra and algebraic topology, an A∞-operad is a parameter space for a multiplication map that is homotopy coherently associative. (An operad that describes a multiplication that is both homotopy coherently associative and homotopy coherently commutative is called an E∞-operad.) Definition In the (usual) setting of operads with an action of the symmetric group on topological spaces, an operad A is said to be an A∞-operad if all of its spaces A(n) are Σn-equivariantly homotopy equivalent to the discrete spaces Σn (the symmetric group) with its multiplication action (where n ∈ N). In the setting of non-Σ operads (also termed nonsymmetric operads, operads without permutation), an operad A is A∞if all of its spaces A(n) are contractible. In other categories than topological spaces, the notions of homotopy and contractibility have to be replaced by suitable analogs, such as homology equivalences in the category of chain complexes. An-operads The letter A in the terminology stands for "associative", and the infinity symbols says that associativity is required up to "all" higher homotopies. More generally, there is a weaker notion of An-operad (n ∈ N), parametrizing multiplications that are associative only up to a certain level of homotopies. In particular, A1-spaces are pointed spaces; A2-spaces are H-spaces with no associativity conditions; and A3-spaces are homotopy associative H-spaces. A∞-operads and single loop spaces A space X is the loop space of some other space, denoted by BX, if and only if X is an algebra over an -operad and the monoid π0(X) of its connected components is a group. An algebra over an -operad is referred to as an -space. There are three consequences of this characterization of loop spaces. First, a loop space is an -space. Second, a connected -space X is a loop space. Third, the group completion of a possibly disconnected -space is a loop space. The importance of -operads in hom
https://en.wikipedia.org/wiki/SeqA%20protein
In molecular biology the SeqA protein is found in bacteria and archaea. The function of this protein is highly important in DNA replication. The protein negatively regulates the initiation of DNA replication at the origin of replication, in Escherichia coli, OriC. Additionally the protein plays a further role in sequestration. The importance of this protein is vital, without its help in DNA replication, cell division and other crucial processes could not occur. This protein domain is thought to be part of a much larger protein complex which includes other proteins such as SeqB. Function DNA replication is an energy consuming process and hence in bacteria the process only occurs at a specific checkpoint in the cell cycle. The binding of SeqA protein to hemimethylated GATC sequences is important in the negative modulation of chromosomal initiation at oriC, and in the formation of SeqA foci necessary for Escherichia coli chromosome segregation. SeqA tetramers are able to aggregate or multimerize in a reversible, concentration-dependent manner. Apart from its function in the control of DNA replication, SeqA may also be a specific transcription factor. Additionally, SeqA is also thought to have a role in chromosome organisation and gene regulation. Localisation Most of the SeqA in the cell is found bound to new DNA, at the replication fork. Structure N terminal domain The N-terminal domain folds into two alpha-helices and one beta-strand. This protein domain is vital in assisting multimerisation. C terminal domain The C-terminal protein domain has an important role in binding to DNA. It binds to fully methylated and hemimethylated GATC sequences at oriC. The structure of the C-terminal domain consists of seven alpha-helices and a three-stranded beta-sheet.
https://en.wikipedia.org/wiki/PICO%20process
The PICO process (or framework) is a mnemonic used in evidence-based practice (and specifically evidence-based medicine) to frame and answer a clinical or health care related question, though it is also argued that PICO "can be used universally for every scientific endeavour in any discipline with all study designs". The PICO framework is also used to develop literature search strategies, for instance in systematic reviews. The PICO acronym has come to stand for: PPatient, problem or population IIntervention CComparison, control or comparator OOutcome(s) (e.g. pain, fatigue, nausea, infections, death) An application that covers clinical questions about interventions, as well as exposures, risk/ prognostic factors, and test accuracy, is: PPatient, problem or population IInvestigated condition (e.g. intervention, exposure, risk/ prognostic factor, or test result) CComparison condition (e.g. intervention, exposure, risk/ prognostic factor, or test result respectively) OOutcome(s) (e.g. symptom, syndrome, or disease of interest) Alternatives such as SPICE and PECO (among many others) can also be used. Some authors suggest adding T and S, as follows: T - Timing (e.g. duration of intervention, or date of publication) S - Study type (e.g. randomized controlled trial, cohort study, etc.) PICO as a universal technique It was argued that PICO may be useful for every scientific endeavor even beyond clinical settings. This proposal is based on a more abstract view of the PICO mnemonic, equating them with four components that is inherent to every single research, namely (1) research object; (2) application of a theory or method; (3) alternative theories or methods (or the null hypothesis); and (4) the ultimate goal of knowledge generation. This proposition would imply that the PICO technique could be used for teaching academic writing even beyond medical disciplines. Examples Clinical question: "In children with headache, is paracetamol more effective than
https://en.wikipedia.org/wiki/Molecular%20modeling%20on%20GPUs
Molecular modeling on GPU is the technique of using a graphics processing unit (GPU) for molecular simulations. In 2007, NVIDIA introduced video cards that could be used not only to show graphics but also for scientific calculations. These cards include many arithmetic units (, up to 3,584 in Tesla P100) working in parallel. Long before this event, the computational power of video cards was purely used to accelerate graphics calculations. What was new is that NVIDIA made it possible to develop parallel programs in a high-level application programming interface (API) named CUDA. This technology substantially simplified programming by enabling programs to be written in C/C++. More recently, OpenCL allows cross-platform GPU acceleration. Quantum chemistry calculations and molecular mechanics simulations (molecular modeling in terms of classical mechanics) are among beneficial applications of this technology. The video cards can accelerate the calculations tens of times, so a PC with such a card has the power similar to that of a cluster of workstations based on common processors. GPU accelerated molecular modelling software Programs Abalone – Molecular Dynamics (Benchmark) ACEMD on GPUs since 2009 Benchmark AMBER on GPUs version Ascalaph on GPUs version – Ascalaph Liquid GPU AutoDock – Molecular docking BigDFT Ab initio program based on wavelet BrianQC Quantum chemistry (HF and DFT) and molecular mechanics Blaze ligand-based virtual screening CHARMM – Molecular dynamics CP2K Ab initio molecular dynamics Desmond (software) on GPUs, workstations, and clusters Firefly (formerly PC GAMESS) FastROCS GOMC – GPU Optimized Monte Carlo simulation engine GPIUTMD – Graphical processors for Many-Particle Dynamics GPUMD - A light weight general-purpose molecular dynamics code GROMACS on GPUs HALMD – Highly Accelerated Large-scale MD package HOOMD-blue – Highly Optimized Object-oriented Many-particle Dynamics—Blue Edition LAMMPS on GPUs version – lammps for accelerators LIO
https://en.wikipedia.org/wiki/International%20Knockout%20Mouse%20Consortium
The International Knockout Mouse Consortium (IKMC) is a scientific endeavour to produce a collection of mouse embryonic stem cell lines that together lack every gene in the genome, and then to distribute the cells to scientific researchers to create knockout mice to study. Many of the targeted alleles are designed so that they can generate both complete and conditional gene knockout mice. The IKMC was initiated on March 15, 2007, at a meeting in Brussels. By 2011, Nature reported that approximately 17,000 different genes have already been disabled by the consortium, "leaving only around 3,000 more to go". The consortium encompasses four major, high-throughput gene-targeted mutagenesis programs: the National Institutes of Health (NIH)-sponsored Knockout Mouse Program (KOMP) and state-funded Texas Institute for Genomic Medicine (TIGM) in the U.S., the North American Conditional Mouse Mutagenesis (NorCOMM) Program in Canada, and the European Conditional Mouse Mutagenesis (EUCOMM) Programme in Europe. The first of its annual meetings of members and funders, hosted by the country of its rotating chair, was held at the NIH in Bethesda, Maryland, in the United States for 2007–2008, with Toronto, Canada, hosting for 2008–2009.
https://en.wikipedia.org/wiki/Atomic%20Energy%20Project
Atomic Energy Project was started at the University of Rochester as a graduate teaching program. Also known as the University of Rochester Atomic Energy Project or URAEP. Atomic Energy Project genesis The Atomic Energy Project was the continuation of the Manhattan Project after the end of World War II. The idea was to continue the work of training the people necessary for the peaceful use of atomic energy and nuclear materials. Key personnel Stafford L. Warren directed the Department of Radiology at Rochester and the Atomic Energy Project had three divisions. William Freer Bale headed the Radiology and Biophysics division that worked largely on radioactive materials — for example, Radium, Radon, Plutonium, and Polonium. James Newell Stannard was responsible for 2 sections, the Radiation Toxicology section and the Radioautography section. Harold Hodge headed the Pharmacology and Toxicology division that focused on Uranium including inhalation studies. Joe Wiseman Howland, M.D., Ph.D. headed the Medical Services division. Herbert Mermagen worked in the Medical Physics section as a radiological physicist, known today as a health physicist. Radiation Dose Reconstruction The National Institute for Occupational Safety and Health (NIOSH) part of the Centers for Disease Control and Prevention started the Radiation Dose Reconstruction NIOSH program area for the University of Rochester Atomic Energy Project. Employees of the U.S. Department of Energy, earlier agencies, and contractors and subcontractors that worked at the University of Rochester Atomic Energy Project in Rochester, New York, from 1 September 1943 - 30 October 1971, for at least 250 work days were included in the study. Additional employees included Laboratory Technicians that worked in the University of Rochester Atomic Energy Project laboratory building from 1 September 1943 to 19 June 1945.
https://en.wikipedia.org/wiki/Road%20ecology
Road ecology is the study of the ecological effects (both positive and negative) of roads and highways (public roads). These effects may include local effects, such as on noise, water pollution, habitat destruction/disturbance and local air quality; and the wider environmental effects of transport such as habitat fragmentation, ecosystem degradation, and climate change from vehicle emissions. The design, construction and management of roads, parking and other related facilities as well as the design and regulation of vehicles can change their effect. Roads are known to cause significant damage to forests, prairies, streams and wetlands. Besides the direct habitat loss due to the road itself, and the roadkill of animal species, roads alter water-flow patterns, increase noise, water, and air pollution, create disturbance that alters the species composition of nearby vegetation thereby reducing habitat for local native animals, and act as barriers to animal movements. Roads are a form of linear infrastructure intrusion that has some effects similar to infrastructure such as railroads, power lines, and canals, particularly in tropical forests. Road ecology is practiced as a field of inquiry by a variety of ecologists, biologists, hydrologists, engineers, and other scientists. There are several global centers for the study of road ecology: 1) The Road Ecology Center at the University of California, Davis, which was the first of its kind in the world; 2) the Centro Brasileiro de Estudos em Ecologia de Estradas at the Federal University of Lavras, Brazil; 3) The Center for Transportation and the Environment, North Carolina State University; and 4) the Road Ecology Program at the Western Transportation Institute, Montana State University. There are also several important global conferences for road ecology research: 1) Infra-Eco Network Europe (IENE), which is international, but focused primarily on Europe; 2) International Conference on Ecology and Transportation (ICOET
https://en.wikipedia.org/wiki/Digital%20Postmarks
A Digital Postmark (DPM) is a technology that applies a trusted time stamp issued by a postal operator to an electronic document, validates electronic signatures, and stores and archives all non-repudiation data needed to support a potential court challenge. It guarantees the certainty of date and time of the postmarking. This global standard was renamed the Electronic Postal Certification Mark (EPCM) in 2007 shortly after a new iteration of the technology was developed by Microsoft and Poste Italiane. The key addition to the traditional postmarking technology was integrity of the electronically postmarked item, meaning any kind of falsification and tampering will be easily and definitely detected. Additionally, content confidentiality is guaranteed since document certification is carried out without access or reading by the postal operator. The EPCM will eventually be available through the UPU to all international postal operators in the 191 member countries willing to be compliant with this standard, thus granting interoperability in certified communications between postal operators. In the United States, the US Postal Service operates a non-global standard called the Electronic Postmark, although it is soon expected to provide services utilizing the EPCM. Providers In the United States, until the end of 2010, Authentidate was the only authorized USPS EPM provider. However, this contract was allowed to expire. The process An electronic document is created Digital Postmarking client software signs the document locally The signed document is sent to the Digital Postmarking (DPM) service for postmarking Upon receipt, the DPM service first validates the authenticity of the signature If the signature is valid then a timestamp is generated by the DPM service as a counter-signature that includes the date and time The document, signature, validation results and timestamp are stored in the Digital Postmark non-repudiation database A Digital Postmark Receipt, including th
https://en.wikipedia.org/wiki/Relict%20%28biology%29
In biogeography and paleontology, a relict is a population or taxon of organisms that was more widespread or more diverse in the past. A relictual population is a population currently inhabiting a restricted area whose range was far wider during a previous geologic epoch. Similarly, a relictual taxon is a taxon (e.g. species or other lineage) which is the sole surviving representative of a formerly diverse group. Definition A relict (or relic) plant or animal is a taxon that persists as a remnant of what was once a diverse and widespread population. Relictualism occurs when a widespread habitat or range changes and a small area becomes cut off from the whole. A subset of the population is then confined to the available hospitable area, and survives there while the broader population either shrinks or evolves divergently. This phenomenon differs from endemism in that the range of the population was not always restricted to the local region. In other words, the species or group did not necessarily arise in that small area, but rather was stranded, or insularized, by changes over time. The agent of change could be anything from competition from other organisms, continental drift, or climate change such as an ice age. When a relict is representative of taxa found in the fossil record, and yet is still living, such an organism is sometimes referred to as a living fossil. However, a relict need not be currently living. An evolutionary relict is any organism that was characteristic of the flora or fauna of one age and that persisted into a later age, with the later age being characterized by newly evolved flora or fauna significantly different from those that came before. Examples A notable example is the thylacine of Tasmania, a relict marsupial carnivore that survived into modern times on an island, whereas the rest of its species on mainland Australia had gone extinct between 3000 and 2000 years ago. Another example is Omma, a genus of beetle with a fossil record e
https://en.wikipedia.org/wiki/Mach%27s%20principle
In theoretical physics, particularly in discussions of gravitation theories, Mach's principle (or Mach's conjecture) is the name given by Albert Einstein to an imprecise hypothesis often credited to the physicist and philosopher Ernst Mach. The hypothesis attempted to explain how rotating objects, such as gyroscopes and spinning celestial bodies, maintain a frame of reference. The proposition is that the existence of absolute rotation (the distinction of local inertial frames vs. rotating reference frames) is determined by the large-scale distribution of matter, as exemplified by this anecdote: You are standing in a field looking at the stars. Your arms are resting freely at your side, and you see that the distant stars are not moving. Now start spinning. The stars are whirling around you and your arms are pulled away from your body. Why should your arms be pulled away when the stars are whirling? Why should they be dangling freely when the stars don't move? Mach's principle says that this is not a coincidence—that there is a physical law that relates the motion of the distant stars to the local inertial frame. If you see all the stars whirling around you, Mach suggests that there is some physical law which would make it so you would feel a centrifugal force. There are a number of rival formulations of the principle, often stated in vague ways like "mass out there influences inertia here". A very general statement of Mach's principle is "local physical laws are determined by the large-scale structure of the universe". Mach's concept was a guiding factor in Einstein's development of the general theory of relativity. Einstein realized that the overall distribution of matter would determine the metric tensor which indicates which frame is stationary with respect to rotation. Frame-dragging and conservation of gravitational angular momentum makes this into a true statement in the general theory in certain solutions. But because the principle is so vague, many disti
https://en.wikipedia.org/wiki/Enterprise%20appliance%20transaction%20module
An enterprise appliance transaction module (EATM) is a device, typically used in the manufacturing automation marketplace, for the transfer of plant floor equipment and product status to manufacturing execution systems (MES), enterprise resource planning (ERP) systems and the like. Solutions that deliver manufacturing floor integration have evolved over time. Initially they took the form of custom integrated systems, designed and delivered by system integrators. These solutions were largely based on separate commercial off-the-shelf (COTS) products integrated into a custom system. Modern EATM products might not needing any software development or custom integration. Components Hardware platform – embedded computer, computer appliance Device communications software – Support for the automation protocols from which data will be extracted. Device communications software typically operates through polled or change based protocols that are vendor specific. Data to be extracted is typically organized into related items, and transferred based on a machine status such as Cycle Complete, Job Start, System Downtime Event, Operator Change, etc. Typical protocols; Rockwell Automation CIP, ControlLogix backplane, EtherNet/IP, Siemens Industrial Ethernet, Modbus TCP. There are hundreds of automation device protocols and EATM solutions are typically targeting certain market segments and will be based on automation vendor relationships. Enterprise communications software – Software that will enable communications to enterprise systems. Communications at this level are typically transaction oriented and require data transactions to be sent and acknowledged to ensure the data integrity. Examples include; Relational Database Adapters, Java Message Services (JMS), Oracle Database Interfaces and proprietary interfaces to specific products. Transaction application – Software that is configured to watch and collect device variables, formats them into required transactions,
https://en.wikipedia.org/wiki/Neuroligin
Neuroligin (NLGN), a type I membrane protein, is a cell adhesion protein on the postsynaptic membrane that mediates the formation and maintenance of synapses between neurons. Neuroligins act as ligands for β-neurexins, which are cell adhesion proteins located presynaptically. Neuroligin and β-neurexin "shake hands", resulting in the connection between two neurons and the production of a synapse. Neuroligins also affect the properties of neural networks by specifying synaptic functions, and they mediate signalling by recruiting and stabilizing key synaptic components. Neuroligins interact with other postsynaptic proteins to localize neurotransmitter receptors and channels in the postsynaptic density as the cell matures. Additionally, neuroligins are expressed in human peripheral tissues and have been found to play a role in angiogenesis. In humans, alterations in genes encoding neuroligins are implicated in autism and other cognitive disorders. Antibodies in a mother from previous male pregnancies against neuroligin 4 from the Y chromosome increase the probability of homosexuality in male offspring. Structure Neuroligins bind with the aid of Ca2+ to the α-neurexin LNS (laminin, neurexin and sex hormone-binding globulin-like folding units) domains and to the β-neurexin LNS domain which then establishes a heterophilic trans-synaptic recognition code. Through the observation of the crystal structure of neuroligin-1, it was determined that neuroligin-1 forms a protein dimer when two neurexin-1 beta monomers bind to the neuroligin-1's two opposite surfaces. This forms a heterotetramer, which contains an interface for binding Ca2+. The interaction of neuroligin and neurexin to form a heterotetramer is monitored by alternatively spliced sites located near the binding interface for Ca2+ in both the neuroligin-1 and the neurexin-1 beta. Subsequently, the presence of native neuroligin dimers was confirmed in neurons through biochemical detection, which included heterodimer
https://en.wikipedia.org/wiki/Cherrybrook%20Kitchen
Cherrybrook Kitchen is a privately held company producing baking products for the food allergy market. The company was founded in 2004 by Patsy Rosenberg and is headquartered in Burlington, Massachusetts. The company was acquired by Cell-nique in 2011. Cherrybrook Kitchen was founded in response to the growing number of children and adults diagnosed with food allergies. 11 million adults and children are affected by peanut, dairy, egg and nut allergens, while approximately one in 133 Americans suffers from Celiac Disease. Cherrybrook Kitchen offers two lines of gourmet baking mixes: Original and Gluten Free. The Original Line includes all-natural gourmet baking mixes that are free of peanuts, dairy, eggs and nuts; the gluten free line, introduced in 2005 as Gluten Free Dreams, is made with rice flour and is free of gluten and wheat in addition to peanuts, dairy, eggs and nuts. . In 2009, the firm announced a new partnership with Arthur, the award-winning book series and PBS program. The firm received a 2008 Kids Food Award from Kiwi Magazine and was recently named one of Parents Magazine's "Best Snacks for Kids with Food Allergies." Manufacturing All of Cherrybrook Kitchen's products are produced in a completely nut-free facility using dedicated equipment that do not share dairy or egg products. The mixes are routinely tested for allergen cross-contaminants. All of its products have been accepted by the Feingold Association, an organization dedicated to helping children and adults establish allergen-free diets . The mixes are also kosher, vegan certified and cholesterol free.
https://en.wikipedia.org/wiki/Minkowski%27s%20second%20theorem
In mathematics, Minkowski's second theorem is a result in the geometry of numbers about the values taken by a norm on a lattice and the volume of its fundamental cell. Setting Let be a closed convex centrally symmetric body of positive finite volume in -dimensional Euclidean space . The gauge or distance Minkowski functional attached to is defined by Conversely, given a norm on we define to be Let be a lattice in . The successive minima of or on are defined by setting the -th successive minimum to be the infimum of the numbers such that contains linearly-independent vectors of . We have . Statement The successive minima satisfy Proof A basis of linearly independent lattice vectors can be defined by . The lower bound is proved by considering the convex polytope with vertices at , which has an interior enclosed by and a volume which is times an integer multiple of a primitive cell of the lattice (as seen by scaling the polytope by along each basis vector to obtain -simplices with lattice point vectors). To prove the upper bound, consider functions sending points in to the centroid of the subset of points in that can be written as for some real numbers . Then the coordinate transform has a Jacobian determinant . If and are in the interior of and (with ) then with , where the inclusion in (specifically the interior of ) is due to convexity and symmetry. But lattice points in the interior of are, by definition of , always expressible as a linear combination of , so any two distinct points of cannot be separated by a lattice vector. Therefore, must be enclosed in a primitive cell of the lattice (which has volume ), and consequently .
https://en.wikipedia.org/wiki/A%20Guide%20to%20the%20Classification%20Theorem%20for%20Compact%20Surfaces
A Guide to the Classification Theorem for Compact Surfaces is a textbook in topology, on the classification of two-dimensional surfaces. It was written by Jean Gallier and Dianna Xu, and published in 2013 by Springer-Verlag as volume 9 of their Geometry and Computing series (, ). The Basic Library List Committee of the Mathematical Association of America has recommended its inclusion in undergraduate mathematics libraries. Topics The classification of surfaces (more formally, compact two-dimensional manifolds without boundary) can be stated very simply, as it depends only on the Euler characteristic and orientability of the surface. An orientable surface of this type must be topologically equivalent (homeomorphic) to a sphere, torus, or more general handlebody, classified by its number of handles. A non-orientable surface must be equivalent to a projective plane, Klein bottle, or more general surface characterized by an analogous number, its number of cross-caps. For compact surfaces with boundary, the only extra information needed is the number of boundary components. This result is presented informally at the start of the book, as the first of its six chapters. The rest of the book presents a more rigorous formulation of the problem, a presentation of the topological tools needed to prove the result, and a formal proof of the classification. Other topics in topology discussed as part of this presentation include simplicial complexes, fundamental groups, simplicial homology and singular homology, and the Poincaré conjecture. Appendices include additional material on embeddings and self-intersecting mappings of surfaces into three-dimensional space such as the Roman surface, the structure of finitely generated abelian groups, general topology, the history of the classification theorem, and the Hauptvermutung (the theorem that every surface can be triangulated). Audience and reception This is a textbook aimed at the level of advanced undergraduates or beginning gr