source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Biliverdin%20reductase
Biliverdin reductase (BVR) is an enzyme () found in all tissues under normal conditions, but especially in reticulo-macrophages of the liver and spleen. BVR facilitates the conversion of biliverdin to bilirubin via the reduction of a double-bond between the second and third pyrrole ring into a single-bond. There are two isozymes, in humans, each encoded by its own gene, biliverdin reductase A (BLVRA) and biliverdin reductase B (BLVRB). Mechanism of catalysis BVR acts on biliverdin by reducing its double-bond between the pyrrole rings into a single-bond. It accomplishes this using NADPH + H+ as an electron donor, forming bilirubin and NADP+ as products. BVR catalyzes this reaction through an overlapping binding site including Lys18, Lys22, Lys179, Arg183, and Arg185 as key residues. This binding site attaches to biliverdin, and causes its dissociation from heme oxygenase (HO) (which catalyzes reaction of ferric heme --> biliverdin), causing the subsequent reduction to bilirubin. Structure BVR is composed of two closely packed domains, between 247-415 amino acids long and containing a Rossmann fold. BVR has also been determined to be a zinc-binding protein with each enzyme protein having one strong-binding zinc atom. The C-terminal half of BVR contains the catalytic domain, which adopts a structure containing a six-stranded beta-sheet that is flanked on one face by several alpha-helices. This domain contains the catalytic active site, which reduces the gamma-methene bridge of the open tetrapyrrole, biliverdin IX alpha, to bilirubin with the concomitant oxidation of a NADH or NADPH cofactor. Function BVR works with the biliverdin/bilirubin redox cycle. It converts biliverdin to bilirubin (a strong antioxidant), which is then converted back into biliverdin through the actions of reactive oxygen species (ROS). This cycle allows for the neutralization of ROS, and the reuse of biliverdin products. Biliverdin also is replenished in the cycle with its format
https://en.wikipedia.org/wiki/Franti%C5%A1ek%20Wald
František "Franz" Wald (9 January 1861 – 19 October 1931) was a Czech professor of chemistry who contributed to metallurgy, analytical and physical chemistry. He questioned atomic and molecular approaches to understanding chemical phenomena. Wald was born at Brandýsek, near Slaný, where his father, originally from Chemnitz, Germany, was a foreman of a workshop of the Austrian Railways. His mother was from Karlsbad. Wald went to school at Kladno and received a grant from the Austrian State Railways to study at Prague. Although German adopted a Czech nationality. He worked at the laboratory of Pražská železářská společnost, the main ironworks in Kladno. He became a chief chemist in 1886. In 1908 he became a professor at the Czech Technical University, Prague. Wald examined chemical phenomena using the laws of thermodynamics, rather than examine them through ideas from atomic theory. He wrote on this in his Die Energie und ihre Entwertung (1888). His second book Chemie fází (Prague, 1918) examined his idea of phase as a fundamental concept rather than atoms.
https://en.wikipedia.org/wiki/Gene%20transfer%20agent
Gene transfer agents (GTAs) are DNA-containing virus-like particles that are produced by some bacteria and archaea and mediate horizontal gene transfer. Different GTA types have originated independently from viruses in several bacterial and archaeal lineages. These cells produce GTA particles containing short segments of the DNA present in the cell. After the particles are released from the producer cell, they can attach to related cells and inject their DNA into the cytoplasm.  The DNA can then become part of the recipient cells' genome. Discovery of gene transfer agents The first GTA system was discovered in 1974, when mixed cultures of Rhodobacter capsulatus strains produced a high frequency of cells with new combinations of genes. The factor responsible was distinct from known gene-transfer mechanisms in being independent of cell contact, insensitive to deoxyribonuclease, and not associated with phage production. Because of its presumed function it was named gene transfer agent (GTA, now RcGTA) More recently other gene transfer agent systems have been discovered by incubating filtered (cell-free) culture medium with a genetically distinct strain. GTA genes and evolution The genes specifying GTAs are derived from bacteriophage (phage) DNA that has integrated into a host chromosome. Such prophages often acquire mutations that make them defective and unable to produce phage particles.  Many bacterial genomes contain one or more defective prophages that have undergone more-or less-extensive mutation and deletion. Gene transfer agents, like defective prophages, arise by mutation of prophages, but they retain functional genes for the head and tail components of the phage particle (structural genes) and the genes for DNA packaging. The phage genes specifying its regulation and DNA replication have typically been deleted, and expression of the cluster of structural genes is under the control of cellular regulatory systems. Additional genes that contribute to GT
https://en.wikipedia.org/wiki/Tripartite%20motif%20family
The tripartite motif family (TRIM) is a protein family. Function Many TRIM proteins are induced by interferons, which are important component of resistance to pathogens and several TRIM proteins are known to be required for the restriction of infection by lentiviruses. TRIM proteins are involved in pathogen-recognition and by regulation of transcriptional pathways in host defence. Structure The tripartite motif is always present at the N-terminus of the TRIM proteins. The TRIM motif includes the following three domains: (1) a RING finger domain (2) one or two B-box zinc finger domains when only one B-box is present, it is always a type-2 B-box when two B-boxes are present the type-1 B-Box always precedes the type-2 B-Box (3) coiled coil region The C-terminus of TRIM proteins contain either: Group 1 proteins: a C-terminal domain selected from the following list: NHL and IGFLMN domains, either in association or alone PHD domain associated with a bromodomain MATH domain (in e.g., TRIM37) ARF domain (in e.g., TRIM23) EXOIII domain (in e.g., TRIM19) or Group 2 proteins: a SPRY C-terminal domain e.g. TRIM21 Family members The TRIM family is split into two groups that differ in domain structure and genomic organization: Group 1 members possess a variety of C-terminal domains, and are represented in both vertebrate and invertebrates Group 2 is absent in invertebrates, possess a C-terminal SPRY domain Members of the family include: Group 1 PHD-BROMO domain containing: TRIM24 (TIF1α), TRIM28 (TIF1β), TRIM33 (TIF1γ)– act as corepressors 1-10: TRIM1, TRIM2, TRIM3, TRIM8, TRIM9 11-20: TRIM12, TRIM13, TRIM14, TRIM16, TRIM18, TRIM19 21-30: TRIM23, TRIM25, TRIM29, TRIM30 31-40: TRIM32, TRIM36, TRIM37 41-50: TRIM42, TRIM44, TRIM45, TRIM46, TRIM47 51-60: TRIM51, TRIM53, TRIM54, TRIM55, TRIM56, TRIM57, TRIM59 61-70: TRIM62, TRIM63, TRIM65, TRIM66, TRIM67, TRIM69, TRIM70 71-75: TRIM71 Group 2 1-10: TRIM4, TRIM5, TRIM6, TRIM7, TRIM10 11-2
https://en.wikipedia.org/wiki/World%20History%20Encyclopedia
World History Encyclopedia (formerly Ancient History Encyclopedia) is a nonprofit educational company created in 2009 by Jan van der Crabben. The organization publishes and maintains articles, images, videos, podcasts, and interactive educational tools related to history. All users may contribute content to the site, although submissions are reviewed by an editorial team before publication. In 2021, the organization was renamed from the Ancient History Encyclopedia to World History Encyclopedia to reflect its broadened scope, covering world history from all time periods, as opposed to just ancient history. Original articles are written in English and later translated into other languages, mainly French and Spanish. Organization history The Ancient History Encyclopedia was founded in 2009 by van der Crabben with the stated goal of improving history education worldwide by creating a freely accessible and reliable history source. The nonprofit organization is based in Godalming, United Kingdom and Montreal, Canada, although it has no office and its team is globally distributed. The site had an emphasis on ancient history when it was founded, but it later shifted to cover the Medieval and early Modern periods as well. In 2021, the organization renamed itself World History Encyclopedia to reflect this change. Reception The website has received praise by educational organizations and has been recommended by the School Library Journal, the Internet Scout Research Group at the University of Wisconsin–Madison, MERLOT, and the European Commission's Open Education Europa initiative. In 2016, it won the .eu Web Award for education from the organization EURid.
https://en.wikipedia.org/wiki/Roadside%20America
Roadside America was an indoor miniature village and railway covering . Created by Laurence Gieringer in 1935, it was first displayed to the public in his Hamburg, Pennsylvania, home. The miniature village's popularity increased after stories were published about it in local newspapers, which prompted Gieringer to move it to a recently-closed local amusement park called Carsonia Park. This location, which supported more visitors, was open from 1938 to about 1940. To accommodate growing interest and build a larger display, Geringer then purchased land at what would be the miniature village's final location, a former dance hall in Shartlesville, Pennsylvania off of Interstate 78, approximately west of the Lehigh Valley, where the display reopened in 1953. After being closed since March 2020 due to the COVID-19 pandemic, Roadside America announced on November 21, 2020, that they were closing permanently after trying, unsuccessfully, to find a buyer for the past three years, and that they would be auctioning off the display. The display The 3/8 inch to one foot scale display contains: A 7,450 square foot, fully landscaped village diorama displaying over 300 miniature structures Up to 18 O gauge trains, trolleys and cable cars running throughout the display 10,000 hand-made trees 4,000 miniature people engaged in everyday daily pursuits Many rivers, streams and waterways Interactive animations such as a circus parade, construction workers, saw mill workers and more that can be activated by visitors 600 miniature light bulbs The display is constructed with: 21,500 feet of electrical wiring 17,700 board feet of lumber 6,000 feet of building paper 4,000 feet of sheet metal under the plaster work 2,250 feet of railroad track 648 feet of canvas for waterproofing 450 feet of pipe 18,000 pounds of plaster 4,000 pounds of sheet iron 900 pounds of nails 600 pounds of rubber roofing material 75 pounds of dry paint 75 gallons of liquid paint 225 bushels o
https://en.wikipedia.org/wiki/Myrmecophily
Myrmecophily ( , ) is the term applied to positive interspecies associations between ants and a variety of other organisms, such as plants, other arthropods, and fungi. Myrmecophily refers to mutualistic associations with ants, though in its more general use, the term may also refer to commensal or even parasitic interactions. The term "myrmecophile" is used mainly for animals that associate with ants. An estimated 10,000 species of ants (Formicidae) are known, with a higher diversity in the tropics. In most terrestrial ecosystems, ants are ecologically and numerically dominant, being the main invertebrate predators. As a result, ants play a key role in controlling arthropod richness, abundance, and community structure. Some evidence shows that the evolution of myrmecophilous interactions has contributed to the abundance and ecological success of ants, by ensuring a dependable and energy-rich food supply, thus providing a competitive advantage for ants over other invertebrate predators. Most myrmecophilous associations are opportunistic, unspecialized, and facultative (meaning both species are capable of surviving without the interaction), though obligate mutualisms (those in which one or both species are dependent on the interaction for survival) have also been observed for many species. As ant nests grow, they are more likely to house more and greater varieties of myrmecophiles. This is partly because larger colonies have greater specializations, so more diversity of ecology within the nests, allowing for more diversity and population sizes among the myrmecophiles. Myrmecophile A "myrmecophile" is an organism that lives in association with ants. Myrmecophiles may have various roles in their host ant colony. Many consume waste materials in the nests, such as dead ants, dead larvae, or fungi growing in the nest. Some myrmecophiles, however, feed on the stored food supplies of ants, and a few are predatory on ant eggs, larvae, or pupae. Others benefit the ants by
https://en.wikipedia.org/wiki/Sergey%20Mergelyan
Sergey Mergelyan (; 19 May 1928 – 20 August 2008) was a Soviet and Armenian mathematician, who made major contributions to the Approximation theory. The modern Complex Approximation Theory is based on Mergelyan's classical work. Corresponding Member of the Academy of Sciences of the Soviet Union (since 1953), member of NAS ASSR (since 1956). The surname "Mergelov" given at birth was changed for patriotic reasons to the more Armenian-sounding "Mergelyan" by the mathematician himself before his trip to Moscow. He was a laureate of the Stalin Prize (1952) and the Order of St. Mesrop Mashtots (2008). He was the youngest Doctor of Sciences in the history of the USSR (at the age of 20), and the youngest corresponding member of the Academy of Sciences of the Soviet Union (the title was conferred at the age of 24). During his postgraduate studies, the 20-year-old Mergelyan solved one of the fundamental problems of the mathematical theory of functions, which had not been solved for more than 70 years. His theorem on the possibility of uniform polynomial approximation of functions of a complex variable is recognized by the classical Mergelyan theorem, and is included in the course of the theory of functions. Although he himself was not a computer designer, Mergelyan was a pioneer in Soviet computational mathematics. Biography Early years Sergey Mergelyan was born on 19 May 1928 in Simferopol in an Armenian family. His father Nikita (Mkrtich) Ivanovich Mergelov, a former private entrepreneur (Nepman), his mother Lyudmila Ivanovna Vyrodova, the daughter of the manager of the Azov-Black Sea bank, who was shot in 1918. In 1936 Sergey's father was building a paper mill in Yelets, but soon together with his family was deported to the Siberian settlement of Narym, Tomsk Oblast. In the Siberian frost, Sergey suffered from a serious illness and narrowly survived. In 1937, the mother and son were acquitted by the court's decision and returned to Kerch, and in 1938 Lyudmila Ivanov
https://en.wikipedia.org/wiki/TOPIO%20Dio
The TOPIO Dio is a robot designed to serve in a restaurant or coffee shop, or as a cocktail bartender, which is manufactured by Vietnam-based company, Tosy. TOPIO Dio has a 3-wheel moving platform, 28 degrees of freedom and is operated remotely with a built-in camera and an obstacle detector. It is 125 cm high and weighs 45 kg. Development history Specifications Technologies Remote control via wireless internet Integrate 3D vision via 2 cameras 3D operation space of robot defined by the controlling software Processes pre-defined images Detects obstacles by Ultrasonic Sensor Three-wheeled base with omnidirectional and balanced motion See also TOPIO
https://en.wikipedia.org/wiki/Chebyshev%27s%20bias
In number theory, Chebyshev's bias is the phenomenon that most of the time, there are more primes of the form 4k + 3 than of the form 4k + 1, up to the same limit. This phenomenon was first observed by Russian mathematician Pafnuty Chebyshev in 1853. Description Let (x; n, m) denote the number of primes of the form nk + m up to x. By the prime number theorem (extended to arithmetic progression), That is, half of the primes are of the form 4k + 1, and half of the form 4k + 3. A reasonable guess would be that (x; 4, 1) > (x; 4, 3) and (x; 4, 1) < (x; 4, 3) each also occur 50% of the time. This, however, is not supported by numerical evidence — in fact, (x; 4, 3) > (x; 4, 1) occurs much more frequently. For example, this inequality holds for all primes x < 26833 except 5, 17, 41 and 461, for which (x; 4, 1) = (x; 4, 3). The first x such that (x; 4, 1) > (x; 4, 3) is 26861, that is, (x; 4, 3) ≥ (x; 4, 1) for all x < 26861. In general, if 0 < a, b < n are integers, gcd(a, n) = gcd(b, n) = 1, a is a quadratic residue mod n, b is a quadratic nonresidue mod n, then (x; n, b) > (x; n, a) occurs more often than not. This has been proved only by assuming strong forms of the Riemann hypothesis. The stronger conjecture of Knapowski and Turán, that the density of the numbers x for which (x; 4, 3) > (x; 4, 1) holds is 1 (that is, it holds for almost all x), turned out to be false. They, however, do have a logarithmic density, which is approximately 0.9959.... Generalizations This is for k = −4 to find the smallest prime p such that (where is the Kronecker symbol), however, for a given nonzero integer k (not only k = −4), we can also find the smallest prime p satisfying this condition. By the prime number theorem, for every nonzero integer k, there are infinitely many primes p satisfying this condition. For positive integers k = 1, 2, 3, ..., the smallest primes p are 2, 11100143, 61981, 3, 2082927221, 5, 2, 11100143, 2, 3, 577, 61463, 2083, 11, 2, 3, 2, 11100121, 5, 20829
https://en.wikipedia.org/wiki/Helly%27s%20selection%20theorem
In mathematics, Helly's selection theorem (also called the Helly selection principle) states that a uniformly bounded sequence of monotone real functions admits a convergent subsequence. In other words, it is a sequential compactness theorem for the space of uniformly bounded monotone functions. It is named for the Austrian mathematician Eduard Helly. A more general version of the theorem asserts compactness of the space BVloc of functions locally of bounded total variation that are uniformly bounded at a point. The theorem has applications throughout mathematical analysis. In probability theory, the result implies compactness of a tight family of measures. Statement of the theorem Let (fn)n ∈ N be a sequence of increasing functions mapping the real line R into itself, and suppose that it is uniformly bounded: there are a,b ∈ R such that a ≤ fn ≤ b for every n  ∈  N. Then the sequence (fn)n ∈ N admits a pointwise convergent subsequence. Generalisation to BVloc Let U be an open subset of the real line and let fn : U → R, n ∈ N, be a sequence of functions. Suppose that (fn) has uniformly bounded total variation on any W that is compactly embedded in U. That is, for all sets W ⊆ U with compact closure W̄ ⊆ U, where the derivative is taken in the sense of tempered distributions; and (fn) is uniformly bounded at a point. That is, for some t ∈ U, { fn(t) | n ∈ N } ⊆ R is a bounded set. Then there exists a subsequence fnk, k ∈ N, of fn and a function f : U → R, locally of bounded variation, such that fnk converges to f pointwise; and fnk converges to f locally in L1 (see locally integrable function), i.e., for all W compactly embedded in U, and, for W compactly embedded in U, Further generalizations There are many generalizations and refinements of Helly's theorem. The following theorem, for BV functions taking values in Banach spaces, is due to Barbu and Precupanu: Let X be a reflexive, separable Hilbert space and let E be a closed, convex subset of X. Le
https://en.wikipedia.org/wiki/Shape%20analysis%20%28digital%20geometry%29
This article describes shape analysis to analyze and process geometric shapes. Description Shape analysis is the (mostly) automatic analysis of geometric shapes, for example using a computer to detect similarly shaped objects in a database or parts that fit together. For a computer to automatically analyze and process geometric shapes, the objects have to be represented in a digital form. Most commonly a boundary representation is used to describe the object with its boundary (usually the outer shell, see also 3D model). However, other volume based representations (e.g. constructive solid geometry) or point based representations (point clouds) can be used to represent shape. Once the objects are given, either by modeling (computer-aided design), by scanning (3D scanner) or by extracting shape from 2D or 3D images, they have to be simplified before a comparison can be achieved. The simplified representation is often called a shape descriptor (or fingerprint, signature). These simplified representations try to carry most of the important information, while being easier to handle, to store and to compare than the shapes directly. A complete shape descriptor is a representation that can be used to completely reconstruct the original object (for example the medial axis transform). Application fields Shape analysis is used in many application fields: archeology for example, to find similar objects or missing parts architecture for example, to identify objects that spatially fit into a specific space medical imaging to understand shape changes related to illness or aid surgical planning virtual environments or on the 3D model market to identify objects for copyright purposes security applications such as face recognition entertainment industry (movies, games) to construct and process geometric models or animations computer-aided design and computer-aided manufacturing to process and to compare designs of mechanical parts or design objects. Shape descriptors Shape descr
https://en.wikipedia.org/wiki/Perls%20Prussian%20blue
In histology, histopathology, and clinical pathology, Perls Prussian blue is a commonly used method to detect the presence of iron in tissue or cell samples. Perls Prussian Blue derives its name from the German pathologist Max Perls (1843–1881), who described the technique in 1867. The method does not involve the application of a dye, but rather causes the pigment Prussian blue to form directly within the tissue. The method stains mostly iron in the ferric state which includes ferritin and hemosiderin, rather than iron in the ferrous state. Uses Perls's method is used to indicate "non-heme" iron in tissues such as ferritin and hemosiderin, the procedure does not stain iron that is bound to porphyrin forming heme such as hemoglobin and myoglobin. The stain is an important histochemical stain used to demonstrate the distribution and amount of iron deposits in liver tissue, often in the form of a biopsy. Perls's procedure may be used to identify excess iron deposits such as hemosiderin deposits (hemosiderosis) and in conditions such as hereditary hemochromatosis. Perls Prussian blue is commonly used on bone marrow aspirates to indicate levels of iron storage and may provide reliable evidence of iron deficiency. Method of application Perls did not publish a detailed procedure other than indicating a dilute potassium ferrocyanide solution was applied to the tissue followed by hydrochloric acid. Ferric iron deposits in tissue (present mostly as ferric iron within the storage protein ferritin) then react with the soluble ferrocyanide in the stain to form the insoluble Prussian blue pigment (a complex hydrated ferric ferrocyanide substance). These deposits are then visualizable microscopically as blue or purple deposits. Many methods of performing Perls Prussian blue stain for iron have been published, Drury and Wallington (1980) give a protocol that uses a mixture of 1 part 2% hydrochloric acid and 1 part 2% potassium ferrocyanide that is applied to the section for 2
https://en.wikipedia.org/wiki/Breakthrough%20Initiatives
Breakthrough Initiatives is a science-based program founded in 2015 and funded by Julia and Yuri Milner, also of Breakthrough Prize, to search for extraterrestrial intelligence over a span of at least 10 years. The program is divided into multiple projects. Breakthrough Listen will comprise an effort to search over 1,000,000 stars for artificial radio or laser signals. A parallel project called Breakthrough Message is an effort to create a message "representative of humanity and planet Earth". The project Breakthrough Starshot, co-founded with Mark Zuckerberg, aims to send a swarm of probes to the nearest star at about 20% the speed of light. The project Breakthrough Watch aims to identify and characterize Earth-sized, rocky planets around Alpha Centauri and other stars within 20 light years of Earth. Breakthrough plans to send a mission to Saturn's moon Enceladus, in search for life in its warm ocean, and in 2018 signed a partnership agreement with NASA for the project. History The Breakthrough Initiatives were announced to the public on 20 July 2015, at London's Royal Society by physicist Stephen Hawking. Russian tycoon Yuri Milner created the Initiatives to search for intelligent extraterrestrial life in the Universe and consider a plan for possibly transmitting messages out into space. The announcement included an open letter co-signed by multiple scientists, including Hawking, expressing support for an intensified search for alien radio communications. During the public launch, Hawking said: "In an infinite Universe, there must be other life. There is no bigger question. It is time to commit to finding the answer." The cash infusion is projected to mark up the pace of SETI research over the early 2000s rate, and will nearly double the rate NASA was spending on SETI research annually in approximately 1973–1993. Projects Breakthrough Listen Breakthrough Listen is a program to search for intelligent extraterrestrial communications in the Universe. With $10
https://en.wikipedia.org/wiki/Open%20Syllabus%20Project
The Open Syllabus Project (OSP) is an online open-source platform that catalogs and analyzes millions of college syllabi. Founded by researchers from the American Assembly at Columbia University, the OSP has amassed the most extensive collection of searchable syllabi. Since its beta launch in 2016, the OSP has collected over 7 million course syllabi from over 80 countries, primarily by scraping publicly accessible university websites. The project is directed by Joe Karaganis. History The OSP was formed by a group of data scientists, sociologists, and digital-humanities researchers at the American Assembly, a public-policy institute based at Columbia University. The OSP was partly funded by the Sloan Foundation and the Arcadia Fund. Joe Karaganis, former vice-president of the American Assembly, serves as the project director of the OSP. The project builds on prior attempts to archive syllabi, such as H-Net, MIT OpenCourseWare, and historian Dan Cohen's defunct Syllabus Finder website (Cohen now sits on the OSP's advisory board). The OSP became a non-profit and independent of the American Assembly in November 2019. In January 2016, the OSP launched a beta version of their "Syllabus Explorer," which they had collected data for since 2013. The Syllabus Explorer allows users to browse and search texts from over one million college course syllabi. The OSP launched a more comprehensive version 2.0 of the Syllabus Explorer in July 2019. The newer version includes an interactive visualization that displays texts as dots on a knowledge map. , the OSP has collected over 7 million course syllabi. The Syllabus Explorer represents the "largest collection of searchable syllabi ever amassed." Methodology The OSP has collected syllabi data from over 80 countries dating to 2000. The syllabi stem from over 4,000 worldwide institutions. Most of the OSP's data originates from the United States. Canada, Australia, and the U.K also have large datasets. The OSP primarily collects sy
https://en.wikipedia.org/wiki/Code%20talker
A code talker was a person employed by the military during wartime to use a little-known language as a means of secret communication. The term is most often used for United States service members during the World Wars who used their knowledge of Native American languages as a basis to transmit coded messages. In particular, there were approximately 400 to 500 Native Americans in the United States Marine Corps whose primary job was to transmit secret tactical messages. Code talkers transmitted messages over military telephone or radio communications nets using formally or informally developed codes built upon their Indigenous languages. The code talkers improved the speed of encryption and decryption of communications in front line operations during World War II and are credited with a number of decisive victories. Their code was never broken. There were two code types used during World War II. Type one codes were formally developed based on the languages of the Comanche, Hopi, Meskwaki, and Navajo peoples. They used words from their languages for each letter of the English alphabet. Messages could be encoded and decoded by using a simple substitution cipher where the ciphertext was the Native language word. Type two code was informal and directly translated from English into the Indigenous language. If there was no corresponding word in the Indigenous language for the military word, code talkers used short, descriptive phrases. For example, the Navajo did not have a word for submarine, so they translated it as iron fish. The term Code Talker was originally coined by the United States Marine Corps and used to identify individuals who completed the special training required to qualify as Code Talkers with their service records indicating "642 – Code Talker" as a duty assignment. Today, the term Code Talker is still strongly associated with the bilingual Navajo speakers trained in the Navajo Code during World War II by the US Marine Corps to serve in all six division
https://en.wikipedia.org/wiki/5-Aza-7-deazaguanine
5-Aza-7-deazaguanine or 2-aminoimidazo[1,2-a][1,3,5]triazin-4(1H)-one is a 5-Aza-7-deazapurine base that is an isomer of guanine. It is used as a nucleobase of hachimoji DNA, in which it pairs with 6-Amino-5-nitropyridin-2-one.
https://en.wikipedia.org/wiki/Neuronal%20cell%20cycle
The Neuronal cell cycle represents the life cycle of the biological cell, its creation, reproduction and eventual death. The process by which cells divide into two daughter cells is called mitosis. Once these cells are formed they enter G1, the phase in which many of the proteins needed to replicate DNA are made. After G1, the cells enter S phase during which the DNA is replicated. After S, the cell will enter G2 where the proteins required for mitosis to occur are synthesized. Unlike most cell types however, neurons are generally considered incapable of proliferating once they are differentiated, as they are in the adult nervous system. Nevertheless, it remains plausible that neurons may re-enter the cell cycle under certain circumstances. Sympathetic and cortical neurons, for example, try to reactivate the cell cycle when subjected to acute insults such as DNA damage, oxidative stress, and excitotoxicity. This process is referred to as “abortive cell cycle re-entry” because the cells usually die in the G1/S checkpoint before DNA has been replicated. Cell cycle regulation Transitions through the cell cycle from one phase to the next are regulated by cyclins binding their respective cyclin dependent kinases (Cdks) which then activate the kinases (Fisher, 2012). During G1, cyclin D is synthesized and binds to Cdk4/6, which in turn phosphorylates retinoblastoma (Rb) protein and induces the release of the transcription factor E2F1 which is necessary for DNA replication (Liu et al., 1998). The G1/S transition is regulated by cyclin E binding to Cdk2 which phosphorylates Rb as well (Merrick and Fisher, 2011). S phase is then driven by the binding of cyclin A with Cdk2. In late S phase, cyclin A binds with Cdk1 to promote late replication origins and also initiates the condensation of the chromatin in the late G2 phase. The G2/M phase transition is regulated by the formation of the Cdk1/cyclin B complex. Inhibition through the cell cycle is maintained by cyclin-dep
https://en.wikipedia.org/wiki/Quadrature%20%28astronomy%29
In spherical astronomy, quadrature is the configuration of a celestial object in which its elongation is perpendicular to the direction of the Sun. It is applied especially to the position of a superior planet or the Moon at its first and last quarter phases. This is not to be confused with the Moon at dichotomy (exactly half-lit) as viewed from Earth, which occurs at 89.85 degrees and 270.15 degrees. As shown in the diagram, a planet (or other object) can be at the western quadrature (when it is to the west of the Sun when viewed from the Earth) or at the eastern quadrature (when it is to the east of the Sun when viewed from the Earth). Note that an inferior planet can never be at quadrature to the reference planet. At quadrature, the shadow that a planet casts on its planetary rings or moons appears most offset from the planet (e.g., Saturn's rings); the dark side of a planet (e.g., Mars) is maximally visible. See also Astrological aspect
https://en.wikipedia.org/wiki/Yoda%20conditions
In programming jargon, Yoda conditions (also called Yoda notation) is a programming style where the two parts of an expression are reversed from the typical order in a conditional statement. A Yoda condition places the constant portion of the expression on the left side of the conditional statement. Yoda conditions are part of the coding standards for Symfony and WordPress. Origin The name for this programming style is derived from the Star Wars character Yoda, who speaks English with a non-standard syntax (e.g., "When 900 years old you reach, look as good you will not."). Thomas M. Tuerke claims to have coined the term Yoda notation and first published it online in 2006. According to him, the term Yoda condition was later popularized by Félix Cloutier in 2010. Example Usually a conditional statement would be written as: if ($value == 42) { /* ... */ } // Reads like: "If the value equals 42..." Yoda conditions describe the same expression, but reversed: if (42 == $value) { /* ... */ } // Reads like: "If 42 equals the value..." Advantage Error detections Placing the constant value in the expression does not change the behavior of the program (unless the values evaluate to false—see below). In programming languages that use a single equals sign (=) for assignment expressions and not for comparison, a possible mistake is to assign a value unintentionally instead of writing a conditional statement. if (myNumber = 42) { /* ... */ } // This assigns 42 to myNumber instead of evaluating the desired condition Using Yoda conditions: if (42 = myNumber) { /* ... */ } // A syntax error this is and compile it will not Since 42 is a constant and cannot be changed, this error will be caught by the compiler. Boolean myBoolean = null; if (myBoolean == true) { /* ... */ } // This causes a NullPointerException in Java Runtime, but legal in compilation. // This happens because Java will try to call myBoolean.booleanValue() on a null Object. Avoiding some types of unsafe n
https://en.wikipedia.org/wiki/Modeccin
Modeccin is a toxic lectin, a group of glycoproteins capable of binding specifically to sugar moieties. Different toxic lectins are present in seeds of different origin. Modeccin is found in the roots of the African plant Adenia digitata. These roots are often mistaken for edible roots, which has led to some cases of intoxication. Sometimes the fruit is eaten, or a root extract is drunk as a manner of suicide. Structure and reactivity Modeccin consists of two subunits that are bound by a disulfide linkage, the intact protein has a molecular weight of approximately 57-63 kDa. When treated with mercaptoethanol the chains can be dissociated into two subunits, subunit A with a mass of 25-28 kDa and subunit B with a mass of 31-35 kDa. The A-chain is called the effectomer and possesses ribosomal-inactivating properties. The B-chain contains the carbohydrate binding site and it is termed the haptomer. While the intact toxin molecules have potent cytotoxic effects on cells, they exhibit no ribosomal inactivating activity on ribosomes in a cell-free system. By contrast, reduction of the toxin with a disulfide reducing agent creates the opposite effects. Reduced, dissociated toxin subunits inhibit ribosomal activity in cell-free systems, but they have no effect on intact cells. The reason for these properties is due to the toxin's mode of action. Toxin molecules bind through saccharide recognition sites on the B-chain to particular β-galactosyl-containing glycoprotein or glycolipid components on the surface of cell membranes. In animals that are sensitive to these toxins these polysaccharides are present in virtually all cell types. The toxin binds to cell-surface polysaccharide receptors with a high affinity (Ka in the range of 107–108/M). When the toxin binds to the cell, the A-chain enters through either active transport or endocytosis. Once inside the cell the A-chain enters the cytoplasmic space, binds to the 60S ribosomal subunit and enzymatically inactivates it.
https://en.wikipedia.org/wiki/Cyazofamid
Cyazofamid is a fungicide that is highly-specific in controlling oomycete plant pathogens such as Phytophthora infestans, the organism which causes late blight in potato. Its mode of action is thought to involve binding to the Qi center of Coenzyme Q – cytochrome c reductase. Cyazofamid is most often sold under the brand name Ranman. Synthesis Processes to manufacture cyazofamid were disclosed in patents from Ishihara Sangyo Kaisha, Ltd. An acetophenone derivative was first treated with aqueous glyoxal and hydroxylamine to form an oxime-substituted imidazole ring system. This intermediate was treated with thionyl chloride and disulfur dichloride to convert the oxime to a cyano group and chlorinate the imidazole in the position next to the phenyl ring. Finally, treatment with dimethylsulfamoyl chloride gave cyazofamid, with the desired regiochemistry. This placed the sulfamoyl group on the nitrogen adjacent to the phenyl ring rather than the chlorine atom. The structure of the fungicide has been confirmed by X-ray crystallography.
https://en.wikipedia.org/wiki/Skhul%20and%20Qafzeh%20hominins
The Skhul and Qafzeh hominins or Qafzeh–Skhul early modern humans are hominin fossils discovered in Es-Skhul and Qafzeh caves in Israel. They are today classified as Homo sapiens, among the earliest of their species in Eurasia. Skhul Cave is on the slopes of Mount Carmel; Qafzeh Cave is a rockshelter near Nazareth in Lower Galilee. The remains found at Es Skhul, together with those found at the Nahal Me'arot Nature Reserve and Mugharet el-Zuttiyeh, were classified in 1939 by Arthur Keith and Theodore D. McCown as Palaeoanthropus palestinensis, a descendant of Homo heidelbergensis. History The remains exhibit a mix of traits found in archaic and anatomically modern humans. They have been tentatively dated at about 80,000–120,000 years old using electron paramagnetic resonance and thermoluminescence dating techniques. The brain case is similar to modern humans, but they possess brow ridges and a projecting facial profile like Neanderthals. They were initially regarded as transitional from Neanderthals to anatomically modern humans, or as hybrids between Neanderthals and modern humans. Neanderthal remains have been found nearby at Kebara Cave that date to 61,000–48,000 years ago, but it has been hypothesised that the Skhul/Qafzeh hominids had died out by 80,000 years ago because of drying and cooling conditions, favouring a return of a Neanderthal population suggesting that the two types of hominids never made contact in the region. A more recent hypothesis is that Skhul/Qafzeh hominids represent the first exodus of modern humans from Africa around 125,000 years ago, probably via the Sinai Peninsula, and that the robust features exhibited by the Skhul/Qafzeh hominids represent archaic sapiens features rather than Neanderthal features. The discovery of modern human made tools from about 125,000 years ago at Jebel Faya, United Arab Emirates, in the Arabian Peninsula, may be from an even earlier exit of modern humans from Africa. In January 2018 it was announced that m
https://en.wikipedia.org/wiki/Team%20software%20process
In combination with the personal software process (PSP), the team software process (TSP) provides a defined operational process framework that is designed to help teams of managers and engineers organize projects and produce software for products that range in size from small projects of several thousand lines of code (KLOC) to very large projects greater than half a million lines of code. The TSP is intended to improve the levels of quality and productivity of a team's software development project, in order to help them better meet the cost and schedule commitments of developing a software system. The initial version of the TSP was developed and piloted by Watts Humphrey in the late 1990s and the Technical Report for TSP sponsored by the U.S. Department of Defense was published in November 2000. The book by Watts Humphrey, Introduction to the Team Software Process, presents a view of the TSP intended for use in academic settings, that focuses on the process of building a software production team, establishing team goals, distributing team roles, and other teamwork-related activities. Introduction to TSP The primary goal of TSP is to create a team environment for establishing and maintaining a self-directed team, and supporting disciplined individual work as a base of PSP framework. Self-directed team means that the team manages itself, plans and tracks their work, manages the quality of their work, and works proactively to meet team goals. TSP has two principal components: team-building and team-working. Team-building is a process that defines roles for each team member and sets up teamwork through TSP launch and periodical relaunch. Team-working is a process that deals with engineering processes and practices utilized by the team. TSP, in short, provides engineers and managers with a way that establishes and manages their team to produce the high-quality software on schedule and budget. How TSP works Before engineers can participate in the TSP, it is require
https://en.wikipedia.org/wiki/Henyey%20track
The Henyey track is a path taken by pre-main-sequence stars with masses greater than 0.5 solar masses in the Hertzsprung–Russell diagram after the end of the Hayashi track. The astronomer Louis G. Henyey and his colleagues in the 1950s showed that the pre-main-sequence star can remain in radiative equilibrium throughout some period of its contraction to the main sequence. The Henyey track is characterized by a slow collapse in near hydrostatic equilibrium, approaching the main sequence almost horizontally in the Hertzsprung–Russell diagram (i.e. the luminosity remains almost constant). See also Historical brightest stars List of brightest stars List of most luminous stars List of nearest bright stars List of Solar System objects in hydrostatic equilibrium Stellar evolution Stellar birthline Stellar isochrone
https://en.wikipedia.org/wiki/DLX
The DLX (pronounced "Deluxe") is a RISC processor architecture designed by John L. Hennessy and David A. Patterson, the principal designers of the Stanford MIPS and the Berkeley RISC designs (respectively), the two benchmark examples of RISC design (named after the Berkeley design). The DLX is essentially a cleaned up (and modernized) simplified Stanford MIPS CPU. The DLX has a simple 32-bit load/store architecture, somewhat unlike the modern MIPS architecture CPU. As the DLX was intended primarily for teaching purposes, the DLX design is widely used in university-level computer architecture courses. There are two known "softcore" hardware implementations: ASPIDA and VAMP. The ASPIDA project resulted in a core with many nice features: it is open source, supports Wishbone, has an asynchronous design, supports multiple ISAs, and is ASIC proven. VAMP is a DLX-variant that was mathematically verified as part of Verisoft project. It was specified with PVS, implemented in Verilog, and runs on a Xilinx FPGA. A full stack from compiler to kernel to TCP/IP was built on it. History In the Stanford MIPS architecture, one of the methods used to gain performance was to force all instructions to complete in one clock cycle. This forced compilers to insert "no-ops" in cases where the instruction would definitely take longer than one clock cycle. Thus input and output activities (like memory accesses) specifically forced this behaviour, leading to artificial program bloat. In general MIPS programs were forced to have a lot of wasteful NOP instructions, a behaviour that was an unintended consequence. The DLX architecture does not force single clock cycle execution, and is therefore immune to this problem. In the DLX design a more modern approach to handling long instructions was used: data-forwarding and instruction reordering. In this case the longer instructions are "stalled" in their functional units, and then re-inserted into the instruction stream when they can complete. Ex
https://en.wikipedia.org/wiki/Cooperative%20Observer%20Program
The NOAA Cooperative Observer Program (COOP) is a citizen weather observer network run by the U.S. National Weather Service (NWS) and National Centers for Environmental Information (NCEI). Over 8,700 volunteers from the fifty states and all territories report at least daily a variety of weather conditions such as daily maximum and minimum temperatures, 24-hour precipitation totals, including snowfall, and significant weather occurrences throughout a day that are recorded via remarks in observer logs. Some stations also report stream stage or tidal levels. Daily observations are reported electronically or over the phone, and monthly logs are submitted electronically or via the mail. Many stations are located in rural areas but the network also includes long-term stations in most urban centers. Observation locations include farms, in urban and suburban areas, National Parks, seashores, and mountaintops. Volunteers are trained by local NWS offices who provide rain gauges, snowsticks, thermometers, or other instruments. Data is initially received and analyzed by local NWS offices then ultimately stored and analyzed by NCEI, which also does final data quality checks. The program began with act of Congress in 1890 and grew out a network of observers developed by the Smithsonian Institution. It was a backbone of the U.S. climatological observation network and remains an important network in providing long-term observations of particular locations. The Cooperative Weather Observer network consists of manual observations of only a few variables and consists of daily summaries rather than being continuous (i.e. real-time). Because of these limitations and other sensor limitations, as well as to attain a denser network of observations, there has been a move to supplement the coop program using automated weather stations since the 1990s. NWS sponsored programs include the Citizen Weather Observer Program (CWOP) and Community Collaborative Rain, Hail and Snow Network (CoCoRaH
https://en.wikipedia.org/wiki/Evolution%20of%20Infectious%20Disease
Evolution of Infectious Disease is a 1993 book by the evolutionary biologist Paul W. Ewald. In this book, Ewald contests the traditional view that parasites should evolve toward benign coexistence with their hosts. He draws on various studies that contradict this dogma and asserts his theory based on fundamental evolutionary principles. This book provides one of the first in-depth presentations of insights from evolutionary biology on various fields in health science, including epidemiology and medicine. Infectious diseases Infectious disease are illnesses induced by another organism. Such diseases range from mild to severe cases. The onset of infectious disease can be induced by bacteria, viruses, fungi, and parasites. Several examples of infectious diseases are as follows: tuberculosis, chickenpox, mumps, meningitis, measles, and malaria. Infectious diseases can be obtained through many routes of transmission such as inhalation, open wounds, sores, ingestion, sexual intercourse, and insect bites. Author, Paul Ewald used his book to expound upon infectious diseases in humans and animals, explain various routes of transmission as well as epidemiology as a whole. Epidemiology is defined as the study of the onset, distribution, and control of diseases. Evolutionary epidemiology focuses on the distribution of infectious diseases whereas Darwinian epidemiology focuses on human beings as hosts of infectious diseases. To fully comprehend both aspects of epidemiology, it is necessary to understand how organisms induce these diseases as well as how infected organisms counteract. Evolution The extensive research about pathogens shows that they can evolve within a month, whereas animal hosts such as humans take centuries to make large evolutionary changes. Parasite virulence and host resistance are variables that strongly impact a pathogen's ability to replicate and be distributed to many hosts. Parasite virulence is the level of harm a host endures due to a virus, bact
https://en.wikipedia.org/wiki/Logo%20of%20the%20BBC
The logo of the BBC has been a brand identity for the corporation and its work since the 1950s in a variety of designs. Until the introduction of a logo in 1958, the corporation had relied on its coat of arms for official documentation and correspondence, although it rarely appeared onscreen. With the increased role of television for the BBC in the 1960s, particularly after the foundation of ITV, the corporation used its logo to increase viewer familiarity and to standardise its image and content. The logo has since been redesigned a number of times, most recently in 2021 with the BBC blocks, a logo designed to work across media. From 1958, there have been six different BBC logos. The first logo of the network was used from 1958 to 1963, the second from 1963 to 1971, the third from 1971 to 1992, the fourth from 1988 to 1997, the fifth from 1997 to 2021, while the sixth and current logo was adopted in October 2021. History Before the logo Before the BBC introduced its logo itself, in the form of the slanted boxes, the BBC used a variety of different symbols with which to represent itself. In printed media and corporation correspondence, it used the BBC coat of arms, while on screen, it used a different logo type. Originally, it used a stylised BBC text on early equipment, not unlike the caption that accompanied the BBC1 COW globe. This logo was rarely seen on screen, with captions containing the words "BBC Television Service" along with matching clock. In 1932, when the original reception room of the BBC Broadcasting House in London opened, a logo was laid in mosaic on the floor. This logo was merely a stylized entwining of two capital B's, one facing either direction, linked by a C in the centre. This mosaic logo is still visible on the floor today, though the area no longer serves as the BBC's main reception room. 1950s In 1953, Abram Games was commissioned to design an on-air image. Nicknamed the 'bat's wings', it consisted of a rounded brass contraption with
https://en.wikipedia.org/wiki/The%20Sean-Bhean%20bhocht
"The Sean-Bhean bhocht" (; Irish for "Poor old woman"), often spelled phonetically as "Shan Van Vocht", is a traditional Irish song from the period of the Irish Rebellion of 1798, and dating in particular to the lead up to a French expedition to Bantry Bay, that ultimately failed to get ashore in 1796. The Sean-Bhean bhocht is used to personify Ireland, a poetic motif which heralds back to the aisling of native Irish language poetry. Many different versions of the song have been composed by balladeers over the years, with the lyrics adapted to reflect the political climate at the time of composition. The title of the song, tune and narration of the misfortunes of the Shean Bhean bhocht remain a constant however, and this version, probably the best known, expresses confidence in the victory of the United Irishmen in the looming rebellion upon the arrival of French aid. W. B. Yeats and Augusta, Lady Gregory based their 1902 nationalist play, Cathleen Ní Houlihan, on the legend dramatized in this song. A more recent version of the character is the speaker in Tommy Makem's "Four Green Fields," a song composed in the modern context of Northern Ireland's partition from the Republic of Ireland. The character also appears as the old lady selling milk in the opening passage of Ulysses by James Joyce. See also The Shan Van Vocht as the title of a late nineteenth century Irish nationalist novel and magazine. Mise Éire Róisín Dubh (song) Hibernia (personification) Kathleen Ni Houlihan Four Green Fields
https://en.wikipedia.org/wiki/Fifth%20principal%20meridian
The Fifth principal meridian, often denoted the "5th Meridian" or "PM 05," is a principal meridian survey line used in the United States for land claims in the Public Land Survey System (PLSS). It was first surveyed in 1815. The meridian, a north-south line, starts from the old mouth of the Arkansas River and runs north. Another survey line related to it is the base line running west from the old mouth of the St. Francis River. These survey lines govern all land surveys in four states and a large portion of the land surveys for two more. Monuments have been erected where the two lines meet at , and the surveyors' skill has been commemorated at the Louisiana Purchase State Park in eastern Arkansas. The Fifth principal meridian is nearly coincident with 91° 3′ 42″ longitude west from the Greenwich meridian. In Arkansas, Missouri, Iowa, and North Dakota, it is the sole principal meridian used in the PLSS. It is used in Minnesota west of the Mississippi River and west of the third guide meridian north of the river. In South Dakota the meridian covers roughly the eastern half of the state: all land east of the Missouri River plus the surveys on the west side of the river to a limiting line following the third guide meridian of the sixth principal meridian system, White River, and the west and north boundaries of the Lower Brule Indian Reservation. This was the first meridian of the new Louisiana Purchase, which covered about . There was much discussion regarding where the baseline would be, with the primary candidate originally being an extension of the Third principal meridian baseline. However, it was ultimately decided to use the confluence of the St. Francis River and the Mississippi River for the base line, and the confluence of the Arkansas and Mississippi for the meridian. Two survey teams started on the same day and proceed north (led by Prospect K. Robbins) and west (led by Joseph C. Brown), respectively. The initial point was located as the place where these
https://en.wikipedia.org/wiki/Kingdom%20Hearts%20Mobile
was an online community-based social gaming networking service developed by Square Enix for the NTT DoCoMo in partnership with Disney Mobile Japan. It was launched on December 15, 2008 in Japan in conjunction with the video game Kingdom Hearts coded for mobile phones. Mobile is not part of the main Kingdom Hearts storyline. It consists of various mini-games as well as downloadable Kingdom Hearts related content such as ringtones and wallpapers. The service ended on April 30, 2013. Gameplay Kingdom Hearts Mobile is a 2D online world known as the Avatar Kingdom where players can explore and roam around freely as their avatars. Avatars Players roam the Avatar Kingdom by controlling their own customized avatar which can be personalized through various means and items. These include outfits, weapons, and items which are purchasable in the game's online store using munny, which is the game's virtual currency. Avatars are used to play mini-games, meet up and chat with friends as well as do activities together within the Avatar Kingdom. Outfits as well as backgrounds can be earned by players whenever they complete an episode of Kingdom Hearts coded. Areas There are many areas located within the Avatar Kingdom, including areas for downloading items, socializing, and playing mini-games. The areas are as follows: Event Hall which serves as the main area for people to participate together with friends in activities; Point Bank where players can keep track of their points as well as munny that they have earned; Mini-game Item Shop where avatars can by outfits based on Kingdom Hearts characters which they have earned by winning certain mini-games; Point Item Shop where players can buy different clothes for their avatars as well as customizing the avatar's facial features and hairstyles; Mini-game Shop where players can buy new mini-games; Information Center which provides help and support for players; V Net Room which is the player's own personal area where he can customize
https://en.wikipedia.org/wiki/Gilles%20Holst
Gilles Holst (20 March 1886 – 11 October 1968) was a Dutch physicist, known worldwide for his invention in 1932 of the low-pressure sodium lamp. Early life His father was a manager of a shipyard. In 1904 he went to ETH Zurich to study mechanical engineering, changing after a year to mathematics and physics. Career He worked with Balthasar van der Pol, known for the Van der Pol oscillator, and Frans Michel Penning, known for Penning ionization and the Penning mixture. In 1908 he became a geprüfter Fachlehrer, or qualified teacher. And most important, he became the science director of the Philips Physics Laboratory in Eindhoven. In 1909 he became an assistant to Heike Kamerlingh Onnes at Leiden University. At Leiden, it is believed that he was the first to witness the phenomenon of superconductivity. In 1926 he became a member of the Royal Netherlands Academy of Arts and Sciences. The Gilles Holst Award was first awarded in 1939. Personal life He died in the Netherlands at the age of 82.
https://en.wikipedia.org/wiki/Identity%20change
Identity change describes the intentional changes to an identity document or digital identity. The topic is of particular interest in "faceless" financial transactions and computer security. There are several different parties who may initiate the change: A first party. the original bearer of an identity may initiate the change A second party, who wishes to use the identity, may initiate the change A third party may initiate an identity change In some instances, multiple parties cooperate to change an identity. Identity change can be categorized in several ways: Identity takeover (identity theft / identity fraud) Identity delegation Identity exchange Identity deletion Identity restoration See also Identity management Biometrics Witness protection
https://en.wikipedia.org/wiki/Thermodiscus
Thermodiscus is a genus of archaea in the family Desulfurococcaceae. The only species is Thermodiscus maritimus.
https://en.wikipedia.org/wiki/Topographic%20Abney%20level
An Abney level and clinometer is an instrument used in surveying which consists of a fixed sighting tube, a movable spirit level that is connected to a pointing arm, and a protractor scale. An internal mirror allows the user to see the bubble in the level while sighting a distant target. It can be used as a hand-held instrument or mounted on a Jacob's staff for more precise measurement, and it is small enough to carry in a coat pocket. The Abney level is an easy to use, relatively inexpensive, and, when used correctly, an accurate surveying tool. Abney levels typically include scales graduated in measure degrees of arc, percent grade, and in topographic Abney levels, grade in feet per surveyor's chain, and chainage correction. The latter is the cosine of the angle, used to convert distances measured along the slope to horizontal distances. By using trigonometry the user of an Abney level can determine height, volume, and grade. Abney levels are made with square tubular bodies so that they may also be used to directly measure the slopes of plane surfaces by simply placing the body of the level on the surface, adjusting the level, and then reading the angle off of the scale. Origins The Abney level was invented by Sir William de Wiveleslie Abney (24 July 1843 – 3 December 1920) who was an English astronomer and chemist best known for his pioneering of color photography and color vision. Abney invented this instrument under the employment of the School of Military Engineering in Chatham, England, prior to late 1870. It is described by W. & L. E. Gurley as an English modification of the Locke hand level. Elliott Brothers of London registered an "improved clinometer and spirit level combined" in December 1870 based on "the old form as originally designed by Lieutenant Abney." By 1871, a committee of the Royal Geographical Society recommended a long list of instruments that explorers should carry. Along with necessary tools such as a watch, compass, sextant a
https://en.wikipedia.org/wiki/AM%20expanded%20band
The extended mediumwave broadcast band, commonly known as the AM expanded band, refers to the broadcast station frequency assignments immediately above the earlier upper limits of 1600 kHz in International Telecommunication Union (ITU) Region 2 (the Americas), and 1602 kHz in ITU Regions 1 (Europe, northern Asia and Africa) and 3 (southern Asia and Oceania). In Region 2, this consists of ten additional frequencies, spaced 10 kHz apart, and running from 1610 kHz to 1700 kHz. In Regions 1 and 3, where frequency assignments are spaced nine kHz apart, the result is eleven additional frequencies, from 1611 kHz to 1701 kHz. ITU Region 1 Europe The extended band is not officially allocated in Europe, and the trend of national broadcasters in the region has been to reduce the number of their AM band stations in favor of FM and digital transmissions. However, new Low-Power AM (LPAM) stations have recently come on the air from countries like Finland, Sweden, Norway, Denmark, the Netherlands and Italy. These frequencies are also used by a number of "hobby" pirate radio stations, particularly in the Netherlands, Greece, and Serbia. Vatican Radio for many years transmitted on 1611 kHz, before ceasing broadcasts on this frequency in 2012. Since 2014 a licensed Norwegian project has been broadcasting both Radio Northern Star and The Sea on 1611 kHz. ITU Region 2 In 1979, a World Administrative Radio Conference (WARC-79) adopted "Radio Regulation No. 480", which stated that "In Region 2, the use of the band 1605-1705 kHz by stations of the broadcasting service shall be subject to a plan to be established by a regional administrative radio conference..." As a consequence, on June 8, 1988 an ITU-sponsored conference held at Rio de Janeiro, Brazil adopted provisions, effective July 1, 1990, to extend the upper end of the Region 2 AM broadcast band, by adding ten frequencies which spanned from 1610 kHz to 1700 kHz. The agreement provided for a standard transmitter power of 1 kilowa
https://en.wikipedia.org/wiki/Minimum%20relevant%20variables%20in%20linear%20system
MINimum Relevant Variables in Linear System (Min-RVLS) is a problem in mathematical optimization. Given a linear program, it is required to find a feasible solution in which the number of non-zero variables is as small as possible. The problem is known to be NP-hard and even hard to approximate. Definition A Min-RVLS problem is defined by: A binary relation R, which is one of {=, ≥, >, ≠}; An m-by-n matrix A (where m is the number of constraints and n the number of variables); An m-by-1 vector b. The linear system is given by: A x R b. It is assumed to be feasible (i.e., satisfied by at least one x). Depending on R, there are four different variants of this system: A x = b, A x ≥ b, A x > b, A x ≠ b. The goal is to find an n-by-1 vector x that satisfies the system A x R b, and subject to that, contains as few as possible nonzero elements. Special case The problem Min-RVLS[=] was presented by Garey and Johnson, who called it "minimum weight solution to linear equations". They proved it was NP-hard, but did not consider approximations. Applications The Min-RVLS problem is important in machine learning and linear discriminant analysis. Given a set of positive and negative examples, it is required to minimize the number of features that are required to correctly classify them. The problem is known as the minimum feature set problem. An algorithm that approximates Min-RVLS within a factor of could substantially reduce the number of training samples required to attain a given accuracy level. The shortest codeword problem in coding theory is the same problem as Min-RVLS[=] when the coefficients are in GF(2). Related problems In MINimum Unsatisfied Linear Relations (Min-ULR), we are given a binary relation R and a linear system A x R b, which is now assumed to be infeasible. The goal is to find a vector x that violates as few relations as possible, while satisfying all the others. Min-ULR[≠] is trivially solvable, since any system with real variables and
https://en.wikipedia.org/wiki/Accounting%20machine
An accounting machine, or bookkeeping machine or recording-adder, was generally a calculator and printer combination tailored for a specific commercial activity such as billing, payroll, or ledger. Accounting machines were widespread from the early 1900s to 1980s, but were rendered obsolete by the availability of low-cost computers such as the IBM PC. This type of machine is generally distinct from unit record equipment (some unit record machines were also called accounting machines). List of Vendors/Accounting Machines Burroughs Sensimatic Burroughs Sensitronic Burroughs B80 Burroughs E103 Burroughs Computer F2000 Burroughs L500 Burroughs E1400 Electronic Computing/Accounting Machine with Magnetic Striped Ledger Dalton Adding Machine Company Electronics Corporation of America: Magnefile-B Electronics Corporation of America: Magnefile-D Elliott-Fisher Federal Adding Machines IBM 632 IBM 858 Cardatype Accounting Machine IBM 6400 Series Laboratory for Electronics: The Inventory Machine II (TIM-II) Monroe Calculator Company: Model 200 Monroe Calculator Company: Synchro-Monroe President Monroe Calculator Company: Monrobot IX NCR Post-Tronic Bookkeeping Machine - Class 29 NCR Compu-Tronic Accounting Machine NCR Accounting Machine - Class 33 NCR Window Posting Machine - Class 42 Olivetti: General Bookkeeping Machine (GBM) J. B. Rea Company: READIX, c. 1955 Sundstrand Adding Machines Underwood ELECOM 50 "The First Electronic Accounting Machine" Underwood ELECOM 125, 125 FP (File Processor), 1956 See also Unit record equipment
https://en.wikipedia.org/wiki/Cure%20Rare%20Disease
Cure Rare Disease is a non-profit biotechnology company based in Boston, Massachusetts that is working to create novel therapeutics using gene therapy, gene editing (CRISPR technology) and antisense oligonucleotides to treat people impacted by rare and ultra-rare genetic neuromuscular conditions. History Richard Horgan founded Terry's Foundation for Muscular Dystrophy in 2017, which became Cure Rare Disease in 2018, in order to develop a cure for Duchenne muscular dystrophy for his brother who has been battling the disease since childhood. Leveraging his network from Harvard Business School, Horgan formed a collaboration consisting of leading researchers and clinicians around the country to develop this cure for his brother, and eventually founded Cure Rare Disease. Horgan connected first with a scientist at Boston Children's Hospital, Dr. Timothy Yu, who had just successfully created a custom drug for a girl with the neurodegenerative condition, Batten disease using antisense oligonucleotide (ASO) technology. Horgan's brother's mutation is not amenable to ASO technology, so Horgan adopted the process and instead used CRISPR as the technology to attempt to cure his brother. This collaboration has expanded over the past three years and has led to the addition of notable researchers and institutions collaborating with Cure Rare Disease on their mission to treat rare disease. Research There are currently three drugs approved by the FDA for Duchenne muscular dystrophy to treat the patients with mutations on the dystrophin gene encompassing exon 51, 53, and 45. However, people with DMD have mutations impacting different exons of the gene, so these do not work to treat all patients. Cure Rare Disease is developing novel therapeutics using gene replacement, gene editing (CRISPR gene-editing) and antisense oligonucleotide technologies. To systemically deliver  a subset of therapeutics, including CRISPR, the therapeutic is inserted into the adeno-associated virus (AAV
https://en.wikipedia.org/wiki/Toda%20field%20theory
In mathematics and physics, specifically the study of field theory and partial differential equations, a Toda field theory, named after Morikazu Toda, is specified by a choice of Lie algebra and a specific Lagrangian. Formulation Fixing the Lie algebra to have rank , that is, the Cartan subalgebra of the algebra has dimension , the Lagrangian can be written The background spacetime is 2-dimensional Minkowski space, with space-like coordinate and timelike coordinate . Greek indices indicate spacetime coordinates. For some choice of root basis, is the th simple root. This provides a basis for the Cartan subalgebra, allowing it to be identified with . Then the field content is a collection of scalar fields , which are scalar in the sense that they transform trivially under Lorentz transformations of the underlying spacetime. The inner product is the restriction of the Killing form to the Cartan subalgebra. The are integer constants, known as Kac labels or Dynkin labels. The physical constants are the mass and the coupling constant . Classification of Toda field theories Toda field theories are classified according to their associated Lie algebra. Toda field theories usually refer to theories with a finite Lie algebra. If the Lie algebra is an affine Lie algebra, it is called an affine Toda field theory (after the component of φ which decouples is removed). If it is hyperbolic, it is called a hyperbolic Toda field theory. Toda field theories are integrable models and their solutions describe solitons. Examples Liouville field theory is associated to the A1 Cartan matrix, which corresponds to the Lie algebra in the classification of Lie algebras by Cartan matrices. The algebra has only a single simple root. The sinh-Gordon model is the affine Toda field theory with the generalized Cartan matrix and a positive value for β after we project out a component of φ which decouples. The sine-Gordon model is the model with the same Cartan matrix but an i
https://en.wikipedia.org/wiki/Acta%20Physica%20Sinica
Acta Physica Sinica (abbreviation: Acta. Phys. Sin., or also APS) is a peer-reviewed scientific journal in the fields of physics published by the Chinese Physical Society. Established in 1933 as Chinese Journal of Physics, the journal was published in English, French and German at first. In 1953, the name of the journal was changed to Acta Physica Sinica and the journal became published in Chinese. Translations From 1974 to 1977, articles on Acta Physica Sinica were translated in English and published as a bimonthly Chinese Journal of Physics (Acta Physica Sinica). From 1981 to 1992, American Institute of Physics published a quarterly named Chinese Physics translating selected works selected from 12 physical journals in Chinese including Acta Physica Sinica. From 1987 to 1989, selected works from Acta Physica Sinica were translated and published as a quarterly under the name of Acta Physica Sinica - Journal of Chinese Physics. Related publications Acta Physica Sinica is a part of a small group of journals from the Chinese Physical Society, the others including: Communications in Theoretical Physics (Chinese Physics A), Chinese Physics B (established in 1991 as Acta Physica Sinica (Overseas Edition), renamed as Chinese Physics in 2000, renamed again in 2008 as Chinese Physics B. Though once named as the overseas edition of Acta Physica Sinica, it was never an English translation of works on Acta Physica Sinica.), and Chinese Physics C.
https://en.wikipedia.org/wiki/Canadian%20Society%20for%20Mass%20Spectrometry
The Canadian Society for Mass Spectrometry is an organization that promotes mass spectrometry in Canada. The goal of the society is to stimulate interest and collaborations in the Canadian mass spectrometry community. The society organizes conferences, awards prices and runs an online job board. The society is an affiliate society of the International Mass Spectrometry Foundation. Its current president is Derek Wilson. The society awards the annual Fred P. Lossing Award.
https://en.wikipedia.org/wiki/Carotid%E2%80%93Kundalini%20function
The Carotid–Kundalini function is closely associated with Carotid-Kundalini fractals coined by popular science columnist Clifford A. Pickover and it is defined as follows: See also arccos
https://en.wikipedia.org/wiki/CAS%20Registry%20Number
A CAS Registry Number (also referred to as CAS RN or informally CAS Number) is a unique identification number assigned by the Chemical Abstracts Service (CAS) in the US to every chemical substance described in the open scientific literature. It includes all substances described since 1957, plus some substances from as far back as the early 1800s. It is a chemical database that includes organic and inorganic compounds, minerals, isotopes, alloys, mixtures, and nonstructurable materials (UVCBs, substances of unknown or variable composition, complex reaction products, or biological origin). CAS RNs are generally serial numbers (with a check digit), so they do not contain any information about the structures themselves the way SMILES and InChI strings do. The registry maintained by CAS is an authoritative collection of disclosed chemical substance information. It identifies more than 204 million unique organic and inorganic substances and 70 million protein and DNA sequences, plus additional information about each substance. It is updated with around 15,000 additional new substances daily. A collection of almost 500 thousand CAS registry numbers are made available under a CC BY-NC license at ACS Commons Chemistry. Use Historically, chemicals have been identified by a wide variety of synonyms. Frequently these are arcane and constructed according to regional naming conventions relating to chemical formulae, structures or origins. Well-known chemicals may additionally be known via multiple generic, historical, commercial, and/or (black)-market names. CAS Registry Numbers (CAS RN) are simple and regular, convenient for database searches. They offer a reliable, common and international link to every specific substance across the various nomenclatures and disciplines used by branches of science, industry, and regulatory bodies. Almost all molecule databases today allow searching by CAS Registry Number. Format A CAS Registry Number has no inherent meaning, but is assi
https://en.wikipedia.org/wiki/XhoI
In molecular biology, XhoI is a type II restriction enzyme EC that recognise the double-stranded DNA sequence CTCGAG and cleaves after C-1. Type II restriction endonucleases (EC) are components of prokaryotic DNA restriction-modification mechanisms that protect the organism against invading foreign DNA. These site-specific deoxyribonucleases catalyse the endonucleolytic cleavage of DNA to give specific double-stranded fragments with terminal 5'-phosphates.
https://en.wikipedia.org/wiki/Steven%20Salzberg
Steven Lloyd Salzberg (born 1960) is an American computational biologist and computer scientist who is a Bloomberg Distinguished Professor of Biomedical Engineering, Computer Science, and Biostatistics at Johns Hopkins University, where he is also Director of the Center for Computational Biology. Early life and education Salzberg was born in 1960 as one of four children to Herman Salzberg, a Distinguished Professor Emeritus of Psychology, and Adele Salzberg, a retired school teacher. Salzberg did his undergraduate studies at Yale University where he received his Bachelor of Arts degree in English in 1980. In 1981 he returned to Yale, and he received his Master of Science and Master of Philosophy degrees in Computer Science in 1982 and 1984, respectively. After several years in a startup company, he enrolled at Harvard University, where he earned a Ph.D. in Computer Science in 1989. Career After obtaining his undergraduate degree, he worked for a local power company in South Carolina, where he gained programming experience using IBM mainframe. He also learned COBOL and IBM Assembler. He then joined a Boston-based AI startup upon completion of his masters degree in Computer Science. After earning his Ph.D., Salzberg joined Johns Hopkins University as an assistant professor in the Department of Computer Science, and was promoted to associate professor in 1997. From 1998 to 2005, he was the head of the Bioinformatics department at The Institute for Genomic Research, one of the world's largest genome sequencing centers. Salzberg then joined the Department of Computer Science at the University of Maryland, College Park, where he was the Horvitz Professor of Computer Science as well as the Director of the Center for Bioinformatics and Computational Biology. In 2011, Salzberg returned to Johns Hopkins University as a professor in the Department of Medicine. From 2014, he was a professor in the Department of Biomedical Engineering in the School of Medicine; the Departme
https://en.wikipedia.org/wiki/Arid%20Lands%20Ecology%20Reserve
The Arid Land Ecology Reserve (ALE) is the largest tract of shrub-steppe ecosystem remaining in the U.S. state of Washington. It is managed for the U.S. Department of Energy by the Pacific Northwest National Laboratory (which is operated for the U.S. Department of Energy by Battelle Memorial Institute). The 320 km² area is a portion of the 1500 km² National Environmental Research Park located on the Hanford Site on the northwest boundary of Richland, Washington. On June 27, 2000, a range fire destroyed most of the native sagebrush and bunchgrass as well as damaged the microbiotic crust. Though the US Fish and Wildlife Service has attempted to re-introduce native flora, the Arid Lands Ecology Reserve is currently dominated by non-native species such as cheatgrass, knapweeds, and Russian thistle (tumbleweed) which flourished after the 2000 fire. Other species such as spiny hop sage and Wyoming big sagebrush were decimated by the fire and in its aftermath. Vegetation Shrub-steppe This vegetation type describes plant communities found in and around arid mountains, ridges and slopes. In the ALE this includes shrubs (sagebrush and rabbit brush), perennial bunchgrasses (Sandberg's blue grass and bluebunch wheat grass) as well as both annual and perennial forbs (balsamroot, phlox and fleabane). Riparian This vegetation type refers to plant communities located around springs and streamflow. In the ALE these areas are dominated by willow, black cottonwood, chokecherry, and mock Orange. Point of Interest North facing slope of Rattlesnake Mountain is the highest “treeless” mountain in the United States. Rare plant species such as Mountain Milk Vetch and Piper's daisy can be found in this area. History and Significant Dates From the early 1800s to around the 1940s this area was used as animal pasture, human homesteading, oil drilling and development of infrastructure such as roads. In 1943 the United States Department of Energy (DOE) gained ownership of the land. T
https://en.wikipedia.org/wiki/IMSI-catcher
An international mobile subscriber identity-catcher, or IMSI-catcher, is a telephone eavesdropping device used for intercepting mobile phone traffic and tracking location data of mobile phone users. Essentially a "fake" mobile tower acting between the target mobile phone and the service provider's real towers, it is considered a man-in-the-middle (MITM) attack. The 3G wireless standard offers some risk mitigation due to mutual authentication required from both the handset and the network. However, sophisticated attacks may be able to downgrade 3G and LTE to non-LTE network services which do not require mutual authentication. IMSI-catchers are used in a number of countries by law enforcement and intelligence agencies, but their use has raised significant civil liberty and privacy concerns and is strictly regulated in some countries such as under the German Strafprozessordnung (StPO / Code of Criminal Procedure). Some countries do not have encrypted phone data traffic (or very weak encryption), thus rendering an IMSI-catcher unnecessary. Overview A virtual base transceiver station (VBTS) is a device for identifying the temporary mobile subscriber identity (TMSI), international mobile subscriber identity (IMSI) of a nearby GSM mobile phone and intercepting its calls, some are even advanced enough to detect the international mobile equipment identity (IMEI). It was patented and first commercialized by Rohde & Schwarz in 2003. The device can be viewed as simply a modified cell tower with a malicious operator, and on 4 January 2012, the Court of Appeal of England and Wales held that the patent is invalid for obviousness. IMSI-catchers are often deployed by court order without a search warrant, the lower judicial standard of a pen register and trap-and-trace order being preferred by law enforcement. They can also be used in search and rescue operation for missing persons. Police departments have been reluctant to reveal use of these programs and contracts with vendors
https://en.wikipedia.org/wiki/Sensometrics
Sensometrics is a scientific discipline that involves the application of mathematical, statistical, and computational methods to the field of Sensory and Consumer Science. It is an interdisciplinary field, closely related to disciplines such as sensory evaluation, statistics, psychometrics, and chemometrics, among others. Sensometrics deals with all aspects of data generation and analysis, from design of experiments and methods used to investigate perceptions and preferences, to specific tools to analyze and model the data resulting from these methods. Sensometrics plays a key role in Sensory and Consumer Science, providing important tools with applications in product development, quality assurance, market research, and consumer behavior. History Sensometrics emerged in response to the increasing attention on measurement in the more general field of Sensory Science—which is typically described as concerned with “evoking, measuring, analyzing, and interpreting” humans’ sensory perceptions of consumer products. The exchange of knowledge between statisticians and sensory scientists contributed to the growth of both fields, by encouraging the development of new data collection methods as well as extending and inspiring the development of statistical techniques In 1992 a group of statisticians and sensory scientists started organizing Sensometrics meetings and the Sensometrics Society was created. Whereas the broader field of Sensory Science focuses on all of the above aspects, Sensometrics is particularly concerned with the development of methodologies for measuring and analyzing sensory experiences quantitatively. See also Sensory evaluation European Sensory Network Food Quality and Preference Journal of Sensory Studies Food science
https://en.wikipedia.org/wiki/Posterior%20talocalcaneal%20ligament
The posterior talocalcaneal ligament (posterior calcaneo-astragaloid ligament) connects the lateral tubercle of the talus with the upper and medial part of the calcaneus; it is a short band, and its fibers radiate from their narrow attachment to the talus.
https://en.wikipedia.org/wiki/Planetarium%20hypothesis
The planetarium hypothesis, conceived in 2001 by Stephen Baxter, attempts to provide a solution to the Fermi paradox by holding that our astronomical observations represent an illusion, created by a Type III civilization capable of manipulating matter and energy on galactic scales. He postulates that we do not see evidence of extraterrestrial life because the universe has been engineered so that it appears empty of other life. Background There is no reliable or reproducible evidence that aliens have visited Earth. No transmissions or evidence of intelligent extraterrestrial life have been detected or observed anywhere other than Earth in the Universe. This runs counter to the knowledge that the Universe is filled with a very large number of planets, some of which likely hold the conditions hospitable for life. Life on Earth has shown the tendency to typically expand until it fills all available niches. These contradictory facts form the basis for the Fermi paradox, of which the Planetarium hypothesis is one proposed solution. Criticism The hypothesis has been considered by some authors as speculative and even next to useless in any practical scientific sense and more related to the theological mode of thinking along with the zoo hypothesis.
https://en.wikipedia.org/wiki/Experimental%20Cell%20Research
Experimental Cell Research is a peer-reviewed scientific journal covering cell biology. It was established in 1950 by Academic Press and is currently published by Elsevier. External links Academic journals established in 1950 Molecular and cellular biology journals Elsevier academic journals English-language journals
https://en.wikipedia.org/wiki/Common%20modeling%20infrastructure
Common modeling infrastructure refers to software libraries that can be shared across multiple institutions in order to increase software reuse and interoperability in complex modeling systems. Early initiatives were in the climate and weather domain, where software components representing distinct physical domains (for example, ocean or atmosphere) tended to be developed by domain specialists, often at different organizations. In order to create complete applications, these needed to be combined, using for instance a general circulation model, that transfers data between different components. An additional challenge is that these models generally require supercomputers to run, to account for the collected data and for data analyses. Thus, it was important to provide an efficient massively parallel computer system, and the processing hardware and software, to account for all the different workloads and communication channels. General Common modeling infrastructure projects include the Network Common Data Form (NetCDF) library, the Spherical Coordinate Remapping and Interpolation Package (SCRIP), the Flexible Modeling System (FMS), the OASIS coupler developed at CERFACS, and the multi-agency Earth System Modeling Framework (ESMF). The Earth System Modeling Framework (ESMF) is considered a technical layer, integrated into a common modeling infrastructure. Other aspects of interoperability and shared infrastructure include: common experimental protocols, common analytic methods, common documentation standards for data and data provenance, shared workflow, and shared model components. History In the late 1990s and early 2000s, a series of journal papers and government reports described common modeling infrastructure as necessary to the competitiveness and evolution of the U.S. Earth science modeling community. These reports resulted in a number of new community projects. The Earth System Modeling Framework (ESMF) and the Earth System Modeling (ESM) are two of t
https://en.wikipedia.org/wiki/Hopfian%20object
In the branch of mathematics called category theory, a hopfian object is an object A such that any epimorphism of A onto A is necessarily an automorphism. The dual notion is that of a cohopfian object, which is an object B such that every monomorphism from B into B is necessarily an automorphism. The two conditions have been studied in the categories of groups, rings, modules, and topological spaces. The terms "hopfian" and "cohopfian" have arisen since the 1960s, and are said to be in honor of Heinz Hopf and his use of the concept of the hopfian group in his work on fundamental groups of surfaces. Properties Both conditions may be viewed as types of finiteness conditions in their category. For example, assuming Zermelo–Fraenkel set theory with the axiom of choice and working in the category of sets, the hopfian and cohopfian objects are precisely the finite sets. From this it is easy to see that all finite groups, finite modules and finite rings are hopfian and cohopfian in their categories. Hopfian objects and cohopfian objects have an elementary interaction with projective objects and injective objects. The two results are: An injective hopfian object is cohopfian. A projective cohopfian object is hopfian. The proof for the first statement is short: Let A be an injective hopfian object, and let f be an injective morphism from A to A. By injectivity, f factors through the identity map IA on A, yielding a morphism g such that gf=IA. As a result, g is a surjective morphism and hence an automorphism, and then f is necessarily the inverse automorphism to g. This proof can be dualized to prove the second statement. Hopfian and cohopfian groups Hopfian and cohopfian modules Here are several basic results in the category of modules. It is especially important to remember that RR being hopfian or cohopfian as a module is different from R being hopfian or cohopfian as a ring. A Noetherian module is hopfian, and an Artinian module is cohopfian. The module R
https://en.wikipedia.org/wiki/Substrate-level%20phosphorylation
Substrate-level phosphorylation is a metabolism reaction that results in the production of ATP or GTP supported by the energy released from another high-energy bond that leads to phosphorylation of ADP or GDP to ATP or GTP (note that the reaction catalyzed by creatine kinase is not considered as "substrate-level phosphorylation"). This process uses some of the released chemical energy, the Gibbs free energy, to transfer a phosphoryl (PO3) group to ADP or GDP. Occurs in glycolysis and in the citric acid cycle. Unlike oxidative phosphorylation, oxidation and phosphorylation are not coupled in the process of substrate-level phosphorylation, and reactive intermediates are most often gained in the course of oxidation processes in catabolism. Most ATP is generated by oxidative phosphorylation in aerobic or anaerobic respiration while substrate-level phosphorylation provides a quicker, less efficient source of ATP, independent of external electron acceptors. This is the case in human erythrocytes, which have no mitochondria, and in oxygen-depleted muscle. Overview Adenosine triphosphate (ATP) is a major "energy currency" of the cell. The high energy bonds between the phosphate groups can be broken to power a variety of reactions used in all aspects of cell function. Substrate-level phosphorylation occurs in the cytoplasm of cells during glycolysis and in mitochondria either during the Krebs cycle or by MTHFD1L (EC 6.3.4.3), an enzyme interconverting ADP + phosphate + 10-formyltetrahydrofolate to ATP + formate + tetrahydrofolate (reversibly), under both aerobic and anaerobic conditions. In the pay-off phase of glycolysis, a net of 2 ATP are produced by substrate-level phosphorylation. Glycolysis The first substrate-level phosphorylation occurs after the conversion of 3-phosphoglyceraldehyde and Pi and NAD+ to 1,3-bisphosphoglycerate via glyceraldehyde 3-phosphate dehydrogenase. 1,3-bisphosphoglycerate is then dephosphorylated via phosphoglycerate kinase, producing 3-ph
https://en.wikipedia.org/wiki/Rayleigh%27s%20quotient%20in%20vibrations%20analysis
The Rayleigh's quotient represents a quick method to estimate the natural frequency of a multi-degree-of-freedom vibration system, in which the mass and the stiffness matrices are known. The eigenvalue problem for a general system of the form in absence of damping and external forces reduces to The previous equation can be written also as the following: where , in which represents the natural frequency, M and K are the real positive symmetric mass and stiffness matrices respectively. For an n-degree-of-freedom system the equation has n solutions , that satisfy the equation By multiplying both sides of the equation by and dividing by the scalar , it is possible to express the eigenvalue problem as follow: for . In the previous equation it is also possible to observe that the numerator is proportional to the potential energy while the denominator depicts a measure of the kinetic energy. Moreover, the equation allow us to calculate the natural frequency only if the eigenvector (as well as any other displacement vector) is known. For academic interests, if the modal vectors are not known, we can repeat the foregoing process but with and taking the place of and , respectively. By doing so we obtain the scalar , also known as Rayleigh's quotient: Therefore, the Rayleigh's quotient is a scalar whose value depends on the vector and it can be calculated with good approximation for any arbitrary vector as long as it lays reasonably far from the modal vectors , i = 1,2,3,...,n. Since, it is possible to state that the vector differs from the modal vector by a small quantity of first order, the correct result of the Rayleigh's quotient will differ not sensitively from the estimated one and that's what makes this method very useful. A good way to estimate the lowest modal vector , that generally works well for most structures (even though is not guaranteed), is to assume equal to the static displacement from an applied force that has the same relative distr
https://en.wikipedia.org/wiki/Pasta%20processing
Pasta processing is the process in which wheat semolina or flour is mixed with water and the dough is extruded to a specific shape, dried and packaged. Durum wheat semolina or flour, common farina or flour, or combination of both is mixed with water and eggs (for egg noodles) and other optional ingredients (like spinach, tomato, herbs, etc.). Usually 25–30 kg of water is added per 100 kg of semolina. The amounts are measured by computerized dispensers. The mixture is then kneaded by auger extruder equipped with mixing paddles and kneading blades to obtain a homogeneous mass, and after that is extruded through various shaped dies. Drying process begins immediately after the products are shaped to prevent deformation and sticking. The pastas are dried completely in drying chambers and stabilized, then ready for packaging. In modern factories, dry pasta is processed using automatic continuous lines. Mixing In this level wheat semolina and water are mixed by the ratio of 3 to 1. Water should be pure, with no off-flavor and suitable for drinking. Its temperature is about 35-45 °C to help speed up absorption. For egg noodles, eggs are added in the form of fresh eggs, frozen eggs, dry eggs, egg yolks or dry egg solids. If eggs are added to the mixture, the amount of water is modified. Adding egg improves the nutritional quality and richness of the pasta. Disodium phosphate is also added to reduce the cooking time. Mixing the semolina and water takes place in two stages. First, the ingredients are measured and added to a pre-mixer, and then they are transferred to a mixing chamber which finalizes the mixing process and produces a homogeneous mass. Measuring the raw material The exact amount of raw materials is very important. Semolina dosing is done by two methods: volumetric feed (measurement by volume), and gravimetric feed (measurement by weight). In volumetric feed, a specific volume of semolina is measured by variable speed screws or rotary air-lock valves. This m
https://en.wikipedia.org/wiki/2Blades
2Blades is an agricultural phytopathology non-profit which performs research to improve durable genetic resistance in crops, and funds other researchers to do the same. 2Blades was co-founded by Dr. Roger Freedman and Dr. Diana Horvath in 2004. Funding source 2Blades is partly funded by the Gatsby Charitable Foundation does its research at The Sainsbury Laboratory, among other locations . One co-founder, Chairman Roger Freedman also works for Gatsby, which was founded by Lord David Sainsbury. Freedman had pitched an idea to Sainsbury's venture capital company to begin investing in plant genetic engineering technologies, and although the board did so, they found someone else to lead it. Freedman had wanted to run it, but was told that was not for him by Sainsbury. Indeed, soon thereafter Sainsbury set up another early investment company specifically for Freedman and a colleague, and a separate non-profit for Freedman to grant money, both for plant science. The non-profit was 2Blades. Research activities 2Blades routinely works in partnership with other crop disease organizations like CIMMYT and BGRI. The foundation also conducts research in partnerships with the industry, including with Bayer CropScience and Monsanto. The organisation's End the Blight campaign has been joined by CIP (the International Potato Center) and Chairman of Joseph P. Kennedy Enterprises Christopher Kennedy. This campaign is advancing research and delivering cultivars specifically for Phytophthora infestans in Africa. Mr Kennedy is chairman of 2Blades African Potato Initiative which is funding the delivery of a Victoria-based cultivar to East African markets. Crops and pathogens of research interest to the foundation include P. infestans on potato, rye, Phakopsora pachyrhizi on soybean, Puccinia graminis f. sp. tritici on wheat, and Fusarium oxysporum f.sp. cubense on Musa spp.
https://en.wikipedia.org/wiki/Software%20Product%20Line%20Conference
The Software Product Line Conference SPLC is an international conference which is held annually. The conference was started based on previous conferences, a Software Product Line Conference organised by Software Engineering Institute in the US and Product Family Engineering Workshops. Today the SPLC is established as knowledge exchange for existing and new software product lines within different industries such as automotive, avionics, medicine which share the same targets such as reducing development cost, increasing quality and at the same time to maintain flexibility. Previous conferences were held in Denver, Bilbao, San Diego, Siena, Boston, Rennes, Baltimore, Kyoto, Limerick, San Francisco, Jeju Island, Munich, Salvador, Tokyo, Florence, Nashville, Beijing, and Sevilla. The following companies have been elected in the Hall of Fame for their achievement in software product line engineering: Boeing, Robert Bosch GmbH, CelsiusTech, Cummins, Danfoss, Ericsson, FISCAN, General Motors, Hewlett Packard, HomeAway, Lockheed Martin, LSI Logic, Lucent, Market Maker, Nokia, Philips, Salion, Siemens, Toshiba, U.S. Army Live Training Transformation, and U.S. Naval Research Laboratory. Since 2016 an award for the most influential paper has been established. 2016 Don Batory was awarded, 2017 the award was given to Krzysztof Czarnecki, Ulrich Eisenecker of the University of Leipzig.
https://en.wikipedia.org/wiki/Gell-Mann%20matrices
The Gell-Mann matrices, developed by Murray Gell-Mann, are a set of eight linearly independent 3×3 traceless Hermitian matrices used in the study of the strong interaction in particle physics. They span the Lie algebra of the SU(3) group in the defining representation. Matrices {| border="0" cellpadding="8" cellspacing="0" | | | |- | | | |- | | | |} Properties These matrices are traceless, Hermitian, and obey the extra trace orthonormality relation (so they can generate unitary matrix group elements of SU(3) through exponentiation). These properties were chosen by Gell-Mann because they then naturally generalize the Pauli matrices for SU(2) to SU(3), which formed the basis for Gell-Mann's quark model. Gell-Mann's generalization further extends to general SU(n). For their connection to the standard basis of Lie algebras, see the Weyl–Cartan basis. Trace orthonormality In mathematics, orthonormality typically implies a norm which has a value of unity (1). Gell-Mann matrices, however, are normalized to a value of 2. Thus, the trace of the pairwise product results in the ortho-normalization condition where is the Kronecker delta. This is so the embedded Pauli matrices corresponding to the three embedded subalgebras of SU(2) are conventionally normalized. In this three-dimensional matrix representation, the Cartan subalgebra is the set of linear combinations (with real coefficients) of the two matrices and , which commute with each other. There are three significant SU(2) subalgebras: and where the and are linear combinations of and . The SU(2) Casimirs of these subalgebras mutually commute. However, any unitary similarity transformation of these subalgebras will yield SU(2) subalgebras. There is an uncountable number of such transformations. Commutation relations The 8 generators of SU(3) satisfy the commutation and anti-commutation relations with the structure constants The structure constants are completely antisymmetric in
https://en.wikipedia.org/wiki/Stanford%20Web%20Credibility%20Project
The Stanford Web Credibility Project, which involves assessments of website credibility conducted by the Stanford University Persuasive Technology Lab, is an investigative examination of what leads people to believe in the veracity of content found on the Web. The goal of the project is to enhance website design and to promote further research on the credibility of Web resources. Origins The Web has become an important channel for exchanging information and services, resulting in a greater need for methods to ascertain the credibility of websites. In response, since 1998, the Stanford Persuasive Technology Lab (SPTL) has investigated what causes people to believe, or not, what they find online. SPTL provides insight into how computers can be designed to change what people think and do, an area called captology. Directed by experimental psychologist B.J. Fogg, the Stanford team includes social scientists, designers, and technologists who research and design interactive products that motivate and influence their users. Objectives The ongoing research of the Stanford Web Credibility Project includes: Performing quantitative research on Web credibility Collecting all public information on Web credibility Acting as a clearinghouse for this information Facilitating research and discussion about Web credibility Collaborating with academic and industry research groups How Do People Evaluate a Web Site's Credibility? A study by the Stanford Web Credibility Project, How Do People Evaluate a Web Site's Credibility? Results from a Large Study, published in 2002, invited 2,684 "average people" to rate the credibility of websites in ten content areas. The study evaluated the credibility of two live websites randomly assigned from one of ten content categories: e-commerce, entertainment, finance, health, news, nonprofit, opinion or review, search engines, sports, and travel. A total of one hundred sites were assessed. This study was launched jointly with a para
https://en.wikipedia.org/wiki/The%20Petersen%20Graph
The Petersen Graph is a mathematics book about the Petersen graph and its applications in graph theory. It was written by Derek Holton and John Sheehan, and published in 1993 by the Cambridge University Press as volume 7 in their Australian Mathematical Society Lecture Series. Topics The Petersen graph is an undirected graph with ten vertices and fifteen edges, commonly drawn as a pentagram within a pentagon, with corresponding vertices attached to each other. It has many unusual mathematical properties, and has frequently been used as a counterexample to conjectures in graph theory. The book uses these properties as an excuse to cover several advanced topics in graph theory where this graph plays an important role. It is heavily illustrated, and includes both open problems on the topics it discusses and detailed references to the literature on these problems. After an introductory chapter, the second and third chapters concern graph coloring, the history of the four color theorem for planar graphs, its equivalence to 3-edge-coloring of planar cubic graphs, the snarks (cubic graphs that have no such colorings), and the conjecture of W. T. Tutte that every snark has the Petersen graph as a graph minor. Two more chapters concern closely related topics, perfect matchings (the sets of edges that can have a single color in a 3-edge-coloring) and nowhere-zero flows (the dual concept to planar graph coloring). The Petersen graph shows up again in another conjecture of Tutte, that when a bridgeless graph does not have the Petersen graph as a minor, it must have a nowhere-zero 4-flow. Chapter six of the book concerns cages, the smallest regular graphs with no cycles shorter than a given length. The Petersen graph is an example: it is the smallest 3-regular graph with no cycles of length shorter than 5. Chapter seven is on hypohamiltonian graphs, the graphs that do not have a Hamiltonian cycle through all vertices but that do have cycles through every set of all but one v
https://en.wikipedia.org/wiki/Timeline%20of%20extinctions%20in%20the%20Holocene
This article is a list of biological species, subspecies, and evolutionary significant units that are known to have become extinct during the Holocene, the current geologic epoch, ordered by their known or approximate date of disappearance from oldest to most recent. The Holocene is considered to have started with the Holocene glacial retreat around 11650 years Before Present ( BC). It is characterized by a general trend towards global warming, the expansion of anatomically modern humans (Homo sapiens) to all emerged land masses, the appearance of agriculture and animal husbandry, and a reduction in global biodiversity. The latter, dubbed the sixth mass extinction in Earth history, is largely attributed to increased human population and activity, and may have started already during the preceding Pleistocene epoch with the demise of the Pleistocene megafauna. The following list is incomplete by necessity, since the majority of extinctions are thought to be undocumented, and for many others there isn't a definitive, widely accepted last, or most recent record. According to the species-area theory, the present rate of extinction may be up to 140,000 species per year. 10th millennium BC 9th millennium BC 8th millennium BC 7th millennium BC 6th millennium BC 5th millennium BC 4th millennium BC 3rd millennium BC 2nd millennium BC 1st millennium BC 1st millennium CE 1st–5th centuries 6th–10th centuries 2nd millennium CE 11th-12th century 13th-14th century 15th-16th century 17th century 18th century 19th century 1800s-1820s 1830s-1840s 1850s-1860s 1870s 1880s 1890s 20th century 1900s 1910s 1920s 1930s 1940s 1950s 1960s 1970s 1980s 1990s 3rd millennium CE 21st century 2000s 2010s See also List of extinct animals Extinction event Quaternary extinction event Holocene extinction Timeline of the evolutionary history of life Timeline of environmental history Index of environmental articles List of environmental issues
https://en.wikipedia.org/wiki/ADONE
ADONE (big AdA) was a high-energy (beam energy 1.5 GeV, center-of-mass energy 3 GeV) particle collider. It collided electrons with their antiparticles, positrons. It was 105 meters in circumference. It was operated from 1969 to 1993, by the National Institute of Nuclear Physics (INFN) at the Frascati National Laboratory (LNF), in Frascati, Italy. See also ADA collider Laboratori Nazionali di Frascati Istituto Nazionale di Fisica Nucleare
https://en.wikipedia.org/wiki/Xml%3Atm
xml:tm (XML-based Text Memory) is a standard for XML to allow ease of translation of XML documents. xml:tm forms part of the Open Architecture for XML Authoring and Localization reference architecture. External links xml:tm page on the LISA web site Technical communication XML-based standards XML markup languages Internationalization and localization
https://en.wikipedia.org/wiki/Multibeam%20echosounder
A multibeam echosounder (MBES) is a type of sonar that is used to map the seabed. It emits acoustic waves in a fan shape beneath its transceiver. The time it takes for the sound waves to reflect off the seabed and return to the receiver is used to calculate the water depth. Unlike other sonars and echo sounders, MBES uses beamforming to extract directional information from the returning soundwaves, producing a swathe of depth soundings from a single ping. History and progression Multibeam sonar sounding systems, also known as swathe (British English) or swath (American English) , originated for military applications. The concept originated in a radar system that was intended for the Lockheed U-2 high altitude reconnaissance aircraft, but the project was derailed when the aircraft flown by Gary Powers was brought down by a Soviet missile in May 1960. A proposal for using the "Mills Cross" beamforming technique adapted for use with bottom mapping sonar was made to the US Navy. Data from each ping of the sonar would be automatically processed, making corrections for ship motion and transducer depth sound velocity and refraction effects, but at the time there was insufficient digital data storage capacity, so the data would be converted into a depth contour strip map and stored on continuous film. The Sonar Array Sounding System (SASS) was developed in the early 1960s by the US Navy, in conjunction with General Instrument to map large swathes of the ocean floor to assist the underwater navigation of its submarine force. SASS was tested aboard the USS Compass Island (AG-153). The final array system, composed of sixty-one one degree beams with a swathe width of approximately 1.15 times water depth, was then installed on the USNS Bowditch (T-AGS-21), USNS Dutton (T-AGS-22) and USNS Michelson (T-AGS-23). At the same time, a Narrow Beam Echo Sounder (NBES) using 16 narrow beams was also developed by Harris ASW and installed on the Survey Ships Surveyor, Discoverer and Res
https://en.wikipedia.org/wiki/Pair%20by%20association
In relation to psychology, pair by association is the action of associating a stimulus with an arbitrary idea or object, eliciting a response, usually emotional. This is done by repeatedly pairing the stimulus with the arbitrary object. For example, repeatedly pairing images of beautiful women in bathing suits elicits a sexual response in most men. Advertising agencies repeatedly pair products with attractive women in television commercials with the intention of eliciting an emotional or sexually aroused response in the consumer. This causes the consumer to be more likely to buy the product than when presented with a similar product without such an association. Hippocampus The hippocampal area, beyond its importance in episodic memory is in part responsible in the creation and storage of associations in the memory, especially for item associations. Furthermore, as stated by Gilbert & Kesner, the associations that are created are those that might be “critical” in paired-associative learning. Through studies on rats, it has been found that lesions to the hippocampus lead to object-place associative learning impairments. The findings of hippocampal damage that lead to impairments in the association between object-place as Gilbert & Kesner state have been seen in not only rodents but also non-human primates and humans . Previously learned associations made before the damage to the hippocampal area were not affected with impairment. Gilbert & Kesner have suggested in their work that this ability to still recall previously stored associations may be due to modified synapses in an “auto associative network”. Pair by Association Task Paired association learning can be defined as a system of learning in which items (such as words, letters, numbers, symbols etc.) are matched so that presentation of one member of the pair will cue the recall of the other member. It is this learning which constitutes the basics in a paired-associate task. These tasks can be divided into the
https://en.wikipedia.org/wiki/NeuroMat
The Research, Innovation, and Dissemination Center for Neuromathematics (RIDC NeuroMat, or simply NeuroMat) is a Brazilian research center established in 2013 at the University of São Paulo that is dedicated to integrating mathematical modeling and theoretical neuroscience. Among the core missions of NeuroMat are the creation of a new mathematical system to understanding neural data and the development of neuroscientific open-source computational tools, keeping an active role under the context of open knowledge, open science and scientific dissemination. The research center is headed by Antonio Galves, from USP's Institute of Mathematics and Statistics, and is funded by the São Paulo Research Foundation (FAPESP). As of 2019, the co-principal investigators are Oswaldo Baffa Filho (USP), Pablo A. Ferrari (USP/UBA), Fernando da Paixão (UNICAMP), Antonio Carlos Roque (USP), Jorge Stolfi (UNICAMP), and Cláudia D. Vargas (UFRJ). Ernst W. Hamburger (USP) was the former director of scientific dissemination. NeuroMat's International Advisory Board consists of David R. Brillinger (UC Berkeley), Leonardo G. Cohen (NIH), Markus Diesmann (Jülich), Francesco Guerra (La Sapienza), Wojciech Szpankowski (Purdue). Research NeuroMat has been involved in the development of what has been called the Galves-Löcherbach model, a model with intrinsic stochasticity for biological neural nets, in which the probability of a future spike depends on the evolution of the complete system since the last spike.[1] This model of spiking neurons was developed by mathematicians Antonio Galves and Eva Löcherbach. In the first article on the model, in 2013, they called it a model of a "system with interacting stochastic chains with memory of variable length. See also Among the current large-scale international brain initiatives: Allen Institute – from the USA AusBrain – from Australia BRAIN Initiative – from the USA BRAINN - Brazilian Institute of Neuroscience and Neurotechnology – from Brazil Brain
https://en.wikipedia.org/wiki/Beta%20negative%20binomial%20distribution
In probability theory, a beta negative binomial distribution is the probability distribution of a discrete random variable  equal to the number of failures needed to get successes in a sequence of independent Bernoulli trials. The probability of success on each trial stays constant within any given experiment but varies across different experiments following a beta distribution. Thus the distribution is a compound probability distribution. This distribution has also been called both the inverse Markov-Pólya distribution and the generalized Waring distribution or simply abbreviated as the BNB distribution. A shifted form of the distribution has been called the beta-Pascal distribution. If parameters of the beta distribution are and , and if where then the marginal distribution of is a beta negative binomial distribution: In the above, is the negative binomial distribution and is the beta distribution. Definition and derivation Denoting the densities of the negative binomial and beta distributions respectively, we obtain the PMF of the BNB distribution by marginalization: Noting that the integral evaluates to: we can arrive at the following formulas by relatively simple manipulations. If is an integer, then the PMF can be written in terms of the beta function,: . More generally, the PMF can be written or . PMF expressed with Gamma Using the properties of the Beta function, the PMF with integer can be rewritten as: . More generally, the PMF can be written as . PMF expressed with the rising Pochammer symbol The PMF is often also presented in terms of the Pochammer symbol for integer Properties Factorial Moments The -th factorial moment of a beta negative binomial random variable is defined for and in this case is equal to Non-identifiable The beta negative binomial is non-identifiable which can be seen easily by simply swapping and in the above density or characteristic function and noting that it is unchanged. Thus estimation demands th
https://en.wikipedia.org/wiki/Proofs%20of%20convergence%20of%20random%20variables
This article is supplemental for “Convergence of random variables” and provides proofs for selected results. Several results will be established using the portmanteau lemma: A sequence {Xn} converges in distribution to X if and only if any of the following conditions are met: <li> for all bounded, continuous functions ; <li> for all bounded, Lipschitz functions ; <li> for all closed sets ; Convergence almost surely implies convergence in probability Proof: If converges to almost surely, it means that the set of points has measure zero. Now fix and consider a sequence of sets This sequence of sets is decreasing () towards the set The probabilities of this sequence are also decreasing, so ; we shall show now that this number is equal to zero. Now for any point outside of we have , which implies that for all for some . In particular, for such the point will not lie in , and hence won't lie in . Therefore, and so . Finally, by continuity from above, which by definition means that converges in probability to . Convergence in probability does not imply almost sure convergence in the discrete case If Xn are independent random variables assuming value one with probability 1/n and zero otherwise, then Xn converges to zero in probability but not almost surely. This can be verified using the Borel–Cantelli lemmas. Convergence in probability implies convergence in distribution Proof for the case of scalar random variables Lemma. Let X, Y be random variables, let a be a real number and ε > 0. Then Proof of lemma: Shorter proof of the lemma: We have for if and , then . Hence by the union bound, Proof of the theorem: Recall that in order to prove convergence in distribution, one must show that the sequence of cumulative distribution functions converges to the FX at every point where FX is continuous. Let a be such a point. For every ε > 0, due to the preceding lemma, we have: So, we have Taking the limit as n → ∞, we obtain:
https://en.wikipedia.org/wiki/Anatoli%20Bugorski
Anatoli Petrovich Bugorski (; born 25 June 1942) is a Russian retired particle physicist. He is known for surviving a radiation accident in 1978, when a high-energy proton beam from a particle accelerator passed through his brain. Accident As a researcher at the Institute for High Energy Physics in Protvino, Russian SSR, Anatoli Bugorski worked with the largest particle accelerator in the Soviet Union, the U-70 synchrotron. On 13 July 1978, Bugorski was checking a malfunctioning piece of equipment when the safety mechanisms failed. Bugorski was leaning over the equipment when he stuck his head in the path of the 76 GeV proton beam. Reportedly, he saw a flash "brighter than a thousand suns" but did not feel any pain. The beam passed through the back of his head, the occipital and temporal lobes of his brain, the left middle ear, and out through the left hand side of his nose. The exposed parts of his head received a local dose of 200,000 to 300,000 roentgens (2,000 to 3,000 Sieverts). Bugorski understood the severity of what had happened, but continued working on the malfunctioning equipment, and initially opted not to tell anyone what had happened. Aftermath The left half of Bugorski's face swelled up beyond recognition and, over the next several days, the skin started to peel, revealing the path that the proton beam had burned through parts of his face, his bone, and the brain tissue underneath. As it was believed that he had received far in excess of a fatal dose of radiation, Bugorski was taken to a clinic in Moscow where the doctors could observe his expected demise. However, Bugorski survived, completed his PhD, and continued working as a particle physicist. There was virtually no damage to his intellectual capacity, but the fatigue of mental work increased markedly. Bugorski completely lost hearing in the left ear, replaced by a form of tinnitus. The left half of his face was paralysed due to the destruction of nerves. He was able to function well, except
https://en.wikipedia.org/wiki/Nerve%20complex
In topology, the nerve complex of a set family is an abstract complex that records the pattern of intersections between the sets in the family. It was introduced by Pavel Alexandrov and now has many variants and generalisations, among them the Čech nerve of a cover, which in turn is generalised by hypercoverings. It captures many of the interesting topological properties in an algorithmic or combinatorial way. Basic definition Let be a set of indices and be a family of sets . The nerve of is a set of finite subsets of the index set . It contains all finite subsets such that the intersection of the whose subindices are in is non-empty: In Alexandrov's original definition, the sets are open subsets of some topological space . The set may contain singletons (elements such that is non-empty), pairs (pairs of elements such that ), triplets, and so on. If , then any subset of is also in , making an abstract simplicial complex. Hence N(C) is often called the nerve complex of . Examples Let X be the circle and , where is an arc covering the upper half of and is an arc covering its lower half, with some overlap at both sides (they must overlap at both sides in order to cover all of ). Then , which is an abstract 1-simplex. Let X be the circle and , where each is an arc covering one third of , with some overlap with the adjacent . Then . Note that {1,2,3} is not in since the common intersection of all three sets is empty; so is an unfilled triangle. The Čech nerve Given an open cover of a topological space , or more generally a cover in a site, we can consider the pairwise fibre products , which in the case of a topological space are precisely the intersections . The collection of all such intersections can be referred to as and the triple intersections as . By considering the natural maps and , we can construct a simplicial object defined by , n-fold fibre product. This is the Čech nerve. By taking connected components we get a simplicial s
https://en.wikipedia.org/wiki/Leverage%20cycle
Leverage is defined as the ratio of the asset value to the cash needed to purchase it. The leverage cycle can be defined as the procyclical expansion and contraction of leverage over the course of the business cycle. The existence of procyclical leverage amplifies the effect on asset prices over the business cycle. Why is leverage significant? Conventional economic theory suggests that interest rates determine the demand and supply of loans. This convention does not take into account the concept of default and hence ignores the need for collateral. When an investor buys an asset, they may use the asset as a collateral and borrow against it, however the investor will not be able to borrow the entire amount. The investor has to finance with their own capital the difference between the value of the collateral and the asset price, known as the margin. Thus the asset becomes leveraged. The need to partially finance the transaction with the investor's own capital implies that their ability to buy assets is limited by their capital at any given time. Impatient borrowers drive the interest rate higher while nervous lenders demand more collateral, a borrower's willingness to pay a higher interest to ease the concerns of the nervous lender may not necessarily satisfy the lender. Before the financial crisis of 2008 hit, lenders were less nervous. As a result, they were willing to make subprime mortgage loans. Consider an individual who took out a subprime mortgage loan paying a high interest relative to a prime mortgage loan and putting up only 5% collateral, a leverage of 20. During the crisis, lenders become more nervous. As a result, they demand 20% as collateral, even though there is sufficient liquidity in the system. The individual who took out a subprime loan is probably not in a position to buy a house now, regardless of how low the interest rates are. Therefore, in addition to interest rates, collateral requirements should also be taken into consideration in d
https://en.wikipedia.org/wiki/Majority%20problem
The majority problem, or density classification task, is the problem of finding one-dimensional cellular automaton rules that accurately perform majority voting. Using local transition rules, cells cannot know the total count of all the ones in system. In order to count the number of ones (or, by symmetry, the number of zeros), the system requires a logarithmic number of bits in the total size of the system. It also requires the system send messages over a distance linear in the size of the system and for the system to recognize a non-regular language. Thus, this problem is an important test case in measuring the computational power of cellular automaton systems. Problem statement Given a configuration of a two-state cellular automaton with i + j cells total, i of which are in the zero state and j of which are in the one state, a correct solution to the voting problem must eventually set all cells to zero if i > j and must eventually set all cells to one if i < j. The desired eventual state is unspecified if i = j. The problem can also be generalized to testing whether the proportion of zeros and ones is above or below some threshold other than 50%. In this generalization, one is also given a threshold ; a correct solution to the voting problem must eventually set all cells to zero if and must eventually set all cells to one if . The desired eventual state is unspecified if . Approximate solutions Gács, Kurdyumov, and Levin found an automaton that, although it does not always solve the majority problem correctly, does so in many cases. In their approach to the problem, the quality of a cellular automaton rule is measured by the fraction of the possible starting configurations that it correctly classifies. The rule proposed by Gacs, Kurdyumov, and Levin sets the state of each cell as follows. If a cell is 0, its next state is formed as the majority among the values of itself, its immediate neighbor to the left, and its neighbor three spaces to the left.
https://en.wikipedia.org/wiki/Economics%20and%20Human%20Biology
Economics and Human Biology is a quarterly peer-reviewed academic journal published by Elsevier since 2003. It is an interdisciplinary periodical covering research on biological economics — economics in the context of human biology and health. The current editors-in-chief are Susan Averett, Joerg Baten and Pinka Chatterji. According to the Journal Citation Reports, the journal has a 2020 impact factor of 2.184. See also Anthropometry Antebellum Puzzle Body mass index History of anthropometry Human height Human body weight
https://en.wikipedia.org/wiki/NUTS%20statistical%20regions%20of%20Luxembourg
In the NUTS (Nomenclature of Territorial Units for Statistics) codes of Luxembourg (LU), the three levels are: NUTS codes LU0 Luxembourg LU00 Luxembourg LU000 Luxembourg Local administrative units Below the NUTS levels, the two LAU (Local Administrative Units) levels are: The LAU codes of Luxembourg can be found in this spread sheet of all 28 EU member states LAU codes here, current as of 2018: See also Subdivisions of Luxembourg ISO 3166-2 codes of Luxembourg FIPS region codes of Luxembourg Sources Hierarchical list of the Nomenclature of territorial units for statistics - NUTS and the Statistical regions of Europe Overview map of EU Countries - NUTS level 1 LUXEMBOURG - NUTS level 2 LUXEMBOURG - NUTS level 3 Correspondence between the NUTS levels and the national administrative units List of current NUTS codes Download current NUTS codes (ODS format) Luxembourg Nuts
https://en.wikipedia.org/wiki/Image%20translation
Image translation is the machine translation of images of printed text (posters, banners, menus, screenshots etc.). This is done by applying optical character recognition (OCR) technology to an image to extract any text contained in the image, and then have this text translated into a language of their choice, and the applying digital image processing on the original image to get the translated image with a new language. General Machine translation made available on the internet (web and mobile) is a notable advance in multilingual communication eliminating the need for an intermediary translator/interpreter, translating foreign texts still poses a problem to the user as they cannot be expected to be able to type the foreign text they wish to translate and understand. Manually entering the foreign text may prove to be a difficulty especially in cases where an unfamiliar alphabet is used from a script which user can't read, e.g. Cyrillic, Chinese, Japanese etc. for an English speaker or any speaker of a Latin-based language or vice versa. The technical advancements in OCR made it possible to recognize text from images. The possibility to use one's mobile device's camera to capture and extract printed text is also known as mobile OCR and was first introduced in Japanese manufactured mobile telephones in 2004. Using the handheld's camera one could take a picture of (a line of) text and have it extracted (digitalized) for further manipulation such as storing the information in their contacts list, as a web page address (URL) or text to use in an SMS/email message etc. Presently, mobile devices having a camera resolution of 2 megapixels or above with an auto-focus ability, often feature the text scanner service. Taking the text scanning facility one step further, image translation emerged, giving users the ability to capture text with their mobile phone's camera, extract the text, and have it translated in their own language. More and more applications emerged on thi
https://en.wikipedia.org/wiki/Nuremberg%20kitchen
Nuremberg kitchen is the traditional English name for a specific type of dollhouse, similar to a room box, usually limited to a single room depicting a kitchen. The name references the city of Nuremberg, the center of the nineteenth-century German toy industry. In German the toy is known as a Puppenküche (literally “dolls’ kitchen”). Description Most surviving examples show variations on a standard form: a single reduced-scale room with the front wall and ceiling missing, rather like a miniaturized stage set, allowing convenient access to the interior and an unobstructed view of the minuscule items within. Often the side walls flair out from the back at wide angles, creating a trapezoid floorplan and presenting a more dramatic display of the contents. Some might have a roof above or a pantry to one side, but these are exceptions. Typically, but not always, the fittings are arranged symmetrically, with a cooking range in the center of the rear wall (a raised masonry hearth with a chimney in early versions, or a metal stove in later ones), with cupboards, shelves, and other storage furniture to either side. They often house an abundant collection of pots, pans, and dishes filling or even overflowing the space. Later nineteenth-century examples are often highly embellished with decorative trimmings. Many of these features pertain more to making these miniatures seem attractive than they do to accurately depicting full-scale kitchens. History Nuremberg kitchens date back at least to 1572, when one was given to Dorothea and Anna, the Princesses of Saxony, daughters of Augustus, Elector of Saxony aged five and ten. Since then, many adult collectors as well as children have owned multi-room dollhouses, but these one-room kitchens seem to have almost always been thought of as girls’ playthings. They reached the height of their popularity in the 1800s. In the early part of the century they were assembled by artisans working from their homes, who produced a remarkably larg
https://en.wikipedia.org/wiki/Argillipedoturbation
Argillipedoturbation, sometimes referred to as self-mulching, is a process of soil mixing caused by the shrinking and swelling of the smectite clays contained in soil. It is an effect specific to soils of the vertisolic variety, and is triggered by the constant cycles of wetting and drying It is characterized by wide (up to ), deep ( or more) vertical cracks in the solum that contain differing materials from the rest of the soil layer they are found in, as well as sloughed-in surface materials. In order for argillipedoturbation to occur, the soil must be at least 30% clay content. The expression of argillipedoturbation depends to a large degree on the exact clay content of the soil, as well as on what other minerals make up the soil composition. Argillipedoturbation can be strong enough that it can affect the soil horizons by combining the different horizons, making them difficult to distinguish. It can also result in a gently-rolling surface referred to as gilgai topography and in the dramatic soil inclusions known as slickensides. In addition, argillipedoturbation sometimes results in a chernozemic-like A-type horizon, or one resembling a gleysolic order soil. This process can also affect the distribution of rock fragments, by moving fragments at the surface to lower soil layers and vice versa. The effects of this process are useful in agriculture, as the organic surface materials fertilize the soil and cause them to become very productive when irrigated. However, they are very difficult to plow and manage due to the high, thoroughly mixed clay content.
https://en.wikipedia.org/wiki/Zacusc%C4%83
Zacuscă () is a vegetable spread popular in Romania and Moldova. Similar spreads are found in other countries in the Balkan region, and bordering regions. Ingredients The main ingredients are roasted eggplant, sauteed onions, tomato paste, and roasted Paprika Pepper (Romanian pepper called gogoșari). Some add mushrooms, carrots, or celery. Bay leaves are added as spice, as well as other ingredients (oil, salt, and pepper). Traditionally, a family will cook a large quantity of it after the fall harvest and preserve it through canning. Use Zacuscă can be eaten as a relish or spread, typically on bread. It is said to improve in taste after some months of maturing but must be used within days of opening. Although traditionally prepared at home, it is also commercially available. Some Bulgarian and Middle Eastern brands are available in the United States. In the Orthodox Christian majority countries, it is sometimes eaten during fasting seasons due to the absence of meat, eggs or dairy products. Etymology The word zacuscă is of Slavic origin which means simply "appetizer", "breakfast" or "snack" (see zakuska) See also Ajvar, pindjur and ljutenica, similar spreads in Balkan cuisine Kyopolou, a similar Bulgarian dish Biber salçası, a Turkish paste made from red peppers alone List of eggplant dishes List of spreads
https://en.wikipedia.org/wiki/TLS-SRP
Transport Layer Security Secure Remote Password (TLS-SRP) ciphersuites are a set of cryptographic protocols that provide secure communication based on passwords, using an SRP password-authenticated key exchange. There are two classes of TLS-SRP ciphersuites: The first class of cipher suites uses only SRP authentication. The second class uses SRP authentication and public key certificates together for added security. Usually, TLS uses only public key certificates for authentication. TLS-SRP uses a value derived from a password (the SRP verifier) and a salt, shared in advance among the communicating parties, to establish a TLS connection. There are several possible reasons one may choose to use TLS-SRP: Using password-based authentication does not require reliance on certificate authorities. The end user does not need to check the URL being certified. If the server does not know data derived from the password then the connection simply cannot be made. This prevents Phishing. Password authentication is less prone than certificate authentication to certain types of configuration mistakes, such as expired certificates or mismatched common name fields. TLS-SRP provides mutual authentication (the client and server both authenticate each other), while TLS with server certificates only authenticates the server to the client. Client certificates can authenticate the client to the server, but it may be easier for a user to remember a password than to install a certificate. Implementations TLS-SRP is implemented in GnuTLS, OpenSSL as of release 1.0.1, Apache mod_gnutls and mod_ssl, cURL, TLS Lite SecureBlackbox and wolfSSL. Standards RFC 2945: “The SRP Authentication and Key Exchange System”. RFC 5054: “Using the Secure Remote Password (SRP) Protocol for TLS Authentication”. See also Transport Layer Security
https://en.wikipedia.org/wiki/Tengion
Tengion, Inc. is an American development-stage regenerative medicine company founded in 2003 with financing from J&J Development Corporation, HealthCap and Oak Investment Partners, which is headquartered in Winston-Salem, North Carolina. Its goals are discovering, developing, manufacturing and commercializing a range of replacement organs and tissues, or neo-organs and neo-tissues, to address unmet medical needs in urologic, renal, gastrointestinal, and vascular diseases and disorders. The company creates these human neo-organs from a patient’s own cells or autologous cells, in conjunction with its Organ Regeneration Platform. The company declared Chapter 7 bankruptcy in December 2014, and it, along with its assets and tissue engineering samples, have been bought back by its creditors and former executives as of March 2015. The purchase was expedited, so that time-sensitive research can continue. History Founded in 2003 and formerly headquartered in East Norriton, Pennsylvania before moving to Winston-Salem, North Carolina in 2012, Tengion went public in 2010, after its stock has been approved for listing on the NASDAQ, through a $26 million IPO to help advance its research and development activities. Some of the groundbreaking regenerative medicine technologies of Dr. Anthony Atala, director of the Wake Forest Institute for Regenerative Medicine, were the core from where those research and development activities developed. On September 4, 2012, Tengion received a notice from NASDAQ stating that the company had not regained compliance with NASDAQ Listing Rule 5550(b)(1) and that its common stock would cease trading on the NASDAQ Capital Market effective on September 6, 2012, and would begin trading on the OTCQB tier of the OTC Marketplace. The company was bought by former executives and creditors after declaring bankruptcy in 2014. Products All current Tengion's regenerative medicine product candidates are investigational and will not be commercially available unt
https://en.wikipedia.org/wiki/Cooperative%20breeding
Cooperative breeding is a social system characterized by alloparental care: offspring receive care not only from their parents, but also from additional group members, often called helpers. Cooperative breeding encompasses a wide variety of group structures, from a breeding pair with helpers that are offspring from a previous season, to groups with multiple breeding males and females (polygynandry) and helpers that are the adult offspring of some but not all of the breeders in the group, to groups in which helpers sometimes achieve co-breeding status by producing their own offspring as part of the group's brood. Cooperative breeding occurs across taxonomic groups including birds, mammals, fish, and insects. Costs for helpers include a fitness reduction, increased territory defense, offspring guarding and an increased cost of growth. Benefits for helpers include a reduced chance of predation, increased foraging time, territory inheritance, increased environmental conditions and an inclusive fitness. Inclusive fitness is the sum of all direct and indirect fitness, where direct fitness is defined as the amount of fitness gained through producing offspring. Indirect fitness is defined as the amount of fitness gained through aiding the offspring of related individuals, that is, relatives are able to indirectly pass on their genes through increasing the fitness of related offspring. This is also called kin selection. For the breeding pair, costs include increased mate guarding and suppression of subordinate mating. Breeders receive benefits as reductions in offspring care and territory maintenance. Their primary benefit is an increased reproductive rate and survival. Cooperative breeding causes the reproductive success of all sexually mature adults to be skewed towards one mating pair. This means the reproductive fitness of the group is held within a select few breeding members and helpers have little to no reproductive fitness. With this system, breeders gain an incre
https://en.wikipedia.org/wiki/15.ai
15.ai is a non-commercial freeware artificial intelligence web application that generates natural emotive high-fidelity text-to-speech voices from an assortment of fictional characters from a variety of media sources. Developed by a pseudonymous MIT researcher under the name 15, the project uses a combination of audio synthesis algorithms, speech synthesis deep neural networks, and sentiment analysis models to generate and serve emotive character voices faster than real-time, particularly those with a very small amount of trainable data. Launched in early 2020, 15.ai began as a proof of concept of the democratization of voice acting and dubbing using technology. Its gratis and non-commercial nature (with the only stipulation being that the project be properly credited when used), ease of use, no user account registration requirement, and substantial improvements to current text-to-speech implementations have been lauded by users; however, some critics and voice actors have questioned the legality and ethicality of leaving such technology publicly available and readily accessible. Credited as the impetus behind the popularization of AI voice cloning (also known as audio deepfakes) in content creation and as the first publicly available AI vocal synthesis project to involve the use of existing popular fictional characters, 15.ai has had a significant impact on multiple Internet fandoms, most notably the My Little Pony: Friendship Is Magic, Team Fortress 2, and SpongeBob SquarePants fandoms. Furthermore, 15.ai has inspired the use of 4chan's Pony Preservation Project in other generative artificial intelligence projects. Several commercial alternatives have spawned with the rising popularity of 15.ai, leading to cases of misattribution and theft. In January 2022, it was discovered that Voiceverse NFT, a company that voice actor Troy Baker announced his partnership with, had plagiarized 15.ai's work as part of their platform. On September 8, 2022, 15.ai was temporari
https://en.wikipedia.org/wiki/Proteomyxa
Proteomyxa is a name given by E. Ray Lankester to a group of Sarcodina. This is an obsolete group. Many of the species are endoparasites in living cells, mostly of algae or fungi, but not exclusively. At least two species of Pseudospora have been taken for reproductive stages in the life history of their hosts—whence indeed the generic name. Plasmodiophora brassicae gives rise to the disease known as Hanburies or fingers and toes in Cruciferae; Lymphosporidium causes a virulent epidemic among the American brook trout (Salvelinus fontinalis). Archerina boltoni is remarkable for containing a pair of chlorophyll corpuscles in each cell; no nucleus has been made out, but the chlorophyll bodies divide previous to fission. It is a fresh-water form. The cells of this species form loose aggregates or filoplasmodia, like those of Mikrogromia or Leydenia.
https://en.wikipedia.org/wiki/Mayhaw
Mayhaw is the name given to the fruit of the species of Crataegus series Aestivales that are common in wetlands throughout the southern United States. The principal species are C. aestivalis, the eastern mayhaw, and C. opaca, the western mayhaw. Mayhaws grow in moist soil in river and creek bottoms under hardwood trees. The fruit is also found in bayous surrounding lakes, such as Caddo Lake on the Texas/Louisiana border. The fruit ripens in late April through May, thus the name may-haw. Mayhaws are often collected out of the water from boats, and the fruit is used to make jelly. Families would go on outings to collect mayhaws and create stockpiles of the jelly to last throughout the year, but the tradition has declined with the increasing urbanization of the South and the destruction of the mayhaw's native habitat. The fruit has also been cultivated to grow outside of wetlands, and this is increasingly the source of the jelly. In culture Many communities associate themselves with the fruit because of its reputation as a celebrated delicacy of Southern U.S. cuisine. For example, Colquitt, Georgia, holds a mayhaw festival in April. Daisetta, Texas; El Dorado, Arkansas; Marion, Louisiana; and Starks, Louisiana, all celebrate a mayhaw festival each May. The mayhaw is the state fruit tree of Louisiana.
https://en.wikipedia.org/wiki/Chaffing%20and%20winnowing
Chaffing and winnowing is a cryptographic technique to achieve confidentiality without using encryption when sending data over an insecure channel. The name is derived from agriculture: after grain has been harvested and threshed, it remains mixed together with inedible fibrous chaff. The chaff and grain are then separated by winnowing, and the chaff is discarded. The cryptographic technique was conceived by Ron Rivest and published in an on-line article on 18 March 1998. Although it bears similarities to both traditional encryption and steganography, it cannot be classified under either category. This technique allows the sender to deny responsibility for encrypting their message. When using chaffing and winnowing, the sender transmits the message unencrypted, in clear text. Although the sender and the receiver share a secret key, they use it only for authentication. However, a third party can make their communication confidential by simultaneously sending specially crafted messages through the same channel. How it works The sender (Alice) wants to send a message to the receiver (Bob). In the simplest setup, Alice enumerates the symbols in her message and sends out each in a separate packet. If the symbols are complex enough, such as natural language text, an attacker may be able to distinguish the real symbols from poorly-faked chaff symbols, posing a similar problem as steganography in needing to generate highly realistic fakes; to avoid this, the symbols can be reduced to just single 0/1 bits, and realistic fakes can then be simply randomly generated 50:50 and are indistinguishable from real symbols. In general the method requires each symbol to arrive in-order and to be authenticated by the receiver. When implemented over networks that may change the order of packets, the sender places the symbol's serial number in the packet, the symbol itself (both unencrypted), and a message authentication code (MAC). Many MACs use a secret key Alice shares with Bob, bu
https://en.wikipedia.org/wiki/SoundGrid
SoundGrid is a networking and processing platform audio application made by Waves Audio and developed in cooperation with DiGiCo. It consists of a Linux-based server that runs the SoundGrid environment, compatible plug-ins, a Mac or Windows control computer, and an audio interface for input/output (I/O). It is used for live sound, broadcast, and post-production and provides a low-latency environment for audio processing on certain hardware audio mixing consoles, e.g., DiGiCo, Allen & Heath, and Yamaha. Features Low latency (less than 1ms) for SoundGrid-compatible plugins. Runs on Intel CPUs and 1Gbit/s Ethernet networks. Integrates with analog and digital mixing consoles. Provides redundancy and recovery. Splits output to record on a standard digital audio workstation (DAW). Comprises network infrastructure for sound installations. Audio transport and system control SoundGrid is a proprietary Ethernet Layer 2 protocol and EtherType. It routes audio between networked I/O devices and processes it on plugin servers connected to the same network. The I/O device converts SoundGrid packets to standard and proprietary audio protocols. Audio processing Using native processing, SoundGrid runs on standard CPUs under a modified Linux operating system (OS). Waves Audio says this provides "predictability, stability, and low latency," previously exclusive to dedicated DSP-based systems. Separate computers provide SoundGrid processing: One or more SoundGrid servers are dedicated to audio processing in a customized Linux OS optimized for audio processing. A Windows or Mac computer runs SoundGrid Host, the host application, and the user interface. Audio interfacing Audio interfaces with SoundGrid by integrating a SoundGrid-programmed FPGA (Xilinx Spartan 3) into a mixing console's I/O ports. The FPGA receives I2S or other audio signal formats and converts them to the SoundGrid format. The FPGA also transfers control messages between control nodes external to the So
https://en.wikipedia.org/wiki/Pacemaker%20crosstalk
Pacemaker crosstalk results when the pacemaker-generated electrical event in one chamber is sensed by the lead in another chamber, resulting in inappropriate inhibition of the pacing artifact in the second chamber. Cause Crosstalk can only occur in dual chamber or biventricular pacemaker. It happens less often in more recent models of dual chamber pacemakers due to the addition of a ventricular blanking period, which coincides with the atrial stimulus. This helps to prevent ventricular channel oversensing of atrial output. Newer dual chamber pacemakers also use bipolar leads with a smaller pacing spike, and steroid eluting leads with lower pacing thresholds. Crosstalk is more common in unipolar systems since they require a larger pacing spike. Crosstalk is sometimes referred to as crosstalk inhibition, far-field sensing, or self-inhibition. In some cases, crosstalk can occur in the pulse generator circuit itself, though more common causes include atrial lead dislodgement into the ventricle, ventricular lead dislodgement into the atrium, high atrial output current, high ventricular sensitivity, and short ventricular blanking period. Treatment In general, the treatment of crosstalk includes decreasing atrial pacing output, decreasing atrial pulse width, decreasing ventricular sensitivity, increasing the ventricular blanking period, activating ventricular safety pacing, and new atrial lead implant if insulation failure mandates unipolar programming. See also Pacemaker failure Electrical conduction system of the heart
https://en.wikipedia.org/wiki/Pericorneal%20plexus
The pericorneal plexus refers to a network of blood vessels in the eye; specifically to branches of the anterior conjunctival arteries. These vessels are arranged around the cornea in superficial and deep layers.
https://en.wikipedia.org/wiki/Infant%20food%20safety
Foodborne illness (also foodborne disease and colloquially referred to as food poisoning) is any illness resulting from the food spoilage of contaminated food, pathogenic bacteria, viruses, or parasites that contaminate food. Infant food safety is the identification of risky food handling practices and the prevention of illness in infants. Foodborne illness is a serious health issue, especially for babies and children. Infants and young children are particularly vulnerable to foodborne illness because their immune systems are not developed enough to fight off foodborne bacterial infections. In fact, 800,000 illnesses affect children under the age of 10 in the U.S. each year. Therefore, extra care should be taken when handling and preparing their food. Prevention Handwashing is the first step in maintaining the safety of infant food. Caregivers hands can pick up bacteria and spread bacteria to the baby. Situations in which one can encounter high levels of bacteria are: Diapers containing feces and urine Raw meat and raw poultry Uncooked seafood, and eggs Dogs and cats, turtles, snakes, birds, and lizards, among other animals. Soil Other children Handwashing can remove harmful bacteria and will help to prevent foodborne illness. Instructing other children in a family on good handwashing will help to limit the spread of bacteria that cause illness. Handwashing is most effective in providing safe food for the infant during 'key times': Before preparing and feeding bottles or foods to the baby. Before touching the baby's mouth. Before touching pacifiers or other things that go into the baby's mouth. After using the toilet or changing diapers. Infant formula Though breastfeeding helps prevent many kinds of sicknesses among infants, caregivers often choose to use infant formula. Promoting food safety in infants requires safe preparation and use. Use infant formula within two hours of preparation. If the infant does not finish the entire bottle, the remainder
https://en.wikipedia.org/wiki/Energy%20recovery%20linac
An energy recovery linac (ERL) is a type of linear particle accelerator that provides a beam of electrons used to produce x-rays by synchrotron radiation. First proposed in 1965 the idea gained interest since the early 2000s. Spectral radiance The usefulness of an x-ray beam for scientific experiments depends upon the beam's spectral radiance, which tells how much power of a given wavelength is concentrated on a spot. Most scientific literature on x-ray sources uses a closely related term called brilliance, which counts the rate of photons produced, rather than their power. The energy of a photon is inversely proportional to the photon's wavelength. Very high power is usually achieved by delivering the energy in short pulses, allowing the apparatus to work within reasonable power demands and cooling limits. Depending upon the pulse length and repetition rate, the average spectral radiance will be much lower than the peak spectral radiance. The peak spectral radiance and the average spectral radiance are both important properties of an x-ray beam. For some experiments, the peak value is most important, but for other experiments, the average value is most important. As a synchrotron light source, the performance of an energy recovery linac falls between a storage ring and a free-electron laser (FEL). Energy recovery linacs have high repetition rates and therefore high average spectral radiance, but lower peak spectral radiance than a FEL. Mechanism While using a recirculating charged particle beam with a magnet lattice resembling that of a storage ring, each particle travels through the recirculating arc before being decelerated in a linac structure. The same linac structure also accelerates new low-energy particles that are continuously injected into the linac. Thus, instead of recycling the particle beam continuously, while its emittance increases by synchrotron radiation emission, only its kinetic energy is recycled, enabling a low beam emittance while m
https://en.wikipedia.org/wiki/Gastric%20pits
Gastric pits are indentations in the stomach which denote entrances to 3-5 tubular shaped gastric glands. They are deeper in the pylorus than they are in the other parts of the stomach. The human stomach has several million of these pits which dot the surface of the lining epithelium. Surface mucous cells line the pits themselves but give way to a series of other types of cells which then line the glands themselves. Gastric acid Gastric acid also known as gastric juice is secreted from gastric glands, which are located in gastric pits. Gastric juice contains hydrochloric acid, pepsinogen and mucus in a healthy adult. Hydrochloric acid is secreted by parietal cells, pepsinogen is secreted by gastric chief cells and mucus is secreted by mucus neck cells.
https://en.wikipedia.org/wiki/Haplogroup%20P%20%28mtDNA%29
In human mitochondrial genetics, Haplogroup P is a human mitochondrial DNA (mtDNA) haplogroup. Origin Haplogroup P is a descendant of Haplogroup R. Distribution Today, P is most commonly found in Oceania, especially in Papuans, Melanesians, indigenous Australians, It's 1.4% in mainstream Filipinos but 1.13% in Luzon, 1.78% in Visayas, 1.43% in Mindanao. It is much higher in Sub-Filipinos groups, 6.67% in Bugkalot and 11.2% in Maranao. It was found in the Philippines Negrito Aeta of Bataan at 40%. It is also found in the Malaysians at 0.9%, including Indonesians. Subclades Tree This phylogenetic tree of haplogroup P subclades is based on the paper by Mannis van Oven and Manfred Kayser Updated comprehensive phylogenetic tree of global human mitochondrial DNA variation and subsequent published research. P (16176) P1 P1d P1d1 P2'10 P2 P10 P8 P3 P3a P3b P3b1 P4 P4a P4a1 P4b P4b1 P5 P6 P7 P9 See also Genealogical DNA test Genetic genealogy Human mitochondrial genetics Population genetics
https://en.wikipedia.org/wiki/Wilson%20quotient
The Wilson quotient W(p) is defined as: If p is a prime number, the quotient is an integer by Wilson's theorem; moreover, if p is composite, the quotient is not an integer. If p divides W(p), it is called a Wilson prime. The integer values of W(p) are : W(2) = 1 W(3) = 1 W(5) = 5 W(7) = 103 W(11) = 329891 W(13) = 36846277 W(17) = 1230752346353 W(19) = 336967037143579 ... It is known that where is the k-th Bernoulli number. Note that the first relation comes from the second one by subtraction, after substituting and . See also Fermat quotient
https://en.wikipedia.org/wiki/Mutual%20coherence%20%28linear%20algebra%29
In linear algebra, the coherence or mutual coherence of a matrix A is defined as the maximum absolute value of the cross-correlations between the columns of A. Formally, let be the columns of the matrix A, which are assumed to be normalized such that The mutual coherence of A is then defined as A lower bound is A deterministic matrix with the mutual coherence almost meeting the lower bound can be constructed by Weil's theorem. This concept was reintroduced by David Donoho and Michael Elad in the context of sparse representations. A special case of this definition for the two-ortho case appeared earlier in the paper by Donoho and Huo. The mutual coherence has since been used extensively in the field of sparse representations of signals. In particular, it is used as a measure of the ability of suboptimal algorithms such as matching pursuit and basis pursuit to correctly identify the true representation of a sparse signal. Joel Tropp introduced a useful extension of Mutual Coherence, known as the Babel function, which extends the idea of cross-correlation between pairs of columns to the cross-correlation from one column to a set of other columns. The Babel function for two columns is exactly the Mutual coherence, but it also extends the coherence relationship concept in a way that is useful and relevant for any number of columns in the sparse representation matix as well. See also Compressed sensing Restricted isometry property Babel function
https://en.wikipedia.org/wiki/Neural%20modeling%20fields
Neural modeling field (NMF) is a mathematical framework for machine learning which combines ideas from neural networks, fuzzy logic, and model based recognition. It has also been referred to as modeling fields, modeling fields theory (MFT), Maximum likelihood artificial neural networks (MLANS). This framework has been developed by Leonid Perlovsky at the AFRL. NMF is interpreted as a mathematical description of the mind's mechanisms, including concepts, emotions, instincts, imagination, thinking, and understanding. NMF is a multi-level, hetero-hierarchical system. At each level in NMF there are concept-models encapsulating the knowledge; they generate so-called top-down signals, interacting with input, bottom-up signals. These interactions are governed by dynamic equations, which drive concept-model learning, adaptation, and formation of new concept-models for better correspondence to the input, bottom-up signals. Concept models and similarity measures In the general case, NMF system consists of multiple processing levels. At each level, output signals are the concepts recognized in (or formed from) input, bottom-up signals. Input signals are associated with (or recognized, or grouped into) concepts according to the models and at this level. In the process of learning the concept-models are adapted for better representation of the input signals so that similarity between the concept-models and signals increases. This increase in similarity can be interpreted as satisfaction of an instinct for knowledge, and is felt as aesthetic emotions. Each hierarchical level consists of N "neurons" enumerated by index n=1,2..N. These neurons receive input, bottom-up signals, X(n), from lower levels in the processing hierarchy. X(n) is a field of bottom-up neuronal synaptic activations, coming from neurons at a lower level. Each neuron has a number of synapses; for generality, each neuron activation is described as a set of numbers, , where D is the number or dimensions
https://en.wikipedia.org/wiki/Aczel%27s%20anti-foundation%20axiom
In the foundations of mathematics, Aczel's anti-foundation axiom is an axiom set forth by , as an alternative to the axiom of foundation in Zermelo–Fraenkel set theory. It states that every accessible pointed directed graph corresponds to exactly one set. In particular, according to this axiom, the graph consisting of a single vertex with a loop corresponds to a set that contains only itself as element, i.e. a Quine atom. A set theory obeying this axiom is necessarily a non-well-founded set theory. Accessible pointed graphs An accessible pointed graph is a directed graph with a distinguished vertex (the "root") such that for any node in the graph there is at least one path in the directed graph from the root to that node. The anti-foundation axiom postulates that each such directed graph corresponds to the membership structure of exactly one set. For example, the directed graph with only one node and an edge from that node to itself corresponds to a set of the form x = {x}. See also von Neumann universe
https://en.wikipedia.org/wiki/Snatcher%20%28video%20game%29
Snatcher is a cyberpunk graphic adventure game developed and published by Konami. It was written and designed by Hideo Kojima and first released in 1988 for the PC-8801 and MSX2 in Japan. Snatcher is set in a future East Asian metropolis where humanoid robots dubbed "Snatchers" have been discovered killing humans and replacing them in society. The game follows Gillian Seed, an amnesiac who joins an anti-Snatcher agency in search of his past. Gameplay takes place primarily through a menu-based interface through which the player can choose to examine items, search rooms, speak to characters, explore a semi-open world, and perform other actions. Kojima wanted Snatcher to have a cinematic feel, so the setting and story are heavily influenced by science fiction films, like Blade Runner, Akira, The Terminator, and Invasion of the Body Snatchers. Development on the PC versions took more than twice as long as the average game of the time, even after Kojima was asked to trim more than half his initial story. The game was released to positive reviews, but poor sales. It garnered a cult following, and was remade as a role-playing game called SD Snatcher for the MSX2 in 1990. This was followed by a remake of the original adventure game using CD-ROM technology, released for the PC Engine Super CD-ROM² System in 1992. Looking to provide a more interactive experience to gamers in the West, Konami developed a Sega CD version of Snatcher specifically for North America and Europe in 1994. Although it was a commercial failure, the Sega CD version received mostly positive reviews for its cinematic presentation and mature themes uncommon in games at the time. Snatcher has been retrospectively acclaimed as both one of the best adventure and cyberpunk games of all time, and identified as a foundation for the themes Kojima explored later in the Metal Gear series. The game was a significant inspiration on Goichi Suda, who worked with Kojima to produce a radio drama prequel, Sdatcher. Snat