source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Fastigial%20nucleus
The fastigial nucleus is located in the cerebellum. It is one of the four deep cerebellar nuclei (the others being the nucleus dentatus, nucleus emboliformis and nucleus globosus), and is grey matter embedded in the white matter of the cerebellum. It refers specifically to the concentration of gray matter nearest to the middle line at the anterior end of the superior vermis, and immediately over the roof of the fourth ventricle (the peak of which is called the fastigium), from which it is separated by a thin layer of white matter. It is smaller than the nucleus dentatus, but somewhat larger than the nucleus emboliformis and nucleus globosus. Although it is one dense mass, it is made up of two sections: the rostral fastigial nucleus and the caudal fastigial nucleus. Structure The Purkinje cells of the cerebellar cortex project into the deep cerebellar nuclei and inhibit the excitatory output system via GABAergic synapses. The fastigial nucleus receives its input from Purkinje cells in the vermis. Most of its efferent connections travel via the inferior cerebellar peduncle to the vestibular nuclei, which are located at the junction of the pons and the medulla oblongata. The fastigial nucleus sends excitatory projections beyond the cerebellum. The likely neurotransmitters of fastigial nucleus axons are glutamate and aspartate. Rostral fastigial nucleus The rostral fastigial nucleus (rFN) is related to the vestibular system. It receives input from the vestibular nuclei and contributes to vestibular neuronal activity. The rFN interprets body motion and places it on spatial planes to estimate the movement of the body through space. It deals with antigravity muscle groups and other synergies involved with standing and walking. Caudal fastigial nucleus The caudal fastigial nucleus (cFN) is related to saccadic eye movements. The Purkinje cell output from the oculomotor vermis relays through the cFN, where neurons directly related to saccadic eye movements are locate
https://en.wikipedia.org/wiki/Thermoanaerobacter%20kivui
Thermoanaerobacter kivui (formerly Acetogenium kivui) is a thermophilic, anaerobic, nonspore-forming species of bacteria. T. kivui was originally isolated from Lake Kivu in Africa. The growth range for the organism is 50 to 72°C at pH 5.3-7.3, with optimal growth conditions at 66°C and pH 6.4. Although the organism stains Gram-negative, it shows a Gram-positive cell structure. The original genus Acetogenium was named because this organism principally produces acetic acid from substrates. T. kivui was originally named Acetogenium kivui, which was the only species within a new genus. However, further 16S ribosomal RNA studies put this bacterium into genus Thermoanaerobacter and the previous genus was no longer necessary.
https://en.wikipedia.org/wiki/Bioasphalt
Bioasphalt is an asphalt alternative made from non-petroleum based renewable resources. These sources include sugar, molasses and rice, corn and potato starches, natural tree and gum resins, natural latex rubber and vegetable oils, lignin, cellulose, palm oil waste, coconut waste, peanut oil waste, canola oil waste, dried sewerage effluent and so on. Bitumen can also be made from waste vacuum tower bottoms produced in the process of cleaning used motor oils, which are normally burned or dumped into land fills. Non-petroleum based bitumen binders can be colored, which can reduce the temperatures of road surfaces and reduce the Urban heat islands. Petroleum, environmental, and heat concerns Because of concerns over Peak oil, pollution and climate change, as well the oil price increases since 2003, non-petroleum alternatives have become more popular. This has led to the introduction of biobitumen alternatives that are more environmentally friendly and nontoxic. For millions of people living in and around cities, heat islands are of growing concern. This phenomenon describes urban and suburban temperatures that are 1 to 6 °C (2 to 10 °F) hotter than nearby rural areas. Elevated temperatures can impact communities by increasing peak energy demand, air conditioning costs, air pollution levels, and heat-related illness and mortality. There are common-sense measures that communities can take to reduce the negative effects of heat islands, such as replacing conventional black asphalt road surfaces with the new pigmentable bitumen that gives lighter colors. History and implementation Asphalt made with vegetable oil based binders was patented by Colas SA in France in 2004. A number of homeowners seeking an environmentally friendly alternative to asphalt for paving have experimented with waste vegetable oil as a binder for driveways and parking areas in single-family applications. The earliest known test occurred in 2002 in Ohio, where the homeowner combined waste ve
https://en.wikipedia.org/wiki/Platinum%E2%80%93samarium
Platinum-samarium is a binary inorganic compound of platinum and samarium with the chemical formula PtSm. This intermetallic compound forms crystals. Synthesis Fusion of stoichiometric amounts of pure substances: Physical properties Platinum-samarium forms crystals of rhombic crystal system, space group P nma, cell parameters a = 0.7148 nm, b = 0.4501 nm, c = 0.5638 nm, Z = 4, structure similar to that of iron boride (FeB). The compound melts congruently at a temperature of ≈1810 °C.
https://en.wikipedia.org/wiki/Formation%20evaluation%20gamma%20ray
The formation evaluation gamma ray log is a record of the variation with depth of the natural radioactivity of earth materials in a wellbore. Measurement of natural emission of gamma rays in oil and gas wells are useful because shales and sandstones typically have different gamma ray levels. Shales and clays are responsible for most natural radioactivity, so gamma ray log often is a good indicator of such rocks. In addition, the log is also used for correlation between wells, for depth correlation between open and cased holes, and for depth correlation between logging runs. Physics Natural radioactivity is the spontaneous decay of the atoms of certain isotopes into other isotopes. If the resultant isotope is not stable, it undergoes further decay until a stable isotope forms. The decay process is usually accompanied by emissions of alpha, beta, and gamma radiation. Natural gamma ray radiation is one form of spontaneous radiation emitted by unstable nuclei. Gamma (γ) radiation may be considered either as an electromagnetic wave similar to visible light or X-rays, or as a particle of photon. Gamma rays are electromagnetic radiations emitted from an atomic nucleus during radioactive decay, with the wavelength in the range of 10−9 to 10−11cm Natural radioactivity in rocks Isotopes naturally found on earth are usually those that are stable or have a decay time larger than, or at least a significant fraction of the age of the earth (about 5 x 109 years). Isotopes with shorter halflifes mainly exist as decay products from longer lived isotopes, and, as in C14, from irradiation of the upper atmosphere. Radioisotopes with a sufficiently long halflife, and whose decay produces an appreciable amount of gamma rays are: Potassium 40K with half-life of 1.3 x 109 years, which emits 0 α, 1 β, and 1 γ-ray Thorium 232Th with half-life of 1.4 x 1010 years, which emits 7 α, 5 β, and numerous γ-ray with different energies Uranium 238U with half-life of 4.4 x 109 years, which
https://en.wikipedia.org/wiki/Alan%20Beddoe
Lieutenant-Commander Alan Brookman Beddoe, OC, OBE, HFHS, FHSC (June 1, 1893 – December 2, 1975) was a Canadian artist, war artist, consultant in heraldry and founder and first president of the Heraldry Society of Canada in 1965. Born in Ottawa, Ontario, in 1893, he studied at Ashbury College. During World War I, he was captured at Second Battle of Ypres in 1915 and spent two and a half years in the prisoner of war camps at Gießen and Zerbst. He studied art at the Ecole des Beaux-Arts in Paris. After the war, he studied at the Art Students League of New York under DuMond and Bridgman. In 1925, he opened the first commercial art studio in Ottawa. He was also an expert in heraldry. The Alan Beddoe collection at Library and Archives Canada contains designs and studies for the Book of Remembrance, postage stamps, posters, crests, money, architecture, coats-of-arms, and a new Canadian flag. His fonds include slides, colour transparencies, prints, watercolours and drawings related to Canadian heraldry. Books of Remembrance Beddoe was instrumental in the creation of the major Books of Remembrance, now housed in the Peace Tower on Parliament Hill in Ottawa. The artist originally chosen for the job, James Purves, died in 1940, at which time Beddoe took on the task. He supervised a team of artists for about the next 2 years to illuminate and hand-letter the books, listing the names of Canadians who died in Canada's military service during World War I and after World War II he supervised another team of artists to create the Book of Remembrance for World War II. He was inducted to the OBE and received the Allied Arts Medal awarded by the Royal Architectural Institute for his work on the Books of Remembrance and made an officer of the Order of Canada. He also was instrumental in the creation of the South Africa Book of Remembrance 1956–1966; Yvonne Diceman, who had worked with him on the Book of Remembrance WWII, produced the Korea Book of Remembrance 1957–1958 and the Ne
https://en.wikipedia.org/wiki/Sequent%20calculus
In mathematical logic, sequent calculus is a style of formal logical argumentation in which every line of a proof is a conditional tautology (called a sequent by Gerhard Gentzen) instead of an unconditional tautology. Each conditional tautology is inferred from other conditional tautologies on earlier lines in a formal argument according to rules and procedures of inference, giving a better approximation to the natural style of deduction used by mathematicians than to David Hilbert's earlier style of formal logic, in which every line was an unconditional tautology. More subtle distinctions may exist; for example, propositions may implicitly depend upon non-logical axioms. In that case, sequents signify conditional theorems in a first-order language rather than conditional tautologies. Sequent calculus is one of several extant styles of proof calculus for expressing line-by-line logical arguments. Hilbert style. Every line is an unconditional tautology (or theorem). Gentzen style. Every line is a conditional tautology (or theorem) with zero or more conditions on the left. Natural deduction. Every (conditional) line has exactly one asserted proposition on the right. Sequent calculus. Every (conditional) line has zero or more asserted propositions on the right. In other words, natural deduction and sequent calculus systems are particular distinct kinds of Gentzen-style systems. Hilbert-style systems typically have a very small number of inference rules, relying more on sets of axioms. Gentzen-style systems typically have very few axioms, if any, relying more on sets of rules. Gentzen-style systems have significant practical and theoretical advantages compared to Hilbert-style systems. For example, both natural deduction and sequent calculus systems facilitate the elimination and introduction of universal and existential quantifiers so that unquantified logical expressions can be manipulated according to the much simpler rules of propositional calculus. In a typic
https://en.wikipedia.org/wiki/Time-sharing%20system%20evolution
This article covers the evolution of time-sharing systems, providing links to major early time-sharing operating systems, showing their subsequent evolution. The meaning of the term time-sharing has shifted from its original usage. From 1949 to 1960, time-sharing was used to refer to multiprogramming; it evolved to mean multi-user interactive computing. Time-sharing Time-sharing was first proposed in the mid- to late-1950s and first implemented in the early 1960s. The concept was born out of the realization that a single expensive computer could be efficiently utilized by enabling multiprogramming, and, later, by allowing multiple users simultaneous interactive access. In 1984, Christopher Strachey wrote he considered the change in the meaning of the term time-sharing to be a source of confusion and not what he meant when he wrote his original paper in 1959. Without time-sharing, an individual user would enter bursts of information followed by long pauses; but with a group of users working at the same time, the pauses of one user would be filled by the activity of the others. Similarly, small slices of time spent waiting for disk, tape, or network input could be granted to other users. Given an optimal group size, the overall process could be very efficient. Each user would use their own computer terminal, initially electromechanical teleprinters such as the Teletype Model 33 ASR or the Friden Flexowriter; from about 1970 these were progressively superseded by CRT-based units such as the DEC VT05, Datapoint 2200 and Lear Siegler ADM-3A. Terminals were initially linked to a nearby computer via current loop or serial cables, by conventional telegraph circuits provided by PTTs and over specialist digital leased lines such T1. Modems such as the Bell 103 and successors, allowed remote and higher-speed use over the analogue voice telephone network. Family tree of major systems See details and additional systems in the table below. Relationships shown here are for
https://en.wikipedia.org/wiki/Speck%20%28cipher%29
Speck is a family of lightweight block ciphers publicly released by the National Security Agency (NSA) in June 2013. Speck has been optimized for performance in software implementations, while its sister algorithm, Simon, has been optimized for hardware implementations. Speck is an add–rotate–xor (ARX) cipher. The NSA began working on the Simon and Speck ciphers in 2011. The agency anticipated some agencies in the US federal government would need a cipher that would operate well on a diverse collection of Internet of Things devices while maintaining an acceptable level of security. Cipher description Speck supports a variety of block and key sizes. A block is always two words, but the words may be 16, 24, 32, 48 or 64 bits in size. The corresponding key is 2, 3 or 4 words. The round function consists of two rotations, adding the right word to the left word, xoring the key into the left word, then xoring the left word into the right word. The number of rounds depends on the parameters selected, as follows: The key schedule uses the same round function as the main block cipher. Reference code The following is the designers' reference implementation, written in C, of the Speck variant with a 128-bit block size and key, where key = (K[1], K[0]). It is adapted from their IACR ePrint. #include <stdint.h> #define ROR(x, r) ((x >> r) | (x << (64 - r))) #define ROL(x, r) ((x << r) | (x >> (64 - r))) #define R(x, y, k) (x = ROR(x, 8), x += y, x ^= k, y = ROL(y, 3), y ^= x) #define ROUNDS 32 void encrypt(uint64_t ct[2], uint64_t const pt[2], uint64_t const K[2]) { uint64_t y = pt[0], x = pt[1], b = K[0], a = K[1]; R(x, y, b); for (int i = 0; i < ROUNDS - 1; i++) { R(a, b, i); R(x, y, b); } ct[0] = y; ct[1] = x; } Note that this code computes the round keys (key schedule) on-demand. In practice, as with other block ciphers it is common for implementations to compute the round keys just once and
https://en.wikipedia.org/wiki/Yael%20Tauman%20Kalai
Yael Tauman Kalai is a cryptographer and theoretical computer scientist who works as a Senior Principal Researcher at Microsoft Research New England and as an adjunct professor at MIT in the Computer Science and Artificial Intelligence Lab. Education and career Kalai graduated from the Hebrew University of Jerusalem in 1997. She worked with Adi Shamir at the Weizmann Institute of Science, earning a master's degree there in 2001, and then moved to the Massachusetts Institute of Technology, where she completed her PhD in 2006 with Shafi Goldwasser as her doctoral advisor. She did postdoctoral study at Microsoft Research and the Weizmann Institute before becoming a faculty member at the Georgia Institute of Technology. She took a permanent position at Microsoft Research in 2008. She serves on the Scientific Advisory Board for the Institute for Computational and Experimental Research in Mathematics (ICERM). Contributions Kalai is known for co-inventing ring signatures, which has become a key component of numerous systems such as Cryptonote and Monero (cryptocurrency). Subsequently, together with her advisor Shafi Goldwasser, she demonstrated an insecurity in the widely used Fiat–Shamir heuristic. Her work on delegating computation has applications to cloud computing. Recognition Kalai was an invited speaker on mathematical aspects of computer science at the 2018 International Congress of Mathematicians. Her master's thesis introducing ring signatures won an outstanding master's thesis award and MIT PhD dissertation was awarded the George M. Sprowls Award for Outstanding PhD Thesis in Computer Science. She was co-chair of the Theory of Cryptography Conference in 2017. She was awarded the 2022 ACM Prize in Computing "for breakthroughs in verifiable delegation of computation and fundamental contributions to cryptography". Personal Kalai is the daughter of game theorist Yair Tauman. Her husband, Adam Tauman Kalai, also works at Microsoft Research.
https://en.wikipedia.org/wiki/Laue%20equations
In crystallography and solid state physics, the Laue equations relate incoming waves to outgoing waves in the process of elastic scattering, where the photon energy or light temporal frequency does not change upon scattering by a crystal lattice. They are named after physicist Max von Laue (1879–1960). The Laue equations can be written as as the condition of elastic wave scattering by a crystal lattice, where is the scattering vector, , are an incoming and outgoing wavevectors (to the crystal and from the crystal, by scattering), and is a crystal reciprocal lattice vector. Due to elastic scattering , three vectors. , , and , form a rhombus if the equation is satisfied. If the scattering satisfies this equation, all the crystal lattice points scatter the incoming wave toward the scattering direction (the direction along ). If the equation is not satisfied, then for any scattering direction, only some lattice points scatter the incoming wave. (This physical interpretation of the equation is based on the assumption that scattering at a lattice point is made in a way that the scattering wave and the incoming wave have the same phase at the point.) It also can be seen as the conservation of momentum as since is the wavevector for a plane wave associated with parallel crystal lattice planes. (Wavefronts of the plane wave are coincident with these lattice planes.) The equations are equivalent to Bragg's law; the Laue equations are vector equations while Bragg's law is in a form that is easier to solve, but these tell the same content. The Laue equations Let be primitive translation vectors (shortly called primitive vectors) of a crystal lattice , where atoms are located at lattice points described by with , , and as any integers. (So indicating each lattice point is an integer linear combination of the primitive vectors.) Let be the wavevector of an incoming (incident) beam or wave toward the crystal lattice , and let be the wavevector of an outgoing (d
https://en.wikipedia.org/wiki/D/U%20ratio
In the design of radio broadcast systems, especially television systems, the desired-to-undesired channel ratio (D/U ratio) is a measure of the strength of the broadcast signal for a particular channel compared with the strength of undesired broadcast signals in the same channel (e.g. from other nearby transmitting stations). See also Signal-to-noise ratio
https://en.wikipedia.org/wiki/Leland%20Clark
Leland C. Clark Jr. (December 4, 1918 – September 25, 2005) was an American biochemist born in Rochester, New York. He is most well known as the inventor of the Clark electrode, a device used for measuring oxygen in blood, water and other liquids. Clark is considered the "father of biosensors", and the modern-day glucose sensor used daily by millions of diabetics is based on his research. He conducted pioneering research on heart-lung machines in the 1940s and '50s and was holder of more than 25 patents. Although he developed a fluorocarbon-based liquid that could be breathed successfully by mice in place of air, his lifelong goal of developing artificial blood remained unfulfilled at the time of his death. He is the inventor of Oxycyte, a third-generation perfluorocarbon (PFC) therapeutic oxygen carrier designed to enhance oxygen delivery to damaged tissues. Professional life Clark received his B.S. degree in chemistry from Antioch College in 1941 and his Ph.D. in biochemistry and physiology from the University of Rochester in 1944. Clark began his professional career as an assistant professor of biochemistry at his alma mater, Antioch College, in Yellow Springs, Ohio. When he left Antioch in 1958, he was head of the department. From 1955 to 1958, he held a simultaneous appointment the University of Cincinnati College of Medicine as a Senior Research Associate in Pediatrics and Surgery. In 1958, Clark moved to Alabama to join the Department of Surgery, University of Alabama Medical College as an associate professor of biochemistry. He later became professor of biochemistry in the same department. In 1962, he invented the first biosensor with Champ Lyons. Clark later became professor of research pediatrics at the Cincinnati Children’s Hospital Research Foundation in 1968 and remained there until he retired in 1991. Afterwards, he helped to found the company Synthetic Blood International, now known as Oxygen Biotherapeutics, Inc., which markets his invention Oxycyt
https://en.wikipedia.org/wiki/Observability%20%28software%29
In distributed systems, observability is the ability to collect data about programs' execution, modules' internal states, and the communication among components. To improve observability, software engineers use a wide range of logging and tracing techniques to gather telemetry information, and tools to analyze and use it. Observability is foundational to site reliability engineering, as it is the first step in triaging a service outage. One of the goals of observability is to minimize the amount of prior knowledge needed to debug an issue. Etymology, terminology and definition The term is borrowed from control theory, where the "observability" of a system measures how well its state can be determined from its outputs. Similarly, software observability measures how well a system's state can be understood from the obtained telemetry (metrics, logs, traces, profiling). The definition of observability varies by vendor: The term is frequently referred to as its numeronym O11y (where 11 stands for the number of letters between the first letter and the last letter of the word). This is similar to other computer science abbreviations such as i18n and L10n and k8s. Observability vs. monitoring Observability and monitoring are sometimes used interchangeably. As tooling, commercial offerings and practices evolved in complexity, "monitoring" was re-branded as observability in order to differentiate new tools from the old. The terms are commonly contrasted in that systems are monitored using predefined sets of telemetry, and monitored systems may be observable. Majors et al. suggest that engineering teams that only have monitoring tools end up relying on expert foreknowledge (seniority), whereas teams that have observability tools rely on exploratory analysis (curiosity). Telemetry types Observability relies on three main types of telemetry data: metrics, logs and traces. Those are often referred to as "pillars of observability". Metrics A metric is a point in tim
https://en.wikipedia.org/wiki/Salisbury%20screen
The Salisbury screen is a way of reducing the reflection of radio waves from a surface. It was one of the first concepts in radar absorbent material, an aspect of "stealth technology", used to prevent enemy radar detection of military vehicles. It was first applied to ship radar cross section (RCS) reduction. The Salisbury screen was invented by American engineer Winfield Salisbury in the early 1940s (see patent filing date). The patent was delayed because of wartime security,. Method of operation Salisbury screens operate on the same principle as optical antireflection coatings used on the surface of camera lenses and glasses to prevent them from reflecting light. The easiest to understand Salisbury screen design consists of three layers: a ground plane which is the metallic surface that needs to be concealed, a lossless dielectric of a precise thickness (a quarter of the wavelength of the radar wave to be absorbed), and a thin glossy screen. When the radar wave strikes the front surface of the dielectric, it is split into two waves. One wave is reflected from the glossy surface screen. The second wave passes into the dielectric layer, is reflected from the metal surface, and passes back out of the dielectric into the air. The extra distance the second wave travels causes it to be 180° out of phase with the first wave by the time it emerges from the dielectric surface When the second wave reaches the surface, the two waves combine and cancel each other out due to the phenomenon of interference. Therefore, there is no wave energy reflected back to the radar receiver. To understand the cancellation of the waves requires an understanding of the concept of interference. When two electromagnetic waves that are coherent and are traveling in the same space interact, they combine to form a single resultant wave. If the two waves are "in phase" so their peaks coincide, they add, and the output intensity is the sum of the two waves' intensities. However, if t
https://en.wikipedia.org/wiki/Label%20propagation%20algorithm
Label propagation is a semi-supervised machine learning algorithm that assigns labels to previously unlabeled data points. At the start of the algorithm, a (generally small) subset of the data points have labels (or classifications). These labels are propagated to the unlabeled points throughout the course of the algorithm. Within complex networks, real networks tend to have community structure. Label propagation is an algorithm for finding communities. In comparison with other algorithms label propagation has advantages in its running time and amount of a priori information needed about the network structure (no parameter is required to be known beforehand). The disadvantage is that it produces no unique solution, but an aggregate of many solutions. Functioning of the algorithm At initial condition, the nodes carry a label that denotes the community they belong to. Membership in a community changes based on the labels that the neighboring nodes possess. This change is subject to the maximum number of labels within one degree of the nodes. Every node is initialized with a unique label, then the labels diffuse through the network. Consequently, densely connected groups reach a common label quickly. When many such dense (consensus) groups are created throughout the network, they continue to expand outwards until it is impossible to do so. The process has 5 steps: 1. Initialize the labels at all nodes in the network. For a given node x, Cx (0) = x. 2. Set t = 1. 3. Arrange the nodes in the network in a random order and set it to X. 4. For each x ∈ X chosen in that specific order, let Cx(t) = f(Cxi1(t), ...,Cxim(t),Cxi(m+1) (t − 1), ...,Cxik (t − 1)). Here returns the label occurring with the highest frequency among neighbours. Select a label at random if there are multiple highest frequency labels. 5. If every node has a label that the maximum number of their neighbours have, then stop the algorithm. Else, set t = t + 1 and go to (3). Multiple community stru
https://en.wikipedia.org/wiki/Alexander%20Zaitsev%20%28astronomer%29
Aleksandr Leonidovich Zaitsev (; 19 May 1945 – 29 November 2021) was a Russian and Soviet radio engineer and astronomer from Fryazino. He worked on radar astronomy devices, near-Earth asteroid radar research, and SETI. Education Zaitsev received his M.Sc. degree in radio engineering from the Moscow Mining University in 1967 and his Ph.D. (1981) and his postdoctoral lecture qualification (1997) in radar astronomy from the Institute of Radio Engineering and Electronics, Russian Academy of Science in Moscow. He was a member of the Space Guard Foundation, the SETI League, and The European Radio Astronomy Club (E.R.A.C.). Career Zaitsev was the chief scientist at the Russian Academy of Science's Institute of Radio Engineering and Electronics. He headed the group transmitting Team Encounter's interstellar messages using the Yevpatoria (Evpatoria) Deep Space Center (EDSC). Zaitsev was also serving as the SETI League's Regional Coordinator for Russia. Zaitsev's career has focused on three main topics: the theory, the design and implementation of radar devices used in the study of Venus, Mars, and Mercury; near-Earth asteroid radar research; and interstellar radio messaging, his later field of research. He retired in 2013. Zaitsev observed the asteroid 4179 Toutatis in December 1992 using the 70-m Yevpatorian Planetary Radar in Crimea (Ukraine), as a sounding signal transmitter, and the 100-m radio telescope in Effelsberg, Germany, as a receiver of the asteroid's radar echo. In June 1995, Zaitsev was responsible for initiating the world's first intercontinental radar astronomy experiment; the radar groups participating in this experiment were led by Steven Ostro at JPL, Zaitsev in Yevpatoria, and Yasuhiro Koyama in Kashima, Japan. Ostro's group transmitted and received using the Goldstone site of the Deep Space Network, while Zaitsev received using the Yevpatoria site and Koyama's group received at Kashima. The target asteroid, 6489 Golevka, was later named for th
https://en.wikipedia.org/wiki/Hendrik%20Poinar
Hendrik Nicholas Poinar (born May 31, 1969 in D.C, United States) is an evolutionary biologist specializing in ancient DNA. Poinar first became known for extracting DNA sequences from ground sloth coprolites. He is currently director of the Ancient DNA Centre at McMaster University in Hamilton, Ontario. Education and academic career The son of noted entomologist George Poinar Jr. and Eva Hecht-Poinar, Poinar received his B.S. and M.S. degrees from California Polytechnic University, San Luis Obispo in 1992 and 1999 respectively before earning a Ph.D. in 1999 from the Ludwig Maximilian University of Munich, after which he was a postdoctoral researcher from 2000 to 2003 at the Max Planck Institute for Evolutionary Anthropology in Leipzig, Germany. In 2003 he was hired as an assistant professor in the anthropology department at McMaster University in Canada. In a joint 2000 paper in Science, Poinar and Dr. Alan Cooper argued that much existing work in human ancient DNA has not been sufficiently rigorous to prevent DNA contamination from modern human sources, and that many reported results for ancient human DNA may therefore be suspect. In 2003, Poinar and others from the Max Planck Institute published genetic sequences isolated from coprolites of the extinct Shasta giant ground sloth, with an estimated age of 10500 years using radiocarbon dates. These were the first genetic sequences retrieved from any extinct ground sloth. In September 2008, Poinar's laboratory published results showing that after a long period of separation in the mammoth populations of Siberia and North America, the Siberian mammoth population had been completely replaced by mammoths of North American origin. In 2014, Poinar and colleagues published the first genomic data from victims of the Plague of Justinian in Bavaria, demonstrating that this plague was caused by a strain of Yersinia pestis now extinct.
https://en.wikipedia.org/wiki/Brachyhypopomus%20bennetti
Brachyhypopomus bennetti is a species of electric knifefish. The species was discovered in the Central Amazon.
https://en.wikipedia.org/wiki/Blind%20equalization
Blind equalization is a digital signal processing technique in which the transmitted signal is inferred (equalized) from the received signal, while making use only of the transmitted signal statistics. Hence, the use of the word blind in the name. Blind equalization is essentially blind deconvolution applied to digital communications. Nonetheless, the emphasis in blind equalization is on online estimation of the equalization filter, which is the inverse of the channel impulse response, rather than the estimation of the channel impulse response itself. This is due to blind deconvolution common mode of usage in digital communications systems, as a means to extract the continuously transmitted signal from the received signal, with the channel impulse response being of secondary intrinsic importance. The estimated equalizer is then convolved with the received signal to yield an estimation of the transmitted signal. Problem statement Noiseless model Assuming a linear time invariant channel with impulse response , the noiseless model relates the received signal to the transmitted signal via The blind equalization problem can now be formulated as follows; Given the received signal , find a filter , called an equalization filter, such that where is an estimation of . The solution to the blind equalization problem is not unique. In fact, it may be determined only up to a signed scale factor and an arbitrary time delay. That is, if are estimates of the transmitted signal and channel impulse response, respectively, then give rise to the same received signal for any real scale factor and integral time delay . In fact, by symmetry, the roles of and are Interchangeable. Noisy model In the noisy model, an additional term, , representing additive noise, is included. The model is therefore Algorithms Many algorithms for the solution of the blind equalization problem have been suggested over the years. However, as one usually has access to only a finite number of s
https://en.wikipedia.org/wiki/High-conductance%20state
The term high-conductance state describes a particular state of neurons in specific states of the brain, such as for example during wakefulness, attentive states, or even during some anesthetized states. In individual neurons, the high-conductance state is formally defined by the fact that the total synaptic conductance received by the neuron is larger than its natural resting (or leak) conductance, so in a sense the neuron is "driven" by its inputs rather than being dominated by its intrinsic activity. High-conductance states have been well characterized experimentally, but they also have motivated numerous theoretical studies, in particular in relation to the considerable amount of "noise" present in such states. It is believed that this "synaptic noise" has determinant properties on neuronal processing, and even may confer several computational advantages to neurons (see details in the article High-Conductance State in Scholarpedia). The term high-conductance state is also used to describe specific states of single ion channels. In this case, the high-conductance state corresponds to an open state of the channel which is associated with a particularly high conductance compared to other states.
https://en.wikipedia.org/wiki/Nilpotent%20matrix
In linear algebra, a nilpotent matrix is a square matrix N such that for some positive integer . The smallest such is called the index of , sometimes the degree of . More generally, a nilpotent transformation is a linear transformation of a vector space such that for some positive integer (and thus, for all ). Both of these concepts are special cases of a more general concept of nilpotence that applies to elements of rings. Examples Example 1 The matrix is nilpotent with index 2, since . Example 2 More generally, any -dimensional triangular matrix with zeros along the main diagonal is nilpotent, with index . For example, the matrix is nilpotent, with The index of is therefore 4. Example 3 Although the examples above have a large number of zero entries, a typical nilpotent matrix does not. For example, although the matrix has no zero entries. Example 4 Additionally, any matrices of the form such as or square to zero. Example 5 Perhaps some of the most striking examples of nilpotent matrices are square matrices of the form: The first few of which are: These matrices are nilpotent but there are no zero entries in any powers of them less than the index. Example 6 Consider the linear space of polynomials of a bounded degree. The derivative operator is a linear map. We know that applying the derivative to a polynomial decreases its degree by one, so when applying it iteratively, we will eventually obtain zero. Therefore, on such a space, the derivative is representable by a nilpotent matrix. Characterization For an square matrix with real (or complex) entries, the following are equivalent: is nilpotent. The characteristic polynomial for is . The minimal polynomial for is for some positive integer . The only complex eigenvalue for is 0. The last theorem holds true for matrices over any field of characteristic 0 or sufficiently large characteristic. (cf. Newton's identities) This theorem has several consequences, including: The i
https://en.wikipedia.org/wiki/Optica%20Optics%20Software
Optica is an optical design program used for the design and analysis of both imaging and illumination systems. It works by ray tracing the propagation of rays through an optical system. It performs polarization ray-tracing, non-sequential ray-tracing, energy calculations, and optimization of optical systems in three-dimensional space. It also performs symbolic modeling of optical systems, diffraction, interference, wave-front, and Gaussian beam propagation calculations. In addition to conducting simulations of optical designs, Optica is used by scientists to create illustrations of the simulated results in publications. Some examples of Optica being used in simulations and illustrations include holography, x-ray optics, spectrometers, Cerenkov radiation, microwave optics, nonlinear optics, scattering, camera design, extreme ultraviolet lithography simulations, telescope optics, laser design, ultrashort pulse lasers, eye models, solar concentrators and Ring Imaging CHerenkov (RICH) particle detectors. History Optica was originally developed by Donald Barnhart of Urbana, Illinois, USA, and has been in continual development since 1994. Wolfram Research first sold the original version as a Mathematica application. From 2005 to 2009, Optica Software was sold by iCyt Mission Technology Inc, Champaign, Illinois (renamed Sony Biotechnology Inc in 2010). At iCyt, Optica2 was renamed as Rayica, and Wavica and LensLab were also developed. Later Rayica-Wavica was combined and named back to Optica3. Since 2009, Optica Software has been a subsidiary of Barnhart Optical Research LLC.
https://en.wikipedia.org/wiki/Chemerin%20peptide
Chemerin peptides are short peptides (on the order of 9 amino acids) that are produced from the carboxyl terminus of the chemokine chemerin. They display the same activities as chemerin, although at higher efficacy and potency. A particular synthetic chemerin-derived peptide, termed C15, was developed at Oxford University. It showed anti-inflammatory activities. Intraperitoneal administration of C15 (0.32 ng/kg) to mice before zymosan challenge conferred significant protection against zymosan-induced peritonitis, suppressing neutrophil (63%) and monocyte (62%) recruitment with a concomitant reduction in proinflammatory mediator expression. C15 was found to promote phagocytosis and efferocytosis in peritoneal macrophages at picomolar concentrations. C15 enhanced macrophage clearance of microbial particles and apoptotic cells by factor of 360% in vitro
https://en.wikipedia.org/wiki/Xanthene
Xanthene (9H-xanthene, 10H-9-oxaanthracene) is the organic compound with the formula CH2[C6H4]2O. It is a yellow solid that is soluble in common organic solvents. Xanthene itself is an obscure compound, but many of its derivatives are useful dyes. Xanthene dyes Dyes that contain a xanthene core include fluorescein, eosins, and rhodamines. Xanthene dyes tend to be fluorescent, yellow to pink to bluish red, brilliant dyes. Many xanthene dyes can be prepared by condensation of derivates of phthalic anhydride with derivates of resorcinol or 3-aminophenol. Further reading See also Xanthone Xanthydrol
https://en.wikipedia.org/wiki/White%20jersey
Numerous cycling stage races award a white jersey to signify the current leader and overall winner of a certain competition, or to signify the best young rider in the race. The most prominent of these is the Tour de France, where the jersey is known as the maillot blanc and is awarded to the best-placed rider age under 26. The use of the white jersey to recognize the best young rider in a race is its most common use, though some tours award a white jersey for a different classification. Other stage races, besides the Tour de France, that also award a white jersey for the best young rider include: Deutschland Tour Giro d'Italia (known as the maglia bianca, and the cut-off is 25 years instead of 26) Paris–Nice Tour of California (where the cut-off is age 23 instead of 26) Tour of Ireland Tour de France Giro d'Italia Other white jerseys Some stage races award a white jersey for a different classification than youth. The foremost of these is probably the Vuelta a España, where it recognizes the leader in the Combination classification. In this classification, ranks in the General, Points, and Mountains classifications are added, and whoever has the lowest cumulative total is awarded the white jersey. It is a relatively new award, having existed only since the 2003 edition of the race. In 1941, the Vuelta a España white jersey was for the leader and overall winner of the General classification. The Tour of the Basque Country awards a white jersey to the leader and overall winner of the Points classification. The Volta a Catalunya awards a white jersey with green stripes to the leader and overall winner of the General classification. The 2008 Tour de Suisse awarded a white jersey to a somewhat unusual competition, the Sprints classification, which awards placings not on stage finishes but strictly in intermediate sprints. The Tour Down Under awards a white jersey, with green piping and side panels, to the leader and ultimately the winner of the King of the Mountain
https://en.wikipedia.org/wiki/Relvar
In relational databases, relvar is a term introduced by C. J. Date and Hugh Darwen as an abbreviation for relation variable in their 1995 paper The Third Manifesto, to avoid the confusion sometimes arising from the use of the term relation, by the inventor of the relational model, E. F. Codd, for a variable to which a relation is assigned as well as for the relation itself. The term is used in Date's well-known database textbook An Introduction to Database Systems and in various other books authored or coauthored by him. Some database textbooks use the term relation for both the variable and the data it contains. Similarly, texts on SQL tend to use the term table for both purposes, though the qualified term base table is used in the standard for the variable. A closely related term often used in academic texts is relation schema, this being a set of attributes paired with a set of constraints, together defining a set of relations for the purpose of some discussion (typically, database normalization). Constraints that mention just one relvar are termed relvar constraints, so relation schema can be regarded as a single term encompassing a relvar and its relvar constraints.
https://en.wikipedia.org/wiki/History%20of%20mass%20spectrometry
The history of mass spectrometry has its roots in physical and chemical studies regarding the nature of matter. The study of gas discharges in the mid 19th century led to the discovery of anode and cathode rays, which turned out to be positive ions and electrons. Improved capabilities in the separation of these positive ions enabled the discovery of stable isotopes of the elements. The first such discovery was with the element neon, which was shown by mass spectrometry to have at least two stable isotopes: 20Ne (neon with 10 protons and 10 neutrons) and 22Ne (neon with 10 protons and 12 neutrons). Mass spectrometers were used in the Manhattan Project for the separation of isotopes of uranium necessary to create the atomic bomb. Prout's Hypothesis Prout's hypothesis was an early 19th-century attempt to explain the properties of the chemical elements using the internal structure of the atom. In 1815, the English chemist William Prout observed that the atomic weights that had been measured were integer multiples of the atomic weight of hydrogen. Prout's hypothesis remained influential in chemistry throughout the 1820s. However, more careful measurements of the atomic weights, such as those compiled by Jöns Jakob Berzelius in 1828 or Edward Turner in 1832, appeared to disprove it. In particular the atomic weight of chlorine, which is 35.45 times that of hydrogen, could not at the time be explained in terms of Prout's hypothesis. It would take the better part of a century for this problem to be resolved. Canal rays In the mid-nineteenth century, Julius Plücker investigated the light emitted in discharge tubes and the influence of magnetic fields on the glow. Later, in 1869, Johann Wilhelm Hittorf studied discharge tubes with energy rays extending from a negative electrode, the cathode. These rays produced a fluorescence when they hit a tube's glass walls, and when interrupted by a solid object they cast a shadow. Canal rays, also called anode rays, were observed by
https://en.wikipedia.org/wiki/Coraliomargarita%20akajimensis
Coraliomargarita akajimensis is a Gram-negative, obligately aerobic, non-spore-forming and non-motile bacterium from the genus of Coraliomargarita which has been isolated from seawater from Japan.
https://en.wikipedia.org/wiki/Asa%20Wright%20Nature%20Centre
The Asa Wright Nature Centre and Lodge is a nature resort and scientific research station in the Arima Valley of the Northern Range in Trinidad and Tobago. The centre is one of the top birdwatching spots in the Caribbean; a total of 256 species of birds have been recorded there. The centre is owned by a non-profit trust. The nature centre is on and includes a main estate house with inn and restaurant serving dishes such as callaloo soup with ingredients from an on-site organic garden. Non-adjacent properties have also been added to the centre's land holdings. Description The major properties are the Spring Hill Estate and the adjacent William Beebe Tropical Research Station (also known as Simla), which was established by the famous naturalist and explorer William Beebe as a tropical research station for the New York Zoological Society. Beebe bought the 'Verdant Vale' estate in 1949 and named it after Simla in India, which he had visited in 1910. Beebe stayed at Springhill in 1949 with the Wrights, while water and electricity connections were made to Simla, four miles down the road. His bedroom is now the Springhill manager's office. The owners of the Spring Hill Estate, Newcombe and Asa Wright, hosted the noted ornithologists David Snow and Barbara Snow in the 1950s. They made detailed studies of the oilbirds and the Bearded bellbird. Starting in 1965, Richard ffrench and Don Eckelberry collaborated in eventually publishing A Guide to the Birds of Trinidad and Tobago. The Wrights' home became internationally renowned for its easy access to wildlife, especially the oilbird (Steatornis caripensis) colonies in the nearby Dunston Cave. The Asa Wright Nature Center was established on 22 Dec. 1967, with the Royal Bank Trust Company (Trinidad) Limited appointed as Trustee. The Trust was meant to protect Springhill "in perpetuity", developing its agriculture in an "ecologically sound" way while facilitating its "scientific and educational" potential. Asa was all
https://en.wikipedia.org/wiki/Gigabyte%20Technology
Gigabyte Technology (branded as GIGABYTE or sometimes GIGA-BYTE; formally GIGA-BYTE Technology Co., Ltd.) is a Taiwanese manufacturer and distributor of computer hardware. Gigabyte's principal business is motherboards. It shipped 4.8 million motherboards in the first quarter of 2015, which allowed it to become the leading motherboard vendor. Gigabyte also manufactures custom graphics cards and laptop computers (including thin and light laptops under its Aero sub-brand). In 2010, Gigabyte was ranked 17th in "Taiwan's Top 20 Global Brands" by the Taiwan External Trade Development Council. The company is publicly held and traded on the Taiwan Stock Exchange, stock ID number . History Gigabyte Technology was established in 1986 by Pei-Cheng Yeh. One of Gigabyte's key advertised features on its motherboards is its "Ultra Durable" construction, advertised with "all solid capacitors". On 8 August 2006 Gigabyte announced a joint venture with Asus. Gigabyte developed the world's first software-controlled power supply in July 2007. An innovative method to charge the iPad and iPhone on the computer was introduced by Gigabyte in April 2010. Gigabyte launched the world's first Z68 motherboard on 31 May 2011, with an on-board mSATA connection for Intel SSD and Smart Response Technology. On 2 April 2012, Gigabyte released the world's first motherboard with 60A ICs from International Rectifier. In 2023, researchers at firmware-focused cybersecurity company Eclypsium said 271 models of Gigabyte motherboards are affected by backdoor vulnerabilities. Whenever a computer with the affected Gigabyte motherboard restarts, code within the motherboard's firmware initiates an updater program that downloads and executes another piece of software. Gigabyte has said it plans to fix the issues. Products Gigabyte designs and manufactures motherboards for both AMD and Intel platforms, and also produces graphics cards and notebooks in partnership with AMD and Nvidia, including Nvidia's Tur
https://en.wikipedia.org/wiki/Dietary%20biology%20of%20the%20tawny%20owl
The tawny owl (Strix aluco) is an opportunistic and generalized predator. Peak hunting activity tends to occur largely between dusk to midnight, with owls often following an erratic hunting pattern, perhaps to sites where previous hunts were successful. When feeding young, hunting may need to be prolonged into daylight in the early morning. Based on hand-reared young owls that re-released into the wild, hunting behaviour is quite innate rather than learned. Normally this owl hunts from a perch. Perching bouts usually last from about 8 to 14 minutes depending largely on habitat. Tawny owl's hunting from a perch or pole can recall a buzzard and the two take similar prey sizes as well. However, high initial speed and maneuvering among trees and bushes with great dexterity may allow it to surprise relatively large prey, more like a goshawk. The tawny owl is capable of lifting and carrying off in flight individual prey weighing up to at least . Their middle talon, the most enlarged claw on owls, measures an average of . While not as large as those of the Ural owl, the talons are extremely sharp, stout and quite decurved. The claws are considered to be visibly more overdeveloped than those of other European mid-sized owls and the footspan including the claws is fairly larger as well, at an average of about . The hunting owl often extends its wings to balance and control prey upon impact. Alternatively, this species may hunt from flight. This occurs from over the ground, often over open habitats such as bushes, marsh or grassland, forming a quartering or zigzag pattern over the opening. During these flights they cover about before changing direction. Hunting from flight was surprisingly prevalent in a Swedish study of two radio-tagged birds, with 34% of study time spent hunting from flight while 40% of the study time was spent on hunting from a perch. In a similar study in England, less than 1% of time was spent hunting from flight. In a more deliberate variation of hunt
https://en.wikipedia.org/wiki/Srp%20receptor%20alpha%20subunit
SRP receptor alpha subunit is a protein that in humans is encoded by the SRPRA gene. Function The gene encodes a subunit of the endoplasmic reticulum signal recognition particle receptor that, in conjunction with the signal recognition particle, is involved in the targeting and translocation of signal sequence tagged secretory and membrane proteins across the endoplasmic reticulum. Alternative splicing results in multiple transcript variants.
https://en.wikipedia.org/wiki/Vijai%20Shukla
Dr. Vijai Shukla (born 23 March 1948) is an Indian-Danish food scientist, researcher and professor in lipidology, and a central figure in the study of essential fatty acids. He is also the president of the International Food Science Center based in Denmark, Fellow of the American Oil Chemists' Society (AOCS) and adjunct professor to the Department of Food Science and Nutrition at University of Minnesota. Career Shukla gained his master's degree in Organic Chemistry in 1969 from the University of Allahabad and his Ph.D. in 1973 from the same university. Subsequently, he received a post-doctoral UNESCO award and did work first at the Charles University in Prague and then in 1975 at the Institute of Neurochemistry in Experimental Life Sciences in Copenhagen, and at the Federal Centre for Lipid Research in Münster, Germany. In 1979 he was appointed R&D manager by then Aarhus Oliefabrik (now AarhusKarlshamn), a leading company in the manufacture of vegetable oils and specialty fats. In 1990 he joined Karlshamn as Research Director, and started his own company in 1991, IFSC, of which he is still President. Shukla's research includes both the field of physical phenomena and the mechanisms of autoxidation, isolation of lipids, spectral phenomena related to lipids (e.g. pulsed Nuclear magnetic resonance), modern analytical methodology, and involvement of essential fatty acids in health and disease such as multiple sclerosis and Batten Disease where he has made significant scientific contributions. Shukla has authored close to 120 peer-reviewed papers, reviews and book chapters, and 15 books. He was the Associate Editor for the peer-reviewed journal INFORM from 1989–97 and remains Associate Editor of the Journal of the American Oil Chemists Society and of Lipid Technology. In 1996 he received the American Oil Chemists' Society (AOCS) Herbert Dutton Award for "significant contributions to analytical methodologies" and was given the prestigious Stephen S. Chang Award in 200
https://en.wikipedia.org/wiki/Membrane%20potential
Membrane potential (also transmembrane potential or membrane voltage) is the difference in electric potential between the interior and the exterior of a biological cell. That is, there is a difference in the energy required for electric charges to move from the internal to exterior cellular environments and vice versa, as long as there is no acquisition of kinetic energy or the production of radiation. The concentration gradients of the charges directly determine this energy requirement. For the exterior of the cell, typical values of membrane potential, normally given in units of milli volts and denoted as mV, range from –80 mV to –40 mV. All animal cells are surrounded by a membrane composed of a lipid bilayer with proteins embedded in it. The membrane serves as both an insulator and a diffusion barrier to the movement of ions. Transmembrane proteins, also known as ion transporter or ion pump proteins, actively push ions across the membrane and establish concentration gradients across the membrane, and ion channels allow ions to move across the membrane down those concentration gradients. Ion pumps and ion channels are electrically equivalent to a set of batteries and resistors inserted in the membrane, and therefore create a voltage between the two sides of the membrane. Almost all plasma membranes have an electrical potential across them, with the inside usually negative with respect to the outside. The membrane potential has two basic functions. First, it allows a cell to function as a battery, providing power to operate a variety of "molecular devices" embedded in the membrane. Second, in electrically excitable cells such as neurons and muscle cells, it is used for transmitting signals between different parts of a cell. Signals are generated by opening or closing of ion channels at one point in the membrane, producing a local change in the membrane potential. This change in the electric field can be quickly sensed by either adjacent or more distant ion chann
https://en.wikipedia.org/wiki/Pitteway%20triangulation
In computational geometry, a Pitteway triangulation is a point set triangulation in which the nearest neighbor of any point p within the triangulation is one of the vertices of the triangle containing p. Alternatively, it is a Delaunay triangulation in which each internal edge crosses its dual Voronoi diagram edge. Pitteway triangulations are named after Michael Pitteway, who studied them in 1973. Not every point set supports a Pitteway triangulation. When such a triangulation exists it is a special case of the Delaunay triangulation, and consists of the union of the Gabriel graph and convex hull. History The concept of a Pitteway triangulation was introduced by . See also , who writes "An optimal partition is one in which, for any point within any triangle, that point lies at least as close to one of the vertices of that triangle as to any other data point." The name "Pitteway triangulation" was given by . Counterexamples points out that not every point set supports a Pitteway triangulation. For instance, any triangulation of a regular pentagon includes a central isosceles triangle such that a point p near the midpoint of one of the triangle sides has its nearest neighbor outside the triangle. Relation to other geometric graphs When a Pitteway triangulation exists, the midpoint of each edge interior to the triangulation must have the two edge endpoints as its nearest neighbors, for any other neighbor would violate the Pitteway property for nearby points in one of the two adjacent triangles. Thus, a circle having that edge as diameter must be empty of vertices, so the Pitteway triangulation consists of the Gabriel graph together with the convex hull of the point set. Conversely, when the Gabriel graph and convex hull together form a triangulation, it is a Pitteway triangulation. Since all Gabriel graph and convex hull edges are part of the Delaunay triangulation, a Pitteway triangulation, when it exists, is unique for points in general position and coincides
https://en.wikipedia.org/wiki/CD74
HLA class II histocompatibility antigen gamma chain also known as HLA-DR antigens-associated invariant chain or CD74 (Cluster of Differentiation 74), is a protein that in humans is encoded by the CD74 gene. The invariant chain (Abbreviated Ii) is a polypeptide which plays a critical role in antigen presentation. It is involved in the formation and transport of MHC class II peptide complexes for the generation of CD4+ T cell responses. The cell surface form of the invariant chain is known as CD74. CD74 is a cell surface receptor for the cytokine macrophage migration inhibitory factor (MIF). Function The nascent MHC class II protein in the rough endoplasmic reticulum (RER) binds a segment of the invariant chain (Ii; a trimer) in order to shape the peptide-binding groove and prevent the formation of a closed conformation. The invariant chain also facilitates the export of MHC class II from the RER in a vesicle. The signal for endosomal targeting resides in the cytoplasmic tail of the invariant chain. This fuses with a late endosome containing the endocytosed antigen proteins (from the exogenous pathway). Binding to Ii ensures that no antigen peptides from the endogenous pathway meant for MHC class I molecules accidentally bind to the groove of MHC class II molecules. The Ii is then cleaved by cathepsin S (cathepsin L in cortical thymic epithelial cells), leaving only a small fragment called CLIP remaining bound to the groove of MHC class II molecules. The rest of the Ii is degraded. CLIP blocks peptide-binding until HLA-DM interacts with MHC II, releasing CLIP and allowing other peptides to bind. In some cases, CLIP dissociates without any further molecular interactions, but in other cases the binding to the MHC is more stable. The stable MHC class II + antigen complex is then presented on the cell surface. Without CLIP, MHC class II aggregates disassemble and/or denature in the endosomes, and proper antigen presentation is impaired. Clinical significance Vaccine
https://en.wikipedia.org/wiki/Pierre%20Auger%20%28biologist%29
Pierre Auger is a French bio-mathematician born on March 8, 1953, in Neuilly-sur-Seine. He is a member of the French Academy of sciences and Director of Exceptional Class Research at the Research Institute for Development. Pierre Auger's research field concerns the mathematical modelling of biological systems. Course Student at the University of Paris 6 (1973-1977), he obtained his PhD in Nuclear Physics at the Institut de Physique Nucléaire d'Orsay in 1979 and his Doctorate in Physics. "Mathematical models of hierarchical systems" at the University of Angers in 1982. He was admitted to the CAPES in Physical Sciences in 1983, and became a certified Professor of Physics and Chemistry at the Lycée Technique Pasteur in Hénin-Beaumont, Pas de Calais from 1984 to 1986 and then a certified Professor in Metallurgy at the Lycée Diderot in Paris from 1986 to 1987. He continued his career as a lecturer at the Biophysics Laboratory of the Faculty of Pharmacy of Dijon from 1987 to 1990, and Professor of Universities at the Ecology Laboratory of the University of Burgundy in Dijon from 1990 to 1992 and then Professor at the Claude Bernard University in Lyon from 1993 to 2004. He has been a research director at IRD since 2004. He was elected correspondent (1999) and then member of the French Academy of sciences (2003) in the Integrative Biology section. Scientific work Pierre Auger is a specialist in mathematical modelling in ecology and environmental sciences. He has contributed to the development of "variable aggregation methods", the aim of which is to build from a detailed or "complete" model, a reduced model governing only a few global variables in the long term. In the 1990s, with the mathematicians Robert Roussarie and Jean-Christophe Poggiale of the University of Dijon, he formalized the method as part of the centre variety theorem (2010-2014). Pierre Auger has been particularly interested with his collaborators in the applications of variable aggregation methods to
https://en.wikipedia.org/wiki/Buddy%20%28software%29
Buddy (also known as Buddy.Works) is a web-based and self-hosted continuous integration and delivery software for Git developers that can be used to build, test, and deploy web sites and applications with code from GitHub, Bitbucket, and GitLab. It employs Docker containers with pre-installed languages and frameworks for builds, alongside DevOps, monitoring and notification actions. History Buddy launched as a downloadable VM in May 2015 under the name Meat!. The service was initially free but employed a proprietary license which stirred some concern in the web development community. Meat! was rebranded to Buddy in November 2015 and released as a cloud-only service. The on-premises version, nicknamed Buddy GO, was released in September 2016. Switching from VM to Docker allowed installation on any Linux-based server, including Amazon EC2, DigitalOcean, and Microsoft Azure. Shortly after, the company launched Guides, a dedicated website section with use cases and workflow automation strategies, later reproduced to Medium, a popular blogging platform. On September 21, 2016, the service was featured on Product Hunt. Configuration Configuration is performed by arranging predefined actions into sequences called pipelines. Pipelines can be triggered automatically on push to branch, manually, or recurrently. Actions include Docker-based builds, deployment to FTP/SFTP and IaaS services, delivery to version control, SSH scripts, website monitoring and conditional notifications. Contrary to other CI tools like Jenkins or Travis CI, Buddy does not use YAML files to describe the process, although the company stated support for .yml files is currently in works. Version control Besides the support for third-party hosting services, Buddy features a native code hosting solution with the most popular Git commands (git log, git show, git blame, git diff) reproduced into the GUI. Other features include a cloud editor with blame tool and syntax highlight, push permissions, merge r
https://en.wikipedia.org/wiki/Herbrand%27s%20theorem
Herbrand's theorem is a fundamental result of mathematical logic obtained by Jacques Herbrand (1930). It essentially allows a certain kind of reduction of first-order logic to propositional logic. Herbrand's theorem is the logical foundation for most automatic theorem provers. Although Herbrand originally proved his theorem for arbitrary formulas of first-order logic, the simpler version shown here, restricted to formulas in prenex form containing only existential quantifiers, became more popular. Statement Let be a formula of first-order logic with quantifier-free, though it may contain additional free variables. This version of Herbrand's theorem states that the above formula is valid if and only if there exists a finite sequence of terms , possibly in an expansion of the language, with and , such that is valid. If it is valid, it is called a Herbrand disjunction for Informally: a formula in prenex form containing only existential quantifiers is provable (valid) in first-order logic if and only if a disjunction composed of substitution instances of the quantifier-free subformula of is a tautology (propositionally derivable). The restriction to formulas in prenex form containing only existential quantifiers does not limit the generality of the theorem, because formulas can be converted to prenex form and their universal quantifiers can be removed by Herbrandization. Conversion to prenex form can be avoided, if structural Herbrandization is performed. Herbrandization can be avoided by imposing additional restrictions on the variable dependencies allowed in the Herbrand disjunction. Proof sketch A proof of the non-trivial direction of the theorem can be constructed according to the following steps: If the formula is valid, then by completeness of cut-free sequent calculus, which follows from Gentzen's cut-elimination theorem, there is a cut-free proof of . Starting from leaves and working downwards, remove the inferences that introduce existent
https://en.wikipedia.org/wiki/T%C3%A1chira%20emerald
The Táchira emerald (Amazilia distans) is a hummingbird described in 1956 by Alexander Wetmore and William Phelps as a new species from a specimen from Venezuela. It is now considered an intergeneric hybrid between the glittering-throated emerald (Amazilia fimbriata) and the white-chinned sapphire (Hylocharis cyanus).
https://en.wikipedia.org/wiki/Photoactivatable%20probes
Photoactivatable probes, or caged probes, are cellular players (proteins, nucleic acids, small molecules) that can be triggered by a flash of light. They are used in biological research to study processes in cells. The basic principle is to bring a photoactivatable agent (e.g. a small molecule modified with a light-responsive group: proteins tagged with an artificial photoreceptor protein) to cells, tissues or even living animals and specifically control its activity by illumination. Light is a well-suited external trigger for these types of experiments since it is non-invasive and does not influence normal cellular processes (though care has to be taken when using light in the ultra-violet part of the spectrum to avoid DNA damage. Furthermore, light offers high spatial and temporal control. Usually, the activation stimulus comes from a laser or a UV lamp and can be incorporated into the same microscope used for monitoring of the effect. All these advantages have led to the development of a wide variety of different photoactivatable probes. Even though the light-induced activation step is usually irreversible, reversible changes can be induced in a number of photoswitches. History The first reported use of photoprotected analogues for biological studies was the synthesis and application of caged ATP by Joseph F. Hoffman in 1978 in his study of Na:K pumps. As of 2013, ATP is still the most commonly used caged compound. Hoffman was also the one to coin the term 'caged' for this type of modified molecules. This nomenclature persisted, despite it being scientifically a misnomer, since it suggests the idea of the molecule being in a physical cage (like in a Fullerene). However, scientists have tried to introduce the newer, more accurate term 'photoactivatable probes'. Both nomenclatures are currently in use. Major discoveries were made in the following years with caged neurotransmitters, such as glutamate, which is used to map functional neuronal circuits in mammal
https://en.wikipedia.org/wiki/A%20Mathematician%27s%20Lament
A Mathematician's Lament, often referred to informally as Lockhart's Lament, is a short book on mathematics education by Paul Lockhart, originally a research mathematician at Brown University and U.C. Santa Cruz, and subsequently a math teacher at Saint Ann's School in Brooklyn, New York City for many years. This strongly worded opinion piece is organized into two parts. The first part, "Lamentation", criticizes the way mathematics is typically taught in American schools and argues for an aesthetic, intuitive, and problem-oriented approach to teaching. The second part, "Exultation", gives specific examples of how to teach mathematics as an art. Background This book was developed from a 25-page essay that was written in 2002, originally circulated in typewritten manuscript copies, and subsequently published by Keith Devlin on his online column for the Mathematical Association of America's webzine MAA Online. Quotes "The first thing to understand is that mathematics is an art. The difference between math and the other arts, such as music and painting, is that our culture does not recognize it as such." "Other math courses may hide the beautiful bird, or put it in a cage, but in geometry class, it is openly and cruelly tortured."
https://en.wikipedia.org/wiki/Preimplantation%20genetic%20haplotyping
Preimplantation genetic haplotyping (PGH) is a clinical method of preimplantation genetic diagnosis (PGD) used to determine the presence of single gene disorders in offspring. PGH provides a more feasible method of gene location than whole-genome association experiments, which are expensive and time-consuming. PGH differs from common PGD methods such as fluorescence in situ hybridization (FISH) and polymerase chain reaction (PCR) for two primary reasons. First, rather than focusing on the genetic makeup of an embryo PGH compares the genome of affected and unaffected members of previous generations. This examination of generational variation then allows for a haplotype of genetic markers statistically associated with the target disease to be identified, rather than searching merely for a mutation. PGH is often used to reinforce other methods of genetic testing, and is considered more accurate than certain more common PGD methods because it has been found to reduce risk of misdiagnoses. Studies have found that misdiagnoses due to allele dropout (ADO), one of the most common causes of interpretation error, can be almost eliminated through use of PGH. Further, in the case of mutation due to translocation, PGH is able to detect chromosome abnormality to its full extent by differentiating between embryos carrying balanced forms of a translocation versus those carrying the homologous normal chromosomes. This is an advantage because PGD methods such as FISH are able to reveal whether an embryo will express the phenotypic difference, but not whether an embryo may be a carrier. In 2015, PGH was used in conjunction with a whole-genome amplification (WGA) process to not only diagnose disease but also distinguish meiotic segregation errors from mitotic ones. Studies are being continually performed in an attempt to utilize and improve PGD methods since their initial invention. It has become increasingly popular because it grants individuals the option of detecting embryo abnor
https://en.wikipedia.org/wiki/T-1152
T-1152 is a quaternary carbamate anticholinesterase. It is synthesized by reaction of m-dimethylaminophenol with methyl isocyanate, followed by quaternization with methyl iodide. Since T-1152 is toxic by ingestion, it was patented as a rodenticide in 1932. The chloride and methylsulfate salt of T-1152 is T-1690 (TL-1226) and AR-13, respectively. See also Neostigmine T-1123 T-1194 TL-599 TL-1238
https://en.wikipedia.org/wiki/Tepidisphaeraceae
Tepidisphaeraceae is a family of bacteria. See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/Ohio%20Web%20Library
The Ohio Web Library is a large collection of over 280 electronic information resources, or online databases, provided by Libraries Connect Ohio (LCO), which is composed of four major Ohio library networks — OPLIN, OhioLINK, INFOhio, and the State Library of Ohio. Within these licensed databases are almost 31,000 individual electronic serial titles (i.e., publications with ISSN), and the databases are accessed through a federated search tool or meta search engine with a simple interface. LCO can cooperatively purchase access to these statewide resources at a cost-effective rate whereas most individual libraries would not be able to afford them on their own. These subscription-based research databases are available for free to all Ohio residents regardless of their location, age, education, or economic status and include online publications and research resources, such as scholarly journals, popular magazines, trade publications, newspapers (Ohio and nationwide), encyclopedias, dictionaries, and educational/training tutorials. Funding and Governance The Ohio Web Library is funded by a federal IMLS Library Services and Technology Act (LSTA) grant through the State Library of Ohio, which provides about $1.5 million per year. The three LCO library networks — OPLIN, OhioLINK, and INFOhio — provide an additional $3 million in state funds each year from their budgets. The federal LSTA grant that helps fund the Ohio Web Library is a temporary funding measure that expires every June 30. No permanent funding is in place to sustain these resources or to expand them. The Ohio Web Library resources are purchased by Ohio libraries and library organizations for exclusive use by Ohio residents. This system uses IP address recognition to verify whether someone is located in Ohio, but these resources can also be accessed by providing an Ohio public library card number for user authentication. Collaborative Library Organizations The Ohio Web Library is a collaborative effort of
https://en.wikipedia.org/wiki/Andy%20the%20Clown
Andrew Rozdilsky Jr. (December 6, 1917 – September 21, 1995) was an American clown. As Andy the Clown, he was well-known for performing at Chicago White Sox games at the original Comiskey Park from 1960 to 1990. Early life Andrew Rozdilsky Jr., the youngest of five brothers and one sister, was born in Chicago to a family of Polish descent. Rozdilsky grew up on the Southwest Side of Chicago and began clowning at ten years old to amuse his family. As a teenager, Rozdilsky worked as a hot dog vendor at Comiskey Park. In the early 1940s, he worked as a drill press operator before being drafted into the U.S. Army in March 1942. He completed his military service in June 1943 when he was discharged due to a pre-existing medical condition. Upon returning to civilian life, Rozdilsky worked as a hearse driver before being hired as a research clerk at International Harvester. Rozdilsky also began working on the side as a clown in the 1940s, doing his routine at parties and luncheons and visiting numerous hospital patients. Career Although a lifelong White Sox fan and a clown since childhood, Rozdilsky only made his first appearance in costume at Comiskey Park in 1960. Rozdilsky dressed as a clown one day to amuse his friends from a local bowling league, and they encouraged him to attend a White Sox game with them in the outfit. Performing as Andy the Clown, Rozdilsky became popular with the crowd. By coincidence, Rozdilsky won White Sox season tickets for 1961 in a Knights of Columbus raffle, and soon began to attend almost every game as Andy the Clown. In the early part of his career, Rozdilsky had no official status and was not paid for his performances at White Sox games; he was only able to appear consistently because he happened to have won season tickets. Not restricting himself to baseball, Rozdilsky also occasionally clowned at other professional sporting events, including Chicago Bears games. In addition to his prolific appearances in Chicago, Rozdilsky was hired b
https://en.wikipedia.org/wiki/Multivariate%20optical%20computing
Multivariate optical computing, also known as molecular factor computing, is an approach to the development of compressed sensing spectroscopic instruments, particularly for industrial applications such as process analytical support. "Conventional" spectroscopic methods often employ multivariate and chemometric methods, such as multivariate calibration, pattern recognition, and classification, to extract analytical information (including concentration) from data collected at many different wavelengths. Multivariate optical computing uses an optical computer to analyze the data as it is collected. The goal of this approach is to produce instruments which are simple and rugged, yet retain the benefits of multivariate techniques for the accuracy and precision of the result. An instrument which implements this approach may be described as a multivariate optical computer. Since it describes an approach, rather than any specific wavelength range, multivariate optical computers may be built using a variety of different instruments (including Fourier Transform Infrared (FTIR) and Raman). The "software" in multivariate optical computing is encoded directly into an optical element spectral calculation engine such as an interference filter based multivariate optical element (MOE), holographic grating, liquid crystal tunable filter, spatial light modulator (SLM), or digital micromirror device (DMD) and is specific to the particular application. The optical pattern for the spectral calculation engine is designed for the specific purpose of measuring the magnitude of that multi-wavelength pattern in the spectrum of a sample, without actually measuring a spectrum. Multivariate optical computing allows instruments to be made with the mathematics of pattern recognition designed directly into an optical computer, which extracts information from light without recording a spectrum. This makes it possible to achieve the speed, dependability, and ruggedness necessary for real ti
https://en.wikipedia.org/wiki/Arched-hill%20symbol
The Arched-hill symbol is a symbol on ancient Coinage of India. There are some variations to the number of the hills depicted, or the symbol surmounting the hill, such as a crescent or a star. It is thought that the three-arched hill symbol was initiated during the Maurya Empire (3rd–2nd century BCE). Later, in coins from Taxila dated from 220 BCE, the three-arched symbol appears regularly, and from 185 BCE is regularly associated with the animal figures of the elephant and the lion. In contrast, the Nandipada is generally associated with the zebu bull. On coins of the Shunga period, the three-arched hill can appear among a multitude of other symbols, such as the Nandipada, the tree-in-railing, the elephant, or the empty cross. The symbol is generally considered a representation of a Buddhist Chaitya. It has also been argued that it was the imperial symbol of the Mauryas. The symbol however, appears in many post-Mauryan contexts as seen with the coins of Taxila and the Shungas. Gallery See also Post-Mauryan coinage of Gandhara
https://en.wikipedia.org/wiki/Loco-Motion%20%28video%20game%29
Loco-Motion, known as in Japan, is an arcade puzzle game developed by Konami in 1982 and released by Sega in Japan. The North American rights were licensed to Centuri. In Loco-Motion, the player builds a path for their unstoppable locomotive by moving tracks which will allow it to pick up passengers. The game was ported to Intellivision, the Tomy Tutor, and–under a different name–MSX. A clone programmed by Carol Shaw of Activision, Happy Trails, was published for Intellivision before the official version was released. Gameplay Loco-Motion is an updated version of a sliding block puzzle game in which the player can move tiles horizontally or vertically within a rectangular frame that contains one empty square. The tiles are sections of railroad track and the player must use them to construct a path for a locomotive that never stops moving. Laid out around the edges of the frame are several stations with passengers that must be picked up. The player uses a joystick to slide a piece of the track into the vacant square and can use a button to accelerate the locomotive. However, it is always in motion and cannot be stopped. The player must avoid running into a dead-end barricade, a barrier at the playfield edge or the edge of the empty square; doing so costs one life. As the player moves the pieces of track around, the route the locomotive will take is highlighted in yellow up to any dead end. A countdown timer occasionally appears on a station. If passengers are waiting there and the player picks them up before the countdown reaches zero, the remaining amount is added to the score as bonus points. If not, the passengers send a "Crazy Train" onto the tracks and the player must avoid crashing into it. If a countdown timer at an unoccupied station reaches zero, the station is destroyed and becomes a pair of barrier squares. Crazy Trains can be crashed into each other to destroy them but doing so creates a new pair of dead ends or destroys a station if the collision h
https://en.wikipedia.org/wiki/Collision-induced%20absorption%20and%20emission
In spectroscopy, collision-induced absorption and emission refers to spectral features generated by inelastic collisions of molecules in a gas. Such inelastic collisions (along with the absorption or emission of photons) may induce quantum transitions in the molecules, or the molecules may form transient supramolecular complexes with spectral features different from the underlying molecules. Collision-induced absorption and emission is particularly important in dense gases, such as hydrogen and helium clouds found in astronomical systems. Collision-induced absorption and emission is distinguished from collisional broadening in spectroscopy in that collisional broadening comes from elastic collisions of molecules, whereas collision-induced absorption and emission is an inherently inelastic process. Collision-induced spectra of gases Ordinary spectroscopy is concerned with the spectra of single atoms or molecules. Here we outline the very different spectra of complexes consisting of two or more interacting atoms or molecules: the "interaction-induced" or "collision-induced" spectroscopy. Both ordinary and collision-induced spectra may be observed in emission and absorption and require an electric or magnetic multipole moment - in most cases an electric dipole moment - to exist for an optical transition to take place from an initial to a final quantum state of a molecule or a molecular complex. (For brevity of expression we will use here the term "molecule" interchangeably for atoms as well as molecules). A complex of interacting molecules may consist of two or more molecules in a collisional encounter, or else of a weakly bound van der Waals molecule. On first sight, it may seem strange to treat optical transitions of a collisional complex, which may exist just momentarily, for the duration of a fly-by encounter (roughly 10−13 seconds), in much the same way as this was long done for molecules in ordinary spectroscopy. But even transient complexes of molecules may
https://en.wikipedia.org/wiki/Reduced%20form
In statistics, and particularly in econometrics, the reduced form of a system of equations is the result of solving the system for the endogenous variables. This gives the latter as functions of the exogenous variables, if any. In econometrics, the equations of a structural form model are estimated in their theoretically given form, while an alternative approach to estimation is to first solve the theoretical equations for the endogenous variables to obtain reduced form equations, and then to estimate the reduced form equations. Let Y be the vector of the variables to be explained (endogeneous variables) by a statistical model and X be the vector of explanatory (exogeneous) variables. In addition let be a vector of error terms. Then the general expression of a structural form is , where f is a function, possibly from vectors to vectors in the case of a multiple-equation model. The reduced form of this model is given by , with g a function. Structural and reduced forms Exogenous variables are variables which are not determined by the system. If we assume that demand is influenced not only by price, but also by an exogenous variable, Z, we can consider the structural supply and demand model supply:    demand:   where the terms are random errors (deviations of the quantities supplied and demanded from those implied by the rest of each equation). By solving for the unknowns (endogenous variables) P and Q, this structural model can be rewritten in the reduced form: where the parameters depend on the parameters of the structural model, and where the reduced form errors each depend on the structural parameters and on both structural errors. Note that both endogenous variables depend on the exogenous variable Z. If the reduced form model is estimated using empirical data, obtaining estimated values for the coefficients some of the structural parameters can be recovered: By combining the two reduced form equations to eliminate Z, the structural coef
https://en.wikipedia.org/wiki/Genetic%20recombination
Genetic recombination (also known as genetic reshuffling) is the exchange of genetic material between different organisms which leads to production of offspring with combinations of traits that differ from those found in either parent. In eukaryotes, genetic recombination during meiosis can lead to a novel set of genetic information that can be further passed on from parents to offspring. Most recombination occurs naturally and can be classified into two types: (1) interchromosomal recombination, occurring through independent assortment of alleles whose loci are on different but homologous chromosomes (random orientation of pairs of homologous chromosomes in meiosis I); & (2) intrachromosomal recombination, occurring through crossing over. During meiosis in eukaryotes, genetic recombination involves the pairing of homologous chromosomes. This may be followed by information transfer between the chromosomes. The information transfer may occur without physical exchange (a section of genetic material is copied from one chromosome to another, without the donating chromosome being changed) (see SDSA – Synthesis Dependent Strand Annealing pathway in Figure); or by the breaking and rejoining of DNA strands, which forms new molecules of DNA (see DHJ pathway in Figure). Recombination may also occur during mitosis in eukaryotes where it ordinarily involves the two sister chromosomes formed after chromosomal replication. In this case, new combinations of alleles are not produced since the sister chromosomes are usually identical. In meiosis and mitosis, recombination occurs between similar molecules of DNA (homologous sequences). In meiosis, non-sister homologous chromosomes pair with each other so that recombination characteristically occurs between non-sister homologues. In both meiotic and mitotic cells, recombination between homologous chromosomes is a common mechanism used in DNA repair. Gene conversion – the process during which homologous sequences are made identical
https://en.wikipedia.org/wiki/Hydrolyzed%20protein
Hydrolyzed protein is a solution derived from the hydrolysis of a protein into its component amino acids and peptides. While many means of achieving this exist, most common is prolonged heating with hydrochloric acid, sometimes with an enzyme such as pancreatic protease to simulate the naturally occurring hydrolytic process. Uses Protein hydrolysis is a useful route to the isolation of individual amino acids. Examples include cystine from hydrolysis of hair, tryptophane from casein, histidine from red blood cells, and arginine from gelatin. Common hydrolyzed products used in food are hydrolyzed vegetable protein and yeast extract, which are used as flavor enhancers because the hydrolysis of the protein produces free glutamic acid. Some hydrolyzed beef protein powders are used for specialized diets. Protein hydrolysis can be used to modify the allergenic properties of infant formula. Reducing the size of cow milk proteins in the formula makes it more suitable for consumption by babies suffering from milk protein intolerance. The US FDA has approved a label for this usage of partially-hydrolyzed proteins in 2017, but a meta-analysis published the same year shows insufficient evidence for this use. Hydrolyzed protein is also used in certain specially formulated hypoallergenic pet foods, notably dog foods for dogs and puppies that suffer from allergies caused by certain protein types in standard commercial dog food brands. The protein contents of the foods are split into peptides which reduces the likelihood for an animal's immune system recognizing an allergic threat. Hydrolyzed protein diets for cats are often recommended for felines with food allergies and certain types of digestive issues. See also Acceptable daily intake Acid-hydrolyzed vegetable protein E number Food allergy Food intolerance Food labeling regulations Glutamic acid Monosodium glutamate Protein allergy
https://en.wikipedia.org/wiki/Psychophysics
Psychophysics quantitatively investigates the relationship between physical stimuli and the sensations and perceptions they produce. Psychophysics has been described as "the scientific study of the relation between stimulus and sensation" or, more completely, as "the analysis of perceptual processes by studying the effect on a subject's experience or behaviour of systematically varying the properties of a stimulus along one or more physical dimensions". Psychophysics also refers to a general class of methods that can be applied to study a perceptual system. Modern applications rely heavily on threshold measurement, ideal observer analysis, and signal detection theory. Psychophysics has widespread and important practical applications. For example, in the study of digital signal processing, psychophysics has informed the development of models and methods of lossy compression. These models explain why humans perceive very little loss of signal quality when audio and video signals are formatted using lossy compression. History Many of the classical techniques and theories of psychophysics were formulated in 1860 when Gustav Theodor Fechner in Leipzig published Elemente der Psychophysik (Elements of Psychophysics). He coined the term "psychophysics", describing research intended to relate physical stimuli to the contents of consciousness such as sensations (Empfindungen). As a physicist and philosopher, Fechner aimed at developing a method that relates matter to the mind, connecting the publicly observable world and a person's privately experienced impression of it. His ideas were inspired by experimental results on the sense of touch and light obtained in the early 1830s by the German physiologist Ernst Heinrich Weber in Leipzig, most notably those on the minimum discernible difference in intensity of stimuli of moderate strength (just noticeable difference; jnd) which Weber had shown to be a constant fraction of the reference intensity, and which Fechner referred
https://en.wikipedia.org/wiki/Insulin%20degludec/insulin%20aspart
Insulin degludec/insulin aspart, sold under the brand name Ryzodeg, is a fixed-dose combination medication for the treatment of diabetes mellitus. It contains insulin degludec and insulin aspart. It is given as an injection under the skin in the abdominal wall (at the front of the waist), upper arm or thigh. The most frequently reported side effect is hypoglycemia (low blood glucose levels). It was approved for medical use in the European Union in January 2013, and in Australia in November 2017. Medical uses Insulin degludec/insulin aspart is indicated for the treatment of diabetes mellitus in adults, adolescents and children from the age of two years. Insulin degludec and insulin aspart are slightly different from human insulin. The differences mean that insulin degludec is absorbed more slowly by the body. This means it has a long duration of action. Meanwhile, insulin aspart is absorbed faster by the body than human insulin, and therefore it starts to work as soon as it is injected and has a short duration of action.
https://en.wikipedia.org/wiki/Vlasic%20Pickles
Vlasic is an American brand of pickles that is currently owned by Conagra Brands. Since its introduction in 1942, it has become one of the most popular pickle brands in the United States. History Franjo "Frank" Vlašić, a Bosnian Croat, emigrated from Livno, Bosnia and Herzegovina, at the time part of Austria-Hungary, to Detroit, Michigan in 1912 and started a small creamery with savings from his factory job. His son Joseph acquired a milk route in 1922, which eventually grew into the state's largest dairy distributor. In 1937, Vlasic was approached to distribute a home-style pickle, later marketing their fresh-packed pickle in glass jars. Its first plant was built in Imlay City, Michigan. The business rapidly expanded in the post-war years, corresponding with growth in per capita pickle consumption. A child-bearing stork was introduced as a mascot in 1974, merging the stork baby mythology with the notion that pregnant women have an above average appetite for pickles. Vlasic marketed themselves as "the pickle pregnant women crave...after all, who's a better pickle expert?" The modern Vlasic Stork speaks in a style reminiscent of Groucho Marx and holds a pickle like a cigar. The original voice was Ron Masak, since replaced by Doug Preis. Vlasic Pickles was sold to Campbell Soup Company in 1978. It was spun off to Vlasic Foods International on March 30, 1998. On April 13, 1999, the company introduced the Vlasic Hamburger Stackers, made from naturally grown but specially cultivated cucumbers long and over in diameter, which enabled a single pickle chip to cover an entire hamburger. In early 2001, the company voluntarily filed for Chapter 11 bankruptcy and its pickle and barbecue sauce units were initially planned to be sold to H. J. Heinz. By April of 2001, the company had backed out of the deal and was looking into other offers. Vlasic is owned by Pinnacle Foods, a subsidiary of Conagra Brands, since its name change in 2001. The Vlasic Stork was shown in "Icon
https://en.wikipedia.org/wiki/FreeBSD%20version%20history
FreeBSD 1 Released in November 1993. 1.1.5.1 was released in July 1994. FreeBSD 2 2.0-RELEASE was announced on 22 November 1994. The final release of FreeBSD 2, 2.2.8-RELEASE, was announced on 29 November 1998. FreeBSD 2.0 was the first version of FreeBSD to be claimed legally free of AT&T Unix code with approval of Novell. It was the first version to be widely used at the beginnings of the spread of Internet servers. 2.2.9-RELEASE was released April 1, 2006 as a fully functional April Fools' Day prank. FreeBSD 3 FreeBSD 3.0-RELEASE was announced on 16 October 1998. The final release, 3.5-RELEASE, was announced on 24 June 2000. FreeBSD 3.0 was the first branch able to support symmetric multiprocessing (SMP) systems, using a Giant lock and marked the transition from a.out to ELF executables. USB support was first introduced with FreeBSD 3.1, and the first Gigabit network cards were supported in 3.2-RELEASE. FreeBSD 4 4.0-RELEASE appeared in March 2000 and the last 4-STABLE branch release was 4.11 in January 2005 supported until 31 January 2007. FreeBSD 4 was lauded for its stability, was a favorite operating system for ISPs and web hosting providers during the first dot-com bubble, and is widely regarded as one of the most stable and high-performance operating systems of the whole Unix lineage. Among the new features of FreeBSD 4, kqueue(2) was introduced (which is now part of other major BSD systems) and Jails, a way of running processes in separate environments. Version 4.8 was forked by Matt Dillon to create DragonFly BSD. FreeBSD 5 After almost three years of development, the first 5.0-RELEASE in January 2003 was widely anticipated, featuring support for advanced multiprocessor and application threading, and for the UltraSPARC and IA-64 platforms. The first 5-STABLE release was 5.3 (5.0 through 5.2.1 were cut from -CURRENT). The last release from the 5-STABLE branch was 5.5 in May 2006. The largest architectural development in FreeBSD 5 was a major chang
https://en.wikipedia.org/wiki/Electronic%20discovery
Electronic discovery (also ediscovery or e-discovery) refers to discovery in legal proceedings such as litigation, government investigations, or Freedom of Information Act requests, where the information sought is in electronic format (often referred to as electronically stored information or ESI). Electronic discovery is subject to rules of civil procedure and agreed-upon processes, often involving review for privilege and relevance before data are turned over to the requesting party. Electronic information is considered different from paper information because of its intangible form, volume, transience and persistence. Electronic information is usually accompanied by metadata that is not found in paper documents and that can play an important part as evidence (e.g. the date and time a document was written could be useful in a copyright case). The preservation of metadata from electronic documents creates special challenges to prevent spoliation. In the United States, at the federal level, electronic discovery is governed by common law, case law and specific statutes, but primarily by the Federal Rules of Civil Procedure (FRCP), including amendments effective December 1, 2006, and December 1, 2015. In addition, state law and regulatory agencies increasingly also address issues relating to electronic discovery. In England and Wales, Part 31 of the Civil Procedure Rules and Practice Direction 31B on Disclosure of Electronic Documents apply. Other jurisdictions around the world also have rules relating to electronic discovery. Stages of process The Electronic Discovery Reference Model (EDRM) is an ubiquitous diagram that represents a conceptual view of these stages involved in the ediscovery process. Identification The identification phase is when potentially responsive documents are identified for further analysis and review. In the United States, in Zubulake v. UBS Warburg, Hon. Shira Scheindlin ruled that failure to issue a written legal hold notice whenever li
https://en.wikipedia.org/wiki/List%20of%20scale%20model%20sizes
This is a list of scale model sizes, listing a variety of size ratios for scale models. Model scales
https://en.wikipedia.org/wiki/Behavioral%20epigenetics
Behavioral epigenetics is the field of study examining the role of epigenetics in shaping animal and human behavior. It seeks to explain how nurture shapes nature, where nature refers to biological heredity and nurture refers to virtually everything that occurs during the life-span (e.g., social-experience, diet and nutrition, and exposure to toxins). Behavioral epigenetics attempts to provide a framework for understanding how the expression of genes is influenced by experiences and the environment to produce individual differences in behaviour, cognition, personality, and mental health. Epigenetic gene regulation involves changes other than to the sequence of DNA and includes changes to histones (proteins around which DNA is wrapped) and DNA methylation. These epigenetic changes can influence the growth of neurons in the developing brain as well as modify the activity of neurons in the adult brain. Together, these epigenetic changes in neuron structure and function can have a marked influence on an organism's behavior. Background In biology, and specifically genetics, epigenetics is the study of heritable changes in gene activity which are not caused by changes in the DNA sequence; the term can also be used to describe the study of stable, long-term alterations in the transcriptional potential of a cell that are not necessarily heritable. Examples of mechanisms that produce such changes are DNA methylation and histone modification, each of which alters how genes are expressed without altering the underlying DNA sequence. Gene expression can be controlled through the action of repressor proteins that attach to silencer regions of the DNA. DNA methylation turns a gene "off" – it results in the inability of genetic information to be read from DNA; removing the methyl tag can turn the gene back "on". Histone modification changes the way that DNA is packaged into chromosomes. These changes impact how genes are expressed. Epigenetics has a strong influence o
https://en.wikipedia.org/wiki/History%20of%20pathology
The history of pathology can be traced to the earliest application of the scientific method to the field of medicine, a development which occurred in the Middle East during the Islamic Golden Age and in Western Europe during the Italian Renaissance. Early systematic human dissections were carried out by the Ancient Greek physicians Herophilus of Chalcedon and Erasistratus of Chios in the early part of the third century BC. The first physician known to have made postmortem dissections was the Arabian physician Avenzoar (1091–1161). Rudolf Virchow (1821–1902) is generally recognized to be the father of microscopic pathology. Most early pathologists were also practicing physicians or surgeons. Origins of pathology Early understanding of the origins of diseases constitutes the earliest application of the scientific method to the field of medicine, a development which occurred in the Middle East during the Islamic Golden Age and in Western Europe during the Italian Renaissance. The Greek physician Hippocrates, the founder of scientific medicine, was the first to deal with the anatomy and the pathology of human spine. Galen developed an interest in anatomy from his studies of Herophilus and Erasistratus. The concept of studying disease through the methodical dissection and examination of diseased bodies, organs, and tissues may seem obvious today, but there are few if any recorded examples of true autopsies performed prior to the second millennium. Though the pathology of contagion was understood by Muslim physicians since the time of Avicenna (980–1037) who described it in The Canon of Medicine (), the first physician known to have made postmortem dissections was the Arabian physician Avenzoar (1091–1161) who proved that the skin disease scabies was caused by a parasite, followed by Ibn al-Nafis (b. 1213) who used dissection to discover pulmonary circulation in 1242. In the 15th century, anatomic dissection was repeatedly used by the Italian physician Antonio Benivi
https://en.wikipedia.org/wiki/Gleba
Gleba (, from Latin glaeba, glēba, "lump") is the fleshy spore-bearing inner mass of certain fungi such as the puffball or stinkhorn. The gleba is a solid mass of spores, generated within an enclosed area within the sporocarp. The continuous maturity of the sporogenous cells leave the spores behind as a powdery mass that can be easily blown away. The gleba may be sticky or it may be enclosed in a case (peridiole). It is a tissue usually found in an angiocarpous fruit-body, especially gasteromycetes. Angiocarpous fruit-bodies usually consist of fruit enclosed within a covering that does not form a part of itself; such as the filbert covered by its husk, or the acorn seated in its cupule. The presence of gleba can be found in earthballs and puffballs. The gleba consists of mycelium and basidia and may also contain capillitium threads. Gleba found on the fruit body of species in the family Phallaceae is typically gelatinous, often fetid-smelling, and deliquescent (becoming liquid from the absorption of water). It is formed on the exterior face of the cap or the upper part of the fruit body. The foul smell helps to attract insects that help disperse the spores. Chemicals that contribute to the odor include methylmercaptan and hydrogen sulfide.
https://en.wikipedia.org/wiki/Physiology%20%28journal%29
Physiology is a peer-reviewed scientific journal on physiology published by the American Physiological Society and the International Union of Physiological Societies. Before August 2003, it was named News in Physiological Sciences. The current editor-in-chief is Gary C. Sieck (Mayo Clinic). External links Bimonthly journals English-language journals Academic journals established in 1986 Physiology journals
https://en.wikipedia.org/wiki/Horticulture
Horticulture is the cultivation of plants in gardens or greenhouses, as opposed to the field-scale production of crops characteristic of agriculture. It includes the cultivation of fruits, vegetables, nuts, seeds, herbs, sprouts, mushrooms, algae, flowers, seaweeds and non-food crops such as grass and ornamental trees and plants. It also includes plant conservation, landscape restoration, landscape and garden design, construction, and maintenance, and arboriculture, ornamental trees and lawns. In anthropology, horticulture refers to a subsistence strategy characterized by the small-scale, non-industrial cultivation of plants for food. Horticulture involves the use of hand tools such as digging sticks, hoes and carrying baskets. In contrast to horticulture, agriculture is viewed by anthropologists as a more intensive strategy involving the use of plowing, animal traction and complex techniques of irrigation and soil management. The study and practice of horticulture have been traced back many years. Horticulture contributed to the transition from nomadic human communities to sedentary, or semi-sedentary, horticultural communities. Horticulture is divided into several categories which focus on the cultivation and processing of different types of plants and food items for specific purposes. In order to conserve the science of horticulture, multiple organizations worldwide educate, encourage, and promote the advancement of horticulture. Some notable horticulturists include Luca Ghini and Luther Burbank. Definition Horticulture, from Latin hortus meaning garden and colere meaning to cultivate, focuses on the use of small plots, in contrast to agronomy which involves intensive crop farming and large-scale field crop production of grains and forages or forestry involving forest trees and products derived from them. It deals with garden crops such as ornamental plants grown for their appearance, fruits and vegetable and spices grown for their food value, and medicinal p
https://en.wikipedia.org/wiki/Magnesium%20citrate
Magnesium citrate is a magnesium preparation in salt form with citric acid in a 1:1 ratio (1 magnesium atom per citrate molecule). It contains 11.23% magnesium by weight. The name "magnesium citrate" is ambiguous and sometimes may refer to other salts such as trimagnesium dicitrate which has a magnesium:citrate ratio of 3:2, or monomagnesium dicitrate with a ratio of 1:2, or a mix of two or three of the salts of magnesium and citric acid. Magnesium citrate (sensu lato) is used medicinally as a saline laxative and to completely empty the bowel prior to a major surgery or colonoscopy. It is available without a prescription, both as a generic and under various brand names. It is also used in the pill form as a magnesium dietary supplement. As a food additive, magnesium citrate is used to regulate acidity and is known as E number E345. Mechanism of action Magnesium citrate works by attracting water through the tissues by a process known as osmosis. Once in the intestine, it can attract enough water into the intestine to induce defecation. The additional water stimulates bowel motility. This means it can also be used to treat rectal and colon problems. Magnesium citrate functions best on an empty stomach, and should always be followed with a full (eight ounce or 250 ml) glass of water or juice to help counteract water loss and aid in absorption. Magnesium citrate solutions generally produce bowel movement in one-half to three hours. There is an exothermic heat generation when water is added, which is "most disagreeable when taken orally." Use and dosage The maximum upper tolerance limit (UTL) for magnesium in supplement form for adults is 350 mg of elemental magnesium per day, according to the National Institutes of Health (NIH). In addition, according to the NIH, total dietary requirements for magnesium from all sources (in other words, food and supplements) is 320–420 mg of elemental magnesium per day, though there is no UT for dietary magnesium. Laxative Mag
https://en.wikipedia.org/wiki/Next-Generation%20Secure%20Computing%20Base
The Next-Generation Secure Computing Base (NGSCB; codenamed Palladium and also known as Trusted Windows') is a software architecture designed by Microsoft which claimed to provide users of the Windows operating system with better privacy, security, and system integrity. NGSCB was the result of years of research and development within Microsoft to create a secure computing solution that equaled the security of closed platforms such as set-top boxes while simultaneously preserving the backward compatibility, flexibility, and openness of the Windows operating system. Microsoft's primary stated objective with NGSCB was to "protect software from software." Part of the Trustworthy Computing initiative when unveiled in 2002, NGSCB was to be integrated with Windows Vista, then known as "Longhorn." NGSCB relied on hardware designed by the Trusted Computing Group to produce a parallel operation environment hosted by a new hypervisor (referred to as a sort of kernel in documentation) called the "Nexus" that existed alongside Windows and provided new applications with features such as hardware-based process isolation, data encryption based on integrity measurements, authentication of a local or remote machine or software configuration, and encrypted paths for user authentication and graphics output. NGSCB would facilitate the creation and distribution of digital rights management (DRM) policies pertaining the use of information. NGSCB was subject to much controversy during its development, with critics contending that it would impose restrictions on users, enforce vendor lock-in, and undermine fair use rights and open-source software. It was first demonstrated by Microsoft at WinHEC 2003 before undergoing a revision in 2004 that would enable earlier applications to benefit from its functionality. Reports indicated in 2005 that Microsoft would change its plans with NGSCB so that it could ship Windows Vista by its self-imposed deadline year, 2006; instead, Microsoft would ship
https://en.wikipedia.org/wiki/Morchella%20meiliensis
Morchella meiliensis is a species of fungus in the family Morchellaceae native to China. Taxonomy The species was described as new to science in 2006. The specific epithet meiliensis refers to Meili Snow Mountain in Yunnan, where the type specimen was collected. Description The fruit bodies are with a conical cap measuring tall by wide. The surface has vertically arranged ridges that are dark brown to black in colour, while the rectangular to quadrangular pits between the ridges are merulioid (wrinkled with low, uneven ridges) and yellowish in colour. The flesh is thin, and lacks any distinctive taste or odour. The cylindrical stipe measures tall by thick. Initially whitish, it turns yellowish with a waxy sheen when dry. In deposit, ascospores are smooth, ellipsoid, hyaline (translucent), and measure 4.7–5.1 by 5.2–5.7 µm. They are thin-walled and contain oil droplets. Asci (spore-bearing cells) are eight-spored, cylindrical, and hyaline, and have dimensions of 5.2–5.9 µm long by 91–94 µm long. The paraphyses are dark, club-shaped, and measure 4.2–5.2 by 40–65 µm. Similar species Morchella conica and M. angusticeps are similar in appearance to M. meiliensis, but the latter species can be distinguished by more lightly coloured ridges on the cap surface, the merulioid texture of the pits, and microscopically by the club-shaped paraphyses. Habitat and distribution Morchella meiliensis fruits on the ground in deciduous or mixed forests. It is known from Deqin County, Yunnan Province in China, where it grows at elevations of .
https://en.wikipedia.org/wiki/5-Con%20triangles
In geometry, two triangles are said to be 5-Con or almost congruent if they are not congruent triangles but they are similar triangles and share two side lengths (of non-corresponding sides). The 5-Con triangles are important examples for understanding the solution of triangles. Indeed, knowing three angles and two sides (but not their sequence) is not enough to determine a triangle up to congruence. A triangle is said to be 5-Con capable if there is another triangle which is almost congruent to it. The 5-Con triangles have been discussed by Pawley:, and later by Jones and Peterson. They are briefly mentioned by Martin Gardner in his book Mathematical Circus. Another reference is the following exercise Explain how two triangles can have five parts (sides, angles) of one triangle congruent to five parts of the other triangle, but not be congruent triangles. A similar exercise dates back to 1955, and there an earlier reference is mentioned. It is however not possible to date the first occurrence of such standard exercises about triangles. Examples There are infinitely many pairs of 5-Con triangles, even up to scaling. The smallest 5-Con triangles with integer sides have side lengths (8; 12; 18) and (12; 18; 27). This is an example with obtuse triangles. An example of acute 5-Con triangles is (1000; 1100; 1210) and (1100; 1210; 1331). The 5-Con right triangles are exactly those obtained from scaling the pair and with where φ is the golden ratio. Consequently, these are Kepler triangles and there can be no right 5-Con triangles with integer sides. There are no 5-Con triangles that are equilateral or isosceles because that would require m = 1 and the 5-Con triangles would be congruent. There are no integer 5-Con triangles that are Heronian because the sides of integer 5-Con triangles are in a geometric progression. Results 1. Consider 5-Con triangles with side lengths and where is the scaling factor, which we may suppose to be greater than . We may also
https://en.wikipedia.org/wiki/Graphics%20device%20interface
A graphics device interface is a subsystem that most operating systems use for representing graphical objects and transmitting them to output devices such as monitors and printers. In most cases, the graphics device interface is only able to draw 2D graphics and simple 3D graphics, in order to make use of more advanced graphics and keep performance, an API such as DirectX or OpenGL needs to be installed. In Microsoft Windows, the GDI functionality resides in gdi.exe on 16-bit Windows, and gdi32.dll on 32-bit Windows. Operating system technology
https://en.wikipedia.org/wiki/Co-occurrence%20matrix
A co-occurrence matrix or co-occurrence distribution (also referred to as : gray-level co-occurrence matrices GLCMs) is a matrix that is defined over an image to be the distribution of co-occurring pixel values (grayscale values, or colors) at a given offset. It is used as an approach to texture analysis with various applications especially in medical image analysis. Method Given a grey-level image , co-occurrence matrix computes how often pairs of pixels with a specific value and offset occur in the image. The offset, , is a position operator that can be applied to any pixel in the image (ignoring edge effects): for instance, could indicate "one down, two right". An image with different pixel values will produce a co-occurrence matrix, for the given offset. The value of the co-occurrence matrix gives the number of times in the image that the and pixel values occur in the relation given by the offset. For an image with different pixel values, the co-occurrence matrix C is defined over an image , parameterized by an offset , as: where: and are the pixel values; and are the spatial positions in the image I; the offsets define the spatial relation for which this matrix is calculated; and indicates the pixel value at pixel . The 'value' of the image originally referred to the grayscale value of the specified pixel, but could be anything, from a binary on/off value to 32-bit color and beyond. (Note that 32-bit color will yield a 232 × 232 co-occurrence matrix!) Co-occurrence matrices can also be parameterized in terms of a distance, , and an angle, , instead of an offset . Any matrix or pair of matrices can be used to generate a co-occurrence matrix, though their most common application has been in measuring texture in images, so the typical definition, as above, assumes that the matrix is an image. It is also possible to define the matrix across two different images. Such a matrix can then be used for color mapping. Aliases Co-occurrence mat
https://en.wikipedia.org/wiki/Mycotaxon
Mycotaxon is a peer-reviewed scientific journal that covers the nomenclature and taxonomy of fungi, including lichens. The journal was founded by Grégoire L. Hennebert and Richard P. Korf in 1974. They were frustrated that papers submitted to journals such as Mycologia took a year or longer from submission to publication. Korf and Hennebert introduced a number of innovations to make their journal more efficient and accessible than its contemporaries. Mycotaxon reduced the wait time between submission and publication by requiring authors to submit camera-ready copy. Linotype was the industry standard at the time; Mycotaxon used photo-offset lithography to expedite publication. A quarterly journal, Mycotaxon aimed to publish papers within four months of submission. Mycotaxon took an unusual non-blind approach to refereeing: authors were required to enlist a reviewer outside their institution to peer-review their manuscript prior to its submission. Initially Mycotaxon did not demand page charges from authors, rather relying on subscription fees to finance publication. Papers of all lengths were accepted. See also Persoonia – a journal with similar scope
https://en.wikipedia.org/wiki/Plica%20semilunaris%20of%20the%20fauces
The plica semilunaris is the thin upper part of the fold of mucous membrane in the supratonsillar fossa that reaches across between the two arches. A separate fold is called the plica triangularis which runs inferoposteriorly from the posterior surface of the palatoglossal arch to cover the inferior portion of the tonsil.
https://en.wikipedia.org/wiki/Gelfand%20representation
In mathematics, the Gelfand representation in functional analysis (named after I. M. Gelfand) is either of two things: a way of representing commutative Banach algebras as algebras of continuous functions; the fact that for commutative C*-algebras, this representation is an isometric isomorphism. In the former case, one may regard the Gelfand representation as a far-reaching generalization of the Fourier transform of an integrable function. In the latter case, the Gelfand–Naimark representation theorem is one avenue in the development of spectral theory for normal operators, and generalizes the notion of diagonalizing a normal matrix. Historical remarks One of Gelfand's original applications (and one which historically motivated much of the study of Banach algebras) was to give a much shorter and more conceptual proof of a celebrated lemma of Norbert Wiener (see the citation below), characterizing the elements of the group algebras L1(R) and whose translates span dense subspaces in the respective algebras. The model algebra For any locally compact Hausdorff topological space X, the space C0(X) of continuous complex-valued functions on X which vanish at infinity is in a natural way a commutative C*-algebra: The structure of algebra over the complex numbers is obtained by considering the pointwise operations of addition and multiplication. The involution is pointwise complex conjugation. The norm is the uniform norm on functions. The importance of X being locally compact and Hausdorff is that this turns X into a completely regular space. In such a space every closed subset of X is the common zero set of a family of continuous complex-valued functions on X, allowing one to recover the topology of X from C0(X). Note that C0(X) is unital if and only if X is compact, in which case C0(X) is equal to C(X), the algebra of all continuous complex-valued functions on X. Gelfand representation of a commutative Banach algebra Let be a commutative Banach algeb
https://en.wikipedia.org/wiki/Loader%20%28computing%29
In computer systems a loader is the part of an operating system that is responsible for loading programs and libraries. It is one of the essential stages in the process of starting a program, as it places programs into memory and prepares them for execution. Loading a program involves either memory-mapping or copying the contents of the executable file containing the program instructions into memory, and then carrying out other required preparatory tasks to prepare the executable for running. Once loading is complete, the operating system starts the program by passing control to the loaded program code. All operating systems that support program loading have loaders, apart from highly specialized computer systems that only have a fixed set of specialized programs. Embedded systems typically do not have loaders, and instead, the code executes directly from ROM or similar. In order to load the operating system itself, as part of booting, a specialized boot loader is used. In many operating systems, the loader resides permanently in memory, though some operating systems that support virtual memory may allow the loader to be located in a region of memory that is pageable. In the case of operating systems that support virtual memory, the loader may not actually copy the contents of executable files into memory, but rather may simply declare to the virtual memory subsystem that there is a mapping between a region of memory allocated to contain the running program's code and the contents of the associated executable file. (See memory-mapped file.) The virtual memory subsystem is then made aware that pages with that region of memory need to be filled on demand if and when program execution actually hits those areas of unfilled memory. This may mean parts of a program's code are not actually copied into memory until they are actually used, and unused code may never be loaded into memory at all. Responsibilities In Unix, the loader is the handler for the system call execv
https://en.wikipedia.org/wiki/Visual%20rhetoric
Visual rhetoric is the art of effective communication through visual elements such as images, typography, and texts. Visual rhetoric encompasses the skill of visual literacy and the ability to analyze images for their form and meaning. Drawing on techniques from semiotics and rhetorical analysis, visual rhetoric expands on visual literacy as it examines the structure of an image with the focus on its persuasive effects on an audience. Although visual rhetoric also involves typography and other texts, it concentrates mainly on the use of images or visual texts. Using images is central to visual rhetoric because these visuals help in either forming the case an image alone wants to convey, or arguing the point that a writer formulates, in the case of a multimodal text which combines image and written text, for example. Visual rhetoric has gained more notoriety as more recent scholarly work started exploring alternative media forms that include graphics, screen design, and other hybrid visual representations that does not privilege print culture and conventions. Also, visual rhetoric involves how writers arrange segments of a visual text on the page. In addition to that, visual rhetoric involves the selection of different fonts, contrastive colors, and graphs, among other elements, to shape a visual rhetoric text. One vital component of visual rhetoric is analyzing the visual text. The interactional and commonly hybrid nature of cyber spaces that usually mixes print text and visual images unable some detachment of them as isolated constructs, and scholarship has claimed that especially in virtual spaces where print text and visuals are usually combined, there is no place either for emphasizing one mode over another. One way of analyzing a visual text is to look for its significant meaning. Simply put, the meaning should be deeper than the literal sense that a visual text holds. One way to analyze a visual text is to dissect it in order for the viewer to understand i
https://en.wikipedia.org/wiki/Bicipital%20aponeurosis
The bicipital aponeurosis (also known as lacertus fibrosus) is a broad aponeurosis of the biceps brachii, which is located in the cubital fossa of the elbow. It separates superficial from deep structures in much of the fossa. Structure The bicipital aponeurosis originates from the distal insertion of the biceps brachii, and inserts into the deep fascia of the forearm. The biceps tendon inserts on the radial tuberosity, and the bicipital aponeurosis lies medially to it. It reinforces the cubital fossa, helping to protect the brachial artery and the median nerve running underneath. Variations Some individuals (about 3% of the population) have a superficial ulnar artery that runs superficially to the bicipital aponeurosis instead of underneath it. These individuals are at risk for accidental injury to the ulnar artery during venipuncture. Clinical significance The bicipital aponeurosis is superficial to the brachial artery and the median nerve, but deep to the median cubital vein. This protection is important during venipuncture (taking blood). It is one structure that has to be incised during fasciotomy in the treatment of acute compartment syndrome of the forearm and elbow region.
https://en.wikipedia.org/wiki/Reconciliation%20ecology
Reconciliation ecology is the branch of ecology which studies ways to encourage biodiversity in the human-dominated ecosystems of the anthropocene era. Michael Rosenzweig first articulated the concept in his book Win-Win Ecology, based on the theory that there is not enough area for all of earth's biodiversity to be saved within designated nature preserves. Therefore, humans should increase biodiversity in human-dominated landscapes. By managing for biodiversity in ways that do not decrease human utility of the system, it is a "win-win" situation for both human use and native biodiversity. The science is based in the ecological foundation of human land-use trends and species-area relationships. It has many benefits beyond protection of biodiversity, and there are numerous examples of it around the globe. Aspects of reconciliation ecology can already be found in management legislation, but there are challenges in both public acceptance and ecological success of reconciliation attempts. Theoretical basis Human land use trends Traditional conservation is based on "reservation and restoration"; reservation meaning setting pristine lands aside for the sole purpose of maintaining biodiversity, and restoration meaning returning human impacted ecosystems to their natural state. However, reconciliation ecologists argue that there is too great a proportion of land already impacted by humans for these techniques to succeed. While it is difficult to measure exactly how much land has been transformed by human use, estimates range from 39 to 50%. This includes agricultural land, pastureland, urban areas, and heavily harvested forest systems. An estimated 50% of arable land is already under cultivation. Land transformation has increased rapidly over the last fifty years, and is likely to continue to increase. Beyond direct transformation of land area, humans have impacted the global biogeochemical cycles, leading to human caused change in even the most remote areas. These inc
https://en.wikipedia.org/wiki/Suaeda%20pulvinata
Suaeda pulvinata is an endemic seepweed from Mexico. It lives in the shores of Lake Texcoco and Lake Totolcingo. It lives underwater as an aquatic plant for half of the year and in dry land as a terrestrial plant for the other half due to the changing levels of the lakes that it inhabits. It is a perennial flat herb with prostrate stems. Its leaves and inflorescences are green to reddish in color. This species is important for people that live in the states of Puebla and Tlaxcala, as it is an edible vegetable. The dish that is prepared using this species is known as romeritos. It has been found in molecular phylogenetic studies that this taxon is monophyletic. Due to differences in its phylogenetic position in its nuclear ITS tree and its chloroplast rpl32-trnL tree, it is thought this species is the result of hybridization of ancestral species of Suaeda. The first scientific collector who found this plant was Efraim Hernandez Xolocotzi. Later, it was cited by Guadalupe Ramos in her university degree thesis. However, he misidentified it for S. nigra. It was in 2013 that Ernesto Alvarado Reyes and Hilda Flores Olvera noticed it was a different species.
https://en.wikipedia.org/wiki/Electronic%20circuit%20design
Electronic circuit design comprises the analysis and synthesis of electronic circuits. Methods To design any electrical circuit, either analog or digital, electrical engineers need to be able to predict the voltages and currents at all places within the circuit. Linear circuits, that is, circuits wherein the outputs are linearly dependent on the inputs, can be analyzed by hand using complex analysis. Simple nonlinear circuits can also be analyzed in this way. Specialized software has been created to analyze circuits that are either too complicated or too nonlinear to analyze by hand. Circuit simulation software allows engineers to design circuits more efficiently, reducing the time cost and risk of error involved in building circuit prototypes. Some of these make use of hardware description languages such as VHDL or Verilog. Network simulation software More complex circuits are analyzed with circuit simulation software such as SPICE and EMTP. Linearization around operating point When faced with a new circuit, the software first tries to find a steady state solution wherein all the nodes conform to Kirchhoff's Current Law and the voltages across and through each element of the circuit conform to the voltage/current equations governing that element. Once the steady state solution is found, the software can analyze the response to perturbations using piecewise approximation, harmonic balance or other methods. Piece-wise linear approximation Software such as the PLECS interface to Simulink uses piecewise linear approximation of the equations governing the elements of a circuit. The circuit is treated as a completely linear network of ideal diodes. Every time a diode switches from on to off or vice versa, the configuration of the linear network changes. Adding more detail to the approximation of equations increases the accuracy of the simulation, but also increases its running time. Synthesis Simple circuits may be designed by connecting a number of elements or
https://en.wikipedia.org/wiki/SolarMagic
SolarMagic is a subsidiary brand of Texas Instruments, selling electronics for photovoltaic (PV) systems. It was established in 2008 as a subsidiary brand of National Semiconductor. The brand name was carried through when TI purchased National Semiconductor in 2011. Products The original SolarMagic device was a proprietary brand of power optimizer, a device which isolates the output from each PV-module from the rest of the system, preventing loss of performance by a single module from affecting the rest of the array. Because of the electronic characteristics of solar arrays, real-world conditions such as partial/temporary shading and irregular panel performance can lead to disproportionate and excessive losses of power output from the system. In one study, shading over a day of between 8% and 16% of the PV system surface area led to relative array power losses of between 35% and 40%. Despite National Semiconductor advertisement claims that their power optimizer could recoup up to 57% of the power lost as a result of these phenomena, real-world results typically ranged between 1% and 6% of overall system output improvement.
https://en.wikipedia.org/wiki/KEVA%20Planks
KEVA Planks are cuboid wooden block toys for children and adults. Each block is sized approximately . The blocks are available for sale in maple, that is produced in the United States, and less expensive imported pine versions. KEVA Planks started out as a simple construction set that is unusual because they only use one piece and no connectors in glue, in contrast to other building sets that often have specific instructions and require sorting. It has developed into a teaching tool used in classrooms and enjoyed in homes. A number of museums have KEVA exhibits for hands on experience with design and construction including: Da Vinci Science Center of Allentown, PA Exploration Place of The Sedgwick County Science and Discovery Center of Wichita, Kansas Fleet Science Center in San Diego, CA Kalamazoo Air Zoo of Kalamazoo, MI Lawrence Hall of Science at University of California, Berkeley, CA Rochester Museum and Science Center in Rochester, NY Science World (Vancouver) in Vancouver, BC Kaleideum of Winston-Salem, NC Discovery Park of America in Union City, TN The tallest tower built with KEVA planks was 51 feet, 8 inches constructed at the National Building Museum in 2006. KEVA planks is a privately owned company located in Virginia. Mindware, a division of Oriental Trading subsidiary of Berkshire Hathaway, is the exclusive licensee of many KEVA Planks products. KEVA Planks in Educational Settings KEVA Planks are used in schools, libraries, museums, and maker spaces. They are a teaching tool that can be used as a manipulative to teach subjects including math, science, geography, history, and humanities. They were featured at Destination Imagination Global Finals in Knoxville, Tennessee in 2011. Beginning in 2015, KEVA Planks traveled with Share Fair Nation STEMosphere events and was one of the most popular sessions in the professional development workshops. STEMosphere highlights innovative and creative teaching tools. KEVA Planks were named number 3
https://en.wikipedia.org/wiki/Walter%20Douglas%20Hincks
Walter Douglas Hincks (3 September 1906 – 12 June 1961) was a British entomologist and museum curator. He was a world expert on the Dermaptera. Biography Hincks originally trained as a chemist and worked in the Pharmaceutical sector before his transition to professional entomology. He became passionate for entomology during his time as a member of the Leeds Naturalist's Club and was particularly encourage to take an interest in the Dermaptera by Malcolm Burr. In 1941 he spent some time rearranging the coleoptera collections at the Yorkshire Museum. The following year he was appointed the Honorary Curator of Entomology (excluding the lepidoptera). During 1942 Hincks worked with A. Smith and Reginald Wagstaffe to collect entomological specimens from Askham Bog for the museum collections. He was instrumental in bringing the Ellis collection of insects to the museum in 1945. Hincks replaced Harry Britten as Assistant Keeper at the Manchester Museum in 1947. In 1957 he was promoted to Keeper and remained in this position until his death in 1961. He, along with his technician (and eventual successor) Alan Brindle, were responsible for developing the extent and geographic scope of the Dermaptera collection at the museum. He was president of the Manchester Entomological Society in 1952–1953. Hincks also served as the Assistant Secretary for the North Western Naturalists' Society, and was recorder of the Lancashire and Cheshire Fauna Committee. He was a member of the Yorkshire Naturalists' Union, the Microscopy Society, and the Society for British Entomology. He was a member of the Yorkshire Philosophical Society during his time in York. Hincks published widely on the Dermaptera of the world, often in international journals, and was responsible for naming several new species of that order. He was well known as the co-author, with G S Kloet, of "A Checklist of British Insects". Select publications Hincks, W.D. 1938. "Die Arthropodenfauna von Madeira nach den Ergebniss
https://en.wikipedia.org/wiki/Multiprotocol%20Encapsulation%20over%20ATM
Multiprotocol Encapsulation over ATM is specified in RFC 2684. It defines two mechanisms for identifying the protocol carried in ATM Adaptation Layer 5 (AAL5) frames. It replaces RFC 1483, a standard data link access protocol supported by DSL modems. RFC 2684 describes two encapsulation mechanisms for network traffic: Virtual Circuit Multiplexing and LLC Encapsulation. Either mechanism carries either routed or bridged protocol data units, and DSL modems often include a setting for RFC 1483 bridging. This is distinct from other "bridge modes" commonly found in combined DSL modems and routers, which turn off the router portion of the DSL modem. In VC Multiplexing (VC-MUX), the hosts agree on the high-level protocol for a given circuit. It has the advantage of not requiring additional information in a packet, which minimises the overhead. For example, if the hosts agree to transfer IP, a sender can pass each datagram directly to AAL5 to transfer; nothing needs to be sent besides the datagram and the AAL5 trailer. The chief disadvantage of such a scheme lies in duplication of virtual circuits: a host must create a separate virtual circuit for each high-level protocol if more than one protocol is used. Because most carriers charge for each virtual circuit, customers try to avoid using multiple circuits because it adds unnecessary cost. In LLC Encapsulation the hosts use a single virtual circuit for multiple protocols. This has the advantage of allowing all traffic over the same circuit, but the disadvantage of requiring each packet to contain octets that identify the protocol type, which adds overhead. The scheme also has the disadvantage that packets from all protocols travel with the same delay and priority. RFC 2684 specifies that hosts can choose between the two methods of using AAL5. Both the sender and receiver must agree on how the circuit will be used, and the agreement may involve manual configuration. Furthermore, the standards suggest that when h
https://en.wikipedia.org/wiki/TA%20cloning
TA cloning (also known as rapid cloning or T cloning) is a subcloning technique that avoids the use of restriction enzymes and is easier and quicker than traditional subcloning. The technique relies on the ability of adenine (A) and thymine (T) (complementary basepairs) on different DNA fragments to hybridize and, in the presence of ligase, become ligated together. PCR products are usually amplified using Taq DNA polymerase which preferentially adds an adenine to the 3' end of the product. Such PCR amplified inserts are cloned into linearized vectors that have complementary 3' thymine overhangs. Procedure Creating the insert The insert is created by PCR using Taq polymerase. This polymerase lacks 3' to 5' proofreading activity and, with a high probability, adds a single, 3'-adenine overhang to each end of the PCR product. It is best if the PCR primers have guanines at the 5' end as this maximizes probability of Taq DNA polymerase adding the terminal adenosine overhang. Thermostable polymerases containing extensive 3´ to 5´ exonuclease activity should not be used as they do not leave the 3´ adenine-overhangs. Creating the vector The target vector is linearized and cut with a blunt-end restriction enzyme. This vector is then tailed with dideoxythymidine triphosphate (ddTTP) using terminal transferase. It is important to use ddTTP to ensure the addition of only one T residue. This tailing leaves the vector with a single 3'-overhanging thymine residue on each blunt end. Manufacturers commonly sell TA Cloning "kits" with a wide range of prepared vectors that have already been linearized and tagged with an overhanging thymine. Benefits and drawbacks Given that there is no need for restriction enzymes other than for generating the linearized vector, the procedure is much simpler and faster than traditional subcloning. There is also no need to add restriction sites when designing primers and thus shorter primers can be used saving time and money. In addition, in instanc
https://en.wikipedia.org/wiki/Occipitomastoid%20suture
The occipitomastoid suture or occipitotemporal suture is the cranial suture between the occipital bone and the mastoid portion of the temporal bone. It is continuous with the lambdoidal suture. See also Jugular foramen Additional images
https://en.wikipedia.org/wiki/Kodiak%E2%80%93Bowie%20Seamount%20chain
The Kodiak–Bowie Seamount chain, also called the Pratt–Welker Seamount chain and the Kodiak Seamounts is a seamount chain in the southeastern Gulf of Alaska stretching from the Aleutian Trench in the north to Bowie Seamount, the youngest volcano in the chain, which lies west of Haida Gwaii, British Columbia, Canada. The oldest volcano in the chain is the Kodiak Seamount. Although the Kodiak Seamount is the oldest extant seamount in the Kodiak-Bowie chain, the adjacent lower slope contains transverse scars indicating earlier subduction of seamounts. The Kodiak–Bowie Seamount chain is mostly extinct volcanoes that formed above the Bowie hotspot. This is a 100-to-150-km-wide morphological swell presumably of thickened hotspot generated crust, although there are no seismic refraction data across the swell to define crustal thickness. The crest of one such peak, Patton Seamount originally formed off Washington state as a submerged volcano 33 million years ago. Over time, as the Pacific Plate moved steadily northwest, Patton Seamount has carried off the Bowie hotspot and into the Gulf of Alaska. New volcanoes were formed one after another over the hotspot, creating the Kodiak–Bowie Seamount chain. Explorations of the Kodiak–Bowie Seamount chain have shown that despite the fact that most of the seamounts were created by the Bowie hotspot, all are unique in their size, shape, and volcanic features. The seamounts teem with deep-sea corals, sponges, and fish. Recent expeditions to these seamounts using crewed submersibles and ROVs have discovered many marine species and have greatly expanded the knowledge of the range of deep sea corals in this region. For example, the Bowie Seamount is a biologically rich area with a dynamic and productive ecosystem. Because of this unique biological rich area, Bowie Seamount was declared a Pilot Marine Protected Area on December 8, 1998. The Kodiak–Bowie seamount chain is at the northern triple junction between the Pacific, North Americ
https://en.wikipedia.org/wiki/Advent%20of%20Code
Advent of Code is an annual set of Christmas-themed computer programming challenges that follow an Advent calendar. It has been running since 2015. The programming puzzles cover a variety of skill sets and skill levels and can be solved using any programming language. Participants also compete based on speed on both global and private leaderboards. The event was founded and is maintained by software engineer Eric Wastl. History Advent of Code was created by Wastl, who is still the sole maintainer of the project. The event was initially launched on December 1, 2015. By midnight EST (UTC−05:00), 81 people had signed up for the event, going slightly over Wastl's planned 70-participant capacity. Within 12 hours, about 4,000 people had joined, nearly causing a system crash. After 48 hours, there were about 15,000 people, and by the end of the 2015 event, the total had risen to 52,000. In 2020, perhaps due to the COVID-19 pandemic, the event saw a 50% growth in traffic, with over 180,000 participants worldwide. On December 4, 2022, Wastl announced that the project had reached 1,000,000 registered users. Puzzle design Puzzles consist of two parts that must be solved in order, with the second part not revealed to the user until the first part is solved correctly. Participants earn one golden star for each part they finish, giving a possible total of two stars per day and fifty stars per year. Each puzzle contains a fictional backstory that is the same for all participants, but each person receives a different piece of input data and should generate a different correct result. Puzzles are released on a daily schedule from December 1 to December 25 at midnight EST. There is no time limit to complete the puzzles, and puzzles from past years' events remain available to solve. Some participants have used tools such as GitHub Copilot and ChatGPT to assist in solving the puzzles. In 2023, Wastl requested that users abstain from using these tools until after the leaderb
https://en.wikipedia.org/wiki/Paralepista%20flaccida
Paralepista flaccida (also called Clitocybe flaccida, Clitocybe inversa, Lepista flaccida and Lepista inversa, or in English tawny funnel cap) is a species of mushroom found across the Northern Hemisphere. It is known to form fairy rings. Naming The naming history of this mushroom is complicated by the fact that for a long time it was regarded as two different species, "flaccida" (associated with broad-leaved trees) and "inversa" (associated with conifers and with a smoother shinier cap). These forms can still be differentiated as varieties within P. flaccida if desired. The earliest description was by Giovanni Antonio Scopoli in 1772 as Agaricus inversus in his book Flora Carniolica, then in 1799 James Sowerby created a description under the name Agaricus flaccidus in his major work "Coloured Figures of English Fungi or Mushrooms". In later years there were defined the combinations Clitocybe flaccida (by Paul Kummer, 1871), Clitocybe inversa (by Lucien Quélet, 1872), and Lepista flaccida and Lepista inversa (by Narcisse Patouillard, 1887). There followed long-standing confusion over whether Clitocybe or Lepista was the appropriate genus. Using molecular analysis, in 2012 Alfredo Vizzini published a paper showing (with a detailed phylogram) that these mushrooms together with Lepista gilva form a clade which is separate from other Lepista species (such as Lepista nuda) and also from Clitocybe species (such as Clitocybe fragrans). Jörg H. Raithelhuber had already proposed the name Paralepista as a genus for those mushrooms in 1981, this had become accepted as a subgenus only, and Vizzini accordingly adopted the same name at the genus level. The new genus is recognized by Species Fungorum, and the Global Biodiversity Information Facility, but not necessarily in derivative databases such as the Encyclopedia of Life. Older mushroom references generally listed the flaccida and inversa forms as separate species, but modern authorities merge them into one. In English P
https://en.wikipedia.org/wiki/Sentinel%20outlet
A sentinel outlet in occupational safety and health is a water outlet that is chosen to have its temperature monitored so that risk from Legionella can be controlled. This is typically chosen to be the closest and furthest outlets from the water tank.
https://en.wikipedia.org/wiki/RPMI%201640
RPMI 1640, simply known as RPMI medium, is a cell culture medium commonly used to culture mammalian cells. RPMI 1640 was developed by George E. Moore, Robert E. Gerner, and H. Addison Franklin in 1966 at Roswell Park Comprehensive Cancer Center (formerly known as Roswell Park Memorial Institute), from where it derives its name. A modification of McCoy′s 5A medium (or RPMI 1630), it was originally formulated to support lymphoblastoid cells in suspension cultures, but can also support a wide variety of adherent cells. It was originally developed to culture human leukemic cells. Over the years, the original formulation was modified and refined by researchers and commercial suppliers to enhance its ability to support the growth of many cell types. This medium contains a great deal of phosphate, amino acids and vitamins. RPMI 1640 uses a bicarbonate buffering system and requires a 5–10% CO2 atmosphere to maintain physiological pH. Normally, the medium contains no proteins or growth factors, so it is commonly supplemented with 10% fetal bovine serum. Properly supplemented with serum or an adequate serum replacement, RPMI 1640 allows the cultivation of many cell types, especially human lymphocytes, Jurkat cells, HeLa cells, bone marrow cells, hybridomas and carcinomas. Composition Many different formulations exist. Typically, one liter of RPMI 1640 contains: Glucose (2 g) A pH indicator (phenol red, 5 mg) Salts (6 g sodium chloride, 2 g sodium bicarbonate, 1.512 g disodium phosphate, 400 mg potassium chloride, 100 mg magnesium sulfate, and 100 mg calcium nitrate) Amino acids (300 mg glutamine; 200 mg arginine; 50 mg each asparagine, cystine, leucine, and isoleucine; 40 mg lysine hydrochloride; 30 mg serine; 20 mg each aspartic acid, glutamic acid, hydroxyproline, proline, threonine, tyrosine, and valine; 15 mg each histidine, methionine, and phenylalanine; 10 mg glycine; 5 mg tryptophan; and 1 mg reduced glutathione) Vitamins (35 mg i-inositol; 3 mg choline chloride; 1 
https://en.wikipedia.org/wiki/Cannonball%20problem
In the mathematics of figurate numbers, the cannonball problem asks which numbers are both square and square pyramidal. The problem can be stated as: given a square arrangement of cannonballs, for what size squares can these cannonballs also be arranged into a square pyramid. Equivalently, which squares can be represented as the sum of consecutive squares, starting from 1. Formulation as a Diophantine equation When cannonballs are stacked within a square frame, the number of balls is a square pyramidal number; Thomas Harriot gave a formula for this number around 1587, answering a question posed to him by Sir Walter Raleigh on their expedition to America. Édouard Lucas formulated the cannonball problem as a Diophantine equation or Solution Lucas conjectured that the only solutions are N = 1, M = 1, and N = 24, M = 70, using either 1 or 4900 cannon balls. It was not until 1918 that G. N. Watson found a proof for this fact, using elliptic functions. More recently, elementary proofs have been published. Applications The solution N = 24, M = 70 can be used for constructing the Leech lattice. The result has relevance to the bosonic string theory in 26 dimensions. Although it is possible to tile a geometric square with unequal squares, it is not possible to do so with a solution to the cannonball problem. The squares with side lengths from 1 to 24 have areas equal to the square with side length 70, but they cannot be arranged to tile it. Related problems A triangular-pyramid version of the Cannon Ball Problem, which is to yield a perfect square from the Nth Tetrahedral number, would have N = 48. That means that the (24 × 2 = ) 48th tetrahedral number equals to (702 × 22 = 1402 = ) 19600. This is comparable with the 24th square pyramid having a total of 702 cannon balls. Similarly, a pentagonal-pyramid version of the Cannon Ball problem to produce a perfect square, would have N = 8, yielding a total of (14 × 14 = ) 196 cannon balls. The only numbers that are sim
https://en.wikipedia.org/wiki/Surface%20wave%20detection%20by%20animals
Surface wave detection by animals is the process by which animals, such as surface-feeding fish are able to sense and localize prey and other objects on the surface of a body of water by analyzing features of the ripples generated by objects' movement at the surface. Features analyzed include waveform properties such as frequency, change in frequency, and amplitude, and the curvature of the wavefront. A number of different species are proficient in surface wave detection, including some aquatic insects and toads, though most research is done on the topminnow/surface killifish Aplocheilus lineatus. The fish and other animals with this ability spend large amounts of time near the water surface, some just to feed and others their entire lives. Description of behavior Certain species of fish spend a substantial portion of their lives near the surface of the water in order to feed, usually on insects that are struggling at the surface. Species that detect surface waves typically use them to localize such prey. When the hunting posture is assumed (which may be neutral posture) as specific mechanosensitive organ is held in contact with the surface of the water in order that mechanoreceptors can receive surface waves. The animal will wait a small amount of time (typically <1s) before initiating a response towards the prey, should the surface waves perceived fall within the preferred stimulus range. Response towards prey typically follows the pattern orientation towards prey, swimming towards prey, and then prey capture. This ability is sometimes referred to as a sense of "distant touch." Several species have been shown to use surface wave detection for prey capture. Among these are many species of freshwater fish, notably the groups hatchetfish (Gasteropelecidae), freshwater butterflyfish (Pantodontidae), halfbeaks (Hemiramphidae) and killifish (Aplocheilidae)(list from ). For its consistently stellar performance at the task, the topminnow/killfish (both terms are used i
https://en.wikipedia.org/wiki/Gigabit%20Video%20Interface
Gigabit Video Interface (GVIF) is a digital video serial interface developed by Sony in 1996 for high quality uncompressed video transmission from digital video hardware. It is intended primarily for automotive applications. It is compatible with the HDCP encryption system. GVIF transmits uncompressed serial data at speeds up to 1.95 Gbit/s. GVIF transmits over a single differential pair, as a result, the cable is thin. Transmission distances up to ten meters are possible. Utilisation The GVIF bus was used circa 2000 onwards on many mid range vehicles including Land Rover Discovery 3, Range Rover, Lexus models and Toyota Prius. The GVIF bus was primarily for carrying the video signal between the integrated Satellite Navigation unit and the in vehicle display (also known as Head Unit). GVIF bus in these vehicles only carried video despite the specification allowing for data stuffing of other information. External links GVIF.com Sony GVIF overview page GVIF Technology for Automotive Applications Serial buses Physical layer protocols
https://en.wikipedia.org/wiki/E-Science
E-Science or eScience is computationally intensive science that is carried out in highly distributed network environments, or science that uses immense data sets that require grid computing; the term sometimes includes technologies that enable distributed collaboration, such as the Access Grid. The term was created by John Taylor, the Director General of the United Kingdom's Office of Science and Technology in 1999 and was used to describe a large funding initiative starting in November 2000. E-science has been more broadly interpreted since then, as "the application of computer technology to the undertaking of modern scientific investigation, including the preparation, experimentation, data collection, results dissemination, and long-term storage and accessibility of all materials generated through the scientific process. These may include data modeling and analysis, electronic/digitized laboratory notebooks, raw and fitted data sets, manuscript production and draft versions, pre-prints, and print and/or electronic publications." In 2014, IEEE eScience Conference Series condensed the definition to "eScience promotes innovation in collaborative, computationally- or data-intensive research across all disciplines, throughout the research lifecycle" in one of the working definitions used by the organizers. E-science encompasses "what is often referred to as big data [which] has revolutionized science... [such as] the Large Hadron Collider (LHC) at CERN... [that] generates around 780 terabytes per year... highly data intensive modern fields of science...that generate large amounts of E-science data include: computational biology, bioinformatics, genomics" and the human digital footprint for the social sciences. Turing Award winner Jim Gray imagined "data-intensive science" or "e-science" as a "fourth paradigm" of science (empirical, theoretical, computational and now data-driven) and asserted that "everything about science is changing because of the impact of informati
https://en.wikipedia.org/wiki/David%20Reitze
David Howard Reitze (born 6 January 1961) is an American laser physicist who is professor of physics at the University of Florida and served as the scientific spokesman of the Laser Interferometer Gravitational-Wave Observatory (LIGO) experiment in 2007-2011. In August 2011, he took a leave of absence from the University of Florida to be the Executive Director of LIGO, stationed at the California Institute of Technology, Pasadena, California. He obtained his BA in 1983 from Northwestern University, his PhD in physics from the University of Texas at Austin in 1990, and had positions at Bell Communications Research and Lawrence Livermore National Laboratory, before taking his faculty position at the University of Florida. He is a Fellow of the American Physical Society, the Optical Society, and the American Association for the Advancement of Science. An expert in ultrafast optics and laser spectroscopy, he now specialises in laser-based interferometric gravitational wave detection. This includes the development of new interferometer topologies for next generation gravitational wave detectors, investigations of thermal loading in passive and active optical elements, development of high power optical components, and the design, construction and operation of the LIGO interferometers. As Director of the LIGO Laboratory, one of his main efforts has been planning the proposed extension of the LIGO network of detectors to include one in India. In February 2016, he, as executive director of LIGO, announced that the first direct gravitational wave observation had occurred on 14 September 2015 by the LIGO Scientific Collaboration and Virgo Collaboration using the LIGO detectors in Hanford, WA and Livingston, LA. Reitze, along with other former and present spokespersons of the LIGO Scientific Collaboration, was awarded the National Academy of Sciences Award for Scientific Discovery in 2017. In 2019, Reitze co-edited a book with Peter Saulson and Hartmut Grote on the current
https://en.wikipedia.org/wiki/Gertrude%20Neumark
Gertrude Fanny Neumark, also known as Gertrude Neumark Rothschild, (April 29, 1927 – November 11, 2010) was an American physicist, most noted for her work in material science and physics of semiconductors with emphasis on optical and electrical properties of wide-bandgap semiconductors and their light-emitting devices. Personal life She was born in Nuremberg, Germany in 1927. Her family, who were Jewish, left Germany in 1935. Her father Sigmund’s naturalization papers indicate that she arrived with him and her mother Bertha in Miami from Cuba on SS Florida January 3, 1940. He lists Cuba as their last place of residence. She died on November 11, 2010, at age 83, due to heart failure. Education Neumark graduated B.A. summa cum laude (chemistry) from Barnard College in 1948 and M.A. (chemistry) at Radcliffe College in the following year. She completed her Ph.D. in chemistry at Columbia University in 1951; her thesis entitled "Free cloud approximation to molecular orbital calculations". Career Following her PhD, she joined the Sylvania Research Laboratories in Bayside, NY as a senior physicist. In 1960, she moved to the Philips Laboratories, Briarcliff Manor, New York, where she worked until 1985. She was elected a Fellow of the American Physical Society in 1982. From 1982 to 1985, she was visiting-adjunct orofessor of materials science at Columbia University, and became a full professor of materials science there in 1985. In 1999, she became the Howe Professor Emerita of Materials Science and Engineering and professor emerita of applied physics and mathematics at Columbia University. Research and patents In the 1980s, Neumark began studying the optical properties of wide-bandgap semiconductors and developed diodes capable of using the upper range of the spectrum and serving as a superior light source. The new short-wavelength LEDs, emitting blue, green, violet and ultraviolet light, turned out to be much more energy efficient, reliable and long-lived while the s
https://en.wikipedia.org/wiki/1QBit
1QB Information Technologies, Inc. (1QBit) is a quantum computing software company, based in Vancouver, British Columbia. 1QBit was founded on December 1, 2012 and has established hardware partnerships with Microsoft, IBM, Fujitsu and D-Wave Systems. While 1QBit develops general purpose algorithms for quantum computing hardware, the organization is primarily focused on computational finance, materials science, quantum chemistry, and the life sciences. Technology 1QBit has divisions focused on universal quantum computing, advanced AI techniques, cloud based quantum processing, and hardware innovation. 1QBit's 1Qloud platform is focused on optimization including reformulating optimization problems into the quadratic unconstrained binary optimization (QUBO) format necessary to compute with quantum annealing processors and similar devices from organizations such as Fujitsu, D-Wave, Hitachi and NTT, while their QEMIST platform is focused on advanced materials and quantum chemistry research with universal quantum computing processors. History 1QBit was founded as the first dedicated quantum computing software company in 2012. In 2013, 1QBit raised seed funding from US and Canadian angel investors, before closing a Series A financing round led by the Chicago Mercantile Exchange in 2014. On August 5, 2015 the World Economic Forum announced 1QBit as a recipient of the 2015 Technology Pioneer Award recognizing 1QBit as a leader among the world's most promising technology companies. In 2017, 1QBit raised a $45M Series B financing round led by Fujitsu with participation from Allianz, Accenture, The Royal Bank of Scotland and the Chicago Mercantile Exchange. In April 2018, 1QBit joined the IBM Q Network, a global community of leading Fortune 500 companies, academic institutions, startups, and national research labs designed to explore practical applications for quantum computing. In May 2020, 1QBit and its collaborators serving health authorities from East to West obtained
https://en.wikipedia.org/wiki/D-STAR
D-STAR (Digital Smart Technologies for Amateur Radio) is a digital voice and data protocol specification for amateur radio. The system was developed in the late 1990s by the Japan Amateur Radio League and uses minimum-shift keying in its packet-based standard. There are other digital modes that have been adapted for use by amateurs, but D-STAR was the first that was designed specifically for amateur radio. Several advantages of using digital voice modes are that it uses less bandwidth than older analog voice modes such as amplitude modulation and frequency modulation. The quality of the data received is also better than an analog signal at the same signal strength, as long as the signal is above a minimum threshold and as long as there is no multipath propagation. D-STAR compatible radios are available for HF, VHF, UHF, and microwave amateur radio bands. In addition to the over-the-air protocol, D-STAR also provides specifications for network connectivity, enabling D-STAR radios to be connected to the Internet or other networks, allowing streams of voice or packet data to be routed via amateur radio. D-STAR compatible radios are manufactured by Icom, Kenwood, and FlexRadio Systems. History In 1998 an investigation into finding a new way of bringing digital technology to amateur radio was started. The process was funded by a ministry of the Japanese government, then called the Ministry of Posts and Telecommunications, and administered by the Japan Amateur Radio League. In 2001, D-STAR was published as the result of the research. In September 2003 Icom named Matt Yellen, KB7TSE (now K7DN), to lead its US D-STAR development program. Starting in April 2004 Icom began releasing new "D-STAR optional" hardware. The first to be released commercially, was a 2-meter mobile unit designated IC-2200H. Icom followed up with 2 meter and 440 MHz handheld transceivers the next year. However, the yet to be released UT-118 add-on card was required for these radios to operate i