source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Developmental%20Neurobiology
Developmental Neurobiology is a monthly peer-reviewed scientific journal covering all aspects of neural development. It was established in 1969 as Journal of Neurobiology, covering all of neuroscience, but when the scope become more specialized, it obtained its current name in 2007. The journal is published by Wiley-Blackwell. The editor-in-chief is Bin Chen (University of California, Santa Cruz) and the associate editors are Song-Hai Shi (Tsinghua University, Beijing, China) and Andreas Prokop (University of Manchester). Abstracting and indexing The journals is abstracted and indexed in: According to the Journal Citation Reports, the journal has a 2019 impact factor of 3.935.
https://en.wikipedia.org/wiki/Standing%20crop
A standing crop is the total biomass of the living organisms present in a given environment. This includes both natural ecosystems and agriculture. See also Net Primary Production Standing State Bibliography Boudouresque CF (1973) Les peuplements sciaphiles ; Recherches de bionomie analytique, structurale et expérimentale sur les peuplements benthiques sciaphiles de Méditerranée occidentale (fraction algale). Bulletin du Muséum d'histoire naturelle, 33, 147, PDF, 80 pages. Campbell, Reece, Urry, Cain, et al. (2011) 9th ed. Biology. Benjamin Cummings. pg 1221 Fausch, K. D., Hawkes, C. L., & Parsons, M. G. (1988). Models that predict standing crop of stream fish from habitat variables: 1950-85 (http://www.treesearch.fs.fed.us/pubs/8730 résumé]). Jenkins, R. M. (1968). The influence of some environmental factors on standing crop and harvest of fishes in US reservoirs.
https://en.wikipedia.org/wiki/Copac
Copac (originally an acronym of Consortium of Online Public Access Catalogues) was a union catalogue which provided free access to the merged online catalogues of many major research libraries and specialist libraries in the United Kingdom and Ireland, plus the British Library, the National Library of Scotland and the National Library of Wales. It had over 40 million records from around 90 libraries as of 2019, representing a wide range of materials across all subject areas. Copac was freely available to all, and was widely used, with users mainly coming from Higher Education institutions in the United Kingdom, but also worldwide. Copac was valued by users as a research tool. Copac was searchable through with a web browser or Z39.50 client. It was also accessible through OpenURL and Search/Retrieve via URL (SRU) interfaces. These interfaces could be used to provide links to items on Copac from external sites, such as those used on the Institute of Historical Research website. Copac was a Jisc service provided for the UK community on the basis of an agreement with Research Libraries UK (RLUK). The service used records supplied by RLUK members, as well as an increasing range of specialist libraries with collections of national research interest. A full list of contributors is available including the National Trust for Places of Historic Interest or Natural Beauty, the Royal Botanic Gardens, Kew, the Middle Temple library and Institution of Mechanical Engineers (IMechE) Library. In July 2019, Jisc replaced COPAC with Library Hub Discover. See also OPAC SUNCAT Talis Group
https://en.wikipedia.org/wiki/PDL%20BioPharma
PDL BioPharma (known as Protein Design Labs prior to 2006) is a publicly traded American holding company that since 2008 manages patents and other intellectual property that had been generated by the company. In 2008 in response to shareholder pressure, PDL spun out its active development programs to a company called Facet Biotech that it capitalized with $400 million. Protein Design Labs, founded in 1986 by cell biologist Laurence J. Korn and mathematician Cary Queen, was a pioneer in humanizing monoclonal antibodies. Its humanization technology was used in several monoclonal antibody drugs under licensing and collaboration agreements; such products included trastuzumab (Herceptin), bevacizumab (Avastin), and palivizumab (Synagis). It held its initial public offering in 1992 and by 1994 it had 13 compounds in its pipeline, and had partnered its lead product, the humanized monoclonal antibody daclizumab, with Roche under a license that called for a 15% royalty. Daclizumab became the first humanized monoclonal antibody approved for human use in 1997. PDL made another public offering in 2000, at the peak of the biotech IPO window that ran from 1998 to 2001. In 2003 it settled litigation with Genentech over royalty payments owed to PDL under their license agreement; annual royalties were anticipated at that time to be around $90M per year. These are the kinds of intellectual property and license agreements that PDL manages on behalf of its shareholders since 2008; in 2007 those royalty payments were around $220 million. One of the development programs that PDL had transferred to Facet was the use of daclizumab in indications it had never licensed to Roche, including multiple sclerosis. Facet partnered that development program with Biogen Idec which attempted a hostile buy out of Facet for $350M in 2009; Facet rejected that offer and was purchased by Abbvie for $450 million in cash the next year. The FDA approved daclizumab for multiple sclerosis in 2016 under
https://en.wikipedia.org/wiki/Medea%20hypothesis
The Medea hypothesis is a term coined by paleontologist Peter Ward for a hypothesis that contests the Gaian hypothesis and proposes that multicellular life, understood as a superorganism, may be self-destructive or suicidal. The metaphor refers to the mythological Medea (representing the Earth), who kills her own children (multicellular life). In this view, microbial-triggered mass extinctions result in returns to the microbial-dominated state it has been for most of its history. Examples Possible examples of extinction events induced entirely or partially by biotic activities include: The Great Oxidation Event, 2.45 billion years ago, believed to be responsible for the mass poisoning of anaerobic microbes to which oxygen was toxic, and for the Huronian glaciation that resulted from the reaction of methane with oxygen to form carbon dioxide (a less potent greenhouse gas than methane) and subsequent depletion of atmospheric carbon dioxide by aerobic photosynthesisers The Sturtian and Marinoan Snowball Earth glaciations, 715 to 680 and 650 to 632.3 million years ago, respectively, resulting from the sequestration of atmospheric carbon dioxide during the Neoproterozoic Oxygenation Event The Late Ordovician Mass Extinction (LOME), to , suggested by some studies to have been caused by glaciation resulting from carbon dioxide depletion driven by the radiation of land plants Euxinic events, such as during the Great Dying, , and the aforementioned LOME, caused by sulphur-reducing prokaryotes that produce hydrogen sulphide The list excludes the Cretaceous–Paleogene extinction event, since this was, at least partially, externally induced by a meteor impact. Current status and future extinctions Peter Ward proposes that the current man-made climate change and mass extinction event may be considered to be the most recent Medean event. As these events are anthropogenic, he postulates that Medean events are not necessarily caused by microbes, but by intelligent life as w
https://en.wikipedia.org/wiki/Phycosphere
The phycosphere is a microscale mucus region that is rich in organic matter surrounding a phytoplankton cell. This area is high in nutrients due to extracellular waste from the phytoplankton cell and it has been suggested that bacteria inhabit this area to feed on these nutrients. This high nutrient environment creates a microbiome and a diverse food web for microbes such as bacteria and protists. It has also been suggested that the bacterial assemblages within the phycosphere are species-specific and can vary depending on different environmental factors. In terms of comparison, the phycosphere in phytoplankton has been suggested analogous to the rhizosphere in plants, which is the root zone important for nutrient recycling. Both plant roots and phytoplankton exude chemicals which alter their immediate surrounds drastically – including altering the pH and oxygen levels. In terms of community construction, chemotaxis is used in both environments in order to propagate the recruitment of microbes. In the rhizosphere, chemotaxis is used by the host – the plant – to mediate the motility of the soil which allows for microbial colonization. In the phycosphere, the phytoplankton release of specific chemical exudates elicits a response from bacterial symbionts who exhibit chemotaxis signaling, thereby enabling the recruitment of microbes and subsequent colonization. The interfaces also have a few similar microbes, chemicals, and metabolites involved in the host – symbiont interactions. This includes microbes such as Rhizobium, which in the phycospheres of green algae was found to be the foremost microbe when compared to other abundant community members. Chemicals such as dimethylsuloniopropionate (DMSP) and 2,3-dihydroxypropane-1-sulfonate (DHPS) and metabolites such as sugars and amino acids are implicated in the mechanisms of action of both microbiomes. Phytoplankton-bacteria interactions The microscale interactions between the phytoplankton and bacteria are complex.
https://en.wikipedia.org/wiki/No%20Cross%2C%20No%20Crown
No Cross, No Crown is one of the chief works of William Penn, first published in 1669. It was written during Penn's imprisonment in the Tower of London. Summary Penn exhorts believers to adhere to the spirit of Primitive Christianity. The book is divided into two parts, the first dealing with the importance of self-denial in the Christian life and the second gathering a series of references to men through the ages who have written of the importance of self-denial, including "heathen," professed Christians, and "retired, aged, and dying men, being their last and serious reflections, to which no ostentation or worldly interests could induce them." Penn's view of Christianity is intensely spiritual rather than formal, and in passing he defends several practices typical of the Religious Society of Friends (Quakers) including clothing which was not fashionable and speech which addressed royal and commoner alike in the second person singular "thee" and "thou." Scholarly editions A 1931 scholarly edition was edited by Norman Penney.
https://en.wikipedia.org/wiki/Open%20vote%20network
In cryptography, the open vote network (or OV-net) is a secure multi-party computation protocol to compute the boolean-count function: namely, given a set of binary values 0/1 in the input, compute the total count of ones without revealing each individual value. This protocol was proposed by Feng Hao, Peter Ryan, and Piotr Zieliński in 2010. It extends Hao and Zieliński's anonymous veto network protocol by allowing each participant to count the number of veto votes (i.e., input one in a boolean-OR function) while preserving the anonymity of those who have voted. The protocol can be generalized to support a wider range of inputs beyond just the binary values 0 and 1. Description All participants agree on a group with a generator of prime order in which the discrete logarithm problem is hard. For example, a Schnorr group can be used. Assume there are participants. Unlike other secure multi-party computation protocols that typically require pairwise secret and authenticated channels between participants in addition to an authenticated public channel, OV-net only requires an authenticated public channel available to every participant. Such a channel may be realized by using digital signatures. The protocol runs in two rounds. Round 1: each participant selects a random value and publishes the ephemeral public key together with a zero-knowledge proof for the proof of the knowledge of the exponent . Such proofs may be realized by using Schnorr non-interactive zero-knowledge proofs as described in RFC 8235. After this round, each participant computes: Round 2: each participant publishes where is either 0 or 1, together with a 1-out-of-2 zero knowledge proof for the proof that is one of . Such 1-out-of-2 proofs may be realized by using Cramer, Gennaro, and Schoenmakers' zero-knowledge proof technique. After round 2, each participant computes . Note that all values vanish because . The exponent represents the count of ones. As it is usually a small number,
https://en.wikipedia.org/wiki/G%20protein-coupled%20receptor%20kinase%202
G-protein-coupled receptor kinase 2 (GRK2) is an enzyme that in humans is encoded by the ADRBK1 gene. GRK2 was initially called Beta-adrenergic receptor kinase (βARK or βARK1), and is a member of the G protein-coupled receptor kinase subfamily of the Ser/Thr protein kinases that is most highly similar to GRK3(βARK2). Functions G protein-coupled receptor kinases phosphorylate activated G protein-coupled receptors, which promotes the binding of an arrestin protein to the receptor. Arrestin binding to phosphorylated, active receptor prevents receptor stimulation of heterotrimeric G protein transducer proteins, blocking their cellular signaling and resulting in receptor desensitization. Arrestin binding also directs receptors to specific cellular internalization pathways, removing the receptors from the cell surface and also preventing additional activation. Arrestin binding to phosphorylated, active receptor also enables receptor signaling through arrestin partner proteins. Thus the GRK/arrestin system serves as a complex signaling switch for G protein-coupled receptors. GRK2 and the closely related GRK3 phosphorylate receptors at sites that encourage arrestin-mediated receptor desensitization, internalization and trafficking rather than arrestin-mediated signaling (in contrast to GRK5 and GRK6, which have the opposite effect). This difference is one basis for pharmacological biased agonism (also called functional selectivity), where a drug binding to a receptor may bias that receptor’s signaling toward a particular subset of the actions stimulated by that receptor. GRK2 is expressed broadly in tissues, but generally at higher levels than the related GRK3. GRK2 was originally identified as a protein kinase that phosphorylated the β2-adrenergic receptor, and has been most extensively studied as a regulator of adrenergic receptors (and other GPCRs) in the heart, where it has been proposed as a drug target to treat heart failure. Strategies to inhibit GRK2 include us
https://en.wikipedia.org/wiki/Comparative%20Biochemistry%20and%20Physiology%20D
Comparative Biochemistry and Physiology Part D: Genomics and Proteomics is a peer-reviewed scientific journal that covers research in biochemistry and physiology. External links Biochemistry journals Physiology journals Elsevier academic journals
https://en.wikipedia.org/wiki/Semicircular%20canals
The semicircular canals or semicircular ducts are three semicircular, interconnected tubes located in the innermost part of each ear, the inner ear. The three canals are the horizontal, superior and posterior semicircular canals. Structure The semicircular canals are a component of the bony labyrinth that are at right angles from each other. At one end of each of the semicircular canals is a dilated sac called an osseous ampulla, which is more than twice the diameter of the canal. Each ampulla contains an ampullary crest, the crista ampullaris which consists of a thick gelatinous cap called a cupula and many hair cells. The superior and posterior semicircular canals are oriented vertically at right angles to each other. The lateral semicircular canal is about a 30-degree angle from the horizontal plane. The orientations of the canals cause a different canal to be stimulated by movement of the head in different planes, and more than one canal is stimulated at once if the movement is off those planes. The lateral canal detects angular acceleration of the head when the head is turned and the superior and posterior canals detect vertical head movements when the head is moved up or down. When the head changes position, the endolymph in the canals lags behind due to inertia and this acts on the cupula which bends the cilia of the hair cells. The stimulation of the hair cells sends the message to the brain that acceleration is taking place. The ampullae open into the vestibule by five orifices, one of the apertures being common to two of the canals. Among species of mammals, the size of the semicircular canals is correlated with their type of locomotion. Specifically, species that are agile and have fast, jerky locomotion have larger canals relative to their body size than those that move more cautiously. Horizontal semicircular canal The lateral or horizontal canal (external semicircular canal) is the shortest of the three canals. Movement of fluid within this canal c
https://en.wikipedia.org/wiki/Tetrahydrofurfuryl%20acetate
Tetrahydrofurfuryl acetate is an organic chemical compound used for food flavouring and cosmetics. It has a fruity ethereal flavour, also described as honey, maple, or bread-like. It is generally accepted as safe in the USA. Typical levels of use are 2 ppm in drinks, 8 ppm in ice cream, and 20 ppm in baked products and confectionery. Classified as a heterocyclic ester, it is made by reacting tetrahydrofurfuryl alcohol with acetic anhydride. Related flavouring compounds are tetrahydrofurfuryl butyrate, tetrahydrofurfuryl cinnamate, tetrahydrofurfuryl alcohol, and tetrahydrofurfuryl propionate.
https://en.wikipedia.org/wiki/Presidential%20Award%20for%20Excellence%20in%20Mathematics%20and%20Science%20Teaching
The Presidential Award for Excellence in Mathematics and Science Teaching (PAEMST) is the highest recognition that a kindergarten through 12th-grade mathematics or science teacher may receive for outstanding teaching in the United States. Authorized by the Education for Economic Security Act in 1984, this program authorizes the President to bestow up to 108 awards each year. The National Science Foundation (NSF) administers PAEMST on behalf of The White House Office of Science and Technology Policy. Eligibility The following are the eligibility criteria for nominees. They must: Be highly qualified teachers, as deemed by their states, districts, or schools; teachers in private/independent schools should meet the spirit of the principles and provisions of the No Child Left Behind Act, Public Law 107-110. Hold a degree or appropriate credentials in the category for which they are applying. Teach in one of the 50 states or four U.S. jurisdictions. The jurisdictions are Washington, D.C., Puerto Rico, Department of Defense Schools, and the U.S. territories as a group (American Samoa, Guam, the Commonwealth of the Northern Mariana Islands, and the U.S. Virgin Islands). Be full-time employees of the school or school district. Have at least 5 years of mathematics or science teaching experience prior to application. Teach mathematics or science at the appropriate grade level in a public or private school. The award is made to teachers of kindergarten through 6th grades and 7th through 12 grades in alternate years. Not have received the national PAEMST award in any prior competition or category. See also Awards and decorations of the United States government List of mathematics awards
https://en.wikipedia.org/wiki/Scavenger%20receptor%20%28endocrinology%29
Scavenger receptors in endocrinology are inactive membrane receptors which bind certain hormones such as IGF-1 and do not transmit an intracellular response.
https://en.wikipedia.org/wiki/M82%20X-1
M82 X-1 is an ultra-luminous X-ray source located in the galaxy M82. It is a candidate intermediate-mass black hole, with the exact mass estimate varying from around 100 to 1000. One of the most luminous ULXs ever known, its luminosity exceeds the Eddington limit for a stellar mass object. See also M82 X-2
https://en.wikipedia.org/wiki/Data%20and%20object%20carousel
In digital video broadcasting (DVB), a data and object carousel is used for repeatedly delivering data in a continuous cycle. Carousels allow data to be pushed from a broadcaster to multiple receivers by transmitting a data set repeatedly in a standard format. A set-top box receiver may tune to the data stream at any time and is able to reconstitute the data into a virtual file system. The carousel may therefore be considered as a transport file system or file broadcasting system that allows data files to be transmitted from the broadcaster to multiple receivers or clients simultaneously. In a unidirectional broadcast environment, the receiver is unable to request the retransmission of any data that was missed or received incorrectly. Repeated retransmission of data allows the receiver to cope with random tuning to a channel at an unpredictable time, for instance as the user changes a channel. The carousel cycle period generally determines the maximum time required for a receiver to acquire an application or specific datum. It is possible to reduce the access time for commonly used files by broadcasting some data more often than others. An individual object carousel is also called a service domain in some documents. To be precise, a service domain is a group of related DSM-CC objects. In broadcast systems, there is no difference between an object carousel and a service domain except for the terminology: an object carousel is a service domain, and vice versa. Usage and applications Data and object carousels are most commonly used in DVB, which has standards for broadcasting digital television content using carousels. The standard format for a carousel is defined in the Digital Storage Media Command and Control (DSM-CC) toolkit in ISO/IEC 13818-6 and is part of the Digital Audio Video Council (DAVIC) DVB standard for digital video broadcasting. The specification provides support for a variety of communication models, including provision for interactive transport
https://en.wikipedia.org/wiki/Uncle%20Sam
Uncle Sam (which has the same initials as United States) is a common national personification of the federal government of the United States or the country in general. Since the early 19th century, Uncle Sam has been a popular symbol of the U.S. government in American culture and a manifestation of patriotic emotion. Uncle Sam has also developed notoriety for his appearance in military propaganda, popularized by a famous 1917 World War I recruiting poster by J. M. Flagg. According to legend, the character came into use during the War of 1812 and may have been named for Samuel Wilson. The actual origin is obscure. The first reference to Uncle Sam in formal literature (as distinct from newspapers) was in the 1816 allegorical book The Adventures of Uncle Sam, in Search After His Lost Honor by Frederick Augustus Fidfaddy, Esq. While the figure of Uncle Sam specifically represents the government, the female figure of Columbia represents the United States as a nation. An archaic character, Brother Jonathan, was known to represent the American populace. Earlier personifications The earliest known personification of the United States was as a woman named Columbia, who first appeared in 1738 (pre-US) and sometimes was associated with another female personification, Lady Liberty. With the American Revolutionary War of 1775 came Brother Jonathan, a male personification. Brother Jonathan saw full literary development into the personification of American national character through the 1825 novel Brother Jonathan by John Neal. Uncle Sam finally appeared after the War of 1812. Columbia appeared with either Brother Jonathan or Uncle Sam, but her use declined as a national person in favor of Liberty, and she was effectively abandoned once she became the mascot of Columbia Pictures in the 1920s. A March 24, 1810, journal entry by Isaac Mayo (a midshipman in the US Navy) states: weighed anchor stood down the harbor, passed Sandy Hook, where there are two light-houses, and pu
https://en.wikipedia.org/wiki/Inosine%20triphosphate
Inosine triphosphate (ITP) is an intermediate in the purine metabolism pathway, seen in the synthesis of ATP and GTP. It comprises an inosine nucleotide containing three phosphate groups esterified to the sugar moiety. ITP results from deamination of ATP. Incorporation of ITP into the DNA from the nucleotide pool can lead to DNA damage, mutagenesis and other harmful effects. ITP is processed by the enzyme inosine triphosphate pyrophosphatase (ITPA), which turns it into inosine monophosphate (IMP), to avoid incorporation into DNA.
https://en.wikipedia.org/wiki/Zinc%20finger%20protein%20727
Zinc finger protein 727 is a protein that in humans is encoded by the ZNF727 gene.
https://en.wikipedia.org/wiki/View%20model
A view model or viewpoints framework in systems engineering, software engineering, and enterprise engineering is a framework which defines a coherent set of views to be used in the construction of a system architecture, software architecture, or enterprise architecture. A view is a representation of the whole system from the perspective of a related set of concerns. Since the early 1990s there have been a number of efforts to prescribe approaches for describing and analyzing system architectures. These recent efforts define a set of views (or viewpoints). They are sometimes referred to as architecture frameworks or enterprise architecture frameworks, but are usually called "view models". Usually a view is a work product that presents specific architecture data for a given system. However, the same term is sometimes used to refer to a view definition, including the particular viewpoint and the corresponding guidance that defines each concrete view. The term view model is related to view definitions. Overview The purpose of views and viewpoints is to enable humans to comprehend very complex systems, to organize the elements of the problem and the solution around domains of expertise and to separate concerns. In the engineering of physically intensive systems, viewpoints often correspond to capabilities and responsibilities within the engineering organization. Most complex system specifications are so extensive that no single individual can fully comprehend all aspects of the specifications. Furthermore, we all have different interests in a given system and different reasons for examining the system's specifications. A business executive will ask different questions of a system make-up than would a system implementer. The concept of viewpoints framework, therefore, is to provide separate viewpoints into the specification of a given complex system in order to facilitate communication with the stakeholders. Each viewpoint satisfies an audience with interest in a pa
https://en.wikipedia.org/wiki/Twisted%20geometries
Twisted geometries are discrete geometries that play a role in loop quantum gravity and spin foam models, where they appear in the semiclassical limit of spin networks. A twisted geometry can be visualized as collections of polyhedra dual to the nodes of the spin network's graph. Intrinsic and extrinsic curvatures are defined in a manner similar to Regge calculus, but with the generalisation of including a certain type of metric discontinuities: the face shared by two adjacent polyhedra has a unique area, but its shape can be different. This is a consequence of the quantum geometry of spin networks: ordinary Regge calculus is "too rigid" to account for all the geometric degrees of freedom described by the semiclassical limit of a spin network. The name twisted geometry captures the relation between these additional degrees of freedom and the off-shell presence of torsion in the theory, but also the fact that this classical description can be derived from Twistor theory, by assigning a pair of twistors to each link of the graph, and suitably constraining their helicities and incidence relations.
https://en.wikipedia.org/wiki/743%20%28number%29
743 (seven hundred [and] forty three) is the natural number following 742 and preceding 744. It is a prime number. 743 is a Sophie Germain prime, because 2 × 743 + 1 = 1487 is also prime. 743 is an emirp, because 347 (the reversal of its digits) is prime. There are exactly 743 independent sets in a four-dimensional (16 vertex) hypercube graph, and exactly 743 connected cubic graphs with 16 vertices and girth four.
https://en.wikipedia.org/wiki/I%2A
i* (pronounced "i star") or i* framework is a modeling language suitable for an early phase of system modeling in order to understand the problem domain. i* modeling language allows to model both as-is and to-be situations. The name i* refers to the notion of distributed intentionality which underlines the framework. It is an approach originally developed for modelling and reasoning about organizational environments and their information systems composed of heterogeneous actors with different, often competing, goals that depend on each other to undertake their tasks and achieve these goals. It covers both actor-oriented and Goal modeling. i* models answer the question WHO and WHY, not what. In contrast, the UML Use case approach covers only functional goals, with actors directly involved in operations (typically with software). The KAOS approach covers goals of all types but is less concerned with the intentionality of actors. Elements The model describes dependencies among actors. There are four elements to describe them: goal, soft goal, task and resource. The central concept in i* is in fact that of the intentional actor. Organizational actors are viewed as having intentional properties such as goals, beliefs, abilities, and commitments (concept of distributed intentionality). Actors depend on each other for goals to be achieved, tasks to be performed and resources to be furnished. By depending on others, an actor may be able to achieve goals that are difficult or impossible to achieve on its own; on the other hand, an actor becomes vulnerable if the depended-on actors do not deliver. Actors are strategic in the sense that they are concerned about opportunities and vulnerabilities, and seek rearrangement of their environments that would better serve their interests by restructuring intentional relationships. Models i* framework consists of two main modeling components: Strategic Dependency model (SD) An SD model describes a network of dependency relationshi
https://en.wikipedia.org/wiki/Graphophone
The Graphophone was the name and trademark of an improved version of the phonograph. It was invented at the Volta Laboratory established by Alexander Graham Bell in Washington, D.C., United States. Its trademark usage was acquired successively by the Volta Graphophone Company, the American Graphophone Company, the North American Phonograph Company, and finally by the Columbia Phonograph Company (known today as Columbia Records), all of which either produced or sold Graphophones. Research and development It took five years of research under the directorship of Benjamin Hulme, Harvey Christmas, Charles Sumner Tainter and Chichester Bell at the Volta Laboratory to develop and distinguish their machine from Thomas Edison's Phonograph. Among their innovations, the researchers experimented with lateral recording techniques as early as 1881. Contrary to the vertically-cut grooves of Edison Phonographs, the lateral recording method used a cutting stylus that moved from side to side in a "zig zag" pattern across the record. While cylinder phonographs never employed the lateral cutting process commercially, this later became the primary method of phonograph disc recording. Bell and Tainter also developed wax-coated cardboard cylinders for their record cylinder. Edison's grooved mandrel covered with a removable sheet of tinfoil (the actual recording medium) was prone to damage during installation or removal. Tainter received a separate patent for a tube assembly machine to automatically produce the coiled cardboard tube cores of the wax cylinder records. The shift from tinfoil to wax resulted in increased sound fidelity and record longevity. Besides being far easier to handle, the wax recording medium also allowed for lengthier recordings and created superior playback quality. Additionally the Graphophones initially deployed foot treadles to rotate the recordings, then wind-up clockwork drive mechanisms, and finally migrated to electric motors, instead of the manual cr
https://en.wikipedia.org/wiki/Psychomotor%20learning
Psychomotor learning is the relationship between cognitive functions and physical movement. Psychomotor learning is demonstrated by physical skills such as movement, coordination, manipulation, dexterity, grace, strength, speed—actions which demonstrate the fine or gross motor skills, such as use of precision instruments or tools, and walking. Sports and dance are the richest realms of gross psychomotor skills. Behavioral examples include driving a car, throwing a ball, and playing a musical instrument. In psychomotor learning research, attention is given to the learning of coordinated activity involving the arms, hands, fingers, and feet, while verbal processes are not emphasized. Stages of psychomotor development According to Paul Fitts and Michael Posner's three-stage model, when learning psychomotor skills, individuals progress through the cognitive stages, the associative stage, and the autonomic stage. The cognitive stage is marked by awkward slow and choppy movements that the learner tries to control. The learner has to think about each movement before attempting it. In the associative stage, the learner spends less time thinking about every detail, however, the movements are still not a permanent part of the brain. In the autonomic stage, the learner can refine the skill through practice, but no longer needs to think about the movement. Factors affecting psychomotor skills Psychological feedback Amount of practice Task complexity Work distribution Motive-incentive conditions Environmental factors How motor behaviors are recorded The motor cortices are involved in the formation and retention of memories and skills. When an individual learns physical movements, this leads to changes in the motor cortex. The more practiced a movement is, the stronger the neural encoding becomes. A study cited how the cortical areas include neurons that process movements and that these neurons change their behavior during and after being exposed to tasks. Psychomotor le
https://en.wikipedia.org/wiki/Simatic
SIMATIC is a series of programmable logic controller and automation systems, developed by Siemens. Introduced in 1958, the series has gone through four major generations, the latest being the SIMATIC S7 generation. The series is intended for industrial automation and production. The name SIMATIC is a registered trademark of Siemens. It is a portmanteau of “Siemens” and “Automatic”. Function As with other programmable logic controllers, SIMATIC devices are intended to separate the control of a machine from the machine's direct operation, in a more lightweight and versatile manner than controls hard-wired for a specific machine. Early SIMATIC devices were transistor-based, intended to replace relays attached and customized to a specific machine. Microprocessors were introduced in 1973, allowing programs similar to those on general-purpose digital computers to be stored and used for machine control. SIMATIC devices have input and output modules to connect with controlled machines. The programs on the SIMATIC devices respond in real time to inputs from sensors on the controlled machines, and send output signals to actuators on the machines that direct their subsequent operation. Depending on the device and its connection modules, signals may be a simple binary value ("high" or "low") or more complex. More complex inputs, outputs, and calculations were also supported as the SIMATIC line developed. For example, the SIMATIC 505 could handle floating point quantities and trigonometric functions. Product lines Siemens has developed four product lines to date: 1958: SIMATIC Version G 1973: SIMATIC S3 1979: SIMATIC S5 1995: SIMATIC S7 SIMATIC S5 The S5 line was sold in 90U, 95U, 101U, 100U, 105, 110, 115,115U, 135U, and 155U chassis styles. Within each chassis style, several CPUs were available, with varying speed, memory, and capabilities. Some systems provided redundant CPU operation for ultra-high-reliability control, as used in pharmaceutical manufac
https://en.wikipedia.org/wiki/Run-around%20coil
A run-around coil is a type of energy recovery heat exchanger most often positioned within the supply and exhaust air streams of an air handling system, or in the exhaust gases of an industrial process, to recover the heat energy. Generally, it refers to any intermediate stream used to transfer heat between two streams that are not directly connected for reasons of safety or practicality. It may also be referred to as a run-around loop, a pump-around coil or a liquid coupled heat exchanger. Description A typical run-around coil system comprises two or more multi-row finned tube coils connected to each other by a pumped pipework circuit. The pipework is charged with a heat exchange fluid, normally water, which picks up heat from the exhaust air coil and gives up heat to the supply air coil before returning again. Thus heat from the exhaust air stream is transferred through the pipework coil to the circulating fluid, and then from the fluid through the pipework coil to the supply air stream. The use of this system is generally limited to situations where the air streams are separated and no other type of device can be utilised since the heat recovery efficiency is lower than other forms of air-to-air heat recovery. Gross efficiencies are usually in the range of 40 to 50%, but more significantly seasonal efficiencies of this system can be very low, due to the extra electrical energy used by the pumped fluid circuit. The fluid circuit containing the circulating pump also contains an expansion vessel, to accommodate changes in fluid pressure. In addition, there is a fill device to ensure the system remains charged. There are also controls to bypass and shut down the system when not required, and other safety devices. Pipework runs should be as short as possible, and should be sized for low velocities to minimize frictional losses, hence reducing pump energy consumption. It is possible to recover some of this energy in the form of heat given off by the motor if a glan
https://en.wikipedia.org/wiki/Jarl%20Waldemar%20Lindeberg
Jarl Waldemar Lindeberg (4 August 1876, Helsinki – 24 December 1932, Helsinki) was a Finnish mathematician known for work on the central limit theorem. Life and work Lindeberg was son of a teacher at the Helsinki Polytechnical Institute and at an early age showed mathematical talent and interest. The family was well off and later Jarl Waldemar would prefer to be a reader than a full professor. Lindeberg's career centred on the University of Helsinki. His early interests were in partial differential equations and the calculus of variations but from 1920 he worked in probability and statistics. In 1920 he published his first paper on the central limit theorem. His result was similar to that obtained earlier by Lyapunov whose work he did not then know. However, their approaches were quite different; Lindeberg's was based on a convolution argument while Lyapunov used the characteristic function. Two years later Lindeberg used his method to obtain a stronger result: the so-called Lindeberg condition. His work on probability led to him becoming involved in applied fields. He developed what we know as Kendall's τ and he found the first two moments of its sampling distribution. Lindeberg used line transect methods in forestry, and when in 1926 determining the necessary number of transects to obtain a sufficiently precise confidence interval, he seems to have rediscovered Student's t-distribution. The Swedish mathematician Harald Cramér met Lindeberg in 1922. He later recalled this story about Lindeberg and the beautiful farm he owned. "When he was reproached for not being sufficiently active in his scientific work, he said 'Well, I am really a farmer.' And if somebody happened to say that his farm was not properly cultivated, his answer was 'Of course my real job is to be a professor.' I was very fond of him and saw him often during the following years." Lindeberg's work was unknown to Alan Turing, who proved the central limit theorem in his dissertation in 1935. Writi
https://en.wikipedia.org/wiki/Leo%20Esaki
Reona Esaki (江崎 玲於奈 Esaki Reona, born March 12, 1925), also known as Leo Esaki, is a Japanese physicist who shared the Nobel Prize in Physics in 1973 with Ivar Giaever and Brian David Josephson for his work in electron tunneling in semiconductor materials which finally led to his invention of the Esaki diode, which exploited that phenomenon. This research was done when he was with Tokyo Tsushin Kogyo (now known as Sony). He has also contributed in being a pioneer of the semiconductor superlattices. Early life and education Esaki was born in Takaida-mura, Nakakawachi-gun, Osaka Prefecture (now part of Higashiōsaka City) and grew up in Kyoto, near by Kyoto Imperial University and Doshisha University. He first had contact with American culture in . After graduating from the Third Higher School, he studied physics at Tokyo Imperial University, where he had attended Hideki Yukawa's course in nuclear theory in October 1944. Also, he lived through the Bombing of Tokyo while he was at college. Esaki received his B.Sc. and Ph.D. in 1947 and 1959, respectively, from the University of Tokyo (UTokyo). Career Esaki diode From 1947 to 1960, Esaki joined Kawanishi Corporation (now Denso Ten) and Tokyo Tsushin Kogyo (now Sony). Meanwhile, American physicists John Bardeen, Walter Brattain, and William Shockley invented the transistor, which encouraged Esaki to change fields from vacuum tube to heavily-doped germanium and silicon research in Sony. One year later, he recognized that when the PN junction width of germanium is thinned, the current-voltage characteristic is dominated by the influence of the tunnel effect and, as a result, he discovered that as the voltage is increased, the current decreases inversely, indicating negative resistance. This discovery was the first demonstration of solid tunneling effects in physics, and it was the birth of new electronic devices in electronics called Esaki diode (or tunnel diode). He received a doctorate degree from UTokyo due to this
https://en.wikipedia.org/wiki/Universal%20Exposition%20of%20Wines%20and%20Spirits
The Exposition Universelle des Vins et Spiritueux was a museum dedicated to alcoholic beverages on the island of Bendor. The island is in the Var department in the Provence-Alpes-Côte d'Azur in South Western France. The museum is open in July and August, and admission is free of charge. It is closed on Wednesdays. Mission The collections shows the evolution of brands and drinks throughout time and the cultural influence of wines and spirits. The museum was created to keep a “complete and permanent encyclopaedia of wine and spirits”. At the opening of the museum Ricard pledged that "This is not a simple event, it is permanent, eternal, and its expansion will never stop". History The museum was opened on the 12 July 1958 by the owner of the Île de Bendor, industrialist Paul Ricard (1909-1997). Ricard was the eponymous founder of Ricard, a manufacturer of pastis. The museum has two directors, Anistatia Miller and Jared Brown. In 2007 the museum underwent conservation, carried out by a team of experts and volunteers. The conservation process discovered more than 5000 wine and spirit labels and more than 1000 menus and drinks lists dating back to 1860. The 2007 conservation also discovered dissertations on viniculture and distillation, which had been presented at the EUVS during the 1960s. The papers had been presented before modern technological advancements in spirit production. The museum possesses a collection of 8,000 bottles of wines and spirits from around the world. The museum also includes a collection of alcoholic ephemera dating from the 1860s including glassware, crystal, labels, restaurant menus and wine lists. The museum is also the site for conferences on wines and spirits. The EUVS hosts a freely accessible online library with a database of rare bottles and searchable library of vintage menus, drinks lists, distillation books, and cocktail books. Art The entrance to the museum is guarded by statues of Bacchus and Vigne, which were carved by the Pro
https://en.wikipedia.org/wiki/Dipleurula
Dipleurula is a hypothetical larva of the ancestral echinoderm. It represents the type of basis of all larval forms of, at least, the eleutherozoans (all echinoderms except crinoids), where the starfish, sea urchins, sea cucumbers and brittle stars belong. The dipleurula is a bilaterally symmetrical, ciliated echinoderm larva (cilia devoted to movement, feeding and perception). Etymology Derives from Ancient Greek di, meaning 'two' and the small form of pleura (pleurula), meaning 'little side', i.e. 'little, two-sided [larva]'. History Although the term dipleurula stems from Semon (1888), it was first systematically introduced, described and drawn by Bather (1900) in his monography on the echinoderms. The name dipleurula, two-sided, was given to stress the fact that the larva of the typically five-rayed, (approximately) radially symmetric adults show a bilateral structure. It was this bilateral structure of the larvae that identified echinoderms as bilaterian animals. The original doliolaria schema shows a benthic, crawling, larva. However Bather could not have known yet that larval echinoderms are typically pelagic (free-floating plankton). Structure The hypothetical dipleurula larva bears resemblance to stages of all extant echinoderms, such as the bipinnaria and the brachiolaria of the starfish, the auricularia of the sea cucumbers, the echinopluteus larva of the sea urchins, and the ophiopluteus of the brittle stars. Also the doliolaria of the crinoids (sea-lilies and feather stars) can be attributed to the same basic pattern. Note, that the extant echinoderms represent just a small window on the extraordinary diversity of early echinoderms as known from their rich fossil record. Although there is current research on fossilized larval skeletons, the doliolaria remains a hypothesis. The doliolaria shows a three-partite body, each of which develop paired coelomic spaces which originate from the enterocoel. This, and their close resemblance to the torn
https://en.wikipedia.org/wiki/HD%20domain
In molecular biology, the HD domain is a conserved protein domain, named after the conserved histidine (H) and/or aspartate (D) amino acid residues. It is found in a superfamily of enzymes with a predicted or known phosphohydrolase activity. These enzymes appear to be involved in nucleic acid metabolism, signal transduction and possibly other functions in bacteria, archaea and eukaryotes. The fact that all the highly conserved residues in the HD superfamily are histidines or aspartates suggests that coordination of divalent cations is essential for the activity of these proteins.
https://en.wikipedia.org/wiki/Sharkbook
Sharkbook is a global database for identifying and tracking sharks, particularly whale sharks, using uploaded photos and videos.In addition to identifying and tracking sharks, the site allows people to "adopt a shark" and get updates on specific animals. Creation Sharkbook is the result of collaboration between Simon J Pierce of the Marine Megafauna Foundation and Jason Holmberg of Wild Me. The software is Open Source and is now being used by other biology projects. Identification of individual sharks Whale sharks have unique spot patterning on their sides, similar to a human fingerprint, which allows for individual identification. Scuba divers around the world can photograph sharks and upload their identification photographs to the Sharkbook website, supporting global research and conservation efforts. Additionally, the software automatically searches social media sites like YouTube and Instagram to look for images of whale sharks and adds them to the database. Sharkbook software uses special pattern-matching software to identify the unique spots on each shark. This software and algorithms were originally adapted from NASA star tracking software used on the Hubble Space Telescope. This software uses a scale-invariant feature transform (SIFT) algorithm, which can cope with complications presented by highly variable spot patterns and low contrast photographs. Purpose This citizen science tool is free to use by researchers worldwide. Sharkbook represents a global initiative to centralize shark sightings and facilitate research on these vulnerable species. See also Manta Matcher - For Manta Rays Flukebook - For whales and dolphins
https://en.wikipedia.org/wiki/Hygroscopy
Hygroscopy is the phenomenon of attracting and holding water molecules via either absorption or adsorption from the surrounding environment, which is usually at normal or room temperature. If water molecules become suspended among the substance's molecules, adsorbing substances can become physically changed, e.g. changing in volume, boiling point, viscosity or some other physical characteristic or property of the substance. For example, a finely dispersed hygroscopic powder, such as a salt, may become clumpy over time due to collection of moisture from the surrounding environment. Deliquescent materials are sufficiently hygroscopic that they absorb so much water that they become liquid and form an aqueous solution. Hygroscopy is essential for many plant and animal species' attainment of hydration, nutrition, reproduction and/or seed dispersal. Biological evolution created hygroscopic solutions for water harvesting, filament tensile strength, bonding and passive motion – natural solutions being considered in future biomimetics. Etymology and pronunciation The word hygroscopy () uses combining forms of hygro- and -scopy. Unlike any other -scopy word, it no longer refers to a viewing or imaging mode. It did begin that way, with the word hygroscope referring in the 1790s to measuring devices for humidity level. These hygroscopes used materials, such as certain animal hairs, that appreciably changed shape and size when they became damp. Such materials were then said to be hygroscopic because they were suitable for making a hygroscope. Eventually, the word hygroscope ceased to be used for any such instrument in modern usage, but the word hygroscopic (tending to retain moisture) lived on, and thus also hygroscopy (the ability to do so). Nowadays an instrument for measuring humidity is called a hygrometer (hygro- + -meter). History Early hygroscopy literature began circa 1880. Studies by Victor Jodin (Annales Agronomiques, October 1897) focused on the biological proper
https://en.wikipedia.org/wiki/Project%20Coast
Project Coast was a 1980s top-secret chemical and biological weapons (CBW) program instituted by the apartheid-era government of South Africa. Project Coast was the successor to a limited postwar CBW program, which mainly produced the lethal agents CX powder and mustard gas, as well as non-lethal tear gas for riot control purposes. The program was headed by the cardiologist Wouter Basson, who was also the personal physician of South African Prime Minister P. W. Botha. History From 1975 onwards, the South African Defence Force (SADF) found itself embroiled in conventional battles in Angola as a result of the South African Border War. The perception that its enemies had access to battlefield chemical and biological weapons (CBW) led South Africa to begin expanding its own program, initially as a defensive measure and by carrying out research on vaccines. As the years went on, research shifted to offensive uses. In 1981, President P. W. Botha ordered the SADF to develop CBW technology for use against South Africa's enemies. In response, the head of the South African Medical Service division, which was responsible for defensive CBW capabilities, hired Wouter Basson, a cardiologist, to visit other countries and report back on their respective CBW capabilities. He returned with the recommendation that South Africa's program be expanded. In 1983, Project Coast was formed, with Basson at its head. To hide the program and its procurement of CBW-related substances, Project Coast formed four front companies: Delta G Scientific Company, Roodeplaat Research Laboratories, Protechnik and Infladel. Ben Raubenheimer was appointed as CEO. Project Coast created a progressively larger variety of lethal offensive CBW toxins and biotoxins, in addition to the defensive measures. Initially, they were intended for use by the military in combat as a last resort. To that end, they copied Soviet techniques and designed devices that looked like ordinary objects but had the capability to poiso
https://en.wikipedia.org/wiki/Ernesto%20Illy
Ernesto Illy (July 18, 1925 – February 3, 2008) was an Italian food chemist and businessman, known as the chairman of the Illycaffè S.p.A. coffee manufacturer. He was deeply and widely respected, and is considered "an absolute giant" in the world of coffee and espresso, He spearheaded Illy's research into coffee and espresso quality, as summarized in his son Andrea's book, Espresso: The Chemistry of Quality, which was written with Ernesto's assistance. Biography Born in Trieste, Illy was the son of the Hungarian-Italian chocolate maker Francesco Illy, a World War I officer, who had established the espresso making company in 1933. Ernesto studied chemistry at University of Bologna (1947) and immediately joined Illy, and later took over the company from his father (1956). Ernesto created a high-tech manufacturing process, specializing in the making of espresso coffee, for which he also was an evangelist. He was chairman of Illy 1963-2004 and then honorary chairman. Ernesto Illy took part in the establishment of Association Scientifique Internationale pour le Café (chairman, 1991-). He was married to Anna Rossi. The Illy company is run by their son Andrea Illy (born 1964), and among the board members are the daughter Anna Illy and son Riccardo Illy (born 1955), who has also been mayor of Trieste and President of Friuli-Venezia Giulia region. Their first son, Francesco, is also involved. Awards Cavaliere del Lavoro of Italy (Knight of Industry, 1994) honorary doctorates and master's degrees (such as University of Udine, 2005) Centromarca brand organization (president, 1996-) Specialty Coffee Association of America (S.C.A.A.). "Lifetime Achievement Award" (1997) Monte Carmelo (Brazil) county (honorary citizenship, 2002) International Coffee Organization, London (chairman of the promotion committee, 2002-) Anacafè member of the order, Flor del Cafè (Guatemala, 2004).
https://en.wikipedia.org/wiki/Vectors%20in%20gene%20therapy
Gene therapy utilizes the delivery of DNA into cells, which can be accomplished by several methods, summarized below. The two major classes of methods are those that use recombinant viruses (sometimes called biological nanoparticles or viral vectors) and those that use naked DNA or DNA complexes (non-viral methods). Viruses All viruses bind to their hosts and introduce their genetic material into the host cell as part of their replication cycle. This genetic material contains basic 'instructions' of how to produce more copies of these viruses, hacking the body's normal production machinery to serve the needs of the virus. The host cell will carry out these instructions and produce additional copies of the virus, leading to more and more cells becoming infected. Some types of viruses insert their genome into the host's cytoplasm, but do not actually enter the cell. Others penetrate the cell membrane disguised as protein molecules and enter the cell. There are two main types of virus infection: lytic and lysogenic. Shortly after inserting its DNA, viruses of the lytic cycle quickly produce more viruses, burst from the cell and infect more cells. Lysogenic viruses integrate their DNA into the DNA of the host cell and may live in the body for many years before responding to a trigger. The virus reproduces as the cell does and does not inflict bodily harm until it is triggered. The trigger releases the DNA from that of the host and employs it to create new viruses. Retroviruses The genetic material in retroviruses is in the form of RNA molecules, while the genetic material of their hosts is in the form of DNA. When a retrovirus infects a host cell, it will introduce its RNA together with some enzymes, namely reverse transcriptase and integrase, into the cell. This RNA molecule from the retrovirus must produce a DNA copy from its RNA molecule before it can be integrated into the genetic material of the host cell. The process of producing a DNA copy from an RNA molecul
https://en.wikipedia.org/wiki/Generalized%20Procrustes%20analysis
Generalized Procrustes analysis (GPA) is a method of statistical analysis that can be used to compare the shapes of objects, or the results of surveys, interviews, or panels. It was developed for analysing the results of free-choice profiling, a survey technique which allows respondents (such as sensory panelists) to describe a range of products in their own words or language. GPA is one way to make sense of free-choice profiling data; other ways can be multiple factor analysis (MFA), or the STATIS method. The method was first published by J. C. Gower in 1975. Generalized Procrustes analysis estimates the scaling factor applied to respondent scale usage, generating a weighting factor that is used to compensate for individual scale usage differences. Unlike measures such as a principal component analysis, GPA uses individual level data and a measure of variance is utilized in the analysis. The Procrustes distance provides a metric to minimize in order to superimpose a pair of shape instances annotated by landmark points. GPA applies the Procrustes analysis method to superimpose a population of shapes instead of only two shape instances. The algorithm outline is the following: arbitrarily choose a reference shape (typically by selecting it among the available instances) superimpose all instances to current reference shape compute the mean shape of the current set of superimposed shapes if the Procrustes distance between the mean shape and the reference is above a certain threshold, set the reference to mean shape and continue to step 2. See also Procrustes analysis Orthogonal Procrustes problem
https://en.wikipedia.org/wiki/Social%20influences%20on%20fitness%20behavior
Physical fitness is maintained by a range of physical activities. Physical activity is defined by the World Health Organization as "any bodily movement produced by skeletal muscles that requires energy expenditure." Human factors and social influences are important in starting and maintaining such activities. Social environments can influence motivation and persistence, through pressures towards social conformity. Obesity Obesity is a physical marker of poor health, increasing the likelihood of various diseases. Due to social constructs surrounding health, the belief that being skinny is healthy and discrimination against those perceived to be 'unhealthy', people who are considered overweight or obese on the BMI scale face many social challenges. Challenges can range from basic things such as buying clothes, pressure from society to change their body, and being unable to get a job. This can lead to various problems such as eating disorders, self-esteem issues, and misdiagnosis and improper treatment of physical ailments due to discrimination. People who are obese are also less likely to seek medical care than people who are not obese, even if their weight is caused by medical problems. Adult Children Obesity can lower mood and lower self-esteem. Reasons for inactivity In the US, only 26% of adults engage in vigorous leisure-time activity (which includes a sport) or exercising three or more times per week. In an effort to increase adult involvement and decrease the percentage of adult inactivity, the US Department of Health and Human Services has set a national health objective for 2010 that hopes to "Reduce the prevalence of no leisure time activity from more than 25 percent to 20 percent of US adults" (Berlin, Storti, and Brach 1137). In Australia, the Australian Bureau of Statistics found that in 2011/12 adults spent an average of 33 minutes per day doing physical activity with 60% of the population doing less than 30 minutes and fewer than 20% doing an hour
https://en.wikipedia.org/wiki/Johann%20Friedrich%20Hennert
Johann Friedrich Hennert (19 October 1733 – 30 March 1813) was German-born and lectured in mathematics and physics at the University of Utrecht. He was a significant student of Leonhard Euler. He was known for his inclination towards the British school of philosophy. Work Hennert held the chair of mathematics at the University of Utrecht until 1805. Hennert was an important figure in the history of Dutch mathematics. He wrote a number of textbooks on differential calculus. Jan van Swinden was one of his most important students.
https://en.wikipedia.org/wiki/Audio%20normalization
Audio normalization is the application of a constant amount of gain to an audio recording to bring the amplitude to a target level (the norm). Because the same amount of gain is applied across the entire recording, the signal-to-noise ratio and relative dynamics are unchanged. Normalization is one of the functions commonly provided by a digital audio workstation. Two principal types of audio normalization exist. Peak normalization adjusts the recording based on the highest signal level present in the recording. Loudness normalization adjusts the recording based on perceived loudness. Normalization differs from dynamic range compression, which applies varying levels of gain over a recording to fit the level within a minimum and maximum range. Normalization adjusts the gain by a constant value across the entire recording. Peak normalization One type of normalization is peak normalization, wherein the gain is changed to bring the highest PCM sample value or analog signal peak to a given levelusually 0 dBFS, the loudest level allowed in a digital system. Since it searches only for the highest level, peak normalization alone does not account for the apparent loudness of the content. As such, peak normalization is generally used to change the volume in such a way to ensure optimal use of available dynamic range during the mastering stage of a digital recording. When combined with compression/limiting, however, peak normalization becomes a feature that can provide a loudness advantage over non–peak-normalized material. This feature of digital-recording systems, compression and limiting followed by peak normalization, enables contemporary trends in program loudness. Loudness normalization Another type of normalization is based on a measure of loudness, wherein the gain is changed to bring the average loudness to a target level. This average may be approximate, such as a simple measurement of average power (e.g. RMS), or more accurate, such as a measure that addresses h
https://en.wikipedia.org/wiki/Open%20biopsy
An open biopsy is a procedure in which a surgical incision (cut) is made through the skin to expose and remove tissues. The biopsy tissue is examined under a microscope by a pathologist. An open biopsy may be done in the doctor's office or hospital, and may use local anesthesia or general anesthesia. A lumpectomy to remove a breast tumor is a type of open biopsy.
https://en.wikipedia.org/wiki/Extensions%20of%20First%20Order%20Logic
Extensions of First Order Logic is a book on mathematical logic. It was written by María Manzano, and published in 1996 by the Cambridge University Press as volume 19 of their book series Cambridge Tracts in Theoretical Computer Science. Topics The book concerns forms of logic that go beyond first-order logic, and in particular (following the work of Leon Henkin) the project of unifying them by translating all of these extensions into a specific form of logic, many-sorted logic. Beyond many-sorted logic, its topics include second-order logic (including its incompleteness and relation with Peano arithmetic), second-order arithmetic, type theory (in relational, functional, and equational forms), modal logic, and dynamic logic. It is organized into seven chapters. The first concerns second-order logic in its standard form, and it proves several foundational results for this logic. The second chapter introduces the sequent calculus, a method of making sound deductions in second-order logic, and its incompleteness. The third continues the topic of second-order logic, showing how to formulate Peano arithmetic in it, and using Gödel's first incompleteness theorem to provide a second proof of incompleteness of second-order logic. Chapter four formulates a non-standard semantics for second-order logic (from Henkin), in which quantification over relations is limited to only the definable relations. It defines this semantics in terms of "second-order frames" and "general structures", constructions that will be used to formulate second-order concepts within many-sorted logic. In the fifth chapter, the same concepts are used to give a non-standard semantics to type theory. After these chapters on other types of logic, the final two chapters introduce many-sorted logic, prove its soundness, completeness, and compactness, and describe how to translate the other forms of logic into it. Audience and reception Although the book is intended as a textbook for advanced undergraduates
https://en.wikipedia.org/wiki/Diffusion%20capacitance
Diffusion Capacitance is the capacitance that happens due to transport of charge carriers between two terminals of a device, for example, the diffusion of carriers from anode to cathode in a forward biased diode or from emitter to base in a forward-biased junction of a transistor. In a semiconductor device with a current flowing through it (for example, an ongoing transport of charge by diffusion) at a particular moment there is necessarily some charge in the process of transit through the device. If the applied voltage changes to a different value and the current changes to a different value, a different amount of charge will be in transit in the new circumstances. The change in the amount of transiting charge divided by the change in the voltage causing it is the diffusion capacitance. The adjective "diffusion" is used because the original use of this term was for junction diodes, where the charge transport was via the diffusion mechanism. See Fick's laws of diffusion. To implement this notion quantitatively, at a particular moment in time let the voltage across the device be . Now assume that the voltage changes with time slowly enough that at each moment the current is the same as the DC current that would flow at that voltage, say (the quasistatic approximation). Suppose further that the time to cross the device is the forward transit time . In this case the amount of charge in transit through the device at this particular moment, denoted , is given by . Consequently, the corresponding diffusion capacitance:. is . In the event the quasi-static approximation does not hold, that is, for very fast voltage changes occurring in times shorter than the transit time , the equations governing time-dependent transport in the device must be solved to find the charge in transit, for example the Boltzmann equation. That problem is a subject of continuing research under the topic of non-quasistatic effects. See Liu , and Gildenblat et al. Notes
https://en.wikipedia.org/wiki/161%20%28number%29
161 (one hundred [and] sixty-one) is the natural number following 160 and preceding 162. In mathematics 161 is the sum of five consecutive prime numbers: 23, 29, 31, 37, and 41 161 is a hexagonal pyramidal number. 161 is a semiprime. Since its prime factors 7 and 23 are both Gaussian primes, 161 is a Blum integer. 161 is a palindromic number is a commonly used rational approximation of the square root of 5 and is the closest fraction with denominator <300 to that number. In the military was a U.S. Navy Type T2 tanker during World War II was a U.S. Navy during World War II was a U.S. Navy Trefoil-class concrete barge during World War II was a U.S. Navy during World War II was a U.S. Navy during World War II was a U.S. Navy during World War II was a U.S. Navy wooden yacht during World War I was a U.S. Navy during World War II was a U.S. Navy Achomawi-class fleet ocean tug following World War II was a U.S. Navy fourth-group S-class submarine between 1920 and 1931 is a fictional U.S. Navy diesel engine submarine featured in the 1996 film Down Periscope The 161st Intelligence Squadron unit of the Kansas Air National Guard. Its parent unit is the 184th Intelligence Wing In music The Bose 161 Speaker System (2001) The Kay K-161 ThinTwin guitar In transportation MTA Maryland commuter bus 161 New Jersey Bus Route 161 London Bus route 161 In other fields 161 is also: The year AD 161 or 161 BC 161 AH is a year in the Islamic calendar that corresponds to 777 – 778 CE 161 Athor is an M-type Main belt asteroid E.161 is an ITU-T assigns letters to the 12-key telephone keypad Fiorina Fury 161 is a foundry facility and penal colony from the film Alien 3 161 is used by Anti Fascist Action as a code for AFA (A=1, F=6, by order of the alphabet), sometimes used in 161>88 (88 is code for Heil Hitler among neo-nazis, as H=8) See also Anti-Fascist Action List of highways numbered 161 United Nations Security Council Resolution 161 Unite
https://en.wikipedia.org/wiki/Pseudo-intersection
In mathematical set theory, a pseudo-intersection of a family of sets is an infinite set S such that each element of the family contains all but a finite number of elements of S. The pseudo-intersection number, sometimes denoted by the fraktur letter 𝔭, is the smallest size of a family of infinite subsets of the natural numbers that has the strong finite intersection property but has no pseudo-intersection.
https://en.wikipedia.org/wiki/Guanacaste%20hummingbird
The guanacaste hummingbird or Alfero's hummingbird (Amazilia alfaroana) is a possibly extinct species of hummingbird known only from a holotype collected in 1895 at the Miravalles Volcano in Costa Rica. Taxonomy It is usually treated as a subspecies of the Indigo-capped hummingbird or a hybrid between two unknown hummingbird species, but analysis of the holotype suggests it is its own species. Conservation It is possibly extinct, but the ecological stability of the area where the specimen was found indicates a possible undiscovered population still existing. The IUCN classifies it as critically endangered.
https://en.wikipedia.org/wiki/Cerebral%20organoid
A cerebral organoid, or brain organoid, describes an artificially grown, in vitro, miniature organ resembling the brain. Cerebral organoids are created by culturing pluripotent stem cells in a three-dimensional rotational bioreactor, and they develop over a course of months. The brain is an extremely complex system of heterogeneous tissues and consists of a diverse array of neurons. This complexity has made studying the brain and understanding how it works a difficult task in neuroscience, especially when it comes to neurodegenerative diseases. The purpose of creating an in vitro neurological model is to study these diseases in a more simple and variable space. This 3D model is free of many potential in vivo limitations. The varying physiology between human and other mammalian models limits the scope of study in neurological disorders. Cerebral organoids are synthesized tissues that contain several types of nerve cells and have anatomical features that recapitulate regions of the cortex observed in brains. Cerebral organoids are most similar to layers of neurons called the cortex and choroid plexus. In some cases, structures similar to the retina, meninges and hippocampus can form. Stem cells have the potential to grow into many different types of tissues, and their fate is dependent on many factors. Below is an image showing some of the chemical factors that can lead stem cells to differentiate into various neural tissues; a more in-depth table of generating specific organoid identity has been published since. Similar techniques are used on stem cells used to grow cerebral organoids. Model development Using human pluripotent stem cells to create in vitro cerebral organoids allows researchers to summarize current developmental mechanisms for human neural tissue as well as study the roots of human neurological diseases. Cerebral organoids are an investigative tool used to understand how disease pathology works. These organoids can be used in experiments that curr
https://en.wikipedia.org/wiki/Time-compressed%20speech
Time-compressed speech refers to an audio recording of verbal text in which the text is presented in a much shorter time interval than it would through normally-paced real time speech. The basic purpose is to make recorded speech contain more words in a given time, yet still be understandable. For example: a paragraph that might normally be expected to take 20 seconds to read, might instead be presented in 15 seconds, which would represent a time-compression of 25% (5 seconds out of 20). The term "time-compressed speech" should not be confused with "speech compression", which controls the volume range of a sound, but does not alter its time envelope. Methods While some voice talents are capable of speaking at rates significantly in excess of general norms, the term "time-compressed speech" most usually refers to examples in which the time-reduction has been accomplished through some form of electronic processing of the recorded speech. In general, recorded speech can be electronically time-compressed by: increasing its speed (linear compression); removing silences (selective editing); a combination of the two (non-linear compression). The speed of a recording can be increased, which will cause the material to be presented at a faster rate (and hence in a shorter amount of time), but this has the undesirable side-effect of increasing the frequency of the whole passage, raising the pitch of the voices, which can reduce intelligibility. There are normally silences between words and sentences, and even small silences within certain words, both of which can be reduced or removed ("edited-out") which will also reduce the amount of time occupied by the full speech recording. However, this can also have the effect of removing verbal "punctuation" from the speech, causing words and sentences to run together unnaturally, again reducing intelligibility. Vowels are typically held a minimum of 20 milliseconds, over many cycles of the fundamental pitch. DSP systems can
https://en.wikipedia.org/wiki/Bernstein%E2%80%93Kushnirenko%20theorem
The Bernstein–Kushnirenko theorem (or Bernstein–Khovanskii–Kushnirenko (BKK) theorem), proven by David Bernstein and in 1975, is a theorem in algebra. It states that the number of non-zero complex solutions of a system of Laurent polynomial equations is equal to the mixed volume of the Newton polytopes of the polynomials , assuming that all non-zero coefficients of are generic. A more precise statement is as follows: Statement Let be a finite subset of Consider the subspace of the Laurent polynomial algebra consisting of Laurent polynomials whose exponents are in . That is: where for each we have used the shorthand notation to denote the monomial Now take finite subsets of , with the corresponding subspaces of Laurent polynomials, Consider a generic system of equations from these subspaces, that is: where each is a generic element in the (finite dimensional vector space) The Bernstein–Kushnirenko theorem states that the number of solutions of such a system is equal to where denotes the Minkowski mixed volume and for each is the convex hull of the finite set of points . Clearly, is a convex lattice polytope; it can be interpreted as the Newton polytope of a generic element of the subspace . In particular, if all the sets are the same, then the number of solutions of a generic system of Laurent polynomials from is equal to where is the convex hull of and vol is the usual -dimensional Euclidean volume. Note that even though the volume of a lattice polytope is not necessarily an integer, it becomes an integer after multiplying by . Trivia Kushnirenko's name is also spelt Kouchnirenko. David Bernstein is a brother of Joseph Bernstein. Askold Khovanskii has found about 15 different proofs of this theorem.
https://en.wikipedia.org/wiki/Rod%20calculus
Rod calculus or rod calculation was the mechanical method of algorithmic computation with counting rods in China from the Warring States to Ming dynasty before the counting rods were increasingly replaced by the more convenient and faster abacus. Rod calculus played a key role in the development of Chinese mathematics to its height in Song Dynasty and Yuan Dynasty, culminating in the invention of polynomial equations of up to four unknowns in the work of Zhu Shijie. Hardware The basic equipment for carrying out rod calculus is a bundle of counting rods and a counting board. The counting rods are usually made of bamboo sticks, about 12 cm- 15 cm in length, 2mm to 4 mm diameter, sometimes from animal bones, or ivory and jade (for well-heeled merchants). A counting board could be a table top, a wooden board with or without grid, on the floor or on sand. In 1971 Chinese archaeologists unearthed a bundle of well-preserved animal bone counting rods stored in a silk pouch from a tomb in Qian Yang county in Shanxi province, dated back to the first half of Han dynasty (206 BC – 8AD). In 1975 a bundle of bamboo counting rods was unearthed. The use of counting rods for rod calculus flourished in the Warring States, although no archaeological artefacts were found earlier than the Western Han Dynasty (the first half of Han dynasty; however, archaeologists did unearth software artefacts of rod calculus dated back to the Warring States); since the rod calculus software must have gone along with rod calculus hardware, there is no doubt that rod calculus was already flourishing during the Warring States more than 2,200 years ago. Software The key software required for rod calculus was a simple 45 phrase positional decimal multiplication table used in China since antiquity, called the nine-nine table, which were learned by heart by pupils, merchants, government officials and mathematicians alike. Rod numerals Displaying numbers Rod numerals is the only numeric system that uses
https://en.wikipedia.org/wiki/Lentinula%20novae-zelandiae
Lentinula novae-zelandiae, also known as New Zealand shiitake, is a species of edible saprobic fungus endemic to New Zealand. Phylogenetic research suggests this species forms a monophyletic clade of Laurasian origins. It can be cultivated, with cultures and grow kits available commercially in New Zealand.
https://en.wikipedia.org/wiki/Decimal64%20floating-point%20format
In computing, decimal64 is a decimal floating-point computer numbering format that occupies 8 bytes (64 bits) in computer memory. It is intended for applications where it is necessary to emulate decimal rounding exactly, such as financial and tax computations. Decimal64 supports 16 decimal digits of significand and an exponent range of −383 to +384, i.e. to . (Equivalently, to .) In contrast, the corresponding binary format, which is the most commonly used type, has an approximate range of to . Because the significand is not normalized, most values with less than 16 significant digits have multiple possible representations; , etc. Zero has 768 possible representations (1536 if both signed zeros are included). Decimal64 floating point is a relatively new decimal floating-point format, formally introduced in the 2008 version of IEEE 754 as well as with ISO/IEC/IEEE 60559:2011. Representation of decimal64 values IEEE 754 allows two alternative representation methods for decimal64 values. The standard does not specify how to signify which representation is used, for instance in a situation where decimal64 values are communicated between systems: In the binary representation method, the 16-digit significand is represented as a binary coded positive integer, based on binary integer decimal (BID). In the decimal representation method, the 16-digit significand is represented as a decimal coded positive integer, based on densely packed decimal (DPD) with 5 groups of 3 digits (except the most significant digit encoded specially) are each represented in declets (10-bit sequences). This is pretty efficient, because 210 = 1024, is only little more than needed to still contain all numbers from 0 to 999. Both alternatives provide exactly the same range of representable numbers: 16 digits of significand and possible decimal exponent values. (All the possible decimal exponent values storable in a binary64 number are representable in decimal64, and most bits of the sig
https://en.wikipedia.org/wiki/Robert%20Pollack%20%28biologist%29
Robert Elliot Pollack is an American biologist whose interests cross many academic lines. He grew up in Brooklyn, attended public schools, and majored in physics at Columbia University, where he graduated from the College in 1961. He received a PhD in Biological Sciences from Brandeis University in 1966, and subsequently was a postdoctoral Fellow in Pathology with Howard Green at NYU Medical center, and at the Weizmann Institute in Israel with Ernest Winocour. He was then recruited to Cold Spring Harbor Laboratory by James Watson to establish a research program on reversion of cancer cells. He became a tenured Associate Professor of Microbiology at the Stony Brook University Medical Center before returning to Columbia as a Professor of Biological Sciences in 1978. He served as Dean of Columbia College from 1982 to 1989, overseeing the enrollment of women in the College for the first time.  He remains at Columbia as a Professor of Biological Sciences, and also serves as Director of The University Seminars; he is the fifth Director since its founding in 1944.  He is also a member of the Affiliate Faculty of the American Studies Program. From 1999-2012, he was the Director of the Center for the Study of Science and Religion, a program within Columbia’s Earth Institute. In 2014 his interest in questions that lie at the intersection of science and subjectivity, coupled with the gift of an endowment from College alumnus Harvey Krueger ’51, led him to establish the Research Cluster on Science and Subjectivity, a project within Columbia’s Center for Science and Society.  In addition to these activities, Pollack has authored many research reports, reviews, articles, and opinion pieces on molecular biology, medical ethics and science education. For the academic year 1993–1994 he was awarded a Guggenheim Fellowship in science writing. He has written or edited ten books, including Signs of Life: the Language and Meanings of DNA (1994), which won the Lionel Trilling Award and
https://en.wikipedia.org/wiki/List%20of%20web%20directories
A Web directory is a listing of Websites organized in a hierarchy or interconnected list of categories. The following is a list of notable Web directory services. General DOAJ.org – Directory of Open Access Journals DMOZ (also known as Open Directory Project) – was at one point the largest directory of the Web. Its open content was mirrored at many sites. Offline since March 2017. Continued since August 2018 as Curlie.org. Jasmine Directory - Lists websites by topic and by region, specializing in business websites. Sources – general subject web portal for journalists, freelance writers, editors, authors and researchers; in addition to a search engine it includes a subject-based directory. Starting Point Directory – Founded in 1995, relaunched in 2006, charges a fee. World Wide Web Virtual Library (VLIB) – oldest directory of the Web. Business directories Business.com – Integrated directory of knowledge resources and companies, that charges a fee for listing review and operates as a pay per click search engine. Yell – is a digital marketing and online directory business in the United Kingdom Niche Business.com – Integrated directory of knowledge resources and companies, that charges a fee for listing review and operates as a pay per click search engine. Library and Archival Exhibitions on the Web – international database of online exhibitions which is a service of the Smithsonian Institution Libraries. ProgrammableWeb – resource on APIs that provides a directory of APIs. Virtual Library museums pages – directory of museum websites around the world. Regional 2345.com – Chinese web directory founded in 2005. The website is the second most used web directory in China. Alleba – Filipino search engine website, with directory. Dalilmasr – Egyptian online directory Timway – web portal and directory primarily serving Hong Kong. Defunct directories AboutUs.org – directory from 2005 to 2013. Anime Web Turnpike – was a web directory founded in Augu
https://en.wikipedia.org/wiki/Test%20article%20%28food%20and%20drugs%29
In medicine and food law, a test article is a prototype product specifically manufactured to test the product. Since medical and food test products involve human safety, they are subject to legal regulation. United States In the United States, medical and food test products are regulated by the Food and Drug Administration (FDA). Title 21 "Food and Drugs" , Part 50 "Protection of Human Subjects" defines test article as "drug (including a biological product for human use), medical device for human use, human food additive, color additive, electronic product, or any other article subject to regulation under the act or under sections 351 and 354-360F of the Public Health Service Act (42 U.S.C. 262 and 263b-263n)." Title 21 "Food and Drugs" , Part 58 "Good Laboratory Practice for Nonclinical Laboratory Studies" defines test article similarly: as "any food additive, color additive, drug, biological product, electronic product, medical device for human use, or any other article subject to regulation under the act or under sections 351 and 354-360F of the Public Health Service Act". Generally, prior to the usage of test articles in research involving human subjects, and approval from the institutional review board (IRB), however emergency life-threatening situations are exempt form this requirement. IRB must be notified about an emergency usage of test articles within 5 business days.
https://en.wikipedia.org/wiki/Acyl-protein%20thioesterase
Acyl-protein thioesterases are enzymes that cleave off lipid modifications on proteins, located on the sulfur atom of cysteine residues linked via a thioester bond. Acyl-protein thioesterases are part of the α/β hydrolase superfamily of proteins and have a conserved catalytic triad. For that reason, acyl-protein thioesterases are also able to hydrolyze oxygen-linked ester bonds. Function Acyl-protein thioesterases are involved in the depalmitoylation of proteins, meaning they cleave off palmitoyl modifications on proteins' cysteine residues. Cellular targets include trimeric G-alpha proteins, ion channels and GAP-43. Moreover, human acyl-protein thioesterases 1 and 2 have been identified as major components in controlling the palmitoylation cycle of the oncogene Ras. Depalmitoylation of Ras by acyl-protein thioesterases potentially reduces Ras' affinity to endomembranes, allowing it to be palmitoylated again at the Golgi apparatus and to be directed to the plasma membrane. Acyl-protein thioesterases, therefore, are thought to correct potential mislocalization of Ras. Known enzymes Currently fully validated human acyl-protein thioesterases are APT1 and APT2 which share 66% sequence homology. Additionally there are a handful of putative acyl-protein thioesterases reported, including the ABHD17 enzyme family. In the lysosome, PPT1 of the palmitoyl protein thioesterase family has similar enzymatic activity as acyl-protein thioesterases. Structure Acyl-protein thioesterases feature 3 major structural components that determine protein function and substrate processing: 1. A conserved, classical catalytic triad to break ester and thioester bonds; 2. A long hydrophobic substrate tunnel to accommodate the palmitoyl moiety, as identified in the crystal structures of human acyl-protein thioesterase 1, human acyl-protein thioesterase 2 and Zea mays acyl-protein thioesterase 2; 3. A lid-loop that covers the catalytic site, is highly flexible and is a main factor determin
https://en.wikipedia.org/wiki/Mezcalera%20Ocean
The Mezcalera Ocean is an inferred ancient ocean preserved in rocks in western Mexico. The Mezcalera oceanic plate was likely subducted and consumed into the mantle allowing the Guerrero Terrane to be accreted to western Mexico in the Early Cretaceous. Speculative reconstructions suggest that Mezcalera plate experienced slab rollback in the east along the Mexican Craton and simultaneously subducted in the west beneath the Guerrero Terrane. See also List of ancient oceans
https://en.wikipedia.org/wiki/Supercentenarian
A supercentenarian (sometimes hyphenated as super-centenarian) is a person who has reached the age of 110 years. This age is achieved by about one in 1,000 centenarians. Supercentenarians typically live a life free of major age-related diseases until shortly before the maximum human lifespan is reached. Etymology The term "Supercentenarian", originally hyphenated as Super-centenarian, has existed since 1870. Norris McWhirter, editor of Guinness World Records, used the term in association with age claim's researcher A. Ross Eckler Jr. in 1976, and the term was further popularised in 1991 by William Strauss and Neil Howe in their book Generations. The term "semisupercentenarian", has been used to describe someone from 105-109. Originally the term "supercentenarian" was used to mean someone well over the age of 100, but 110 years and over became the cutoff point of accepted criteria for demographers. Incidence The Gerontology Research Group maintains a top 30–40 list of oldest verified living people. The researchers estimate, based on a 0.15% to 0.25% survival rate of centenarians until the age of 110, that there should be between 300 and 450 living supercentenarians in the world. A study conducted in 2010 by the Max Planck Institute for Demographic Research found 663 validated supercentenarians, living and dead, and showed that the countries with the highest total number (not frequency) of supercentenarians (in decreasing order) were the United States, Japan, England plus Wales, France, and Italy. The first verified supercentenarian in human history was Dutchman Geert Adriaans Boomgaard (1788–1899), and it was not until the 1980s that the oldest verified age surpassed 115. History While claims of extreme age have persisted from the earliest times in history, the earliest supercentenarian accepted by Guinness World Records is Dutchman Thomas Peters (reportedly 1745–1857).. Other scholars, such as French demographer Jean-Marie Robine, consider Geert Adriaans Boom
https://en.wikipedia.org/wiki/International%20Exposition%20of%20Electricity
The first International Exposition of Electricity () ran from 15 August 1881 through to 15 November 1881 at the Palais de l'Industrie on the Champs-Élysées in Paris, France. It served to display the advances in electrical technology since the small electrical display at the 1878 Universal Exposition. Exhibitors came from the United Kingdom, United States, Germany, Italy and the Netherlands, as well as from France. As part of the exhibition, the first International Congress of Electricians presented numerous scientific and technical papers, including definitions of the standard practical units volt, ohm and ampere. History Adolphe Cochery, Minister of Posts and Telegraphs of the time, had initially suggested that an international exposition should be held. This show was a great stir. The public could admire the dynamo of Zénobe Gramme, the incandescent light, the Théâtrophone, the electric tramway of Werner von Siemens, the telephone of Alexander Graham Bell, an electrical distribution network by Marcel Deprez, and an electric boat by Gustave Trouvé. As part of the exhibition, the first International Congress of Electricians, which met in the halls of the Palais du Trocadero, presented numerous scientific and technical papers, including definitions of the standard practical units volt, ohm and ampere, the International System of Electrical and Magnetic Units. George Berger was the Commissioner General. Aside from the provision of the building by the French government, the exhibition was privately financed. Organizers would donate profits to scientific works in the public interest. International Congress of Electricians This congress was a decisive step in the building of the modern International System of Units (SI), since ohm, ampere, coulomb and farad were defined at this occasion. Main participants include Éleuthère Mascart, William Thomson (who later became Lord Kelvin), Hermann von Helmholtz, Rudolf Clausius, Gustav Kirchhoff, Gustav Heinrich Wiedemann,
https://en.wikipedia.org/wiki/Field%20mill%20%28carriage%29
A field mill, also known as a camp mill, was a premodern vehicle which acted as a mobile mill used for grinding grains, which had the very practical use of feeding a moving army. History Later Zhao In the Yezhongji (鄴中記) ('Record of Affairs at the Capital of the Later Zhao Dynasty') by Lu Hui, covering the history of the Later Zhao (319–351 AD) court in China, the text describes various mechanical devices used, including the wheeled odometer for measuring distance and the south-pointing chariot for indicating cardinal direction. Two engineers in particular, the Palace Officer Xie Fei and Director of Imperial Workshops Wei Mengbian, were known for their designs and worked at the court of Shi Hu (r. 334–349). The two had crafted a four-wheeled carriage about 6 m (20 ft) long with water-spouting dragons hanging over a large golden Buddhist statue that had a mechanical wooden statue of a Daoist continually rubbing his front. Other mechanical figures included ten Daoists dressed in monastic robes who continually rotated around the Buddha while periodically bowing, saluting, and throwing incense into a censer. All of these mechanical figures were driven only by the movement of the carriage; once the carriage halted, the figures stopped moving and the water stopped spouting from the artificial dragons. Xie and Wei created a similar device operated by wheel motion called the field mill, although it served a more practical purpose than the theatrical display of moving statues and water-spouting dragons. The Yezhongji states that the two devised a "pounding cart" or "pounding wagon" which had figurine statues armed with real tilt hammers who pounded and hulled rice only when the cart moved. In addition to this they had a "mill cart" (field mill or camp mill) which had rotating millstones mounted on their frames, which would rotate and grind wheat as the cart moved forward. Just like the carriage with mechanical figures mentioned above, when the carriage stopped, the device
https://en.wikipedia.org/wiki/Sand%20bath
A sand bath is a common piece of laboratory equipment made from a container filled with heated sand. It is used to evenly heat another container, most often during a chemical reaction. A sand bath is most commonly used in conjunction with a hot plate or heating mantle. A beaker is filled with sand or metal pellets (called shot) and is placed on the plate or mantle. The reaction vessel is then partially covered by sand or pellets. The sand or shot then conducts the heat from the plate to all sides of the reaction vessel. This technique allows a reaction vessel to be heated throughout with minimal stirring, as opposed to heating the bottom of the vessel and waiting for convection to heat the remainder, cutting down on both the duration of the reaction and the possibility of side reactions that may occur at higher temperatures. A variation on this theme is the water bath in which the sand is replaced with water. It can be used to keep a reaction vessel at the temperature of boiling water until all water is evaporated (see Standard enthalpy change of vaporization). Sand baths are one of the oldest known pieces of laboratory equipment, having been used by the alchemists. In Arabic alchemy, a sand bath was known as a qadr. In Latin alchemy, a sand bath was called balneum siccum, balneum cineritium, or balneum arenosum. See also Heat bath Water bath Oil bath Notes
https://en.wikipedia.org/wiki/Scardovia%20wiggsiae
Scardovia wiggsiae is a species of bacterium in the family Bifidobacteriaceae. In 2011, a study carried out using anaerobic culture conditions allowed the identification of a newly named species, Scardovia wiggsiae [], which was significantly associated with severe ECC (Early childhood caries, a particularly severe manifestation of carious pathology affecting children between birth and 71 months of age). The paper of Bossù et al. 2020 [] shows that S. wiggsiae forms biofilm and illustrates for the first time with high resolution scanning electron microscopy images the morphology of this bacterium and its biofilm. Images were obtained usingn original scanning electron microscopy protocol, the OsO4-RR-TA-IL treatment. The biofilm had an intricate three-dimensional architecture made of Eps trabeculae, in this structure a complex micro-canalicular system was developed. S. wiggsiae has the aspect of an elongated bacterium, without pili or fimbriae. It forms clusters of bacteria embedded in the Eps scaffold.
https://en.wikipedia.org/wiki/Boundary%20value%20problem
In the study of differential equations, a boundary-value problem is a differential equation subjected to constraints called boundary conditions. A solution to a boundary value problem is a solution to the differential equation which also satisfies the boundary conditions. Boundary value problems arise in several branches of physics as any physical differential equation will have them. Problems involving the wave equation, such as the determination of normal modes, are often stated as boundary value problems. A large class of important boundary value problems are the Sturm–Liouville problems. The analysis of these problems, in the linear case, involves the eigenfunctions of a differential operator. To be useful in applications, a boundary value problem should be well posed. This means that given the input to the problem there exists a unique solution, which depends continuously on the input. Much theoretical work in the field of partial differential equations is devoted to proving that boundary value problems arising from scientific and engineering applications are in fact well-posed. Among the earliest boundary value problems to be studied is the Dirichlet problem, of finding the harmonic functions (solutions to Laplace's equation); the solution was given by the Dirichlet's principle. Explanation Boundary value problems are similar to initial value problems. A boundary value problem has conditions specified at the extremes ("boundaries") of the independent variable in the equation whereas an initial value problem has all of the conditions specified at the same value of the independent variable (and that value is at the lower boundary of the domain, thus the term "initial" value). A boundary value is a data value that corresponds to a minimum or maximum input, internal, or output value specified for a system or component. For example, if the independent variable is time over the domain [0,1], a boundary value problem would specify values for at both and , wh
https://en.wikipedia.org/wiki/Advaxis
Advaxis Inc. is an American company devoted to the discovery, development and commercialization of immunotherapies based on a technology platform which uses engineered Listeria monocytogenes (aka Lm). The company is headquartered in Princeton, New Jersey and was incorporated in Delaware in 2006. The Lm-based platform on which the company's products are based involves the use of attenuated Lm which secrete antigen/adjuvant fusion proteins and stimulate a patient's immune system (specifically their T cells) to mount an immune response to the secreted antigen; if the antigen is specifically found on cancerous cells, then the result aims to be an effective immune response targeting and eliminating the cancer. Treatments developed using this paradigm are referred to as Lm-LLO immunotherapies. Today, the Company has over fifteen distinct constructs in various stages of development, directly developed by the Company and through strategic collaborations with centers such as: the National Cancer Institute, Cancer Research – UK, the Wistar Institute, the University of Pennsylvania, and the Department of Homeland Security among others. The Company also has a veterinary medicine program that is evaluating an Lm-LLO based immunotherapy in a Phase 1 study in canine osteosarcoma. Source: www.advaxis.com Corporate history and governance Beginnings Advaxis was a Delaware corporation when it was acquired by a shell corporation (in the official SEC sense) in November 2004. The acquiring company was Great Expectations, which was incorporated in Colorado in June 1987. The only operating company owned by Great Expectations was Advaxis and in 2004, a month after the acquisition, it changed its name to Advaxis, and 18 months after that it reincorporated as a Delaware corporation. The official 'date of inception' for the company is 1 March 2002. Partnerships In 2014, Advaxis entered a co-development and commercialization agreement with India's Biocon for the ADXS-HPV therapeutic
https://en.wikipedia.org/wiki/Roofline%20model
The Roofline model is an intuitive visual performance model used to provide performance estimates of a given compute kernel or application running on multi-core, many-core, or accelerator processor architectures, by showing inherent hardware limitations, and potential benefit and priority of optimizations. By combining locality, bandwidth, and different parallelization paradigms into a single performance figure, the model can be an effective alternative to assess the quality of attained performance instead of using simple percent-of-peak estimates, as it provides insights on both the implementation and inherent performance limitations. The most basic Roofline model can be visualized by plotting floating-point performance as a function of machine peak performance, machine peak bandwidth, and arithmetic intensity. The resultant curve is effectively a performance bound under which kernel or application performance exists, and includes two platform-specific performance ceilings: a ceiling derived from the memory bandwidth and one derived from the processor's peak performance (see figure on the right). Related terms and performance metrics Work The work denotes the number of operations performed by a given kernel or application. This metric may refer to any type of operation, from number of array points updated, to number of integer operations, to number of floating point operations (FLOPs), and the choice of one or another is driven by convenience. In the majority of the cases however, is expressed as FLOPs. Note that the work is a property of the given kernel or application and thus depend just partially on the platform characteristics. Memory traffic The memory traffic denotes the number of bytes of memory transfers incurred during the execution of the kernel or application. In contrast to , is heavily dependent on the properties of the chosen platform, such as for instance the structure of the cache hierarchy. Arithmetic intensity The arithmetic intensi
https://en.wikipedia.org/wiki/Biochemical%20cascade
A biochemical cascade, also known as a signaling cascade or signaling pathway, is a series of chemical reactions that occur within a biological cell when initiated by a stimulus. This stimulus, known as a first messenger, acts on a receptor that is transduced to the cell interior through second messengers which amplify the signal and transfer it to effector molecules, causing the cell to respond to the initial stimulus. Most biochemical cascades are series of events, in which one event triggers the next, in a linear fashion. At each step of the signaling cascade, various controlling factors are involved to regulate cellular actions, in order to respond effectively to cues about their changing internal and external environments. An example would be the coagulation cascade of secondary hemostasis which leads to fibrin formation, and thus, the initiation of blood coagulation. Another example, sonic hedgehog signaling pathway, is one of the key regulators of embryonic development and is present in all bilaterians. Signaling proteins give cells information to make the embryo develop properly. When the pathway malfunctions, it can result in diseases like basal cell carcinoma. Recent studies point to the role of hedgehog signaling in regulating adult stem cells involved in maintenance and regeneration of adult tissues. The pathway has also been implicated in the development of some cancers. Drugs that specifically target hedgehog signaling to fight diseases are being actively developed by a number of pharmaceutical companies. Introduction Signaling cascades Cells require a full and functional cellular machinery to live. When they belong to complex multicellular organisms, they need to communicate among themselves and work for symbiosis in order to give life to the organism. These communications between cells triggers intracellular signaling cascades, termed signal transduction pathways, that regulate specific cellular functions. Each signal transduction occurs with a p
https://en.wikipedia.org/wiki/Oculo-auricular%20phenomenon
The oculo-auricular phenomenon, first described by Kinnier Wilson in 1908, is the phenomenon of an extreme lateral gaze inducing a slight but perceptible backwards movement of the upper part of the pinna. It is a muscle synergy involving the Abducens innervated lateral rectus muscle, an external muscle of the eye, and the facial innervated posterior auricular muscle, an external muscle of the ear. Wilson's phenomenon had attracted attention at the time because of his renown and for its implications regarding Darwin's theory of natural selection. According to (Urban 1993), "In patients with brainstem disease abnormal transverse auricular muscle coactivation is characterized by absence of activity in one or both ear muscles during lateral gaze in either or both directions."
https://en.wikipedia.org/wiki/Direction%20flag
The direction flag is a flag that controls the left-to-right or right-to-left direction of string processing, stored in the FLAGS register on all x86-compatible CPUs. It is bit number 10. This flag is used to determine the direction ('forward' or 'backward') in which several bytes of data will be copied from one place in the memory, to another. The direction is important mainly when the original data position in memory and the target data position overlap. If it is set to 0 (using the clear-direction-flag instruction CLD) — it means that string is processed beginning from lowest to highest address; such instructions mode is called auto-incrementing mode. Both the source index and destination index (like MOVS) will increase them; In case it is set to 1 (using the set-direction-flag instruction STD) — the string is processed from highest to lowest address. This is called auto-decrementing mode.
https://en.wikipedia.org/wiki/Microcrystallization
Microcrystallization (or microcrystal test) is a method for identifying lichen metabolites that was predominantly used before the advent of more advanced techniques such as thin-layer chromatography and high-performance liquid chromatography. Developed primarily by Yasuhiko Asahina, this approach relies on the formation of distinctive crystals from lichen extracts. Although now superseded by modern analytical methods, microcrystallization still holds importance for compound purification and analysis using X-ray crystallography. History Between 1936 and 1940, Japanese chemist and lichenologist Yasuhiko Asahina published a series of papers in the Journal of Japanese Botany detailing the microcrystallization technique. This simple and rapid method allowed for the identification of major metabolites in hundreds of lichen species, contributing significantly to taxonomic research. The technique was introduced to western lichenologists in a 1943 publication by Alexander Evans, and was used regularly until more advanced techniques such as thin-layer chromatography and high-performance liquid chromatography were introduced and integrated into laboratories. Decades of research on the secondary metabolite of lichens culminated in the publication of Identification of Lichen Substances, a 1996 work by Siegfried Huneck and Isao Yoshimura, that summarized analytical data for hundreds of lichen molecules, including images of microcrystals. Ultimately, the microcrystallization method had limitations, as it was unable to detect minor components or analyze complex mixtures of lichen substances. Despite these drawbacks, microcrystallization played a crucial role in the study of correlations between lichen chemistry, morphology, and geographic distribution. Procedure To perform microcrystallization, a small piece of lichen is extracted using acetone or other solvents, filtered, and evaporated to yield a residue. The residue is transferred to a microscope slide, and a drop of microcr
https://en.wikipedia.org/wiki/Animal%20technology
Animal technology refers to the practices of keeping, breeding and providing care for animals that are used for scientific purposes, such as captive in a laboratory. Animal technology is one of the recognized professional areas held by registered science technicians, and animal technologists, also called animal technicians, play a key role in supporting biomedical research by ensuring that animals are available for study. Principal areas of animal technology include animal husbandry and breeding, providing day-to-day care for laboratory animals, ensuring compliance with animal welfare practices and legal issues and performing essential scientific procedures. Technical qualifications for animal technologists and the regulations they must adhere to vary by country, but in many parts of the world animal technology is a highly structured profession that forms part of laboratory management. Animal technology is related to the field of animal management and technologists are often specialize in working with particular species of animals, either in the laboratory or in the field. See also Technology in veterinary medicine History of wildlife tracking technology
https://en.wikipedia.org/wiki/Gene%20delivery
Gene delivery is the process of introducing foreign genetic material, such as DNA or RNA, into host cells. Gene delivery must reach the genome of the host cell to induce gene expression. Successful gene delivery requires the foreign gene delivery to remain stable within the host cell and can either integrate into the genome or replicate independently of it. This requires foreign DNA to be synthesized as part of a vector, which is designed to enter the desired host cell and deliver the transgene to that cell's genome. Vectors utilized as the method for gene delivery can be divided into two categories, recombinant viruses and synthetic vectors (viral and non-viral). In complex multicellular eukaryotes (more specifically Weissmanists), if the transgene is incorporated into the host's germline cells, the resulting host cell can pass the transgene to its progeny. If the transgene is incorporated into somatic cells, the transgene will stay with the somatic cell line, and thus its host organism. Gene delivery is a necessary step in gene therapy for the introduction or silencing of a gene to promote a therapeutic outcome in patients and also has applications in the genetic modification of crops. There are many different methods of gene delivery for various types of cells and tissues. History Viral based vectors emerged in the 1980s as a tool for transgene expression. In 1983, Albert Siegel described the use of viral vectors in plant transgene expression although viral manipulation via cDNA cloning was not yet available. The first virus to be used as a vaccine vector was the vaccinia virus in 1984 as a way to protect chimpanzees against hepatitis B. Non-viral gene delivery was first reported on in 1943 by Avery et al. who showed cellular phenotype change via exogenous DNA exposure. Methods There are a variety of methods available to deliver genes to host cells. When genes are delivered to bacteria or plants the process is called transformation and when it is used to
https://en.wikipedia.org/wiki/Public%20key%20fingerprint
In public-key cryptography, a public key fingerprint is a short sequence of bytes used to identify a longer public key. Fingerprints are created by applying a cryptographic hash function to a public key. Since fingerprints are shorter than the keys they refer to, they can be used to simplify certain key management tasks. In Microsoft software, "thumbprint" is used instead of "fingerprint." Creating public key fingerprints A public key fingerprint is typically created through the following steps: A public key (and optionally some additional data) is encoded into a sequence of bytes. To ensure that the same fingerprint can be recreated later, the encoding must be deterministic, and any additional data must be exchanged and stored alongside the public key. The additional data is typically information which anyone using the public key should be aware of. Examples of additional data include: which protocol versions the key should be used with (in the case of PGP fingerprints); and the name of the key holder (in the case of X.509 trust anchor fingerprints, where the additional data consists of an X.509 self-signed certificate). The data produced in the previous step is hashed with a cryptographic hash function such as SHA-1 or SHA-2. If desired, the hash function output can be truncated to provide a shorter, more convenient fingerprint. This process produces a short fingerprint which can be used to authenticate a much larger public key. For example, whereas a typical RSA public key will be 2048 bits in length or longer, typical MD5 or SHA-1 fingerprints are only 128 or 160 bits in length. When displayed for human inspection, fingerprints are usually encoded into hexadecimal strings. These strings are then formatted into groups of characters for readability. For example, a 128-bit MD5 fingerprint for SSH would be displayed as follows: 43:51:43:a1:b5:fc:8b:b7:0a:3a:a9:b1:0f:66:73:a8 Using public key fingerprints for key authentication When a public key
https://en.wikipedia.org/wiki/Lustre%20%28programming%20language%29
Lustre is a formally defined, declarative, and synchronous dataflow programming language for programming reactive systems. It began as a research project in the early 1980s. A formal presentation of the language can be found in the 1991 Proceedings of the IEEE. In 1993 it progressed to practical, industrial use in a commercial product as the core language of the industrial environment SCADE, developed by Esterel Technologies. It is now used for critical control software in aircraft, helicopters, and nuclear power plants. Structure of Lustre programs A Lustre program is a series of node definitions, written as: node foo(a : bool) returns (b : bool); let b = not a; tel Where foo is the name of the node, a is the name of the single input of this node and b is the name of the single output. In this example the node foo returns the negation of its input a, which is the expected result. Inner variables Additional internal variables can be declared as follows: node Nand(X,Y: bool) returns (Z: bool); var U: bool; let U = X and Y; Z = not U; tel Note: The equations order doesn't matter, the order of lines U = X and Y; and Z = not U; doesn't change the result. Special operators Examples Edge detection node Edge (X : bool) returns (E : bool); let E = false -> X and not pre X; tel See also Esterel SIGNAL (another dataflow-oriented synchronous language) Synchronous programming language Dataflow programming
https://en.wikipedia.org/wiki/Twin%20vector%20quantization
In data compression, twin vector quantization is related to vector quantization, but the speed of the quantization is doubled by the secondary vector analyzer. By using a subdimensional vector space useless hyperspace will be destroyed in the process. The formula for calculating the amount of destroyed hyperspace is: H(x) = 5.22 / 4m Data compression
https://en.wikipedia.org/wiki/Fairy%20Flag
The Fairy Flag (Scottish Gaelic: Am Bratach Sìth) is an heirloom of the chiefs of Clan MacLeod. It is held in Dunvegan Castle along with other notable heirlooms, such as the Dunvegan Cup and Sir Rory Mor's Horn. The Fairy Flag is known for the numerous traditions of celtic fairies, and magical properties associated with it. The flag is made of silk, is yellow or brown in colour, and is a square of side about . It has been examined numerous times in the last two centuries, and its condition has somewhat deteriorated. It is ripped and tattered, and is considered to be extremely fragile. The flag is covered in small red "elf dots". In the early part of the 19th century, the flag was also marked with small crosses, but these have since disappeared. The silk of the flag has been stated to have originated in the Far East, and was therefore extremely precious, which led some to believe that the flag may have been an important relic of some sort. Others have attempted to associate the flag with the Crusades or even a raven banner, which was said to have been used by various Viking leaders in the British Isles. There are numerous traditions and stories associated with the flag, most of which deal with its magical properties and mysterious origins. The flag is said to have originated as: a gift from the fairies to an infant chieftain; a gift to a chief from a departing fairy-lover; a reward for defeating an evil spirit. The various powers attributed to the Fairy Flag include: the ability to multiply a clan's military forces; the ability to save the lives of certain clanfolk; the ability to cure a plague on cattle; the ability to increase the chances of fertility; and the ability to bring herring into the loch at Dunvegan. Some traditions relate that if the flag were to be unfurled and waved more than three times, it would either vanish, or lose its powers forever. Clan tradition, preserved in the early 19th century, tells how the Fairy Flag was entrusted to a family of here
https://en.wikipedia.org/wiki/Cotransporter
Cotransporters are a subcategory of membrane transport proteins (transporters) that couple the favorable movement of one molecule with its concentration gradient and unfavorable movement of another molecule against its concentration gradient. They enable coupled or cotransport (secondary active transport) and include antiporters and symporters. In general, cotransporters consist of two out of the three classes of integral membrane proteins known as transporters that move molecules and ions across biomembranes. Uniporters are also transporters but move only one type of molecule down its concentration gradient and are not classified as cotransporters. Background Cotransporters are capable of moving solutes either up or down gradients at rates of 1000 to 100000 molecules per second. They may act as channels or transporters, depending on conditions under which they are assayed. The movement occurs by binding to two molecules or ions at a time and using the gradient of one solute's concentration to force the other molecule or ion against its gradient. Some studies show that cotransporters can function as ion channels, contradicting the classical models. For instance the wheat HKT1 transporter shows two modes of transport by the same protein. Cotransporters can be classified as antiporters and symporters. Both use electric potential and/or chemical gradients to move protons and ions against their concentration gradient. In plants the proton is considered a secondary substance and high proton concentration in the apoplast powers the inward movement of certain ions by symporters. A Proton gradient moves the ions into the vacuole by proton-sodium antiporter or the proton-calcium antiporter. In plants, sucrose transport is distributed throughout the plant by the proton-pump where the pump, as discussed above, creates a gradient of protons so that there are many more on one side of the membrane than the other. As the protons diffuse back across the membrane, the free energy
https://en.wikipedia.org/wiki/Gapless%20playback
Gapless playback is the uninterrupted playback of consecutive audio tracks, such that relative time distances in the original audio source are preserved over track boundaries on playback. For this to be useful, other artifacts (than timing-related ones) at track boundaries should not be severed either. Gapless playback is common with compact discs, gramophone records, or tapes, but is not always available with other formats that employ compressed digital audio. The absence of gapless playback is a source of annoyance to listeners of music where tracks are meant to segue into each other, such as some classical music (opera in particular), progressive rock, concept albums, electronic music, and live recordings with audience noise between tracks. Causes of gaps Playback latency Various software, firmware, and hardware components may add up to a substantial delay associated with starting playback of a track. If not accounted for, the listener is left waiting in silence as the player fetches the next file (see harddisk access time), updates metadata, decodes the whole first block, before having any data to feed the hardware buffer. The gap can be as much as half a second or more — very noticeable in "continuous" music such as certain classical or dance genres. In extreme cases, the hardware is even reset between tracks, creating a very short "click". To account for the whole chain of delays, the start of the next track should ideally be readily decoded before the currently playing track finishes. The two decoded pieces of audio must be fed to the hardware continuously over the transition, as if the tracks were concatenated in software. Many older audio players on personal computers do not implement the required buffering to play gapless audio. Some of these rely on third-party gapless audio plug-ins to buffer output. Most recent players and newer versions of old players now support gapless playback directly. Compression artifacts Lossy audio compression schemes th
https://en.wikipedia.org/wiki/Intel%202700G
Intel 2700G (code-named Marathon) is a low power (50 mW max) graphics co-processor for the XScale PXA27x processor, announced on April 12, 2004. It is built on both the PowerVR MBX Lite chip design (which is a descendant from the PowerVR2 graphics technology that powered the Dreamcast) and on the MVED1 video encoder/decoder technology. Variants This accelerator comes in 3 variants: The 2700G3, 2700G5 and 2700G7. 2700G3 The 2700G3 is the value version of the accelerator. It has 384 kB of on-die memory suitable for driving an HVGA (320×480) or smaller graphics display. 2700G5 The 2700G5 is the performance version of the accelerator. It has 704 kB of on-die memory suitable for driving a VGA (640×480) resolution display and decoding MPEG-4 video. 2700G7 The 2700G7 is the same as the 2700G5 but is stacked with 16 MB of local memory with a 100 MHz, 32-bit bus (maximum 400 MB/s theoretical bandwidth) in the same package. Features All the chips have a 75 MHz, 32-bit core shipped in BGA packaging. Accelerated dual display The 2700G has its own LCD controller and two LCD outputs. This allows for the XScale processor and 2700G's graphics displays to be used at the same time. The 2700G can drive an external display up to 1024×768 (with 32-bit colour) or 1280×1024 (with 16-bit colour). 2D graphics acceleration The accelerator supports clipping, alpha blending and anti-aliasing. It also has a variety of block level transfer (BLT) functions. The 2D accelerator is capable of up to an 84 million pixels per second fill rate (150 million pixels per second claimed by intel). 2D features of Intel 2700G ROP 2,3,4 Alpha Blending Full-screen antialiasing BitBLT, StretchBLT, CSCBLT Hardware doubling and multiplication of pixels Color depth of 16, 18, and 24 bit 3D graphics acceleration The 2700G has a complete hardware rendering pipeline and is compatible with the OpenGL ES 1.0 standard, using the common-lite profile. The 3D accelerator is capable of processing about 831 00
https://en.wikipedia.org/wiki/Theories%20of%20general%20anaesthetic%20action
A general anaesthetic (or anesthetic) is a drug that brings about a reversible loss of consciousness. These drugs are generally administered by an anaesthetist/anesthesiologist to induce or maintain general anaesthesia to facilitate surgery. General anaesthetics have been widely used in surgery since 1842 when Crawford Long for the first time administered diethyl ether to a patient and performed a painless operation. It has long been believed that general anaesthetics exert their effects (analgesia, unconsciousness, immobility) through a membrane mediated mechanism or by directly modulating the activity of membrane proteins in the neuronal membrane. In general, different anaesthetics exhibit different mechanisms of action such that there are numerous molecular targets at all levels of integration within the central nervous system. However, for certain intravenous anaesthetics, such as propofol and etomidate, the main molecular target has been identified to be GABAA receptor, with particular β subunits playing a crucial role. The concept of specific interactions between receptors and drugs first introduced by Paul Ehrlich in 1897 states that drugs act only when they are bound to their targets (receptors). The identification of concrete molecular targets for general anaesthetics was made possible only with the modern development of molecular biology techniques for single amino acid mutations in proteins of genetically engineered mice. Lipid solubility-anaesthetic potency correlation (the Meyer-Overton correlation) A nonspecific mechanism of general anaesthetic action was first proposed by Emil Harless and Ernst von Bibra in 1847. They suggested that general anaesthetics may act by dissolving in the fatty fraction of brain cells and removing fatty constituents from them, thus changing activity of brain cells and inducing anaesthesia. In 1899 Hans Horst Meyer published the first experimental evidence of the fact that anaesthetic potency is related to lipid solubil
https://en.wikipedia.org/wiki/Crack%20%28password%20software%29
Crack is a Unix password cracking program designed to allow system administrators to locate users who may have weak passwords vulnerable to a dictionary attack. Crack was the first standalone password cracker for Unix systems and the first to introduce programmable dictionary generation as well. Crack began in 1990 when Alec Muffett, a Unix system administrator at the University of Wales Aberystwyth, was trying to improve Dan Farmer's 'pwc' cracker in COPS. Muffett found that by re-engineering the memory management, he got a noticeable performance increase. This led to a total rewrite which became "Crack v2.0" and further development to improve usability. Public Releases The first public release of Crack was version 2.7a, which was posted to the Usenet newsgroups alt.sources and alt.security on 15 July 1991. Crack v3.2a+fcrypt, posted to comp.sources.misc on 23 August 1991, introduced an optimised version of the Unix crypt() function but was still only really a faster version of what was already available in other packages. The release of Crack v4.0a on 3 November 1991, however, introduced several new features that made it a formidable tool in the system administrator's arsenal. Programmable dictionary generator Network distributed password cracking Crack v5.0a released in 2000 did not introduce any new features, but instead concentrated on improving the code and introducing more flexibility, such as the ability to integrate other crypt() variants such as those needed to attack the MD5 password hashes used on more modern Unix, Linux and Windows NT systems. It also bundled Crack v6 - a minimalist password cracker and Crack v7 - a brute force password cracker. Legal issues arising from using Crack Randal L. Schwartz, a notable Perl programming expert, in 1995 was prosecuted for using Crack on the password file of a system at Intel, a case the verdict of which was eventually expunged. Crack was also used by Kevin Mitnick when hacking into Sun Microsystems
https://en.wikipedia.org/wiki/Peierls%20substitution
The Peierls substitution method, named after the original work by Rudolf Peierls is a widely employed approximation for describing tightly-bound electrons in the presence of a slowly varying magnetic vector potential. In the presence of an external magnetic vector potential , the translation operators, which form the kinetic part of the Hamiltonian in the tight-binding framework, are simply and in the second quantization formulation The phases are defined as Properties The number of flux quanta per plaquette is related to the lattice curl of the phase factor, and the total flux through the lattice is with being the magnetic flux quantum in Gaussian units. The flux quanta per plaquette is related to the accumulated phase of a single particle state, surrounding a plaquette: Justification Here we give three derivations of the Peierls substitution, each one is based on a different formulation of quantum mechanics theory. Axiomatic approach Here we give a simple derivation of the Peierls substitution, which is based on The Feynman Lectures (Vol. III, Chapter 21). This derivation postulates that magnetic fields are incorporated in the tight-binding model by adding a phase to the hopping terms and show that it is consistent with the continuum Hamiltonian. Thus, our starting point is the Hofstadter Hamiltonian: The translation operator can be written explicitly using its generator, that is the momentum operator. Under this representation its easy to expand it up to the second order, and in a 2D lattice . Next, we expand up to the second order the phase factors, assuming that the vector potential does not vary significantly over one lattice spacing (which is taken to be small) Substituting these expansions to relevant part of the Hamiltonian yields Generalizing the last result to the 2D case, the we arrive to Hofstadter Hamiltonian at the continuum limit: where the effective mass is and . Semi-classical approach Here we show that the Peierls phase fac
https://en.wikipedia.org/wiki/Aposematism
Aposematism is the advertising by an animal to potential predators that it is not worth attacking or eating. This unprofitability may consist of any defenses which make the prey difficult to kill and eat, such as toxicity, venom, foul taste or smell, sharp spines, or aggressive nature. These advertising signals may take the form of conspicuous coloration, sounds, odours, or other perceivable characteristics. Aposematic signals are beneficial for both predator and prey, since both avoid potential harm. The term was coined in 1877 by Edward Bagnall Poulton for Alfred Russel Wallace's concept of warning coloration. Aposematism is exploited in Müllerian mimicry, where species with strong defences evolve to resemble one another. By mimicking similarly coloured species, the warning signal to predators is shared, causing them to learn more quickly at less of a cost. A genuine aposematic signal that a species actually possesses chemical or physical defences is not the only way to deter predators. In Batesian mimicry, a mimicking species resembles an aposematic model closely enough to share the protection, while many species have bluffing deimatic displays which may startle a predator long enough to enable an otherwise undefended prey to escape. There is good evidence for aposematism in terrestrial animals; its existence in marine animals is possible but disputed. Etymology The term aposematism was coined by the English zoologist Edward Bagnall Poulton in his 1890 book The Colours of Animals. He based the term on the Ancient Greek words ἀπό apo 'away' and σῆμα sēma 'sign', referring to signs that warn other animals away. Defense mechanism The function of aposematism is to prevent attack, by warning potential predators that the prey animal has defenses such as being unpalatable or poisonous. The easily detected warning is a primary defense mechanism, and the non-visible defenses are secondary. Aposematic signals are primarily visual, using bright colors and high-con
https://en.wikipedia.org/wiki/%C8%98tefan%20Emilian
Ștefan Emilian (August 8, 1819 – November 1899) was an Imperial Austrian-born Romanian mathematician and architect. Born in Bonchida, Kolozs County (now Bonțida, Cluj County), in the Principality of Transylvania, he was given the surname Kertész as a child, although his birth name was Emilian. He attended high school in Sibiu. Then, from 1841 to 1845, he studied at the Academy of Fine Arts Vienna, graduating with an architect's degree. Additionally, from 1841 to 1843, he took courses at the Vienna Polytechnic Institute. Emilian returned home shortly before 1848, in time for the Transylvanian Revolution. Pursued by the authorities, he sought refuge in Wallachia. By 1850, he was back in Transylvania, where he taught mathematics at Brașov's Greek Orthodox High School. He remained there until 1858, a period during which he designed the new school building. Additionally, he was the architect for the first paper factory in Zărnești. In 1858, he was invited to Iași, the capital of Moldavia, in order to teach drawing and geometry to the upper classes of Academia Mihăileană. Emilian remained there for two years, until the founding of the University of Iași. Additionally, he taught at the military officers' school and the technical school of arts and professions. At the new university, he was named full professor of descriptive geometry and linear perspective, remaining from October 1860 to October 1892, when he had to retire. Meanwhile, he designed the Iași anatomy institute, the Lipovan Church, and the church in Bosia. A single published book of his is known: the 1886 Curs practic de perspectivă liniară. Emilian's funeral eulogy was delivered by Alexandru Dimitrie Xenopol. He married Cornelia Ederlly de Medve. Notes 1819 births 1899 deaths People from Cluj County Romanian people in the Principality of Transylvania (1711–1867) Emigrants from Austria-Hungary to Romania Romanian mathematicians Romanian architects Romanian schoolteachers Romanian revolutionaries Romanian t
https://en.wikipedia.org/wiki/Triptolide
Triptolide is a diterpenoid epoxide which is produced by the thunder god vine, Tripterygium wilfordii. It has in vitro and in vivo activities against mouse models of polycystic kidney disease and pancreatic cancer, but its physical properties and severe toxicity limit its therapeutic potential. Consequently, a synthetic water-soluble prodrug, minnelide, is being studied clinically instead. Triptolide is a component of ContraPest, a contraceptive pest control liquid used to reduce rat populations in the United States. Mechanism of action Several putative target proteins of triptolide have been reported, including polycystin-2, ADAM10, DCTPP1, TAB1, and XPB. Multiple triptolide-resistant mutations exist in XPB (ERCC3) and its partner protein GTF2H4. However, no triptolide-resistant mutations were found in polycystin-2, ADAM10, DCTPP1 and TAB1. Cys342 of XPB was identified as the residue that undergoes covalent modification by the 12,13-epoxide group of triptolide, and the XPB-C342T mutant rendered the T7115 cell line nearly completely resistant to triptolide. The level of resistance conferred by the C342T mutation is about 100-fold higher than the most triptolide-resistant mutants previously identified. Together, these results validate XPB as a target responsible for the antiproliferative activity of triptolide. The disruption of super-enhancer networks has also been suggested as a mechanism of action. Water-soluble prodrugs Minnelide is a more water-soluble synthetic prodrug of triptolide which is converted to triptolide in vivo. In a preclinical mouse model of pancreatic cancer, it was "even more effective than gemcitabine". Its Phase II clinical trials are expected to conclude in February 2019. Glutriptolide, a glucose conjugate of triptolide with better solubility and lower toxicity, did not inhibit XPB activity in vitro, but exhibited tumor control in vivo, which is likely due to sustained stepwise release of active triptolide within cancer cells. A second
https://en.wikipedia.org/wiki/Recursive%20largest%20first%20algorithm
The Recursive Largest First (RLF) algorithm is a heuristic for the NP-hard graph coloring problem. It was originally proposed by Frank Leighton in 1979. The RLF algorithm assigns colors to a graph’s vertices by constructing each color class one at a time. It does this by identifying a maximal independent set of vertices in the graph, assigning these to the same color, and then removing these vertices from the graph. These actions are repeated on the remaining subgraph until no vertices remain. To form high-quality solutions (solutions using few colors), the RLF algorithm uses specialized heuristic rules to try to identify "good quality" independent sets. These heuristics make the RLF algorithm exact for bipartite, cycle, and wheel graphs. In general, however, the algorithm is approximate and may well return solutions that use more colors than the graph’s chromatic number. Description The algorithm can be described by the following three steps. At the end of this process, gives a partition of the vertices representing a feasible -colouring of the graph . Let be an empty solution. Also, let be the graph we wish to color, comprising a vertex set and an edge set . Identify a maximal independent set . To do this: The first vertex added to should be the vertex in that has the largest number of neighbors. Subsequent vertices added to should be chosen as those that (a) are not currently adjacent to any vertex in , and (b) have a maximal number of neighbors that are adjacent to vertices in . Ties in condition (b) can be broken by selecting the vertex with the minimum number of neighbors not in . Vertices are added to in this way until it is impossible to add further vertices. Now set and remove the vertices of from . If still contains vertices, then return to Step 2; otherwise end. Example Consider the graph shown on the right. This is a wheel graph and will therefore be optimally colored by RLF. Executing the algorithm results in the vertices being
https://en.wikipedia.org/wiki/KRP%20%28biochemistry%29
KRP stands for kinesin related proteins. bimC is a subfamily of KRPs and its function is to separate the duplicated centrosomes during mitosis. Role in mitotic repair Kinesin-13 MCAK (Mitotic Centromere-Associated Kinesin) is a KRP that is involved in resolving errors during mitosis involving kinetochore-microtubules. This process is associated with Aurora B Protein Kinase. When Aurora B's function is disrupted, MCAK ability to locate centromeres, which play a critical role in separation of chromosomes during mitosis, was suppressed. There are other environments in which MCAK's function is impaired, absent impact on its associated kinase. For example, alpha-tubulin detyrosination has been demonstrated to impact MCAK's mitotic repair capabilities, suggesting a potential cause of chromosomal instability.
https://en.wikipedia.org/wiki/Dark%20galaxy
A dark galaxy is a hypothesized galaxy with no (or very few) stars. They received their name because they have no visible stars but may be detectable if they contain significant amounts of gas. Astronomers have long theorized the existence of dark galaxies, but there are no confirmed examples to date. Dark galaxies are distinct from intergalactic gas clouds caused by galactic tidal interactions, since these gas clouds do not contain dark matter, so they do not technically qualify as galaxies. Distinguishing between intergalactic gas clouds and galaxies is difficult; most candidate dark galaxies turn out to be tidal gas clouds. The best candidate dark galaxies to date include HI1225+01, AGC229385, and numerous gas clouds detected in studies of quasars. On 25 August 2016, astronomers reported that Dragonfly 44, an ultra diffuse galaxy (UDG) with the mass of the Milky Way galaxy, but with nearly no discernable stars or galactic structure, is made almost entirely of dark matter. Observational evidence Large surveys with sensitive but low-resolution radio telescopes like Arecibo or the Parkes Telescope look for 21-cm emission from atomic hydrogen in galaxies. These surveys are then matched to optical surveys to identify any objects with no optical counterpart; i.e., sources with no stars. Another way astronomers search for dark galaxies is to look for hydrogen absorption lines in the spectra of background quasars. This technique has revealed many intergalactic clouds of hydrogen, but following up on candidate dark galaxies is difficult, since these sources tend to be too far away and are often optically drowned out by the bright light from the quasars. Nature of dark galaxies Origin In 2005, astronomers discovered gas cloud VIRGOHI21 and attempted to determine what it was and why it exerted such a massive gravitational pull on galaxy NGC 4254. After years of ruling out other possible explanations, some have concluded that VIRGOHI21 is a dark galaxy. Size The actua
https://en.wikipedia.org/wiki/Tashiro%20Masashi%20no%20Princess%20ga%20Ippai
is a Japanese video game for the MSX2 home computer system and Family Computer featuring former comedian Masashi Tashiro released in 1989. Summary The story is about the hero Masashi Tashiro who has to rescue the four princesses in distress. One happy ending and four unhappy endings were used in the game; becoming one of the first video games to have multiple endings. The game was not very successful, but it started appearing frequently and getting high prices on online auction sites like Yahoo! after 2000, when Masashi Tashiro was arrested and convicted several times in connection with voyeurism and drug abuse.
https://en.wikipedia.org/wiki/Agmark
AGMARK is a certification mark employed on agricultural products in India, assuring that they conform to a set of standards approved by the Directorate of Marketing and Inspection an attached Office of the Department of Agriculture, Cooperation and Farmers Welfare under Ministry of Agricultural & Farmers Welfare an agency of the Government of India. The AGMARK Head Office at Faridabad (Haryana) is legally enforced in India by the Agricultural Produce (Grading and Marking) Act of 1937 (and amended in 1986). The present AGMARK standards cover quality guidelines for 224 different commodities spanning a variety of pulses, cereals, essential oils, vegetable oils, Fruits and Vegetables and semi-processed products like vermicelli. Etymology (origin) The term agmark was coined by joining the words 'Ag' to mean agriculture and 'mark' for a certification mark. This term was introduced originally in the bill presented in the parliament of India for the Agricultural Produce (Grading and Marking) Act. The entire system of Agmark, including the name, was created by Archibald MacDonald Livingstone, Agricultural and Marketing Advisory to the Government of India, from 1934 to 1941. He was supported by a staff of several hundred. The system was designed to benefit local growers throughout India who were, in the absence of a certification as to quality, exposed to receiving less for their produce from dealers than its true worth. Agmark Laboratories The Agmark certification is employed through fully state-owned Agmark laboratories located across the nation which act as testing and certifying centres. In addition to the Central AGMARK Laboratory (CAL) in Nagpur, there are Regional AGMARK Laboratories (RALs) in 11 nodal cities (Mumbai, New Delhi, Chennai, Kolkata, Kanpur, Kochi, Guntur, Amritsar, Jaipur, Rajkot, Bhopal). Each of the regional laboratories is equipped with and specializes in the testing of products of regional significance. Hence the product range that could be test
https://en.wikipedia.org/wiki/Signaling%20protocol
A signaling protocol is a type of communications protocol for encapsulating the signaling between communication endpoints and switching systems to establish or terminate a connection and to identify the state of connection. The following is a list of signaling protocols: ALOHA Digital Subscriber System No. 1 (EDSS1) Dual-tone multi-frequency signaling H.248 H.323 H.225.0 Jingle Media Gateway Control Protocol (MGCP) Megaco Regional System R1 NBAP (Node B Application Part) Signalling System R2 Session Initiation Protocol Signaling System No. 5 Signaling System No. 6 Signaling System No. 7 Skinny Client Control Protocol (SCCP, Skinny) Q.931 QSIG Network protocols Telephony signals
https://en.wikipedia.org/wiki/Elevator%20paradox%20%28physics%29
The elevator paradox relates to a hydrometer placed on an "elevator" or vertical conveyor that, by moving to different elevations, changes the atmospheric pressure. In this classic demonstration, the floating hydrometer remains at an equilibrium position. Essentially, a hydrometer measures specific gravity of liquids independent of barometric pressure. This is because the change in air pressure is applied to the entire hydrometer flask. The submerged portion of the flask receives a transmitted force through the liquid, thus no portion of the apparatus receives a net force resulting from a change in air pressure. This is a paradox if the buoyancy of the hydrometer is said to depend on the weight of the liquid that it displaces. At a higher barometric pressure, the liquid occupies a slightly smaller volume, and thus more dense might be considered to have a higher specific gravity. However, the hydrometer also displaces air, and the weight of the liquid and the air are affected equally by elevation. Cartesian divers A Cartesian diver, on the other hand, has an internal space that, unlike a hydrometer, is not rigid, and thus can change its displacement as increasing external air pressure compresses the air in the diver. If the diver, instead of being placed in the classic plastic bottle, were floated in a flask on an elevator, the diver would respond to a change in air pressure. Similarly, a non-rigid container like a toy balloon will be affected, as will the rib cage of a human SCUBA diver, and such systems will vary in buoyancy. A glass hydrometer is rigid under normal pressure, for all practical purposes. The hydrometer in an accelerating frame of reference The upward or downward acceleration of the elevator, as long as the net force is directed downward, will not change the equilibrium point of the hydrometer either. The force due to acceleration acts on the hydrometer exactly as it would on an equal mass of water or other liquid. Experimental physics Physic
https://en.wikipedia.org/wiki/Enumerative%20geometry
In mathematics, enumerative geometry is the branch of algebraic geometry concerned with counting numbers of solutions to geometric questions, mainly by means of intersection theory. History The problem of Apollonius is one of the earliest examples of enumerative geometry. This problem asks for the number and construction of circles that are tangent to three given circles, points or lines. In general, the problem for three given circles has eight solutions, which can be seen as 23, each tangency condition imposing a quadratic condition on the space of circles. However, for special arrangements of the given circles, the number of solutions may also be any integer from 0 (no solutions) to six; there is no arrangement for which there are seven solutions to Apollonius' problem. Key tools A number of tools, ranging from the elementary to the more advanced, include: Dimension counting Bézout's theorem Schubert calculus, and more generally characteristic classes in cohomology The connection of counting intersections with cohomology is Poincaré duality The study of moduli spaces of curves, maps and other geometric objects, sometimes via the theory of quantum cohomology. The study of quantum cohomology, Gromov–Witten invariants and mirror symmetry gave a significant progress in Clemens conjecture. Enumerative geometry is very closely tied to intersection theory. Schubert calculus Enumerative geometry saw spectacular development towards the end of the nineteenth century, at the hands of Hermann Schubert. He introduced it for the purpose the Schubert calculus, which has proved of fundamental geometrical and topological value in broader areas. The specific needs of enumerative geometry were not addressed until some further attention was paid to them in the 1960s and 1970s (as pointed out for example by Steven Kleiman). Intersection numbers had been rigorously defined (by André Weil as part of his foundational programme 1942–6, and again subsequently), but this did
https://en.wikipedia.org/wiki/Multivalued%20dependency
In database theory, a multivalued dependency is a full constraint between two sets of attributes in a relation. In contrast to the functional dependency, the multivalued dependency requires that certain tuples be present in a relation. Therefore, a multivalued dependency is a special case of tuple-generating dependency. The multivalued dependency plays a role in the 4NF database normalization. A multivalued dependency is a special case of a join dependency, with only two sets of values involved, i.e. it is a binary join dependency. A multivalued dependency exists when there are at least three attributes (like X,Y and Z) in a relation and for a value of X there is a well defined set of values of Y and a well defined set of values of Z. However, the set of values of Y is independent of set Z and vice versa. Formal definition The formal definition is as follows: Let be a relation and let and be sets of attributes. The multivalued dependency (" multidetermines ") holds on if, for any legal relation and all pairs of tuples and in such that , there exist tuples and in such that: Informally, if one denotes by the tuple having values for collectively equal to , then whenever the tuples and exist in , the tuples and should also exist in . The multivalued dependency can be schematically depicted as shown below: Example Consider this example of a relation of university courses, the books recommended for the course, and the lecturers who will be teaching the course: Because the lecturers attached to the course and the books attached to the course are independent of each other, this database design has a multivalued dependency; if we were to add a new book to the AHA course, we would have to add one record for each of the lecturers on that course, and vice versa. Put formally, there are two multivalued dependencies in this relation: {course}  {book} and equivalently {course}  {lecturer}. Databases with multivalued dependencies thus exhibit redun
https://en.wikipedia.org/wiki/South%20African%20Marine%20Predator%20Lab
The South African Marine Predator Lab (SAMPLA) is a multidisciplinary research institute aimed at uncovering the lives of the marine predators of Southern Africa. SAMPLA’s primary goal is to produce accurate and essential scientific information on marine predators and create awareness on the need to understand and conserve the marine predators and ecosystems of the world. SAMPLA was established by Ryan Johnson, Enrico Gennari, Stephan Swanson and Toby Keswick in the Summer of 2008 SAMPLA undertakes, disseminates and promotes scientific research aimed at producing unique scientific knowledge on South Africa’s marine predators. Research projects The white shark research program at Mossel Bay is multi-faceted, but centred on three main academic pillars. These are: Collection of data for collaborate projects. Conduction of novel research projects by SAMPLA scientists. Offering research facilities, supervision and guidance for graduate student projects at South African universities. Main projects Horizontal and vertical movements of the white shark Bite kinematics of white sharks Predator-prey dynamics White shark genetic profile White shark population dynamics Thermal eco-physiology of the white shark External links South African Marine Predator Lab home page SAMPLA Photo Gallery SAMPLA research projects SAMPLA Publications Research institutes in South Africa Biological research institutes
https://en.wikipedia.org/wiki/Dawson%E2%80%93G%C3%A4rtner%20theorem
In mathematics, the Dawson–Gärtner theorem is a result in large deviations theory. Heuristically speaking, the Dawson–Gärtner theorem allows one to transport a large deviation principle on a “smaller” topological space to a “larger” one. Statement of the theorem Let (Yj)j∈J be a projective system of Hausdorff topological spaces with maps pij : Yj → Yi. Let X be the projective limit (also known as the inverse limit) of the system (Yj, pij)i,j∈J, i.e. Let (με)ε>0 be a family of probability measures on X. Assume that, for each j ∈ J, the push-forward measures (pj∗με)ε>0 on Yj satisfy the large deviation principle with good rate function Ij : Yj → R ∪ {+∞}. Then the family (με)ε>0 satisfies the large deviation principle on X with good rate function I : X → R ∪ {+∞} given by
https://en.wikipedia.org/wiki/Majority%20logic%20decoding
In error detection and correction, majority logic decoding is a method to decode repetition codes, based on the assumption that the largest number of occurrences of a symbol was the transmitted symbol. Theory In a binary alphabet made of , if a repetition code is used, then each input bit is mapped to the code word as a string of -replicated input bits. Generally , an odd number. The repetition codes can detect up to transmission errors. Decoding errors occur when more than these transmission errors occur. Thus, assuming bit-transmission errors are independent, the probability of error for a repetition code is given by , where is the error over the transmission channel. Algorithm Assumption: the code word is , where , an odd number. Calculate the Hamming weight of the repetition code. if , decode code word to be all 0's if , decode code word to be all 1's This algorithm is a boolean function in its own right, the majority function. Example In a code, if R=[1 0 1 1 0], then it would be decoded as, , , so R'=[1 1 1 1 1] Hence the transmitted message bit was 1.
https://en.wikipedia.org/wiki/Distributed-element%20circuit
Distributed-element circuits are electrical circuits composed of lengths of transmission lines or other distributed components. These circuits perform the same functions as conventional circuits composed of passive components, such as capacitors, inductors, and transformers. They are used mostly at microwave frequencies, where conventional components are difficult (or impossible) to implement. Conventional circuits consist of individual components manufactured separately then connected together with a conducting medium. Distributed-element circuits are built by forming the medium itself into specific patterns. A major advantage of distributed-element circuits is that they can be produced cheaply as a printed circuit board for consumer products, such as satellite television. They are also made in coaxial and waveguide formats for applications such as radar, satellite communication, and microwave links. A phenomenon commonly used in distributed-element circuits is that a length of transmission line can be made to behave as a resonator. Distributed-element components which do this include stubs, coupled lines, and cascaded lines. Circuits built from these components include filters, power dividers, directional couplers, and circulators. Distributed-element circuits were studied during the 1920s and 1930s but did not become important until World War II, when they were used in radar. After the war their use was limited to military, space, and broadcasting infrastructure, but improvements in materials science in the field soon led to broader applications. They can now be found in domestic products such as satellite dishes and mobile phones. Circuit modelling Distributed-element circuits are designed with the distributed-element model, an alternative to the lumped-element model in which the passive electrical elements of electrical resistance, capacitance and inductance are assumed to be "lumped" at one point in space in a resistor, capacitor or inductor, respe
https://en.wikipedia.org/wiki/Wadi%20Rum%20Consultancy
The Wadi Rum Consultancy of Wadi Rum Organic Farms, is an example of desert greening. Begun in 2010, it is located in historic Wadi Rum, in southern Jordan. Overseen by permaculture expert Geoff Lawton, it has established a sustainable agriculture system. Overview The program has achieved success, primarily, by implementing principles of hydrological and permacultural design. The results of the consultancy have been documented in photographs, as well as in several videos. See also Al Baydha Project Sahara Forest Project
https://en.wikipedia.org/wiki/Information%20continuum
The term information continuum is used to describe the whole set of all information, in connection with information management. The term may be used in reference to the information or the information infrastructure of a people, a species, a scientific subject or an institution. Other usages in biological anthropology, term information continuum is related to study of social information transfer and evolution of communication in animals. the Internet is sometimes called an information continuum.