source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Abdominal%20obesity
Abdominal obesity, also known as central obesity and truncal obesity, is the human condition of an excessive concentration of visceral fat around the stomach and abdomen to such an extent that it is likely to harm its bearer's health. Abdominal obesity has been strongly linked to cardiovascular disease, Alzheimer's disease, and other metabolic and vascular diseases. Visceral and central abdominal fat and waist circumference show a strong association with type 2 diabetes. Visceral fat, also known as organ fat or intra-abdominal fat, is located inside the peritoneal cavity, packed in between internal organs and torso, as opposed to subcutaneous fat, which is found underneath the skin, and intramuscular fat, which is found interspersed in skeletal muscle. Visceral fat is composed of several adipose depots including mesenteric, epididymal white adipose tissue (EWAT), and perirenal fat. An excess of adipose visceral fat is known as central obesity, the "pot belly" or "beer belly" effect, in which the abdomen protrudes excessively. This body type is also known as "apple shaped", as opposed to "pear shaped" in which fat is deposited on the hips and buttocks. Researchers first started to focus on abdominal obesity in the 1980s when they realized it had an important connection to cardiovascular disease, diabetes, and dyslipidemia. Abdominal obesity was more closely related with metabolic dysfunctions connected with cardiovascular disease than was general obesity. In the late 1980s and early 1990s insightful and powerful imaging techniques were discovered that would further help advance the understanding of the health risks associated with body fat accumulation. Techniques such as computed tomography and magnetic resonance imaging made it possible to categorize mass of adipose tissue located at the abdominal level into intra-abdominal fat and subcutaneous fat. Abdominal obesity is linked with higher cardiovascular events among South Asian ethnic population. Health risks
https://en.wikipedia.org/wiki/Decacovirus
Decacovirus is a subgenus of viruses in the genus Alphacoronavirus. Species The genus consists of the following two species: Bat coronavirus HKU10 Rhinolophus ferrumequinum alphacoronavirus HuB-2013
https://en.wikipedia.org/wiki/Liquid%20air
Liquid air is air that has been cooled to very low temperatures (cryogenic temperatures), so that it has condensed into a pale blue mobile liquid. It is stored in specialized containers, such as vacuum flasks, to insulate it from room temperature. Liquid air can absorb heat rapidly and revert to its gaseous state. It is often used for condensing other substances into liquid and/or solidifying them, and as an industrial source of nitrogen, oxygen, argon, and other inert gases through a process called air separation. Properties Liquid air has a density of approximately . The density of a given air sample varies depending on the composition of that sample (e.g. humidity & concentration). Since dry gaseous air contains approximately 78% nitrogen, 21% oxygen, and 1% argon, the density of liquid air at standard composition is calculated by the percentage of the components and their respective liquid densities (see liquid nitrogen and liquid oxygen). Although air contains trace amounts of carbon dioxide (about 0.03%), carbon dioxide solidifies from the gas phase without passing through the intermediate liquid phase, and hence will not be present in liquid air at pressures less than . The boiling point of air is , intermediate between the boiling points of liquid nitrogen and liquid oxygen. However, it can be difficult to keep at a stable temperature as the liquid boils, since the nitrogen will boil off first, leaving the mixture oxygen-rich and changing the boiling point. This may also occur in some circumstances due to the liquid air condensing oxygen out of the atmosphere. Liquid air starts to freeze at approximately , precipitating nitrogen-rich solid (but with appreciable amount of oxygen in solid solution). Unless the oxygen is previously accommodated in the solid solution, the eutectic freezes at 50 K. Preparation Principle of production The constituents of air were once known as "permanent gases", as they could not be liquified solely by compression at room
https://en.wikipedia.org/wiki/Simultaneous%20game
In game theory, a simultaneous game or static game is a game where each player chooses their action without knowledge of the actions chosen by other players. Simultaneous games contrast with sequential games, which are played by the players taking turns (moves alternate between players). In other words, both players normally act at the same time in a simultaneous game. Even if the players do not act at the same time, both players are uninformed of each other's move while making their decisions. Normal form representations are usually used for simultaneous games. Given a continuous game, players will have different information sets if the game is simultaneous than if it is sequential because they have less information to act on at each step in the game. For example, in a two player continuous game that is sequential, the second player can act in response to the action taken by the first player. However, this is not possible in a simultaneous game where both players act at the same time. Characteristics In sequential games, players observe what rivals have done in the past and there is a specific order of play. However, in simultaneous games, all players select strategies without observing the choices of their rivals and players choose at the exact same time. A simple example is rock-paper-scissors in which all players make their choice at the exact same time. However moving at exactly the same time isn’t always taken literally, instead players may move without being able to see the choices of other players. A simple example is an election in which not all voters will vote literally at the same time but each voter will vote not knowing what anyone else has chosen. Given that decision makers are rational, then so is individual rationality. An outcome is individually rational if it yields each player at least his security level. The security level for Player i is the amount max min Hi (s) that the player can guarantee themselves unilaterally, that is, without consid
https://en.wikipedia.org/wiki/Tre%20recombinase
Tre recombinase is an experimental enzyme that in lab tests has removed DNA inserted by HIV from infected cells. Through selective mutation, Cre recombinase which recognizes loxP sites are modified to identify HIV long terminal repeats (loxLTR) instead. As a result, instead of performing Cre-Lox recombination, the new enzyme performs recombination at HIV provirus sites. The structure of Tre in complex with loxLTR has been resolved (), allowing for analyzing the roles of individual mutations.
https://en.wikipedia.org/wiki/Software%20peer%20review
In software development, peer review is a type of software review in which a work product (document, code, or other) is examined by author's colleagues, in order to evaluate the work product's technical content and quality. Purpose The purpose of a peer review is to provide "a disciplined engineering practice for detecting and correcting defects in software artifacts, and preventing their leakage into field operations" according to the Capability Maturity Model. When performed as part of each Software development process activity, peer reviews identify problems that can be fixed early in the lifecycle. That is to say, a peer review that identifies a requirements problem during the Requirements analysis activity is cheaper and easier to fix than during the Software architecture or Software testing activities. The National Software Quality Experiment, evaluating the effectiveness of peer reviews, finds, "a favorable return on investment for software inspections; savings exceeds costs by 4 to 1". To state it another way, it is four times more costly, on average, to identify and fix a software problem later. Distinction from other types of software review Peer reviews are distinct from management reviews, which are conducted by management representatives rather than by colleagues, and for management and control purposes rather than for technical evaluation. They are also distinct from software audit reviews, which are conducted by personnel external to the project, to evaluate compliance with specifications, standards, contractual agreements, or other criteria. Review processes Peer review processes exist across a spectrum of formality, with relatively unstructured activities such as "buddy checking" towards one end of the spectrum, and more Informal approaches such as walkthroughs, technical peer reviews, and software inspections, at the other. The IEEE defines formal structures, roles, and processes for each of the last three. Management representatives ar
https://en.wikipedia.org/wiki/10G
10G is a term used by some cable Internet access providers and industry groups in the United States in reference to broadband networks with a maximum potential download rate of ten gigabits per second (10 Gbit/s). The term was first used in this regard by industry association NCTA in January 2019, which said it had filed for a trademark on the term, and expanded on by CableLabs in a summer 2019 white paper. The term "10G" has no connection to the numbered generations of cellular network standards such as 5G (fifth generation). Some articles discussing the term have posited that 10G suggests to casual readers that service would be twice as fast as 5G, when in fact the 5G standard already encompasses even faster speeds of up to 20 Gbit/s. In early 2023, Comcast began referring to its Xfinity Internet service as now being on a "10G network", despite the fact that the top-speed service available in the vast majority of homes served by Comcast was still only 1 Gbit/s. In October 2023, the National Advertising Division (NAD) of the Better Business Bureau ruled that it considered Comcast's use of 10G to be false or misleading, as it constituted an express claim that Comcast was using a tenth-generation network or was promising 10 Gbit/s speeds to all customers. The NAD recommended that Comcast should either discontinue its claims or clarify 10G as an "aspirational" technology. Comcast said it would appeal the decision by the self-regulatory body. See also 10 Gigabit Ethernet – a set of technologies for Ethernet communications that support up to 10 Gbit/s speeds 10G-PON and 10G-EPON – passive optical network standards that support up to 10 Gbit/s speeds
https://en.wikipedia.org/wiki/National%20Society%20of%20Hispanic%20Physicists
The National Society of Hispanic Physicists (NSHP) was established in 1996 with the goal of promoting the participation and advancement of Hispanic-Americans in physics and celebrating the contributions of Hispanic-American physicists to the study and teaching of physics. Brief history A grant from the Sloan Foundation was awarded to the Pan-American Association for Physics to establish the National Society of Hispanic Physicists (NSHP). The Founding Meeting of the Society was held in Austin at the University of Texas in April, 1996 and the first annual meeting was held in Houston, Texas in October 1997. The US-Mexico Workshop on Teaching Introductory Physics, the first major project undertaken by the Society, was held later that year in Monterrey, Mexico. The project was a bilingual joint venture between the NSHP and the Instituto Tecnológico y de Estudios Superiores de Monterrey (ITESM) to explore the goals of the introductory physics sequence and recent pedagogical developments to meet those goals. The Hispanic Physicist, the official newsletter of the NSHP, has been in publication since 1997. The NSHP meets jointly with other societies organizing sessions, hosting social functions, promoting discussions of diversity and inclusion issues in the physics community, and recognizing achievements of Hispanic-American physics students and faculty. The National Society of Hispanic Physicists has met annually with the Society for the Advancement of Chicanos and Native Americans in Science (SACNAS) since 1997 and twice with the American Association of Physics Teachers (AAPT) (in Austin, TX in 2003 and Albuquerque, NM in 2005). In addition, the NSHP has met at sectional meetings of the American Physical Society (APS) and the American Astronomical Society (AAS). More recently the Society has met annually with the National Society of Black Physicists. The organization was incorporated under the umbrella of the Southeastern Universities Research Association (SURA) in Aug
https://en.wikipedia.org/wiki/Timlovirales
Timlovirales is an order of viruses, which infect prokaryotes. Most of these bacteriophages were discovered by metagenomics. Taxonomy Timlovirales contains the following two families: Blumeviridae Steitzviridae
https://en.wikipedia.org/wiki/Storage%20effect
The storage effect is a coexistence mechanism proposed in the ecological theory of species coexistence, which tries to explain how such a wide variety of similar species are able to coexist within the same ecological community or guild. The storage effect was originally proposed in the 1980s to explain coexistence in diverse communities of coral reef fish, however it has since been generalized to cover a variety of ecological communities. The theory proposes one way for multiple species to coexist: in a changing environment, no species can be the best under all conditions. Instead, each species must have a unique response to varying environmental conditions, and a way of buffering against the effects of bad years. The storage effect gets its name because each population "stores" the gains in good years or microhabitats (patches) to help it survive population losses in bad years or patches. One strength of this theory is that, unlike most coexistence mechanisms, the storage effect can be measured and quantified, with units of per-capita growth rate (offspring per adult per generation). The storage effect can be caused by both temporal and spatial variation. The temporal storage effect (often referred to as simply "the storage effect") occurs when species benefit from changes in year-to-year environmental patterns, while the spatial storage effect occurs when species benefit from variation in microhabitats across a landscape. The concept For the storage effect to operate, it requires variation (i.e. fluctuations) in the environment and thus can be termed a "fluctuation-dependent mechanism". This variation can come from a large degree of factors, including resource availability, temperature, and predation levels. However, for the storage effect to function, this variation must change the birth, survival, or recruitment rate of species from year to year (or patch to patch). For competing species within the same community to coexist, they have to meet one fund
https://en.wikipedia.org/wiki/Triad%20%28anatomy%29
In the histology of skeletal muscle, a triad is the structure formed by a T tubule with a sarcoplasmic reticulum (SR) known as the terminal cisterna on either side. Each skeletal muscle fiber has many thousands of triads, visible in muscle fibers that have been sectioned longitudinally. (This property holds because T tubules run perpendicular to the longitudinal axis of the muscle fiber.) In mammals, triads are typically located at the A-I junction; that is, the junction between the A and I bands of the sarcomere, which is the smallest unit of a muscle fiber. Triads form the anatomical basis of excitation-contraction coupling, whereby a stimulus excites the muscle and causes it to contract. A stimulus, in the form of positively charged current, is transmitted from the neuromuscular junction down the length of the T tubules, activating dihydropyridine receptors (DHPRs). Their activation causes 1) a negligible influx of calcium and 2) a mechanical interaction with calcium-conducting ryanodine receptors (RyRs) on the adjacent SR membrane. Activation of RyRs causes the release of calcium from the SR, which subsequently initiates a cascade of events leading to muscle contraction. These muscle contractions are caused by calcium's bonding to troponin and unmasking the binding sites covered by the troponin-tropomyosin complex on the actin myofilament and allowing the myosin cross-bridges to connect with the actin. Function: Helps in muscle contraction and Ca+ Secretion See also Diad, a homologous structure in cardiac muscle
https://en.wikipedia.org/wiki/Major%20histocompatibility%20complex%2C%20class%20II%2C%20DQ%20alpha%201
Major histocompatibility complex, class II, DQ alpha 1, also known as HLA-DQA1, is a human gene present on short arm of chromosome 6 (6p21.3) and also denotes the genetic locus which contains this gene. The protein encoded by this gene is one of two proteins that are required to form the DQ heterodimer, a cell surface receptor essential to the function of the immune system. Function HLA-DQA1 belongs to the HLA class II alpha chain paralogues. This class II molecule is a heterodimer consisting of an alpha (DQA) and a beta chain (DQB), both anchored in the membrane. It plays a central role in the immune system by presenting peptides derived from extracellular proteins. Class II molecules are expressed in antigen-presenting cells (APC: B lymphocytes, dendritic cells, macrophages). Gene structure and polymorphisms The alpha chain contains 5 exons. Exon one encodes the leader peptide, exons 2 and 3 encode the two extracellular protein domains, exon 4 encodes the transmembrane domain and the cytoplasmic tail. Within the DQ molecule both the alpha chain and the beta chain contain the polymorphisms specifying the peptide binding specificities, resulting in up to 4 different molecules. Typing for these polymorphisms is routinely done for bone marrow transplantation. Alleles DQ1 There are four commonly encountered DQA1 alleles: DQA1*0101, *0102, *0103, *0104. These alleles are always found in haplotypes with HLA-DQB1*05 (DQ5) and HLA-DQB1*06 (DQ6). DQ1 is a serotype, rare among serotypes for human class II antigens, in that the antibodies to DQ1 react to the alpha chain of HLA DQ, these DQA1 allele gene products. Other The other DQA1 alleles have no defined serotype. There are 5 groups, DQA1*02, *03, *04, *05, *06. DQA1 within these groups are either invariant or produce the same α-chain subunit. DQA1*02 and DQA1*06 contain only one allele. DQA1*03 has three alleles which each produce nearly identical α3. For DQA1*05, the DQA1*0501 and DQA1*0505 produce identical α5.
https://en.wikipedia.org/wiki/Florian%20Neukart
Florian Neukart is an Austrian business executive, computer scientist, physicist, and scientific author known for his work in quantum computing and artificial intelligence. He has primarily been working on utilizing quantum computers, artificial intelligence, and related technologies for solving industry problems. In his work on artificial intelligence, he describes methods for interpreting signals in the human brain in combination with paradigms from artificial intelligence to create artificial conscious entities. Biography Neukart holds a Ph.D. in computer science from the Transilvania University of Brasov and master's degrees in physics, information technology, and computer science from the Liverpool John Moores University, CAMPUS02 University of Applied Sciences and the Joanneum University of Applied Sciences. Work He is a member of the Board of Management at Terra Quantum AG, and previously worked as Director, Advanced Technologies and IT Innovation at Volkswagen Group of America, where he was concerned with research in the fields of quantum computing, quantum machine learning, artificial intelligence, and materials science. Neukart, born in Bruck/Mur, was also a member of the World Economic Forum's global future council on quantum computing, and an assistant professor for quantum computing at Leiden University. He is the author of the books "Reverse Engineering the Mind Consciously Acting Machines and Accelerated Evolution", in which he elaborates on establishing a symbiotic relationship between a biological brain, sensors, AI, and quantum hard- and software, resulting in solutions for the continuous consciousness problem as well as other state-of-the-art problems, and "Humankind's Hunger for Energy: The journey of a million years, from using flints to harvesting galaxies", in which he describes the evolution of humankind in terms of its energy consumption. He is the co-editor of the book "Chancen und Risiken der Quantentechnologien", in which the potentia
https://en.wikipedia.org/wiki/Thermal%20management%20of%20high-power%20LEDs
High power light-emitting diodes (LEDs) can use 350 milliwatts or more in a single LED. Most of the electricity in an LED becomes heat rather than light (about 70% heat and 30% light). If this heat is not removed, the LEDs run at high temperatures, which not only lowers their efficiency, but also makes the LED less reliable. Thus, thermal management of high power LEDs is a crucial area of research and development. It is necessary to limit both the junction and the phosphor particles temperatures to a value that will guarantee the desired LED lifetime. Thermal management is a universal problem having to do with power density, which occurs both at higher powers or in smaller devices. Many lighting applications wish to combine a high light flux with an extremely small light emitting substrate, causing concerns with LED power management to be particularly acute. Heat transfer procedure In order to maintain a low junction temperature to keep good performance of an LED, every method of removing heat from LEDs should be considered. Conduction, convection, and radiation are the three means of heat transfer. Typically, LEDs are encapsulated in a transparent polyurethane-based resin, which is a poor thermal conductor. Nearly all heat produced is conducted through the back side of the chip. Heat is generated from the p–n junction by electrical energy that was not converted to useful light, and conducted to outside ambience through a long path, from junction to solder point, solder point to board, and board to the heat sink and then to the atmosphere. A typical LED side view and its thermal model are shown in the figures. The junction temperature will be lower if the thermal impedance is smaller and likewise, with a lower ambient temperature. To maximize the useful ambient temperature range for a given power dissipation, the total thermal resistance from junction to ambient must be minimized. The values for the thermal resistance vary widely depending on the materi
https://en.wikipedia.org/wiki/Paris%20meridian
The Paris meridian is a meridian line running through the Paris Observatory in Paris, France – now longitude 2°20′14.02500″ East. It was a long-standing rival to the Greenwich meridian as the prime meridian of the world. The "Paris meridian arc" or "French meridian arc" (French: la Méridienne de France) is the name of the meridian arc measured along the Paris meridian. The French meridian arc was important for French cartography, since the triangulations of France began with the measurement of the French meridian arc. Moreover, the French meridian arc was important for geodesy as it was one of the meridian arcs which were measured to determine the figure of the Earth via the arc measurement method. The determination of the figure of the Earth was a problem of the highest importance in astronomy, as the diameter of the Earth was the unit to which all celestial distances had to be referred. History French cartography and the figure of the Earth In the year 1634, France ruled by Louis XIII and Cardinal Richelieu, decided that the Ferro Meridian through the westernmost of the Canary Islands should be used as the reference on maps, since El Hierro (Ferro) was the most western position of the Ptolemy's world map. It was also thought to be exactly 20 degrees west of Paris. The astronomers of the French Academy of Sciences, founded in 1666, managed to clarify the position of El Hierro relative to the meridian of Paris, which gradually supplanted the Ferro meridian. In 1666, Louis XIV of France had authorized the building of the Paris Observatory. On Midsummer's Day 1667, members of the Academy of Sciences traced the future building's outline on a plot outside town near the Port Royal abbey, with Paris meridian exactly bisecting the site north–south. French cartographers would use it as their prime meridian for more than 200 years. Old maps from continental Europe often have a common grid with Paris degrees at the top and Ferro degrees offset by 20 at the bottom. A Fr
https://en.wikipedia.org/wiki/Fault%20Tolerant%20Ethernet
Fault Tolerant Ethernet (FTE) is proprietary protocol created by Honeywell. Designed to provide rapid network redundancy, on top of spanning tree protocol. Each node is connected twice to a single LAN through the dual network interface controllers. The driver and the FTE enabled components allow network communication to occur over an alternate path when the primary path fails. Default time before failure is detected, is Diagnostic Interval (1000ms) multiplier with Disjoin Multiplier (3), for a 3000ms recovery time. Similar to Switch Fault Tolerance (SFT) in windows and mode=1 (active-backup) in Linux. Supported hardware and software Windows 7/2003 or newer Honeywell Control Firewall (CF9) Honeywell C300 Controller Honeywell Series 8 I/O Technical overview Uses Multicast ( 234.5.6.7), for FTE community. Recommended maximum of 300 FTE nodes and 200 single connected Ethernet nodes (A machine with to network cards is considered as two separate single connected Ethernet nodes). Recommended to have separate broadcast/multicast domain , for different FTE communities. Recommended maximum of 3 tiers of switches. Default UDP Source Port: 47837 Default UDP Destination Port : 51966
https://en.wikipedia.org/wiki/Cytochrome%20P450%20eryF
6-deoxyerythronolide B hydroxylase is an Actinomycetota Cytochrome P450 enzyme originally from Saccharopolyspora erythraea, catalyzes the 6S-hydroxylate of 6-deoxyerythronolide B (6-DEB) to erythronolide B (EB) which is the first step of biosynthesis of the macrolide antibiotic erythromycin. This bacterial enzyme belongs to CYP family CYP107, with the CYP Symbol CYP107A1.
https://en.wikipedia.org/wiki/Lazy%20user%20model
The lazy user model of solution selection (LUM) is a model in information systems proposed by Tétard and Collan that tries to explain how an individual selects a solution to fulfill a need from a set of possible solution alternatives. LUM expects that a solution is selected from a set of available solutions based on the amount of effort the solutions require from the user – the user is supposed to select the solution that carries the least effort. The model is applicable to a number of different types of situations, but it can be said to be closely related to technology acceptance models. The model draws from earlier works on how least effort affects human behaviour in information seeking and in scaling of language. Earlier research within the discipline of information systems especially within the topic of technology acceptance and technology adoption is closely related to the lazy user model. The model structure The model starts from the observation that there is a "user need", i.e. it is expected that there is a "clearly definiable, fully satisfiable want" that the user wants satisfied (it can also be said that the user has a problem that he/she wants solved). So there is a place for a solution, product, or service. The user need defines the set of possible solutions (products, services etc.) that fulfill the user need. The basic model considers for simplicity needs that are 100% satisfiable and services that 100% satisfy the needs. This means that only the solutions that solve the problem are relevant. This logically means that the need defines the possible satisfying solutions – a set of solutions (many different products/services) that all can fulfill the user need. LUM is not limited to looking at one solution separately. All of the solutions in the set that fulfill the need have their own characteristics; some are good and suitable for the user, others unsuitable and unacceptable – for example, if the user is in a train and wants to know what the result
https://en.wikipedia.org/wiki/WinHelp
Microsoft WinHelp is a proprietary format for online help files that can be displayed by the Microsoft Help browser winhelp.exe or winhlp32.exe. The file format is based on Rich Text Format (RTF). It remained a popular Help platform from Windows 3.0 through Windows XP. WinHelp was removed in Windows Vista purportedly to discourage software developers from using the obsolete format and encourage use of newer help formats. Support for WinHelp files would eventually be removed entirely in Windows 10. History 1990 – WinHelp 1.0 shipped with Windows 3.0. 1995 – WinHelp 4.0 shipped with Windows 95 / Windows NT. 2006 – Microsoft announced its intentions to phase out WinHelp as a supported platform. WinHelp is not part of Windows Vista out of the box. WinHelp files come in 16 bit and 32 bit types. Vista treats these files types differently. When starting an application that uses the 32 bit .hlp format, Windows warns that the format is no longer supported. A downloadable viewer for 32 bit .hlp files is available from the Microsoft Download Center. The 16 bit WinHelp files continue to display in Windows Vista (32 bit only) without the viewer download. January 9, 2009 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows Server 2008 at the Microsoft Download Center. October 14, 2009 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows 7 and Windows Server 2008 R2 at the Microsoft Download Center. October 26, 2012 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows 8 at the Microsoft Download Center. November 5, 2013 – Microsoft announced the availability of Windows Help program (WinHlp32.exe) for Windows 8.1 at the Microsoft Download Center. July 15, 2015 - Microsoft completely removed Windows Help from Windows 10. Attempting to open a .hlp file just brings users to a help page detailing that it was removed. File format A WinHelp file has a ".hlp" suffix. It
https://en.wikipedia.org/wiki/Phagosome
In cell biology, a phagosome is a vesicle formed around a particle engulfed by a phagocyte via phagocytosis. Professional phagocytes include macrophages, neutrophils, and dendritic cells (DCs). A phagosome is formed by the fusion of the cell membrane around a microorganism, a senescent cell or an apoptotic cell. Phagosomes have membrane-bound proteins to recruit and fuse with lysosomes to form mature phagolysosomes. The lysosomes contain hydrolytic enzymes and reactive oxygen species (ROS) which kill and digest the pathogens. Phagosomes can also form in non-professional phagocytes, but they can only engulf a smaller range of particles, and do not contain ROS. The useful materials (e.g. amino acids) from the digested particles are moved into the cytosol, and waste is removed by exocytosis. Phagosome formation is crucial for tissue homeostasis and both innate and adaptive host defense against pathogens. However, some bacteria can exploit phagocytosis as an invasion strategy. They either reproduce inside of the phagolysosome (e.g. Coxiella spp.) or escape into the cytoplasm before the phagosome fuses with the lysosome (e.g. Rickettsia spp.). Many Mycobacteria, including Mycobacterium tuberculosis and Mycobacterium avium paratuberculosis, can manipulate the host macrophage to prevent lysosomes from fusing with phagosomes and creating mature phagolysosomes. Such incomplete maturation of the phagosome maintains an environment favorable to the pathogens inside it. Formation Phagosomes are large enough to degrade whole bacteria, or apoptotic and senescent cells, which are usually >0.5μm in diameter. This means a phagosome is several orders of magnitude bigger than an endosome, which is measured in nanometres. Phagosomes are formed when pathogens or opsonins bind to a transmembrane receptor, which are randomly distributed on the phagocyte cell surface. Upon binding, "outside-in" signalling triggers actin polymerisation and pseudopodia formation, which surrounds and fus
https://en.wikipedia.org/wiki/Executable%20architecture
An Executable Architecture (EA), in general, is the description of a system architecture (including software and/or otherwise) in a formal notation together with the tools (e.g. compilers/translators) that allow the automatic or semi-automatic generation of artifacts (e.g. capability gap analysis (CGA), models, software stubs, Military Scenario Definition Language (MSDL)) from that notation and which are used in the analysis, refinement, and/or the implementation of the architecture described. Closely related subjects Subjects closely related to EA include: Object Management Group's Model-driven architecture Object Management Group's Business Process Management Initiative Vanderbilt University's Model Integrated Computing (MIC) Implementations Implementations of EA include: Rational Rose Generic Modeling Environment (GME) Open-Source eGov Reference Architecture (OSERA) See also Business Process Execution Language (BPEL) Business Process Management Initiative (BPMI) Business Process Modeling Language (BPML) Executable Operational Architecture Model-driven architecture (MDA) Model-driven engineering (MDE) Object Management Group (OMG) Semantic Web Unified Process Unified Modeling Language (UML) Vanderbilt University
https://en.wikipedia.org/wiki/Protoflight
Protoflight is a portmanteau of "prototype" and "flight hardware". As defined by NASA Technical Standard NASA-STD-7002A, it refers to a strategy where no test-dedicated qualification article exists and all production (flight) hardware is intended for flight. An example of a program using protoflight methods is the Mars Orbiter Laser Altimeter project. A protoflight approach carries a higher technical risk approach compared to a full qualification test program since it has no demonstrated life capability over the anticipated life cycle of the hardware, but is a technology development design process that utilizes higher risk tolerances, agile management practices, and quick responsiveness that are needed for certain prototype flight projects or missions. Examples Atmospheric Reentry Demonstrator
https://en.wikipedia.org/wiki/Immunoreceptor%20tyrosine-based%20inhibitory%20motif
An immunoreceptor tyrosine-based inhibitory motif (ITIM), is a conserved sequence of amino acids that is found intracellularly in the cytoplasmic domains of many inhibitory receptors of the non-catalytic tyrosine-phosphorylated receptor family found on immune cells. These immune cells include T cells, B cells, NK cells, dendritic cells, macrophages and mast cells.  ITIMs have similar structures of S/I/V/LxYxxI/V/L, where x is any amino acid, Y is a tyrosine residue that can be phosphorylated, S is the amino acide Serine, I is the amino acid Isoleucine, and V is the amino acid Valine. ITIMs recruit SH2 domain-containing phosphatases, which inhibit cellular activation. ITIM-containing receptors often serve to target Immunoreceptor tyrosine-based activation motif(ITAM)-containing receptors, resulting in an innate inhibition mechanism within cells. ITIM bearing receptors have important role in regulation of immune system allowing negative regulation at different levels of the immune response. A list of human candidate ITIM-containing proteins has been generated by proteome-wide scans. It has included more than 135 proteins with ITIM motif. This list is reportedly expanded by studying rare human SNPs that create the consensus signature S/I/V/LxYxxI/V/L motif. Structure ITIM motifs are defined as six amino acid signature with the consensus sequence S/I/V/LxYxxL/V, where x stands for any amino acid, Y for a tyrosine residue that can be phosphorylated and S, I, V for amino acids serine, isoleucine, and valine, respectively. ITIM conserved sequence was first identified in the low affinity IgG receptor FcγRIIB. All ITIM containing receptors are a part of the immunoglobulin superfamily. Signaling by ITIM The main characteristic of ITIM-containing molecules is that they become tyrosyl-phopshorylated. In order to become phosphorylated, the inhibitory receptor has to be brought in close proximity to the kinase. This may be achieved by co-crosslinking with an ITAM motif o
https://en.wikipedia.org/wiki/Rotation%20distance
In discrete mathematics and theoretical computer science, the rotation distance between two binary trees with the same number of nodes is the minimum number of tree rotations needed to reconfigure one tree into another. Because of a combinatorial equivalence between binary trees and triangulations of convex polygons, rotation distance is equivalent to the flip distance for triangulations of convex polygons. Rotation distance was first defined by Karel Čulík II and Derick Wood in 1982. Every two -node binary trees have rotation distance at most , and some pairs of trees have exactly this distance. The computational complexity of computing the rotation distance is unknown. Definition A binary tree is a structure consisting of a set of nodes, one of which is designated as the root node, in which each remaining node is either the left child or right child of some other node, its parent, and in which following the parent links from any node eventually leads to the root node. (In some sources, the nodes described here are called "internal nodes", there exists another set of nodes called "external nodes", each internal node is required to have exactly two children, and each external node is required to have zero children. The version described here can be obtained by removing all the external nodes from such a tree.) For any node in the tree, there is a subtree of the same form, rooted at and consisting of all the nodes that can reach by following parent links. Each binary tree has a left-to-right ordering of its nodes, its inorder traversal, obtained by recursively traversing the left subtree (the subtree at the left child of the root, if such a child exists), then listing the root itself, and then recursively traversing the right subtree. In a binary search tree, each node is associated with a search key, and the left-to-right ordering is required to be consistent with the order of the keys. A tree rotation is an operation that changes the structure of a binary t
https://en.wikipedia.org/wiki/Mean-field%20game%20theory
Mean-field game theory is the study of strategic decision making by small interacting agents in very large populations. It lies at the intersection of game theory with stochastic analysis and control theory. The use of the term "mean field" is inspired by mean-field theory in physics, which considers the behavior of systems of large numbers of particles where individual particles have negligible impacts upon the system. In other words, each agent acts according to his minimization or maximization problem taking into account other agents’ decisions and because their population is large we can assume the number of agents goes to infinity and a representative agent exists. In traditional game theory, the subject of study is usually a game with two players and discrete time space, and extends the results to more complex situations by induction. However, for games in continuous time with continuous states (differential games or stochastic differential games) this strategy cannot be used because of the complexity that the dynamic interactions generate. On the other hand with MFGs we can handle large numbers of players through the mean representative agent and at the same time describe complex state dynamics. This class of problems was considered in the economics literature by Boyan Jovanovic and Robert W. Rosenthal, in the engineering literature by Minyi Huang, Roland Malhame, and Peter E. Caines and independently and around the same time by mathematicians and Pierre-Louis Lions. In continuous time a mean-field game is typically composed of a Hamilton–Jacobi–Bellman equation that describes the optimal control problem of an individual and a Fokker–Planck equation that describes the dynamics of the aggregate distribution of agents. Under fairly general assumptions it can be proved that a class of mean-field games is the limit as of an N-player Nash equilibrium. A related concept to that of mean-field games is "mean-field-type control". In this case, a social planner
https://en.wikipedia.org/wiki/PyBOP
PyBOP (benzotriazol-1-yloxytripyrrolidinophosphonium hexafluorophosphate) is a peptide coupling reagent used in solid phase peptide synthesis. It is used as a substitute for the BOP reagent - avoiding the formation of the carcinogenic waste product HMPA. See also BOP reagent DEPBT, a related reagent that contains no phosphorus-nitrogen bonds HATU HBTU
https://en.wikipedia.org/wiki/Pen-Strep
Pen-Strep (also known as penicillin-streptomycin) is a mixture of penicillin G and streptomycin that is widely used in mammalian cell culture media to prevent bacterial contamination. The solution contains 5,000 units of penicillin G (sodium salt) which acts as the active base, and 5,000 micrograms of streptomycin (sulfate) (base per milliliter), formulated in 0.85% saline. In general, 50-100 units of Pen-Strep per milliliter of media is used to avoid contamination in cell culture. Thus, the retail product is generally 100 times more concentrated.  It is recommended for use in cell culture applications at a concentration of 10 ml per liter. It is the most common antibiotic solution for the culture of mammalian cells and it does not have any adverse effects on the cells themselves. It was first introduced in 1955 in cell culture. Penicillin is a beta-lactam antibiotic that is effective in inhibiting Gram-positive bacteria, whereas streptomycin is an aminoglycoside antibiotic which is effective against most Gram-negative bacteria. Advantages and disadvantages Penicillin is a narrow spectrum antibiotic against Gram-positive bacteria and is relatively inexpensive, but penicillin can cause hypersensitivity reactions and is unstable in acid and alkaline pH environments. Streptomycin is a broad spectrum antibiotic against Gram negative bacteria and is also relatively inexpensive, but can be rapidly destroyed at alkaline pH environments. The cell culture media is generally neutral in pH which is why Pen-Strep is still widely used in mammalian cell culture. For the purpose of contamination control in cell culture, gentamicin sulfate or kanamycin sulfate are also commonly used.
https://en.wikipedia.org/wiki/Komarov%20Botanical%20Institute
The Komarov Botanical Institute of the Russian Academy of Sciences () is a leading botanical institution in Russia, It is located on Aptekarsky Island in St. Petersburg, and is named after the Russian botanist Vladimir Leontyevich Komarov (1869–1945). The institute was established in 1931 as merger of the Botanical Garden and the Botanical Museum of the Academy of Sciences. The institute hosts Saint Petersburg Botanical Garden as well as herbarium collections that house over seven million specimens of plants and fungi. The Institute's LE Herbarium has approximately 200,000-250 000 specimens from Latin America in storage. The most known collections from Latin America include those of German naturalist G. Langsdorff (collected between 1814-1817 & 1821-1829) and German botanist Ludwig Riedel (1821-1828 & 1831-1836) from Brazil, G. Mertens (1826) and German naturalist Wilhelm Friedrich Karwinsky von Karwin (1841-1843) from Mexico, and plants collected in expeditions of the Soviet Rubber Trust in 1926-1928. Among the scientists who have worked there are Yevgenia Georgievna Pobedimova. See also List of Russian biologists
https://en.wikipedia.org/wiki/Stickers%20as%20a%20Service
Stickers as a service (SaaS) is a category of cloud computing services that provides digital stickers to messaging apps and photo editing software. In this model, the stickers are sold through stores integrated in messaging apps, chat and photo editing software. They can be used as a method to generate revenue besides through in-app advertising as they are non-intrusive. SaaS offerings may include handling micro payments, sticker shop integration and content management systems with developing apps, and is similar to Software as a Service (SaaS) and Platform as a Service (PaaS). Stickers as a service enables developers to build a full sticker stores into their apps with just a few lines of code. Stickers Digital stickers, commonly sold as virtual goods, tend to depict either original or well-known characters. They are used during chat sessions between users and act as large sized emoji to express emotions or as decoration in photographs. Stickers vary from original characters to popular manga, anime and gaming characters or movie tie-ins. They can be purchased or downloaded for free in apps with sticker stores installed. Limited edition stickers may also be gained via special occasions or achievements. It is used as an alternative to emoji or emoticons as they are a more expressive form of communication and have a variety of designs and art styles. Apps that use Stickers LINE Kakaotalk Facebook Messenger WhatsApp Instagram (Messages and Stories) Snapchat (Messages and Stories) Viber BlackBerry PhotoStamped See also Virtual economy Gamification Purikura
https://en.wikipedia.org/wiki/Goormaghtigh%20conjecture
In mathematics, the Goormaghtigh conjecture is a conjecture in number theory named for the Belgian mathematician René Goormaghtigh. The conjecture is that the only non-trivial integer solutions of the exponential Diophantine equation satisfying and are and Partial results showed that, for each pair of fixed exponents and , this equation has only finitely many solutions. But this proof depends on Siegel's finiteness theorem, which is ineffective. showed that, if and with , , and , then is bounded by an effectively computable constant depending only on and . showed that for and odd , this equation has no solution other than the two solutions given above. Balasubramanian and Shorey proved in 1980 that there are only finitely many possible solutions to the equations with prime divisors of and lying in a given finite set and that they may be effectively computed. showed that, for each fixed and , this equation has at most one solution. For fixed x (or y), equation has at most 15 solutions, and at most two unless x is either odd prime power times a power of two, or in the finite set {15, 21, 30, 33, 35, 39, 45, 51, 65, 85, 143, 154, 713}, in which case there are at most three solutions. Furthermore, there is at most one solution if the odd part of n is squareful unless n has at most two distinct odd prime factors or n is in a finite set {315, 495, 525, 585, 630, 693, 735, 765, 855, 945, 1035, 1050, 1170, 1260, 1386, 1530, 1890, 1925, 1950, 1953, 2115, 2175, 2223, 2325, 2535, 2565, 2898, 2907, 3105, 3150, 3325, 3465, 3663, 3675, 4235, 5525, 5661, 6273, 8109, 17575, 39151}. Application to repunits The Goormaghtigh conjecture may be expressed as saying that 31 (111 in base 5, 11111 in base 2) and 8191 (111 in base 90, 1111111111111 in base 2) are the only two numbers that are repunits with at least 3 digits in two different bases. See also Feit–Thompson conjecture
https://en.wikipedia.org/wiki/Thai%20baht
The baht (; , ; sign: ฿; code: THB) is the official currency of Thailand. It can be divided into 100 satang (, ). The issuance of currency is the responsibility of the Bank of Thailand. SWIFT ranked the Thai baht as the 10th-most-frequently used world payment currency as of January 2019. History The Thai baht, like the pound, originated from a traditional unit of mass. Its currency value was originally expressed as that of silver of corresponding weight (now defined as 15 grams), and was in use probably as early as the Sukhothai period in the form of bullet coins known in Thai as phot duang. These were pieces of solid silver cast to various weights corresponding to a traditional system of units related by simple fractions and multiples, one of which is the baht. These are listed in the following table: That system was in use up until 1897, when the decimal system devised by Prince Jayanta Mongkol, in which one baht = 100 satang, was introduced by his half-brother King Chulalongkorn along with the demonetization of silver bullet coins on 28 October 1904 after the end of silver bullet coin production by the opening of Sitthikarn Royal Mint in 1857. However, coins denominated in the old units were issued until 1910, and the amount of 25 satang is still commonly referred to as a , as is the 25-satang coin. Until 27 November 1902, the baht was fixed on a purely silver basis, with 15 grams of silver to the baht. This caused the value of the currency to vary relative to currencies on a gold standard. From 1856 to 1864, the values of certain foreign silver coins were fixed by law, with 5 baht = 3 Spanish dollar = 7 Indian rupees. Before 1880 the exchange rate was fixed at 8 baht per pound sterling, falling to 10 to the pound during the 1880s. In 1902, the government began to increase the value of the baht by following all increases in the value of silver against gold but not reducing it when the silver price fell. Beginning at 21.75 baht per pound sterling, the curr
https://en.wikipedia.org/wiki/Register%20of%20data%20controllers
The Register of data controllers was a United Kingdom database under the control of the UK Information Commissioner's Office mandated by the Data Protection Act 1998. The Register of fee payers is the new name of an equivalent register under the Data Protection Act 2018, which implements the European Union's General Data Protection Regulation (GDPR). Registration under both Acts carries a fee, the proceeds of which fund the UK Information Commissioner's Office. Any entry may be inspected by the public at any time at no cost to the enquirer. Data Protection Act 1998 Under the 1998 Act, the name of the data controller was recorded with the purpose(s) for the processing of the data processed by that controller within the meaning of The Act. A data controller may, under some circumstances, be exempt from registration (previously termed notification). When not exempt, failure to notify the Information Commissioner's Office formally before the start of processing data was a strict liability offence for which a prosecution may be brought by the Information Commissioner's Office in the criminal court of the UK. Failure to notify was a criminal offence unless exempt. Exemption from registration does not exempt a data controller from compliance with The Act. Amendments to a data controller's notification could be made at any time, and must have been made before the start of a new processing purpose. Data Protection Act 2018 Under the 2018 Act, the register is called the Register of fee payers, and the purposes for processing are nor supplied, though other trading names and the name of a Data Protection Officer may be given. The enforcement of the Act by the Information Commissioner's Office is supported by a data protection charge on UK data controllers under the Data Protection (Charges and Information) Regulations 2018. Exemptions from the charge were left broadly the same as for 1998 Act: largely some businesses and non-profits internal core purposes (staff or m
https://en.wikipedia.org/wiki/Holstein%E2%80%93Herring%20method
The Holstein–Herring method, also called the surface Integral method, or Smirnov's method is an effective means of getting the exchange energy splittings of asymptotically degenerate energy states in molecular systems. Although the exchange energy becomes elusive at large internuclear systems, it is of prominent importance in theories of molecular binding and magnetism. This splitting results from the symmetry under exchange of identical nuclei (Pauli exclusion principle). The basic idea pioneered by Theodore Holstein, Conyers Herring and Boris M. Smirnov in the 1950-1960. Theory The method can be illustrated for the hydrogen molecular ion or more generally, atom-ion systems or one-active electron systems, as follows. We consider states that are represented by even or odd functions with respect to behavior under space inversion. This is denoted with the suffixes g and u from the German gerade and ungerade and are standard practice for the designation of electronic states of diatomic molecules, whereas for atomic states the terms even and odd are used. The electronic time-independent Schrödinger equation can be written as: where E is the (electronic) energy of a given quantum mechanical state (eigenstate), with the electronic state function depending on the spatial coordinates of the electron and where is the electron-nuclear Coulomb potential energy function. For the hydrogen molecular ion, this is: For any gerade (or even) state, the electronic Schrödinger wave equation can be written in atomic units () as: For any ungerade (or odd) state, the corresponding wave equation can be written as: For simplicity, we assume real functions (although the result can be generalized to the complex case). We then multiply the gerade wave equation by on the left and the ungerade wave equation on the left by and subtract to obtain: where is the exchange energy splitting. Next, without loss of generality, we define orthogonal single-particle functions, and , located at th
https://en.wikipedia.org/wiki/Bidirectional%20transformation
In computer programming, bidirectional transformations (bx) are programs in which a single piece of code can be run in several ways, such that the same data are sometimes considered as input, and sometimes as output. For example, a bx run in the forward direction might transform input I into output O, while the same bx run backward would take as input versions of I and O and produce a new version of I as its output. Bidirectional model transformations are an important special case in which a model is input to such a program. Some bidirectional languages are bijective. The bijectivity of a language is a severe restriction of its power, because a bijective language is merely relating two different ways to present the very same information. More general is a lens language, in which there is a distinguished forward direction ("get") that takes a concrete input to an abstract output, discarding some information in the process: the concrete state includes all the information that is in the abstract state, and usually some more. The backward direction ("put") takes a concrete state and an abstract state and computes a new concrete state. Lenses are required to obey certain conditions to ensure sensible behaviour. The most general case is that of symmetric bidirectional transformations. Here the two states that are related typically share some information, but each also includes some information that is not included in the other. Usage Bidirectional transformations can be used to: Maintain the consistency of several sources of information Provide an 'abstract view' to easily manipulate data and write them back to their source Definition Bidirectional transformations fall into various well-studied categories. A lens is a pair of functions , relating a source and a view . If these functions obey the three lens laws: PutGet: GetPut: PutPut: It is called a well-behaved lens. A related notion is that of a prism, in which the signatures of the functions a
https://en.wikipedia.org/wiki/Augmented%20renal%20clearance
In pharmacology, augmented renal clearance (ARC) is a phenomenon where certain critically ill patients may display increased clearance of a medication through the kidneys. In many cases, it is observed as a measured creatinine clearance above that which is expected given the patient's age, gender, and other factors. The phenomenon is most commonly observed in patients with neurologic damage, sepsis, major trauma, or burns. Augmented renal clearance can be caused by increased fluid administration, certain medications, and critical illnesses. It can lead to failure of treatment in people due to a decrease in drug concentrations, increase in clearance, or shorter half life. Many medications require adjustment to account for the changed clearance in people with ARC, notably some antibiotics. History Normal kidney function measured by creatinine clearance varies in different populations based on age, gender, race, fluid balance, and other factors, but also can be affected by diseases themselves. This makes it challenging to assign an objective number or scale to kidney function. For this reason, kidney function (and thus medication elimination) has been approximated by measuring creatinine clearance, or calculating an estimated glomerular filtration rate (eGFR), since 1976. Beginning in the late 1970s, an increase in the creatinine clearance had been observed in burn patients. This led to the realization that some burn patients required higher than expected doses of aminoglycosides to obtain the same serum concentration of drug. Diagnosis The primary sign of augmented renal clearance is an increase in the creatinine clearance well above that which would be considered normal. Commonly, ARC is defined as a creatinine clearance of greater than 130 mL/min, but the effects of increased clearance on therapy are not directly correlated to a specific number. For this reason, lower cutoffs such as 120 mL/min are used by some, as well as higher cutoffs in young people who typic
https://en.wikipedia.org/wiki/Logical%20form
In logic, logical form of a statement is a precisely-specified semantic version of that statement in a formal system. Informally, the logical form attempts to formalize a possibly ambiguous statement into a statement with a precise, unambiguous logical interpretation with respect to a formal system. In an ideal formal language, the meaning of a logical form can be determined unambiguously from syntax alone. Logical forms are semantic, not syntactic constructs; therefore, there may be more than one string that represents the same logical form in a given language. The logical form of an argument is called the argument form of the argument. History The importance of the concept of form to logic was already recognized in ancient times. Aristotle, in the Prior Analytics, was probably the first to employ variable letters to represent valid inferences. Therefore, Jan Łukasiewicz claims that the introduction of variables was "one of Aristotle's greatest inventions." According to the followers of Aristotle like Ammonius, only the logical principles stated in schematic terms belong to logic, and not those given in concrete terms. The concrete terms man, mortal, and so forth are analogous to the substitution values of the schematic placeholders A, B, C, which were called the "matter" (Greek hyle, Latin materia) of the argument. The term "logical form" itself was introduced by Bertrand Russell in 1914, in the context of his program to formalize natural language and reasoning, which he called philosophical logic. Russell wrote: "Some kind of knowledge of logical forms, though with most people it is not explicit, is involved in all understanding of discourse. It is the business of philosophical logic to extract this knowledge from its concrete integuments, and to render it explicit and pure." Example of argument form To demonstrate the important notion of the form of an argument, substitute letters for similar items throughout the sentences in the original argument. Origin
https://en.wikipedia.org/wiki/Chiral%20gauge%20theory
In quantum field theory, a chiral gauge theory is a quantum field theory with charged chiral (i.e. Weyl) fermions. For instance, the Standard Model is a chiral gauge theory. For topological reasons, chiral charged fermions cannot be given a mass without breaking the gauge symmetry, which will lead to inconsistencies unlike a global symmetry. It is notoriously difficult to construct a chiral gauge theory from a theory which does not already contain chiral fields at the fundamental level. A consistent chiral gauge theory must have no gauge anomaly (or global anomaly). Almost by necessity, regulators will have to break the gauge symmetry. This is responsible for gauge anomalies in the first place. Fermion doubling on a lattice Lattice regularizations suffer from fermion doublings leading to a loss of chirality. See also Chiral anomaly
https://en.wikipedia.org/wiki/Antonio%20Scarpa
Antonio Scarpa (9 May 1752 – 31 October 1832) was an Italian anatomist and professor. Biography Scarpa was born to an impoverished family in the frazione of Lorenzaga, Motta di Livenza, Veneto. An uncle, who was a member of the priesthood, gave him instruction until the age of 15, when he passed the entrance exam for the University of Padua. He was a pupil of Giovanni Battista Morgagni and Marc Antonio Caldani. Under the former, he became doctor of medicine on 19 May 1770; in 1772, he became professor at the University of Modena. For a time he chose to travel, visiting Holland, France and England. When he returned to Italy, he was made professor of anatomy at the University of Pavia in 1783, on the strong recommendation of Emperor Joseph II. His lectures were so popular with students that Emperor Joseph II commissioned Leopoldo Pollack to build a new anatomic theater, now called Aula Scarpa, inside the Old Campus of the University of Pavia. He remained in that post until 1804, when he stepped down to allow his student Santo Fattori to assume the chair. In May 1791, he was elected a Fellow of the Royal Society on account of being the "Author of some ingenious observations on the Ganglions of the Nerves, on the structure of the organs of hearing and smell, and other subjects of anatomy and Physiology" In 1805, Napoleon was made King of Italy. He chose to visit the University of Pavia, upon which he inquired as to the whereabouts of Dr. Scarpa. He was informed that the doctor had been dismissed because of his political opinions and his refusal to take oaths, whereupon Dr. Scarpa was restored to his position as the chair. In 1821, he was elected a foreign member of the Royal Swedish Academy of Sciences. During his lifetime he became a rich man, acquiring a collection of valuable paintings and living a wealthy lifestyle. He was a confirmed bachelor, and fathered several sons out of wedlock (whom he favoured through nepotism). In his career, he earned a reputation fo
https://en.wikipedia.org/wiki/Superfluid%20film
Superfluidity is a phenomenon where a fluid, or a fraction of a fluid, loses all its viscosity and can flow without resistance, which can form thin films. Superfluid helium, for example, forms a 30-nm-thick film on the surface of any container. The film's properties cause the helium to climb the walls of the container and, if this is not closed, flow out. Superfluidity and superconductivity are macroscopic manifestations of quantum mechanics. There is considerable interest, both theoretical and practical, in these quantum phase transitions. There has been a tremendous amount of work done in the field of phase transitions and critical phenomena in two dimensions. Much of the interest in this field is because as the number of dimensions increases, the number of exactly solvable models diminishes drastically. In three or more dimensions one must resort to a mean field theory approach. The theory of superfluid transitions in two dimensions is known as the Kosterlitz-Thouless (KT) theory. The 2D XY model - where the order parameter is characterized by an amplitude and a phase - is the universality class for this transition. Experimental methods In looking at phase transitions in thin films, specifically helium, the two main experimental signatures are the superfluid fraction and heat capacity. If either of these measurements were to be done on a superfluid film in a typical open container, the film signal would be overwhelmed by the background signal from the container. Therefore, when studying superfluid films, it is of paramount importance to study a system of large surface area as to enhance the film signal. There are several ways of doing this. In the first, a long thin strip of material such as PET film is rolled up into a "jelly roll" configuration. The result is a film that is a long continuous plane, referred to as a planar film. A second way is to have a highly porous material such as porous gold, Vycor, or Aerogel. This results in a multiply connected fi
https://en.wikipedia.org/wiki/Quantum%20dot%20cellular%20automaton
Quantum dot cellular automata (QDCA, sometimes referred to simply as quantum cellular automata, or QCA) are a proposed improvement on conventional computer design (CMOS), which have been devised in analogy to conventional models of cellular automata introduced by John von Neumann. Background Any device designed to represent data and perform computation, regardless of the physics principles it exploits and materials used to build it, must have two fundamental properties: distinguishability and conditional change of state, the latter implying the former. This means that such a device must have barriers that make it possible to distinguish between states, and that it must have the ability to control these barriers to perform conditional change of state. For example, in a digital electronic system, transistors play the role of such controllable energy barriers, making it extremely practical to perform computing with them. Cellular automata A cellular automaton (CA) is a discrete dynamical system consisting of a uniform (finite or infinite) grid of cells. Each cell can be in only one of a finite number of states at a discrete time. As time moves forward, the state of each cell in the grid is determined by a transformation rule that factors in its previous state and the states of the immediately adjacent cells (the cell's "neighborhood"). The most well-known example of a cellular automaton is John Horton Conway's "Game of Life", which he described in 1970. Quantum-dot cells Origin Cellular automata are commonly implemented as software programs. However, in 1993, Lent et al. proposed a physical implementation of an automaton using quantum-dot cells. The automaton quickly gained popularity and it was first fabricated in 1997. Lent combined the discrete nature of both cellular automata and quantum mechanics, to create nano-scale devices capable of performing computation at very high switching speeds (order of Terahertz) and consuming extremely small amounts of electrical
https://en.wikipedia.org/wiki/Mendelian%20susceptibility%20to%20mycobacterial%20disease
Mendelian susceptibility to mycobacterial disease (MSMD) is a rare genetic disease. It is a primary immunodeficiency featured by molecular defects in IL12/IFNγ dependent signalling pathway, leading to increased susceptibility to local or disseminated infections by environmental mycobacteria, Mycobacterium bovis Bacille Calmette-Guerin strain, nontyphoidal and typhoidal Salmonella serotypes. Symptoms and signs Normally patients who suffer from this disease are young children under 3 years which have also lack of response to IFN-γ cytokine replacement therapy. This disease is very rare and have high index of mortality. Following symptoms and signs are: recurrent wheezing dyspnea asthma-like symptoms recurring fever productive cough endobronchial mycobacterial infection low hemoglobin Patients with IFNγR1 deficiency can also suffer of disorders of the lung, parenchymal lung diseases caused by mycobacterial infections, hylar lymphadenopathy, or endobronchial disease. If these patients have nontubercular mycobacterial infection there should be suspicion for immunodeficiency. Transplantation of hematopoietic stem cell is the only one curative therapy for these patients. Children with partial MSMD usually have milder clinical phenotype, later onset, less severe infections, better response for IFNγ and antibiotic therapy, better survival rates and normally they don't need hematopoietic stem cell transplant. Pathophysiology Phagocytes are important components of the innate immune system for the body defence against infections by mycobacteria and other intracellular pathogens. The professional phagocytes include neutrophils, dendritic cells, macrophages and monocytes. These cells engulf the pathogens by phagocytosis and activate the adaptive immune system to facilitate the elimination of the infection. Cytokine signalling is the key for the interplay between the innate and adaptive limbs of the immune system, the most important of which is the IL12-dependent, IFN
https://en.wikipedia.org/wiki/Bitwarden
Bitwarden is a freemium open-source password management service that stores sensitive information such as website credentials in an encrypted vault. The platform offers a variety of client applications including a web interface, desktop applications, browser extensions, mobile apps, and a command-line interface. Bitwarden offers a free US or European cloud-hosted service as well as the ability to self-host. Desktop applications are available for Windows, MacOS, and Linux. Browser extensions include Chrome, Firefox, Safari, Edge, Opera, Vivaldi, Arc, Brave and Tor. Mobile apps for Android, iPhone, and iPad are available. Client functionalities include 2FA login, passwordless login, biometric unlock, random password generator, password strength testing tool, login/form/app autofill, syncing across unlimited platforms and devices, storing unlimited number of items, sharing credentials, and storing a variety of information including credit cards. Features Overall security Open-source codebase Zero-knowledge encryption, i.e., the company can't see the vault contents End-to-end encryption of the stored vault data Uses AES-CBC 256-bit to encrypt vault data, and PBKDF2 SHA-256 / Argon2id to derive user's encryption key from the entered password. Third-party independent application/code-library/network-infrastructure audits and bug bounty program Vault storage Cloud synchronization (Microsoft Azure)with free version being able to sync across unlimited platforms and devices Can choose either US cloud or European cloud, and migrate between them Can self-host the Bitwarden server on-premises, or with services such as DigitalOcean Availability Variety of client applications including a web interface, desktop applications, browser extensions, mobile apps, and a command-line interface. Desktop applications are available for Windows, MacOS, and Linux Browser extensions are available for Chrome, Firefox, Safari, Edge, Opera, Vivaldi, Brave and Tor. Mobile apps a
https://en.wikipedia.org/wiki/NADH%20dehydrogenase%20%28ubiquinone%29%2C%20alpha%201
NADH dehydrogenase [ubiquinone] 1 alpha subcomplex subunit 1 is a protein that in humans is encoded by the NDUFA1 gene. The NDUFA1 protein is a subunit of NADH dehydrogenase (ubiquinone), which is located in the mitochondrial inner membrane and is the largest of the five complexes of the electron transport chain. Mutations in the NDUFA1 gene are associated with mitochondrial Complex I deficiency. Structure The NDUFA1 gene is located on the long q arm of the X chromosome at position 24 and it spans 5,176 base pairs. The NDUFA1 gene produces an 8.1 kDa protein composed of 70 amino acids. NDUFA1 is a subunit of the enzyme NADH dehydrogenase (ubiquinone), the largest of the respiratory complexes. The structure is L-shaped with a long, hydrophobic transmembrane domain and a hydrophilic domain for the peripheral arm that includes all the known redox centers and the NADH binding site. NDUFA1 is one of about 31 hydrophobic subunits that form the transmembrane region of Complex I. It has been noted that the N-terminal hydrophobic domain has the potential to be folded into an alpha helix spanning the inner mitochondrial membrane with a C-terminal hydrophilic domain interacting with globular subunits of Complex I. The highly conserved two-domain structure suggests that this feature is critical for the protein function and that the hydrophobic domain acts as an anchor for the NADH dehydrogenase (ubiquinone) complex at the inner mitochondrial membrane. Function The human NDUFA1 gene codes for a subunit of Complex I of the respiratory chain, which transfers electrons from NADH to ubiquinone. However, NDUFA1 is an accessory subunit of the complex that is believed not to be involved in catalysis. Initially, NADH binds to Complex I and transfers two electrons to the isoalloxazine ring of the flavin mononucleotide (FMN) prosthetic arm to form FMNH2. The electrons are transferred through a series of iron-sulfur (Fe-S) clusters in the prosthetic arm and finally to coenzyme Q10 (Co
https://en.wikipedia.org/wiki/Texas%20Instruments%20SN76477
SN76477 "complex sound generator" is a sound chip produced by Texas Instruments (TI). The chip came to market in 1978, and TI ceased production of the part. A compatible version is identified as ICS76477. The chip is typically used as a sound effects generator in arcade games and toys and for hobby projects. The use of the SN76477 in a musical context is limited by the fact that it was difficult to electronically control the pitch of the produced sound. Overview The following quotation summarizes its facilities: There were two different sizes of the SN76477 available. The SN76477N was in a standard 0.6 in (15.24 mm) width dual in-line package (DIP). The SN76477NF was in a less common 0.4 inch (10.16 mm) width DIP. The text below is intended to explain the use of the 76477 in the demo circuit shown at the right. SW1 is a five position rotary switch that selects which capacitor controls the one-shot circuit that sets the envelope timing. SW24 (toggle or momentary) is the one-shot trigger switch, closed position is active. R1 a one megohm linear potentiometer sets the actual one-shot timing (in series with fixed 27k resistor R2) within the range selected by SW1. SW2 is a five position rotary switch which selects which capacitor sets the range for the low frequency oscillator (SLF oscillator in future references). SW3 (toggle or momentary) is the switch used to turn the low frequency oscillator on or off. R3 a one meagohm linear potentiometer (in series with 27k fixed resistor R4) sets the actual frequency for the SLF oscillator. SW4 (toggle or momentary) is the switch that takes control of the voltage-controlled oscillator's (VCO) duty cycle (called pitch control). If the switch is off the VCO operates at 50% duty cycle by default. R6 a 50k ohm linear potentiometer is used (in series with a 50k ohm fixed resistor R5) form a voltage divider used to set the VCO duty cycle, the on time vs off time of the waveform (most variable inputs to this chip are looking
https://en.wikipedia.org/wiki/Maximal%20surface
In the mathematical field of differential geometry, a maximal surface is a certain kind of submanifold of a Lorentzian manifold. Precisely, given a Lorentzian manifold , a maximal surface is a spacelike submanifold of whose mean curvature is zero. As such, maximal surfaces in Lorentzian geometry are directly analogous to minimal surfaces in Riemannian geometry. The difference in terminology between the two settings has to do with the fact that small regions in maximal surfaces are local maximizers of the area functional, while small regions in minimal surfaces are local minimizers of the area functional. In 1976, Shiu-Yuen Cheng and Shing-Tung Yau resolved the "Bernstein problem" for maximal hypersurfaces of Minkowski space which are properly embedded, showing that any such hypersurface is a plane. This was part of the body of work for which Yau was awarded the Fields medal in 1982. The Bernstein problem was originally posed by Eugenio Calabi in 1970, who proved some special cases of the result. Simple examples show that there are a number of hypersurfaces of Minkowski space of zero mean curvature which fail to be spacelike. By an extension of Cheng and Yau's methods, Kazuo Akutagawa considered the case of spacelike hypersurfaces of constant mean curvature in Lorentzian manifolds of positive constant curvature, such as de Sitter space. Luis Alías, Alfonso Romero, and Miguel Sánchez proved a version of Cheng and Yau's result, replacing Minkowski space by the warped product of a closed Riemannian manifold with an interval. As a problem of partial differential equations, Robert Bartnik and Leon Simon studied the boundary-value problem for maximal surfaces in Minkowski space. The general existence of maximal hypersurfaces in asymptotically flat Lorentzian manifolds, due to Bartnik, is significant in Demetrios Christodoulou and Sergiu Klainerman's renowned proof of the nonlinear stability of Minkowski space under the Einstein field equations. They use a maximal slici
https://en.wikipedia.org/wiki/U-Prove
U-Prove is a free and open-source technology and accompanying software development kit for user-centric identity management. The underlying cryptographic protocols were designed by Dr. Stefan Brands and further developed by Credentica and, subsequently, Microsoft. The technology was developed to allow internet users to disclose only the minimum amount of personal data when making electronic transactions as a way to reduce the likelihood of privacy violations. Overview U-Prove enables application developers to reconcile seemingly conflicting security and privacy objectives (including anonymity), and allows for digital identity claims to be efficiently tied to the use of tamper-resistant devices such as smart cards. Application areas of particular interest include cross-domain enterprise identity and access management, e-government SSO and data sharing, electronic health records, anonymous electronic voting, policy-based digital rights management, social networking data portability, and electronic payments. In 2008, Microsoft committed to opening up the U-Prove technology. As the first step, in March 2010 the company released a cryptographic specification and open-source API implementation code for part of the U-Prove technology as a Community Technology Preview under Microsoft's Open Specification Promise. Since then, several extensions have been released under the same terms and the technology has been tested in real-life applications. In 2010, the International Association of Privacy Professionals (IAPP) honored U-Prove with the 2010 Privacy Innovation Award for Technology. Microsoft also won the in European Identity Award in the Best Innovation category for U-Prove at the European Identity Conference 2010. The U-Prove Crypto SDK for C# is licensed under Apache License 2.0 and the source code is available on GitHub. Microsoft also provides a JavaScript SDK that implements the client-side of the U-Prove Cryptographic Specification. See also Blind signature
https://en.wikipedia.org/wiki/Radio%20IP%20Software
Radio IP Software is a privately held software company specializing in mobile data connectivity and mobile virtual private network (Mobile VPN) solutions. Established in 1998, Radio IP Software is headquartered in Montreal, Quebec, Canada and has offices in Florida, USA and London, England. In 2009, Radio IP Software completed the acquisition of ipUnplugged AB, a Stockholm, Sweden based developer of the ipUnplugged Mobile VPN software, which was designed using the Mobile IP standard communications protocol and including the IPSec security standards. Users Radio IP Software's users are primarily in the public safety, utility, mobile government, and enterprise sectors.
https://en.wikipedia.org/wiki/Inductive%20programming
Inductive programming (IP) is a special area of automatic programming, covering research from artificial intelligence and programming, which addresses learning of typically declarative (logic or functional) and often recursive programs from incomplete specifications, such as input/output examples or constraints. Depending on the programming language used, there are several kinds of inductive programming. Inductive functional programming, which uses functional programming languages such as Lisp or Haskell, and most especially inductive logic programming, which uses logic programming languages such as Prolog and other logical representations such as description logics, have been more prominent, but other (programming) language paradigms have also been used, such as constraint programming or probabilistic programming. Definition Inductive programming incorporates all approaches which are concerned with learning programs or algorithms from incomplete (formal) specifications. Possible inputs in an IP system are a set of training inputs and corresponding outputs or an output evaluation function, describing the desired behavior of the intended program, traces or action sequences which describe the process of calculating specific outputs, constraints for the program to be induced concerning its time efficiency or its complexity, various kinds of background knowledge such as standard data types, predefined functions to be used, program schemes or templates describing the data flow of the intended program, heuristics for guiding the search for a solution or other biases. Output of an IP system is a program in some arbitrary programming language containing conditionals and loop or recursive control structures, or any other kind of Turing-complete representation language. In many applications the output program must be correct with respect to the examples and partial specification, and this leads to the consideration of inductive programming as a special area inside aut
https://en.wikipedia.org/wiki/FLEX%20%28operating%20system%29
FLEX is a discontinued single-tasking operating system developed by Technical Systems Consultants (TSC) of West Lafayette, Indiana, for the Motorola 6800 in 1976. Overview The original version was distributed on 8" floppy disks; the (smaller) version for 5.25" floppies is called mini-Flex. It was also later ported to the Motorola 6809; that version is called Flex09. All versions are text-based and intended for use on display devices ranging from printing terminals like the Teletype Model 33 ASR to smart terminals. While no graphic displays are supported by TSC software, some hardware supports elementary graphics and pointing devices. FLEX is a disk-based operating system, using 256-byte sectors on soft-sectored floppies; the disk structure uses linkage bytes in each sector to indicate the next sector in a file or free list. The directory structure is simplified as a result. TSC (and others) provide several programming languages including BASIC in two flavors (standard and extended) and a tokenizing version of extended BASIC called Pre-compiled BASIC, FORTH, C, FORTRAN, and PASCAL. TSC also wrote a version of FLEX, Smoke Signal DOS, for the California hardware manufacturer Smoke Signal Broadcasting; this version uses forward and reverse linkage bytes in each sector which increase disk reliability at the expense of compatibility and speed. Later, TSC introduced the multitasking, multi-user, Unix-like UniFLEX operating system, which requires DMA disk controllers, 8" disk, and sold in small numbers. Several of the TSC computer languages were ported to UniFLEX. During the early 1980s, FLEX was offered by Compusense Ltd as an operating system for the 6809-based Dragon 64 home computer. Commands The following commands are supported by different versions of the FLEX operating system. APPEND ASN BACKUP BUILD CAT COPY COPYNEW C4MAT CLEAN DATE DELETE ECHO EXEC FIX GET I JUMP LINK LIST MEMTEST1 MON N NEWDISK O P P.COR PO PRINT PROT PSP Q
https://en.wikipedia.org/wiki/Vogel%E2%80%93Johnson%20agar
Vogel–Johnson agar is a type of agar growth medium selective for coagulase-positive staphylococci. It is used to isolate Staphylococcus aureus from clinical specimens and food. It was first described by Vogel and Johnson, who modified the Tellurite Glycine Agar recipe by Zebovitz et al. by doubling the mannitol concentration to 1% (w/v) and adding Phenol red as a pH indicator. It is widely available commercially. Typical composition Vogel–Johnson agar typically contains: Tryptone - 10.0 g/L Yeast extract - 5.0 g/L Mannitol - 10.0 g/L Dipotassium phosphate - 5.0 g/L Lithium chloride - 5.0 g/L Glycine - 10.0 g/L Phenol red - 0.025 g/L Agar = 16.0 g/L The Modified Vogel-Johnson Agar contains, in addition to above: 5 g/L of beef extract, 2 g/L of deoxyribonucleic acid, 2 /L g of phosphatidyl choline and 780 units/L of catalase spread on the plates before inoculation.
https://en.wikipedia.org/wiki/Large%20intestine
The large intestine, also known as the large bowel, is the last part of the gastrointestinal tract and of the digestive system in tetrapods. Water is absorbed here and the remaining waste material is stored in the rectum as feces before being removed by defecation. The colon is the longest portion of the large intestine, and the terms are often used interchangeably but most sources define the large intestine as the combination of the cecum, colon, rectum, and anal canal. Some other sources exclude the anal canal. In humans, the large intestine begins in the right iliac region of the pelvis, just at or below the waist, where it is joined to the end of the small intestine at the cecum, via the ileocecal valve. It then continues as the colon ascending the abdomen, across the width of the abdominal cavity as the transverse colon, and then descending to the rectum and its endpoint at the anal canal. Overall, in humans, the large intestine is about long, which is about one-fifth of the whole length of the human gastrointestinal tract. Structure The colon of the large intestine is the last part of the digestive system. It has a segmented appearance due to a series of saccules called haustra. It extracts water and salt from solid wastes before they are eliminated from the body and is the site in which the fermentation of unabsorbed material by the gut microbiota occurs. Unlike the small intestine, the colon does not play a major role in absorption of foods and nutrients. About 1.5 litres or 45 ounces of water arrives in the colon each day. The colon is the longest part of the large intestine and its average length in the adult human is 65 inches or 166 cm (range of 80 to 313 cm) for males, and 61 inches or 155 cm (range of 80 to 214 cm) for females. Sections In mammals, the large intestine consists of the cecum (including the appendix), colon (the longest part), rectum, and anal canal. The four sections of the colon are: the ascending colon, transverse colon, desce
https://en.wikipedia.org/wiki/Failure%20analysis
Failure analysis is the process of collecting and analyzing data to determine the cause of a failure, often with the goal of determining corrective actions or liability. According to Bloch and Geitner, ”machinery failures reveal a reaction chain of cause and effect… usually a deficiency commonly referred to as the symptom…”. Failure analysis can save money, lives, and resources if done correctly and acted upon. It is an important discipline in many branches of manufacturing industry, such as the electronics industry, where it is a vital tool used in the development of new products and for the improvement of existing products. The failure analysis process relies on collecting failed components for subsequent examination of the cause or causes of failure using a wide array of methods, especially microscopy and spectroscopy. Nondestructive testing (NDT) methods (such as industrial computed tomography scanning) are valuable because the failed products are unaffected by analysis, so inspection sometimes starts using these methods. Forensic investigation Forensic inquiry into the failed process or product is the starting point of failure analysis. Such inquiry is conducted using scientific analytical methods such as electrical and mechanical measurements, or by analyzing failure data such as product reject reports or examples of previous failures of the same kind. The methods of forensic engineering are especially valuable in tracing product defects and flaws. They may include fatigue cracks, brittle cracks produced by stress corrosion cracking or environmental stress cracking for example. Witness statements can be valuable for reconstructing the likely sequence of events and hence the chain of cause and effect. Human factors can also be assessed when the cause of the failure is determined. There are several useful methods to prevent product failures occurring in the first place, including failure mode and effects analysis (FMEA) and fault tree analysis (FTA), methods wh
https://en.wikipedia.org/wiki/Gtk-gnutella
gtk-gnutella is a peer-to-peer file sharing application which runs on the gnutella network. gtk-gnutella uses the GTK+ toolkit for its graphical user interface. Released under the GNU General Public License, gtk-gnutella is free software. History Initially gtk-gnutella was written to look like the original Nullsoft Gnutella client. The original author Yann Grossel stopped working on the client in early 2001. After a while Raphael Manfredi took over as the main software architect, and the client has been in active development ever since. Versions released after July 2002 do not look like the original Nullsoft client. Features gtk-gnutella is programmed in C with an emphasis on efficiency and portability without being minimalistic but rather head-on with most of the modern features of the gnutella network. Therefore, it requires fewer resources (such as CPU and/or RAM) than the major gnutella clients. It can also be used as headless gnutella client not requiring GTK+ at all. gtk-gnutella has a filtering engine that can reduce the amount of spam and other irrelevant results. gtk-gnutella supports a large range of the features of modern gnutella clients. gtk-gnutella was the first gnutella client to support IPv6 and encryption using TLS. It can handle and export magnet links. It has strong internationalization features, supporting English, German, Greek, French, Hungarian, Spanish, Japanese, Norwegian, Dutch and Chinese. gtk-gnutella also has support to prevent spamming and other hostile peer activity. Several software distributions provide pre-compiled packages, but they are usually outdated as many distributions version freeze old stable releases. The gnutella network benefits from running the latest version obtainable as peer and hostile IP address lists change rapidly, making building the latest SVN snapshot the best option. There are also pre-compiled packages for many Linux distributions available online. Persons concerned about security might wish to compi
https://en.wikipedia.org/wiki/Enterotoxigenic%20Escherichia%20coli
Enterotoxigenic Escherichia coli (ETEC) is a type of Escherichia coli and one of the leading bacterial causes of diarrhea in the developing world, as well as the most common cause of travelers' diarrhea. Insufficient data exists, but conservative estimates suggest that each year, about 157,000 deaths occur, mostly in children, from ETEC. A number of pathogenic isolates are termed ETEC, but the main hallmarks of this type of bacterium are expression of one or more enterotoxins and presence of fimbriae used for attachment to host intestinal cells. The bacterium was identified by the Bradley Sack lab in Kolkata in 1968. Signs and symptoms Infection with ETEC can cause profuse, watery diarrhea with no blood or leukocytes and abdominal cramping. Fever, nausea with or without vomiting, chills, loss of appetite, headache, muscle aches and bloating can also occur, but are less common. Enterotoxins Enterotoxins produced by ETEC include heat-labile enterotoxin (LT) and heat-stable enterotoxin (ST). Prevention To date, no licensed vaccines specifically target ETEC, though several are in various stages of development. Studies indicate that protective immunity to ETEC develops after natural or experimental infection, suggesting that vaccine-induced ETEC immunity should be feasible and could be an effective preventive strategy. Prevention through vaccination is a critical part of the strategy to reduce the incidence and severity of diarrheal disease due to ETEC, particularly among children in low-resource settings. The development of a vaccine against this infection has been hampered by technical constraints, insufficient support for coordination, and a lack of market forces for research and development. Most vaccine development efforts are taking place in the public sector or as research programs within biotechnology companies. ETEC is a longstanding priority and target for vaccine development for the World Health Organization. Management Treatment for ETEC infection inclu
https://en.wikipedia.org/wiki/Photophobia%20%28biology%29
In biology, photophobia (adjective: photophobic) is negative response to light. Photophobia is a behavior demonstrated by insects or other animals which seek to stay out of the light. In botany, the term photophobia/photophobic describes shade-loving plants (sciophytes) that thrive in low light conditions. Photophobia (or photophobic response) may also refer to a negative phototaxis or phototropism response.
https://en.wikipedia.org/wiki/Certificateless%20cryptography
Certificateless cryptography is a variant of ID-based cryptography intended to prevent the key escrow problem. Ordinarily, keys are generated by a certificate authority or a key generation center (KGC) who is given complete power and is implicitly trusted. To prevent a complete breakdown of the system in the case of a compromised KGC, the key generation process is split between the KGC and the user. The KGC first generates a key pair, where the private key is now the partial private key of the system. The remainder of the key is a random value generated by the user, and is never revealed to anyone, not even the KGC. All cryptographic operations by the user are performed by using a complete private key which involves both the KGC's partial key, and the user's random secret value. One disadvantage of this is that the identity information no longer forms the entire public key. Meaning, the user's public key is not discoverable from only the user's identity string and the KGC's public key. Thus, the user's public key must be published or otherwise obtained by other users. One advantage of the system, is that it is possible to verify that any such obtained public key belongs to the stated identity string. (In other words, the method of distributing the user's public key does not have to be secure.) The identity string and the KGC's public key can be used to verify that the obtained public key belongs to the identity string. (It can be verified that the obtained public key was generated from the identity string, the KGC's private key and some unknown value). Note that multiple public / private key pairs can be generated for any identity string, but attackers would not have access to the KGC's private key in the creation process. To encrypt a message to another user, three pieces of information are needed: 1) the recipient's public key and 2) identity string, and also 3) the KGC's public information (public key). The identity string and the KGC's public key are used to v
https://en.wikipedia.org/wiki/Floquet%20theory
Floquet theory is a branch of the theory of ordinary differential equations relating to the class of solutions to periodic linear differential equations of the form with a piecewise continuous periodic function with period and defines the state of the stability of solutions. The main theorem of Floquet theory, Floquet's theorem, due to , gives a canonical form for each fundamental matrix solution of this common linear system. It gives a coordinate change with that transforms the periodic system to a traditional linear system with constant, real coefficients. When applied to physical systems with periodic potentials, such as crystals in condensed matter physics, the result is known as Bloch's theorem. Note that the solutions of the linear differential equation form a vector space. A matrix is called a fundamental matrix solution if all columns are linearly independent solutions. A matrix is called a principal fundamental matrix solution if all columns are linearly independent solutions and there exists such that is the identity. A principal fundamental matrix can be constructed from a fundamental matrix using . The solution of the linear differential equation with the initial condition is where is any fundamental matrix solution. Floquet's theorem Let be a linear first order differential equation, where is a column vector of length and an periodic matrix with period (that is for all real values of ). Let be a fundamental matrix solution of this differential equation. Then, for all , Here is known as the monodromy matrix. In addition, for each matrix (possibly complex) such that there is a periodic (period ) matrix function such that Also, there is a real matrix and a real periodic (period-) matrix function such that In the above , , and are matrices. Consequences and applications This mapping gives rise to a time-dependent change of coordinates (), under which our original system becomes a linear system with real constant coe
https://en.wikipedia.org/wiki/AES51
AES51 is a standard first published by the Audio Engineering Society in June 2006 that specifies a method of carrying Asynchronous Transfer Mode (ATM) cells over Ethernet physical structure intended in particular for use with AES47 to carry AES3 digital audio transport structure. The purpose of this is to provide an open standard, Ethernet based approach to the networking of linear (uncompressed) digital audio with extremely high quality-of-service alongside standard Internet Protocol connections. This standard specifies a method, also known as "ATM-E", of carrying ATM cells over hardware specified for IEEE 802.3 (Ethernet). It is intended as a companion standard to AES47 (Transmission of digital audio over ATM networks), to provide a standard method of carrying ATM cells and real-time clock over hardware specified for Ethernet.
https://en.wikipedia.org/wiki/Geresh
Geresh (׳ in Hebrew: ‎ or ‎<ref>Even-Shoshan Dictionary, 2003; Shoshana Bahat and Mordechay Mishor, Dictionary of Contemporary Hebrew, 2007.</ref> , or medieval ) is a sign in Hebrew writing. It has two meanings. An apostrophe-like sign (also known colloquially as a chupchik) placed after a letter: as a diacritic that modifies the pronunciation of some letters (only in modern Hebrew), as a diacritic that signifies Yiddish origin of a word or suffix, (examples below) as a punctuation mark to denote initialisms or abbreviations, or to denote a single-digit Hebrew numeral A note of cantillation in the reading of the Torah and other Biblical books, taking the form of a curved diagonal stroke placed above a letter. Diacritic As a diacritic, the Geresh is written immediately after (left of) the letter it modifies. It indicates three sounds native to speakers of modern Hebrew that are common in loan words and slang: as in judge, as in measure and as in church. In transliteration of Arabic, it indicates Arabic phonemes which are usually allophones in modern Hebrew: is distinguished from and is distinguished from . Finally, it indicates other sounds foreign to the phonology of modern Hebrew speakers and used exclusively for the transliteration of foreign words: as in then, as in thin, ; and, in some transliteration systems, also , and . It may be compared to the usage of a following h'' in various Latin digraphs to form other consonant sounds not supported by the basic Latin alphabet, such as "sh", "th", etc. Loanwords, slang, foreign names and transliterations Transcriptions of Arabic There are six additional letters in the Arabic alphabet. They are Ṯāʾ, Ḫāʾ, Ḏāl, Ḍād, Ẓāʾ, and Ghayn. Also, some letters have different sounds in Arabic phonology and modern Hebrew phonology, such as Jīm. ט Tet Ŧ z with glottal stop at end ט׳ tet with . ظ Tran
https://en.wikipedia.org/wiki/Hemolysis%20%28microbiology%29
Hemolysis (from Greek αιμόλυση, meaning 'blood breakdown') is the breakdown of red blood cells. The ability of bacterial colonies to induce hemolysis when grown on blood agar is used to classify certain microorganisms. This is particularly useful in classifying streptococcal species. A substance that causes hemolysis is a hemolysin. Types Alpha-hemolysis When alpha-hemolysis (α-hemolysis) is present, the agar under the colony is light and greenish. Streptococcus pneumoniae and a group of oral streptococci (Streptococcus viridans or viridans streptococci) display alpha hemolysis. This is sometimes called green hemolysis because of the color change in the agar. Other synonymous terms are incomplete hemolysis and partial hemolysis. Alpha hemolysis is caused by hydrogen peroxide produced by the bacterium, oxidizing hemoglobin producing the green oxidized derivative methemoglobin. Beta-hemolysis Beta-hemolysis (β-hemolysis), sometimes called complete hemolysis, is a complete lysis of red cells in the media around and under the colonies: the area appears lightened (yellow) and transparent. Streptolysin, an exotoxin, is the enzyme produced by the bacteria which causes the complete lysis of red blood cells. There are two types of streptolysin: Streptolysin O (SLO) and streptolysin S (SLS). Streptolysin O is an oxygen-sensitive cytotoxin, secreted by most Group A streptococcus (GAS) and Streptococcus dysgalactiae, and interacts with cholesterol in the membrane of eukaryotic cells (mainly red and white blood cells, macrophages, and platelets), and usually results in β-hemolysis under the surface of blood agar. Streptolysin S is an oxygen-stable cytotoxin also produced by most GAS strains which results in clearing on the surface of blood agar. SLS affects immune cells, including polymorphonuclear leukocytes and lymphocytes, and is thought to prevent the host immune system from clearing infection. Streptococcus pyogenes, or Group A beta-hemolytic Strep (GAS), displa
https://en.wikipedia.org/wiki/Transportation%20Science
Transportation Science is a bimonthly peer-reviewed academic journal published by the Institute for Operations Research and the Management Sciences (INFORMS). The studies published in the journal apply operations research techniques to problems in the full range of transportation sectors, including air travel, rail transport, commuter lines, and vehicular travel/traffic.
https://en.wikipedia.org/wiki/Mark%20Geiger
Mark William Geiger (born August 25, 1974) is an American sports administrator and former soccer referee. He is the senior director of match officials at the Professional Referee Organization (PRO), which oversees domestic referees in Major League Soccer (MLS). Geiger previously officiated in MLS and was on the FIFA International Referees List from 2008 to 2019. At the international level, Geiger refereed the 2012 Olympics, the 2013, 2015, and 2017 CONCACAF Gold Cups, and the 2014 and 2018 FIFA World Cups. At the 2014 World Cup, he became the first referee from the United States to officiate a knockout match at a World Cup tournament. Before becoming a full-time referee, Geiger was a high school mathematics teacher. Refereeing career Geiger first took up refereeing in 1988. He became a United States Soccer Federation National Referee in 2003 and officiated in Major League Soccer from 2004 to 2018. He has officiated in CONCACAF tournaments since becoming a FIFA referee in 2008, and has officiated Gold Cup and other international tournaments and fixtures. Geiger was selected for the CONCACAF U-20 Championship in 2011, where he refereed the final. Geiger officiated at the 2011 FIFA U-20 World Cup in Colombia. He officiated the Group E game between eventual tournament champions Brazil and Austria at Estadio Metropolitano Roberto Meléndez in Barranquilla. He then took charge of the Group B game between Uruguay and Cameroon at Estadio El Campín in Bogotá. He was appointed to the Round of 16 match between Spain and South Korea at Estadio Palogrande in Manizales. Finally, he refereed the Final with American Assistant Referee Sean Hurd and Canadian Assistant Referee Joe Fletcher, between Brazil and Portugal at Estadio El Campín in Bogotá. This was the first time a referee from the United States has ever officiated a major men's tournament final. Geiger was selected as the Major League Soccer referee of the year for the 2011 season, and then again in 2014. In 2012, Geiger
https://en.wikipedia.org/wiki/Utimaco%20Atalla
Utimaco Atalla, founded as Atalla Technovation and formerly known as Atalla Corporation or HP Atalla, is a security vendor, active in the market segments of data security and cryptography. Atalla provides government-grade end-to-end products in network security, and hardware security modules (HSMs) used in automated teller machines (ATMs) and Internet security. The company was founded by Egyptian engineer Mohamed M. Atalla in 1972. Atalla HSMs are the payment card industry's de facto standard, protecting 250million card transactions daily (more than billion transactions annually) as of 2013, and securing the majority of the world's ATM transactions as of 2014. Company history 1970s The company was originally founded in 1972, initially as Atalla Technovation, before it was later called Atalla Corporation. The company was founded by Dr. Mohamed M. Atalla, the inventor of the MOSFET (metal–oxide–semiconductor field-effect transistor). In 1972, Atalla filed for a remote PIN verification system, which utilized encryption techniques to assure telephone link security while entering personal ID information, which would be transmitted as encrypted data over telecommunications networks to a remote location for verification. He invented the first hardware security module (HSM), dubbed the "Atalla Box", a security system which encrypted PIN and ATM messages, and protected offline devices with an un-guessable PIN-generating key. He commercially released the "Atalla Box" in 1973. The product was released as the Identikey. It was a card reader and customer identification system, providing a terminal with plastic card and PIN capabilities. The system was designed to let banks and thrift institutions switch to a plastic card environment from a passbook program. The Identikey system consisted of a card reader console, two customer PIN pads, intelligent controller and built-in electronic interface package. The device consisted of two keypads, one for the customer and one for the
https://en.wikipedia.org/wiki/Alor%20Setar%20Tower
Alor Setar Tower (), also known as Kedah Tower is a 4-story, 165.5-meter-tall telecommunication tower in Alor Setar, Kedah, Malaysia, It is Malaysia's third tallest tower, the tallest structure in the state and the tallest structure in the city. Apart from serving the role of a telecommunication tower, it also caters as a tourist destination for the town. The tower also houses some restaurants and a souvenir shop. The tower is an observatory tower to look for the crescent moon to mark the beginning of Muslim months such as Ramadhan, Shawwal, and Zulhijjah, to celebrate Ramadhan, Hari Raya Aidilfitri and Hari Raya Aidiladha, respectively. The observation deck is at a height of from the base of the structure. Also, the open deck or skydeck is at a height of and the antenna tower is with the height of . History Before Alor Setar Tower was built, the tallest structure in Kedah was Holiday Villa Alor Setar, which is . Located at the Alor Setar City Centre, it was built in 1994. It is also the second structure in Northern Region of Peninsular Malaysia to surpass after Penang's Komtar. Alor Setar Tower was started in 1994, topped-out in 1995, it surpassed the previous Holiday Villa Alor Setar, became the third to surpass . The tower was completed in 1996. Opened in 14 August 1997 by Mahathir Mohamad, who was the prime minister at the time. Following the completion of the Alor Setar Tower, many high-rise and low-rise buildings started taking shape, including Amansuri Residences, Aman Central, SADA Tower, G Residence Alor Setar, VIVRE Residence, AS100, and AMEX Alor Mengkudu. Gallery Channels listed by frequency Television TV Alhijrah UHF 559.25 MHz (Ch 32) Radio Fly FM 99.1 MHz Buletin FM 107.3 MHz Menara Alor Setar is a telecommunication tower in Alor Setar, Kedah and a popular tourist destination in the northern region. With a total height of , it is the third tallest telecommunication tower in Malaysia (behind the Kuala Lumpur Tower, and the newly opened Kua
https://en.wikipedia.org/wiki/Pocket%20LOOX
Pocket LOOX was a series of Pocket PC-based personal digital assistants (PDAs) and navigation systems developed by Fujitsu Siemens. It was discontinued in 2007. Product comparison charts Fujitsu Siemens Pocket LOOX PDAs Pocket LOOX 600 was the first PDA by Fujitsu Siemens and HTC, released in 2002. It had a built-in Bluetooth module and two expansion slots, SD/MMC and CF. Fujitsu Siemens later released a GPRS expansion module that provides mobile phone functionality. Pocket LOOX 600 can be upgraded to Windows Mobile 2003. Fujitsu Siemens Pocket LOOX navigation system Pocket LOOX navigation systems were powered by Navigon MobileNavigator|6s. There were 9 maps supplied with device, covering 37 European countries. Main difference between N100 and N110 is that N100 did not have any built-in user accessible memory, but came with 1 GB Mini SD card. N110 on the other hand had 2 GB built-in memory and a free Mini SD expansion slot. Alternative operating systems It is possible to run Linux on Pocket LOOX 600. The latest supported version of Linux kernel is 2.4.19. Linux is loaded from a CF card, allowing a return to PocketPC 2002. But during restart, all contents of RAM are erased. To prevent data loss, users should backup RAM before restarting to Linux. Work is being done to port Android to Pocket Loox 720. A port of Android 2.2 is running on the 720. It is possible to boot from either CF or SD cards; as mentioned above, a backup of all data is highly recommended. See also Fujitsu Siemens Computers List of Fujitsu products Personal digital assistant Windows Mobile Pocket PC
https://en.wikipedia.org/wiki/Localization%20of%20a%20category
In mathematics, localization of a category consists of adding to a category inverse morphisms for some collection of morphisms, constraining them to become isomorphisms. This is formally similar to the process of localization of a ring; it in general makes objects isomorphic that were not so before. In homotopy theory, for example, there are many examples of mappings that are invertible up to homotopy; and so large classes of homotopy equivalent spaces. Calculus of fractions is another name for working in a localized category. Introduction and motivation A category C consists of objects and morphisms between these objects. The morphisms reflect relations between the objects. In many situations, it is meaningful to replace C by another category C''' in which certain morphisms are forced to be isomorphisms. This process is called localization. For example, in the category of R-modules (for some fixed commutative ring R) the multiplication by a fixed element r of R is typically (i.e., unless r is a unit) not an isomorphism: The category that is most closely related to R-modules, but where this map is an isomorphism turns out to be the category of -modules. Here is the localization of R with respect to the (multiplicatively closed) subset S consisting of all powers of r, The expression "most closely related" is formalized by two conditions: first, there is a functor sending any R-module to its localization with respect to S. Moreover, given any category C and any functor sending the multiplication map by r on any R-module (see above) to an isomorphism of C, there is a unique functor such that . Localization of categories The above examples of localization of R-modules is abstracted in the following definition. In this shape, it applies in many more examples, some of which are sketched below. Given a category C and some class W of morphisms in C, the localization C[W−1] is another category which is obtained by inverting all the morphisms in W. More formally,
https://en.wikipedia.org/wiki/Pipeline%20video%20inspection
Pipeline video inspection is a form of telepresence used to visually inspect the interiors of pipelines, plumbing systems, and storm drains. A common application is for a plumber to determine the condition of small diameter sewer lines and household connection drain pipes. Older sewer lines of small diameter, typically , are made by the union of a number of short sections. The pipe segments may be made of cast iron, with to sections, but are more often made of vitrified clay pipe (VCP), a ceramic material, in , & sections. Each iron or clay segment will have an enlargement (a "bell") on one end to receive the end of the adjacent segment. Roots from trees and vegetation may work into the joins between segments and can be forceful enough to break open a larger opening in terra cotta or corroded cast iron. Eventually a root ball will form that will impede the flow and this may cleaned out by a cutter mechanism or plumber's snake and subsequently inhibited by use of a chemical foam - a rooticide. With modern video equipment, the interior of the pipe may be inspected - this is a form of non-destructive testing. A small diameter collector pipe will typically have a cleanout access at the far end and will be several hundred feet long, terminating at a manhole. Additional collector pipes may discharge at this manhole and a pipe (perhaps of larger diameter) will carry the effluent to the next manhole, and so forth to a pump station or treatment plant. Without regular inspection of public sewers, a significant amount of waste may accumulate unnoticed until the system fails. In order to prevent resulting catastrophic events such as pipe bursts and raw sewage flooding onto city streets, municipalities usually conduct pipeline video inspections as a precautionary measure. Inspection equipment Service truck The service truck contains a power supply in the form of a small generator, a small air-conditioned compartment containing video monitoring and recording equipment,
https://en.wikipedia.org/wiki/Lucky%20Lynx
Lucky Lynx, commonly known as Lucky, is the mascot of Dutch football club Ajax Amsterdam playing in the Eredivisie, a role he has filled since 2000. On 20 December 2015, he was replaced by his son Lucky Junior. Biography Since the year 2000, Lucky Lynx has been entertaining Ajax fans at the Amsterdam ArenA as well at any Ajax Kids Club events, school events as well as charity events in and around the city of Amsterdam. Lucky is a Lynx, who according to his fictional biography was born on the top of Mount Olympus in Greece, has yellow eyes. Brown, black and white hair, is 1,75 meters tall, wears a size 58 shoe, and whose hobbies include playing football. As the face for the Ajax Kids Club, a fan club for children initiated on 6 December 2000, Lucky is the central character in the Fan clubs Kids magazine, and is a popular character amongst fans at games and social events. His name was derived from the Lucky Ajax team, which was created by "Mr. Ajax" Sjaak Swart, a beneficiary team consisting of ex-Ajax players who compete in at least one charity match a year. Costume The Lucky Lynx suit consists of a whole-body lynx costume made of black, brown and white fur with large yellow eyes, and a short tail. Over the suit, Lucky wears an authentic uniform (white/red/white home kit), an Ajax baseball cap of the same colors, and team-appointed athletic shoes. His jersey bears the name "Lucky" on the back. Duties In addition to photo opportunities and entertaining the crowd at Ajax games, Lucky also makes public appearances throughout the greater Amsterdam area on behalf of the team. As the key figure of the Ajax Kids Club, he appears at schools to promote tolerance, teamwork and companionship and the importance of a good education. There is a monthly Ajax fanzine for Kids that comes with the kids club membership, of which Lucky is the main character. The magazine also features a comic strip starring Lucky Lynx, which is illustrated by Dutch cartoonist Maarten Rijnen. Lucky
https://en.wikipedia.org/wiki/Newton%E2%80%93Pepys%20problem
The Newton–Pepys problem is a probability problem concerning the probability of throwing sixes from a certain number of dice. In 1693 Samuel Pepys and Isaac Newton corresponded over a problem posed to Pepys by a school teacher named John Smith. The problem was: Pepys initially thought that outcome C had the highest probability, but Newton correctly concluded that outcome A actually has the highest probability. Solution The probabilities of outcomes A, B and C are: These results may be obtained by applying the binomial distribution (although Newton obtained them from first principles). In general, if P(n) is the probability of throwing at least n sixes with 6n dice, then: As n grows, P(n) decreases monotonically towards an asymptotic limit of 1/2. Example in R The solution outlined above can be implemented in R as follows: for (s in 1:3) { # looking for s = 1, 2 or 3 sixes n = 6*s # ... in n = 6, 12 or 18 dice q = pbinom(s-1, n, 1/6) # q = Prob( <s sixes in n dice ) cat("Probability of at least", s, "six in", n, "fair dice:", 1-q, "\n") } Newton's explanation Although Newton correctly calculated the odds of each bet, he provided a separate intuitive explanation to Pepys. He imagined that B and C toss their dice in groups of six, and said that A was most favorable because it required a 6 in only one toss, while B and C required a 6 in each of their tosses. This explanation assumes that a group does not produce more than one 6, so it does not actually correspond to the original problem. Generalizations A natural generalization of the problem is to consider n non-necessarily fair dice, with p the probability that each die will select the 6 face when thrown (notice that actually the number of faces of the dice and which face should be selected are irrelevant). If r is the total number of dice selecting the 6 face, then is the probability of having at least k correct selections when throwing exactly n dice. Then the original Newton
https://en.wikipedia.org/wiki/Multipole%20density%20formalism
The Multipole Density Formalism (also referred to as Hansen-Coppens Formalism) is an X-ray crystallography method of electron density modelling proposed by Niels K. Hansen and Philip Coppens in 1978. Unlike the commonly used Independent Atom Model, the Hansen-Coppens Formalism presents an aspherical approach, allowing one to model the electron distribution around a nucleus separately in different directions and therefore describe numerous chemical features of a molecule inside the unit cell of an examined crystal in detail. Theory Independent Atom Model The Independent Atom Model (abbreviated to IAM), upon which the Multipole Model is based, is a method of charge density modelling. It relies on an assumption that electron distribution around the atom is isotropic, and that therefore charge density is dependent only on the distance from a nucleus. The choice of the radial function used to describe this electron density is arbitrary, granted that its value at the origin is finite. In practice either Gaussian- or Slater-type 1s-orbital functions are used. Due to its simplistic approach, this method provides a straightforward model that requires no additional parameters (other than positional and Debye–Waller factors) to be refined. This allows the IAM to perform satisfactorily while a relatively low amount of data from the diffraction experiment is available. However, the fixed shape of the singular basis function prevents any detailed description of aspherical atomic features. Kappa Formalism In order to adjust some valence shell parameters, the Kappa formalism was proposed. It introduces two additional refineable parameters: an outer shell population (denoted as ) and its expansion/contraction (). Therefore, the electron density is formulated as: While , being responsible for the charge flow part, is linearly coupled with partial charge, the normalised parameter scales radial coordinate . Therefore, lowering the parameter results in expansion of the outer
https://en.wikipedia.org/wiki/Analog%20temperature%20controlled%20crystal%20oscillator
In physics, an Analog Temperature Controlled Crystal Oscillator or Analogue Temperature Compensated Crystal Oscillator (ATCXO) uses analog sampling techniques to correct the temperature deficiencies of a crystal oscillator circuit, its package and its environment. Typically the correction techniques involve the physical and electrical characterisation of the motional inductance and terminal capacitance of a crystal blank, the knowledge of which is used to create a correction polynomial, or algorithm, which in turn is implemented in circuit blocks. These are usually simulated in a mathematical modeling software tool such as SPICE, to verify that the original measured data can be corrected adequately. Once the system performance has been verified, these circuits are then implemented in a silicon die, usually in a bulk CMOS technology. Once fabricated, this die is then embedded into an oscillator module along with the crystal blank. Due to the sub accuracy of this type of crystal oscillator specialist packaging must be used to ensure good ageing and temperature shock characteristics. Example applications are for use in low power or battery operated consumer electronic products such as GSM or CDMA mobile phones, or GPS satellite navigation systems.
https://en.wikipedia.org/wiki/NBR2
NBR2 is a gene best known for its location near the breast cancer associated gene BRCA1. Like BRCA1, NBR2 has been a subject of research, but links to breast cancer are currently inconclusive. NBR2 recently was identified as a glucose starvation-induced long non-coding RNA. NBR2 interacts with AMP-activated protein kinase (AMPK), a critical energy sensor in most eukaryotic cells, and promotes AMPK function to mediate energy stress response. Knockdown of NBR2 attenuates energy stress-induced AMPK activation, resulting in unchecked cell cycling, altered apoptosis/autophagy response, and increased tumour development in vivo. It is now appreciated that NBR2, a former junk gene, plays critical roles in tumor suppression.
https://en.wikipedia.org/wiki/Functional%20equivalence%20%28ecology%29
In ecology, functional equivalence (or functional redundancy) is the ecological phenomenon that multiple species representing a variety of taxonomic groups can share similar, if not identical, roles in ecosystem functionality (e.g., nitrogen fixers, algae scrapers, scavengers). This phenomenon can apply to both plant and animal taxa. The idea was originally presented in 2005 by Stephen Hubbell, a plant ecologist at the University of Georgia. This idea has led to a new paradigm for species-level classification – organizing species into groups based on functional similarity rather than morphological or evolutionary history. In the natural world, several examples of functional equivalence among different taxa have emerged analogously. Plant-pollinator relationships One example of functional equivalence is demonstrated in plant-pollinator relationships, whereby a certain plant species may evolve flower morphology that selects for pollination by a host of taxonomically-unrelated species to provide the same function (fruit production following pollination). For example, the herbaceous plant spiny madwort (Hormathophylla spinosa) grows flowers that are shaped so that taxonomically unrelated pollinators behave almost identically during pollination. From the plant's perspective, each of these pollinators are functionally equivalent and thus are not subjected to specific selective pressures Variation in the shape and structure of both flower and seed morphology can be a source of selective pressure for animal species to evolve a variety of morphological features, yet also provide the same function to the plant. Plant-animal seed dispersal mechanisms Plant-animal interactions in terms of seed dispersal are another example of functional equivalence. Evidence has shown that, over the course of millions of years, most plants have maintained evolutionary trait stability in terms of the size and shape of their fruits. However, the animal species that consume and disperse the
https://en.wikipedia.org/wiki/Anthoxanthin
Anthoxanthins (flavones and flavonols) are a type of flavonoid pigments in plants. Anthoxanthins are water-soluble pigments which range in color from white or colorless to a creamy to yellow, often on petals of flowers. These pigments are generally whiter in an acid medium and yellowed in an alkaline medium. They are very susceptible to color changes with minerals and metal ions, similar to anthocyanins. Uses As with all flavonoids, anthoxanthins have antioxidant properties and are important for nutrition. They are sometimes used as food additives to add color or flavor to foods. One of the most well-known anthoxanthins is quercetin, which is found in many fruits and vegetables, including capers, red onions, and kale. In addition to their use as food additives, anthoxanthins are also used in the production of dyes and pigments. Anthoxanthins can also be used to create yellow, orange, or red dyes for use in textiles, cosmetics, and other products.
https://en.wikipedia.org/wiki/Multilinear%20polynomial
In algebra, a multilinear polynomial is a multivariate polynomial that is linear (meaning affine) in each of its variables separately, but not necessarily simultaneously. It is a polynomial in which no variable occurs to a power of 2 or higher; that is, each monomial is a constant times a product of distinct variables. For example f(x,y,z) = 3xy + 2.5 y - 7z is a multilinear polynomial of degree 2 (because of the monomial 3xy) whereas f(x,y,z) = x² +4y is not. The degree of a multilinear polynomial is the maximum number of distinct variables occurring in any monomial. Definition Multilinear polynomials can be understood as a multilinear map (specifically, a multilinear form) applied to the vectors [1 x], [1 y], etc. The general form can be written as a tensor contraction: For example, in two variables: Properties A multilinear polynomial is linear (affine) when varying only one variable, :where and do not depend on . Note that is generally not zero, so is linear in the "shaped like a line" sense, but not in the "directly proportional" sense of a multilinear map. All repeated second partial derivatives are zero:In other words, its Hessian matrix is a symmetric hollow matrix. In particular, the Laplacian , so is a harmonic function. This implies has maxima and minima only on the boundary of the domain. More generally, every restriction of to a subset of its coordinates is also multilinear, so still holds when one or more variables are fixed. In other words, is harmonic on every "slice" of the domain along coordinate axes. On a rectangular domain When the domain is rectangular in the coordinate axes (e.g. a hypercube), will have maxima and minima only on the vertices of the domain, i.e. the finite set of points with minimal and maximal coordinate values. The value of the function on these points completely determines the function, since the value on the edges of the boundary can be found by linear interpolation, and the value on the rest of the b
https://en.wikipedia.org/wiki/Hot%20pixel%20%28telescopes%29
A hot pixel or bright dot defect is a pixel that outputs many more electrons than others at the same input signal in a charge-coupled device (CCD) or CMOS sensor. In the simulated image, the hot pixels are the sources of the salt-and-pepper noise. In the definition of the HST ACS, A pixel above 0.14 e¯/pixel/second is considered a "hot" pixel. A warm pixel is a pixel that has negative bias values. In the definition of the Hubble Space Telescope, a pixel below the hot pixels range but above 0.06 e¯/pixel/second is considered a "warm" pixel. See also Defective pixel
https://en.wikipedia.org/wiki/Seaweed%20collecting
Seaweed collecting is the process of collecting, drying, and pressing seaweed. It became popular as a pastime in the Victorian era and remains a hobby today. History of seaweed collecting Collecting seaweed can be traced back to at least the 17th century with the pressings found in Hans Sloane's Herbarium. The pastime became increasingly popular during the Victorian Era, where it played to the burgeoning interest in natural history and collection in general. It was especially fashionable with young women, as it allowed a greater level of personal freedom. Indeed, it was so in-style that, as a young girl, Queen Victoria created her own seaweed album. The materials needed for the hobby became readily available at seaside shops. These activities also afforded women the opportunity to display their understanding and appreciation of the natural world. Anna Atkins, thought to be the first female photographer, published the first book using photographs as illustrations. This was Photographs of British Algae: Cyanotype Impressions and contained pictures of seaweed. The actions of some of the collectors earned them recognition and admiration from their male, professional counterparts. These Victorian collections form valuable historical resources for morphological studies and from which genomic DNA can be extracted. Seaweed collecting equipment In his 1881 book, A.B. Hervey recommended the following equipment for collecting and pressing seaweed. See also History of phycology Phycology
https://en.wikipedia.org/wiki/Advanced%20LIGO%20Documentary%20Project
The Advanced LIGO Documentary Project is a collaboration formed in the summer of 2015 among Caltech, the Massachusetts Institute of Technology and Director Les Guthman to make the definitive documentary about the Advanced LIGO project's search for, and expected first detection of, gravitational waves; and to record a longitudinal video archive of the project for future researchers and historians. The feature documentary, "LIGO," was released in the spring of 2019. Mr. Guthman also wrote, produced and directed an eight-part video series on YouTube, LIGO: A DISCOVERY THAT SHOOK THE WORLD, which was released over three years, 2017-2020. The video series remains in production with three more episodes covering the LIGO project's third science run 2019-2020. Nobel Prize On October 3, 2017, Rainer Weiss, Kip Thorne and Barry Barish won the Nobel Prize in Physics for LIGO's historic first observation of gravitational waves in September 2015. History On September 14, 2015, the Advanced LIGO Documentary team was on location filming at the LIGO Livingston Observatory when the detection was made. Over the next five months, it had exclusive media access to document the long, careful process of scientific verification that was conducted by the LIGO Scientific Collaboration to confirm that the received signal was in fact a gravitational wave, as predicted by Albert Einstein more than 100 years ago. The detection was announced by LIGO at the National Press Club in Washington DC on February 11, 2016. In addition to its filming of the secret internal confirmation months of the gravitational wave detection and LIGO's continuing breakthrough science, the Advanced LIGO Documentary Project filmed LIGO's Nobel Prize week in Stockholm; three important post-detection lectures by Thorne, Barish and LIGO astrophysicist Alessandra Buonanno. It also filmed the LIGO Scientific Collaboration semi-annual meeting at CERN two weeks after its August 2017 detection of two colliding neutron st
https://en.wikipedia.org/wiki/DSS%20%28NMR%20standard%29
Sodium trimethylsilylpropanesulfonate (DSS) is the organosilicon compound with the formula (CH3)3SiCH2CH2CH2SO3−Na+. It is the sodium salt of trimethylsilylpropanesulfonic acid. A white, water-soluble solid, it is used as a chemical shift standard for proton NMR spectroscopy of aqueous solutions. The chemical shift, specifically the signal for the trimethylsilyl group, is relatively insensitive to pH. The proton spectrum of DSS also exhibits resonances at 2.91 ppm (m), 1.75 ppm (m), and 0.63 ppm (m) at an intensity of 22% of the reference resonance at 0 ppm. Alternatives Sodium trimethylsilyl propionate (TSP) is a related compound used as an NMR standard. It uses a carboxylic acid instead of the sulfonic acid found in DSS to confer water solubility. As a weak acid, TSP is more sensitive to changes in pH. 4,4-Dimethyl-4-silapentane-1-ammonium trifluoroacetate (DSA) has also been proposed as an alternative, to overcome certain drawbacks of DSS.
https://en.wikipedia.org/wiki/HIST1H4A
Histone H4 is a protein that in humans is encoded by the HIST1H4A gene. Histones are basic nuclear proteins that are responsible for the nucleosome structure of the chromosomal fiber in eukaryotes. Two molecules of each of the four core histones (H2A, H2B, H3, and H4) form an octamer, around which approximately 146 bp of DNA is wrapped in repeating units, called nucleosomes. The linker histone, H1, interacts with linker DNA between nucleosomes and functions in the compaction of chromatin into higher order structures. This gene is intronless and encodes a member of the histone H4 family. Transcripts from this gene lack polyA tails but instead contain a palindromic termination element. This gene is found in the large histone gene cluster on chromosome 6.
https://en.wikipedia.org/wiki/Roman%20pot
The Roman pot is the name of a technique (and of the relevant device) used in accelerator physics. Named after its implementation by the CERN-Rome collaboration in the early 1970s, it is an important tool to measure the total cross section of two particle beams in a collider. They are called pots because the detectors are housed in cylindrical vessels. The first generation of Roman pots was purpose-built by the CERN Central Workshops and used in the measurement of the total cross-section of proton-proton inter-actions in the ISR. Roman pots are located as close to the beamline as possible, to capture the accelerated particles which scatter by very small angles. Roman pots used at the Large Hadron Collider (LHC) Roman pots were first used in the TOTEM experiment and later by the ATLAS and the CMS collaborations at the LHC. The figure below shows a detector used on the beamline near IP5 (interaction point 5), the location of the CMS detector. Three of these are used per Roman pot unit. Each is shoved into place to within 10 microns of the beamline. Two detectors are placed above and below the beamline, and a third to the side. These detectors will record any protons that are not travelling precisely along the beamline, and thus record the elastic scattering of the protons. This is used to measure the total elastic cross-section, including Coulomb scattering as well as diffractive scattering (i.e. diffraction because the protons are not point particles, and have an internal structure (i.e. quarks)). Effectively, these are detectors for studying Regge theory. The goal is to search for elastic scattering effects beyond the Standard Model, such as hypothetical "colorless gluons", as well confirming ideas of pomeron exchange, and the possible existence of an odderon. Odderons were potentially observed only in 2017 by the TOTEM experiment at the LHC. This observation was later confirmed in a joint analysis with the DØ experiment at the Tevatron. The figure below shows a
https://en.wikipedia.org/wiki/Helen%20MacGillivray
Helen Louise MacGillivray is an Australian statistician and statistics educator. She is the former president of the International Statistical Institute, the International Association for Statistical Education, and the Statistical Society of Australia, and chair of the United Nations Global Network of Institutions for Statistical Training. Education and career MacGillivray entered her studies at the University of Queensland planning to work in physics, but ended up earning a bachelor's degree with honours in mathematics, in the course of which she discovered her love for statistics. She remained at the University of Queensland for graduate study, and completed a Ph.D. in statistics there. Her dissertation was Moment inequalities with applications to particle size distributions. She was a professor of statistics and director of the Maths Access Centre at Queensland University of Technology (QUT), until her retirement. She continues to hold an adjunct professorship at QUT. Service MacGillivray is the editor of the journal Teaching Statistics. She was president of the International Statistical Institute for the 2017–2019 term. When she was elected president she became both the second woman and the second Australian to hold the position, after Denise Lievesley and Dennis Trewin. She was the first female president of the Statistical Society of Australia. She was president of the International Association for Statistical Education for 2009–2011, and is the founding chair of the Global Network of Institutions for Statistical Training of the United Nations. Books With Peter Petocz, MacGillivray is the coauthor of the two-volume textbook Statistics and Probability in the Australian Curriculum (Years 7 and 8, and Years 9 and 10), and is the author of Utts & Heckard's Mind on Statistics (Nelson Australia, 2010, adapted from previous work by Jessica Utts and Robert Heckard). Recognition MacGillivray is a Fellow of the Royal Statistical Society, an Australian Learning and
https://en.wikipedia.org/wiki/Attitude-behavior%20consistency
Attitude-behavior consistency is when a person's attitude is consistent with their behavior. This is not true in many cases. The fact that people often express attitudes that are inconsistent with how they act may surprise those unfamiliar with social and behavioral science, but it is an important fact to understand because facts are often reported as if they are about people's actions when they may only be known to be true about their words. It is often much easier to conduct interviews or surveys than to obtain records of how people behave in situations. Sometimes attitudes, such as voting, are measurably consistent with behavior. In such cases it may be possible to obtain accurate estimates of behavior. However, there is no general method for correcting for attitude-behavior inconsistency. Applications to research methodology Attitude-behavior consistency is an important concept for social science research because claims are often made about behavior based on evidence which is really about attitudes. The attitudinal fallacy is committed when verbal data are used to support claims not about what people believe or say, but what they do. Data collection methods based on self-report like surveys, and interviews are vulnerable to the attitudinal fallacy if they attempt to measure behavior and if reported attitudes are inconsistent with the behavior. Research methods that directly observe behaviors avoid the attitudinal fallacy as a matter of course. However many kinds of behavior are not easily observed. Ethnography can make rich observations and descriptions of behavior and allow for comparison between behavior and attitude. Unfortunately, in general ethnographic data cannot be used to draw statistically generalizable conclusions about behavior in a population. Ethnographers can still commit the attitudinal fallacy if they rely on quotations as evidence for behaviors. Experiments in laboratories make it possible to observe behavior, with the limitations th
https://en.wikipedia.org/wiki/Azinphos-methyl
Azinphos-methyl (Guthion) (also spelled azinophos-methyl) is a broad spectrum organophosphate insecticide manufactured by Bayer CropScience, Gowan Co., and Makhteshim Agan. Like other pesticides in this class, it owes its insecticidal properties (and human toxicity) to the fact that it is an acetylcholinesterase inhibitor (the same mechanism is responsible for the toxic effects of the V-series nerve agent chemical weapons). It is classified as an extremely hazardous substance in the United States as defined in Section 302 of the U.S. Emergency Planning and Community Right-to-Know Act (42 U.S.C. 11002), and is subject to strict reporting requirements by facilities which produce, store, or use it in significant quantities. History and uses Azinphos-methyl is a neurotoxin derived from nerve agents developed during World War II. It was first registered in the US in 1959 as an insecticide and is also used as active ingredient in organophosphate (OP) pesticides. It is not registered for consumer or residential use. It has been linked to health problems of farmers who apply it, and the U.S. Environmental Protection Agency (EPA) considered a denial of reregistration, citing, “concern to farm workers, pesticide applicators, and aquatic ecosystems. The use of AZM has been fully banned in the USA since 30 September 2013, ending a phase-out period of twelve years. Azinphos-methyl has been banned in the European Union since 2006 and in Turkey since 2013. The New Zealand Environmental Risk Management Authority made a decision to phase out azinphos-methyl over a five-year period starting from 2009. In 2014, it was still used in Australia and partly in New Zealand. Available forms AzM is often used as active ingredient in organophosphate pesticides like Guthion, Gusathion (GUS), Gusathion-M, Crysthyron, Cotnion, Cotnion-methyl, Metriltrizotion, Carfene, Bay 9027, Bay 17147, and R-1852. This is why Guthion is often used as a nickname for AzM. Studies have shown that pure AzM is
https://en.wikipedia.org/wiki/Brodel%27s%20bloodless%20line
Brodel's line, also known as Brodel's white line, is a relatively avascular line between the anterior and posterior segmental branches of the renal artery of the body. It has been proved that Brodel's line is also vascular. Josef Hyrtl, in 1882, and Max Brödel, in 1902, described this relatively avascular plane near the midline (5 mm posterior) of the convex border of the kidney through which the collecting system of the kidney could be entered. In continental Europe, credit for the plane was given to Hyrtl; but in England and the United States it was called the Brödel bloodless line or the Brödel white line. This plane is used for nephrolithotomy, more specifically while performing anatrophic nephrolithotomy.
https://en.wikipedia.org/wiki/Mixture
In chemistry, a mixture is a material made up of two or more different chemical substances which are not chemically bonded. A mixture is the physical combination of two or more substances in which the identities are retained and are mixed in the form of solutions, suspensions and colloids. Mixtures are one product of mechanically blending or mixing chemical substances such as elements and compounds, without chemical bonding or other chemical change, so that each ingredient substance retains its own chemical properties and makeup. Despite the fact that there are no chemical changes to its constituents, the physical properties of a mixture, such as its melting point, may differ from those of the components. Some mixtures can be separated into their components by using physical (mechanical or thermal) means. Azeotropes are one kind of mixture that usually poses considerable difficulties regarding the separation processes required to obtain their constituents (physical or chemical processes or, even a blend of them). Characteristics of mixtures All mixtures can be characterized as being separable by mechanical means (e.g. purification, distillation, electrolysis, chromatography, heat, filtration, gravitational sorting, centrifugation). Mixtures differ from chemical compounds in the following ways: the substances in a mixture can be separated using physical methods such as filtration, freezing, and distillation. there is little or no energy change when a mixture forms (see Enthalpy of mixing). The substances in a mixture keep its separate properties. In the example of sand and water, neither one of the two substances changed in any way when they are mixed. Although the sand is in the water it still keeps the same properties that it had when it was outside the water. mixtures have variable compositions, while compounds have a fixed, definite formula. when mixed, individual substances keep their properties in a mixture, while if they form a compound their properties
https://en.wikipedia.org/wiki/Pumping%20lemma%20for%20regular%20languages
In the theory of formal languages, the pumping lemma for regular languages is a lemma that describes an essential property of all regular languages. Informally, it says that all sufficiently long strings in a regular language may be pumped—that is, have a middle section of the string repeated an arbitrary number of times—to produce a new string that is also part of the language. Specifically, the pumping lemma says that for any regular language there exists a constant such that any string in with length at least can be split into three substrings , and (, with being non-empty), such that the strings constructed by repeating zero or more times are still in . This process of repetition is known as "pumping". Moreover, the pumping lemma guarantees that the length of will be at most , imposing a limit on the ways in which may be split. Languages with a finite number of strings vacuously satisfy the pumping lemma by having equal to the maximum string length in plus one. By doing so, zero strings in have length greater than . The pumping lemma is useful for disproving the regularity of a specific language in question. It was first proven by Michael Rabin and Dana Scott in 1959, and rediscovered shortly after by Yehoshua Bar-Hillel, Micha A. Perles, and Eli Shamir in 1961, as a simplification of their pumping lemma for context-free languages. Formal statement Let be a regular language. Then there exists an integer depending only on such that every string in of length at least ( is called the "pumping length") can be written as (i.e., can be divided into three substrings), satisfying the following conditions: is the substring that can be pumped (removed or repeated any number of times, and the resulting string is always in ). (1) means the loop to be pumped must be of length at least one, that is, not an empty string; (2) means the loop must occur within the first characters. must be smaller than (conclusion of (1) and (2)), but apa
https://en.wikipedia.org/wiki/Trait%C3%A9%20de%20Zoologie
The , complete title popularly known as is a 52 volume synthesis of Zoology published between 1948 and 1979 originally under the direction of Pierre-Paul Grassé. A new edition commenced in 1980. The books were not published in order, and some promised parts are yet to appear. The books are written in the French language. Tomes Tome I, Fascicule 1 (1952) : Phylogénie. Protozoaires : généralités. Flagellés. Tome I, Fascicule 2 (1953) : Rhizopodes, , Sporozoaires, Cnidosporidies. Tome II, Fascicule 1 (1984) : Infusoires ciliés. Tome co-dirigé avec Pierre de Puytorac. Tome II, Fascicule 2 (1994) : Infusoires ciliés. Tome III, Fascicule 1 (1994) : Spongiaires, anatomie, physiologie systématique. Tome III, Fascicule 2 (1997) : Cnidaires (Hydrozoaires, Scyphozoaires et Cubozoaires), Cténaires. Tome co-dirigé avec André Franc. Tome III, Fascicule 3 (1953, 1997) : Cnidaires anthozoaires. Tome co-dirigé avec Dominique Doumenc. Tome IV, Fascicule 1 (1961, 1997) : Plathelminthes, Mésozoaires, Acanthocéphales, Némertiens. Tome IV, Fascicule 2 (1965) : Némathelminthes (Nématodes). Tome IV, Fascicule 3 (1965) : Némathelminthes (Nématodes, Gordiacés), Rotifères, Gastrotriches, Kinorhynques. Tome V, Fascicule 1 (1959, 2007) : Annélides, Myzostomides, Sipunculiens, Echiuriens, Priapuliens, Endoproctes, Phoronidiens. Tome V, Fascicule 2 (1959, 2007) : Bryozoaires, Brachiopodes, Chétognathes, Pogonophores, Mollusques (généralités, Aplacophores, Polyplacophores, Monoplacophores, Bivalves). Tome V, Fascicule 3 (1959, 1997) : Mollusques, Gastéropodes et Scaphopodes. Tome V, Fascicule 4 (1997) : Céphalopodes. Tome co-dirigé avec Katarina Mangold. Tome VI (1949) : Onychophores, Tardigrades, Arthropodes, Trilobitomorphes, Chélicérates. Tome VII, Fascicule 2 (1996) : Crustacés. Tome VIII, Fascicule 1 (1973) : Insectes. Tête, Aile. Tome VIII, Fascicule 2 (1979) : Insectes. Thorax, Abdomen. Tome VIII, Fascicule 3 (1975) : Insectes. Téguments, système nerveux, organes sens
https://en.wikipedia.org/wiki/Bisulfite%20sequencing
Bisulfite sequencing (also known as bisulphite sequencing) is the use of bisulfite treatment of DNA before routine sequencing to determine the pattern of methylation. DNA methylation was the first discovered epigenetic mark, and remains the most studied. In animals it predominantly involves the addition of a methyl group to the carbon-5 position of cytosine residues of the dinucleotide CpG, and is implicated in repression of transcriptional activity. Treatment of DNA with bisulfite converts cytosine residues to uracil, but leaves 5-methylcytosine residues unaffected. Therefore, DNA that has been treated with bisulfite retains only methylated cytosines. Thus, bisulfite treatment introduces specific changes in the DNA sequence that depend on the methylation status of individual cytosine residues, yielding single-nucleotide resolution information about the methylation status of a segment of DNA. Various analyses can be performed on the altered sequence to retrieve this information. The objective of this analysis is therefore reduced to differentiating between single nucleotide polymorphisms (cytosines and thymidine) resulting from bisulfite conversion (Figure 1). Methods Bisulfite sequencing applies routine sequencing methods on bisulfite-treated genomic DNA to determine methylation status at CpG dinucleotides. Other non-sequencing strategies are also employed to interrogate the methylation at specific loci or at a genome-wide level. All strategies assume that bisulfite-induced conversion of unmethylated cytosines to uracil is complete, and this serves as the basis of all subsequent techniques. Ideally, the method used would determine the methylation status separately for each allele. Alternative methods to bisulfite sequencing include Combined Bisulphite Restriction Analysis and methylated DNA immunoprecipitation (MeDIP). Methodologies to analyze bisulfite-treated DNA are continuously being developed. To summarize these rapidly evolving methodologies, numero
https://en.wikipedia.org/wiki/VG-1%20%28cell%20line%29
VG-1 is a B cell line which was derived from primary effusion lymphoma (PEL) . It was first established in 2000 by David T. Scadden’s group at Massachusetts General Hospital. It is infected with Kaposi's sarcoma-associated herpesvirus (KSHV), but negative with Epstein–Barr virus (EBV) .
https://en.wikipedia.org/wiki/YqeY%20protein%20domain
In molecular biology, YqeY is a type of protein domain of unknown function. It is thought to have a role in protein synthesis, facilitating the production of charged transfer RNA used in the process of translating mRNA into protein. It is present as a domain of glutaminyl-tRNA synthetase (GlnRS) in almost all eukaryotes. Function The YqeY domain has been found to be involved in the recognition of tRNA charged with the amino acid glutamine (tRNA-Gln). In some cases YqeY also increases the affinity of GlnRS for tRNA-Gln, but only when present in cis (that is, as part of the GlnRS polypeptide chain). However, the presence of YqeY as a standalone domain in organisms without GlnRS suggests that YqeY domains may have additional cellular functions. Homology This protein domain shares sequence homology with the C-terminal domain of GatB and GatE, the tRNA-binding subunits of bacterial and archaeal glutamine amidotransferases.
https://en.wikipedia.org/wiki/Sievert%20integral
The Sievert integral, named after Swedish medical physicist Rolf Sievert, is a special function commonly encountered in radiation transport calculations. It plays a role in the sievert (symbol: Sv) unit of ionizing radiation dose in the International System of Units (SI). Definition
https://en.wikipedia.org/wiki/SSE2
SSE2 (Streaming SIMD Extensions 2) is one of the Intel SIMD (Single Instruction, Multiple Data) processor supplementary instruction sets introduced by Intel with the initial version of the Pentium 4 in 2000. It extends the earlier SSE instruction set, and is intended to fully replace MMX. Intel extended SSE2 to create SSE3 in 2004. SSE2 added 144 new instructions to SSE, which has 70 instructions. Competing chip-maker AMD added support for SSE2 with the introduction of their Opteron and Athlon 64 ranges of AMD64 64-bit CPUs in 2003. Features Most of the SSE2 instructions implement the integer vector operations also found in MMX. Instead of the MMX registers they use the XMM registers, which are wider and allow for significant performance improvements in specialized applications. Another advantage of replacing MMX with SSE2 is avoiding the mode switching penalty for issuing x87 instructions present in MMX because it is sharing register space with the x87 FPU. The SSE2 also complements the floating-point vector operations of the SSE instruction set by adding support for the double precision data type. Other SSE2 extensions include a set of cache control instructions intended primarily to minimize cache pollution when processing infinite streams of information, and a sophisticated complement of numeric format conversion instructions. AMD's implementation of SSE2 on the AMD64 (x86-64) platform includes an additional eight registers, doubling the total number to 16 (XMM0 through XMM15). These additional registers are only visible when running in 64-bit mode. Intel adopted these additional registers as part of their support for x86-64 architecture (or in Intel's parlance, "Intel 64") in 2004. Differences between x87 FPU and SSE2 FPU (x87) instructions provide higher precision by calculating intermediate results with 80 bits of precision, by default, to minimise roundoff error in numerically unstable algorithms (see IEEE 754 design rationale and references therein
https://en.wikipedia.org/wiki/Domain%20coloring
In complex analysis, domain coloring or a color wheel graph is a technique for visualizing complex functions by assigning a color to each point of the complex plane. By assigning points on the complex plane to different colors and brightness, domain coloring allows for a function from the complex plane to itself — whose graph would normally require four space dimensions — to be easily represented and understood. This provides insight to the fluidity of complex functions and shows natural geometric extensions of real functions. Motivation A graph of a real function can be drawn in two dimensions because there are two represented variables, and . However, complex numbers are represented by two variables and therefore two dimensions; this means that representing a complex function (more precisely, a complex-valued function of one complex variable ) requires the visualization of four dimensions. One way to achieve that is with a Riemann surface, but another method is by domain coloring. Method Representing a four dimensional complex mapping with only two variables is undesirable, as methods like projections can result in a loss of information. However, it is possible to add variables that keep the four-dimensional process without requiring a visualization of four dimensions. In this case, the two added variables are visual inputs such as color and brightness because they are naturally two variables easily processed and distinguished by the human eye. This assignment is called a "color function". There are many different color functions used. A common practice is to represent the complex argument, , (also known as "phase" or "angle") with a hue following the color wheel, and the magnitude by other means, such as brightness or saturation. Simple color function The following example colors the origin in black, in green, in magenta, and a point at infinity in white: There are a number of choices for the function . should be strictly monotonic and continuous. Anot
https://en.wikipedia.org/wiki/Outer%20membrane%20protein%20G
Outer membrane protein G (OmpG) is a porin, a channel proteins in the outer membrane of Gram-negative bacteria. Escherichia coli OmpG forms a 14-stranded beta-barrel and in contrast to most porins, appears to function as a monomer. The central pore of OmpG is wider than other E. coli porins and it is speculated that it may form a non-specific channel for the transport of larger oligosaccharides.
https://en.wikipedia.org/wiki/NOTCH2NL
Notch homolog 2 N-terminal-like is a family of proteins that in humans consists of 3 proteins (NOTCH2NLA, NOTCH2NLB, and NOTCH2NLC) and is encoded by NOTCH2NL gene. It appears to play a key role in the development of the prefrontal cortex, a part of the brain. NOTCH2NL increases the number of cortical stem cells, which while delaying the generation of neurons ultimately leads to a greater number of neurons and larger brains. NOTCH2NL copy number loss and gain is associated with various neurological disorders, and they showed that loss of NOTCH2NL in cortical organoids leads to the organoids being smaller, while resulting in premature differentiation of cortical stem cells into neurons. The role of NOTCH2NL in the development of the human brain together with the evolutionary history of NOTCH2NL genes, suggests that the emergence of NOTCH2NL genes may have contributed to the increase in size of the human neocortex which tripled over the last two million years.
https://en.wikipedia.org/wiki/Spherically%20symmetric%20spacetime
In physics, spherically symmetric spacetimes are commonly used to obtain analytic and numerical solutions to Einstein's field equations in the presence of radially moving matter or energy. Because spherically symmetric spacetimes are by definition irrotational, they are not realistic models of black holes in nature. However, their metrics are considerably simpler than those of rotating spacetimes, making them much easier to analyze. Spherically symmetric models are not entirely inappropriate: many of them have Penrose diagrams similar to those of rotating spacetimes, and these typically have qualitative features (such as Cauchy horizons) that are unaffected by rotation. One such application is the study of mass inflation due to counter-moving streams of infalling matter in the interior of a black hole. Formal definition A spherically symmetric spacetime is a spacetime whose isometry group contains a subgroup which is isomorphic to the rotation group SO(3) and the orbits of this group are 2-spheres (ordinary 2-dimensional spheres in 3-dimensional Euclidean space). The isometries are then interpreted as rotations and a spherically symmetric spacetime is often described as one whose metric is "invariant under rotations". The spacetime metric induces a metric on each orbit 2-sphere (and this induced metric must be a multiple of the metric of a 2-sphere). Conventionally, the metric on the 2-sphere is written in polar coordinates as , and so the full metric includes a term proportional to this. Spherical symmetry is a characteristic feature of many solutions of Einstein's field equations of general relativity, especially the Schwarzschild solution and the Reissner–Nordström solution. A spherically symmetric spacetime can be characterised in another way, namely, by using the notion of Killing vector fields, which, in a very precise sense, preserve the metric. The isometries referred to above are actually local flow diffeomorphisms of Killing vector fields and thus
https://en.wikipedia.org/wiki/Sieve%20theory
Sieve theory is a set of general techniques in number theory, designed to count, or more realistically to estimate the size of, sifted sets of integers. The prototypical example of a sifted set is the set of prime numbers up to some prescribed limit X. Correspondingly, the prototypical example of a sieve is the sieve of Eratosthenes, or the more general Legendre sieve. The direct attack on prime numbers using these methods soon reaches apparently insuperable obstacles, in the way of the accumulation of error terms. In one of the major strands of number theory in the twentieth century, ways were found of avoiding some of the difficulties of a frontal attack with a naive idea of what sieving should be. One successful approach is to approximate a specific sifted set of numbers (e.g. the set of prime numbers) by another, simpler set (e.g. the set of almost prime numbers), which is typically somewhat larger than the original set, and easier to analyze. More sophisticated sieves also do not work directly with sets per se, but instead count them according to carefully chosen weight functions on these sets (options for giving some elements of these sets more "weight" than others). Furthermore, in some modern applications, sieves are used not to estimate the size of a sifted set, but to produce a function that is large on the set and mostly small outside it, while being easier to analyze than the characteristic function of the set. Basic sieve theory For information on notation see at the end. We start with some countable sequence of non-negative numbers . In the most basic case this sequence is just the indicator function of some set we want to sieve. However this abstraction allows for more general situations. Next we introduce a general set of prime numbers called the sifting range and their product up to as a function . The goal of sieve theory is to estimate the sifting function In the case of this just counts the cardinality of a subset of numbers, that ar
https://en.wikipedia.org/wiki/Pandya%20theorem
The Pandya theorem is a good illustration of the richness of information forthcoming from a judicious use of subtle symmetry principles connecting vastly different sectors of nuclear systems. It is a tool for calculations regarding both particles and holes. Description Pandya theorem provides a theoretical framework for connecting the energy levels in jj coupling of a nucleon-nucleon and nucleon-hole system. It is also referred to as Pandya Transformation or Pandya Relation in literature. It provides a very useful tool for extending shell model calculations across shells, for systems involving both particles and holes. The Pandya transformation, which involves angular momentum re-coupling coefficients (Racah-Coefficient), can be used to deduce one-particle one-hole (ph) matrix elements. By assuming the wave function to be "pure" (no configuration mixing), Pandya transformation could be used to set an upper bound to the contributions of 3-body forces to the energies of nuclear states. History It was first published in 1956 as follows: Nucleon-Hole Interaction in jj Coupling S.P. Pandya, Phys. Rev. 103, 956 (1956). Received 9 May 1956 A theorem connecting the energy levels in jj coupling of a nucleon-nucleon and nucleon-hole system is derived, and applied in particular to Cl38 and K40. Shell model Monte Carlo approaches to nuclear level densities Since it is by no means obvious how to extract "pairing correlations" from the realistic shell-model calculations, Pandya transform is applied in such cases. The "pairing Hamiltonian" is an integral part of the residual shell-model interaction. The shell-model Hamiltonian is usually written in the p-p representation, but it also can be transformed to the p-h representation by means of the Pandya transformation. This means that the high-J interaction between pairs can translate into the low-J interaction in the p-h channel. It is only in the mean-field theory that the division into "particle-hole" and "particle-parti
https://en.wikipedia.org/wiki/Wisconsin%20Integrally%20Synchronized%20Computer
The Wisconsin Integrally Synchronized Computer (WISC) was an early digital computer designed and built at the University of Wisconsin–Madison. Operational in 1954, it was the first digital computer in the state. Pioneering computer designer Gene Amdahl drafted the WISC's design as his PhD thesis. The computer was built over the period 1951-1954. It had 1,024 50-bit words (equivalent to about 6 KB) of drum memory, with an operation time of 1/15 second and throughput of 60 operations per second, which was achieved by an early form of instruction pipeline. It was capable of both fixed and floating point operation. It weighed about . Part of it was at the Computer History Museum until about 2020, when it was moved to an unknown location.