source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/CYP170B1
Cytochrome P450 family 170 subfamily B member 1 (abbreviated CYP170B1) is an actinobacterial Cytochrome P450 enzyme originally from Streptomyces albus, which catalyzes the biosynthesis of the tricyclic sesquiterpene antibiotic albaflavenone.
https://en.wikipedia.org/wiki/Tropical%20Eastern%20Pacific
The Tropical Eastern Pacific is one of the twelve marine realms that cover the coastal waters and continental shelves of the world's oceans. The Tropical Eastern Pacific extends along the Pacific Coast of the Americas, from the southern tip of the Baja California Peninsula in the north to northern Peru in the south. It is bounded on the north by the Temperate Northern Pacific realm, and on the south by the Temperate South America realm. It includes a number of oceanic islands and oceanic island groups like the Galápagos Islands, Revillagigedo Islands, Cocos Island, and Clipperton Island. Clipperton has been labelled as a meeting point between the Oceanian realm and the Tropical Eastern Pacific realm. The Galápagos Islands are also believed to have much higher percentages of Indo-West Pacific species in their marine faunas when compared to the continental American coasts. A 2018 study revealed interconnectivity of species between the Line Islands, Hawaii and Clipperton, as well as interconnectivity between the Galápagos Islands and Marquesas Islands in French Polynesia. The study identified the Northern Galápagos Islands as a "critical stepping-stone connecting the Central Pacific and Eastern Pacific." The WWF and Nature Conservancy divide the Tropical Eastern Pacific realm into two marine provinces, Tropical East Pacific and Galápagos, which are further subdivided into marine ecoregions. Ecoregions Tropical East Pacific Marine Province Revillagigedos Clipperton Mexican Tropical Pacific Chiapas–Nicaragua Nicoya Cocos Island Panama Bight Guayaquil Galápagos Marine Province Northern Galápagos Islands Eastern Galápagos Islands Western Galápagos Islands
https://en.wikipedia.org/wiki/Edward%20Teller%20Award
The Edward Teller Award (or the Edward Teller Medal) is an award presented every two years by the American Nuclear Society for "pioneering research and leadership in the use of laser and ion-particle beams to produce unique high-temperature and high-density matter for scientific research and for controlled thermonuclear fusion". It was established in 1999 and is named after Edward Teller. The award carries a $2000 cash prize and an engraved silver medal. Recipients See also List of physics awards List of prizes named after people Publications
https://en.wikipedia.org/wiki/Rhodopirellula%20baltica
Rhodopirellula baltica is a bacterium from the genus of Rhodopirellula which has been isolated from brackish water from the Baltic Sea.
https://en.wikipedia.org/wiki/Lempel%E2%80%93Ziv%E2%80%93Welch
Lempel–Ziv–Welch (LZW) is a universal lossless data compression algorithm created by Abraham Lempel, Jacob Ziv, and Terry Welch. It was published by Welch in 1984 as an improved implementation of the LZ78 algorithm published by Lempel and Ziv in 1978. The algorithm is simple to implement and has the potential for very high throughput in hardware implementations. It is the algorithm of the Unix file compression utility compress and is used in the GIF image format. Algorithm The scenario described by Welch's 1984 paper encodes sequences of 8-bit data as fixed-length 12-bit codes. The codes from 0 to 255 represent 1-character sequences consisting of the corresponding 8-bit character, and the codes 256 through 4095 are created in a dictionary for sequences encountered in the data as it is encoded. At each stage in compression, input bytes are gathered into a sequence until the next character would make a sequence with no code yet in the dictionary. The code for the sequence (without that character) is added to the output, and a new code (for the sequence with that character) is added to the dictionary. The idea was quickly adapted to other situations. In an image based on a color table, for example, the natural character alphabet is the set of color table indexes, and in the 1980s, many images had small color tables (on the order of 16 colors). For such a reduced alphabet, the full 12-bit codes yielded poor compression unless the image was large, so the idea of a variable-width code was introduced: codes typically start one bit wider than the symbols being encoded, and as each code size is used up, the code width increases by 1 bit, up to some prescribed maximum (typically 12 bits). When the maximum code value is reached, encoding proceeds using the existing table, but new codes are not generated for addition to the table. Further refinements include reserving a code to indicate that the code table should be cleared and restored to its initial state (a "clear
https://en.wikipedia.org/wiki/Agile%20testing
Agile testing is a software testing practice that follows the principles of agile software development. Agile testing involves all members of a cross-functional agile team, with special expertise contributed by testers, to ensure delivering the business value desired by the customer at frequent intervals, working at a sustainable pace. Specification by example is used to capture examples of desired and undesired behavior and guide coding. Overview Agile development recognizes that testing is not a separate phase, but an integral part of software development, along with coding. Agile teams use a "whole-team" approach to "baking quality in" to the software product. Testers on agile teams lend their expertise in eliciting examples of desired behavior from customers, collaborating with the development team to turn those into executable specifications that guide coding. Testing and coding are done incrementally and interactively, building up each feature until it provides enough value to release to production. Agile testing covers all types of testing. The Agile Testing Quadrants provide a helpful taxonomy to help teams identify and plan the testing needed. The model of the Agile Testing Quadrants was originally described by Brian Marick, and was popularized by Lisa Crispin and Janet Gregory in their book Agile Testing: A Practical Guide for Testers and Agile Teams. It places different test types on two axis: Technology Facing vs Business Facing, and Support Programming vs Critique Product. Traditional testing methodologies (often employed in the Waterfall model of software development) usually involve a two-team, two-phase process in which the development team builds the product to as near perfection as possible. The software product is delivered late in the software development life cycle at which point the test team strives to find as many bugs/errors as possible. In contrast with these traditional methodologies, Agile testing focuses on repairing faults immediatel
https://en.wikipedia.org/wiki/Video%20camera%20tube
Video camera tubes were devices based on the cathode ray tube that were used in television cameras to capture television images, prior to the introduction of charge-coupled device (CCD) image sensors in the 1980s. Several different types of tubes were in use from the early 1930s, and as late as the 1990s. In these tubes, an electron beam was scanned across an image of the scene to be broadcast focused on a target. This generated a current that was dependent on the brightness of the image on the target at the scan point. The size of the striking ray was tiny compared to the size of the target, allowing 480–486 horizontal scan lines per image in the NTSC format, 576 lines in PAL, and as many as 1035 lines in Hi-Vision. Cathode ray tube Any vacuum tube which operates using a focused beam of electrons, originally called cathode rays, is known as a cathode ray tube (CRT). These are usually seen as display devices as used in older (i.e., non-flat panel) television receivers and computer displays. The camera pickup tubes described in this article are also CRTs, but they display no image. Early research In June 1908, the scientific journal Nature published a letter in which Alan Archibald Campbell-Swinton, fellow of the Royal Society (UK), discussed how a fully electronic television system could be realized by using cathode ray tubes (or "Braun" tubes, after their inventor, Karl Braun) as both imaging and display devices. He noted that the "real difficulties lie in devising an efficient transmitter", and that it was possible that "no photoelectric phenomenon at present known will provide what is required". A cathode ray tube was successfully demonstrated as a displaying device by the German Professor Max Dieckmann in 1906; his experimental results were published by the journal Scientific American in 1909. Campbell-Swinton later expanded on his vision in a presidential address given to the Röntgen Society in November 1911. The photoelectric screen in the proposed transmi
https://en.wikipedia.org/wiki/Auth-Code
An Auth-Code, also known as an EPP code, authorization code, transfer code, or Auth-Info Code, is a generated passcode required to transfer an Internet domain name between domain registrars; the code is intended to indicate that the domain name owner has authorized the transfer. Auth-Codes are created by the current registrar of the domain. The registrar is required to provide the Auth-Code to the domain name owner within five calendar days of the owner's request, and ICANN accepts complaints about registrars that do not. Some registrars allow Auth-Codes to be generated by the domain owners through the registrar's website. All Generic top-level domains use an Auth-Code in their transfer process. The .nz domain registry used an eight-character Auth-Code called Unique Domain Authentication Identifier (UDAI) for domain transfers and name conflict procedures. The UDAI was provided to the domain owner by the domain's current registrar, and expired after 30 days. With the .nz registry update in 2022 the term UDAI was retired, and the passcode is now also referred to as an Auth-Code. Alternative systems The .uk and .co.uk domain registry, instead of using a passcode, has the domain owner specify the new registrar using the old registrar. The destination registrar is specified using the destination's registrar tag, also known as an Internet Provider Security (IPS) tag or Nominet Provider tag. Some registries use a document based approach either in conjunction with or instead of an Auth-Code. An example for that is .hu for which the registrant has to fill out a document and send it to the new registrar, who sends it to the registry to fulfill the domain transfer. The .is domain registry uses the domains admin NIC handle and the old registrar has to update it to the new registrars NIC handle. Some other registries use an email template (that may or may not be in part processed automatically) like .lr or .jm. In this case the technical contact is set to the registrar and
https://en.wikipedia.org/wiki/Personal%20initiative
Personal initiative (PI) is self-starting and proactive behavior that overcomes barriers to achieve a goal. The concept was developed by Michael Frese and coworkers in the 1990s . The three facets of PI – self-starting, future oriented, and overcoming barriers form a syndrome of proactive behaviors relating to each other empirically. Self-starting implies that the goals are set by an individual themselves and not by someone else. These self-started goals are often related to future orientation that involves having long-term focus and preparation for future demands and problems. Future demands can be met by proactive actions – 'pro' meaning preparatory or beforehand in Greek. Thus, a proactive approach attempts to get pre-signals signifying future obstacles and developing plans to prevent them. Implementation of long-term goals often leads to new setbacks. Initiative, therefore, implies that one will overcome these barriers actively and persistently. PI stands in contrast to a passive approach, which is characterized by doing what one is told, giving up when faced with difficulties, and reacting to environmental demands. Proponents of PI have argued that it may become more important in future workplaces as they require a high degree of self-reliance. PI is often conceptualized as the behavioral component of the general proactivity concept; it is also related but not identical to work engagement. Relevance PI is developed as a performance outcome within the action (regulation) theory tradition. While PI is consistently linked to higher work performance and innovativeness of individuals, it also interacts with other types of constructs of intrinsic motivation. The PI concept was used to create an effective training tool, which is now being used to help thousands of entrepreneurs and organisations in developing countries to improve their businesses. Climate PI Climate refers to formal and informal organizational practices which guide and support a proactive, sel
https://en.wikipedia.org/wiki/Alexander%20Anderson%20%28mathematician%29
Alexander Anderson ( in Aberdeen – in Paris) was a Scottish mathematician. Life He was born in Aberdeen, possibly in 1582, according to a print which suggests he was aged 35 in 1617. It is unknown where he was educated, but it is likely that he initially studied writing and philosophy (the "belles lettres") in his home city of Aberdeen. He then went to the continent, and was a professor of mathematics in Paris by the start of the seventeenth century. There he published or edited, between the years 1612 and 1619, various geometric and algebraic tracts. He described himself as having "more wisdom than riches" in the dedication of Vindiciae Archimedis (1616). He was first cousin of David Anderson of Finshaugh, a celebrated mathematician, and David Anderson's daughter was the mother of mathematician James Gregory. Work He was selected by the executors of François Viète to revise and edit Viète's manuscript works. Viète died in 1603, and it is unclear if Anderson knew him, but his eminence was sufficient to attract the attention of the dead man's executors. Anderson corrected and expanded upon Viète's manuscripts, which extended known geometry to the new algebra, which used general symbols to represent quantities. Publications The known works of Anderson amount to six thin quarto volumes, and as the last of them was published in 1619, it is probable that the author died soon after that year, but the precise date is unknown. He wrote other works that have since been lost. From his last work it appears he wrote another piece, "A Treatise on the Mensuration of Solids," and copies of two other works, Ex. Math. and Stereometria Triangulorum Sphæricorum, were in the possession of Sir Alexander Hume until the after the middle of the seventeenth century. 1612: Supplementum Apollonii Redivivi 1615: Ad Angularum Sectionem Analytica Theoremata F. Vieta 1615: Pro Zetetico Apolloniani 1615: Francisci Vietae Fontenaeensis 1616: Vindiciae Archimedis 1619: Alexandri Andersoni Exe
https://en.wikipedia.org/wiki/Identity-based%20cryptography
Identity-based cryptography is a type of public-key cryptography in which a publicly known string representing an individual or organization is used as a public key. The public string could include an email address, domain name, or a physical IP address. The first implementation of identity-based signatures and an email-address based public-key infrastructure (PKI) was developed by Adi Shamir in 1984, which allowed users to verify digital signatures using only public information such as the user's identifier. Under Shamir's scheme, a trusted third party would deliver the private key to the user after verification of the user's identity, with verification essentially the same as that required for issuing a certificate in a typical PKI. Shamir similarly proposed identity-based encryption, which appeared particularly attractive since there was no need to acquire an identity's public key prior to encryption. However, he was unable to come up with a concrete solution, and identity-based encryption remained an open problem for many years. The first practical implementations were finally devised by Sakai in 2000, and Boneh and Franklin in 2001. These solutions were based on bilinear pairings. Also in 2001, a solution was developed independently by Clifford Cocks. Closely related to various identity-based encryption schemes are identity based key agreement schemes. One of the first identity based key agreement algorithms was published in 1986, just two years after Shamir's identity based signature. The author was E. Okamoto. Identity based key agreement schemes also allow for "escrow free" identity based cryptography. A notable example of such an escrow free identity based key agreement is the McCullagh-Barreto's "Authenticated Key Agreement without Escrow" found in section 4 of their 2004 paper, "A New Two-Party Identity-Based Authenticated Key Agreement". A variant of this escrow free key exchange is standardized as the identity based key agreement in the Chinese
https://en.wikipedia.org/wiki/Nakai%20conjecture
In mathematics, the Nakai conjecture is an unproven characterization of smooth algebraic varieties, conjectured by Japanese mathematician Yoshikazu Nakai in 1961. It states that if V is a complex algebraic variety, such that its ring of differential operators is generated by the derivations it contains, then V is a smooth variety. The converse statement, that smooth algebraic varieties have rings of differential operators that are generated by their derivations, is a result of Alexander Grothendieck. The Nakai conjecture is known to be true for algebraic curves and Stanley–Reisner rings. A proof of the conjecture would also establish the Zariski–Lipman conjecture, for a complex variety V with coordinate ring R. This conjecture states that if the derivations of R are a free module over R, then V is smooth.
https://en.wikipedia.org/wiki/Van%20Arkel%E2%80%93de%20Boer%20process
The van Arkel–de Boer process, also known as the iodide process or crystal-bar process, was the first industrial process for the commercial production of pure ductile titanium, zirconium and some other metals. It was developed by Anton Eduard van Arkel and Jan Hendrik de Boer in 1925. Now it is used in the production of small quantities of ultrapure titanium and zirconium. It primarily involves the formation of the metal iodides and their subsequent decomposition to yield pure metal. This process was superseded commercially by the Kroll process. Process As seen in the diagram below, impure titanium, zirconium, hafnium, vanadium, thorium or protactinium is heated in an evacuated vessel with a halogen at 50–250 °C. The patent specifically involved the intermediacy of TiI4 and ZrI4, which were volatilized (leaving impurities as solid). At atmospheric pressure TiI4 melts at 150 °C and boils at 377 °C, while ZrI4 melts at 499 °C and boils at 600 °C. The boiling points are lower at reduced pressure. The gaseous metal tetraiodide is decomposed on a white hot tungsten filament (1400 °C). As more metal is deposited the filament conducts better and thus a greater electric current is required to maintain the temperature of the filament. The process can be performed in the span of several hours or several weeks, depending on the particular setup. Generally, the crystal bar process can be performed using any number of metals using whichever halogen or combination of halogens is most appropriate for that sort of transport mechanism, based on the reactivities involved. The only metals it has been used to purify on an industrial scale are titanium, zirconium and hafnium, and in fact is still in use today on a much smaller scale for special purity needs.
https://en.wikipedia.org/wiki/Equational%20logic
First-order equational logic consists of quantifier-free terms of ordinary first-order logic, with equality as the only predicate symbol. The model theory of this logic was developed into universal algebra by Birkhoff, Grätzer, and Cohn. It was later made into a branch of category theory by Lawvere ("algebraic theories"). The terms of equational logic are built up from variables and constants using function symbols (or operations). Syllogism Here are the four inference rules of logic. denotes textual substitution of expression for variable in expression . Next, denotes equality, for and of the same type, while , or equivalence, is defined only for and of type boolean. For and of type boolean, and have the same meaning. History Equational logic was developed over the years (beginning in the early 1980s) by researchers in the formal development of programs, who felt a need for an effective style of manipulation, of calculation. Involved were people like Roland Carl Backhouse, Edsger W. Dijkstra, Wim H.J. Feijen, David Gries, Carel S. Scholten, and Netty van Gasteren. Wim Feijen is responsible for important details of the proof format. The axioms are similar to those used by Dijkstra and Scholten in their monograph Predicate calculus and program semantics (Springer Verlag, 1990), but our order of presentation is slightly different. In their monograph, Dijkstra and Scholten use the three inference rules Leibniz, Substitution, and Transitivity. However, Dijkstra/Scholten system is not a logic, as logicians use the word. Some of their manipulations are based on the meanings of the terms involved, and not on clearly presented syntactical rules of manipulation. The first attempt at making a real logic out of it appeared in A Logical Approach to Discrete Math, however the inference rule Equanimity is missing there, and the definition of theorem is contorted to account for it. The introduction of Equanimity and its use in the proof format is due to Gries a
https://en.wikipedia.org/wiki/Spin%20polarization
In particle physics, spin polarization is the degree to which the spin, i.e., the intrinsic angular momentum of elementary particles, is aligned with a given direction. This property may pertain to the spin, hence to the magnetic moment, of conduction electrons in ferromagnetic metals, such as iron, giving rise to spin-polarized currents. It may refer to (static) spin waves, preferential correlation of spin orientation with ordered lattices (semiconductors or insulators). It may also pertain to beams of particles, produced for particular aims, such as polarized neutron scattering or muon spin spectroscopy. Spin polarization of electrons or of nuclei, often called simply magnetization, is also produced by the application of a magnetic field. Curie law is used to produce an induction signal in electron spin resonance (ESR or EPR) and in nuclear magnetic resonance (NMR). Spin polarization is also important for spintronics, a branch of electronics. Magnetic semiconductors are being researched as possible spintronic materials. The spin of free electrons is measured either by a LEED image from a clean wolfram-crystal (SPLEED) or by an electron microscope composed purely of electrostatic lenses and a gold foil as a sample. Back scattered electrons are decelerated by annular optics and focused onto a ring shaped electron multiplier at about 15°. The position on the ring is recorded. This whole device is called a Mott-detector. Depending on their spin the electrons have the chance to hit the ring at different positions. 1% of the electrons are scattered in the foil. Of these 1% are collected by the detector and then about 30% of the electrons hit the detector at the wrong position. Both devices work due to spin orbit coupling. The circular polarization of electromagnetic fields is due to spin polarization of their constituent photons. In the most generic context, spin polarization is any alignment of the components of a non-scalar (vectorial, tensorial, spinor) field w
https://en.wikipedia.org/wiki/Metepa
Metepa is a chemosterilant, with the capability to restrict ovarian development. Metepa can also result in carcinogenesis, in particular the formation of teratomas. It is sometimes used as a Rocket propellant.
https://en.wikipedia.org/wiki/Michael%20A.%20B.%20Deakin
Michael Andrew Bernard Deakin (12 August 1939 – 5 August 2014) was an Australian mathematician and mathematics educator. He was known for his work as a writer and editor of Function, a mathematics magazine aimed at high school students, and as a biographer of ancient Greek mathematician Hypatia. He won the B. H. Neumann award of the Australian Mathematics Trust in 2003 for his "rich and varied commitment to mathematics enrichment". Education and career Deakin was born 12 August 1939. He grew up in Tasmania, and moved to Melbourne late in his high school education, taking a second matriculation year studying Latin at St Patrick's College, East Melbourne before entering the University of Melbourne in 1957. He completed a bachelor's degree with second-class honours in mathematics at Melbourne in 1961. He went on to earn a master's degree there in 1963, with a thesis on integral equations supervised by Russell Love. Deakin moved to the University of Chicago in 1963 for graduate study, and completed his Ph.D. in 1966. He became a lecturer at Monash University in Melbourne in 1967, but then in 1970 moved to Papua New Guinea to become reader-in-charge in the mathematics department of the Institute of Higher Technical Education. He returned to Monash as a senior reader in 1973. He earned a master's degree in education in 1975 from the University of Exeter, and remained at Monash for the rest of his career. He died on 5 August 2014, survived by his widow, Rayda, and the children of his first marriage. Function In 1976 a group of mathematicians at Monash University led by department chair Gordon Preston recognized the need for a journal focused on "mathematics as mathematicians themselves would recognise it, but addressed to secondary students". A secondary but explicit goal was to encourage young women in mathematics, as at that time their under-representation was already recognized. Later, over beers with friends from other disciplines, Deakin found the name for the new
https://en.wikipedia.org/wiki/Rhodothermus
Rhodothermus is a genus of bacteria. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI) See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/Neutrophil
Neutrophils (also known as neutrocytes, heterophils or polymorphonuclear leukocytes) are a type of white blood cell. More specifically, they form the most abundant type of granulocytes and make up 40% to 70% of all white blood cells in humans. They form an essential part of the innate immune system, with their functions varying in different animals. They are formed from stem cells in the bone marrow and differentiated into subpopulations of neutrophil-killers and neutrophil-cagers. They are short-lived (between 5 and 135 hours, see ) and highly mobile, as they can enter parts of tissue where other cells/molecules cannot. Neutrophils may be subdivided into segmented neutrophils and banded neutrophils (or bands). They form part of the polymorphonuclear cells family (PMNs) together with basophils and eosinophils. The name neutrophil derives from staining characteristics on hematoxylin and eosin (H&E) histological or cytological preparations. Whereas basophilic white blood cells stain dark blue and eosinophilic white blood cells stain bright red, neutrophils stain a neutral pink. Normally, neutrophils contain a nucleus divided into 2–5 lobes. Neutrophils are a type of phagocyte and are normally found in the bloodstream. During the beginning (acute) phase of inflammation, particularly as a result of bacterial infection, environmental exposure, and some cancers, neutrophils are one of the first responders of inflammatory cells to migrate toward the site of inflammation. They migrate through the blood vessels and then through interstitial space, following chemical signals such as interleukin-8 (IL-8), C5a, fMLP, Leukotriene B4, and H2O2 in a process called chemotaxis. They are the predominant cells in pus, accounting for its whitish/yellowish appearance. Neutrophils are recruited to the site of injury within minutes following trauma and are the hallmark of acute inflammation; however, due to some pathogens being indigestible, they might not be able to resolve certain i
https://en.wikipedia.org/wiki/Bornyl%20acetate
Bornyl acetate is a chemical compound. Its molecular formula is C12H20O2 and its molecular weight is 196.29 g/mol. It is the acetate ester of borneol. It is used as a food additive, flavouring agent, and odour agent. It is a component of the essential oil from pine needles (from the family Pinaceae) and primarily responsible for its odor.
https://en.wikipedia.org/wiki/Endostatin
Endostatin is a naturally occurring, 20-kDa C-terminal fragment derived from type XVIII collagen. It is reported to serve as an anti-angiogenic agent, similar to angiostatin and thrombospondin. Endostatin is a broad-spectrum angiogenesis inhibitor and may interfere with the pro-angiogenic action of growth factors such as basic fibroblast growth factor (bFGF/FGF-2) and vascular endothelial growth factor (VEGF). Background Endostatin is an endogenous inhibitor of angiogenesis. It was first found secreted in the media of non-metastasizing mouse cells from a hemangioendothelioma cell line in 1997 and was subsequently found in humans, e.g. in platelets. It is produced by proteolytic cleavage of collagen XVIII, a member of the multiplexin family that is characterized by interruptions in the triple helix creating multiple domains, by proteases such as cathepsins. Collagen is a component of epithelial and endothelial basement membranes. Endostatin, as a fragment of collagen 18, demonstrates a role of the ECM in suppression of neoangiogenesis. Pro-angiogenic and anti-angiogenic factors can also be created by proteolysis during coagulation cascades. Endogenous inhibitors of angiogenesis are present in both normal tissue and cancerous tissue. Overall, endostatin down regulates many signaling cascades like ephrin, TNF-α, and NFκB signaling as well as coagulation and adhesion cascades. Other collagen derived antiangiogenic factors include arresten, canstatin, tumstatin, α 6 collagen type IV antiangiogenic fragment, and restin. Structure Human monomeric endostatin is a globular protein containing two disulfide bonds: Cys162−302 and Cys264−294. It folds tightly, has a zinc binding domain at the N-terminus of the protein, and has a high affinity for heparin through an 11 arginine basic patch. Endostatin also binds all heparan sulfate proteoglycans with low affinity. Oligomeric endostatin (trimer or dimer) binds mainly with laminin of the basal lamina. Biological a
https://en.wikipedia.org/wiki/2-Phenylphenol
2-Phenylphenol, or o-phenylphenol, is an organic compound. In terms of structure, it is one of the monohydroxylated isomers of biphenyl. It is a white solid. It is a biocide used as a preservative with E number E231 and under the trade names Dowicide, Torsite, Fungal, Preventol, Nipacide and many others. Uses The primary use of 2-phenylphenol is as an agricultural fungicide. It is generally applied post-harvest. It is a fungicide used for waxing citrus fruits. It is no longer a permitted food additive in the European Union, but is still allowed as a post-harvest treatment in 4 EU countries. It is also used for disinfection of seed boxes. It is a general surface disinfectant, used in households, hospitals, nursing homes, farms, laundries, barber shops, and food processing plants. It can be used on fibers and other materials. It is used to disinfect hospital and veterinary equipment. Other uses are in rubber industry and as a laboratory reagent. It is also used in the manufacture of other fungicides, dye stuffs, resins and rubber chemicals. 2-Phenylphenol is found in low concentrations in some household products such as spray disinfectants and aerosol or spray underarm deodorants. The sodium salt of orthophenyl phenol, sodium orthophenyl phenol, is a preservative, used to treat the surface of citrus fruits. Orthophenyl phenol is also used as a fungicide in food packaging and may migrate into the contents. Preparation It is prepared by condensation of cyclohexanone to give cyclohexenylcyclohexanone. The latter undergoes dehydrogenation to give 2-phenylphenol. Safety LD50 (rats) is 2700 to 3000 mg/kg.
https://en.wikipedia.org/wiki/Gnodal
Gnodal was a computer networking company headquartered in Bristol, UK. The company designed and sold network switches for datacenter, high-performance computing and high-frequency trading environments. Gnodal's products were based on its own Peta ASIC, which was the basis of a family of 1RU and 2RU 10 Gigabit Ethernet low latency switches, including the GS7200 switch. Gnodal's products ran a Linux-based network operating system, GnoS OS), which was based on Aricent's ISS product. Corporate history Gnodal was founded in March 2007 by Fred Homewood, Tony Ford, Jon Beecroft, David Hewson and Ed Turner (all formerly of Quadrics), Matt Hatch joined the founding group in July 2007. The company was funded by this group until April 2008, then receiving £1.1m seed funding from Adrian Beecroft (former CIO of Apax Partners), NESTA, South West Ventures Fund and Finance South West Growth Fund (both managed at the time by YFM). Will Leonard (formerly at Xyratex) was appointed as a non-executive director and chairman by the investors at this time. Dawn Capital (an ECF) joined the other funds in an A round of unspecified size completed in July 2009. On 1 September 2010, Eddie Minshull joined the company as chairman, replacing Will Leonard. At the time Minshull was also CEO of MLLTelecom and was formerly EVP of Worldwide Field Operations for Juniper Networks. In 2011, Gnodal launched its first products at Interop in Las Vegas, May 2011. Also in 2011, Mike Cartwright joined the company as CFO. In March 2012, Bob Fernander joined the company, replacing Fred Homewood as CEO. Fred took the role of full-time CTO after Bob's appointment. At the Interop trade show in May 2012 Gnodal won Best of Interop, for networking with its 72 port 40GbE switch in Las Vegas. At 2RU for 72x 40GbE ports, this switch was the most dense at the time. In March 2013, Beb Fernander left Gnodal and Eddie Minshull was appointed Executive Chairman, in a significant reorganisation that included a major r
https://en.wikipedia.org/wiki/Eukaryotic%20translation%20termination%20factor%201
Eukaryotic translation termination factor 1 (eRF1), also known as TB3-1, is a protein that in humans is encoded by the ETF1 gene. In eukaryotes and archaea, this is the sole class 1 release factor (eRF) which recognizes all three stop codons. The overall process of termination is similar in bacteria, but in the latter 2 separate codon-recognizing release factors exist, RF1 and RF2. Function Termination of protein biosynthesis and release of the nascent polypeptide chain are signaled by the presence of an in-frame stop codon at the aminoacyl site of the ribosome. The process of translation termination is universal and is mediated by protein release factors (RFs) and GTP. A class 1 RF recognizes the stop codon and promotes the hydrolysis of the ester bond linking the polypeptide chain with the peptidyl site tRNA, a reaction catalyzed at the peptidyl transferase center of the ribosome. Class 2 RFs, which are not codon specific and do not recognize codons, stimulate class 1 RF activity and confer GTP dependency upon the process. In bacteria, both class 1 RFs, RF1 and RF2, recognize UAA; however, UAG and UGA are decoded specifically by RF1 and RF2, respectively. In eukaryotes, eRF1, or ETF1, the functional counterpart of RF1 and RF2, functions as an omnipotent RF, decoding all 3 stop codons.
https://en.wikipedia.org/wiki/Eve%20Online
Eve Online (stylised EVE Online) is a space-based, persistent world massively multiplayer online role-playing game (MMORPG) developed and published by CCP Games. Players of Eve Online can participate in a number of in-game professions and activities, including mining, piracy, manufacturing, trading, exploration, and combat (both player versus environment and player versus player). The game contains a total of 7,800 star systems that can be visited by players. The game is renowned for its scale and complexity with regards to player interactions. In its single, shared game world, players engage in unscripted economic competition, warfare, and political schemes with other players. The Bloodbath of B-R5RB, a battle involving thousands of players in a single star system, took 21 hours and was recognized as one of the largest and most expensive battles in gaming history. Eve Online was exhibited at the Museum of Modern Art with a video including the historical events and accomplishments of the playerbase. Eve Online was released in North America and Europe in May 2003. It was published from May to December 2003 by Simon & Schuster Interactive, after which CCP purchased the rights and began to self-publish via a digital distribution scheme. On January 22, 2008, it was announced that Eve Online would be distributed via Steam. On March 10, 2009, the game was again made available in boxed form in stores, released by Atari. In February 2013, Eve Online reached over 500,000 subscribers. On November 11, 2016, Eve Online added a limited free-to-play version. Background Set more than 21,000 years in the future, the background story of Eve Online explains that humanity, having used up most of Earth's resources through centuries of explosive population growth, began colonizing the rest of the Milky Way. As on Earth, this expansion also led to competition and fighting over available resources, but everything changed with the discovery of a natural wormhole leading to an unexplored
https://en.wikipedia.org/wiki/Roland%20Dobrushin
Roland Lvovich Dobrushin () (July 20, 1929 – November 12, 1995) was a mathematician who made important contributions to probability theory, mathematical physics, and information theory. Life and work Dobrushin received his Ph.D. at Moscow State University under the supervision of Andrey Kolmogorov. In statistical mechanics, he introduced (simultaneously with Lanford and Ruelle) the DLR equations for the Gibbs measure. Together with Kotecký and Shlosman, he studied the formation of droplets in Ising-type models, providing mathematical justification of the Wulff construction. He was a foreign member of the American Academy of Arts and Sciences, Academia Europæa and US National Academy of Sciences. The Dobrushin prize was established in his honour. Notes
https://en.wikipedia.org/wiki/Gap%20analysis
In management literature, gap analysis involves the comparison of actual performance with potential or desired performance. If an organization does not make the best use of current resources, or forgoes investment in capital or technology, it may produce or perform below an idealized potential. This concept is similar to an economy's production being below the production possibilities frontier. Gap analysis identifies gaps between the optimized allocation and integration of the inputs (resources), and the current allocation-level. This reveals areas that can be improved. Gap analysis involves determining, documenting and improving the difference between business requirements and current capabilities. Gap analysis naturally flows from benchmarking and from other assessments. Once the general expectation of performance in an industry is understood, it is possible to compare that expectation with the company's current level of performance. This comparison becomes the gap analysis. Such analysis can be performed at the strategic or at the operational level of an organization. Gap analysis is a formal study of what a business is doing currently and where it wants to go in the future. It can be conducted, in different perspectives, as follows: Organization (e.g., Human Resources) Business direction Business processes Information technology Gap analysis provides a foundation for measuring investment of time, money and human resources required to achieve a particular outcome (e.g. to turn the salary payment process from paper-based to paperless with the use of a system). Note that "GAP analysis" has also been used as a means of classifying how well a product or solution meets a targeted need or set of requirements. In this case, "GAP" can be used as a ranking of "Good", "Average" or "Poor". (This terminology appears in the PRINCE2 project management publication.) Gap analysis and new products The need for new products or additions to existing lines may emerge from
https://en.wikipedia.org/wiki/Surface%20plasmon%20polariton
Surface plasmon polaritons (SPPs) are electromagnetic waves that travel along a metal–dielectric or metal–air interface, practically in the infrared or visible-frequency. The term "surface plasmon polariton" explains that the wave involves both charge motion in the metal ("surface plasmon") and electromagnetic waves in the air or dielectric ("polariton"). They are a type of surface wave, guided along the interface in much the same way that light can be guided by an optical fiber. SPPs have a shorter wavelength than light in vacuum at the same frequency (photons). Hence, SPPs can have a higher momentum and local field intensity. Perpendicular to the interface, they have subwavelength-scale confinement. An SPP will propagate along the interface until its energy is lost either to absorption in the metal or scattering into other directions (such as into free space). Application of SPPs enables subwavelength optics in microscopy and photolithography beyond the diffraction limit. It also enables the first steady-state micro-mechanical measurement of a fundamental property of light itself: the momentum of a photon in a dielectric medium. Other applications are photonic data storage, light generation, and bio-photonics. Excitation SPPs can be excited by both electrons and photons. Excitation by electrons is created by firing electrons into the bulk of a metal. As the electrons scatter, energy is transferred into the bulk plasma. The component of the scattering vector parallel to the surface results in the formation of a surface plasmon polariton. For a photon to excite an SPP, both must have the same frequency and momentum. However, for a given frequency, a free-space photon has less momentum than an SPP because the two have different dispersion relations (see below). This momentum mismatch is the reason that a free-space photon from air cannot couple directly to an SPP. For the same reason, an SPP on a smooth metal surface cannot emit energy as a free-space photon in
https://en.wikipedia.org/wiki/Weak-Link%20Approach
The Weak-Link Approach (WLA) is a supramolecular coordination-based assembly methodology, first introduced in 1998 by the Mirkin Group at Northwestern University. This method takes advantage of hemilabile ligands -ligands that contain both strong and weak binding moieties- that can coordinate to metal centers and quantitatively assemble into a single condensed ‘closed’ structure (Figure 1). Unlike other supramolecular assembly methods, the WLA allows for the synthesis of supramolecular complexes that can be modulated from rigid ‘closed’ structures to flexible ‘open’ structures through reversible binding of allosteric effectors at the structural metal centers. The approach is general and has been applied to a variety of metal centers and ligand designs including those with utility in catalysis and allosteric regulation. Weak-Link Approach components There are three main components of the WLA methodology that enable the in situ control of supramolecular architecture: 1) the utilization of hemilabile ligands, 2) the choice of metal centers, and 3) the type of allosteric effector. Hemilabile ligands utilized in the WLA A key component of the WLA is the use of hemilabile ligands. Hemilabile ligands are polydentate chelates that contain at least two different types of bonding groups, denoted X and Y (Figure 2). The first group (X) bonds strongly to the metal center, while the other group (Y) is weakly bonding and easily displaced by coordinating ligands or solvent molecules (Z). In this way, the substitutionally labile group (Y) can be displaced from the metal center yet remain available for recoordination. For WLA-generated structures, a typical ligand design consists of a phosphine-based strong binding group and a weak-binding group containing O, S, Se, or N. More recent reports have utilized N-heterocyclic carbenes (NHC) as the strong-binding moiety. By using a combination of NHC- and phosphine-based hemilabile ligands, heteroligated complexes, and macrocycles have b
https://en.wikipedia.org/wiki/Static%20grass
Static grass is used in scale models and miniatures to create realistic-looking grass textures. It consists of small coloured fibres charged with static electricity, making them stand on end when sprinkled onto a surface coated with glue that then hardens, holding the fibres in place. Static grass is usually prepared by applying a layer of glue on the surface, then pouring the fibres on and tipping off the excess. The fibres can also be applied with a shaker, also known as a puffer. Static grass consists of man-made fibres selected for their ability to hold a static electric charge. They are usually a blend of coloured nylon, rayon, or polyester fibres that are used to more realistically replicate grass on a modeller's layout. The fibres are usually sold by weight in 2, 4, 6, 10 and 12 millimetre lengths, although fibres can be found from as little as 0.5 mm in length. If using an electronic applicator, the fibres are attracted to the adhesive vertically and "end-on", giving the grass-like effect the modeller requires. The application sequence is as follows: Apply adhesive to the area to be covered with grass; Ground the applicator to the adhesive area; Load the applicator with fibres; Apply the fibres; Allow the adhesive to dry; Remove excess fibres. Once the basic technique is mastered, advanced techniques can be learned, such as developing differing lengths, dead grass and creating grass tufts, to enhance realism.. Several companies produce static grass products, including PECO, Woodland Scenics and WW Scenics. See also Rail transport modelling#Scatter (modeling)), alternative that may be just dyed sawdust External links How to Make a Homemade Static Grass Applicator for Model Train Scenery Easy Static Grass application by Craig Stocks Static Grass Applicator & Static Grass specialists — DoubleO Scenics Static Grass Applicator — GrassTech USA Scale modeling
https://en.wikipedia.org/wiki/Elliptic%20boundary%20value%20problem
In mathematics, an elliptic boundary value problem is a special kind of boundary value problem which can be thought of as the stable state of an evolution problem. For example, the Dirichlet problem for the Laplacian gives the eventual distribution of heat in a room several hours after the heating is turned on. Differential equations describe a large class of natural phenomena, from the heat equation describing the evolution of heat in (for instance) a metal plate, to the Navier-Stokes equation describing the movement of fluids, including Einstein's equations describing the physical universe in a relativistic way. Although all these equations are boundary value problems, they are further subdivided into categories. This is necessary because each category must be analyzed using different techniques. The present article deals with the category of boundary value problems known as linear elliptic problems. Boundary value problems and partial differential equations specify relations between two or more quantities. For instance, in the heat equation, the rate of change of temperature at a point is related to the difference of temperature between that point and the nearby points so that, over time, the heat flows from hotter points to cooler points. Boundary value problems can involve space, time and other quantities such as temperature, velocity, pressure, magnetic field, etc. Some problems do not involve time. For instance, if one hangs a clothesline between the house and a tree, then in the absence of wind, the clothesline will not move and will adopt a gentle hanging curved shape known as the catenary. This curved shape can be computed as the solution of a differential equation relating position, tension, angle and gravity, but since the shape does not change over time, there is no time variable. Elliptic boundary value problems are a class of problems which do not involve the time variable, and instead only depend on space variables. The main example In two dim
https://en.wikipedia.org/wiki/Holochroal%20eye
Holochroal eyes are compound eyes with many tiny lenses (sometimes more than 15,000, each 30-100μm, rarely larger). They are the oldest and most common type of trilobite eye, and found in all orders of trilobite from the Cambrian to the Permian periods. Lenses (composed of calcite) covered a curved, kidney-shaped visual surface in a hexagonal close packing system, with a single corneal membrane covering all lenses. Unlike in schizochroal eyes, adjacent lenses were in direct contact with one another. Lens shape generally depended on cuticle thickness. The lenses of trilobites with thin cuticles were thin and biconvex, whereas those with thick cuticles had thick lenses, which in extreme cases, could be thick columns with the outer surface flattened and the inner surface hemispherical. Regardless of lens thickness, however, the point at which light was focused was roughly the same distance below the lens.
https://en.wikipedia.org/wiki/Penny%20bun
A penny bun or a penny loaf was a small bread bun or loaf which cost one old penny at the time when there were 240 pence to the pound. A penny loaf was a common size loaf of bread in England regulated by the Assize of Bread and Ale act of 1266. The size of the loaf could vary depending on the prevailing cost of the flour used in the baking. The nursery rhyme London Bridge Is Falling Down has a version which includes the line "Build it up with penny loaves". The term appears in the first complete published version of the "To Market, To Market" rhyme in 1805 as "To market, to market, to buy a penny bun, Home again, home again, market is done" in Songs for the Nursery. The expression "penny bun" is Cockney rhyming slang for one, sun and son. "Penny bun" is also the common English name for the cep (French), or Boletus edulis, an edible basidiomycete mushroom. Native to Europe and North America, it is Europe's second most sought-after fungus after truffles. "Cockle to a penny bun" is British slang for racing odds of 10 to 1. See also List of buns
https://en.wikipedia.org/wiki/Bicircular%20matroid
In the mathematical subject of matroid theory, the bicircular matroid of a graph G is the matroid B(G) whose points are the edges of G and whose independent sets are the edge sets of pseudoforests of G, that is, the edge sets in which each connected component contains at most one cycle. The bicircular matroid was introduced by and explored further by and others. It is a special case of the frame matroid of a biased graph. Circuits The circuits, or minimal dependent sets, of this matroid are the bicircular graphs (or bicycles, but that term has other meanings in graph theory); these are connected graphs whose circuit rank is exactly two. There are three distinct types of bicircular graph: The theta graph consists of three paths joining the same two vertices but not intersecting each other. The figure eight graph (or tight handcuff) consists of two cycles having just one common vertex. The loose handcuff (or barbell) consists of two disjoint cycles and a minimal connecting path. All these definitions apply to multigraphs, i.e., they permit multiple edges (edges sharing the same endpoints) and loops (edges whose two endpoints are the same vertex). Flats The closed sets (flats) of the bicircular matroid of a graph can be described as the forests of such that in the induced subgraph of , every connected component has a cycle. Since the flats of a matroid form a geometric lattice when partially ordered by set inclusion, these forests of also form a geometric lattice. In the partial ordering for this lattice, that if each component tree of is either contained in or vertex-disjoint from every tree of , and each vertex of is a vertex of . For the most interesting example, let be with a loop added to every vertex. Then the flats of are all the forests of , spanning or nonspanning. Thus, all forests of a graph form a geometric lattice, the forest lattice of G . As transversal matroids Bicircular matroids can be characterized as the transversal matro
https://en.wikipedia.org/wiki/Cesare%20Burali-Forti
Cesare Burali-Forti (13 August 1861 – 21 January 1931) was an Italian mathematician, after whom the Burali-Forti paradox is named. Biography Burali-Forti was born in Arezzo, and was an assistant of Giuseppe Peano in Turin from 1894 to 1896, during which time he discovered a theorem which Bertrand Russell later realised contradicted a previously proved result by Georg Cantor. The contradiction came to be called the Burali-Forti paradox of Cantorian set theory. He died in Turin. Books by C. Burali-Forti Analyse vectorielle générale: Applications à la mécanique et à la physique. with Roberto Marcolongo (Mattéi & co., Pavia, 1913). Corso di geometria analitico-proiettiva per gli allievi della R. Accademia Militare (G. B. Petrini di G. Gallizio, Torino, 1912). Geometria descrittiva (S. Lattes & c., Torino, 1921). Introduction à la géométrie différentielle, suivant la méthode de H. Grassmann (Gauthier-Villars, 1897). Lezioni Di Geometria Metrico-Proiettiva (Fratelli Bocca, Torino, 1904). Meccanica razionale with Tommaso Boggio (S. Lattes & c., Torino, 1921). Logica Matematica (Hoepli, Milano, 1894). Complete listing of publications and bibliography, 8 pages. Bibliography Primary literature in English translation: Jean van Heijenoort, 1967. A Source Book in Mathematical Logic, 1879-1931. Harvard Univ. Press. 1897. "A question on transfinite numbers," 104-11. 1897. "On well-ordered classes," 111-12. Secondary literature: Ivor Grattan-Guinness, 2000. The Search for Mathematical Roots 1870-1940. Princeton Uni. Press.
https://en.wikipedia.org/wiki/Avant-Garde%20Computing
Avant-Garde Computing, Inc., was a publicly traded American software and computer hardware company active from 1978 to 1990 and based in Mount Laurel, New Jersey. It was most well known for its Net/Command, Net/Adviser, Net/Alert, and Net/Guard suite of network management, monitoring, and security products. The company was acquired by Boole & Babbage in 1990 after a five-year string of losses. Beginning Avant-Garde Computing was founded by Timothy P. Ahlstrom and F. Morgan LaMarche and incorporated in 1978. Ahlstrom and LaMarche were previously 20 year veterans of IBM, both working in that company's marketing department. In their off-time in the early 1970s, the duo built a device that would warn computer operators when a data tape was close to the end of its reel, founding Ahlstrom LaMarche & Co. to market it. The device proliferated rapidly in the computer rooms of various companies, and the duo later sold their company and its patents to Telegentics of Cherry Hill, New Jersey, to reportedly modest profit. After several years, the duo regrouped and discussed starting another business. They decided to invite several top executives of companies who ran large mainframe computer networks to dinner at a restaurant. When Ahlstrom and LaMarche asked them what troubles they frequently encountered, network management was cited as the most challenging task. These conversations inspired the duo to raise the capital to incorporate Avant-Garde Computing in Cherry Hill; in order to secure adequate financing, the two also put second mortgages on their homes. The company's first product, Net/Alert, was announced in October 1979. It was a hardware–software network management suite comprising a light-pen-capable, color CRT monitor, logic analyzer, and graphical software to analyze traffic on a mainframe network and display and print reports based on the collected data. Net/Alert took two years to develop and was primarily the brainchild of Ahlstrom; LaMarche meanwhile possessed t
https://en.wikipedia.org/wiki/Lumped-element%20model
The lumped-element model (also called lumped-parameter model, or lumped-component model) is a simplified representation of a physical system or circuit that assumes all components are concentrated at a single point and their behavior can be described by idealized mathematical models. The lumped-element model simplifies the system or circuit behavior description into a topology. It is useful in electrical systems (including electronics), mechanical multibody systems, heat transfer, acoustics, etc. This is in contrast to distributed parameter systems or models in which the behaviour is distributed spatially and cannot be considered as localized into discrete entities. The simplification reduces the state space of the system to a finite dimension, and the partial differential equations (PDEs) of the continuous (infinite-dimensional) time and space model of the physical system into ordinary differential equations (ODEs) with a finite number of parameters. Electrical systems Lumped-matter discipline The lumped-matter discipline is a set of imposed assumptions in electrical engineering that provides the foundation for lumped-circuit abstraction used in network analysis. The self-imposed constraints are: The change of the magnetic flux in time outside a conductor is zero. The change of the charge in time inside conducting elements is zero. Signal timescales of interest are much larger than propagation delay of electromagnetic waves across the lumped element. The first two assumptions result in Kirchhoff's circuit laws when applied to Maxwell's equations and are only applicable when the circuit is in steady state. The third assumption is the basis of the lumped-element model used in network analysis. Less severe assumptions result in the distributed-element model, while still not requiring the direct application of the full Maxwell equations. Lumped-element model The lumped-element model of electronic circuits makes the simplifying assumption that the attribut
https://en.wikipedia.org/wiki/Cortinarius%20praestans
Cortinarius praestans, also known as the goliath webcap, is a basidiomycete mushroom of the genus Cortinarius. The mushroom has orangish-yellow caps that reach up to in diameter, and thick club-shaped stipes up to long. The edible mushroom is found in Europe. Taxonomy The species was first described as Agaricus praestans by the French botanist François Simon Cordier in 1870. It is commonly known as the "goliath webcap". Description Mature caps are convex, usually grooved at the rolled-in margins, and typically reach diameters between . The surface color of the cap is chestnut or chocolate-brown with violet-copper tint. It has a light covering of surface fibrils and veil remnants that may appear as if pressed against the surface, or like small scales. The gills are whitish with an amethyst tint then later creamy clay- to rust-colored. They are crowded closely together, and have edges that are usually wavy and scalloped. The stem is long and thick, solid, and thickened at the base in an emarginate bulb that is roughly club-shaped to ventricose. The stem surface is covered with silky fibrils, and is whitish-violet when very young, later losing the violet tones. The surface becomes fibrillosely floccose or whitish at the base and violet at the top, later becoming covered with the violet to whitish silky cortina (a cobwebby partial veil). The flesh is thick, whitish, and has a mild taste and no odor. The mushroom is considered a good edible species. The spore print is ochre-rust colored. The spores are fusiformly lemon-shaped, densely covered with small "warts", and measure 15–17 by 8–10 μm. The edges of the gills have cystidiate hairs 4–6 μm wide, protruding 20-40 μm, and degenerate sterile basidia 10–11 μm wide protruding 10–12 μm. The typical basidia measure 30–35 by 12–13 μm. Distribution and habitat The fruit bodies of the fungus grow in deciduous forests, mainly on calcareous soils. It is common in western Europe, particularly England and France. It was on
https://en.wikipedia.org/wiki/Variance-based%20sensitivity%20analysis
Variance-based sensitivity analysis (often referred to as the Sobol’ method or Sobol’ indices, after Ilya M. Sobol’) is a form of global sensitivity analysis. Working within a probabilistic framework, it decomposes the variance of the output of the model or system into fractions which can be attributed to inputs or sets of inputs. For example, given a model with two inputs and one output, one might find that 70% of the output variance is caused by the variance in the first input, 20% by the variance in the second, and 10% due to interactions between the two. These percentages are directly interpreted as measures of sensitivity. Variance-based measures of sensitivity are attractive because they measure sensitivity across the whole input space (i.e. it is a global method), they can deal with nonlinear responses, and they can measure the effect of interactions in non-additive systems. Decomposition of variance From a black box perspective, any model may be viewed as a function Y=f(X), where X is a vector of d uncertain model inputs {X1, X2, ... Xd}, and Y is a chosen univariate model output (note that this approach examines scalar model outputs, but multiple outputs can be analysed by multiple independent sensitivity analyses). Furthermore, it will be assumed that the inputs are independently and uniformly distributed within the unit hypercube, i.e. for . This incurs no loss of generality because any input space can be transformed onto this unit hypercube. f(X) may be decomposed in the following way, where f0 is a constant and fi is a function of Xi, fij a function of Xi and Xj, etc. A condition of this decomposition is that, i.e. all the terms in the functional decomposition are orthogonal. This leads to definitions of the terms of the functional decomposition in terms of conditional expected values, From which it can be seen that fi is the effect of varying Xi alone (known as the main effect of Xi), and fij is the effect of varying Xi and Xj simultaneously, addi
https://en.wikipedia.org/wiki/167th%20meridian%20east
The meridian 167° east of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Asia, the Pacific Ocean, New Zealand, the Southern Ocean, and Antarctica to the South Pole. The 167th meridian east forms a great circle with the 13th meridian west. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 167th meridian east passes through: {| class="wikitable plainrowheaders" ! scope="col" width="130" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | East Siberian Sea | style="background:#b0e0e6;" | |-valign="top" | ! scope="row" | | Chukotka Autonomous Okrug Kamchatka Krai — from |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Bering Sea | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Pacific Ocean | style="background:#b0e0e6;" | |- | ! scope="row" | | Rongelap Atoll |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Pacific Ocean | style="background:#b0e0e6;" | |- | ! scope="row" | | Kwajalein Atoll |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Pacific Ocean | style="background:#b0e0e6;" | Passing just east of (at ) |-valign="top" | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Coral Sea | style="background:#b0e0e6;" | Passing just west of the Duff Islands, (at ) Passing just east of the Vanikolo Islands, (at ) Passing just east of the Torres Islands, (at ) Passing just west of the Banks Islands, (at ) |- | ! scope="row" | | Island of Espiritu Santo |-valign="top" | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Coral Sea | style="background:#b
https://en.wikipedia.org/wiki/Edward%20Flatau
Edward Flatau (27 December 1868, Płock – 7 June 1932, Warsaw) was a Polish neurologist and psychiatrist. He was a co-founder of the modern Polish neurology, an authority on the physiology and pathology of meningitis, co-founder of medical journals Neurologia Polska and Warszawskie Czasopismo Lekarskie, and member of the Polish Academy of Learning. His name in medicine is linked to Redlich-Flatau syndrome, Flatau-Sterling torsion dystonia (type 1), Flatau-Schilder disease, and Flatau's law. His publications greatly influenced the developing field of neurology. He published a human brain atlas (1894), wrote a fundamental book on migraine (1912), established the localization principle of long fibers in the spinal cord (1893), and with Sterling published an early paper (1911) on progressive torsion spasm in children and suggested that the disease has a genetic component. Life He was born in 1868 in Płock, the son of Anna and Ludwik Flatau of assimilated Jewish family. In 1886, he graduated from high school (gymnasium) in Płock (now Marshal Stanisław Małachowski High School, Płock, also known as "Małachowianka"). From 1886, Flatau attended medical school at the University of Moscow, where he graduated eximia cum laude. In Moscow, he was greatly influenced by the psychiatrist Sergei Sergeievich Korsakoff (1854–1900) and the neurologist Alexis Jakovlevich Kozhevnikof (1836–1902). Flatau became a medical doctor in 1892. He spent the years 1893–1899 in Berlin in the laboratories of Emanuel Mendel (1839–1907) and in the University of Berlin under Wilhelm von Waldeyer-Hartz (1836–1921). During that time, he worked together with Alfred Goldscheider (1858–1935), Ernst Viktor von Leyden (1832–1910), Hermann Oppenheim, Louis Jacobsohn, Ernst Remak, and Hugo Liepmann. Though he was offered a position of professorship of neurology in Buenos Aires, he returned to Poland and in 1899 settled in Warsaw. He was married twice. He had two daughters, Anna and Joanna Flatau. His first
https://en.wikipedia.org/wiki/Multiscale%20turbulence
Multiscale turbulence is a class of turbulent flows in which the chaotic motion of the fluid is forced at different length and/or time scales. This is usually achieved by immersing in a moving fluid a body with a multiscale, often fractal-like, arrangement of length scales. This arrangement of scales can be either passive or active As turbulent flows contain eddies with a wide range of scales, exciting the turbulence at particular scales (or range of scales) allows one to fine-tune the properties of that flow. Multiscale turbulent flows have been successfully applied in different fields., such as: Reducing acoustic noise from wings by modifying the geometry of spoilers; Enhancing heat transfer from impinging jets passing through grids; Reducing the vortex shedding intensity of flows past normal plates without changing the shedding frequency; Enhancing mixing by energy-efficient stirring; Improving flow metering and flow conditioning in pipes; Improving combustion. Multiscale turbulence has also played an important role into probing the internal structure of turbulence. This sort of turbulence allowed researchers to unveil a novel dissipation law in which the parameter in is not constant, as required by the Richardson-Kolmogorov energy cascade. This new law can be expressed as , with , where and are Reynolds numbers based, respectively, on initial/global conditions (such as free-stream velocity and the object's length scale) and local conditions (such as the rms velocity and integral length scale). This new dissipation law characterises non-equilibrium turbulence apparently universally in various flows (not just multiscale turbulence) and results from non-equilibrium unsteady energy cascade. This imbalance implies that new mean flow scalings exist for free shear turbulent flows, as already observed in axisymmetric wakes
https://en.wikipedia.org/wiki/Product%20of%20experts
Product of experts (PoE) is a machine learning technique. It models a probability distribution by combining the output from several simpler distributions. It was proposed by Geoffrey Hinton in 1999, along with an algorithm for training the parameters of such a system. The core idea is to combine several probability distributions ("experts") by multiplying their density functions—making the PoE classification similar to an "and" operation. This allows each expert to make decisions on the basis of a few dimensions without having to cover the full dimensionality of a problem. This is related to (but quite different from) a mixture model, where several probability distributions are combined via an "or" operation, which is a weighted sum of their density functions. The experts may be understood as each being responsible for enforcing a constraint in a high-dimensional space. A data point is considered likely iff none of the experts say that the point violates a constraint. To optimize it, he proposed the contrastive divergence minimization algorithm. This algorithm is most often used for learning restricted Boltzmann machines. See also Mixture of experts Boltzmann machine
https://en.wikipedia.org/wiki/Michael%20Francis%20Tompsett
Michael Tompsett (born 1939) is a British-born physicist, engineer, and inventor, and the founder director of the US software company TheraManager. He is a former researcher at the English Electric Valve Company, who later moved to Bell Labs in the United States. Tompsett invented CCD imagers and designed and built the first ever video camera with a solid-state (CCD) sensor. Tompsett received the Queen Elizabeth Prize for Engineering in 2017, with Eric Fossum, George Smith, and Nobukazu Teranishi. Tompsett has also received two other lifetime awards; the New Jersey Inventors Hall of Fame 2010 Pioneer Award, and the 2012 IEEE Edison Medal. The thermal-imaging camera tube developed from his invention also earned a Queen's Award in 1987. Tompsett is known particularly for his work on infrared imagers and CCD imagers. He pioneered compact, low power, high performance, and low cost solid-state infrared imagers, CCD imagers, and digital cameras and made contributions in several fields with patents and publications over an extended period of time. He is credited with applying the principle behind the charge-coupled device to invent the CCD imager, used in devices such as digital cameras. Education and career He studied physics at the University of Cambridge and also completed an engineering PhD there (1962–66). Tompsett built a reflection high-energy electron diffraction (RHEED) system to study surfaces. While at English Electric Valve (EEV), he built the first ultra-high-vacuum RHEED system with in-situ deposition to study the structures of thin-films of lead oxide as they were deposited. This understanding was needed to make Plumbicon television camera-tubes. He consulted with VacGen (now VG Scienta) to make a commercial system, the first of which was sold to IBM Labs. In 1968 while still at EEV, Tompsett invented the un-cooled Pyro-electric thermal-imaging camera tube. He also invented a solid-state version, which is now the basis for thermal imagers made today.
https://en.wikipedia.org/wiki/Martin%20measure
In descriptive set theory, the Martin measure is a filter on the set of Turing degrees of sets of natural numbers, named after Donald A. Martin. Under the axiom of determinacy it can be shown to be an ultrafilter. Definition Let be the set of Turing degrees of sets of natural numbers. Given some equivalence class , we may define the cone (or upward cone) of as the set of all Turing degrees such that ; that is, the set of Turing degrees that are "at least as complex" as under Turing reduction. In order-theoretic terms, the cone of is the upper set of . Assuming the axiom of determinacy, the cone lemma states that if A is a set of Turing degrees, either A includes a cone or the complement of A contains a cone. It is similar to Wadge's lemma for Wadge degrees, and is important for the following result. We say that a set of Turing degrees has measure 1 under the Martin measure exactly when contains some cone. Since it is possible, for any , to construct a game in which player I has a winning strategy exactly when contains a cone and in which player II has a winning strategy exactly when the complement of contains a cone, the axiom of determinacy implies that the measure-1 sets of Turing degrees form an ultrafilter. Consequences It is easy to show that a countable intersection of cones is itself a cone; the Martin measure is therefore a countably complete filter. This fact, combined with the fact that the Martin measure may be transferred to by a simple mapping, tells us that is measurable under the axiom of determinacy. This result shows part of the important connection between determinacy and large cardinals.
https://en.wikipedia.org/wiki/Botanical%20Latin
Botanical Latin is a technical language based on Neo-Latin, used for descriptions of botanical taxa. Until 2012, International Code of Botanical Nomenclature mandated Botanical Latin to be used for the descriptions of most new taxa. It is still the only language other than English accepted for descriptions. The names of organisms governed by the Code also have forms based on Latin. Botanical Latin is primarily a written language. It includes taxon names derived from any language or even arbitrarily derived, and consequently there is no single consistent pronunciation system. When speakers of different languages use Botanical Latin in speech, they use pronunciations influenced by their own languages, or, notably in French, there may be variant spellings based on the Latin. There are at least two pronunciation systems used for Latin by English speakers. Neither system, however, works across the full spectrum of botanical names, because many non-Latin words, such as people's names, have been used. Origin Alphonse Pyramus de Candolle described the language in 1880: C'est le latin arrangé par Linné à l'usage des descriptions et, j'oserai dire, à l'usage de ceux qui n'aiment ni les complications grammaticales, ni les phrases disposées sens desus dessous." (Quoted by W. T. Stearn) [It is the Latin chosen by Linnaeus for the purpose of descriptions, and, I dare to say, for the use of those who love neither grammatical complications nor phrases arranged with senses on top of one another.] De Candolle estimated that to learn Botanical Latin would take three months' work for an English speaker not already familiar with any language of Latin origin, and one month for an Italian. William T. Stearn wrote: Botanical Latin is best described as a modern Romance language of special technical application, derived from Renaissance Latin with much plundering of ancient Greek, which has evolved, mainly since 1700 and primarily through the work of Carl Linnaeus (1707–78), to serve as
https://en.wikipedia.org/wiki/DNA%20Specimen%20Provenance%20Assignment
DNA Specimen Provenance Assignment (DSPA) also known as DNA Specimen Provenance Assay, is a molecular diagnostic test used to definitively assign biopsy specimen identity and establish specimen purity during the diagnostic testing cycle for cancer and other histopathological conditions. The term first appeared in the 2011 scientific paper, “The Changing Spectrum of DNA-Based Specimen Provenance Testing in Surgical Pathology,” published in the American Journal of Clinical Pathology, which built upon concepts described in an earlier paper published in the Journal of Urology. Testing description A panel of 16 microsatellite short tandem repeat (STR) markers that recognizes highly variable loci of human DNA is used in a PCR-based assay and analyzed by capillary electrophoresis. The 16 STR loci range from approximately 75 to 400 base pairs in length, and are highly polymorphic in the human population. The collective data from this panel provides a genetic profile, or ‘DNA fingerprint,’ representing the individual. To complete the diagnostic testing cycle for cancer, the genetic profile from the putative malignant specimen(s) is compared with the genetic profile derived from the patient’s DNA reference sample taken via cheek swab at the time of the biopsy procedure. Indications Conventional histopathology is inherently subject to a certain degree of ambiguity due to the possibility of Specimen Provenance Complications (SPCs) resulting from instances of specimen transposition, extraneous/foreign cell contamination or misidentification of cells used in clinical or anatomical pathology. These types of errors are considered the most rapidly growing category of malpractice claims involving pathologists, and most claims involve a specimen switch between patients. In terms of frequency of occurrence, a 2013 study published in the American Journal of Clinical Pathology indicates that 0.93% of biopsies could be subject to an SPC. Furthermore, each case involves at least two
https://en.wikipedia.org/wiki/Atypia
Atypia (from Greek, a + typos, without type; a condition of being irregular or nonstandard) is a histopathologic term for a structural abnormality in a cell, i.e. it is used to describe atypical cells. Atypia can be caused by an infection or irritation if diagnosed in a Pap smear, for example. In the uterus it is more likely to be precancerous. The related concept of dysplasia refers to an abnormality of development, and includes abnormalities on larger, histopathologic scales. Example features Features that constitute atypia have different definitions for different diseases, but often include the following nucleus abnormalities: Enlargement Pleomorphism Nuclear polychromasia, which means variability in nuclear chromatin content. Polychromasia otherwise refers to a disease of immature red blood cells. Numerous mitotic figures Examples for Barrett's esophagus In Barrett's esophagus, features that are classified as atypia but not as dysplasia are mainly: Nuclear stratification, wherein cell nuclei, which are normally located nearly at the same level between adjacent cells, are instead located at different levels. Crowding Hyperchromatism Prominent nucleoli Prognosis It may or may not be a precancerous indication associated with later malignancy, but the level of appropriate concern is highly dependent on the context with which it is diagnosed. For example, already differentiated, specialised cells such as epithelia displaying "cellular atypia" are far less likely to become problematic (cancerous/malignant) than are myeloid progenitor cells of the immune system. The 'further back' in an already specialised, differentiated cell's lineage, the more problematic cellular atypia is likely to be. This is due to the conferring of such atypia to progeny-cells further down the lineage of that cell type. See also Irregularity List of biological development disorders
https://en.wikipedia.org/wiki/Ipomoea%20cairica
Ipomoea cairica is a vining, herbaceous, perennial plant with palmate leaves and large, showy white to lavender flowers. A species of morning glory, it has many common names, including mile-a-minute vine, Messina creeper, Cairo morning glory, coast morning glory and railroad creeper. The species name cairica translates to "from Cairo", the city where this species was first collected. Description A hairless, slim climber with bulbous roots and lignescented base, its leaves are stalked with 2 to 6 cm long petioles. The leaf blade is ovate to circular in outline, 3 to 10 cm long and 6 to 9 cm wide. It is divided into five to seven segments, these are lanceolate, ovate or elliptic, entire and pointed at the tip and base. Often pseudo side-leaves are formed. The lavender-coloured inflorescences are one to a little bloody cymes. The flower stalks are 12 to 20 mm long, the sepals are 6 to 8 mm long, ovate and sting-pointed. The crown is funnel-shaped, 4 to 6 cm long and violet colored. The stamens and the stylus do not protrude beyond the crown. The ovary is hairless. The fruits are spherical capsules approximately 1 cm in diameter containing one or two hairy seeds. Each fruit matures at about 1 cm across and contains hairy seeds. The vine blooms occasionally throughout the months, but more profusely from spring to summer. Range Its exact native range is uncertain, though it is believed to originate from a rather wide area, ranging from Cape Verde to the Arabian Peninsula, including northern Africa, tropical Africa and the Mediterranean. It covers walls, fences or trees, with stems that can measure more than 10 m in length. The altitude at which it has been recorded ranges from 250 to 2250 m. Invasive species Because of human dispersal, it occurs today on most continents as an introduced species and is sometimes a noxious weed and an invasive species, such as along the coast of New South Wales. As well as in the United States, where it occurs in Hawaii, California, all
https://en.wikipedia.org/wiki/Nucular
Nucular is a common, proscribed pronunciation of the word "nuclear". It is a rough phonetic spelling of . The Oxford English Dictionarys entry dates the word's first published appearance to 1943. Dictionary notes This is one of two contentious pronunciations which receive particular mention in the FAQ of the Merriam-Webster Dictionary: Though disapproved of by many, pronunciations ending in \-kyə-lər\ have been found in widespread use among educated speakers, including scientists, lawyers, professors, congressmen, United States cabinet members, and at least two United States presidents and one vice president. While most common in the United States, these pronunciations have also been heard from British and Canadian speakers. The American Heritage Dictionary notes: The pronunciation (noo'kyə-lər), which is generally considered incorrect, is an example of how a familiar phonological pattern can influence an unfamiliar one … [since] much more common is the similar sequence (-kyə-lər), which occurs in words like particular, circular, spectacular, and in many scientific words like molecular, ocular, and vascular. The Oxford English Dictionary notes: The colloquial pronunciation British /ˈnjuːkjʊlə/, U.S. /ˈn(j)ukjələr/ (frequently rendered in written form as nucular[...]) has been criticized in usage guides since at least the mid 20th century [...] although it is now commonly given as a variant in modern dictionaries. In his 1999 book The Big Book of Beastly Mispronunciations, logophile Charles Harrington Elster noted that the vast majority of those he spoke with during the writing of his book as well as 99 percent of the 1985 usage panel of Morris & Morris' Harper Dictionary of Contemporary Usage specifically condemned the use of the word and characterized it as a mispronunciation. Elster's own view on the matter derives from the root of the word: "nucleus". Arguing by analogy, Elster suggests that "Molecular comes from molecule, and particular comes from particle
https://en.wikipedia.org/wiki/Puzzle%20jug
A puzzle jug is a puzzle in the form of a jug, popular in the 18th and 19th centuries. Puzzle jugs of varying quality were popular in homes and taverns. An inscription typically challenges the drinker to consume the contents without spilling them, which, because the neck of the jug is perforated, is impossible to do conventionally. The solution to the puzzle is that the jug has a hidden tube, one end of which is the spout. The tube usually runs around the rim and then down the handle, with its other opening inside the jug and near the bottom. To solve the puzzle, the drinker must suck from the spout end of the tube. To make the puzzle more interesting, it was common to provide a number of additional holes along the tube, which must be closed off before the contents could be sucked. Some jugs even have a hidden hole to make the challenge still more confounding. History The earliest example in England is the Exeter puzzle jug—an example of medieval pottery in Britain. The Exeter puzzle jug dates from about AD 1300 and was originally made in Saintonge, Western France. The puzzle jug is a descendant of earlier drinking puzzles, such as the fuddling cup and the pot crown, each of which has a different solution. Known inscriptions include: Come drink of me and merry be. Come drink your fill, but do not spill. Fill me up with licker sweet / For it is good when fun us do meet. Gentlemen, now try your Skill / I'll hold your Sixpence if you Will / That you don't drink unless you spill. Here, Gentlemen, come try your skill / I'll hold a wager if you will / That you don't drink this liquor all / Without you spill and let some fall. Within this jug there is good liquor / 'tis fit for Parson or for Vicar / but how to drink and not to spill / will test the utmost of your skill See also Bridge-spouted vessel Dribble glass Fuddling cup Pythagorean cup
https://en.wikipedia.org/wiki/Null%20coalescing%20operator
The null coalescing operator (called the Logical Defined-Or operator in Perl) is a binary operator that is part of the syntax for a basic conditional expression in several programming languages, including C#, PowerShell as of version 7.0.0, Perl as of version 5.10, Swift, and PHP 7.0.0. While its behavior differs between implementations, the null coalescing operator generally returns the result of its left-most operand if it exists and is not null, and otherwise returns the right-most operand. This behavior allows a default value to be defined for cases where a more specific value is not available. In contrast to the ternary conditional if operator used as x ? x : y, but like the binary Elvis operator used as x ?: y, the null coalescing operator is a binary operator and thus evaluates its operands at most once, which is significant if the evaluation of x has side-effects. Examples by languages Bourne-like Shells In Bourne shell (and derivatives), "If parameter is unset or null, the expansion of word is substituted. Otherwise, the value of parameter is substituted": #supplied_title='supplied title' # Uncomment this line to use the supplied title title=${supplied_title:-'Default title'} echo "$title" # prints: Default title C# In C#, the null coalescing operator is ??. It is most often used to simplify expressions as follows: possiblyNullValue ?? valueIfNull For example, if one wishes to implement some C# code to give a page a default title if none is present, one may use the following statement: string pageTitle = suppliedTitle ?? "Default Title"; instead of the more verbose string pageTitle = (suppliedTitle != null) ? suppliedTitle : "Default Title"; or string pageTitle; if (suppliedTitle != null) { pageTitle = suppliedTitle; } else { pageTitle = "Default Title"; } The three forms result in the same value being stored into the variable named pageTitle. suppliedTitle is referenced only once when using the ?? operator, and twice in the other
https://en.wikipedia.org/wiki/Bootstrap%20curriculum
Bootstrap is based at Brown University (USA), and builds on the research and development done there. Bootstrap curriculum consists of 4 research-based curricular computer science modules for grades 6-12. The 4 modules are Bootstrap:Algebra, Bootstrap:Reactive, Bootstrap:Data Science, and Bootstrap:Physics. Bootstrap materials reinforce core concepts from mainstream subjects like Math, Physics and more, enabling non-CS teachers to adopt the introductory materials while delivering rigorous and engaging computing content drawn from Computer Science classes at universities like Brown, WPI, and Northeastern. Bootstrap:Algebra is the flagship curriculum for students ages 12–16, teaching algebraic concepts through coding. By the end of the curriculum, each student has designed their own video game using the concepts (e.g. - order of operations, linear functions, function composition, the pythagorean theorem, inequalities in the plane, piecewise functions, and more). Their mission is to take students' excitement around gaming and drive it towards mathematics and computer programming. Beyond simply expanding students’ interest in math, Bootstrap:Algebra is among the first curricula to demonstrate real improvement in students' algebra performance. Bootstrap:Algebra can be integrated into a standalone CS or mainstream math class, and aligns with national and state math standards. And since every child takes algebra - regardless of gender or background - Bootstrap is one of the largest providers of formal CS education to girls and underrepresented students nationwide. The other modules model physics, data science, and sophisticated interactive programs, and can be integrated into Social Studies, Science, Math, Intro and even AP CS Principles courses. Teachers can mix-and-match content across various modules to fit their needs. Bootstrap works with schools, districts and organizations across the United States, reaching hundreds of teachers and tens of thousands of students
https://en.wikipedia.org/wiki/Essential%20dimension
In mathematics, essential dimension is an invariant defined for certain algebraic structures such as algebraic groups and quadratic forms. It was introduced by J. Buhler and Z. Reichstein and in its most generality defined by A. Merkurjev. Basically, essential dimension measures the complexity of algebraic structures via their fields of definition. For example, a quadratic form q : V → K over a field K, where V is a K-vector space, is said to be defined over a subfield L of K if there exists a K-basis e1,...,en of V such that q can be expressed in the form with all coefficients aij belonging to L. If K has characteristic different from 2, every quadratic form is diagonalizable. Therefore, q has a field of definition generated by n elements. Technically, one always works over a (fixed) base field k and the fields K and L in consideration are supposed to contain k. The essential dimension of q is then defined as the least transcendence degree over k of a subfield L of K over which q is defined. Formal definition Fix an arbitrary field k and let /k denote the category of finitely generated field extensions of k with inclusions as morphisms. Consider a (covariant) functor F : /k → . For a field extension K/k and an element a of F(K/k) a field of definition of a is an intermediate field K/L/k such that a is contained in the image of the map F(L/k) → F(K/k) induced by the inclusion of L in K. The essential dimension of a, denoted by ed(a), is the least transcendence degree (over k) of a field of definition for a. The essential dimension of the functor F, denoted by ed(F), is the supremum of ed(a) taken over all elements a of F(K/k) and objects K/k of /k. Examples Essential dimension of quadratic forms: For a natural number n consider the functor Qn : /k → taking a field extension K/k to the set of isomorphism classes of non-degenerate n-dimensional quadratic forms over K and taking a morphism L/k → K/k (given by the inclusion of L in K) to the map sending the isomo
https://en.wikipedia.org/wiki/Gadsden%20flag
The Gadsden flag is a historical American flag with a yellow field depicting a timber rattlesnake coiled and ready to strike. Beneath the rattlesnake are the words "DONT TREAD ON ME". Some modern versions of the flag include an apostrophe. The flag is named for Christopher Gadsden, a South Carolina delegate to the Continental Congress and brigadier general in the Continental Army who designed the flag in 1775 during the American Revolution. He gave the flag to Commodore Esek Hopkins, and it was unfurled on the main mast of Hopkins's flagship USS Alfred on December 20, 1775. Two days later, Congress made Hopkins commander-in-chief of the Continental Navy. He adopted the Gadsden banner as his personal flag, flying it from the mainmast of the flagship while he was aboard. The Continental Marines also flew the flag during the early part of the war. The rattlesnake was a symbol of the unity of the Thirteen Colonies at the start of the Revolutionary War, and it had a long history as a political symbol in America. Benjamin Franklin used it for his Join, or Die woodcut in 1754. Gadsden intended his flag as a warning to Britain not to violate the liberties of its American subjects. The flag has been described as the "most popular symbol of the American Revolution." Its design proclaims an assertive warning of vigilance and willingness to act in defense against coercion. This has led it to be associated with the ideas of individualism and liberty. It is often used in the United States as a symbol for right-libertarianism, classical liberalism, and small government; for distrust or defiance against authorities and government. Appearance and symbolism Variations in appearance Many variations of the Gadsden flag exist. The motto sometimes includes an apostrophe in the word "Don't" and sometimes does not; the typeface used for the motto is sometimes a serif typeface and other times sans-serif. The rattlesnake sometimes is shown as resting on a green ground; representations d
https://en.wikipedia.org/wiki/Adenosine%20thiamine%20diphosphate
Adenosine thiamine diphosphate (AThDP), or thiaminylated adenosine diphosphate (ADP) is a naturally occurring thiamine adenine nucleotide. It was chemically synthesized and exists in small amounts in vertebrate liver. Its biological significance remains unknown. See also Adenosine thiamine triphosphate
https://en.wikipedia.org/wiki/CD28%20family%20receptor
CD28 family receptors are a group of regulatory cell surface receptors expressed on immune cells. The CD28 family in turn is a subgroup of the immunoglobulin superfamily. Two family members, CD28 and ICOS, act as positive regulators of T cell function while another three, BTLA, CTLA-4 and PD-1 act as inhibitors. Ligands for the CD28 receptor family include B7 family proteins. CD28 receptors play a role in the development and proliferation of T cells. The CD28 receptors enhance signals from the T cell receptors (TCR) in order to stimulate an immune response and an anti-inflammatory response on regulatory T cells. Through the promotion of T cell function, CD28 receptors allow effector T cells to combat regulatory T cell-mediated suppression from adaptive immunity. CD28 receptors also elicit the prevention of spontaneous autoimmunity. Function CD28 receptors aid in other T cell processes such as cytoskeletal remodeling, production of cytokines and chemokines and intracellular biochemical reactions (i.e. phosphorylation, transcriptional signaling, and metabolism) that are key for T cell proliferation and differentiation. Ligation of CD28 receptors causes epigenetic, transcriptional and post-translational alterations in T cells. Specifically, CD28 costimulation controls many aspects within T cells, one being the expression of proinflammatory cytokine genes. A particular cytokine gene encodes for IL-2, which influences T cell proliferation, survival, and differentiation. The absence of CD28 costimulation results in the loss of IL-2 production causing the T cells to be anergic. Additionally, CD28 ligation causes arginine-methylation for many proteins. CD28 also drives transcription within T cells and produce signals that lead to IL-2 production and Bcl-xL regulation, an antiapoptotic protein, which are essential for T cell survival. CD28 receptors can be seen on 80% of human CD4+ and 50% of CD8+ T cells, in which this percentage decreases with age. Clinical signifi
https://en.wikipedia.org/wiki/Interpretability
In mathematical logic, interpretability is a relation between formal theories that expresses the possibility of interpreting or translating one into the other. Informal definition Assume T and S are formal theories. Slightly simplified, T is said to be interpretable in S if and only if the language of T can be translated into the language of S in such a way that S proves the translation of every theorem of T. Of course, there are some natural conditions on admissible translations here, such as the necessity for a translation to preserve the logical structure of formulas. This concept, together with weak interpretability, was introduced by Alfred Tarski in 1953. Three other related concepts are cointerpretability, logical tolerance, and cotolerance, introduced by Giorgi Japaridze in 1992–93. See also Interpretation (logic) Interpretation (model theory) Interpretability logic
https://en.wikipedia.org/wiki/Visual%20cycle
The visual cycle is a process in the retina that replenishes the molecule retinal for its use in vision. Retinal is the chromophore of most visual opsins, meaning it captures the photons to begin the phototransduction cascade. When the photon is absorbed, the 11-cis retinal photoisomerizes into all-trans retinal as it is ejected from the opsin protein. Each molecule of retinal must travel from the photoreceptor cell to the RPE and back in order to be refreshed and combined with another opsin. This closed enzymatic pathway of 11-cis retinal is sometimes called Wald's visual cycle after George Wald (1906–1997), who received the Nobel Prize in 1967 for his work towards its discovery. Retinal Retinal is a chromophore that forms photosensitive Retinylidene proteins when covalently bound to proteins called opsins. Retinal can be photoisomerized by itself, but requires to be bound to an opsin protein to both trigger the phototransduction cascade and tune the spectral sensitivity to longer wavelengths, which enable color vision. Retinal is a species of retinoid and the aldehyde form of Vitamin A. Retinal is interconvertible with retinol, the transport and storage form of vitamin A. During the visual cycle, retinal moves between several different isomers and is also converted to retinol and retinyl ester. Retinoids can be derived from the oxidation of carotenoids like beta carotene or can be consumed directly. To reach the retina, it is bound to Retinol Binding Protein (RBP) and Transthyretin, which prevents its filtration in the glomeruli. As in transport via the RBP-Transthyretin pathway, retinoids must always be bound to Chaperone molecules, for several reasons. Retinoids are toxic, insoluble in aqueous solutions, and prone to oxidation, and as such they must be bound and protected when within the body. The body uses a variety of chaperones, particularly in the retina, to transport retinoids. Overview The visual cycle is consistent within mammals, and is summarized
https://en.wikipedia.org/wiki/Sciensus
Sciensus (formerly Healthcare at Home) is a pharmaceutical supplier established in 1992 by founder and former chairman Charles Walsh based in Burton on Trent. The company supplies a wide range of pharmaceuticals and home healthcare with around 1,600 employees dealing with more than 230,000 patients a year. It works with every NHS trust in the UK. In Spring 2015, Healthcare at Home reverted to using in house logistics rather than working with a third party for all deliveries. In Autumn 2015, the company went through a full corporate rebranding, including new logos, literature, strap line and website in an attempt to re-establish the company's status in the market. The company's name was changed in 2022 from "Healthcare at Home LTD" to "Sciensus Pharma Limited", using only "Sciensus" to refer to themselves from then on. In July 2023, 1 patient died and 3 were hospitalised after being administered unlicensed versions of cabazitaxel chemotherapy provided by Sciensus. This follows reports in May 2023 of complaints of poor service and delays raised about the company, sparking a possible CQC review. This resulted in "partial suspension" of its licence by the Medicines and Healthcare products Regulatory Agency (MHRA). History In 2013 Healthcare at Home took over almost 3,000 patients from another drug delivery company, Medco Health Solutions which pulled out of the UK market. Subsequently, the company struggled to maintain services. In March 2014 Healthcare at Home said that it was no longer accepting new ‘high risk’ patients such as those suffering from haemophilia or respiratory diseases like cystic fibrosis because it could not guarantee getting their drugs to them on time or in full. The company blamed IT issues and a ‘system failure’ relating to the firm's outsourcing of its logistics and warehousing departments. Dave Roberts, chief executive of the National Clinical Homecare Association, says there are questions about whether the rapid growth and low profi
https://en.wikipedia.org/wiki/Z22%20%28computer%29
The Z22 was the seventh computer model Konrad Zuse developed (the first six being the Z1, Z2, Z3, Z4, Z5 and Z11, respectively). One of the early commercial computers, the Z22's design was finished about 1955. The major version jump from Z11 to Z22 was due to the use of vacuum tubes, as opposed to the electromechanical systems used in earlier models. The first machines built were shipped to Berlin and Aachen. By the end of 1958 the ZMMD-group had built a working ALGOL 58 compiler for the Z22 computer. ZMMD was an abbreviation for Zürich (where Rutishauser worked), München (workplace of Bauer and Samelson), Mainz (location of the Z22 computer), Darmstadt (workplace of Bottenbruch). In 1961, the Z22 was followed by a logically very similar transistorized version, the Z23. Already in 1954, Zuse had come to an agreement with Heinz Zemanek that his Zuse KG would finance the work of Rudolf Bodo, who helped Zemanek build the early European transistorized computer Mailüfterl, and that after that project Bodo should work for the Zuse KG—there he helped build the transistorized Z23. Furthermore, all circuit diagrams of the Z22 were supplied to Bodo and Zemanek. The University of Applied Sciences, Karlsruhe still has an operational Z22 which is on permanent loan at the ZKM in Karlsruhe. Altogether 55 Z22 computers were produced. In the 1970s, clones of the Z22 using TTL were built by the company Thiemicke Computer. Technical data The typical setup of a Z22 was: 14 words of 38-bit as fast access RAM implemented as core memory 8192 word (38-bit each) magnetic drum memory as RAM One teletype as console and main input/output device Additional punch tape devices as fast input/output devices 600 tubes working as flip-flops electrical cooling unit, needing a water tap connection (water cooling, so to say) 380 V 16 A three-phase power supply The Z22 operated at 3 kHz operating frequency, which was synchronous with the speed of the drum storage. The input of data and
https://en.wikipedia.org/wiki/Arthur%20Besse
Arthur Besse is a pseudonym chosen by a group of French differential geometers, led by Marcel Berger, following the model of Nicolas Bourbaki. A number of monographs have appeared under the name. Bibliography Actes de la Table Ronde de Géométrie Différentielle. [Proceedings of the Roundtable on Differential Geometry] En l'honneur de Marcel Berger. [In honor of Marcel Berger] Held in Luminy, July 12–18, 1992. Edited by Arthur L. Besse. Séminaires et Congrès [Seminars and Congresses], 1. Société Mathématique de France, Paris; distributed by American Mathematical Society, Providence, RI, 1996. Besse, Arthur L.: Some trends in Riemannian geometry. Duration and change, 71–105, Springer, Berlin, 1994 . Besse, A. Многообразия Эйнштейна. Том I,II. (Russian) [Einstein manifolds. Vol. I, II] Translated from the English and with a preface by D. V. Alekseevskiĭ. "Mir", Moscow, 1990. Vol. I: 320 pp.; Vol. II: pp. 321–704. Besse, Arthur L.: Einstein manifolds. Ergebnisse der Mathematik und ihrer Grenzgebiete (3) [Results in Mathematics and Related Areas (3)], 10. Springer-Verlag, Berlin, 1987 . Четырехмерная риманова геометрия. (Russian) [Riemannian geometry in dimension 4] Семинар Артура Бессе 1978/79. [The Arthur Besse seminar 1978/79] Translated from the French by G. B. Shabat. Translation edited by A. N. Tyurin. "Mir", Moscow, 1985. Géométrie riemannienne en dimension 4. (French) [Riemannian geometry in dimension 4] Papers from the Arthur Besse seminar held at the Université de Paris VII, Paris, 1978/1979. Edited by Lionel Bérard-Bergery, Marcel Berger and Christian Houzel. Textes Mathématiques [Mathematical Texts], 3. CEDIC, Paris, 1981. Besse, Arthur L. Многообразия с замкнутыми геодезическими. (Russian) [Manifolds all of whose geodesics are closed] Translated from the English by Yu. S. Osipov, I. D. Novikov and Yu. P. Solovʹev. Edited and with a preface by Vladimir Mikhaĭlovich Alekseev. "Mir", Moscow, 1981. Besse, Arthur L. Manifolds all of whose geodesics are close
https://en.wikipedia.org/wiki/Resolvent%20set
In linear algebra and operator theory, the resolvent set of a linear operator is a set of complex numbers for which the operator is in some sense "well-behaved". The resolvent set plays an important role in the resolvent formalism. Definitions Let X be a Banach space and let be a linear operator with domain . Let id denote the identity operator on X. For any , let A complex number is said to be a regular value if the following three statements are true: is injective, that is, the corestriction of to its image has an inverse ; is a bounded linear operator; is defined on a dense subspace of X, that is, has dense range. The resolvent set of L is the set of all regular values of L: The spectrum is the complement of the resolvent set: The spectrum can be further decomposed into the point/discrete spectrum (where condition 1 fails), the continuous spectrum (where conditions 1 and 3 hold but condition 2 fails) and the residual/compression spectrum (where condition 1 holds but condition 3 fails). If is a closed operator, then so is each , and condition 3 may be replaced by requiring that is surjective. Properties The resolvent set of a bounded linear operator L is an open set. More generally, the resolvent set of a densely defined closed unbounded operator is an open set.
https://en.wikipedia.org/wiki/WFNA%20%28TV%29
WFNA (channel 55) is a television station licensed to Gulf Shores, Alabama, United States, serving as the CW outlet for southwest Alabama and northwest Florida. It is owned and operated by network majority owner Nexstar Media Group alongside Mobile-licensed CBS affiliate WKRG-TV (channel 5). The two stations share studios with several radio stations owned by iHeartMedia on Broadcast Drive in southwest Mobile; WFNA's transmitter is located in unincorporated Baldwin County near Spanish Fort, Alabama. History Prior to the station's sign-on, WFNA's call letters were originally planned to be WGMP (standing for "Gulf Shores, Mobile, Pensacola"). The station first signed on the air as WBPG on September 2, 2001; it replaced WFGX (channel 35) as the area's WB affiliate after the station reverted to independent status four days earlier on August 31. The station was originally owned by Pegasus Broadcasting. At the time, WFGX's signal was all but unviewable over-the-air on the Alabama side of the market, but WBPG's signal decently covered the entire market. In 2003, Emmis Communications purchased the station, which created a duopoly with Fox affiliate WALA-TV (channel 10); WBPG's operations were subsequently merged with WALA at the latter station's facility on Satchel Paige Drive. LIN TV Corporation acquired WALA-TV on November 30, 2005; instead of acquiring WBPG directly along with it, the company instead began to operate the station under a local marketing agreement. Just over seven months later, on July 7, 2006, LIN purchased WBPG outright. On January 24, 2006, CBS Corporation and Time Warner announced the shutdown of both UPN and The WB effective that fall. In place of these two networks, a new "fifth" network—"The CW Television Network" (its name representing the first initials of parent companies CBS and Warner Bros.), jointly owned by both companies, would launch, with a lineup primarily featuring the most popular programs from both networks. WBPG joined The CW on Se
https://en.wikipedia.org/wiki/Dihedral%20symmetry%20in%20three%20dimensions
In geometry, dihedral symmetry in three dimensions is one of three infinite sequences of point groups in three dimensions which have a symmetry group that as an abstract group is a dihedral group Dihn (for n ≥ 2). Types There are 3 types of dihedral symmetry in three dimensions, each shown below in 3 notations: Schönflies notation, Coxeter notation, and orbifold notation. Chiral Dn, [n,2]+, (22n) of order 2n – dihedral symmetry or para-n-gonal group (abstract group: Dihn). Achiral Dnh, [n,2], (*22n) of order 4n – prismatic symmetry or full ortho-n-gonal group (abstract group: Dihn × Z2). Dnd (or Dnv), [2n,2+], (2*n) of order 4n – antiprismatic symmetry or full gyro-n-gonal group (abstract group: Dih2n). For a given n, all three have n-fold rotational symmetry about one axis (rotation by an angle of 360°/n does not change the object), and 2-fold rotational symmetry about a perpendicular axis, hence about n of those. For n = ∞, they correspond to three Frieze groups. Schönflies notation is used, with Coxeter notation in brackets, and orbifold notation in parentheses. The term horizontal (h) is used with respect to a vertical axis of rotation. In 2D, the symmetry group Dn includes reflections in lines. When the 2D plane is embedded horizontally in a 3D space, such a reflection can either be viewed as the restriction to that plane of a reflection through a vertical plane, or as the restriction to the plane of a rotation about the reflection line, by 180°. In 3D, the two operations are distinguished: the group Dn contains rotations only, not reflections. The other group is pyramidal symmetry Cnv of the same order, 2n. With reflection symmetry in a plane perpendicular to the n-fold rotation axis, we have Dnh, [n], (*22n). Dnd (or Dnv), [2n,2+], (2*n) has vertical mirror planes between the horizontal rotation axes, not through them. As a result, the vertical axis is a 2n-fold rotoreflection axis. Dnh is the symmetry group for a regular n-sided prism and also for a
https://en.wikipedia.org/wiki/Contact%20order
The contact order of a protein is a measure of the locality of the inter-amino acid contacts in the protein's native state tertiary structure. It is calculated as the average sequence distance between residues that form native contacts in the folded protein divided by the total length of the protein. Higher contact orders indicate longer folding times, and low contact order has been suggested as a predictor of potential downhill folding, or protein folding that occurs without a free energy barrier. This effect is thought to be due to the lower loss of conformational entropy associated with the formation of local as opposed to nonlocal contacts. Relative contact order (CO) is formally defined as: where N is the total number of contacts, ΔSi,j is the sequence separation, in residues, between contacting residues i and j, and L is the total number of residues in the protein. The value of contact order typically ranges from 5% to 25% for single-domain proteins, with lower contact order belonging to mainly helical proteins, and higher contact order belonging to proteins with a high beta-sheet content. Protein structure prediction methods are more accurate in predicting the structures of proteins with low contact orders. This may be partly because low contact order proteins tend to be small, but is likely to be explained by the smaller number of possible long-range residue-residue interactions to be considered during global optimization procedures that minimize an energy function. Even successful structure prediction methods such as the Rosetta method overproduce low-contact-order structure predictions compared to the distributions observed in experimentally determined protein structures. The percentage of the natively folded contact order can also be used as a measure of the "nativeness" of folding transition states. Phi value analysis in concert with molecular dynamics has produced transition-state models whose contact order is close to that of the folded state in pr
https://en.wikipedia.org/wiki/History%20of%20alternative%20medicine
The history of alternative medicine refers to the history of a group of diverse medical practices that were collectively promoted as "alternative medicine" beginning in the 1970s, to the collection of individual histories of members of that group, or to the history of western medical practices that were labeled "irregular practices" by the western medical establishment. It includes the histories of complementary medicine and of integrative medicine. "Alternative medicine" is a loosely defined and very diverse set of products, practices, and theories that are perceived by its users to have the healing effects of medicine, but do not originate from evidence gathered using the scientific method, are not part of biomedicine, or are contradicted by scientific evidence or established science. "Biomedicine" is that part of medical science that applies principles of anatomy, physics, chemistry, biology, physiology, and other natural sciences to clinical practice, using scientific methods to establish the effectiveness of that practice. Much of what is now categorized as alternative medicine was developed as independent, complete medical systems, was developed long before biomedicine and use of scientific methods, and was developed in relatively isolated regions of the world where there was little or no medical contact with pre-scientific western medicine, or with each other's systems. Examples are Traditional Chinese medicine, European humoral theory and the Ayurvedic medicine of India. Other alternative medicine practices, such as homeopathy, were developed in western Europe and in opposition to western medicine, at a time when western medicine was based on unscientific theories that were dogmatically imposed by western religious authorities. Homeopathy was developed prior to discovery of the basic principles of chemistry, which proved homeopathic remedies contained nothing but water. But homeopathy, with its remedies made of water, was harmless compared to the unscientif
https://en.wikipedia.org/wiki/Joel%20Lebowitz
Joel Louis Lebowitz (born May 10, 1930) is a mathematical physicist widely acknowledged for his outstanding contributions to statistical physics, statistical mechanics and many other fields of Mathematics and Physics. Lebowitz has published more than five hundred papers concerning statistical physics and science in general, and he is one of the founders and editors of the Journal of Statistical Physics, one of the most important peer-reviewed journals concerning scientific research in this area. He has been president of the New York Academy of Sciences. Lebowitz is the George William Hill Professor of Mathematics and Physics at Rutgers University. He is also an active member of the human rights community and a long-term co-chair of the Committee of Concerned Scientists. Biography Lebowitz was born in Taceva, then in Czechoslovakia, now Ukraine, in 1930 into a Jewish family. During World War II he was deported with his family to Auschwitz, where his father, his mother, and his younger sister were killed in 1944. After being liberated from the camp, he moved to United States by boat, and he studied in an Orthodox Jewish school and Brooklyn College. He earned his PhD at Syracuse University in 1956 under the supervision of Peter G. Bergmann. Then he continued his research with Lars Onsager, at Yale University, where he got a faculty position. He moved to the Stevens Institute of Technology in 1957 and to the Belfer Graduate School of Science of Yeshiva University in 1959. Finally he got a faculty position at Rutgers University in 1977, where he holds the prestigious George William Hill Professor position. During his years at the Yeshiva University and Rutgers University he has been in contact with several scientists, and artists, like Fumio Yoshimura and Kate Millett. In 1975 he founded the Journal of Statistical Physics. In 1979 he was president of the New York Academy of Sciences. He has been one of the most active supporters of dissident scientists in the forme
https://en.wikipedia.org/wiki/Maxwell%27s%20demon
Maxwell's demon is a thought experiment that would hypothetically violate the second law of thermodynamics. It was proposed by the physicist James Clerk Maxwell in 1867. In his first letter, Maxwell referred to the entity as a "finite being" or a "being who can play a game of skill with the molecules". Lord Kelvin would later call it a "demon". In the thought experiment, a demon controls a small massless door between two chambers of gas. As individual gas molecules (or atoms) approach the door, the demon quickly opens and closes the door to allow only fast-moving molecules to pass through in one direction, and only slow-moving molecules to pass through in the other. Because the kinetic temperature of a gas depends on the velocities of its constituent molecules, the demon's actions cause one chamber to warm up and the other to cool down. This would decrease the total entropy of the system, without applying any work, thereby violating the second law of thermodynamics. The concept of Maxwell's demon has provoked substantial debate in the philosophy of science and theoretical physics, which continues to the present day. It stimulated work on the relationship between thermodynamics and information theory. Most scientists argue that, on theoretical grounds, no practical device can violate the second law in this way. Other researchers have implemented forms of Maxwell's demon in experiments, though they all differ from the thought experiment to some extent and none have been shown to violate the second law. Origin and history of the idea The thought experiment first appeared in a letter Maxwell wrote to Peter Guthrie Tait on 11 December 1867. It appeared again in a letter to John William Strutt in 1871, before it was presented to the public in Maxwell's 1872 book on thermodynamics titled Theory of Heat. In his letters and books, Maxwell described the agent opening the door between the chambers as a "finite being". William Thomson (Lord Kelvin) was the first to use th
https://en.wikipedia.org/wiki/Capsule%20%28geometry%29
A capsule (from Latin capsula, "small box or chest"), or stadium of revolution, is a basic three-dimensional geometric shape consisting of a cylinder with hemispherical ends. Another name for this shape is spherocylinder. It can also be referred to as an oval although the sides (either vertical or horizontal) are straight parallel. Usages The shape is used for some objects like containers for pressurised gases, windows of places like a jet, software buttons, building domes (like the U.S. Capitol, having the windows of the top hat that depict The Apotheosis of Washington inside designed with the appearance of the shape & placed in an omnidirectional pattern), mirrors, and pharmaceutical capsules. In chemistry and physics, this shape is used as a basic model for non-spherical particles. It appears, in particular as a model for the molecules in liquid crystals or for the particles in granular matter. Formulas The volume of a capsule is calculated by adding the volume of a ball of radius (that accounts for the two hemispheres) to the volume of the cylindrical part. Hence, if the cylinder has height , . The surface area of a capsule of radius whose cylinder part has height is . Generalization A capsule can be equivalently described as the Minkowski sum of a ball of radius with a line segment of length . By this description, capsules can be straightforwardly generalized as Minkowski sums of a ball with a polyhedron. The resulting shape is called a spheropolyhedron. Related shapes A capsule is the three-dimensional shape obtained by revolving the two-dimensional stadium around the line of symmetry that bisects the semicircles.
https://en.wikipedia.org/wiki/Vaccine%20resistance
Vaccine resistance is the evolutionary adaptation of pathogens to infect and spread through vaccinated individuals, analogous to antimicrobial resistance. It concerns both human and animal vaccines. Although the emergence of a number of vaccine resistant pathogens has been well documented, this phenomenon is nevertheless much more rare and less of a concern than antimicrobial resistance. Vaccine resistance may be considered a special case of immune evasion, from the immunity conferred by the vaccine. Since the immunity conferred by a vaccine may be different from that induced by infection by the pathogen, the immune evasion may also be easier (in case of an inefficient vaccine) or more difficult (would be the case of the universal flu vaccine). We speak of vaccine resistance only if the immune evasion is a result of evolutionary adaptation of the pathogen (and not a feature of the pathogen that it had before any evolutionary adaptation to the vaccine) and the adaptation is driven by the selective pressure induced by the vaccine (this would not be the case of an immune evasion that is the result of genetic drift that would be present even without vaccinating the population). Some of the causes advanced for less frequent emergence of resistance are that vaccines are mostly used for prophylaxis, that is before infection occurs, and usually act to suppress the pathogen before the host becomes infectious most vaccines target multiple antigenic sites of the pathogen different hosts may produce different immune responses to the same pathogen For diseases that confer long lasting immunity after exposure, typically childhood diseases, it was argued that a vaccine may provide the same immune response as natural infection, so it is expected that there should be no vaccine resistance. If vaccine resistance emerges the vaccine may retain some level of protection against serious infection, possibly by modifying the immune response of the host away from immunopathology. T
https://en.wikipedia.org/wiki/Morchella%20angusticeps
Morchella angusticeps is a species of fungus in the family Morchellaceae native to eastern North America. Described by Charles Horton Peck in 1879, the name M. angusticeps was clarified in 2012 prior to which this species may have been referred to as either M. angusticeps or M. elata. M. angusticeps is one of the black morels, and is found in eastern North America, where it occurs in association with various hardwoods in the spring. A similar, although smaller, black morel occurs in northeastern North America, M. septentrionalis.
https://en.wikipedia.org/wiki/Ulpian%27s%20life%20table
Ulpian's life table is an ancient Roman annuities table. It is known through a passage, originating from the jurist Aemilius Macer, preserved in edited form in Justinian's Digest. The table appears to provide a rough outline of ancient Roman life expectancy. Although it is not clear what population the table refers to, or how its data was gathered, Richard Duncan-Jones has suggested that it refers to slaves and ex-slaves, who were often the object of testamentary maintenance grants. Aemilius Macer probably lived in the 230s AD. He records the table in his systematic commentary on the lex Julia de vicesima hereditatium, an Augustan law of 6 AD that put a 5 percent tax on inheritances. Despite its many numbers, the fragment does not appear to be afflicted by any serious textual corruption. Table Macer's text provides two figures: a forma, or schedule, presented by Ulpian (d. 223), and a customary (solitum est) schedule that antedates Ulpian's. The forma is described as a means of calculating tax for alimenta and usufructs. The age of the legatee is checked against the table; the figure recorded on the table is multiplied by annuity's annual value. Five percent of this last figure is what is owed in tax. Ulpian's life table gives figures broadly consistent with the Coale–Demeny Model West life table: female life expectancy at birth is 22.5 years, male life expectancy is 20.4. Its mortality figures are thus higher than those of most models, though the statistical flaws in the evidence itself has encouraged interpretative caution. Although, among moderns, "life expectancy" tends to mean "the average number of years lived after age x", the table figures probably represent median life expectancy (i.e., the number of years elapsed before half the selected population is dead). After childhood, the two figures are quite close, but childhood mortality causes the figures for the first years of life to diverge. The figures given are too high to represent the predicted market
https://en.wikipedia.org/wiki/Plant%E2%80%93soil%20feedback
Plant–soil feedback is a process where plants alter the biotic and abiotic qualities of soil they grow in, which then alters the ability of plants to grow in that soil in the future. Negative plant–soil feedback occurs when plants are less able to grow in soil that was previously occupied by a member of the same species, and positive plant–soil feedback occurs when plants are more able to grow in soil that was previously occupied by a member of the same species. Although it was originally assumed that negative plant–soil feedback was caused by plants depleting the soil of nutrients, recent work has suggested that a major cause of plant–soil feedback is a buildup of soil-borne pathogens. Mutualism and allelopathy are also thought to cause plant–soil feedback. Studies have shown that, on average, plant–soil feedback tends to be negative; however, there have been many notable exceptions, such as many invasive species. Negative plant–soil feedback is thought to be an important factor in helping plants to coexist. If a plant is over-abundant, then soil pathogens and other negative factors will become common, hurting its growth. Similarly, if a plant becomes overly rare, then so too will its soil pathogens and other negative factors, helping its growth. This negative feedback will help populations to stay in the community. Negative plant–soil feedback has been called a particular case of the Janzen–Connell hypothesis. Measuring plant–soil feedback Plant–soil feedback is best measured using Bever's interaction coefficient, Is. This value quantifies how much each plant's growth is limited by its own soil community compared to how much it limits others. It is for two-species comparisons. To measure this quantity, one must measure the growth of two plants, both in soil conditioned by members of their own species (Gx(home) for species x), and in soil conditioned by members of the other species (Gx(away) for plant species x). Then, the interaction coefficient is c
https://en.wikipedia.org/wiki/Ectodomain
An ectodomain is the domain of a membrane protein that extends into the extracellular space (the space outside a cell). Ectodomains are usually the parts of proteins that initiate contact with surfaces, which leads to signal transduction. A notable example of an ectodomain is the S protein, commonly known as the spike protein, of the viral particle responsible for the COVID-19 pandemic. The ectodomain region of the spike protein (S) is essential for attachment and eventual entry of the viral protein into the host cell. Ectodomains play a crucial part in the signaling pathways of viruses. Recent findings have indicated that certain antibodies including the anti-receptor binding domain (anti-RBD) or anti-spike ectodomain (anti-ECD) IgG titers can act as virus neutralization titers (VN titers) which can be identified in individuals with diseases, dyspnea and hospitalizations. In perspective of severe acute respiratory syndrome corona virus 2 (SARS-Cov-2) these specific ectodomains may detect antibody efficacy against SARS-Cov-2, in which VN titers can classify eligible plasma donors. Protective measures against diseases and respiratory conditions can further be advanced through ongoing research on ectodomains. Ectodomain's play a crucial part in the signaling pathways of viruses. In perspective of severe acute respiratory syndrome corona virus 2 (SARS-Cov-2) these specific ectodomains may detect antibody efficacy against SARS-Cov-2, in which VN titers can classify eligible plasma donors. Protective measures against diseases and respiratory conditions can further be advanced through ongoing research on ectodomains. Ectodomains also interact with membrane systems inducing vesicle aggregation, lipid mixing and liposome leakage which provides information as to how certain viruses spread infection throughout the cellular domain. Specifically, the hepatitis C virus (HCV) utilize a fusion process in which the ectodomain of HCV E2 envelope protein confers fusogenic prope
https://en.wikipedia.org/wiki/Hanan%20Mohamed%20Abdelrahman
Hanan Mohamed Abdelrahman is an Egyptian-Norwegian mathematics educator, the 2017 winner of the . Life and career Hanan Mohamed Abdelrahman is originally from Egypt, where her father was an engineer and her mother was a banker. She came to Norway with her husband, a Sudanese immigrant to Norway, in 2002. She has a master's degree in mathematics education, and is a teacher at , a secondary school near Oslo. Contributions Abdelrahman's school has many immigrant students of varying backgrounds. Abdelrahman provides them with mathematical games, puzzles, and videos as well as the more traditional coursework of mathematics. She takes the point of view that with individual attention all her students should be able to learn mathematics, and that all should be equally challenged. Abdelrahman is the author of a Norwegian-language book, Mattehjelperen - Leksehjelp for foreldre og elever på ungdomsskolen, aimed at getting parents to help students with mathematics, and the founder of an online web site for assisting students with mathematics. She is also a member of a national committee to investigate the effects of gender differences on school outcomes. Prize The Holmboe Prize, which Abdelrahman won in 2017, is an annual award for the top mathematics teachers at the primary and secondary school level in Norway, given by the Norwegian Mathematics Council. It is presented in conjunction with the Abel Prize, whose 2017 winner was Yves Meyer. The Holmboe Prize is named after Norwegian mathematics teacher Bernt Michael Holmboe, the teacher of Niels Henrik Abel. It offers a prize of 100,000 Norwegian krone to be split between its winner and the winner's school.
https://en.wikipedia.org/wiki/Muometric%20navigation
Muometric navigation is positioning, navigation and timing using cosmic ray muons and other cosmic particles. It is possible to determine locations with GNSS satellites with well-known positions and time. GNSS is often used by critically important governmental organizations for navigating ships and planes, but the signals can be easily jammed and spoofed. In 2020 Hiroyuki K.M. Tanaka created an entirely new approach from GNSS that locates the receiver's position with cosmic-ray muons. Muometric techniques include the muometric positioning system (muPS), the muometric wireless navigation system (MuWNS) or muPS Wireless Navigation System (muWNS), cosmic time synchronizer (CTS) and cosmic time calibrator (CTC). Positioning and navigation The muometric positioning and navigation techniques are based on the time-of-flight of relativistic cosmic-ray muons between reference detectors and the receiver detector usually located indoor, underground, or underwater. Instead of receiving a GNSS signal, they detect cosmic-ray muons. Three or more reference detectors are deployed with known positions and time-references. Like GNSS, clocks between the reference receivers and the receiver must be well-synchronized. Unlike GNSS, this technology enables navigation in Arctic areas where GNSS satellite access is limited due to orbital constraints of these satellites. The initial prototype required wiring between the receiver and each reference detector for accurate time synchronization. However, this configuration restricted the range of applicability of the system. Efforts to find a way to navigate without wires, growing out of the success of this initial system replaced wires with a clock. muWNS is expected to be applied to rescue teams, for example, to guide robots underwater and underground by positioning inside tunnels, in a building or mine collapse. The indoor muometric positioning accuracy is 3.9 cm as of 2023. Timing Precise timekeeping generally requires GNSS and atomic
https://en.wikipedia.org/wiki/Dinosauria%20Park
Dinosauria Park (, O Kosmos ton Dinosauron) is a dinosaur park located near the town of Gournes in Crete, Greece, east of the city of Heraklion. The park features some fossil replicas and several animatronic dinosaur models. Dinosaur models are displayed along an approximately long walk, each accompanied by an information board. The Dinosauria Park started as an idea in 2013 and was first housed in Gournes Δekk, located in the former American base. It displayed 30 different dinosaur species from all periods of the Mesozoic era and other prehistoric animals. In 2014 the exhibits were transferred to their own space, in which they were displayed in what would have been their natural environment. New species of dinosaurs were added, along with new sections, such as a "dinosaur hospital", depicting the hatching of a dinosaur egg. A "5D" cinema screen shows films to visitors. See also List of dinosaur parks Cretaquarium External links Dinosauria Park Miniature parks Heraklion (regional unit) Dinosaurs in amusement parks
https://en.wikipedia.org/wiki/NAPA%20%28gene%29
N-ethylmaleimide-sensitive factor Attachment Protein Alpha, also known as SNAP-α, is a SNAP protein that is involved in the intra-cellular trafficking and fusing of vesicles to target membranes in cells. Function The 'SNARE hypothesis' is a model explaining the process of docking and fusion of vesicles to their target membranes. According to this model, membrane proteins from the vesicle (v-SNAREs) and proteins from the target membrane (t-SNAREs) govern the specificity of vesicle targeting and docking through mutual recognition. Once the 2 classes of SNAREs bind to each other, they form a complex that recruits the general elements of the fusion apparatus, namely NSF (N-ethylmaleimide-sensitive factor) and SNAPs (soluble NSF-attachment proteins), to the site of membrane fusion, thereby forming the 20S fusion complex. Alpha- and gamma-SNAP are found in a wide range of tissues and act synergistically in intra-Golgi transport. The sequence of the predicted 295-amino acid human protein encoded by NAPA shares 37%, 60%, and 67% identity with the sequences of yeast, Drosophila, and squid alpha-SNAP, respectively. Platelets contain some of the same proteins, including NSF, p115/TAP, alpha-SNAP (this protein), gamma-SNAP, and the t-SNAREs syntaxin-2 and syntaxin-4, that are used in many vesicular transport processes in other cell types. Platelet exocytosis uses a molecular mechanism similar to that used by other secretory cells, such as neurons, although the proteins used by the platelet and their modes of regulation may be quite different. Clinical significance NAPA is abnormally expressed in fetuses of both IVF and ICSI, which may contribute to the increased risk of birth defects in these methods of assisted reproductive technology (ART). Interactions NAPA has been shown to interact with: NSF, SNAP23, STX1A, STX4, STX5. ORAI1, STIM1.
https://en.wikipedia.org/wiki/Ball%20lightning
Ball lightning is a rare and unexplained phenomenon described as luminescent, spherical objects that vary from pea-sized to several meters in diameter. Though usually associated with thunderstorms, the observed phenomenon is reported to last considerably longer than the split-second flash of a lightning bolt, and is a phenomenon distinct from St. Elmo's fire. Some 19th-century reports describe balls that eventually explode and leave behind an odor of sulfur. Descriptions of ball lightning appear in a variety of accounts over the centuries and have received attention from scientists. An optical spectrum of what appears to have been a ball lightning event was published in January 2014 and included a video at high frame rate. Laboratory experiments have produced effects that are visually similar to reports of ball lightning, but how these relate to the supposed phenomenon remains unclear. Scientists have proposed a number of hypotheses to explain reports of ball lightning over the centuries, but scientific data on ball lightning remain scarce. The presumption of its existence has depended on reported public sightings, which have produced inconsistent findings. Owing to the lack of reproducible data, the existence of ball lightning as a distinct physical phenomenon remains unproven. Historical accounts Ball lightning is a possible source of legends that describe luminous balls, such as the mythological Anchimayen from Argentinean and Chilean Mapuche culture. According to statistical investigations in 1960, ball lightning had been seen by 5% of the population of the Earth. Another study analyzed reports of 10,000 cases. Gervase of Canterbury The chronicle of Gervase of Canterbury, an English monk, contains what is possibly the earliest known reference to ball lightning, dated 7 June 1195. He states, "A marvellous sign descended near London", consisting of a dense and dark cloud, emitting a white substance that grew into a spherical shape under the cloud, from whic
https://en.wikipedia.org/wiki/Ambiguity
Ambiguity is the type of meaning in which a phrase, statement, or resolution is not explicitly defined, making several interpretations plausible. A common aspect of ambiguity is uncertainty. It is thus an attribute of any idea or statement whose intended meaning cannot be definitively resolved, according to a rule or process with a finite number of steps. (The prefix ambi- reflects the idea of "two," as in "two meanings.") The concept of ambiguity is generally contrasted with vagueness. In ambiguity, specific and distinct interpretations are permitted (although some may not be immediately obvious), whereas with vague information it is difficult to form any interpretation at the desired level of specificity. Linguistic forms Lexical ambiguity is contrasted with semantic ambiguity. The former represents a choice between a finite number of known and meaningful context-dependent interpretations. The latter represents a choice between any number of possible interpretations, none of which may have a standard agreed-upon meaning. This form of ambiguity is closely related to vagueness. Ambiguity in human language is argued to reflect principles of efficient communication. Languages that communicate efficiently will avoid sending information that is redundant with information provided in the context. This can be shown mathematically to result in a system which is ambiguous when context is neglected. In this way, ambiguity is viewed as a generally useful feature of a linguistic system. Linguistic ambiguity can be a problem in law, because the interpretation of written documents and oral agreements is often of paramount importance. Lexical ambiguity The lexical ambiguity of a word or phrase applies to it having more than one meaning in the language to which the word belongs. "Meaning" here refers to whatever should be represented by a good dictionary. For instance, the word "bank" has several distinct lexical definitions, including "financial institution" and "edge of
https://en.wikipedia.org/wiki/Newtonian%20limit
In physics, the Newtonian limit is a mathematical approximation applicable to physical systems exhibiting (1) weak gravitation, (2) objects moving slowly compared to the speed of light, and (3) slowly changing (or completely static) gravitational fields. Under these conditions, Newton's law of universal gravitation may be used to obtain values that are accurate. In general, and in the presence of significant gravitation, the general theory of relativity must be used. In the Newtonian limit, spacetime is approximately flat and the Minkowski metric may be used over finite distances. In this case 'approximately flat' is defined as space in which gravitational effect approaches 0, mathematically actual spacetime and Minkowski space are not identical, Minkowski space is an idealized model. Special relativity In special relativity, Newtonian behaviour can in most cases be obtained by performing the limit . In this limit, the often appearing gamma factor becomes 1 and the Lorentz transformations between reference frames turn into Galileo transformations General relativity The geodesic equation for a free particle on curved spacetime with metric can be derived from the action If the spacetime-metric is then, ignoring all contributions of order the action becomes which is the action that reproduces the Newtonian equations of motion of a particle in a gravitational potential See also Classical limit
https://en.wikipedia.org/wiki/List%20of%20emulators
This article lists software emulators. Central processing units ARM ARMulator Aemulor QEMU MIPS SPIM: The OVPsim 500 mips MIPS32 emulator, can be used to develop software using virtual platforms, emulators including MIPS processors running at up to 500 MIPS for MIPS32 processors running many OSes including Linux. OVP is used to build emulators of single MIPS processors or multiple - homogeneous MP or heterogenous MP. x86 architecture Bochs DOSBox FX!32 PCem QEMU – an opensource emulator that emulates 7 architectures including ARM, x86, MIPS, and others box86 Rosetta 2: Apple's emulator for macOS allowing to run x86_64 applications on arm64 platform Motorola 680x0 Mac 68K emulator: For PowerPC classic Mac OS PowerPC PearPC Rosetta: Apple's emulator for PowerPC processors, built into Mac OS X WarpUP: Amiga system for PowerPC expansion cards built into MorphOS and available for AmigaOS SheepShaver: Emulates the PowerPC processor. Can run Mac OS 7 to Mac OS 9. Computer system emulators Full system simulators Simics CPU Sim: A Java application that allows the user to design and create a simple architecture and instruction set and then run programs of instructions from the set through simulation GXemul: Framework for full-system computer architecture emulation Mobile phones and PDAs Palm OS Emulator Adobe Device Central BlueStacks Windows Subsystem for Android Blisk (browser) touchHLE Multi-system emulators blueMSX: Emulates Z80 based computers and consoles MAME: Emulates multiple arcade machines, video game consoles and computers DAPHNE is an arcade emulator application that emulates a variety of laserdisc video games with the intent of preserving these games and making the play experience as faithful to the originals as possible. The developer calls DAPHNE the "First Ever Multiple Arcade Laserdisc Emulator" ("FEMALE"). It derives its name from Princess Daphne, the heroine of Dragon's Lair. HYPSEUS is a modern SDL2 update to the DAPH
https://en.wikipedia.org/wiki/Circulating%20mitochondrial%20DNA
Circulating mitochondrial DNA, also called cell-free circulating mitochondrial DNA and circulating cell-free mitochondrial DNA (ccf mtDNA), are short sections of mitochondrial DNA (mtDNA) that are released by cells undergoing stress or other damaging or pathological events. Circulating mitochondrial DNA is recognized by the immune system and activates inflammatory reactions. It is also a biomarker that can be used to detect the degree of damage from myocardial infarctions, cancers and ordinary stress. In certain situations it acts as a hormone. Plasma or serum levels of ccf mtDNA have been seen to be of marked difference in people with cancer. These higher levels have shown to be a reliable non-invasive biomarker in the diagnosis and prognosis of many kinds of tumours. Specific analysis of tumor-derived circulating mitochondrial DNA is challenging in human samples as it requires to track in plasma defined mutations, or alterations from the mitochondrial genome. In animal models, separating tumor-derived DNA in plasma from non-tumor derived DNA is easier. A proof of principle demonstrated the sensitivity of detecting tumor-derived circulating mitochondrial DNA in the plasma, CSF and urine of xenografted animals. See also Circulating tumor DNA NucPosDB: a database of nucleosome positioning in vivo and nucleosomics of cell-free DNA
https://en.wikipedia.org/wiki/Floer%20homology
In mathematics, Floer homology is a tool for studying symplectic geometry and low-dimensional topology. Floer homology is a novel invariant that arises as an infinite-dimensional analogue of finite-dimensional Morse homology. Andreas Floer introduced the first version of Floer homology, now called Lagrangian Floer homology, in his proof of the Arnold conjecture in symplectic geometry. Floer also developed a closely related theory for Lagrangian submanifolds of a symplectic manifold. A third construction, also due to Floer, associates homology groups to closed three-dimensional manifolds using the Yang–Mills functional. These constructions and their descendants play a fundamental role in current investigations into the topology of symplectic and contact manifolds as well as (smooth) three- and four-dimensional manifolds. Floer homology is typically defined by associating to the object of interest an infinite-dimensional manifold and a real valued function on it. In the symplectic version, this is the free loop space of a symplectic manifold with the symplectic action functional. For the (instanton) version for three-manifolds, it is the space of SU(2)-connections on a three-dimensional manifold with the Chern–Simons functional. Loosely speaking, Floer homology is the Morse homology of the function on the infinite-dimensional manifold. A Floer chain complex is formed from the abelian group spanned by the critical points of the function (or possibly certain collections of critical points). The differential of the chain complex is defined by counting the function's gradient flow lines connecting certain pairs of critical points (or collections thereof). Floer homology is the homology of this chain complex. The gradient flow line equation, in a situation where Floer's ideas can be successfully applied, is typically a geometrically meaningful and analytically tractable equation. For symplectic Floer homology, the gradient flow equation for a path in the loop
https://en.wikipedia.org/wiki/Uroscopy
Uroscopy is the historical medical practice of visually examining a patient's urine for pus, blood, or other symptoms of disease. The first records of uroscopy as a method for determining symptoms of an illness date back to the 4th millennium BC, and became common practice in Classical Greece. Later reaching medical predominance during the Byzantine Era & High Middle Ages, the practice eventually was replaced with more accurate methods during the early modern period, with uroscopy being considered inadequate due to the lack of empirical evidence and higher standards of post-Renaissance medicine. In modern medicine, visual examination of a patient's urine may provide preliminary evidence for a diagnosis, but is generally limited to conditions that specifically affect the urinary system such as urinary tract infections, kidney and bladder issues, and liver failure. History Records of urinalysis for uroscopy date back as far as 4000 BC, originating with Babylonian and Sumerian physicians. At the outset of the 4th century BC Greek physician Hippocrates hypothesized that urine was a "filtrate" of the four humors, and limited possible the diagnoses resulting from this method to issues dealing with the bladder, kidneys, and urethra. This in turn led another Greek physician, Galen, to refine the idea down to urine being a filtrate of only blood, and not of black bile, yellow bile, or phlegm. Byzantine medicine followed, though it maintained its roots from Greco-Roman antiquity, and continued the application and study of uroscopy – it eventually becoming the primary form of ailment diagnosis. Byzantine physicians created some of the foundational codifications of uroscopy, with the most well known example being a 7th-century guide on uroscopic methods: Theophilus Protospatharius's On Urines. The work, along with others, became widely popular and accelerated the rate at which uroscopy spread throughout the Mediterranean. Over time these Byzantine works inspired further in
https://en.wikipedia.org/wiki/Midbrain
The midbrain or mesencephalon is the rostral-most portion of the brainstem connecting the diencephalon and cerebrum with the pons. It consists of the cerebral peduncles, tegmentum, and tectum. It is functionally associated with vision, hearing, motor control, sleep and wakefulness, arousal (alertness), and temperature regulation. The name comes from the Greek mesos, "middle", and enkephalos, "brain". Anatomy The midbrain is the shortest segment of the brainstem, measuring at less than 2cm in length. It is situated mostly in the posterior cranial fossa, with its superior part extending above the tentorial notch. Structure The principal regions of the midbrain are the tectum, the cerebral aqueduct, tegmentum, and the cerebral peduncles. Rostrally the midbrain adjoins the diencephalon (thalamus, hypothalamus, etc.), while caudally it adjoins the hindbrain (pons, medulla and cerebellum). In the rostral direction, the midbrain noticeably splays laterally. Sectioning of the midbrain is usually performed axially, at one of two levels – that of the superior colliculi, or that of the inferior colliculi. One common technique for remembering the structures of the midbrain involves visualizing these cross-sections (especially at the level of the superior colliculi) as the upside-down face of a bear, with the cerebral peduncles forming the ears, the cerebral aqueduct the mouth, and the tectum the chin; prominent features of the tegmentum form the eyes and certain sculptural shadows of the face. Tectum The tectum (Latin for roof) is the part of the midbrain dorsal to the cerebral aqueduct. The position of the tectum is contrasted with the tegmentum, which refers to the region in front of the ventricular system, or floor of the midbrain. It is involved in certain reflexes in response to visual or auditory stimuli. The reticulospinal tract, which exerts some control over alertness, takes input from the tectum, and travels both rostrally and caudally f
https://en.wikipedia.org/wiki/Line%20of%20action
In physics, the line of action (also called line of application) of a force () is a geometric representation of how the force is applied. It is the straight line through the point at which the force is applied in the same direction as the vector . The concept is essential, for instance, for understanding the net effect of multiple forces applied to a body. For example, if two forces of equal magnitude act upon a rigid body along the same line of action but in opposite directions, they cancel and have no net effect. But if, instead, their lines of action are not identical, but merely parallel, then their effect is to create a moment on the body, which tends to rotate it. Calculation of torque For the simple geometry associated with the figure, there are three equivalent equations for the magnitude of the torque associated with a force directed at displacement from the axis whenever the force is perpendicular to the axis: where is the cross-product, is the component of perpendicular to , is the moment arm, and is the angle between and
https://en.wikipedia.org/wiki/Depth%20of%20focus
Depth of focus is a lens optics concept that measures the tolerance of placement of the image plane (the film plane in a camera) in relation to the lens. In a camera, depth of focus indicates the tolerance of the film's displacement within the camera and is therefore sometimes referred to as "lens-to-film tolerance". Depth of focus versus depth of field The phrase depth of focus is sometimes erroneously used to refer to the depth of field (DOF), which is the area in front of the lens in acceptable focus, whereas the true meaning of depth of focus refers to the zone behind the lens wherein the film plane or sensor is placed to produce an in-focus image. Depth of focus can have two slightly different meanings. The first is the distance over which the image plane can be displaced while a single object plane remains in acceptably sharp focus; the second is the image-side conjugate of depth of field. With the first meaning, the depth of focus is symmetrical about the image plane; with the second, the depth of focus is greater on the far side of the image plane, though in most cases the distances are approximately equal. Where depth of field often can be measured in macroscopic units such as meters and feet, depth of focus is typically measured in microscopic units such as fractions of a millimeter or thousandths of an inch. The same factors that determine depth of field also determine depth of focus, but these factors can have different effects than they have in depth of field. Both depth of field and depth of focus increase with smaller apertures. For distant subjects (beyond macro range), depth of focus is relatively insensitive to focal length and subject distance, for a fixed f-number. In the macro region, depth of focus increases with longer focal length or closer subject distance, while depth of field decreases. Determining factors In small-format cameras, the smaller circle of confusion limit yields a proportionately smaller depth of focus. In motion-picture
https://en.wikipedia.org/wiki/15th%20meridian%20east
The meridian 15° east of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Europe, Africa, the Atlantic Ocean, the Southern Ocean, and Antarctica to the South Pole. The 15th meridian east forms a great circle with the 165th meridian west. The meridian is the central axis of time zones with the UTC+01:00 offset, including Central European Time and West Africa Time. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 15th meridian east passes through: {| class="wikitable plainrowheaders" ! scope="col" width="125" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | ! scope="row" | | Island of Spitsbergen, Svalbard |-valign="top" | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Atlantic Ocean | style="background:#b0e0e6;" | Greenland Sea Norwegian Sea |- | ! scope="row" | | Islands of Langøya, Austvågøy and Hinnøya |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Vestfjorden | style="background:#b0e0e6;" | |- | ! scope="row" | | Island of Engeløya and the mainland |- | ! scope="row" | | for about 1111 km |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Baltic Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | Island of Bornholm, for about 21 km. There is a small memorial at |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Baltic Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | |- | ! scope="row" | | For about 17 km between Rothenburg and Görlitz. |- | ! scope="row" | | For 1500 m in Zgorzelec. |- | ! scope="row" | | For 300 m in Görlitz. |- | ! scope="row" | | |- | ! scope="row" | | For about 3 km |- | ! scope="row" | | For about 6 km |-valign="top" | ! scope="row" | | A mark
https://en.wikipedia.org/wiki/Uriel%20Frisch
Uriel Frisch (born in Agen, in France, on December 10, 1940) is a French mathematical physicist known for his work on fluid dynamics and turbulence. Biography From 1959 to 1963 Frisch was a student at the École Normale Supérieure. Early in his graduate studies, he became interested in turbulence, under the mentorship of Robert Kraichnan, a former assistant to Albert Einstein. Frisch earned a Ph.D. in 1967 from the University of Paris, and since then he has worked at the French National Centre for Scientific Research (CNRS). He retired in 2006, and became a director of research emeritus at CNRS. Frisch's wife Hélène is also a physicist, and the grand daughter of mathematician Paul Lévy. Research Frisch is the author of a 1995 book on turbulence and of over 200 research publications. One of his most cited works, published in 1986, concerns the lattice gas automaton method of simulating fluid dynamics using a cellular automaton. The method used until that time, the HPP model, simulated particles moving in axis-parallel directions in a square lattice, but this model was unsatisfactory because it obeyed unwanted and unphysical conservation laws (the conservation of momentum within each axis-parallel line). Frisch and his co-authors Brosl Hasslacher and Yves Pomeau introduced a model using instead the hexagonal lattice which became known as the FHP model after the initials of its inventors and which much more accurately simulated the behavior of actual fluids. Frisch is also known for his work with Giorgio Parisi on the analysis of the fine structure of turbulent flows, for his early advocacy of multifractal systems in modeling physical processes, and for his research on using transportation theory to reconstruct the distribution of matter in the early universe. Awards and honors Frisch won the Peccot Prize of the Collège de France for his doctoral thesis in 1967, the Bazin Prize of the French Academy of Sciences in 1985, and the Lewis Fry Richardson Medal of the E
https://en.wikipedia.org/wiki/AMD%2010h
The AMD Family 10h, or K10, is a microprocessor microarchitecture by AMD based on the K8 microarchitecture. The first third-generation Opteron products for servers were launched on September 10, 2007, with the Phenom processors for desktops following and launching on November 11, 2007 as the immediate successors to the K8 series of processors (Athlon 64, Opteron, 64-bit Sempron). Nomenclature It appears that AMD has not used K-nomenclature (which originally stood for "Kryptonite" in the K5 processor) from the time after the use of the codename K8 for the AMD K8 or Athlon 64 processor family, since no K-nomenclature naming convention beyond K8 has appeared in official AMD documents and press releases after the beginning of 2005. The name "K8L" was first coined by Charlie Demerjian in 2005, at the time a writer at The Inquirer, and was used by the wider IT community as a convenient shorthand while according to AMD official documents, the processor family was termed "AMD Next Generation Processor Technology". The microarchitecture has also been referred to as Stars, as the codenames for desktop line of processors was named under stars or constellations (the initial Phenom models being codenamed Agena and Toliman). In a video interview, Giuseppe Amato confirmed that the codename is K10. It was revealed, by The Inquirer itself, that the codename "K8L" referred to a low-power version of the K8 family, later named Turion 64, and that K10 was the official codename for the microarchitecture. AMD refers to it as Family 10h Processors, as it is the successor of the Family 0Fh Processors (codename K8). 10h and 0Fh refer to the main result of the CPUID x86 processor instruction. In hexadecimal numbering, 0Fh (h represents hexadecimal numbering) equals the decimal number 15, and 10h equals decimal 16. (The "K10h" form that sometimes pops up is an improper hybrid of the "K" code and Family identifier number.) Schedule of launch and delivery Timeline Historical information
https://en.wikipedia.org/wiki/Beacon%20Pharmaceuticals
Beacon Pharmaceuticals PLC is a Bangladeshi pharmaceutical company that develops generic version of medications. Beacon manufactures more than 200 generic drugs and 65 oncology products. Beacon is the first company in Bangladesh to start export of cancer drugs. The company is exporting its products to Asia, Africa, Europe and Latin America. Beacon is public limited company listed in Dhaka & Chittagong stock exchange. About 2000 people are working in this company. Beacon has introduced a number of global first generics. The company's commercially available products include velpatasvir/sofosbuvir, sofosbuvir/daclatasvir, osimertinib, crizotinib, daclatasvir, sofosbuvir, afatinib, axitinib, brigatinib, baricitinib, cabozantinib, dasatinib, neratinib, eltrombopag, ibrutinib, lenvatinib, palbociclib, regorafenib, tofacitinib, and trelagliptin. BEACON Medicare Limited(BML) is the exclusive global marketing & distribution partner of Beacon Pharmaceuticals Limited.
https://en.wikipedia.org/wiki/Royal%20Entomological%20Society%20Handbooks
Handbooks for the Identification of British Insects is a series of books produced by the Royal Entomological Society (RES). The aim of the Handbooks is to provide illustrated identification keys to the insects of Britain, together with concise morphological, biological and distributional information. The series also includes several Check Lists of British Insects. All books contain line drawings, with the most recent volumes including colour photographs. In recent years, new volumes in the series have been published by Field Studies Council, and benefit from association with the AIDGAP identification guides and Synopses of the British Fauna. Full list of titles Vol : 1 - Small Orders Vol : 2 - Hemiptera Vol : 4 - Coleoptera Vol : 5 - Coleoptera Vol : 6 - Hymenoptera Vol : 7 - Hymenoptera: Ichneumonoidea Vol : 8 - Hymenoptera: Cynipoidea, Chalcidoidea & Proctotrupoidea Vol : 9 - Diptera: Nematocera & Brachycera Vol : 10 - Diptera: Cyclorrhapha Vol : 11 & 12 - Checklists of British Insects External links Handbooks - Royal Entomological Society Out of print handbooks - Royal Entomological Society (free to download) Entomological literature Handbooks Taxonomy (biology) books
https://en.wikipedia.org/wiki/Wait%20%28system%20call%29
In computer operating systems, a process (or task) may wait for another process to complete its execution. In most systems, a parent process can create an independently executing child process. The parent process may then issue a wait system call, which suspends the execution of the parent process while the child executes. When the child process terminates, it returns an exit status to the operating system, which is then returned to the waiting parent process. The parent process then resumes execution. Modern operating systems also provide system calls that allow a process's thread to create other threads and wait for them to terminate ("join" them) in a similar fashion. An operating system may provide variations of the wait call that allow a process to wait for any of its child processes to exit, or to wait for a single specific child process (identified by its process ID) to exit. Some operating systems issue a signal (SIGCHLD) to the parent process when a child process terminates, notifying the parent process and allowing it to retrieve the child process's exit status. The exit status returned by a child process typically indicates whether the process terminated normally or abnormally. For normal termination, this status also includes the exit code (usually an integer value) that the process returned to the system. During the first 20 years of UNIX, only the low 8 bits of the exit code have been available to the waiting parent. In 1989 with SVR4, a new call waitid has been introduced that returns all bits from the exit call in a structure called siginfo_t in the structure member si_status. Waitid is a mandatory part of the POSIX standard since 2001. Zombies and orphans When a child process terminates, it becomes a zombie process, and continues to exist as an entry in the system process table even though it is no longer an actively executing program. Under normal operation it will typically be immediately waited on by its parent, and then reaped by the syste
https://en.wikipedia.org/wiki/Astronomical%20symbols
Astronomical symbols are abstract pictorial symbols used to represent astronomical objects, theoretical constructs and observational events in European astronomy. The earliest forms of these symbols appear in Greek papyrus texts of late antiquity. The Byzantine codices in which many Greek papyrus texts were preserved continued and extended the inventory of astronomical symbols. New symbols have been invented to represent many planets and minor planets discovered in the 18th to the 21st centuries. These symbols were once commonly used by professional astronomers, amateur astronomers, alchemists, and astrologers. While they are still commonly used in almanacs and astrological publications, their occurrence in published research and texts on astronomy is relatively infrequent, with some exceptions such as the Sun and Earth symbols appearing in astronomical constants, and certain zodiacal signs used to represent the solstices and equinoxes. Unicode has encoded many of these symbols, mainly in the Miscellaneous Symbols, Miscellaneous Symbols and Arrows, Miscellaneous Symbols and Pictographs, and Alchemical Symbols blocks. Symbols for the Sun and Moon The use of astronomical symbols for the Sun and Moon dates to antiquity. The forms of the symbols that appear in the original papyrus texts of Greek horoscopes are a circle with one ray () for the Sun and a crescent for the Moon. The modern Sun symbol, a circle with a dot (☉), first appeared in Europe in the Renaissance. In modern academic writing, the Sun symbol is used for astronomical constants relating to the Sun. Teff☉ represents the solar effective temperature, and the luminosity, mass, and radius of stars are often represented using the corresponding solar constants (, , and , respectively) as units of measurement. Symbols for the planets Symbols for the classical planets appear in many medieval Byzantine codices in which many ancient horoscopes were preserved. The written symbols for Mercury, Venus, Jupit
https://en.wikipedia.org/wiki/Two-state%20quantum%20system
In quantum mechanics, a two-state system (also known as a two-level system) is a quantum system that can exist in any quantum superposition of two independent (physically distinguishable) quantum states. The Hilbert space describing such a system is two-dimensional. Therefore, a complete basis spanning the space will consist of two independent states. Any two-state system can also be seen as a qubit. Two-state systems are the simplest quantum systems that are of interest, since the dynamics of a one-state system is trivial (as there are no other states the system can exist in). The mathematical framework required for the analysis of two-state systems is that of linear differential equations and linear algebra of two-dimensional spaces. As a result, the dynamics of a two-state system can be solved analytically without any approximation. The generic behavior of the system is that the wavefunction's amplitude oscillates between the two states. A very well known example of a two-state system is the spin of a spin-1/2 particle such as an electron, whose spin can have values +ħ/2 or −ħ/2, where ħ is the reduced Planck constant. The two-state system cannot be used as a description of absorption or decay, because such processes require coupling to a continuum. Such processes would involve exponential decay of the amplitudes, but the solutions of the two-state system are oscillatory. Analytical solutions for stationary state energies and time-dependence Representation Supposing the two available basis states of the system are and , in general the state can be written as a superposition of these two states with probability amplitudes , Since the basis states are orthonormal, where and is the Kronecker delta, so . These two complex numbers may be considered coordinates in a two-dimensional complex Hilbert space. Thus the state vector corresponding to the state is and the basis states correspond to the basis vectors, and If the state is normalized, the norm
https://en.wikipedia.org/wiki/Trans-regulatory%20element
Trans-regulatory elements (TRE) are DNA sequences encoding upstream regulators (ie. trans-acting factors), which may modify or regulate the expression of distant genes. Trans-acting factors interact with cis-regulatory elements to regulate gene expression. TRE mediates expression profiles of a large number of genes via trans-acting factors. While TRE mutations affect gene expression, it is also one of the main driving factors for evolutionary divergence in gene expression. Trans vs cis elements Trans-regulatory elements work through an intermolecular interaction between two different molecules and so are said to be "acting in trans". For example (1) a transcribed and translated transcription factor protein derived from the trans-regulatory element; and a (2) DNA regulatory element that is adjacent to the regulated gene. This is in contrast to cis-regulatory elements that work through an intramolecular interaction between different parts of the same molecule: (1) a gene; and (2) an adjacent regulatory element for that gene in the same DNA molecule. Additionally, each trans-regulatory element affects a large number of genes on both alleles, while cis-regulatory element is allele specific and only controls genes nearby. Exonic and promoter sequences of the genes are significantly more conserved than the genes in cis- and trans- regulatory elements. Hence, they have higher resistance to genetic divergence, yet retains its susceptibility to mutations in upstream regulators. This accentuates the significance of genetic divergence within species due to cis- and trans-regulatory variants. Trans- and cis-regulatory elements co-evolved rapidly in large-scale to maintain gene expression. They often act in opposite directions, one up-regulates while another down-regulates, to compensate for their effects on the exonic and promoter sequences they act on. Other evolutionary models, such as the independent evolution of trans- or cis-regulatory elements, were deemed incompati
https://en.wikipedia.org/wiki/Ehrenfeucht%E2%80%93Fra%C3%AFss%C3%A9%20game
In the mathematical discipline of model theory, the Ehrenfeucht–Fraïssé game (also called back-and-forth games) is a technique based on game semantics for determining whether two structures are elementarily equivalent. The main application of Ehrenfeucht–Fraïssé games is in proving the inexpressibility of certain properties in first-order logic. Indeed, Ehrenfeucht–Fraïssé games provide a complete methodology for proving inexpressibility results for first-order logic. In this role, these games are of particular importance in finite model theory and its applications in computer science (specifically computer aided verification and database theory), since Ehrenfeucht–Fraïssé games are one of the few techniques from model theory that remain valid in the context of finite models. Other widely used techniques for proving inexpressibility results, such as the compactness theorem, do not work in finite models. Ehrenfeucht–Fraïssé-like games can also be defined for other logics, such as fixpoint logics and pebble games for finite variable logics; extensions are powerful enough to characterise definability in existential second-order logic. Main idea The main idea behind the game is that we have two structures, and two players – Spoiler and Duplicator. Duplicator wants to show that the two structures are elementarily equivalent (satisfy the same first-order sentences), whereas Spoiler wants to show that they are different. The game is played in rounds. A round proceeds as follows: Spoiler chooses any element from one of the structures, and Duplicator chooses an element from the other structure. In simplified terms, the Duplicator's task is to always pick an element "similar" to the one that the Spoiler has chosen, whereas the Spoiler's task is to choose an element for which no "similar" element exists in the other structure. Duplicator wins if there exists an isomorphism between the eventual substructures chosen from the two different structures; otherwise, Spoiler wins.
https://en.wikipedia.org/wiki/Debug%20symbol
A debug symbol is a special kind of symbol that attaches additional information to the symbol table of an object file, such as a shared library or an executable. This information allows a symbolic debugger to gain access to information from the source code of the binary, such as the names of identifiers, including variables and routines. The symbolic information may be compiled together with the module's binary file, or distributed in a separate file, or simply discarded during the compilation and/or linking. This information can be helpful while trying to investigate and fix a crashing application or any other fault. Embedded symbols Debug symbols typically include not only the name of a function or global variable, but also the name of the source code file in which the symbol occurs, as well as the line number at which it is defined. Other information includes the type of the symbol (integer, float, function, exception, etc.), the scope (block scope or global scope), the size, and, for classes, the name of the class, and the methods and members in it. All of this additional information can take up quite a bit of space, especially the filenames and line numbers. Thus, binaries with debug symbols can become quite large, often several times the stripped file size. To avoid this extra size, most operating system distributions ship binaries that are stripped, i.e. from which all of the debugging symbols have been removed. This is accomplished, for example, with the strip command in Unix. Some compilers will output the symbolic debugging information into a separate file, rather than placing it together with the binary. SysV ABI The SysV application binary interface (ABI) includes a specification for the format of debug symbols. This allows any compatible compiler or assembler to create debug symbols in a standardized format, and for any debugger, such as the GNU Debugger (GDB), to gain access and display these symbols. For example, part of the important debug inf