source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Meroplankton
Meroplankton are a wide variety of aquatic organisms which have both planktonic and benthic stages in their life cycles. Much of the meroplankton consists of larval stages of larger organism. Meroplankton can be contrasted with holoplankton, which are planktonic organisms that stay in the pelagic zone as plankton throughout their entire life cycle. After a period of time in the plankton, many meroplankton graduate to the nekton or adopt a benthic (often sessile) lifestyle on the seafloor. The larval stages of benthic invertebrates make up a significant proportion of planktonic communities. The planktonic larval stage is particularly crucial to many benthic invertebrate in order to disperse their young. Depending on the particular species and the environmental conditions, larval or juvenile-stage meroplankton may remain in the pelagic zone for durations ranging from hour to months. Not all meroplankton are larvae or juvenile stages of larger organisms. Many dinoflagellates are meroplanktonic, undergoing a seasonal cycle of encystment and dormancy in the benthic zone followed by excystment and reproduction in the pelagic zone before returning to the benthic zone once more. There also exist meroplanktonic diatoms; these have a seasonal resting phase below the photic zone and can be found commonly amongst the benthos of lakes and coastal zones. Spatial distribution Meroplankton species composition depends on spatial distribution and reproductive habits of adults in a given area. Biotic and abiotic factors such as tidal and lunar cycles and availability of food determine adult spawning schedules, in turn, determining subsequent meroplankton populations. Behavioural factors, such as predator avoidance are also important. Freshwater inputs play a key role in meroplankton species composition in estuarine environments. Effects of tides contribute greatly to meroplankton species distribution. One study conducted in a Patagonian Fjord found that species composition of the
https://en.wikipedia.org/wiki/American%20Information%20Exchange
The American Information Exchange (AMIX) was a platform for the buying and selling of information, goods and services as well as the exchange of information, ideas, and certain kinds of intellectual work product, created by economist and futurist Phil Salin in the 1980s, together with Chip Morningstar (chief architect) and Randy Farmer, and involvement from Esther Dyson and Mitch Kapor. Economist Bill Tulloh was market manager. Salin began thinking about information marketplaces in the 1970s, and was inspired by Friedrich Hayek's idea of spontaneous order. Starting in 1984, Salin worked on AMIX as a tool with the goal of elevating individual decision making over central planning, and improving human coordination to help reduce transaction costs in the economy. AMIX would be an international network for the exchange of information, consulting contracts, computer code and research. He envisaged a world in which the ready exchange of expertise would reduce transaction costs, with wide-ranging beneficial effects. In particular, he predicted that information markets would reduce the need for redundant employees at different organizations, so that companies would become smaller and more efficient, relying on each other as external sources of expertise. He also expected revolutionary political changes as the markets became widely adopted. The AMIX platform was accessed via dial-up and client PC software written in C. The project originated long before the widespread deployment of the Internet, so the challenge of creating the market was compounded by the technical difficulty of creating the network on which it would run. AMIX developed the early mechanics of reputation systems, payment processing, online dispute resolution, as well smart contracts. AMIX had markets for information in primarily technical or business-related fields: on products, technologies, companies and industries, market research reports, and software (including shareware). Services offered included c
https://en.wikipedia.org/wiki/Triton%20%28content%20delivery%29
Triton was a digital delivery and digital rights management service created by Digital Interactive Streams, which abruptly went out of business in early October 2006. Triton was a new competitor in the rapidly growing market for electronic distribution of video games. Triton was being used to serve budget-oriented games from such publishers as Strategy First and Global Star Software, and was most known for distributing Prey. History Triton was launched on November 10, 2004, under the name Game xStream. The service signed several smaller publishers shortly thereafter, and announced its first high-profile deal in May 2005, signing 3D Realms and its then in-development FPS Prey. Game xStream was renamed to its current title of Triton on May 8, 2006. In early October, 2006, owners of Prey who had purchased it via Triton began to complain about problems purchasing the game, activating it, and reaching customer service. 3D Realms' webmaster Joe Siegler managed to find out that Triton and Digital Interactive Streams had gone out of business suddenly and apparently without warning. A follow-up from Royal O'Brien of Triton said that Prey owners who use Triton won't lose their game. A patch was in development to remove the dependency from the live system and allow you to back up/copy and play your games. However, customers who purchased the game through Triton will receive a retail copy. Prey was released on Valve's Steam service, which allows any existing Prey owners to register their game through Steam by entering the activation code, including those who bought Prey through Triton. The game is, however, no longer available for purchase through Steam. Technology Although similar to competing services, the primary selling point of Triton was its "dynamic streaming" technology, which allows for games to be played before they have been completely downloaded - new content is sent to the client as it is needed. All games on the service required the user to be online to be
https://en.wikipedia.org/wiki/Perpendicular%20plate%20of%20ethmoid%20bone
The perpendicular plate of the ethmoid bone (vertical plate) is a thin, flattened lamina, polygonal in form, which descends from the under surface of the cribriform plate, and assists in forming the septum of the nose; it is generally deflected a little to one or other side. The anterior border articulates with the spine of the frontal bone and the crest of the nasal bones. The posterior border articulates by its upper half with the sphenoidal crest, by its lower with the vomer. The inferior border is thicker than the posterior, and serves for the attachment of the septal nasal cartilage of the nose. The surfaces of the plate are smooth, except above, where numerous grooves and canals are seen; these lead from the medial foramina on the cribriform plate and lodge filaments of the olfactory nerves. Additional images
https://en.wikipedia.org/wiki/Bidiakis%20cube
In the mathematical field of graph theory, the bidiakis cube is a 3-regular graph with 12 vertices and 18 edges. Construction The bidiakis cube is a cubic Hamiltonian graph and can be defined by the LCF notation [-6,4,-4]4. The bidiakis cube can also be constructed from a cube by adding edges across the top and bottom faces which connect the centres of opposite sides of the faces. The two additional edges need to be perpendicular to each other. With this construction, the bidiakis cube is a polyhedral graph, and can be realized as a convex polyhedron. Therefore, by Steinitz's theorem, it is a 3-vertex-connected simple planar graph. Algebraic properties The bidiakis cube is not a vertex-transitive graph and its full automorphism group is isomorphic to the dihedral group of order 8, the group of symmetries of a square, including both rotations and reflections. The characteristic polynomial of the bidiakis cube is . Gallery
https://en.wikipedia.org/wiki/Medallion%20%28architecture%29
A medallion is a round or oval ornament that frames a sculptural or pictorial decoration in any context, but typically a façade, an interior, a monument, or a piece of furniture or equipment. Ancient Roman round versions are called an imago clipeata, from the clipeus or Roman round shield. This was a popular form of decoration in neoclassical architecture. The frame and portrait were carved as one, in marble for interiors, and in stone for exterior walls. It is also the name of a scene that is inset into a larger stained glass window. Gallery The following gallery shows how medallions changed over time, from style to style, and how decorated or simple they were. Sometimes they were one of the key ornaments of a style, like the Louis XVI style of the 18th century and the Beaux Arts architecture of the Belle Époque. They also came in different shapes, not just circles and ovals. Many Art Deco medallions are octagonal, showing the use of angular and stylized shapes that characterize the style, inspired by Cubism. They also had different reliefs inside over time. For example, some medieval Moldavian churches are decorated with colourful medallions that feature animals and mythological creatures, while many oval Neoclassical ones feature profiles, inspired by Roman cameos. See also Floor medallion Tondo (art): round (circular) Cartouche (design): oval
https://en.wikipedia.org/wiki/GlowHost
GlowHost.com, LLC. is a privately held web hosting company headquartered in the United States of America and is currently registered in the State of Florida providing shared, reseller, virtual private server, and dedicated web hosting Company history The company was founded in Crested Butte, Colorado by Matt Lundstrom in 2002. GlowHost is a Registered Trademark registered with the United States Patent and Trademark Office. This trademark was first used in commerce on April 1, 2002, later on April 16, 2009 an application for the trademark was submitted to the USPTO which was published for opposition on September 1, 2009. Finally the Trademark was registered on November 17, 2009. GlowHost is accredited with the Better Business Bureau, and had an A+ rating as of October 21, 2015. Press GlowHost introduces Cloud hosting technology for small and medium-sized business, geared to replace dedicated servers. Webhostdir.com interviewed founder and President Matt Lundstrom in 2011. The article is available on their site. In 2010 GlowHost sponsored The Annual Web Hosting Awards at hostreview.com and placed 3rd for the best VPS Hosting service. According to a press release from April 15, 2009, GlowHost acquired DataCities.com, a rival web hosting company. Awards and achievements GlowHost has received numerous awards for its web hosting services. An interesting point is that the company has received an award outside of the web hosting industry. In 2010 GlowHost wrote a software firewall to combat forum Spam for a popular 3rd Party forum software called vBulletin. Known as Spam-O-Matic Spam Firewall, this software was download by over 2500 users in the first couple of months that it was available in the public domain. The software was awarded "Mod of the Month" by vBulletin.org.
https://en.wikipedia.org/wiki/Bubble%20raft
A bubble raft is an array of bubbles. It demonstrates materials' microstructural and atomic length-scale behavior by modelling the {111} plane of a close-packed crystal. A material's observable and measurable mechanical properties strongly depend on its atomic and microstructural configuration and characteristics. This fact is intentionally ignored in continuum mechanics, which assumes a material to have no underlying microstructure and be uniform and semi-infinite throughout. Bubble rafts assemble bubbles on a water surface, often with the help of amphiphilic soaps. These assembled bubbles act like atoms, diffusing, slipping, ripening, straining, and otherwise deforming in a way that models the behavior of the {111} plane of a close-packed crystal. The ideal (lowest energy) state of the assembly would undoubtedly be a perfectly regular single crystal, but just as in metals, the bubbles often form defects, grain boundaries, and multiple crystals. History of bubble rafts The concept of bubble raft modelling was first presented in 1947 by Nobel Laureate Sir William Lawrence Bragg and John Nye of Cambridge University's Cavendish Laboratory in Proceedings of the Royal Society A. Legend claims that Bragg conceived of bubble raft models while pouring oil into his lawn mower. He noticed that bubbles on the surface of the oil assembled into rafts resembling the {111} plane of close-packed crystals. Nye and Bragg later presented a method of generating and controlling bubbles on the surface of a glycerine-water-oleic acid-triethanolamine solution, in assemblies of 100,000 or more sub-millimeter sized bubbles. In their paper, they go on at length about the microstructural phenomena observed in bubble rafts and hypothesized in metals. Dynamics Bubble rafts exhibit complex dynamics, as illustrated in the video. This is triggered by rupture of a first bubble, driven by thermal fluctuations and a cascade of subsequent bursting bubbles, which can give rise to self-organized
https://en.wikipedia.org/wiki/Ground%20glass%20hepatocyte
In liver pathology, a ground glass hepatocyte, abbreviated GGH, is a liver parenchymal cell with a flat hazy and uniformly dull appearing cytoplasm on light microscopy. The cytoplasm's granular homogeneous eosinophilic staining is caused by the presence of HBsAg. The appearance is classically associated with abundant hepatitis B antigen in the endoplasmic reticulum, but may also be drug-induced. In the context of hepatitis B, GGHs are only seen in chronic infections, i.e. they are not seen in acute hepatitis B. GGHs were first described by Hadziyannis et al. Types Several different types of GGHs are recognized: Type I - morphologically consist of GGHs that are scattered singly and have weak Pre-S2 positive immunostaining. Type II - morphologically consist of GGHs that are in clusters and have Pre-S2 negative immunostaining. There is some evidence to suggest that type II GGHs predispose to hepatocellular carcinoma. See also Drug reaction Mallory body Viral hepatitis Additional images
https://en.wikipedia.org/wiki/Laser%20SETI
Laser SETI is an instrument that could continuously survey the entire night sky for brief laser pulses. The instrument can look everywhere simultaneously.  The technology, which consists of a robust assembly of straightforward optical and mechanical components, has been prototyped and subjected to preliminary tests. It will observe all of the sky, all of the time so even relatively rare events can be found. Laser SETI can discover pulses over a wide range of pulse durations, and is especially sensitive to millisecond singleton pulses which may have been overlooked in previous astronomical surveys. As of October 2017, the team had spent close to $50k thus far, have 21 components in hand, 5 on order or in transit, 3 ready to order, and 7 waiting on test results or TBD. In 2018, the SETI Institute announced that they were going to be able to deploy 8 cameras instead of four, meaning that they can fully monitor two independent fields-of-view. In 2019, the entity announced that the final logistics were being worked out for the placement of LaserSETI's first observatory at RFO's (Robert Ferguson Observatory) idyllic facility, in Sonoma County. In summer 2021 a second LaserSETI observatory was being installed in Hawaii, and was operational by Dec 2021. Two of four cameras are fully functional. Cameras are installed in pairs with their diffraction gratings at 90 degrees to each other. Images are read out more than a thousand times a second.
https://en.wikipedia.org/wiki/European%20Girls%27%20Mathematical%20Olympiad
The European Girls' Mathematical Olympiad (EGMO) is a mathematical olympiad for girls which started in 2012, and is held in April each year. It was inspired by the China Girls Mathematical Olympiad (CGMO). Although the competition is held in Europe, it is open to female participants from all over the world, and is considered the most prestigious mathematics competition for girls. In recent years, participants from around 55 countries have been invited to the competition. Process and scoring The competition is similar in style to the International Mathematical Olympiad (IMO), with two papers, each consisting of three problems to be solved in 4.5 hours, taken on consecutive days. Participating countries send teams consisting of four female mathematicians below the age of 20 who are not enrolled at a university. Each of the six problems are marked out of 7, making the maximum possible score 42 points. The first edition was held in Cambridge, UK. Since then, 10 other countries in Europe have organized the EGMO. The number of participating countries have grown from 19 in the first edition to 57 in the eleventh edition, and the number of contestants from 61 in the first edition to 226 in the eleventh edition. The competitors participate as a team of 4 under the national flag but the contest itself is individual. The selection process varies between countries, but it often involves the results obtained in the national Mathematical Olympiads and in other tests, which become progressively more selective. Medals are awarded according to this criterion: The top 1/12 of the competitors receive a gold medal The following 1/6 of the general classification receive a silver medal The subsequent 1/4 of the general classification receive a bronze medal All those who have not received a medal but have scored the maximum points in at least one of the six problems receive an honorable mention. Summary Medal table The 61 countries that have won a medal are as follows: The indi
https://en.wikipedia.org/wiki/Hong%20Kong%20Olympiad%20in%20Informatics
Hong Kong Olympiad in Informatics (HKOI; 香港電腦奧林匹克競賽) is an annual programming competition for secondary school students in Hong Kong, emphasizing on problem solving techniques and programming skills. It is co-organized by the Hong Kong Association for Computer Education (HKACE) and the Hong Kong Education Bureau (EDB). It serves as a preliminary contest to international, national and regional competitions such as the China National Olympiad in Informatics (NOI) and the International Olympiad in Informatics (IOI). The first HKOI was held in 1997. History Hong Kong first participated in IOI in 1992. In order select representatives for the Hong Kong Delegation Team, a selection test was held a few months before the competition. In the following years, Hong Kong started sending teams to other competitions, including the SEARCC International Schools' Software Competition (ISSC) in 1993, the Software Competition for the Youths (SCY) in 1994 and the China National Olympiad in Informatics in 1995. Selection tests were separately administered for these competitions, and the purpose of each test was solely to select team members for the competitions. A considerable amount of resources were used to organize these tests. The tests were not very popular among students in Hong Kong. In 1996, the Hong Kong Association for Computer Education, the Hong Kong Computer Society and the Education Department of Hong Kong (now the Education Bureau) jointly organized the Joint Selection Contest to replace all the selection tests. 39 students were selected as seeds for the Hong Kong teams. They received intensive training on topics like data structures and algorithms. After that, a Team Formation Test was conducted to select the Hong Kong representatives in IOI and NOI among the seeds. Another Team Formation Test was conducted for the SEARCC-ISSC and SCY. In 1997, the Joint Selection Contest was renamed as the Hong Kong Olympiad in Informatics. Prizes are awarded to students with good re
https://en.wikipedia.org/wiki/Axial%20pen%20force
In graphonomics, Axial pen force is the component of the normal pen force that is parallel to the pen. It is dependent upon pen tilt. In the special case of a perfectly vertical orientation of the writing instrument the axial pen force equals the normal pen force. See also Graphonomics External links http://hwr.nici.ru.nl/~miami/taxonomy/node54.html http://hwr.nici.kun.nl/~miami/taxonomy/node57.html Pen ergonomics Penmanship Force
https://en.wikipedia.org/wiki/Complex%20oxide
A complex oxide is a chemical compound that contains oxygen and at least two other elements (or oxygen and just one other element that's in at least two oxidation states). Complex oxide materials are notable for their wide range of magnetic and electronic properties, such as ferromagnetism, ferroelectricity, and high-temperature superconductivity. These properties often come from their strongly correlated electrons in d or f orbitals. Natural occurrence Many minerals found in the ground are complex oxides. Commonly studied mineral crystal families include spinels and perovskites. Applications Complex oxide materials are used in a variety of commercial applications. Magnets Magnets made of the complex oxide ferrite are commonly used in transformer cores and in inductors. Ferrites are ideal for these applications because they are magnetic, electrically insulating, and inexpensive. Transducers and actuators Piezoelectric transducers and actuators are often made of the complex oxide PZT (lead zirconate titanate). These transducers are used in applications such ultrasound imaging and some microphones. PZT is also sometimes used for piezo ignition in lighters and gas grills. Capacitors Complex oxide materials are the dominant dielectric material in ceramic capacitors. About one trillion ceramic capacitors are produced each year to be used in electronic equipment. Fuel cells Solid oxide fuel cells often use complex oxide materials as their electrolytes, anodes, and cathodes. Gemstone jewelry Many precious gemstones, such as emerald and topaz, are complex oxide crystals. Historically, some complex oxide materials (such as strontium titanate, yttrium aluminium garnet, and gadolinium gallium garnet) were also synthesized as inexpensive diamond simulants, though after 1976 they were mostly eclipsed by cubic zirconia. New electronic devices As of 2015, there is research underway to commercialize complex oxides in new kinds of electronic devices, such as ReRAM, FeRA
https://en.wikipedia.org/wiki/Belgrade
Belgrade ( , ; , ; names in other languages) is the capital and largest city of Serbia. It is located at the confluence of the Sava and Danube rivers and at the crossroads of the Pannonian Plain and the Balkan Peninsula. The population of the Belgrade metropolitan area is 1,681,405 according to the 2022 census. It is one of the major cities of Southeast Europe and the third most populous city on the Danube river. Belgrade is one of the oldest continuously inhabited cities in Europe and the world. One of the most important prehistoric cultures of Europe, the Vinča culture, evolved within the Belgrade area in the 6th millennium BC. In antiquity, Thraco-Dacians inhabited the region and, after 279 BC, Celts settled the city, naming it Singidūn. It was conquered by the Romans under the reign of Augustus and awarded Roman city rights in the mid-2nd century. It was settled by the Slavs in the 520s, and changed hands several times between the Byzantine Empire, the Frankish Empire, the Bulgarian Empire, and the Kingdom of Hungary before it became the seat of the Serbian king Stefan Dragutin in 1284. Belgrade served as capital of the Serbian Despotate during the reign of Stefan Lazarević, and then his successor Đurađ Branković returned it to the Hungarian king in 1427. Noon bells in support of the Hungarian army against the Ottoman Empire during the siege in 1456 have remained a widespread church tradition to this day. In 1521, Belgrade was conquered by the Ottomans and became the seat of the Sanjak of Smederevo. It frequently passed from Ottoman to Habsburg rule, which saw the destruction of most of the city during the Ottoman–Habsburg wars. Following the Serbian Revolution, Belgrade was once again named the capital of Serbia in 1841. Northern Belgrade remained the southernmost Habsburg post until 1918, when it was attached to the city, due to former Austro-Hungarian territories becoming part of the new Kingdom of Serbs, Croats and Slovenes after World War I. Belgrade was
https://en.wikipedia.org/wiki/Realizability
In mathematical logic, realizability is a collection of methods in proof theory used to study constructive proofs and extract additional information from them. Formulas from a formal theory are "realized" by objects, known as "realizers", in a way that knowledge of the realizer gives knowledge about the truth of the formula. There are many variations of realizability; exactly which class of formulas is studied and which objects are realizers differ from one variation to another. Realizability can be seen as a formalization of the BHK interpretation of intuitionistic logic; in realizability the notion of "proof" (which is left undefined in the BHK interpretation) is replaced with a formal notion of "realizer". Most variants of realizability begin with a theorem that any statement that is provable in the formal system being studied is realizable. The realizer, however, usually gives more information about the formula than a formal proof would directly provide. Beyond giving insight into intuitionistic provability, realizability can be applied to prove the disjunction and existence properties for intuitionistic theories and to extract programs from proofs, as in proof mining. It is also related to topos theory via realizability topoi. Example: Kleene's 1945-realizability Kleene's original version of realizability uses natural numbers as realizers for formulas in Heyting arithmetic. A few pieces of notation are required: first, an ordered pair (n,m) is treated as a single number using a fixed primitive recursive pairing function; second, for each natural number n, φn is the computable function with index n. The following clauses are used to define a relation "n realizes A" between natural numbers n and formulas A in the language of Heyting arithmetic, known as Kleene's 1945-realizability relation: Any number n realizes an atomic formula s=t if and only if s=t is true. Thus every number realizes a true equation, and no number realizes a false equation. A pair (n,m
https://en.wikipedia.org/wiki/Xerophyte
A xerophyte (from Greek ξηρός xeros 'dry' + φυτόν phuton 'plant') is a species of plant that has adaptations to survive in an environment with little liquid water. Examples are typically desert regions like the Sahara, and places in the Alps or the Arctic. Popular examples of xerophytes are cacti, pineapple and some Gymnosperm plants. The structural features (morphology) and fundamental chemical processes (physiology) of xerophytes are variously adapted to conserve water, also common to store large quantities of water, during dry periods. Other species are able to survive long periods of extreme dryness or desiccation of their tissues, during which their metabolic activity may effectively shut down. Plants with such morphological and physiological adaptations are . Xerophytes such as cacti are capable of withstanding extended periods of dry conditions as they have deep-spreading roots and capacity to store water. Their waxy, thorny leaves prevent loss of moisture. Even their fleshy stems can store water. Introduction Plants absorb water from the soil, which then evaporates from their shoots and leaves; this process is known as transpiration. If placed in a dry environment, a typical mesophytic plant would evaporate water faster than the rate of water uptake from the soil, leading to wilting and even death. Xerophytic plants exhibit a diversity of specialized adaptations to survive in such water-limiting conditions. They may use water from their own storage, allocate water specifically to sites of new tissue growth, or lose less water to the atmosphere and so channel a greater proportion of water from the soil to photosynthesis and growth. Different plant species possess different qualities and mechanisms to manage water supply, enabling them to survive. Cacti and other succulents are commonly found in deserts, where there is little rainfall. Other xerophytes, such as certain bromeliads, can survive through both extremely wet and extremely dry periods and can
https://en.wikipedia.org/wiki/Fumed%20silica
Fumed silica (CAS number 112945-52-5), also known as pyrogenic silica because it is produced in a flame, consists of microscopic droplets of amorphous silica fused into branched, chainlike, three-dimensional secondary particles which then agglomerate into tertiary particles. The resulting powder has an extremely low bulk density and high surface area. Its three-dimensional structure results in viscosity-increasing, thixotropic behavior when used as a thickener or reinforcing filler. Properties Fumed silica has a very strong thickening effect. Primary particle size is 5–50 nm. The particles are non-porous and have a surface area of 50–600 m2/g. The density is 160–190 kg/m3. Production Fumed silica is made from flame pyrolysis of silicon tetrachloride or from quartz sand vaporized in a 3000 °C electric arc. Major global producers are Evonik (who sells it under the name Aerosil), Cabot Corporation (Cab-O-Sil), Wacker Chemie (HDK), Dow Corning, Heraeus (Zandosil), Tokuyama Corporation (Reolosil), OCI (Konasil), Orisil (Orisil) and Xunyuchem(XYSIL). Applications Fumed silica serves as a universal thickening agent and an anticaking agent (free-flow agent) in powders. Like silica gel, it serves as a desiccant. It is used in cosmetics for its light-diffusing properties. It is used as a light abrasive, in products like toothpaste. Other uses include filler in silicone elastomer and viscosity adjustment in paints, coatings, printing inks, adhesives and unsaturated polyester resins. Fumed silica readily forms a network structure within bitumen and enhances its elasticity. Health issues Fumed silica is not listed as a carcinogen by OSHA, IARC, or NTP. Due to its fineness and thinness, fumed silica can easily become airborne, making it an inhalation hazard, capable of causing irritation. See also Hydrophobic silica Precipitated silica Nanoparticles
https://en.wikipedia.org/wiki/Reductive%20dechlorination
In organochlorine chemistry, reductive dechlorination describes any chemical reaction which cleaves the covalent bond between carbon and chlorine via reductants, to release chloride ions. Many modalities have been implemented, depending on the application. Reductive dechlorination is often applied to remediation of chlorinated pesticides or dry cleaning solvents. It is also used occasionally in the synthesis of organic compounds, e.g. as pharmaceuticals. Chemical Dechlorination is a well-researched reaction in organic synthesis, although it is not often used. Usually stoichiometric amounts of dechlorinating agent are required. In one classic application, the Ullmann reaction, chloroarenes are coupled to biphenyl]]s. For example, the activated substrate 2-chloronitrobenzene is converted into 2,2'-dinitrobiphenyl with a copper - bronze alloy. Zerovalent iron effects similar reactions. Organophosphorus(III) compounds effect gentle dechlorinations. The products are alkenes and phosphorus(V). Alkaline earth metals and zinc are used for more difficult dechlorinations. The side product is zinc chloride. Biological Vicinal reduction involves the removal of two halogen atoms that are adjacent on the same alkane or alkene, leading to the formation of an additional carbon-carbon bond. Biological reductive dechlorination is often effected by certain species of bacteria. Sometimes the bacterial species are highly specialized for organochlorine respiration and even a particular electron donor, as in the case of Dehalococcoides and Dehalobacter. In other examples, such as Anaeromyxobacter, bacteria have been isolated that are capable of using a variety of electron donors and acceptors, with a subset of possible electron acceptors being organochlorines. These reactions depend on a molecule which tends to be very aggressively sought after by some microbes, vitamin B12. Bioremediation using reductive dechlorination Reductive dechlorination of chlorinated organic molecule
https://en.wikipedia.org/wiki/Ultramicrotomy
Ultramicrotomy is a method for cutting specimens into extremely thin slices, called ultra-thin sections, that can be studied and documented at different magnifications in a transmission electron microscope (TEM). It is used mostly for biological specimens, but sections of plastics and soft metals can also be prepared. Sections must be very thin because the 50 to 125 kV electrons of the standard electron microscope cannot pass through biological material much thicker than 150 nm. For best resolutions, sections should be from 30 to 60 nm. This is roughly the equivalent to splitting a 0.1 mm-thick human hair into 2,000 slices along its diameter, or cutting a single red blood cell into 100 slices. Ultramicrotomy process Ultra-thin sections of specimens are cut using a specialized instrument called an "ultramicrotome". The ultramicrotome is fitted with either a diamond knife, for most biological ultra-thin sectioning, or a glass knife, often used for initial cuts. There are numerous other pieces of equipment involved in the ultramicrotomy process. Before selecting an area of the specimen block to be ultra-thin sectioned, the technician examines semithin or "thick" sections range from 0.5 to 2 μm. These thick sections are also known as survey sections and are viewed under a light microscope to determine whether the right area of the specimen is in a position for thin sectioning. "Ultra-thin" sections from 50 to 100 nm thick are able to be viewed in the TEM. Tissue sections obtained by ultramicrotomy are compressed by the cutting force of the knife. In addition, interference microscopy of the cut surface of the blocks reveals that the sections are often not flat. With Epon or Vestopal as embedding medium the ridges and valleys usually do not exceed 0.5 μm in height, i.e., 5–10 times the thickness of ordinary sections (1). A small sample is taken from the specimen to be investigated. Specimens may be from biological matter, like animal or plant tissue, or from inorgani
https://en.wikipedia.org/wiki/Krrb
Krrb (pronounced 'curb') was a hyperlocal classified advertising website that allowed individuals and businesses to sell items. Members of the website could also list real estate, housing, events, services, jobs and community notices. The website was based on the idea of neighborhood garage sale or flea market, prioritizing proximity over categorization so that users could see listings nearest to them. Each seller got a personal storefronts (known as 'corners') where all their listings were displayed. As of October 2014, the company had 31,332 listings from 92,341 members in 3,211 cities and 118 countries. The website's domain was no longer operational as of March 2017. History George Eid created an "online flea market" designed to be a competitor to craigslist. Krrb.com was launched in November 2010 and incorporated as Krrb, Inc. in January 2011. The company is headquartered in Brooklyn, New York City and is privately owned by Area 17, an interactive agency based in New York City and Paris, France. Luis Lavena, core developer of the programming language Ruby, designed and developed the software and led the engineering team. Interactive art director Arnaud Mercier (now deceased) designed the logo and user interface. Andrew Wagner, former editor of ReadyMade, American Craft and Dwell Magazine and columnist for the New York Times, joined the company in 2011 as director and editor-in-chief. In 2014, Phil Jeffs, director of product at Area 17, joined the team to lead product development. Krrb classifieds network In 2012, Krrb partnered with Apartment Therapy to create classifieds for the site. Following this, Krrb created a classifieds network of other targeted classified sites. The Krrb Classified Network was released in October 2013 with the launch of the Chicago Magazine Classifieds. Conflict with Craigslist In February 2013, Craigslist attorneys sent a cease-and-desist letter demanding that Krrb permanently disable its "Krrb It" button, a bookmarklet tool tha
https://en.wikipedia.org/wiki/Bioinorganic%20Chemistry%20Award
The Bioinorganic Chemistry Award has been awarded by the Dalton division of the Royal Society of Chemistry every two years since 2009. The winner receives £2000 and undertakes a lecture tour in the UK. The award was discontinued in 2020. Winners Source: See also List of chemistry awards
https://en.wikipedia.org/wiki/Okta
In meteorology, an okta is a unit of measurement used to describe the amount of cloud cover at any given location such as a weather station. Sky conditions are estimated in terms of how many eighths of the sky are covered in cloud, ranging from 0 oktas (completely clear sky) through to 8 oktas (completely overcast). In addition, in the SYNOP code there is an extra cloud cover indicator '9' indicating that the sky is totally obscured (i.e. hidden from view), usually due to dense fog or heavy snow. When used in weather charts, okta measurements are shown by means of graphic symbols (rather than numerals) contained within weather circles, to which are attached further symbols indicating other measured data such as wind speed and wind direction. Although relatively straightforward to measure (visually, for instance, by using a mirror), oktas only estimate cloud cover in terms of the area of the sky covered by clouds. They do not account for cloud type or thickness, and this limits their use for estimating cloud albedo or surface solar radiation receipt. Cloud oktas can also be measured using satellite imagery from geostationary satellites equipped with high-resolution image sensors such as Himawari-8. Similar to traditional approaches, satellite images do not account for cloud composition. Oktas are often referenced in aviation weather forecasts and low level forecasts: SKC = Sky clear (0 oktas); FEW = Few (1 to 2 oktas); SCT = Scattered (3 to 4 oktas); BKN = Broken (5 to 7 oktas); OVC = Overcast (8 oktas); NSC = nil significant cloud; CAVOK = ceiling and visibility okay. Hand-drawn maps In the early 20th century, it was common for weather maps to be hand drawn. The symbols for cloud cover on these maps, like the modern symbols, were drawn inside the circle marking the position of the weather station making the measurements. Unlike the modern symbols, these ones consisted of straight lines only rather than filled in blocks which would have been less practical
https://en.wikipedia.org/wiki/Lateral%20condyle%20of%20femur
The lateral condyle is one of the two projections on the lower extremity of the femur. The other one is the medial condyle. The lateral condyle is the more prominent and is broader both in its front-to-back and transverse diameters. Clinical significance The most common injury to the lateral femoral condyle is an osteochondral fracture combined with a patellar dislocation. The osteochondral fracture occurs on the weight-bearing portion of the lateral condyle. Typically, the condyle will fracture (and the patella may dislocate) as a result of severe impaction from activities such as downhill skiing and parachuting. Open reduction and internal fixation surgery is typically used to repair an osteochondral fracture. For a Type B1 partial articular fracture of the lateral condyle, interfragmentary lag screws are used to secure the bone back together. Supplementation of buttress screws or a buttress plate is used if the fracture extends to the proximal metaphysis or distal diaphysis. Additional images
https://en.wikipedia.org/wiki/Enneagram%20of%20Personality
The Enneagram of Personality, or simply the Enneagram (from the Greek words [, meaning "nine"] and [, meaning something "written" or "drawn"]), is a model of the human psyche which is principally understood and taught as a typology of nine interconnected personality types. Although the origins and history of ideas associated with the Enneagram of Personality are disputed, contemporary approaches are principally derived from the teachings of the Bolivian psycho-spiritual teacher Oscar Ichazo from the 1950s and the Chilean psychiatrist Claudio Naranjo from the 1970s. Naranjo's theories were also influenced by earlier teachings about personality by George Gurdjieff and the Fourth Way tradition in the first half of the 20th century. As a typology, the Enneagram defines nine personality types (sometimes called "enneatypes"), which are represented by the points of a geometric figure called an enneagram, which indicate some of the principal connections between the types. There have been different schools of thought among Enneagram teachers and their understandings are not always in agreement. The Enneagram of Personality is promoted in both business management and spirituality contexts through seminars, conferences, books, magazines, and DVDs. In business contexts, it is often promoted as a means to gain insights into workplace interpersonal dynamics; in spirituality it is commonly presented as a path to states of enlightenment and essence. Proponents in both contexts say it has aided in self-awareness, self-understanding, and self-development. There has been limited formal psychometric analysis of the Enneagram, and the peer-reviewed research that has been done is not accepted within the relevant academic communities. Though the Enneagram integrates some concepts that parallel other theories of personality, it has been dismissed by personality assessment experts as pseudoscience. History The origins and historical development of the Enneagram of Personality are
https://en.wikipedia.org/wiki/Barometer%20question
The barometer question is an example of an incorrectly designed examination question demonstrating functional fixedness that causes a moral dilemma for the examiner. In its classic form, popularized by American test designer professor Alexander Calandra (1911–2006), the question asked the student to "show how it is possible to determine the height of a tall building with the aid of a barometer." The examiner was confident that there was one, and only one, correct answer, which is found by measuring the difference in pressure at the top and bottom of the building and solving for height. Contrary to the examiner's expectations, the student responded with a series of completely different answers. These answers were also correct, yet none of them proved the student's competence in the specific academic field being tested. The barometer question achieved the status of an urban legend; according to an internet meme, the question was asked at the University of Copenhagen and the student was Niels Bohr. The Kaplan, Inc. ACT preparation textbook describes it as an "MIT legend", and an early form is found in a 1958 American humor book. However, Calandra presented the incident as a real-life, first-person experience that occurred during the Sputnik crisis. Calandra's essay, "Angels on a Pin", was published in 1959 in Pride, a magazine of the American College Public Relations Association. It was reprinted in Current Science in 1964, in Saturday Review in 1968 and included in the 1969 edition of Calandra's The Teaching of Elementary Science and Mathematics. Calandra's essay became a subject of academic discussion. It was frequently reprinted since 1970, making its way into books on subjects ranging from teaching, writing skills, workplace counseling and investment in real estate to chemical industry, computer programming and integrated circuit design. Calandra's account A colleague of Calandra posed the barometer question to a student, expecting the correct answer: "the heigh
https://en.wikipedia.org/wiki/Balich%C3%A3o
Balichão (Cantonese: , haam4 haa1 zoeng3, "salt shrimp sauce") is an ingredient that is used in a number of dishes in Macanese cuisine. It is made with shrimp, aguardente, salt, pepper, bay leaves and malagueta chillies. Characteristics Despite being commonly described as a shrimp paste, some sources state balichão is actually made with krill. Some describe Balichão as a fish sauce. Balichão is noted for having a distinct odor that some find foul. Some, however, find balichão's smell and taste milder and mellower than shrimp pastes from Southeast Asia. History There are at least two theories on where the word balichão originated. One theory states the word is believed to have evolved from the Portuguese word for Whale (baleia). Another theory states the word is a corruption of the word belachang, which is one of a number of Malay words for shrimp paste. Balichão is noted as originating from shrimp pastes that were brought to Macao by Portuguese sailors who discovered it during their voyage to the city, and later developed by locals as a special variant. Balichãos are noted to have a tradition of being home made, and can take up to a month to mature. In 2009, Balichão is described as an almost extinct condiment, as people look for its substitute among other, more readily available ingredients. Ingredients Besides krill, Balichão contains bay leaves, chili, clove, lime, peppercorn, and is laced with fortified wine. Culture Balichão is noted to be so central to Macanese culture, that Macanese people call themselves balichung as an affectionate demonym. There is also a street named after balichão in Macao, named Travessa do Balichão ().
https://en.wikipedia.org/wiki/Wendy%20Young
Wendy B. Young is a medicinal chemist and pharmaceutical executive currently employed at Genentech. Education Young received her B.S. and M.S. from Wake Forest University, working with Prof. Huw Davies. She was co-author on an early application of Davies' rhodium(II) carbenoid insertion - Cope rearrangement chemistry, leading to the total synthesis of three small tropane natural products. Young received her Ph.D. from Princeton in 1993, working with Edward C. Taylor on heterocycles derived from natural pigments, one of which ultimately became pemetrexed (Alimta), an oncology treatment. In her postdoctoral fellowship with Samuel Danishefsky, Young was among one of a handful of groups in the mid-1990s to synthesize paclitaxel (Taxol), a highly-oxygenated terpenoid natural product used to treat cancer. Career Despite multiple employment offers on the East Coast of the United States, Young chose to remain in the San Francisco Bay Area for her professional career. From 1995 to 2006, Young worked at Celera Genomics, studying inhibitor compounds of human plasma proteins such as kallikrein and Factors VIIa and IXa. She was recruited to Genentech in 2006, and in 2018 was promoted to Senior Vice President of Small Molecule drug discovery. One of her major research successes was development of a chemistry campaign against Bruton's tyrosine kinase, leading to molecules to potentially treat rheumatoid arthritis and B-cell lymphomas. Her team developed fenebrutinib, currently in Phase II trials for several autoimmune disorders. Awards 2018 - William S. Johnson Symposium, Stanford University 2017 - Elected Chair of ACS Medicinal Chemistry Division 2015 - "Most Influential Woman of 2015" - San Francisco Business Times 1995 - American Cancer Society Postdoctoral Fellowship 1993 - H.W. Dodds Top Thesis Award, Princeton University
https://en.wikipedia.org/wiki/Methylthiouracil
Methylthiouracil is an organosulfur compound that is used antithyroid preparation. It is a thioamide, closely related to propylthiouracil. Methylthiouracil is not used clinically in the United States, it has a similar mechanism of action and side effect to that of propylthiouricil. The drug acts to decrease the formation of stored thyroid hormone, as thyroglobulin in the thyroid gland. The clinical effects of the drug to treat the hyperthyroid state can have a lag period of up to two weeks, depending on the stores of thyroglobulin and other factors. Synthesis Methylthiouracil is prepared quite simply by condensation of ethyl acetoacetate with thiourea. Further work in this series shows that better activity was obtained by incorporation of a lipophilic side chain.
https://en.wikipedia.org/wiki/2013%20in%20paleobotany
This list of 2013 in paleobotany records new fossil plant taxa that were described during 2013, as well as other significant discoveries and events related to paleobotany that occurred in the year. Chlorophyta Bryophyta Ferns and fern allies Ginkgophytes Gnetophyta Pachytestopsida Pinophyta Angiosperms Other seed plants Other plants
https://en.wikipedia.org/wiki/Radio-frequency%20welding
Radio-frequency welding, also known as dielectric welding and high-frequency welding, is a plastic welding process that utilizes high-frequency electric fields to induce heating and melting of thermoplastic base materials. The electric field is applied by a pair of electrodes after the parts being joined are clamped together. The clamping force is maintained until the joint solidifies. Advantages of this process are fast cycle times (on the order of a few seconds), automation, repeatability, and good weld appearance. Only plastics which have dipoles can be heated using radio waves and therefore not all plastics are able to be welded using this process. Also, this process is not well suited for thick or overly complex joints. The most common use of this process is lap joints or seals on thin plastic sheets or parts. Heating mechanism Four types of polarization can occur in materials subjected to high-frequency alternating electric fields: Electronic or electric polarization is the redistribution of electrons Ionic polarization is the redistribution of charged particles—cations and anions Maxwell–Wagner polarization is a charge buildup at the interfaces of non-homogeneous materials Dipole polarization is the realignment of permanent dipoles Dipole polarization is the phenomenon that is responsible for the heating mechanism in Radio Frequency plastic welding, dielectric heating. When an electric field is applied to a molecule with an asymmetric distribution of charge, or dipole, the electric forces cause the molecule to align itself with the electrical field. When an alternating electrical field is applied, the molecule will continuously reverse its alignment, leading to molecular rotation. This process is not instantaneous, therefore if the frequency is high enough, the dipole will be unable to rotate quickly enough to stay aligned with the electric field resulting in random motion as the molecule attempts to follow the electrical field. This motion cau
https://en.wikipedia.org/wiki/Carl-Gustav%20Esseen
Carl-Gustav Esseen (13 September 1918 in Linköping – 10 November 2001) was a Swedish mathematician. His work was in the theory of probability. The Berry–Esseen theorem is named after him. Life Carl-Gustav Esseen attended school in Linköping. Starting in 1936, he studied mathematics, astronomy, physics and chemistry at the University of Uppsala. Inspired by the harmonic-analytic research of Harald Cramér and Arne Beurling, Esseen examined the accuracy of the approximation to the normal distribution in the central limit theorem in the case of independent and identically distributed summands. Esseen's bound is now called "the Berry-Esseen theorem", because it was independently proved by Andrew C. Berry, also. In 1944 Esseen received his doctorate with a thesis on the Fourier analysis of probability distributions. In 1949 he was appointed full professor of applied mathematics at the Royal Institute of Technology in Stockholm. In 1962 his professorship moved to the field of mathematical statistics and in 1967, he became first holder of the chair of mathematical statistics at the University of Uppsala. He retired in 1984. Scientific work Although Esseen worked mostly on the central limit theorem and related topics, he also worked in other areas. Some industrial applications were considered in his writings, for example, his studies on control theory and in telecommunications. After retirement, Esseen worked on topics from number theory, especially factorization, a topic of importance in cryptology. Esseen supervised several doctoral students. His lectures and writings were meticulously prepared and delivered. Honours 1963: Elected Member of the Royal Swedish Academy of Engineering Sciences. Selected works "On the Liapounoff limit of error in the theory of probability", Arkiv för Mathematik, Astronomi och Fysik 28A, #9 (1942), 19 pp., . "Determination of the maximum deviation from the Gaussian law", Arkiv för Mathematik, Astronomi och Fysik 29A, #20 (1943), 10
https://en.wikipedia.org/wiki/Motifs%20targeted%20by%20APC/C
The anaphase- promoting complex or cyclosome (APC/C) is a highly specific ubiquitin protein ligase responsible for triggering events of late mitosis. In early mitosis, Cdc20 levels rise and APC/C binds to form active APC/CCdc20. This then leads to the destruction of mitotic cyclins, securin, and other proteins to trigger chromosome separation in anaphase. In early anaphase, Cdk1 is inactivated, leading to the activation of Cdh1, the other activator subunit of APC/C. This then triggers the degradation of Cdc20 and leads to the activation of APC/CCdh1 through G1 to suppress S- phase cyclin-Cdk activity. At the end of G1, APC/CCdh1 is inactivated and S- phase and mitotic cyclins gets reaccumulate as the cell progresses to S phase. APC/C finds its substrates via short, linear sequence motifs or degrons found in its substrates. These degrons can be split into three major groups: destruction boxes, KEN boxes, and ABBA motifs. Binding pockets on the surface of activator subunits of Cdc20 or Cdh1 interact with the degrons on the substrates to target them for ubiquitination and degradation to trigger specific events in the cell cycle. Degrons targeted by APC/C The APC/C is a binding platform that associates with an activator subunit (Cdc20 or Cdh1) which has binding pockets on its surface to interact with degrons on the substrate. The three main degrons that the APC/C interacts with are destruction box (also known as D box), the KEN box, and the ABBA motif. Most APC targets contain the D and the KEN box. These degrons are typically located in largely unstructured regions of the substrates. Destruction Box (D Box) The D box is a sequence motif found in many targets of the APC with the consensus RxxLxxxxN amino acid sequence. R is arginine, X is any amino acid, L is leucine, and N is asparagine. This binding pocket involves the activator WD40 domain, which is located in the C- terminal half of the APC/C activator subunit, and a site on the Apc10/ Doc1 subunit of the APC
https://en.wikipedia.org/wiki/Gnome%20Ranger
Gnome Ranger is a text adventure game designed by Peter Austin and released by Level 9 in 1987. The Atari 8-bit family and Apple II ports are text-only. The disk version for other systems includes still graphic images to accompany the text. A sequel was published in 1988: Ingrid's Back. Plot The gnome Ingrid Bottomlow has displeased her family by her un-gnomelike behaviour, such as going off to university and getting an education. She has been teleported from her village by a faulty scroll, and must find her way back… Gameplay The game is a text adventure with limited graphics on some systems. It comes with a short novella by Peter McBride ("The Gnettlefield Journal") explaining Ingrid's predicament and setting the background to the story. Gameplay is similar to the earlier Level 9 adventure Knight Orc, which uses the same game engine (KAOS). The player must explore the settings while collecting useful items and interacting with various non-player characters to solve puzzles and problems. The game takes place in three areas, each characterized by the non-player characters Ingrid will meet. The first contains characters of an animal nature, the second of a vegetable nature, and the third of a mineral nature. Reception Gnome Ranger received mostly positive reviews. Atari ST User magazine reviewer rated it 9/10 points overall, concluding: "I am convinced it is Level 9's best ever. It is funny, well-plotted, teasing, very friendly, but powerful, literate and big". Similarly, the review for ACE magazine also ended on a positive note: "Gnome Ranger has improved on Knight Orc in the puzzle stakes (more logical and enjoyable), the atmosphere (more enjoyable and less confusing), the plot (more comprehensible), and the characters (more interesting). You can't ask for much more than that".
https://en.wikipedia.org/wiki/Digital%20Audio%20Access%20Protocol
The Digital Audio Access Protocol (DAAP) is the proprietary protocol introduced by Apple in its iTunes software to share media across a local network. DAAP addresses the same problems for Apple as the UPnP AV standards address for members of the Digital Living Network Alliance (DLNA). Description The DAAP protocol was originally introduced in iTunes version 4.0. Initially, Apple did not officially release a protocol description, but it has been reverse-engineered to a sufficient degree that reimplementations of the protocol for non-iTunes platforms have been possible. A DAAP server is a specialized HTTP server, which performs two functions. It sends a list of songs and it streams requested songs to clients. There are also provisions to notify the client of changes to the server. Requests are sent to the server by the client in form of URLs and are responded to with data in mime-type, which can be converted to XML by the client. iTunes uses the zeroconf (also known as Bonjour) service to announce and discover DAAP shares on a local subnet. The DAAP service uses TCP port 3689 by default. DAAP is one of two media sharing schemes that Apple has currently released. The other, Digital Photo Access Protocol (DPAP), is used by iPhoto for sharing images. They both rely on an underlying protocol, Digital Media Access Protocol (DMAP). Early versions of iTunes allowed users to connect to shares across the Internet, however, in recent versions only computers on the same subnet can share music (workarounds such as port tunneling are possible). The Register speculates that Apple made this move in response to pressure from the record labels. More recent versions of iTunes also limit the number of clients to 5 unique IP addresses within a 24-hour period. DAAP has also been implemented in other non-iTunes media applications such as Banshee, Amarok, Exaile (with a plugin), Songbird (with a plugin), Rhythmbox, and WiFiTunes. DAAP authentication Beginning with iTunes 4.2, Appl
https://en.wikipedia.org/wiki/Cladonia%20alaskana
Cladonia alaskana, commonly known as the Alaskan cup lichen, is a species of cup lichen in the Cladoniaceae family. It is found in and around the arctic circle, growing in acidic soils. 40-80 mm tall and up to 2 mm in diameter. This species grows over boulders in heath and tussock tundras.
https://en.wikipedia.org/wiki/Eosinophiluria
Eosinophiluria is the abnormal presence of eosinophils in the urine. It can be measured by detecting levels of eosinophil cationic protein. Associated conditions It can be associated with a wide variety of conditions, including: Kidney disorders such as acute interstitial nephritis and acute kidney injury from cholesterol embolism Urinary tract infection Eosinophilic granulomatosis with polyangiitis Eosinophiluria (>5% of urine leukocytes ) is a common finding (~90%) in antibiotic induced allergic nephritis, however lymphocytes predominate in allergic interstitial nephritis induced by NSAIDs. Eosinophiluria is a feature of atheroembolic ARF. In PAN, microscopic polyangitis, eosinophiluria is rare.
https://en.wikipedia.org/wiki/Dynamic%20semantics
Dynamic semantics is a framework in logic and natural language semantics that treats the meaning of a sentence as its potential to update a context. In static semantics, knowing the meaning of a sentence amounts to knowing when it is true; in dynamic semantics, knowing the meaning of a sentence means knowing "the change it brings about in the information state of anyone who accepts the news conveyed by it." In dynamic semantics, sentences are mapped to functions called context change potentials, which take an input context and return an output context. Dynamic semantics was originally developed by Irene Heim and Hans Kamp in 1981 to model anaphora, but has since been applied widely to phenomena including presupposition, plurals, questions, discourse relations, and modality. Dynamics of anaphora The first systems of dynamic semantics were the closely related File Change Semantics and discourse representation theory, developed simultaneously and independently by Irene Heim and Hans Kamp. These systems were intended to capture donkey anaphora, which resists an elegant compositional treatment in classic approaches to semantics such as Montague grammar. Donkey anaphora is exemplified by the infamous donkey sentences, first noticed by the medieval logician Walter Burley and brought to modern attention by Peter Geach. Donkey sentence (relative clause): Every farmer who owns a donkey beats it. Donkey sentence (conditional): If a farmer owns a donkey, he beats it. To capture the empirically observed truth conditions of such sentences in first order logic, one would need to translate the indefinite noun phrase "a donkey" as a universal quantifier scoping over the variable corresponding to the pronoun "it". FOL translation of donkey sentence: : While this translation captures (or approximates) the truth conditions of the natural language sentences, its relationship to the syntactic form of the sentence is puzzling in two ways. First, indefinites in non-donkey context
https://en.wikipedia.org/wiki/Terpene%20synthase%20C%20terminal%20domain
In molecular biology, this protein domain belongs to the terpene synthase family (TPS). Its role is to synthesize terpenes, which are part of primary metabolism, such as sterols and carotene, and also part of the secondary metabolism. This entry will focus on the C terminal domain of the TPS protein. Function Terpenes synthases have a role in producing important molecules in metabolism, these molecules are part of a large group called terpenoids . In particular, the C terminal domain catalyzes the cyclization of geranyl diphosphate, orienting and stabilizing multiple reactive carbocation intermediates. Or in simpler terms, the C terminal aids the synthesis of new molecules. Structure It is thought to have at least two alpha helices. Conservation Sequences containing this protein domain belong to the terpene synthase family. It has been suggested that this gene family be designated tps (for terpene synthase). Sequence comparisons reveal similarities between the monoterpene (C10) synthases, sesquiterpene (C15) synthases and the diterpene (C20) synthases. It has been split into six subgroups on the basis of phylogeny, called Tpsa-Tpsf . Tpsa includes vetispiridiene synthase. Tpsb includes (-)-limonene synthase. Tpsc includes copalyl diphosphate synthase (kaurene synthase A). Tpsd includes taxadiene synthase, pinene synthase, and myrcene synthase. Tpse includes ent-kaurene synthase B. Tpsf includes linalool synthase. In the fungus Phaeosphaeria sp. (strain L487) the synthesis of ent-kaurene from geranylgeranyl dophosphate is promoted by a single bifunctional protein. See also Terpene synthase N terminal domain
https://en.wikipedia.org/wiki/Bond%20graph
A bond graph is a graphical representation of a physical dynamic system. It allows the conversion of the system into a state-space representation. It is similar to a block diagram or signal-flow graph, with the major difference that the arcs in bond graphs represent bi-directional exchange of physical energy, while those in block diagrams and signal-flow graphs represent uni-directional flow of information. Bond graphs are multi-energy domain (e.g. mechanical, electrical, hydraulic, etc.) and domain neutral. This means a bond graph can incorporate multiple domains seamlessly. The bond graph is composed of the "bonds" which link together "single-port", "double-port" and "multi-port" elements (see below for details). Each bond represents the instantaneous flow of energy () or power. The flow in each bond is denoted by a pair of variables called power variables, akin to conjugate variables, whose product is the instantaneous power of the bond. The power variables are broken into two parts: flow and effort. For example, for the bond of an electrical system, the flow is the current, while the effort is the voltage. By multiplying current and voltage in this example you can get the instantaneous power of the bond. A bond has two other features described briefly here, and discussed in more detail below. One is the "half-arrow" sign convention. This defines the assumed direction of positive energy flow. As with electrical circuit diagrams and free-body diagrams, the choice of positive direction is arbitrary, with the caveat that the analyst must be consistent throughout with the chosen definition. The other feature is the "causality". This is a vertical bar placed on only one end of the bond. It is not arbitrary. As described below, there are rules for assigning the proper causality to a given port, and rules for the precedence among ports. Causality explains the mathematical relationship between effort and flow. The positions of the causalities show which of the power va
https://en.wikipedia.org/wiki/Non-positive%20curvature
In mathematics, spaces of non-positive curvature occur in many contexts and form a generalization of hyperbolic geometry. In the category of Riemannian manifolds, one can consider the sectional curvature of the manifold and require that this curvature be everywhere less than or equal to zero. The notion of curvature extends to the category of geodesic metric spaces, where one can use comparison triangles to quantify the curvature of a space; in this context, non-positively curved spaces are known as (locally) CAT(0) spaces. Riemann Surfaces If is a closed, orientable Riemann surface then it follows from the Uniformization theorem that may be endowed with a complete Riemannian metric with constant Gaussian curvature of either , or . As a result of the Gauss–Bonnet theorem one can determine that the surfaces which have a Riemannian metric of constant curvature i.e. Riemann surfaces with a complete, Riemannian metric of non-positive constant curvature, are exactly those whose genus is at least . The Uniformization theorem and the Gauss–Bonnet theorem can both be applied to orientable Riemann surfaces with boundary to show that those surfaces which have a non-positive Euler characteristic are exactly those which admit a Riemannian metric of non-positive curvature. There is therefore an infinite family of homeomorphism types of such surfaces whereas the Riemann sphere is the only closed, orientable Riemann surface of constant Gaussian curvature . The definition of curvature above depends upon the existence of a Riemannian metric and therefore lies in the field of geometry. However the Gauss–Bonnet theorem ensures that the topology of a surface places constraints on the complete Riemannian metrics which may be imposed on a surface so the study of metric spaces of non-positive curvature is of vital interest in both the mathematical fields of geometry and topology. Classical examples of surfaces of non-positive curvature are the Euclidean plane and flat torus (for
https://en.wikipedia.org/wiki/Bucket%20argument
Isaac Newton's rotating bucket argument (also known as Newton's bucket) was designed to demonstrate that true rotational motion cannot be defined as the relative rotation of the body with respect to the immediately surrounding bodies. It is one of five arguments from the "properties, causes, and effects" of "true motion and rest" that support his contention that, in general, true motion and rest cannot be defined as special instances of motion or rest relative to other bodies, but instead can be defined only by reference to absolute space. Alternatively, these experiments provide an operational definition of what is meant by "absolute rotation", and do not pretend to address the question of "rotation relative to what?" General relativity dispenses with absolute space and with physics whose cause is external to the system, with the concept of geodesics of spacetime. Background These arguments, and a discussion of the distinctions between absolute and relative time, space, place and motion, appear in a scholium at the end of Definitions sections in Book I of Newton's work, The Mathematical Principles of Natural Philosophy (1687) (not to be confused with General Scholium at the end of Book III), which established the foundations of classical mechanics and introduced his law of universal gravitation, which yielded the first quantitatively adequate dynamical explanation of planetary motion. Despite their embrace of the principle of rectilinear inertia and the recognition of the kinematical relativity of apparent motion (which underlies whether the Ptolemaic or the Copernican system is correct), natural philosophers of the seventeenth century continued to consider true motion and rest as physically separate descriptors of an individual body. The dominant view Newton opposed was devised by René Descartes, and was supported (in part) by Gottfried Leibniz. It held that empty space is a metaphysical impossibility because space is nothing other than the extension of matte
https://en.wikipedia.org/wiki/Sabot%20%28firearms%29
A sabot (, ) is a supportive device used in firearm/artillery ammunitions to fit/patch around a projectile, such as a bullet/slug or a flechette-like projectile (such as a kinetic energy penetrator), and keep it aligned in the center of the barrel when fired. It allows a narrower projectile with high sectional density to be fired through a barrel of much larger bore diameter with maximal accelerative transfer of kinetic energy. After leaving the muzzle, the sabot typically separates from the projectile in flight, diverting only a very small portion of the overall kinetic energy. The sabot component in projectile design is the relatively thin, tough and deformable seal known as a driving band or obturation ring needed to trap propellant gases behind a projectile, and also keep the projectile centered in the barrel, when the outer shell of the projectile is only slightly smaller in diameter than the caliber of the barrel. Driving bands and obturators are used to seal these full-bore projectiles in the barrel because of manufacturing tolerances; there always exists some gap between the projectile outer diameter and the barrel inner diameter, usually a few thousandths of an inch; enough of a gap for high pressure gasses to slip by during firing. Driving bands and obturator rings are made from material that will deform and seal the barrel as the projectile is forced from the chamber into the barrel. Sabots use driving bands and obturators, because the same manufacturing tolerance issues exist when sealing the saboted projectile in the barrel, but the sabot itself is a more substantial structural component of the in-bore projectile configuration. Refer to the two armor-piercing fin-stabilized discarding sabot (APFSDS) pictures on the right to see the substantial material nature of a sabot to fill the bore diameter around the sub-caliber arrow-type flight projectile, compared to the very small gap sealed by a driving band or obturator to mitigate what is known classical
https://en.wikipedia.org/wiki/Sherwood%20conferences
The Sherwood Conferences were a series of classified conferences that were held between 1952 and 1958 in the United States. These conferences were a part of the United States controlled nuclear program called Project Sherwood. These conferences were established in order to entice experienced personnel to join the newly developed Project Sherwood. There were three different plasma confinement designs that were being researched in three different locations: the stellarator at Princeton Plasma Physics Laboratory, the toroidal pinch at Los Alamos National Laboratory, and the magnetic mirror at the Livermore National Laboratory. Because these individual projects operated in separate facilities, these conferences were helpful to strengthen communication of information between all three projects. The First Conference The first Sherwood conference was organized by the AEC director of the Division of Research, Thomas Johnson. The conference was held at the University of Denver in Denver, Colorado on June 28, 1952. To attract more experienced personnel, workers already participating on the projects attended the conference. Overall, there were about eighty scientists that attended the first conference. Dates and Locations of the Sherwood Conferences
https://en.wikipedia.org/wiki/Digital%20Compression%20System
Digital Compression System, or DCS, is a sound system developed by Williams Electronics. This advanced sound board was used in Williams and Bally pinball games, coin-op arcade video games by Midway Manufacturing, and mechanical and video slot machines by Williams Gaming. This sound system became the standard for these game platforms. The DCS Sound system was created by Williams sound engineers Matt Booty and Ed Keenan, and further developed by Andrew Eloff. Versions of DCS DCS ROM-based mono: The first version of DCS used an Analog Devices ADSP2105 DSP (clocked at 10 MHz) and a DMA-driven DAC, outputting in mono. This was used for the majority of Williams and Midway's pinball games (starting with 1993's Indiana Jones: The Pinball Adventure), as well as Midway's video games, up until the late 1990s. The pinball game, The Twilight Zone, was originally supposed to use the DCS System, but because the DCS board was still in development at the time, all of the music and sounds for this game were reprogrammed for the Yamaha YM2151 / Harris CVSD sound board. DCS-95: This was a revised version of the original DCS System (allowing for 16MB of data instead of 8MB to be addressed), used for Williams and Midway's WPC-95 pinball system. DCS2 ROM-based stereo: This version used the ADSP2104 DSP and two DMA-driven DACs, outputting in stereo. This was used in Midway's Zeus-based hardware, and in the short-lived Pinball 2000 platform. DCS2 RAM-based stereo: This version used the ADSP2115 DSP and two DMA-driven DACs, outputting in stereo. This was used in Midway's 3DFX-based hardware (NFL Blitz, etc.). This system would be adopted by Atari Games, following their acquisition by WMS Industries. DCS2 RAM-based multi-channel: This version used the ADSP2181 DSP and up to six DMA-driven DACs, outputting in multichannel sound. Pinball games using DCS Attack From Mars (1995) (DCS95) Cactus Canyon (1998) (DCS95) The Champion Pub (1998) (DCS95) Cirqus Voltaire (1997) (DCS95) Congo (1995) (DC
https://en.wikipedia.org/wiki/Unitary%20matrix
In linear algebra, an invertible complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose is referred to as the Hermitian adjoint of a matrix and is denoted by a dagger (†), so the equation above is written For real numbers, the analogue of a unitary matrix is an orthogonal matrix. Unitary matrices have significant importance in quantum mechanics because they preserve norms, and thus, probability amplitudes. Properties For any unitary matrix of finite size, the following hold: Given two complex vectors and , multiplication by preserves their inner product; that is, . is normal (). is diagonalizable; that is, is unitarily similar to a diagonal matrix, as a consequence of the spectral theorem. Thus, has a decomposition of the form where is unitary, and is diagonal and unitary. . That is, will be on the unit circle of the complex plane. Its eigenspaces are orthogonal. can be written as , where indicates the matrix exponential, is the imaginary unit, and is a Hermitian matrix. For any nonnegative integer , the set of all unitary matrices with matrix multiplication forms a group, called the unitary group . Any square matrix with unit Euclidean norm is the average of two unitary matrices. Equivalent conditions If U is a square, complex matrix, then the following conditions are equivalent: is unitary. is unitary. is invertible with . The columns of form an orthonormal basis of with respect to the usual inner product. In other words, . The rows of form an orthonormal basis of with respect to the usual inner product. In other words, . is an isometry with respect to the usual norm. That is, for all , where . is a normal matrix (equivalently, there is an orthonormal basis formed by eigenvectors of ) with eigenvalues lying on the unit circle. Elementary constructions 2 × 2 unitary matrix On
https://en.wikipedia.org/wiki/Adolph%20Winkler%20Goodman
Adolph Winkler Goodman (July 20, 1915 – July 30, 2004) was an American mathematician who contributed to number theory, graph theory and to the theory of univalent functions: The conjecture on the coefficients of multivalent functions named after him is considered the most interesting challenge in the area after the Bieberbach conjecture, proved by Louis de Branges in 1985. Life and work In 1948, he made a mathematical conjecture on coefficients of -valent functions, first published in his Columbia University dissertation thesis and then in a closely following paper. After the proof of the Bieberbach conjecture by Louis de Branges, this conjecture is considered the most interesting challenge in the field, and he himself and coauthors answered affirmatively to the conjecture for some classes of -valent functions. His researches in the field continued in the paper Univalent functions and nonanalytic curves, published in 1957: in 1968, he published the survey Open problems on univalent and multivalent functions, which eventually led him to write the two-volume book Univalent Functions. Apart from his research activity, He was actively involved in teaching: he wrote several college and high school textbooks including Analytic Geometry and the Calculus, and the five-volume set Algebra from A to Z. He retired in 1993, became a Distinguished Professor Emeritus in 1995, and died in 2004. Selected works Notes Biographical references
https://en.wikipedia.org/wiki/Lymphatic%20pump
The lymphatic pump is a method of manipulation used by physicians who practice manual medicine (primarily osteopathic physicians). Manual lymphatic drainage techniques remain a clinical art founded upon hypotheses, theory, and preliminary evidence. History The term lymphatic pump was invented by Earl Miller, D.O. to describe what was formerly known in osteopathic medicine as the thoracic pump technique. Technique The technique is applied to a person lying down by holding their ankle and applying gentle pressure repeatedly using the leg as a "lever" to rock the pelvis. Relative contraindications While no firmly established absolute contraindications exist for lymphatic techniques, the following cases are examples of relative contraindications: bone fractures, bacterial infections with fever, abscesses, and cancer.
https://en.wikipedia.org/wiki/Evolution%20of%20ageing
Enquiry into the evolution of ageing, or aging, aims to explain why a detrimental process such as ageing would evolve, and why there is so much variability in the lifespans of organisms. The classical theories of evolution (mutation accumulation, antagonistic pleiotropy, and disposable soma) suggest that environmental factors, such as predation, accidents, disease, and/or starvation, ensure that most organisms living in natural settings will not live until old age, and so there will be very little pressure to conserve genetic changes that increase longevity. Natural selection will instead strongly favor genes which ensure early maturation and rapid reproduction, and the selection for genetic traits which promote molecular and cellular self-maintenance will decline with age for most organisms. Theories and hypotheses The beginning August Weismann was responsible for interpreting and formalizing the mechanisms of Darwinian evolution in a modern theoretical framework. In 1889, he theorized that ageing was part of life's program to make room for the next generation in order to sustain the turnover that is necessary for evolution. The idea that the ageing characteristic was selected (an adaptation) because of its deleterious effect was largely discounted for much of the 20th century, but a theoretical model suggests that altruistic ageing could evolve if there is little migration among populations. Weismann later abandoned his theory and after some time followed up with his "programmed death" theory. Natural selection is a process that allows organisms to better adapt to the environment, it is the survival of the fittest which are predicted to produce more offsprings. Natural selection acts on life history traits in order to optimize reproductive success and lifetime fitness. Fitness in this context refers to how likely an organism is to survive and reproduce. It is based on the environment and is also relative to other individuals in the population. Examples of lif
https://en.wikipedia.org/wiki/1960%E2%80%9361%20Silver%20Hut%20expedition
The 1960–61 Silver Hut expedition or formally the Himalayan Scientific and Mountaineering Expedition was initiated by Edmund Hillary and Griffith Pugh with John Dienhart of World Books in America (producers of a children’s encyclopaedia). The expedition lasted from September 1960 to June 1961. In 1958 Hillary and Pugh had discussed whether Everest could be climbed without oxygen; with improved acclimatising by wintering at say for six months beforehand. But Pugh’s plans involving two bases on Everest (Base camp, and on the Western Cwm at feet) had been dropped by Hillary as a request to the Chinese had been rebuffed because of troubles in Tibet. And finance was needed; Hillary wrote to Pugh in 1959 "I’m damn certain that we’d get someone on the top (of Everest) without oxygen but we’d need a lot of cash". In 1959 Hillary was awarded the Explorer of the Year Award by Argosy magazine; $US1000 and a trip to New York to address the award banquet. His speech and personality impressed Dienhart who invited him to their Chicago headquarters. Hillary proposed a "Yeti search" plus a party of climbers who would winter for the first time at () and then attempt the summit of Makalu () without oxygen. Hillary estimated the expedition cost at $US120,000 and after meeting him in Chicago in October 1959 the World Book board gave him $US125,000 and a "practically free hand". Nepal and the Silver Hut base The expedition gathered in Kathmandu in September 1960, destination the Rolwaling Valley; reputed to be a Yeti stronghold and where Shipton with Ward and the Sherpa Sen Tenzing had photographed the footprint in 1951. At the end of October the expedition went via the Tashi Laptsa Pass to the Khumbu region. Norman Hardie led a party of 310 Nepali porters with parts of the hut, the laboratory equipment and winter supplies. He set up a base at the village of Changmatang near the entrance to the Mingbo Valley. Research accommodation was in the prefabricated Silver Hut and an adj
https://en.wikipedia.org/wiki/Quotient
In arithmetic, a quotient (from 'how many times', pronounced ) is a quantity produced by the division of two numbers. The quotient has widespread use throughout mathematics. It has two definitions: either the integer part of a division (in the case of Euclidean division), or as a fraction or a ratio (in the case of a general division). For example, when dividing 20 (the dividend) by 3 (the divisor), the quotient is 6 (with a remainder of 2) in the first sense, and (a repeating decimal) in the second sense. In metrology (International System of Quantities and the International System of Units), "quotient" refers to the general case with respect to the units of measurement of physical quantities. Ratios is the special case for dimensionless quotients of two quantities of the same kind. Quotients with a non-trivial dimension and compound units, especially when the divisor is a duration (e.g., "per second"), are known as rates. For example, density (mass divided by volume, in units of kg/m3) is said to be a "quotient", whereas mass fraction (mass divided by mass, in kg/kg or in percent) is a "ratio". Specific quantities are intensive quantities resulting from the quotient of a physical quantity by mass, volume, or other measures of the system "size". Notation The quotient is most frequently encountered as two numbers, or two variables, divided by a horizontal line. The words "dividend" and "divisor" refer to each individual part, while the word "quotient" refers to the whole. Integer part definition The quotient is also less commonly defined as the greatest whole number of times a divisor may be subtracted from a dividend—before making the remainder negative. For example, the divisor 3 may be subtracted up to 6 times from the dividend 20, before the remainder becomes negative: 20 − 3 − 3 − 3 − 3 − 3 − 3 ≥ 0, while 20 − 3 − 3 − 3 − 3 − 3 − 3 − 3 < 0. In this sense, a quotient is the integer part of the ratio of two numbers. Quotient of two integers A rati
https://en.wikipedia.org/wiki/Taxonomic%20rank
In biology, taxonomic rank is the relative level of a group of organisms (a taxon) in an ancestral or hereditary hierarchy. A common system of biological classification (taxonomy) consists of species, genus, family, order, class, phylum, kingdom, and domain. While older approaches to taxonomic classification were phenomenological, forming groups on the basis of similarities in appearance, organic structure and behaviour, methods based on genetic analysis have opened the road to cladistics. A given rank subsumes less general categories under it, that is, more specific descriptions of life forms. Above it, each rank is classified within more general categories of organisms and groups of organisms related to each other through inheritance of traits or features from common ancestors. The rank of any species and the description of its genus is basic; which means that to identify a particular organism, it is usually not necessary to specify ranks other than these first two. Consider a particular species, the red fox, Vulpes vulpes: the specific name or specific epithet vulpes (small v) identifies a particular species in the genus Vulpes (capital V) which comprises all the "true" foxes. Their close relatives are all in the family Canidae, which includes dogs, wolves, jackals, and all foxes; the next higher major rank, the order Carnivora, includes caniforms (bears, seals, weasels, skunks, raccoons and all those mentioned above), and feliforms (cats, civets, hyenas, mongooses). Carnivorans are one group of the hairy, warm-blooded, nursing members of the class Mammalia, which are classified among animals with backbones in the phylum Chordata, and with them among all animals in the kingdom Animalia. Finally, at the highest rank all of these are grouped together with all other organisms possessing cell nuclei in the domain Eukarya. The International Code of Zoological Nomenclature defines rank as: "The level, for nomenclatural purposes, of a taxon in a taxonomic hierarchy (
https://en.wikipedia.org/wiki/Consumer%20Electronics%20Control
Consumer Electronics Control (CEC) is a feature of HDMI designed to control HDMI connected devices by using only one remote controller; so, individual CEC enabled devices can command and control each other without user intervention, for up to 15 devices. For example, a television set remote controller can also control a set-top box and a DVD player. It is a one-wire bidirectional serial bus that is based on the CENELEC standard AV.link protocol to perform remote control functions. CEC wiring is mandatory, although implementation of CEC in a product is optional. It was defined in HDMI Specification 1.0 and updated in HDMI 1.2, HDMI 1.2a and HDMI 1.3a (which added timer and audio commands to the bus). USB-to-CEC adapters exist that allow a computer to control CEC-enabled devices. Trade names for CEC technology Trade names for CEC include: 1-Touch Play (Roku) Anynet+ (Samsung) Aquos Link (Sharp) BRAVIA Link, BRAVIA Sync, Control for HDMI (Sony) CEC (Hisense, Vizio) CE-Link, Regza Link (Toshiba) E-link (AOC) EasyLink (Philips) Fun-Link (Funai, Sylvania, Emerson, Magnavox, Philips) HDMI-CEC (Hitachi) INlink (Insignia, Westinghouse) Kuro Link (Pioneer) NetCommand for HDMI, Realink for HDMI (Mitsubishi) RIHD (Remote Interactive over HDMI) (Onkyo) RuncoLink (Runco International) SimpLink (LG) T-Link (ITT, Thomson) VIERA Link, HDAVI Control, EZ-Sync (Panasonic) CEC commands The following is a list of the most commonly used HDMI-CEC commands: One Touch Play allows devices to switch the TV to use it as the active source when playback starts System Standby enables users to switch multiple devices to standby mode with the press of one button Preset Transfer transfers the tuner channel setup to another TV set One Touch Record allows users to record whatever is currently being shown on the HDTV screen on a selected recording device Timer Programming allows users to use the electronic program guides (EPGs) that are built into many HDTVs and set-top-boxes
https://en.wikipedia.org/wiki/Dental%20abrasion
Abrasion is the non-carious, mechanical wear of tooth from interaction with objects other than tooth-tooth contact. It most commonly affects the premolars and canines, usually along the cervical margins. Based on clinical surveys, studies have shown that abrasion is the most common but not the sole aetiological factor for development of non-carious cervical lesions (NCCL) and is most frequently caused by incorrect toothbrushing technique. Abrasion frequently presents at the cemento-enamel junction and can be caused by many contributing factors, all with the ability to affect the tooth surface in varying degrees. The appearance may vary depending on the cause of abrasion, however most commonly presents in a V-shaped caused by excessive lateral pressure whilst tooth-brushing. The surface is shiny rather than carious, and sometimes the ridge is deep enough to see the pulp chamber within the tooth itself. Non-carious cervical loss due to abrasion may lead to consequences and symptoms such as increased tooth sensitivity to hot and cold, increased plaque trapping which will result in caries and periodontal disease, and difficulty of dental appliances such as retainers or dentures engaging the tooth. It may also be aesthetically unpleasant to some people. For successful treatment of abrasion, the cause first needs to be identified and ceased (e.g. overzealous brushing). Once this has occurred, subsequent treatment may involve the changes in oral hygiene, application of fluoride to reduce sensitivity, or the placement of a restoration to help prevent further loss of tooth structure and aid plaque control. Cause Cause of abrasion may arise from interaction of teeth with other objects such as toothbrushes, toothpicks, floss, and ill-fitting dental appliance like retainers and dentures. Apart from that, people with habits such as nail biting, chewing tobacco, lip or tongue piercing, and having occupation such as joiner, are subjected to higher risks of abrasion. The ae
https://en.wikipedia.org/wiki/ViralZone
ViralZone is a Swiss Institute of Bioinformatics web-resource for all viral genus and families, providing general molecular and epidemiological information, along with virion and genome figures.
https://en.wikipedia.org/wiki/James%20P.%20Muldoon%20River%20Center
The James P. Muldoon River Center, located in Saint Mary's City, Maryland, is a 13,000 square foot eco-friendly facility of the St. Mary's College of Maryland. The facility houses a marine biology laboratory classroom, a seminar room, boat repair facilities, a safety monitor station, water sports equipment storage, small locker rooms, offices for waterfront staff and biology faculty, a multi-purpose room, instructional space for water-based activities, and a deck that overlooks the waterfront basin and Horseshoe Bend on the St. Mary’s River. It is the home venue of the St. Mary's Seahawks sailing team. St. Mary's River Project The St. Mary's River Project, which is a part of the Muldoon River Center, is a state and federally funded marine biology research program, administered through St. Mary's College of Maryland. Using the St. Mary's River and other nearby waters as field laboratories, the project investigates and monitors the water quality and the ecological health of both the St. Mary's River and the Chesapeake Bay. The project also promotes environment awareness and stewardship in Chesapeake Bay communities. Students work and study in all aspects of the programs activities, including classroom and hands-on field and laboratory learning. The laboratories and offices of the project are located in the Muldoon River Center, a geothermally heated and cooled building on the campus waterfront. Geothermal project at the Muldoon River Center The Muldoon River Center has a geothermal heating and cooling system, with special pipes running 300 feet down into the ground, to tap the deep soil's energy management potential. The system cools the building in the summer and warms it in the wintertime with an extremely low impact on the environment.
https://en.wikipedia.org/wiki/Partition%20of%20a%20set
In mathematics, a partition of a set is a grouping of its elements into non-empty subsets, in such a way that every element is included in exactly one subset. Every equivalence relation on a set defines a partition of this set, and every partition defines an equivalence relation. A set equipped with an equivalence relation or a partition is sometimes called a setoid, typically in type theory and proof theory. Definition and notation A partition of a set X is a set of non-empty subsets of X such that every element x in X is in exactly one of these subsets (i.e., the subsets are nonempty mutually disjoint sets). Equivalently, a family of sets P is a partition of X if and only if all of the following conditions hold: The family P does not contain the empty set (that is ). The union of the sets in P is equal to X (that is ). The sets in P are said to exhaust or cover X. See also collectively exhaustive events and cover (topology). The intersection of any two distinct sets in P is empty (that is ). The elements of P are said to be pairwise disjoint or mutually exclusive. See also mutual exclusivity. The sets in are called the blocks, parts, or cells, of the partition. If then we represent the cell containing by . That is to say, is notation for the cell in which contains . Every partition may be identified with an equivalence relation on , namely the relation such that for any we have if and only if (equivalently, if and only if ). The notation evokes the idea that the equivalence relation may be constructed from the partition. Conversely every equivalence relation may be identified with a partition. This is why it is sometimes said informally that "an equivalence relation is the same as a partition". If P is the partition identified with a given equivalence relation , then some authors write . This notation is suggestive of the idea that the partition is the set X divided in to cells. The notation also evokes the idea that, from the equivale
https://en.wikipedia.org/wiki/Syra%20Madad
Syra Madad (, née Sikandar; born October 22, 1986) is an American pathogen preparedness expert and infectious disease epidemiologist. Madad is the Senior Director of the System-wide Special Pathogens Program at NYC Health + Hospitals where she is part of the executive leadership team which oversees New York City's response to the Coronavirus disease 2019 pandemic in the city's 11 public hospitals. She was featured in the Netflix documentary series Pandemic: How to Prevent an Outbreak and the Discovery Channel documentary The Vaccine: Conquering COVID. Early life and education Madad was born in the United States. Her family is of Pakistani background. Her mother and father married and then immigrated to the United States from Pakistan; her mother, Rehana Sikandar, was 18 at the time. Madad's interest in public health and infectious disease began at a young age, including after she saw the 1995 film Outbreak. In 2008, Madad received a B.S. in Psychology from the University of Maryland, College Park. In 2010, she received a master's degree in Biotechnology with a concentration in Biodefense and Biosecurity, also from the University of Maryland, College Park. In 2014, Madad received an DHSc degree in health science with a concentration in Global Health from Nova Southeastern University. Career In 2014, Madad was Lead Continuity of Operations Liaison and State Trainer for the BioThreat and Chemical Threat Teams in the Emergency Preparedness Branch at the Texas Department of State Health Services. In this position she worked in the Ebola and Other Infectious Disease Agent Surge Team in response to a 2014-2015 Ebola outbreak in Texas. During this time, Madad volunteered at the Texas State Medical Operations Center as a Planning and Intelligence Specialist. She also volunteered as Logistics Specialist and Trauma Medical Responder at Sector 7 of the Texas Emergency Medical Task Force. In 2015, Madad was hired as Senior Director of the System-wide Special Pathogens Progr
https://en.wikipedia.org/wiki/Compressive%20strength
In mechanics, compressive strength (or compression strength) is the capacity of a material or structure to withstand loads tending to reduce size (as opposed to tensile strength which withstands loads tending to elongate). In other words, compressive strength resists compression (being pushed together), whereas tensile strength resists tension (being pulled apart). In the study of strength of materials, tensile strength, compressive strength, and shear strength can be analyzed independently. Some materials fracture at their compressive strength limit; others deform irreversibly, so a given amount of deformation may be considered as the limit for compressive load. Compressive strength is a key value for design of structures. Compressive strength is often measured on a universal testing machine. Measurements of compressive strength are affected by the specific test method and conditions of measurement. Compressive strengths are usually reported in relationship to a specific technical standard. Introduction When a specimen of material is loaded in such a way that it extends it is said to be in tension. On the other hand, if the material compresses and shortens it is said to be in compression. On an atomic level, the molecules or atoms are forced apart when in tension whereas in compression they are forced together. Since atoms in solids always try to find an equilibrium position, and distance between other atoms, forces arise throughout the entire material which oppose both tension or compression. The phenomena prevailing on an atomic level are therefore similar. The "strain" is the relative change in length under applied stress; positive strain characterizes an object under tension load which tends to lengthen it, and a compressive stress that shortens an object gives negative strain. Tension tends to pull small sideways deflections back into alignment, while compression tends to amplify such deflection into buckling. Compressive strength is measured on materi
https://en.wikipedia.org/wiki/Papain-like%20protease
Papain-like proteases (or papain-like (cysteine) peptidases; abbreviated PLP or PLCP) are a large protein family of cysteine protease enzymes that share structural and enzymatic properties with the group's namesake member, papain. They are found in all domains of life. In animals, the group is often known as cysteine cathepsins or, in older literature, lysosomal peptidases. In the MEROPS protease enzyme classification system, papain-like proteases form Clan CA. Papain-like proteases share a common catalytic dyad active site featuring a cysteine amino acid residue that acts as a nucleophile. The human genome encodes eleven cysteine cathepsins which have a broad range of physiological functions. In some parasites papain-like proteases have roles in host invasion, such as cruzipain from Trypanosoma cruzi. In plants, they are involved in host defense and in development. Studies of papain-like proteases from prokaryotes have lagged their eukaryotic counterparts. In cellular organisms they are synthesized as preproenzymes that are not enzymatically active until mature, and their activities are tightly regulated, often by the presence of endogenous protease inhibitors such as cystatins. In many RNA viruses, including significant human pathogens such as the coronaviruses SARS-CoV and SARS-CoV-2, papain-like protease protein domains often have roles in processing of polyproteins into mature viral nonstructural proteins. Many papain-like proteases are considered potential drug targets. Classification The MEROPS system of protease enzyme classification defines clan CA as containing the papain-like proteases. They are thought to have a shared evolutionary origin. As of 2021, the clan contained 45 families. Structure The structure of papain was among the earliest protein structures experimentally determined by X-ray crystallography. Many papain-like protease enzymes function as monomers, though a few, such as cathepsin C (Dipeptidyl-peptidase I), are homotetramers. The matu
https://en.wikipedia.org/wiki/Picone%20identity
In the field of ordinary differential equations, the Picone identity, named after Mauro Picone, is a classical result about homogeneous linear second order differential equations. Since its inception in 1910 it has been used with tremendous success in association with an almost immediate proof of the Sturm comparison theorem, a theorem whose proof took up many pages in Sturm's original memoir of 1836. It is also useful in studying the oscillation of such equations and has been generalized to other type of differential equations and difference equations. The Picone identity is used to prove the Sturm–Picone comparison theorem. Picone identity Suppose that u and v are solutions of the two homogeneous linear second order differential equations in self-adjoint form and Then, for all x with v(x) ≠ 0, the following identity holds Proof Notes
https://en.wikipedia.org/wiki/Robert%20R.%20Jensen
Robert Ronald Jensen (born 6 April 1949) is an American mathematician, specializing in nonlinear partial differential equations with applications to physics, engineering, game theory, and finance. Jensen graduated in 1971 with B.S. in mathematics from Illinois Institute of Technology. He received in 1975 his Ph.D. from Northwestern University with thesis Finite difference approximation to the free boundary of a parabolic variational inequality under the supervision of Avner Friedman. Jensen was from 1975 to 1977 an assistant professor at the University of California, Los Angeles and from 1977 to 1980 a visiting assistant professor at the University of Wisconsin's Mathematics Research Center. At the University of Kentucky he was from 1977 to 1980 an assistant professor and from 1980 to 1987 an associate professor. At Loyola University Chicago he was from 1985 to 1986 a visiting associate professor and is since 1986 a full professor. At Loyola he was from 2007 to 2012 the chair of the department of mathematics and statistics. From 1982 to 1986 Jensen held a Sloan Fellowship. He was a visiting member of Berkeley's Mathematical Sciences Research Institute in 1992, 2005, and 2013. He has given invited talks at universities and conferences around the world. In 1998 he was an Invited Speaker at the International Congress of Mathematicians in Berlin. Selected publications
https://en.wikipedia.org/wiki/Regius%20Professor%20of%20Mathematics
The Regius Professorship of Mathematics is the name given to three chairs in mathematics at British universities, one at the University of St Andrews, founded by Charles II in 1668, the second one at the University of Warwick, founded in 2013 to commemorate the Diamond Jubilee of Elizabeth II and the third one at the University of Oxford, founded in 2016. University of St Andrews (1668) From 1997 to 2015 there was no Regius Professor of Mathematics. In April 2013 the post was advertised, and in 2015 Igor Rivin was appointed. He was succeeded by Kenneth Falconer in 2017. List of Regius Professors of Mathematics The following list may be incomplete. 1668–1674 James Gregory 1674–1688 William Sanders 1689–1690 James Fenton 1690–1707 vacant 1707–1739 Charles Gregory 1739–1765 David Gregory 1765–1807 Nicolas Vilant 1807–1809 vacant 1809–1820 Robert Haldane 1820–1858 Thomas Duncan 1857–1858 John Couch Adams 1859–1877 William L F Fischer 1877–1879 George Chrystal 1879–1921 Peter Redford Scott Lang 1921–1950 Herbert Westren Turnbull 1950–1969 Edward Thomas Copson 1970–1997 John Mackintosh Howie 1997–2015 vacant 2015–2017 Igor Rivin 2017–present Kenneth Falconer University of Warwick (2013) The creation of the post of the Regius Professor of Mathematics was announced in January 2013, in March 2014 Martin Hairer was appointed to the position. University of Oxford (2016) The creation of the post of the Regius Professor of Mathematics was announced in June 2016 and Andrew Wiles was appointed as the first holder of the chair in May 2018. In August 2020, it was announced that the Regius Professorship in Mathematics at the University of Oxford will become a permanent fixture at Merton College.
https://en.wikipedia.org/wiki/Fecal%20incontinence
Fecal incontinence (FI), or in some forms encopresis, is a lack of control over defecation, leading to involuntary loss of bowel contents, both liquid stool elements and mucus, or solid feces. When this loss includes flatus (gas), it is referred to as anal incontinence. FI is a sign or a symptom, not a diagnosis. Incontinence can result from different causes and might occur with either constipation or diarrhea. Continence is maintained by several interrelated factors, including the anal sampling mechanism, and incontinence usually results from a deficiency of multiple mechanisms. The most common causes are thought to be immediate or delayed damage from childbirth, complications from prior anorectal surgery (especially involving the anal sphincters or hemorrhoidal vascular cushions), altered bowel habits (e.g., caused by irritable bowel syndrome, Crohn's disease, ulcerative colitis, food intolerance, or constipation with overflow incontinence). An estimated 2.2% of community-dwelling adults are affected. However, reported prevalence figures vary. A prevalence of 8.39% among non-institutionalized U.S adults between 2005 and 2010 has been reported, and among institutionalized elders figures come close to 50%. Fecal incontinence has three main consequences: local reactions of the perianal skin and urinary tract, including maceration (softening and whitening of the skin due to continuous moisture), urinary tract infections, or decubitus ulcers (pressure sores); a financial expense for individuals (due to the cost of medication and incontinence products, and loss of productivity), employers (days off), and medical insurers and society generally (health care costs, unemployment); and an associated decrease in quality of life. There is often reduced self-esteem, shame, humiliation, depression, a need to organize life around easy access to a toilet, and avoidance of enjoyable activities. FI is an example of a stigmatized medical condition, which creates barriers to successf
https://en.wikipedia.org/wiki/Richter%20tuning
Richter tuning is a system of choosing the reeds for a diatonic wind instrument (such as a harmonica or accordion). It is named after Joseph Richter, a Bohemian instrument maker who adopted the tuning for his harmonicas in the early 19th century and is credited with inventing the blow/draw mechanism that allows the harmonica to play different notes when the air is drawn instead of blown. Richter tuning is designed as a compromise between diatonic melody and harmony. The lower portion of the harmonica is designed to play the tonic and dominant chords on the blow and draw respectively (in the key of C, this would be the C major and G major chords). The remainder of the instrument is tuned to, in this example, blow entirely in the key of C major, with each successive note following the sequence C E G and the four notes not in the C major chord arranged on the draw in the sequence D F A B. For example: {| class="wikitable" style="text-align:center" ! hole ||  1 ||  2 ||  3 ||  4 ||  5 ||  6 ||  7 ||  8 ||  9 || 10 || 11 || 12 || 13 || 14 || 15 || 16 |- ! blow note || G || C || E || G || C || E || G || C || E || G || C || E || G || C || E || G |- ! draw note || B || D || G || B || D || F || A || B || D || F || A || B || D || F || A || B |} The above diagram shows that Richter tuning has some missing notes, notably A and F are absent from the lowest octave (draw notes in holes 3 and 4). By comparison, solo tuning includes all the major scale notes (C D E F G A B C) for all octaves. Variants There have been many variants of Richter tuning. Country tuning Country tuning raises the initial F by a semitone to an F. This primarily aids in harmony in the key of G, facilitating easy play of the G-C-D (I-IV-V) chord progression, while maintaining a partial G7 (minus the root) higher on the harmonica. It can occasionally be helpful in some melodies, most notably "The Star-Spangled Banner," which has a low F# and a high F. For example: {| class="wikitable" style="
https://en.wikipedia.org/wiki/Ivar%20Jacobson
Ivar Hjalmar Jacobson (born 1939) is a Swedish computer scientist and software engineer, known as major contributor to UML, Objectory, Rational Unified Process (RUP), aspect-oriented software development and Essence. Biography Ivar Jacobson was born in Ystad, Sweden, on September 2, 1939. He received his Master of Electrical Engineering degree at Chalmers Institute of Technology in Gothenburg in 1962. After his work at Ericsson, he formalized the language and method he had been working on in his PhD at the Royal Institute of Technology in Stockholm in 1985 on the thesis Language Constructs for Large Real Time Systems. After his master's degree, Jacobson joined Ericsson and worked in R&D on computerized switching systems AKE and AXE including PLEX. After his PhD thesis in April 1987, he started Objective Systems with Ericsson as a major customer. A majority stake of the company was acquired by Ericsson in 1991, and the company was renamed Objectory AB. Jacobson developed the software method Object-Oriented Software Engineering (OOSE) published 1992, which was a simplified version of the commercial software process Objectory (short for Object Factory). In October, 1995, Ericsson divested Objectory to Rational Software and Jacobson started working with Grady Booch and James Rumbaugh, known collectively as the Three Amigos. When IBM bought Rational in 2003, Jacobson decided to leave, after he stayed on until May 2004 as an executive technical consultant. In mid-2003 Jacobson formed Ivar Jacobson International (IJI) which operates across three continents with offices in the UK, the US, Sweden, Switzerland, China, and Singapore. Work Ericsson In 1967 at Ericsson, Jacobson proposed the use of software components in the new generation of software controlled telephone switches Ericsson was developing. In doing this he invented sequence diagrams, and developed collaboration diagrams. He also used state transition diagrams to describe the message flows between comp
https://en.wikipedia.org/wiki/ICES%20Statistical%20Rectangles
ICES Statistical Rectangles (aka "ICES Rectangles") is a gridded, latitude-longitude based area notation system covering the north-east Atlantic region developed by the International Council for the Exploration of the Sea (ICES) in the 1970s, for simplified analysis and visualization of spatial data of relevance to that organization's interests. The individual rectangles that make up the system each measure 1 degree of longitude by 0.5 degrees of latitude and are intended to be roughly square in real world use in the ICES region of interest, approximately 30 nautical miles by 30 nautical miles (55×55 km) at 60°N, although the actual width varies with latitude, gradually becoming wider than they are high south of 60°N, and narrower further north. The grid covers the region from 36°N to 85°30'N and from 44°W to 69°E (quoted as 68°30') using a set of alphanumeric identifiers, with row of latitude (identifiers 01 through 99, from south to north) cited first, then column of longitude (identifiers A0-A3, then B0-B9, C0-C9 etc., from west to east). The last used column identifier is M8; column identifiers A4-A9, and prefix "I" (uppercase "i") i.e. columns "I"0-"I"9 are not used. The resulting grid is 113 columns by 99 rows, comprising 11,187 labelled 1×0.5 degree cells. An example cell designation is 37F3, which designates the 1×0.5 degree rectangle of which the south-west corner is 54°00'N, 03°00'E. The grid covers both land and sea areas across its designated region, but as per the interests of its originating body, is typically employed for use with marine data such as analysis of marine resources, fishing activities, seabed habitat, etc., refer example references below. The full extent of the grid is visible in published figures such as Figs. 5-8 in Williamson et al., 2017. To accommodate the visualization and/or analysis of finer scale data than is supported by the "standard" (1×0.5 degree) rectangles, an optional fifth character (as digits 1 through 9) can be used w
https://en.wikipedia.org/wiki/Macronovirus
Macronovirus is the only genus of the family Sarthroviridae and only contains the species Macrobrachium satellite virus 1 It is found in The French West Indies, Thailand, Taiwan, China, and India. Etymology The genus name, Macronovirus, is a combination of Macro, from type species host Macrobrachium rosenbergii and no, from helper virus nodavirus. The family name, Sarthroviridae, is a combination of S, from Small and arthro, from host arthropoda. Hosts Macronovirus'''s cell tropism is muscle and connective cells of diseased animals, and its natural hosts are arthropods Structure The virion Macrobrachium satellite virus 1 has a genome consisting of linear single-stranded RNA of positive polarity, 0.8kb in size, with two genes. This encodes two capsid proteins, CP-17 and CP-16. The virion is non-enveloped, spherical, with a capsid of about 15 nm with icosahedral symmetry. The virion is constructed from two capsid proteins CP-17 and CP-16. It has a Monopartite, linear, ssRNA(+) genome. Gene expression The virion RNA is infectious and serves as both the genome and viral messenger RNA. Replication Its replication is cytoplasmic, and has 8 steps. Attachement to host receptors mediates entry into the host cell. Uncoating, and release of the viral genomic RNA into the cytoplasm. Viral RNA is translated in a polyprotein to produce replication proteins. Replication by helper virus occurs in viral factories made of membrane vesicles derived from the ER. A dsRNA genome is synthesized from the genomic ssRNA(+). The dsRNA genome is transcribed/replicated thereby providing viral mRNAs/new ssRNA(+) genomes. Expression of the capsid proteins. Assembly of new virus particles. Virus release. Disease Whitish muscle disease, which develops in post-larvae of freshwater prawn Macrobrachium rosenbergii and is caused by Macrobrachium rosenbergii nodavirus (MrNV) and its associate Macrobrachium satellite virus 1''. Main symptom is a whitish appearance of the muscl
https://en.wikipedia.org/wiki/Code%20page%201102
Code page 1102 (CCSID 1102), also known as CP1102 or NL7DEC, is an IBM code page number assigned to the Dutch variant of DEC's National Replacement Character Set (NRCS). The 7-bit character set was introduced for DEC's computer terminal systems, starting with the VT200 series in 1983, but is also used by IBM for their DEC emulation. It is called DUTCH by Kermit. Although NL7DEC complies with the ISO 646 invariant layout (and is hence a close derivation from ASCII, with only nine code points differing), it is not ISO646-NL, which is otherwise unrelated (Code page 1019). Code page layout See also National Replacement Character Set (NRCS) Code page 1019
https://en.wikipedia.org/wiki/Nuclei%20in%20the%20Cosmos
Nuclei in the Cosmos (NIC) is an internationally hosted series of biennial nuclear astrophysics conferences. Bringing together nuclear scientists and astronomers, it has served as the primary forum within the field leading it to be called "the most important international meeting in the field of nuclear astrophysics". Prior to the conference, a school for graduate students and postdocs is held along with a pre-workshop. The conference series was initiated by Heinz Oberhummer and Claus Rolfs with the first held in 1990 in Austria and rotates internationally. Conferences 1990 NIC-I Austria 1992 NIC-II Germany 1994 NIC-III Italy 1996 NIC-IV United States 1998 NIC-V Greece 2000 NIC-VI Denmark 2002 NIC-VII Japan 2004 NIC-VIII Canada 2006 NIC-IX Switzerland/France 2008 NIC X United States 2010 NIC-XI Germany 2012 NIC-XII Australia 2014 NIC-XIII Hungary 2016 NIC-XIV Japan 2018 NIC-XV Italy 2021 NIC-XVI China 2023 NIC-XVII South Korea 2025 NIC-XVIII Spain See also Joint Institute for Nuclear Astrophysics
https://en.wikipedia.org/wiki/Cofaco
Cofaco is a Portuguese canned fish company, headquartered in Lisbon with industrial facilities in the Azores. History In 1961, Cofaco was created in the Algarve. Its head office was in Vila Real de Santo António. The tuna fish is rare in the Algarve. Cofaco fishes in the Azores where the tuna passes and its migration routes begin. Currently, Cofaco is mainly located in the Azores and its industrial poles are concentrated in the islands of Pico, (Madalena do Pico), and S. Miguel (Rabo de Peixe). Its main brands include Bom Petisco, pitéu, líder, Ás do Mar, Bon Appetit, and Santamaria. See also Fishing in Portugal External links Official site Food and drink companies of Portugal Canned food Companies based in Lisbon Fishing in Portugal
https://en.wikipedia.org/wiki/Co-receptor
A co-receptor is a cell surface receptor that binds a signalling molecule in addition to a primary receptor in order to facilitate ligand recognition and initiate biological processes, such as entry of a pathogen into a host cell. Properties The term co-receptor is prominent in literature regarding signal transduction, the process by which external stimuli regulate internal cellular functioning. The key to optimal cellular functioning is maintained by possessing specific machinery that can carry out tasks efficiently and effectively. Specifically, the process through which intermolecular reactions forward and amplify extracellular signals across the cell surface has developed to occur by two mechanisms. First, cell surface receptors can directly transduce signals by possessing both serine and threonine or simply serine in the cytoplasmic domain. They can also transmit signals through adaptor molecules through their cytoplasmic domain which bind to signalling motifs. Secondly, certain surface receptors lacking a cytoplasmic domain can transduce signals through ligand binding. Once the surface receptor binds the ligand it forms a complex with a corresponding surface receptor to regulate signalling. These categories of cell surface receptors are prominently referred to as co-receptors. Co-receptors are also referred to as accessory receptors, especially in the fields of biomedical research and immunology. Co-receptors are proteins that maintain a three-dimensional structure. The large extracellular domains make up approximately 76–100% of the receptor. The motifs that make up the large extracellular domains participate in ligand binding and complex formation. The motifs can include glycosaminoglycans, EGF repeats, cysteine residues or ZP-1 domains. The variety of motifs leads to co-receptors being able to interact with two to nine different ligands, which themselves can also interact with a number of different co-receptors. Most co-receptors lack a cytoplasmic domain
https://en.wikipedia.org/wiki/SCHEMA%20%28bioinformatics%29
SCHEMA is a computational algorithm used in protein engineering to identify fragments of proteins (called schemas) that can be recombined without disturbing the integrity of the proteins' three-dimensional structure. The algorithm calculates the interactions between a protein's different amino acid residues to determine which interactions may be disrupted by swapping structural domains of the protein. By minimizing these disruptions, SCHEMA can be used to engineer chimeric proteins that stably fold and may have altered function relative to their parent proteins. SCHEMA algorithm has been applied in the recombinant libraries of distantly related β-lactamases.
https://en.wikipedia.org/wiki/Adrian%20Smith%20%28statistician%29
Sir Adrian Frederick Melhuish Smith, PRS (born 9 September 1946) is a British statistician who is chief executive of the Alan Turing Institute and president of the Royal Society. Early life and education Smith was born on 9 September 1946 in Dawlish. He was educated at Selwyn College, Cambridge, and University College London, where his PhD supervisor was Dennis Lindley. Career From 1977 until 1990, he was professor of statistics and head of department of mathematics at the University of Nottingham. He was subsequently at Imperial College, London, where he was head of the mathematics department. Smith is a former deputy vice-chancellor of the University of London and became vice-chancellor of the university on 1 September 2012. He stood down from the role in August 2018 to become the director of the Alan Turing Institute. Smith is a member of the governing body of the London Business School. He served on the Advisory Council for the Office for National Statistics from 1996 to 1998, was statistical advisor to the Nuclear Waste Inspectorate from 1991 to 1998 and was advisor on Operational Analysis to the Ministry of Defence from 1982 to 1987. He is a former president of the Royal Statistical Society. He was elected a Fellow of the Royal Society in 2001. His FRS citation included "his diverse contributions to Bayesian statistics. His monographs are the most comprehensive available and his work has had a major impact on the development of monitoring tools for clinicians." In statistical theory, Smith is a proponent of Bayesian statistics and evidence-based practice—a general extension of evidence-based medicine into all areas of public policy. With Antonio Machi, he translated Bruno de Finetti's Theory of Probability into English. He wrote an influential paper in 1990 along with Alan E. Gelfand, which drew attention to the significance of the Gibbs sampler technique for Bayesian numerical integration problems. He was also co-author of the seminal paper on the partic
https://en.wikipedia.org/wiki/Chemesthesis
Chemesthesis is the chemical sensitivity of the skin and mucous membranes. Chemesthetic sensations arise when chemical compounds activate receptors associated with other senses that mediate pain, touch, and thermal perception. These chemical-induced reactions do not fit into the traditional sense categories of taste and smell. Examples of chemesthetic sensations include the burn-like irritation from capsaicin and related compounds in foods like chili peppers; the coolness of menthol in mouthwashes and topical analgesic creams; the stinging or tingling of carbonated beverages in the nose and mouth; the tear-induction of cut onions; and the pungent, cough-inducing sensation in the back of the throat elicited by the oleocanthal in high-quality extra virgin olive oil. Some of these sensations may be referred to as spiciness, pungency, or piquancy. Chemesthetic sensations sometimes arise by direct chemical activation of ion channels on sensory nerve fibers, for example of transient receptor potential channels including those of the TRPV, TRPA or TRPM subtypes. Alternatively, irritant chemicals may activate cells of the epithelium to release substances that indirectly activate the nerve fibers. The respiratory passages, including the nose and trachea, possess specialized cells called solitary chemosensory cells which release acetylcholine or other activators to excite nearby nerve fibers. Because chemoresponsive nerve fibers are present in all types of skin, chemesthetic sensations can be stimulated from anywhere on the body's surface as well as from mucosal surfaces in the nose, mouth, eyes, etc. Mucus membranes are generally more sensitive to chemesthetic stimuli because they lack the barrier function of cornified skin. Much of the chemesthetic flavor sensations are mediated by the trigeminal nerves, which are relatively large and important nerves. Flavors that stimulate the trigeminal nerves are therefore important - for example, carbon dioxide is the trigemin
https://en.wikipedia.org/wiki/Carrier%20preselect
Carrier preselect is a term relating to the telecommunications industry. It is a method of routing calls for least-cost routing (LCR) without the need for programming of PBX telephone system. This is the process whereby a telephone subscriber whose telephone line is maintained by one company, usually a former monopoly provider (e.g. BT), can choose to have some of their calls automatically routed across a different telephone company's network (e.g. Talk Talk) without needing to enter a special code or special equipment. See also Local loop unbundling Wholesale line rental External links Ofcom - Carrier Pre-Selection and Wholesale Line Rental - Outbound Carrier Pre-Selection Services from GCI - Outbound Carrier Pre-Selection Services from Six Degrees Group Telephony Local loop Broadband Telecommunications economics Teletraffic Telephone exchanges
https://en.wikipedia.org/wiki/Field%20capacity
Field capacity is the amount of soil moisture or water content held in the soil after excess water has drained away and the rate of downward movement has decreased. This usually takes place 2–3 days after rain or irrigation in pervious soils of uniform structure and texture. The physical definition of field capacity (expressed symbolically as θfc) is the bulk water content retained in soil at −33 kPa (or −0.33 bar) of hydraulic head or suction pressure. The term originated from Israelsen and West and Frank Veihmeyer and Arthur Hendrickson. Veihmeyer and Hendrickson realized the limitation in this measurement and commented that it is affected by so many factors that, precisely, it is not a constant (for a particular soil), yet it does serve as a practical measure of soil water-holding capacity. Field capacity improves on the concept of moisture equivalent by Lyman Briggs. Veihmeyer & Hendrickson proposed this concept as an attempt to improve water-use efficiency for farmers in California during 1949. Field capacity is characterized by measuring water content after wetting a soil profile, covering it (to prevent evaporation) and monitoring the change soil moisture in the profile. Water content when the rate of change is relatively small is indicative of when drainage ceases and is called Field Capacity, it is also termed drained upper limit (DUL). Lorenzo A. Richards and Weaver found that water content held by soil at a potential of −33 kPa (or −0.33 bar) correlate closely with field capacity (−10 kPa for sandy soils). Criticism There is also criticism of this concept; field capacity is a static measurement: in a field it depends upon the initial water content and the depth of wetting before the commencement of redistribution and the rate of change in water content over time. These conditions are not unique for a given soil. See also Available water capacity Integral energy Nonlimiting water range Pedotransfer function Permanent wilting point Water poten
https://en.wikipedia.org/wiki/List%20of%20medical%20roots%2C%20suffixes%20and%20prefixes
This is a used in medical terminology, their meanings, and their etymologies. Most of them are combining forms in Neo-Latin and hence international scientific vocabulary. There are a few general rules about how they combine. First, prefixes and suffixes, most of which are derived from ancient Greek or classical Latin, have a droppable vowel, usually -o-. As a general rule, this vowel almost always acts as a joint-stem to connect two consonantal roots (e.g. + + -logy = arthrology), but generally, the -o- is dropped when connecting to a vowel-stem (e.g. + -itis = arthritis, instead of ). Second, medical roots generally go together according to language, i.e., Greek prefixes occur with Greek suffixes and Latin prefixes with Latin suffixes. Although international scientific vocabulary is not stringent about segregating combining forms of different languages, it is advisable when coining new words not to mix different lingual roots. Prefixes and suffixes The following is an alphabetical list of medical prefixes and suffixes, along with their meanings, origins, and English examples. A B C D E F G H I J–K L M N O P Q–R S T U V X–Z English meanings This section contains lists of different root classification (e.g. body components, quantity, description, etc.). Each list is alphabetized by English meanings, with the corresponding Greek and Latin roots given. Roots of the body Roots of bodily concepts Body parts and substances Roots of color Roots of description Roots of position Prefixes of quantity or amount See also Glossary of medicine Classical compound International scientific vocabulary List of medical abbreviations Medical dictionary Medicine List of commonly used taxonomic affixes
https://en.wikipedia.org/wiki/Ethylparaben
Ethylparaben (ethyl para-hydroxybenzoate) is the ethyl ester of p-hydroxybenzoic acid. Its formula is HO-C6H4-CO-O-CH2CH3. It is a member of the class of compounds known as parabens. It is used as an antifungal preservative. As a food additive, it has E number E214. Sodium ethyl para-hydroxybenzoate, the sodium salt of ethylparaben, has the same uses and is given the E number E215.
https://en.wikipedia.org/wiki/Samalamig
Samalamig, also known as palamig, is a collective term for various Filipino sweet chilled beverages that usually include jelly-like ingredients. They come in various flavors, and are commonly sold by street vendors as refreshments. Typical ingredients of the drinks include gulaman (agar), sago pearls, kaong, tapioca pearls, nata de coco, and coconut (including macapuno). They are usually anglicized as pearl coolers or pearl and jelly coolers. Samalamig may also include various chilled fruit juices (usually with chunks of fruit), chocolate, and coffee drinks, regardless if jellies are added, that are also typically sold by samalamig vendors. Name The name "samálamig" comes from sa, meaning "for; to; at", and malamíg, an adjective meaning "cold, chilly" in Tagalog. "Sa malamig" may thus loosely mean "for cold (drinks); at a cold place; chilled". "Sa malamig" might have come from the calls of ambulant vendors, telling people to come and get cold drinks, i.e. "[Dito] sa malamig", loosely "here for cold drinks". Thus, "sa malamig" could be taken as a qualifier for the various types of drinks stored in cold containers, i.e. buko juice is "buko sa malamig" and sago't gulaman is "sago't gulaman sa malamig", but these full phrases are no longer in habitual use. An alternate name is palamig which means "cooler" or "chiller". Description Samalamig does not refer to a specific drink, but to a class of drinks that are served cold by street vendors. Thus they can come in a wide variety of flavors and types. They are traditionally sold by street vendors during summer months, but are now also offered by restaurants. The restaurant versions typically top the drinks with shaved ice. Types The main types of samalamig are listed below. The recipes however can be combined at the discretion of the maker. There are no set recipes for samalamig. The only common theme is that they are served cold with ice cubes or shaved ice. They also usually include jelly-like ingredients or pieces
https://en.wikipedia.org/wiki/Contingency%20management
Contingency management (CM) is the application of the three-term contingency (or operant conditioning), which uses stimulus control and consequences to change behavior. CM originally derived from the science of applied behavior analysis (ABA), but it is sometimes implemented from a cognitive-behavior therapy (CBT) framework as well (such as in dialectical behavior therapy, or DBT). Incentive-based contingency management is well-established when used as a clinical behavior analysis (CBA) treatment for substance use disorders, which entails that patients' earn money (vouchers) or other incentives (i.e., prizes) as a reward to reinforce drug abstinence (and, less often, punishment if they fail to adhere to program rules and regulations or their treatment plan). Another popular approach based on CM for alcoholism is the community reinforcement approach and family training (CRAFT) model, which uses self-management and shaping techniques. By most evaluations, its procedures produce one of the largest effect sizes out of all mental health and educational interventions. Token economies One form of contingency management is the token economy system. Token systems can be used in an individual or group format. Token systems are successful with a diverse array of populations including those suffering from addiction, those with special needs, and those experiencing delinquency. However, recent research questions the use of token systems with very young children. The exception to the last would be the treatment of stuttering. The goal of such systems is to gradually thin out and to help the person begin to access the natural community of reinforcement (the reinforcement typically received in the world for performing the behavior). Walker (1990) presents an overview of token systems and combining such procedures with other interventions in the classroom. He relates the comprehensiveness of token systems to the child's level of difficulty. Voucher programs and related applica
https://en.wikipedia.org/wiki/Phosphatidylinositol%205-phosphate
Phosphatidylinositol 5-phosphate (PtdIns5P) is a phosphoinositide, one of the phosphorylated derivatives of phosphatidylinositol (PtdIns), that are well-established membrane-anchored regulatory molecules. Phosphoinositides participate in signaling events that control cytoskeletal dynamics, intracellular membrane trafficking, cell proliferation and many other cellular functions. Generally, phosphoinositides transduce signals by recruiting specific phosphoinositide-binding proteins to intracellular membranes. Phosphatidylinositol 5-phosphate is one of the 7 known cellular phosphoinositides with less understood functions. It is phosphorylated on position D-5 of the inositol head group, which is attached via phosphodiester linkage to diacylglycerol (with varying chemical composition of the acyl chains, frequently 1-stearoyl-2-arachidonoyl chain). In quiescent cells, on average, PtdIns5P is of similar or higher abundance as compared to PtdIns3P and ~20-100-fold below the levels of PtdIns4P (Phosphatidylinositol 4-phosphate and PtdIns(4,5)P2 (Phosphatidylinositol 4,5-bisphosphate). Notably, steady-state PtdIns5P levels are more than 5-fold higher than those of PtdIns(3,5)P2. PtdIns5P was first demonstrated by HPLC (high pressure liquid chromatography) in mouse fibroblasts as a substrate for PtdIns(4,5)P2 synthesis by type II PIP kinases (1-phosphatidylinositol-5-phosphate 4-kinase). In many cell types, however, PtdIns5P is not detected by HPLC due to technical limitations associated with its poor separation from the abundant PtdIns4P. Rather, PtdIns5P is measured by the "mass assay", where PtdIns5P (as a part of the extracted cellular lipids) is converted in vitro by purified PtdIns5P 4-kinase to PtdIns(4,5)P2 that is subsequently quantified. Based on studies with the mass assay and an improved HPLC technique, PtdIns5P is detected in all studied mammalian cells. Most of the cellular PtdIns5P is found on cytoplasmic membranes whereas a smaller fraction resides in the nu
https://en.wikipedia.org/wiki/Biomolecular%20structure
Biomolecular structure is the intricate folded, three-dimensional shape that is formed by a molecule of protein, DNA, or RNA, and that is important to its function. The structure of these molecules may be considered at any of several length scales ranging from the level of individual atoms to the relationships among entire protein subunits. This useful distinction among scales is often expressed as a decomposition of molecular structure into four levels: primary, secondary, tertiary, and quaternary. The scaffold for this multiscale organization of the molecule arises at the secondary level, where the fundamental structural elements are the molecule's various hydrogen bonds. This leads to several recognizable domains of protein structure and nucleic acid structure, including such secondary-structure features as alpha helixes and beta sheets for proteins, and hairpin loops, bulges, and internal loops for nucleic acids. The terms primary, secondary, tertiary, and quaternary structure were introduced by Kaj Ulrik Linderstrøm-Lang in his 1951 Lane Medical Lectures at Stanford University. Primary structure The primary structure of a biopolymer is the exact specification of its atomic composition and the chemical bonds connecting those atoms (including stereochemistry). For a typical unbranched, un-crosslinked biopolymer (such as a molecule of a typical intracellular protein, or of DNA or RNA), the primary structure is equivalent to specifying the sequence of its monomeric subunits, such as amino acids or nucleotides. The primary structure of a protein is reported starting from the amino N-terminus to the carboxyl C-terminus, while the primary structure of DNA or RNA molecule is known as the nucleic acid sequence reported from the 5' end to the 3' end. The nucleic acid sequence refers to the exact sequence of nucleotides that comprise the whole molecule. Often, the primary structure encodes sequence motifs that are of functional importance. Some examples of such motif
https://en.wikipedia.org/wiki/Aviator%20call%20sign
An aviator call sign or aviator callsign is a call sign given to a military pilot, flight officer, and even some enlisted aviators. The call sign is a specialized form of nickname that is used as a substitute for the aviator's given name. It is used on flight suit and flight jacket name tags, painted/displayed beneath the officer's or enlisted aircrewman's name on aircraft fuselages or canopy rails, and in radio conversations. They are most commonly used in tactical jet aircraft communities (i.e., fighter, bomber, attack) than in other aircraft communities (i.e., airlift, mobility, maritime patrol), but their use is not totally exclusive to the former. Many NASA Astronauts with military aviator backgrounds are referred to during spaceflights by their call signs rather than their first names. The origins of aviator call signs are varied. Most call signs play on or reference on variants of the aviator's firstname or surname. Other inspirations for call signs may include personality traits, middle name, references to historical figures, or past exploits during the pilot's career. Aviator call signs nearly always must come from a member or members of the aviator's squadron, training class, or other cohort. It is considered bad form to try to give oneself a call sign and it is also common for aviators to be given a fairly derogatory call sign, and the more they complain about it, the more likely it is to stick. Some aviators use the same call sign throughout their careers; in other cases an aviator might have a series of call signs. For example, U.S. Navy Lieutenant Kara Hultgreen was originally given the call sign "Hulk" because of her ability to bench-press 200 pounds. Later, after a television appearance in which she wore noticeable makeup, she received the call sign "Revlon", and a 1998 biography was entitled Call Sign Revlon. In fiction Film The 1986 film Top Gun, set at the United States Navy Fighter Weapons School, featured several aviators with call signs,
https://en.wikipedia.org/wiki/Gustatory%20cortex
The primary gustatory cortex (GC) is a brain structure responsible for the perception of taste. It consists of two substructures: the anterior insula on the insular lobe and the frontal operculum on the inferior frontal gyrus of the frontal lobe. Because of its composition the primary gustatory cortex is sometimes referred to in literature as the AI/FO(Anterior Insula/Frontal Operculum). By using extracellular unit recording techniques, scientists have elucidated that neurons in the AI/FO respond to sweetness, saltiness, bitterness, and sourness, and they code the intensity of the taste stimulus. Role in the taste pathway Like the olfactory system, the taste system is defined by its specialized peripheral receptors and central pathways that relay and process taste information. Peripheral taste receptors are found on the upper surface of the tongue, soft palate, pharynx, and the upper part of the esophagus. Taste cells synapse with primary sensory axons that run in the chorda tympani and greater superficial petrosal branches of the facial nerve (cranial nerve VII), the lingual branch of the glossopharyngeal nerve (cranial nerve IX), and the superior laryngeal branch of the vagus nerve (Cranial nerve X) to innervate the taste buds in the tongue, palate, epiglottis, and esophagus respectively. The central axons of these primary sensory neurons in the respective cranial nerve ganglia project to rostral and lateral regions of the nucleus of the solitary tract in the medulla, which is also known as the gustatory nucleus of the solitary tract complex. Axons from the rostral (gustatory) part of the solitary nucleus project to the ventral posterior complex of the thalamus, where they terminate in the medial half of the ventral posterior medial nucleus. This nucleus projects in turn to several regions of the neocortex which includes the gustatory cortex (the frontal operculum and the insula), which becomes activated when the subject is consuming and experiencing t
https://en.wikipedia.org/wiki/Diaspore%20%28botany%29
In botany, a diaspore is a plant dispersal unit consisting of a seed or spore plus any additional tissues that assist dispersal. In some seed plants, the diaspore is a seed and fruit together, or a seed and elaiosome. In a few seed plants, the diaspore is most or all of the plant, and is known as a tumbleweed. Diaspores are common in weedy and ruderal species. Collectively, diaspores, seeds, and spores that have been modified for migration are known as disseminules. Role in dispersal A diaspore of seed plus elaiosome is a common adaptation to seed dispersal by ants (myrmecochory). This is most notable in Australian and South African sclerophyll plant communities. Typically, ants carry the diaspore to their nest, where they may eat the elaiosome and discard the seed, and the seed may subsequently germinate. A diaspore of seed(s) plus fruit is common in plants dispersed by frugivores. Fruit-eating bats typically carry the diaspore to a favorite perch, where they eat the fruit and discard the seed. Fruit-eating birds typically swallow small seeds but, like bats, may carry larger seeded fruits to a perch where they eat the fruit and discard the seed. Diaspores such as achenes and samarae are dispersed primarily by wind; samaras are dispersed also by sailing or tumbling as they fall in still air. Drift fruits and some others are dispersed by water. Tumbleweeds are dispersed by wind, sometimes over very long distances. These occur in a variety of weedy and ruderal species native to steppes and deserts. Grasses have various units of dispersal: rarely the caryopsis alone, often a diaspore. Disarticulation occurs below, between, or above the glumes and at all nodes. Although in some species the diaspore is a foxtail, in a few (the "tumble grasses") it is like a tumbleweed. See also Foxtail (diaspore) Lichen Propagule Seedbank Soil seed bank Sporogenesis Wolffia
https://en.wikipedia.org/wiki/Ultimate%20Chicken%20Horse
Ultimate Chicken Horse is a multiplayer competitive platform video game developed and published by Canadian studio Clever Endeavour Games. It was released for Microsoft Windows on 4 March 2016, PlayStation 4 and Xbox One in December 2017, and Nintendo Switch on 25 September 2018. Gameplay Ultimate Chicken Horse is a platform video game where players assume the role of one of various animals. The goal of each game is to score points by building a platform level one piece at a time (per player) and race each other to a flag on the other side of the level. Players add obstacles designed to challenge their opponents, while also making sure they themselves can handle their handiwork. If nobody reaches the goal, or if everybody does, nobody gets any points. Points are awarded each round for various achievements, such as reaching the goal first, or setting a successful trap. The winner of the match is the player who reaches a set number of points or has the most points after a set number of rounds. Each round is estimated to last a minute. The PC, Nintendo Switch, Xbox and PlayStation versions of the game support cross-play online. Development The concept for Ultimate Chicken Horse emerged in September 2014 during a game jam. Developer Richard Atlas stated in an interview with Gamasutra that he was in the process of forming Clever Endeavour Games when it came about. The final team included personnel who had previously worked on games such as Gardenarium and Rimworld. Ultimate Chicken Horse was primarily built in the Unity engine. Reception The PlayStation 4 version of Ultimate Chicken Horse received "generally favorable" reviews, according to review aggregator Metacritic. The Escapist said the game "does everything it tries to do very well, but a few quibbles keep it from greatness," and recommended "a better guiding hand for stage building, especially for new players".
https://en.wikipedia.org/wiki/Journal%20of%20Industrial%20Ecology
The Journal of Industrial Ecology is a bimonthly peer-reviewed academic journal covering industrial ecology. It is published by Wiley-Blackwell on behalf of the Yale School of the Environment and is an official journal of the International Society for Industrial Ecology. The editor-in-chief is Reid Lifset. According to the Journal Citation Reports, the journal had an impact factor of 6.946 in 2020. Abstracting and indexing The journal is abstracted and indexed in:
https://en.wikipedia.org/wiki/Fermentek
Fermentek Ltd. is a biotechnological company in the Atarot industrial zone of Jerusalem, Israel. It specializes in the research, development and manufacture of biologically active, natural products isolated from microorganisms as well as from other natural sources such as plants and algae. The main microorganisms used are nonpathogenic actinomycetes, Nocardia and Streptomycetes. The fungi used are: Penicillium, Aspergillus, Fusarium and the like. None of these is a human pathogen. Fermentek does not sell to individuals. Most of its products are marketed through major international distributors specializing in chemicals, under their own brand names. Nevertheless, Fermentek has specific impact on the biochemical market, especially in the field of mycotoxins. Mycotoxins are toxic compounds produced by molds in human food and farm animal feeds, thus being economically important factors. Fermentek manufactures an extensive line of pure mycotoxins used as standards in food analysis. In some cases, such as Aflatoxin M2, Fermentek supplies the entire world's requirements. In 2009, Fermentek announced a product family of highly standardized calibrant solutions of main mycotoxins. These are marketed under the brand name FermaSol. In 2010, it obtained ISO 13485 accreditation in connection with the production of starting materials for experimental drug production, and with manufacturing of reference standards of food contaminants. None of Fermentek's products have been invented by it. Fermentek's aim is to make known compounds affordable to the scientific community. Fermentek was founded by Dr. Yosef Behrend in 1994. It moved in 2004 to its new building, quadrupling its working space and greatly enlarging its manufacturing capacities. Technology Fermentek operates fermentors ranging in size from 10 to 15,000 liters, with filter presses and centrifuges of matching capacity. According to the company policy as declared at its official website, Fermentek uses only the "Cla
https://en.wikipedia.org/wiki/UniFrac
UniFrac, a shortened version of unique fraction metric, is a distance metric used for comparing biological communities. It differs from dissimilarity measures such as Bray-Curtis dissimilarity in that it incorporates information on the relative relatedness of community members by incorporating phylogenetic distances between observed organisms in the computation. Both weighted (quantitative) and unweighted (qualitative) variants of UniFrac are widely used in microbial ecology, where the former accounts for abundance of observed organisms, while the latter only considers their presence or absence. The method was devised by Catherine Lozupone, when she was working with Rob Knight of the University of Colorado at Boulder in 2005. Research methods The distance is calculated between pairs of samples (each sample represents an organismal community). All taxa found in one or both samples are placed on a phylogenetic tree. A branch leading to taxa from both samples is marked as "shared" and branches leading to taxa which appears only in one sample are marked as "unshared". The distance between the two samples is then calculated as: This definition satisfies the requirements of a distance metric, being non-negative, zero only when entities are identical, transitive, and conforming to the triangle inequality. If there are several different samples, a distance matrix can be created by making a tree for each pair of samples and calculating their UniFrac measure. Subsequently, standard multivariate statistical methods such as data clustering and principal co-ordinates analysis can be used. One can determine the statistical significance of the UniFrac distance between two samples using Monte Carlo simulations. By randomizing the sample classification of each taxon on the tree (leaving the branch structure unchanged) and creating a distribution of UniFrac distance values, one can obtain a distribution of UniFrac values. From this, a p-value can be given to the actual distanc
https://en.wikipedia.org/wiki/Interactions%20of%20actors%20theory
In information theory, Interactions of actors theory is a theory developed by Gordon Pask and Gerard de Zeeuw. It is a generalisation of Pask's earlier conversation theory: The chief distinction being that conversation theory focuses on analysing the specific features that allow a conversation to emerge between two participants, whereas interaction of actor's theory focuses on the broader domain of conversation in which conversations may appear, disappear, and reappear over time. Overview Interactions of actors theory was developed late in Pask's career. It is reminiscent of Freud's psychodynamics, Bateson's panpsychism (see "Mind and Nature: A Necessary Unity" 1980). Pask's nexus of analogy, dependence and mechanical spin produces the differences that are central to cybernetics. While working with clients in the last years of his life, Pask produced an axiomatic scheme for his interactions of actors theory, less well-known than his conversation theory. Interactions of Actors, Theory and Some Applications, as the manuscript is entitled, is essentially a concurrent spin calculus applied to the living environment with strict topological constraints. One of the most notable associates of Gordon Pask, Gerard de Zeeuw, was a key contributor to the development of interactions of actors theory. Interactions of actors theory is a process theory. As a means to describe the interdisciplinary nature of his work, Pask would make analogies to physical theories in the classic positivist enterprises of the social sciences. Pask sought to apply the axiomatic properties of agreement or epistemological dependence to produce a "sharp-valued" social science with precision comparable to the results of the hard sciences. It was out of this inclination that he would develop his interactions of actors theory. Pask's concepts produce relations in all media and he regarded IA as a process theory. In his complementarity principle he stated "Processes produce products and all products (fin
https://en.wikipedia.org/wiki/The%20Physical%20Principles%20of%20the%20Quantum%20Theory
The Physical Principles of the Quantum Theory ( publisher: S. Hirzel Verlag, 1930) by Nobel laureate (1932) Werner Heisenberg and subsequently translated by Carl Eckart and Frank C. Hoyt. The book was first published in 1930 by University of Chicago Press. Then in 1949, according to its copyright page, Dover Publications reprinted the "unabridged and unaltered" 1930's version. The book is collection of 1929 university lectures by Heisenberg but with more detailed mathematics. The book discusses quantum mechanics and one 1931 review states that this is a "less technical and less involved account of the theor[y]". This book has been cited more than 2,000 times. In the book, after briefly discussing various theories, including quantum theory, Heisenberg discusses the basis for the fundamental concepts of quantum theory. Also by this time Heisenberg has stated, "the interaction between observer and object causes uncontrollable and large changes in the [atomic] system being observed...". In this work Heisenberg also discusses his uncertainty principle or uncertainty relations. About the author Werner Heisenberg (b. 1901 - d. 1976) was a renowned German theoretical physicist whose work pioneered and advanced quantum mechanics. He received his PhD in 1923 from Ludwig Maximilian University of Munich under Arnold Sommerfeld. He was awarded the 1932 Nobel Prize in Physics "for the creation of quantum mechanics, the application of which has led to the discovery of the allotropic forms of hydrogen".
https://en.wikipedia.org/wiki/Cryptoperiod
A cryptoperiod is the time span during which a specific cryptographic key is authorized for use. Common government guidelines range from 1 to 3 years for asymmetric cryptography, and 1 day to 7 days for symmetric cipher traffic keys. Factors to consider include the strength of the underlying encryption algorithm, key length, the likelihood of compromise through a security breach and the availability of mechanisms of revoking keys. In traditional cryptographic practice, keys were changed at regular intervals, typically at the same time each day. The code word for a key change, in NSA parlance, is HJ or Hotel Juliet in the NATO phonetic alphabet. When cryptographic devices began to be used in large scale, those who had to update the key had to set a specific time to synchronize the re-key. This was accomplished at the hour (H) the Julian (J) Date changed, among crypto-accountants, managers and users the jargon "HJ" became the accepted term meaning it was time to change the crypto-key. NESTOR in Vietnam During the Vietnam War, the United States issued its forces a series of secure voice encryption equipment code-named NESTOR. According to a U.S. Army history: "To maintain compatibility, key changes had to occur simultaneously in all units. The time chosen for this change was midnight, tactically the worst possible time because the greatest number of enemy contacts occurred from 2200 to 0200. Moreover, where several units shared the same keying device, having to move at night to change key settings was inconvenient and dangerous and added to the reasons for not using the equipment. Later the time of the daily NESTOR key change was moved to 0600."
https://en.wikipedia.org/wiki/Analytic%20narrative
An analytic narrative is a social science research method seeking to combine historical narratives with the rigor of rational choice theory, particularly through the use of game theory. The goal of analytic narratives is to provide several forms of discipline on the structure of case studies, such as a game, out of sample tests, and additional predictions that can be tested even if the main assertion about the case cannot. Analytic narratives (Bates et al. 1998, 2000; Levi 2002, 2004) involve choosing a problem or puzzle, then building a model to explicate the logic of an explanation for the puzzle or problem, often in the context of a unique case. The method involves: the use of narrative to elucidate the principal players, their preferences, the key decision points and possibilities, and the rules of game in a textured and sequenced account; and the evaluation of the model through comparative statics and the testable implications the model generates. The analytic narrative approach is most attractive to scholars who seek to evaluate the strength of parsimonious causal mechanisms in the context of a specific and often unique case. The requirement of explicit formal theorizing (or at least theory that could be formalized) compels scholars to make causal statements and to identify a small number of variables as central to understanding the case. This approach provides two methods for establishing the generalizability of the theory. First, the model in an analytic narrative often affords a range of explanations and predictions. Although the main account of a unique case may not be testable, the model may yield other predictions that can be tested, either in this case or in other cases. Second, as with other methods, out-of-sample tests constitute an important route to generalization. The presumption today in social science research is that the authors will provide those tests themselves. However, seldom does the level of knowledge for the out of sample case rival t
https://en.wikipedia.org/wiki/Intrinsic%20immunity
Intrinsic immunity refers to a set of cellular-based anti-viral defense mechanisms, notably genetically encoded proteins which specifically target eukaryotic retroviruses. Unlike adaptive and innate immunity effectors, intrinsic immune proteins are usually expressed at a constant level, allowing a viral infection to be halted quickly. Intrinsic antiviral immunity refers to a form of innate immunity that directly restricts viral replication and assembly, thereby rendering a cell non-permissive to a specific class or species of viruses. Intrinsic immunity is conferred by restriction factors preexisting in certain cell types, although these factors can be further induced by virus infection. Intrinsic viral restriction factors recognize specific viral components, but unlike other pattern recognition receptors that inhibit viral infection indirectly by inducing interferons and other antiviral molecules, intrinsic antiviral factors block viral replication immediately and directly. Background Eukaryotic organisms have been exposed to viral infections for millions of years. The development of the innate and adaptive immune system reflects the evolutionary importance of fighting infection. Some viruses, however, have proven to be so deadly or refractory to conventional immune mechanisms that specific, genetically encoded cellular defense mechanisms have evolved to combat them. Intrinsic immunity comprises cellular proteins which are always active and have evolved to block infection by specific viruses or viral taxa. The recognition of intrinsic immunity as a potent anti-viral defense mechanism is a recent discovery and is not yet discussed in most immunology courses or texts. Though the extent of protection intrinsic immunity affords is still unknown, it is possible that intrinsic immunity may eventually be considered a third branch of the traditionally bipartite immune system. Relationship to the immune system Intrinsic Immunity combines aspects of the two traditional b
https://en.wikipedia.org/wiki/Thyroiditis
Thyroiditis is the inflammation of the thyroid gland. The thyroid gland is located on the front of the neck below the laryngeal prominence, and makes hormones that control metabolism. Signs and symptoms There are many different signs and symptoms for thyroiditis, none of which are exclusively limited to this disease. Many of the signs imitate symptoms of other diseases, so thyroiditis can sometimes be difficult to diagnose. Common hypothyroid symptoms manifest when thyroid cell damage is slow and chronic, and may include fatigue, weight gain, feeling "fuzzy headed", depression, dry skin, and constipation. Other, rarer symptoms include swelling of the legs, vague aches and pains, decreased concentration and so on. When conditions become more severe, depending on the type of thyroiditis, one may start to see puffiness around the eyes, slowing of the heart rate, a drop in body temperature, or even incipient heart failure. On the other hand, if the thyroid cell damage is acute, the thyroid hormone within the gland leaks out into the bloodstream causing symptoms of thyrotoxicosis, which is similar to those of hyperthyroidism. These symptoms include weight loss, irritability, anxiety, insomnia, fast heart rate, and fatigue. Elevated levels of thyroid hormone in the bloodstream cause both conditions, but thyrotoxicosis is the term used with thyroiditis since the thyroid gland is not overactive, as in the case of hyperthyroidism. Causes Thyroiditis is generally caused by an immune system attack on the thyroid, resulting in inflammation and damage to the thyroid cells. This disease is often considered a malfunction of the immune system and can be associated with IgG4-related systemic disease, in which symptoms of autoimmune pancreatitis, retroperitoneal fibrosis and noninfectious aortitis also occur. Such is also the case in Riedel thyroiditis, an inflammation in which the thyroid tissue is replaced by fibrous tissue which can extend to neighbouring structures. Antibodie
https://en.wikipedia.org/wiki/Nicolaas%20Govert%20de%20Bruijn
Nicolaas Govert "Dick" de Bruijn (; 9 July 1918 – 17 February 2012) was a Dutch mathematician, noted for his many contributions in the fields of analysis, number theory, combinatorics and logic. Biography De Bruijn was born in The Hague where he attended elementary school between 1924 and 1930 and secondary school until 1934. He started studies in mathematics at Leiden University in 1936 but his studies were interrupted by the outbreak of World War II in 1939. He became a full-time Assistant in the Department of Mathematics of the Technological University of Delft in September 1939 while continuing his studies. He completed his undergraduate studies at the University of Leiden in 1941. He received his PhD in 1943 from the Vrije Universiteit Amsterdam with a thesis entitled "Over modulaire vormen van meer veranderlijken" advised by Jurjen Ferdinand Koksma. From June 1944 he was a Scientific Associate working in Philips Research Laboratories in Eindhoven. He married Elizabeth de Groot on 30 August 1944. The couple had four children: Jorina Aleida (born 19 January 1947), Frans Willem (born 13 April 1948), Elisabeth (born 24 November 1950), and Judith Elizabeth (born 31 March 1963). De Bruijn started his academic career at the University of Amsterdam, where he was Professor of Mathematics from 1952 to 1960. In 1960 he moved to the Technical University Eindhoven where he was Professor of Mathematics until his retirement in 1984. Among his graduate students were Johannes Runnenburg (1960), Antonius Levelt (1961), S. Ackermans (1964), Jozef Beenakker (1966), W. van der Meiden (1967), Matheus Hautus (1970), Robert Nederpelt Lazarom (1973), Lambert van Benthem Jutting (1977), A. Janssen (1979), Diederik van Daalen (1980), and Harmannus Balsters (1986). In 1957 he was appointed member of the Royal Netherlands Academy of Arts and Sciences. He was Knighted with the Order of the Netherlands Lion. Work De Bruijn covered many areas of mathematics. He is especially noted f
https://en.wikipedia.org/wiki/Cloacibacillus%20porcorum
Cloacibacillus porcorum is a Gram-negative, anaerobic, mesophilic and non-motile bacterium from the genus of Cloacibacillus which has been isolated from the intestinal tract of a pig from Ames in the United States.
https://en.wikipedia.org/wiki/F.%20Lynn%20McNulty
Lt. Col. Frederick “F.” Lynn McNulty (1939–2012) was the first Director of Information Systems Security for the U.S. State Department, a pioneer in the field of cybersecurity, and has been characterized as the “father” of U.S. federal information security. His peers eulogized him as a tireless advocate for the security of the nation. He was an early and persistent champion of information security in government. Cybersecurity McNulty played a key role in influencing U.S. national security policy on everything from export controls on commercial encryption products to the deployment of key federal cybersecurity infrastructure. He was repeatedly called to testify on cybersecurity matters by U.S. House and Senate subcommittees. According to SCMagazine.com’s Robert Bigman, McNulty’s contributions to cybersecurity are “woven into the DNA of almost every government information security policy and program,” and his security initiatives became a model within the federal government and served as the basis for The Computer Security Act of 1987. In its June 15, 2012 edition, Federal Computer Weekly identified McNulty as one of the key thought leaders in the field of cybersecurity during the preceding quarter century. A Fellow of the International Information Systems Security Consortium (ISC)², McNulty was the recipient of the Federal 100 Award and inducted into the Information Systems Security Association (ISSA) Hall of Fame. His publications include F. Lynn McNulty “Encryption's Importance to Economic and Infrastructure Security”, 9 Duke Journal of Comparative & International Law 427–450 (1999). Background & Postscript Born in Alameda, California, McNulty received his B.A. in international affairs from the University of California, Berkeley. He was awarded 2 master's degrees, an M.A. in international affairs from San Jose State and an M.P.A. from George Washington University. A United States Army Reserve officer from 1963 until 1999, Frederick Lynn McNulty was activated for
https://en.wikipedia.org/wiki/Kaid%C4%81%20glyphs
are a set of pictograms once used in the Yaeyama Islands of southwestern Japan. The word kaidā was taken from Yonaguni, and most studies on the pictographs focused on Yonaguni Island. However, there is evidence for their use in Yaeyama's other islands, most notably on Taketomi Island. They were used primarily for tax notices, thus were closely associated with the poll tax imposed on Yaeyama by Ryūkyū on Okinawa Island, which was in turn dominated by Satsuma Domain on Southern Kyushu. Etymology Sudō (1944) hypothesized that the etymology of kaidā was , which meant "government office" in Satsuma Domain. This term was borrowed by Ryūkyū on Okinawa and also by the bureaucrats of Yaeyama (karja: in Modern Ishigaki). Standard Japanese /j/ regularly corresponds to /d/ in Yonaguni, and /r/ is often dropped when surrounded by vowels. This theory is in line with the primary impetus for Kaidā glyphs, taxation. History Immediately after conquering Ryūkyū, Satsuma conducted a land survey in Okinawa in 1609 and in Yaeyama in 1611. By doing so, Satsuma decided the amount of tribute to be paid annually by Ryūkyū. Following that, Ryūkyū imposed a poll tax on Yaeyama in 1640. A fixed quota was allocated to each island and then was broken up into each community. Finally, quotas were set for the individual islanders, adjusted only by age and gender. Community leaders were notified of quotas in the government office on Ishigaki. They checked the calculation using warazan (barazan in Yaeyama), a straw-based method of calculation and recording numerals that was reminiscent of Incan Quipu. After that, the quota for each household was written on a wooden plate called . That was where Kaidā glyphs were used. Although sōrō-style Written Japanese had the status of administrative language, the remote islands had to rely on pictograms to notify illiterate peasants. According to a 19th-century document cited by the Yaeyama rekishi (1954), an official named Ōhama Seiki designed "perfect ideograp