source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Characteristic%20length
In physics, a characteristic length is an important dimension that defines the scale of a physical system. Often, such a length is used as an input to a formula in order to predict some characteristics of the system, and it is usually required by the construction of a dimensionless quantity, in the general framework of dimensional analysis and in particular applications such as fluid mechanics. In computational mechanics, a characteristic length is defined to force localization of a stress softening constitutive equation. The length is associated with an integration point. For 2D analysis, it is calculated by taking the square root of the area. For 3D analysis, it is calculated by taking the cubic root of the volume associated to the integration point. Examples A characteristic length is usually the volume of a system divided by its surface: For example, it is used to calculate flow through circular and non-circular tubes in order to examine flow conditions (i.e., the Reynolds number). In those cases, the characteristic length is the diameter of the pipe or, in case of non-circular tubes, its hydraulic diameter : Where is the cross-sectional area of the pipe and is its wetted perimeter. It is defined such that it reduces to a circular diameter of D for circular pipes. For flow through a square duct with a side length of a, the hydraulic diameter is: For a rectangular duct with side lengths a and b: For free surfaces (such as in open-channel flow), the wetted perimeter includes only the walls in contact with the fluid. Similarly, in the combustion chamber of a rocket engine, the characteristic length is defined as the chamber volume divided by the throat area. Because the throat of a de Laval nozzle is smaller than the cross section of the combustion chamber, the characteristic length is greater than the physical length of the combustion chamber.
https://en.wikipedia.org/wiki/Inscribed%20angle
In geometry, an inscribed angle is the angle formed in the interior of a circle when two chords intersect on the circle. It can also be defined as the angle subtended at a point on the circle by two given points on the circle. Equivalently, an inscribed angle is defined by two chords of the circle sharing an endpoint. The inscribed angle theorem relates the measure of an inscribed angle to that of the central angle subtending the same arc. The inscribed angle theorem appears as Proposition 20 on Book 3 of Euclid's Elements. Theorem Statement The inscribed angle theorem states that an angle θ inscribed in a circle is half of the central angle 2θ that subtends the same arc on the circle. Therefore, the angle does not change as its vertex is moved to different positions on the circle. Proof Inscribed angles where one chord is a diameter Let O be the center of a circle, as in the diagram at right. Choose two points on the circle, and call them V and A. Draw line VO and extended past O so that it intersects the circle at point B which is diametrically opposite the point V. Draw an angle whose vertex is point V and whose sides pass through points A and B. Draw line OA. Angle BOA is a central angle; call it θ. Lines OV and OA are both radii of the circle, so they have equal lengths. Therefore, triangle VOA is isosceles, so angle BVA (the inscribed angle) and angle VAO are equal; let each of them be denoted as ψ. Angles BOA and AOV add up to 180°, since line VB passing through O is a straight line. Therefore, angle AOV measures 180° − θ. It is known that the three angles of a triangle add up to 180°, and the three angles of triangle VOA are: 180° − θ ψ ψ. Therefore, Subtract from both sides, where θ is the central angle subtending arc AB and ψ is the inscribed angle subtending arc AB. Inscribed angles with the center of the circle in their interior Given a circle whose center is point O, choose three points V, C, and D on the circle. Draw lines
https://en.wikipedia.org/wiki/Mac%20OS%20Gaelic
Mac OS Gaelic is a character encoding created for the Irish Gaelic language, based on the Welsh Mac OS Celtic encoding but replacing 23 characters with Gaelic characters. It was developed by Michael Everson, and was in his CeltScript fonts and on some fonts included with the Irish localization of Mac OS 6.0.8 and 7.1 and on. Like ISO 8859-14, this codepage represents the Irish Gaelic and Welsh languages. Layout Before Mac OS 8.5, 0x26 mapped to both & (ampersand) and ⁊ (Tironian et, Unicode character U+20A4), which were unified. Before Mac OS 8.5, the character 0xDB mapped to currency sign (¤), Unicode character U+00A4. Before Mac OS 8.5, the character 0xE4 mapped to ‰, Unicode character U+2030. Before Unicode 4.1, the character 0xF0 mapped to ♣ Unicode character U+2663.
https://en.wikipedia.org/wiki/Racism%20on%20the%20Internet
Racist rhetoric is distributed through computer-mediated means and includes some or all of the following characteristics: ideas of racial uniqueness, racist attitudes towards specific social categories, racist stereotypes, hate-speech, nationalism and common destiny, racial supremacy, superiority and separation, conceptions of racial otherness, and anti-establishment world-view. Racism online can have the same effects as offensive remarks made face-to-face. Definitions The term "cyber racism" was coined by English sociologist Les Back in 2002. Cyber racism has been interpreted to be more than a phenomenon featuring racist acts displayed online. According to the Australian Human Rights Commission, Cyber-Racism involves online activity that can include "jokes or comments that cause offense or hurt; name-calling or verbal abuse; harassment or intimidation, or public commentary that inflames hostility towards certain groups". Roots and enabling factors Institutional racism Though there have been studies and strategies for thwarting and confronting cyber racism on the individual level there have not been many studies that expand on how cyber racism's roots in institutional racism can be combated. An increase in literature on cyber racism's relationship with institutional racism will provide new avenues for research on combatting cyber racism on a systemic level. For example, cyber racism's connections to institutional racism have been noted in the work of Jessie Daniels, a professor of sociology at Hunter College. Although some tech companies have taken steps to combat cyber racism on their sites, most tech companies are hesitant to take action over fears of limiting free speech. A Declaration of the Independence of Cyberspace, a document that declares the internet as a place free from control by "governments of the industrial world", continues to influence and reflect the views of Silicon Valley. Online stereotypes Online stereotypes can cause racist prejudice a
https://en.wikipedia.org/wiki/FiberHome
FiberHome Telecommunication Technologies Co., Ltd. () is a major networking and telecommunication equipment provider in the People's Republic of China. Its headquarters is in Hongshan District, Wuhan, Hubei province, China. Founded in 1999, Fiberhome Networks was one of the 8 affiliated companies and highly specializing on IP networks under the management of Fiberhome Company. In May 2020, the U.S. Commerce Department's Bureau of Industry and Security said it was adding Fiberhome to its Entity List of organizations subject to Export Administration Regulations governing exports and other transactions. The Commerce Department said Fiberhome was “complicit” in alleged human rights abuses involving Muslim minority groups in the Xinjiang Uighur Autonomous Region (XUAR) in northwest China. U.S. companies will have to obtain special, hard-to-get licenses to do business with Fiberhome.
https://en.wikipedia.org/wiki/Korteweg%E2%80%93De%20Vries%20equation
In mathematics, the Korteweg–De Vries (KdV) equation is a partial differential equation (PDE) which serves as a mathematical model of waves on shallow water surfaces. It is particularly notable as the prototypical example of an integrable PDE and exhibits many of the expected behaviors for an integrable PDE, such as a large number of explicit solutions, in particular soliton solutions, and an infinite number of conserved quantities, despite the nonlinearity which typically renders PDEs intractable. The KdV can be solved by the inverse scattering method (ISM). In fact, Gardner, Greene, Kruskal and Miura developed the classical inverse scattering method to solve the KdV equation. The KdV equation was first introduced by and rediscovered by , who found the simplest solution, the one-soliton solution. Understanding of the equation and behavior of solutions was greatly advanced by the computer simulations of Zabusky and Kruskal in 1965 and then the development of the inverse scattering transform in 1967. Definition The KdV equation is a nonlinear, dispersive partial differential equation for a function of two dimensionless real variables, and which are proportional to space and time respectively: with and denoting partial derivatives with respect to and . For modelling shallow water waves, is the height displacement of the water surface from its equilibrium height. The constant in front of the last term is conventional but of no great significance: multiplying , , and by constants can be used to make the coefficients of any of the three terms equal to any given non-zero constants. The introduces dispersion while is an advection term. Soliton solutions One-soliton solution Consider solutions in which a fixed wave form (given by ) maintains its shape as it travels to the right at phase speed . Such a solution is given by . Substituting it into the KdV equation gives the ordinary differential equation or, integrating with respect to , where is a con
https://en.wikipedia.org/wiki/Local%20structure
The local structure is a term in nuclear spectroscopy that refers to the structure of the nearest neighbours around an atom in crystals and molecules. E.g. in crystals the atoms order in a regular fashion on wide ranges to form even gigantic highly ordered crystals (Naica Mine). However, in reality, crystals are never perfect and have impurities or defects, which means that a foreign atom resides on a lattice site or in between lattice sites (interstitials). These small defects and impurities cannot be seen by methods such as X-ray diffraction or neutron diffraction, because these methods average in their nature of measurement over a large number of atoms and thus are insensitive to effects in local structure. Methods in nuclear spectroscopy use specific nuclei as probe. The nucleus of an atom is about 10,000 to 150,000 times smaller than the atom itself. It experiences the electric fields created by the atom's electrons that surround the nucleus. In addition, the electric fields created by neighbouring atoms also influence the fields that the nucleus experiences. The interactions between the nucleus and these fields are called hyperfine interactions that influence the nucleus' properties. The nucleus therefore becomes very sensitive to small changes in its hyperfine structure, which can be measured by methods of nuclear spectroscopy, such as e.g. nuclear magnetic resonance, Mössbauer spectroscopy, and perturbed angular correlation. With the same methods, the local magnetic fields in a crystal structure can also be probed and provide a magnetic local structure. This is of great importance for the understanding of defects in magnetic materials, which have wide range of applications such as modern magnetic materials or the giant magnetoresistance effect, that is used in materials in the reader heads of harddrives. Research of the local structure of materials has become an important tool for the understanding of properties especially in functional materials, such a
https://en.wikipedia.org/wiki/Inviscid%20flow
In fluid dynamics, inviscid flow is the flow of an inviscid (zero-viscosity) fluid, also known as a superfluid. The Reynolds number of inviscid flow approaches infinity as the viscosity approaches zero. When viscous forces are neglected, such as the case of inviscid flow, the Navier–Stokes equation can be simplified to a form known as the Euler equation. This simplified equation is applicable to inviscid flow as well as flow with low viscosity and a Reynolds number much greater than one. Using the Euler equation, many fluid dynamics problems involving low viscosity are easily solved, however, the assumed negligible viscosity is no longer valid in the region of fluid near a solid boundary (the boundary layer) or, more generally in regions with large velocity gradients which are evidently accompanied by viscous forces. Inviscid flows are broadly classified into potential flows (or, irrotational flows) and rotational inviscid flows. Prandtl hypothesis Ludwig Prandtl developed the modern concept of the boundary layer. His hypothesis establishes that for fluids of low viscosity, shear forces due to viscosity are evident only in thin regions at the boundary of the fluid, adjacent to solid surfaces. Outside these regions, and in regions of favorable pressure gradient, viscous shear forces are absent so the fluid flow field can be assumed to be the same as the flow of an inviscid fluid. By employing the Prandtl hypothesis it is possible to estimate the flow of a real fluid in regions of favorable pressure gradient by assuming inviscid flow and investigating the irrotational flow pattern around the solid body. Real fluids experience separation of the boundary layer and resulting turbulent wakes but these phenomena cannot be modelled using inviscid flow. Separation of the boundary layer usually occurs where the pressure gradient reverses from favorable to adverse so it is inaccurate to use inviscid flow to estimate the flow of a real fluid in regions of unfavorable pre
https://en.wikipedia.org/wiki/Domain%20Name%20System%20Security%20Extensions
The Domain Name System Security Extensions (DNSSEC) are a suite of extension specifications by the Internet Engineering Task Force (IETF) for securing data exchanged in the Domain Name System (DNS) in Internet Protocol (IP) networks. The protocol provides cryptographic authentication of data, authenticated denial of existence, and data integrity, but not availability or confidentiality. Overview The original design of the Domain Name System did not include any security features. It was conceived only as a scalable distributed system. The Domain Name System Security Extensions (DNSSEC) attempt to add security, while maintaining backward compatibility. documents some of the known threats to the DNS, and their solutions in DNSSEC. DNSSEC was designed to protect applications using DNS from accepting forged or manipulated DNS data, such as that created by DNS cache poisoning. All answers from DNSSEC protected zones are digitally signed. By checking the digital signature, a DNS resolver is able to check if the information is identical (i.e. unmodified and complete) to the information published by the zone owner and served on an authoritative DNS server. While protecting IP addresses is the immediate concern for many users, DNSSEC can protect any data published in the DNS, including text records (TXT) and mail exchange records (MX), and can be used to bootstrap other security systems that publish references to cryptographic certificates stored in the DNS such as Certificate Records (CERT records, ), SSH fingerprints (SSHFP, ), IPSec public keys (IPSECKEY, ), TLS Trust Anchors (TLSA, ), or Encrypted Client Hello (SVCB/HTTPS records for ECH ). DNSSEC does not provide confidentiality of data; in particular, all DNSSEC responses are authenticated but not encrypted. DNSSEC does not protect against DoS attacks directly, though it indirectly provides some benefit (because signature checking allows the use of potentially untrustworthy parties). Other standards (not DNSSEC) a
https://en.wikipedia.org/wiki/Chern%E2%80%93Simons%20theory
The Chern–Simons theory is a 3-dimensional topological quantum field theory of Schwarz type developed by Edward Witten. It was discovered first by mathematical physicist Albert Schwarz. It is named after mathematicians Shiing-Shen Chern and James Harris Simons, who introduced the Chern–Simons 3-form. In the Chern–Simons theory, the action is proportional to the integral of the Chern–Simons 3-form. In condensed-matter physics, Chern–Simons theory describes the topological order in fractional quantum Hall effect states. In mathematics, it has been used to calculate knot invariants and three-manifold invariants such as the Jones polynomial. Particularly, Chern–Simons theory is specified by a choice of simple Lie group G known as the gauge group of the theory and also a number referred to as the level of the theory, which is a constant that multiplies the action. The action is gauge dependent, however the partition function of the quantum theory is well-defined when the level is an integer and the gauge field strength vanishes on all boundaries of the 3-dimensional spacetime. It is also the central mathematical object in theoretical models for topological quantum computers (TQC). Specifically, an SU(2) Chern–Simons theory describes the simplest non-abelian anyonic model of a TQC, the Yang–Lee–Fibonacci model. The dynamics of Chern–Simons theory on the 2-dimensional boundary of a 3-manifold is closely related to fusion rules and conformal blocks in conformal field theory, and in particular WZW theory. The classical theory Mathematical origin In the 1940s S. S. Chern and A. Weil studied the global curvature properties of smooth manifolds M as de Rham cohomology (Chern–Weil theory), which is an important step in the theory of characteristic classes in differential geometry. Given a flat G-principal bundle P on M there exists a unique homomorphism, called the Chern–Weil homomorphism, from the algebra of G-adjoint invariant polynomials on g (Lie algebra of G) to the c
https://en.wikipedia.org/wiki/Disposable%20camera
A disposable or single-use camera is a simple box camera meant to be used once. Most use fixed-focus lenses. Some are equipped with an integrated flash unit, and there are even waterproof versions for underwater photography. Internally, the cameras use a 135 film or an APS cartridge. While some disposables contain an actual cartridge as used for loading normal, reusable cameras, others just have the film wound internally on an open spool. The whole camera is handed in for processing. Some of the cameras are recycled, i.e. refilled with film and resold. The cameras are returned for "processing" in the same fashion as film cameras. In general the one-time-use camera represents a return to the business model pioneered by Kodak for their Kodak camera, predecessor to the Brownie camera; it is particularly popular in situations where a reusable camera would be easily stolen or damaged, when one's regular camera is forgotten, or if one cannot afford a regular camera. History A company called Photo-Pac produced a cardboard camera beginning in 1949 which shot eight exposures and which was mailed-in for processing. Cameras were expensive, and would often have been left safely at home when photo opportunities presented themselves. Frustrated with missing photo opportunities, H. M. Stiles had invented a way to enclose 35mm film in an inexpensive enclosure without the expensive precision film transport mechanism. It cost . Though incredibly similar to the familiar single-use cameras today, Photo-Pac failed to make a permanent impression on the market. In 1966, French company FEX introduced a disposable bakelite camera called "Photo Pack Matic", featuring 12 photos (4×4 cm). The currently familiar disposable camera was developed by Fujifilm in 1986. Their QuickSnap line, known as 写ルンです (Utsurun-Desu, "It takes pictures") in Japan, used 35 mm film, while Eastman Kodak's 1987 Fling was based on 110 film. Kodak released a 35 mm version in 1988, and in 1989 renamed the 35 mm
https://en.wikipedia.org/wiki/TrueAllele
TrueAllele is a software program by Cybergenetics that analyzes DNA using statistical methods, a process called probabilistic genotyping. It is used in forensic identification. The program can be used in situations unsuited to traditional methods, such as when a mixture of multiple people's DNA is in a sample. Some studies, mostly conducted by Cybergenetics' Chief Scientific Officer Mark W. Perlin, have validated the program's accuracy. In one study, TrueAllele distinguished between the genetic code of first-degree relatives with "great accuracy". The President's Council of Advisors on Science and Technology has noted that many validation studies were made by people affiliated with TrueAllele and are therefore not independent, demanding more independent research. In one case, TrueAllele's results differed from the results of STRMix, another probabilistic genotyping program, leading to the judge rejecting the DNA evidence. The proprietary nature of the code has led to concerns over its reliability. Unlike some similar programs, TrueAllele is not open source, so judges and attorneys cannot check the program's code.
https://en.wikipedia.org/wiki/Zeaxanthin
Zeaxanthin is one of the most common carotenoids in nature, and is used in the xanthophyll cycle. Synthesized in plants and some micro-organisms, it is the pigment that gives paprika (made from bell peppers), corn, saffron, goji (wolfberries), and many other plants and microbes their characteristic color. The name (pronounced zee-uh-zan'-thin) is derived from Zea mays (common yellow maize corn, in which zeaxanthin provides the primary yellow pigment), plus xanthos, the Greek word for "yellow" (see xanthophyll). Xanthophylls such as zeaxanthin are found in highest quantity in the leaves of most green plants, where they act to modulate light energy and perhaps serve as a non-photochemical quenching agent to deal with triplet chlorophyll (an excited form of chlorophyll) which is overproduced at high light levels during photosynthesis. Zeaxanthin in guard cells acts as a blue light photoreceptor which mediates the stomatal opening. Animals derive zeaxanthin from a plant diet. Zeaxanthin is one of the two primary xanthophyll carotenoids contained within the retina of the eye. Zeaxanthin supplements are typically taken on the supposition of supporting eye health. Although there are no reported side effects from taking zeaxanthin supplements, the actual health effects of zeaxanthin and lutein are not proven, and, as of 2018, there is no regulatory approval in the European Union or the United States for health claims about products that contain zeaxanthin. As a food additive, zeaxanthin is a food dye with E number E161h. Isomers and macular uptake Lutein and zeaxanthin have identical chemical formulas and are isomers, but they are not stereoisomers. The only difference between them is in the location of the double bond in one of the end rings. This difference gives lutein three chiral centers whereas zeaxanthin has two. Because of symmetry, the (3R,3′S) and (3S,3′R) stereoisomers of zeaxanthin are identical. Therefore, zeaxanthin has only three stereoisomeric forms. T
https://en.wikipedia.org/wiki/Huberta%20%28hippopotamus%29
Huberta (initially named Hubert; the sex was discovered after death) was a hippopotamus which travelled for a large distance across South Africa. In November 1928, Huberta left her waterhole in the St. Lucia Estuary in Zululand and over the next three years, travelled to the Eastern Cape. In that time, Huberta became a minor celebrity in South Africa and attracted crowds wherever she went. She was initially thought to be a male and was nicknamed Hubert by the press. The first report in the press was on 23 November 1928 in the Natal Mercury and reported the appearance of a hippo in Natal. The report was accompanied by the only photograph of Huberta in life. Huberta stopped for a while at the mouth of the Mhlanga River about north of Durban, and a failed attempt was made to capture her and put her in Johannesburg Zoo. After this, she headed south to Durban where she visited a beach and a country club. Moving on to the Umgeni River, she became revered by Zulus and Xhosas alike. Finally, Huberta arrived in East London in March 1931. Despite her having been declared to be protected royal game by the Natal Provincial Council, she was shot by farmers a month later. After a public outcry, the farmers were arrested and fined £25. Huberta's body was recovered and sent to a taxidermist in London. Upon her return to South Africa in 1932, she was greeted by 20,000 people and was displayed at the Amathole Museum (previously known as the Kaffrarian Museum) in King William's Town. Huberta is the subject of the children's book Hubert The Traveling Hippopotamus by Edmund Lindop and illustrated by Jane Carlson. The book was published in 1961 by Little, Brown and Company.
https://en.wikipedia.org/wiki/Formula%20for%20primes
In number theory, a formula for primes is a formula generating the prime numbers, exactly and without exception. No such formula which is efficiently computable is known. A number of constraints are known, showing what such a "formula" can and cannot be. Formulas based on Wilson's theorem A simple formula is for positive integer , where is the floor function, which rounds down to the nearest integer. By Wilson's theorem, is prime if and only if . Thus, when is prime, the first factor in the product becomes one, and the formula produces the prime number . But when is not prime, the first factor becomes zero and the formula produces the prime number 2. This formula is not an efficient way to generate prime numbers because evaluating requires about multiplications and reductions modulo . In 1964, Willans gave the formula for the th prime number . This formula reduces to ; that is, it tautologically defines as the smallest integer m for which the prime-counting function is at least n. This formula is also not efficient. In addition to the appearance of , it computes by adding up copies of ; for example, . The articles What is an Answer? by Herbert Wilf (1982) and Formulas for Primes by Underwood Dudley (1983) have further discussion about the worthlessness of such formulas. Formula based on a system of Diophantine equations Because the set of primes is a computably enumerable set, by Matiyasevich's theorem, it can be obtained from a system of Diophantine equations. found an explicit set of 14 Diophantine equations in 26 variables, such that a given number k + 2 is prime if and only if that system has a solution in nonnegative integers: The 14 equations α0, …, α13 can be used to produce a prime-generating polynomial inequality in 26 variables: That is, is a polynomial inequality in 26 variables, and the set of prime numbers is identical to the set of positive values taken on by the left-hand side as the vari
https://en.wikipedia.org/wiki/Azure%20Stream%20Analytics
Microsoft Azure Stream Analytics is a serverless scalable complex event processing engine by Microsoft that enables users to develop and run real-time analytics on multiple streams of data from sources such as devices, sensors, web sites, social media, and other applications. Users can set up alerts to detect anomalies, predict trends, trigger necessary workflows when certain conditions are observed, and make data available to other downstream applications and services for presentation, archiving, or further analysis. Query Language Users can author real-time analytics using a simple declarative SQL-like language with embedded support for temporal logic. Callouts to custom code with JavaScript user defined functions extend the streaming logic written in SQL. Callouts to Azure Machine Learning helps with predictive scoring on streaming data. Scalability Azure Stream Analytics is a serverless job service on Azure that eliminates the need for infrastructure, servers, virtual machines, or managed clusters. Users only pay for the processing used for the running jobs. IoT applications Azure Stream Analytics integrates with Azure IoT Hub to enable real-time analytics on data from IoT devices and applications. Real-time Dashboards Users can build real-time dashboards with Power BI for a live command and control view. Real-time dashboards help transform live data into actionable and insightful visuals. Data Input Sources Stream Analytics supports three different types of input sources - Azure Event Hubs, Azure IoT Hubs, and Azure Blob Storage. Additionally, stream analytics supports Azure Blob storage as the input reference data to help augment fast moving event data streams with static data. Stream analytics supports a wide variety of output targets. Support for Power BI allows for real-time dashboarding. Event Hub, Service bus topics and queues help trigger downstream workflows. Support for Azure Table Storage, Azure SQL Databases, Azure SQL Data Warehouse, A
https://en.wikipedia.org/wiki/Florida%20Platform
The Florida Platform is a flat geological feature with the emergent portion forming the Florida peninsula. Structure The platform forms a rampart between the Gulf of Mexico and the Atlantic Ocean. The platform's western edge, or Florida Escarpment, is normally defined where water depths at drop dramatically and in a short distance to . The Florida peninsula is located on the eastern side of the platform, where in places it lies only from the platform's edge. On the Gulf side the platform ends over to the west of the modern shoreline, where a massive cliff rises over from the depth of the Gulf floor. The western reaches of the platform just off Tampa were explored by the submersible DSV Alvin. Examination has placed the depth of carbonate rocks at greater than . Geology Basement The platform's basement rocks include Precambrian-Cambrian igneous rocks, Ordovician-Devonian sedimentary rocks, and Triassic-Jurassic volcanic rock. Florida's igneous and sedimentary foundation separated from what is now the African Plate when the super-continent Pangea rifted apart in the Triassic and possibly pre-Middle Jurassic. It then secured to the North American craton. One of the early geologic structures is the Peninsular Arch which formed during the Jurassic. Emergent portion The oldest sediments that are exposed are Middle Eocene carbonates found in the Avon Park Formation. Most of the state of Florida is covered by Pliocene, Pleistocene, and Holocene siliciclastic-bearing sediments deposited during sea-level fluctuations and filling in of the Gulf Trough beginning in the late Tertiary and Quaternary. Notes External links Geologic Map of the State of Florida General Facts about the Gulf of Mexico Geology of Florida Physical oceanography Terranes Landforms of the Gulf of Mexico
https://en.wikipedia.org/wiki/Algae
Algae (, ; : alga ) is an informal term for a large and diverse group of photosynthetic, eukaryotic organisms. It is a polyphyletic grouping that includes species from multiple distinct clades. Included organisms range from unicellular microalgae, such as Chlorella, Prototheca and the diatoms, to multicellular forms, such as the giant kelp, a large brown alga which may grow up to in length. Most are aquatic and lack many of the distinct cell and tissue types, such as stomata, xylem and phloem that are found in land plants. The largest and most complex marine algae are called seaweeds, while the most complex freshwater forms are the Charophyta, a division of green algae which includes, for example, Spirogyra and stoneworts. Algae that are carried by water are plankton, specifically phytoplankton. Algae constitute a polyphyletic group since they do not include a common ancestor, and although their plastids seem to have a single origin, from cyanobacteria, they were acquired in different ways. Green algae are examples of algae that have primary chloroplasts derived from endosymbiotic cyanobacteria. Diatoms and brown algae are examples of algae with secondary chloroplasts derived from an endosymbiotic red alga. Algae exhibit a wide range of reproductive strategies, from simple asexual cell division to complex forms of sexual reproduction. Algae lack the various structures that characterize land plants, such as the phyllids (leaf-like structures) of bryophytes, rhizoids of non-vascular plants, and the roots, leaves, and other organs found in tracheophytes (vascular plants). Most are phototrophic, although some are mixotrophic, deriving energy both from photosynthesis and uptake of organic carbon either by osmotrophy, myzotrophy, or phagotrophy. Some unicellular species of green algae, many golden algae, euglenids, dinoflagellates, and other algae have become heterotrophs (also called colorless or apochlorotic algae), sometimes parasitic, relying entirely on external e
https://en.wikipedia.org/wiki/Area%20code%20split
In telecommunications, an area code split is the practice of introducing a new telephone area code by geographically dividing an existing numbering plan area (NPA), and assigning area codes to the resulting divisions, but retaining the existing area code only for one of the divisions. The purpose of this practice is to provide more central office prefixes, and therefore more telephone numbers, in an area with high demand for telecommunication services, and prevent a shortage of telephone numbers. An increasing demand for telephone numbers has existed since the development of automatic telephony in the early 20th century, but was spurred especially since the 1990s, with the proliferation of fax machines, pager systems, mobile telephones, computer modems and, eventually, smart phones. The implementation of an area code split typically involves the establishment of a Class-4 toll switching center for each division of the existing numbering plan area that receive a new area code. The local seven-digit telephone numbers in any of the areas are typically not changed. The existing central office prefixes are maintained and only the central offices of the new divisions are reassigned to a new area code. The impact of a split on the general public involves the printing of new stationery, advertisements, and signage for many customers, and the dissemination of the new area code to family, friends, and customers. Computer systems, and telephone equipment may require updates in address books, speed dialing directories, and other automated equipment. Since area code splits have substantial impact in the involved communities, and involve substantial cost in telephone plant and exchange equipment, they are planned carefully well ahead of implementation with the intent that an area is not again affected by a subsequent realignment for at least a decade. The new boundaries of the numbering plan areas are drawn in a manner that minimizes splitting communities and should coincide
https://en.wikipedia.org/wiki/Persistence%20module
A persistence module is a mathematical structure in persistent homology and topological data analysis that formally captures the persistence of topological features of an object across a range of scale parameters. A persistence module often consists of a collection of homology groups (or vector spaces if using field coefficients) corresponding to a filtration of topological spaces, and a collection of linear maps induced by the inclusions of the filtration. The concept of a persistence module was first introduced in 2005 as an application of graded modules over polynomial rings, thus importing well-developed algebraic ideas from classical commutative algebra theory to the setting of persistent homology. Since then, persistence modules have been one of the primary algebraic structures studied in the field of applied topology. Definition Single Parameter Persistence Modules Let be a totally ordered set and let be a field. The set is sometimes called the indexing set. Then a single-parameter persistence module is a functor from the poset category of to the category of vector spaces over and linear maps. A single-parameter persistence module indexed by a discrete poset such as the integers can be represented intuitively as a diagram of spaces: To emphasize the indexing set being used, a persistence module indexed by is sometimes called a -persistence module, or simply a -module. Common choices of indexing sets include , etc. One can alternatively use a set-theoretic definition of a persistence module that is equivalent to the categorical viewpoint: A persistence module is a pair where is a collection of -vector spaces and is a collection of linear maps where for each , such that for any (i.e., all the maps commute). Multiparameter Persistence Modules Let be a product of totally ordered sets, i.e., for some totally ordered sets . Then by endowing with the product partial order given by only if for all , we can define a multiparameter persiste
https://en.wikipedia.org/wiki/Battenburg%20markings
Battenburg markings or Battenberg markings are a pattern of high-visibility markings developed in the United Kingdom in the 1990s and currently seen on many types of emergency service vehicles in the UK, Crown dependencies, British Overseas Territories and several other European countries such as the Czech Republic, Iceland, Sweden, Germany, Romania, Spain, Ireland, and Belgium as well as in Australia, New Zealand, Hong Kong, Pakistan, Trinidad and Tobago, and more recently, Canada. The name comes from its similarity in appearance to the cross-section of a Battenberg cake. History Battenburg markings were developed in the mid-1990s in the United Kingdom by the Police Scientific Development Branch (now the Home Office Centre for Applied Science and Technology) at the request of the national motorway policing sub-committee of the Association of Chief Police Officers. They were first developed for traffic patrol cars for United Kingdom police forces; private organisations and civil emergency services have also used them since then. The brief was to design a livery for motorway and trunk road police vehicles that would maximise the vehicles' visibility, from a distance of up to , when stopped either in daylight or under headlights, and which distinctively marked them as police vehicles. The primary objectives were to design markings that: Made officers and vehicles more conspicuous (e.g. to prevent collisions when stopped) Made police vehicles recognisable at a distance of up to in daylight Assisted in high-visibility policing for public reassurance and deterrence of traffic violations Made police vehicles nationally recognisable Were an equal-cost option compared to existing markings Were acceptable to at least 75% of the staff Conspicuity Battenburg design uses a regular pattern and the contrast between a light and a dark colour to increase conspicuity for the human eye. The lighter colour is daylight-fluorescent (such as fluorescent-yellow) for better vi
https://en.wikipedia.org/wiki/Topness
Topness (T, also called truth), a flavour quantum number, represents the difference between the number of top quarks (t) and number of top antiquarks () that are present in a particle: By convention, top quarks have a topness of +1 and top antiquarks have a topness of −1. The term "topness" is rarely used; most physicists simply refer to "the number of top quarks" and "the number of top antiquarks". Conservation Like all flavour quantum numbers, topness is preserved under strong and electromagnetic interactions, but not under weak interaction. However the top quark is extremely unstable, with a half-life under 10−23 s, which is the required time for the strong interaction to take place. For that reason the top quark does not hadronize, that is it never forms any meson or baryon, so the topness of a meson or a baryon is always zero. By the time it can interact strongly it has already decayed to another flavour of quark (usually to a bottom quark).
https://en.wikipedia.org/wiki/Ryegrass%20mosaic%20virus
Ryegrass mosaic virus (RMV) is a virus in the genus Rymovirus. As the name suggests its hosts include ryegrass, but also other relatives in the family Poaceae. RMV's genome was sequenced in 2015.
https://en.wikipedia.org/wiki/Hamiltonian%20cycle%20polynomial
In mathematics, the Hamiltonian cycle polynomial of an n×n-matrix is a polynomial in its entries, defined as where is the set of n-permutations having exactly one cycle. This is an algebraic option useful, in a number of cases, for determining the existence of a Hamiltonian cycle in a directed graph. It is a generalization of the number of Hamiltonian cycles of a digraph as the sum of the products of its Hamiltonian cycles' arc weights (all of which equal unity) for weighted digraphs with arc weights taken from a given commutative ring. In the meantime, for an undirected weighted graph the sum of the products of the edge weights of its Hamiltonian cycles containing any fixed edge (i,j) can be expressed as the product of the weight of (i,j) and the Hamiltonian cycle polynomial of a matrix received from its weighted adjacency matrix via subjecting its rows and columns to any permutation mapping i to 1 and j to 2 and then removing its 1-st row and 2-nd column. In () it was shown that where is the submatrix of induced by the rows and columns of indexed by , and is the complement of in , while the determinant of the empty submatrix is defined to be 1. Due to this and Borchardt's identities, for a non-singular n×n Cauchy matrix where are diagonal matrices that make unitary (in a real field or a field of a finite characteristic, or orthogonal in the field of complex numbers), is the Hadamard (entry-wise) square of , and is the identity n×n-matrix with the entry of indexes 1,1 replaced by 0. In a field of characteristic 2 the equality turns into what therefore provides an opportunity to polynomial-time calculate the Hamiltonian cycle polynomial of any unitary matrix (i.e. such that where is the identity n×n-matrix), because in such a field each minor of a unitary matrix coincides with its algebraic complement: where is the identity n×n-matrix with the entry of indexes 1,1 replaced by 0. Hence if it's possible to polynomial-time assign
https://en.wikipedia.org/wiki/Windows%203.0
Windows 3.0 is the third major release of Microsoft Windows, launched in 1990. It features a new graphical user interface (GUI) where applications are represented as clickable icons, as opposed to the list of file names seen in its predecessors. Later updates would expand the software's capabilities, one of which added multimedia support for sound recording and playback, as well as support for CD-ROMs. Windows 3.0 is the first version of Windows to perform well both critically and commercially. Critics and users considered its GUI to be a challenger to those of Apple Macintosh and Unix. Other praised features were the improved multitasking, customizability, and especially the utilitarian management of computer memory that troubled the users of Windows 3.0's predecessors. Microsoft was criticized by third-party developers for the bundling of its separate software with the operating environment, which they viewed as an anticompetitive practice. Windows 3.0 sold 10 million copies before it was succeeded by Windows 3.1 in 1992. On December 31, 2001, Microsoft declared Windows 3.0 obsolete and stopped providing support and updates for the system. Development history Before Windows 3.0, Microsoft had a partnership with IBM, where the latter had sold personal computers running on the former's MS-DOS since 1981. Microsoft had made previous attempts to develop a successful operating environment called Windows, and IBM declined to include the project in its product line. As MS-DOS was entering its fifth iteration, IBM demanded a version of DOS that could run in "protected mode", which would allow it to execute multiple programs at once, among other benefits. MS-DOS was originally designed to run in real mode and run only one program at a time, due to the capability limitations of the Intel 8088 microprocessor. Intel had later released the Intel 80286, which was designed to support such multitasking efficiently (with several different hardware features, including memory pr
https://en.wikipedia.org/wiki/London%20School%20of%20Medicine%20for%20Women
The London School of Medicine for Women (LSMW) established in 1874 was the first medical school in Britain to train women as doctors. The patrons, vice-presidents, and members of the committee that supported and helped found the London School of Medicine for Women wanted to provide educated women with the necessary facilities for learning and practicing midwifery and other branches of medicine while also promoting their future employment in the fields of midwifery and other fields of treatment for women and children. History The school was formed in 1874 by an association of pioneering women physicians Sophia Jex-Blake, Elizabeth Garrett Anderson, Emily Blackwell and Elizabeth Blackwell with Thomas Henry Huxley. The founding was motivated at least in part by Jex-Blake's frustrated attempts at getting a medical degree at a time when women were not admitted to British medical schools, thus being expelled from Edinburgh University. Other women who had studied with Jex-Blake in Edinburgh joined her at the London school, including Isabel Thorne who succeeded her as honorary secretary in 1877. She departed to start a medical practice in Edinburgh where she would found the Edinburgh School of Medicine for Women in 1886. The UK Medical Act of 1876 (39 and 40 Vict, Ch. 41) was an act which repealed the previous Medical Act in the United Kingdom and allowed the medical authorities to license all qualified applicants irrespective of gender. In 1877 an agreement was reached with the Royal Free Hospital that allowed students at the London School of Medicine for Women to complete their clinical studies there. The Royal Free Hospital was the first teaching hospital in London to admit women for training. Elizabeth Garrett Anderson was Dean (1883–1903) while the school was rebuilt, became part of the University of London and consolidated association with the Royal Free Hospital. In 1896, the School was officially renamed the London (Royal Free Hospital) School of Medicine fo
https://en.wikipedia.org/wiki/Peter%20F.%20C.%20Gilbert
Peter Francis Cecil Gilbert (born 7 July 1944) is an English neuroscientist and biophysicist. He is known for his pioneering work on motor learning in the cerebellum. Early life and education Gilbert was educated at the Royal Grammar School, Newcastle upon Tyne. He went on to attend Gonville and Caius College, Cambridge on a college scholarship, graduating in 1966 with a degree in natural sciences. Remaining in Cambridge, he pursued a PhD at the MRC Laboratory of Molecular Biology, under the supervision of Aaron Klug, researching the structure of tobacco mosaic virus protein. Career and research Upon finishing his PhD in 1970, Gilbert continued his research at the MRC Laboratory of Molecular Biology as a staff scientist. During this time, Gilbert developed Simultaneous Iterative Reconstruction Technique (SIRT) – an iterative method for reconstructing the three-dimensional structure of objects from two-dimensional projections. The original SIRT, as proposed by Gilbert, along with the family of variants that it spawned (known as SIRT-like), have widespread applications across medicine and biology, such as CT scans and cryo-EM. In 1974, Gilbert moved into the field of neuroscience, joining the laboratory of Sir John Eccles at the State University of New York, Buffalo as a visiting professor, where he investigated the neurophysiology of the monkey cerebellum. Towards the end of 1974 and after completing his work in the Eccles laboratory, Gilbert partnered with Dr W. T. Thach at Yale School of Medicine. Over the next two years, and later, at Washington University School of Medicine, the pair carried out pioneering work on cerebellar motor learning. By recording from the Purkinje cells of conscious monkeys learning a manual task, Gilbert and Thach tested and confirmed theoretical predictions on motor learning in the cerebellum. Besides his experimental work, Gilbert has produced a number of theories on cerebellar motor learning and has published his theories in Natu
https://en.wikipedia.org/wiki/Boletus%20loyo
Boletus loyo is a species of bolete fungus in the family Boletaceae that is found in South America. It was described as new to science by Carlos Luigi Spegazzini in 1912, who made the first scientifically documented collections in Argentina. The bolete is edible. See also List of Boletus species
https://en.wikipedia.org/wiki/Metabidiminished%20icosahedron
In geometry, the metabidiminished icosahedron is one of the Johnson solids (). The name refers to one way of constructing it, by removing two pentagonal pyramids () from a regular icosahedron, replacing two sets of five triangular faces of the icosahedron with two adjacent pentagonal faces. If two pentagonal pyramids are removed to form nonadjacent pentagonal faces, the result is instead the pentagonal antiprism.
https://en.wikipedia.org/wiki/Right-hand%20rule
In mathematics and physics, the right-hand rule is a convention and a mnemonic for deciding the orientation of axes in three-dimensional space. It is a convenient method for determining the direction of the cross product of two vectors. There are two ways of applying the right hand rule. The first one is conventionally called the Right hand rule or the Flemming's right hand rule. It involves the index finger, the middle finger and the thumb of the right hand. By arranging them as shown in the diagram, the direction of cross product or vector product can be calculated. The other way, known as Amperes right hand grip rule, right-hand screw rule, coffee-mug rule or the corkscrew-rule involves pointing all fingers of the right hand along the first vector and curling the fingers along the second vector, the direction which the thumb makes is the direction of vector product. For example, If the curling motion of the fingers represents a movement from the first (x-axis) to the second (y-axis), then the third (z-axis) can point along either thumb in a right handed coordinate system. Both these rules can be used interchangeably. The rule can be used to find the direction of the magnetic field, rotation, spirals, electromagnetic fields, mirror images, and enantiomers in mathematics and chemistry. The sequence is often: index finger along the first vector, then middle finger along the second, then thumb along the third. Two other sequences also work because they preserve the cyclic nature of the cross product (and the underlying Levi-Civita symbol): Middle finger, thumb, index finger. Thumb, index finger, middle finger. Coordinates For right-handed coordinates, if the thumb of a person's right hand points along the z-axis in the positive direction (third coordinate vector), then the fingers curl from the positive x-axis (first coordinate vector) toward the positive y-axis (second coordinate vector). When viewed at a position along the positive z-axis, the ¼ tu
https://en.wikipedia.org/wiki/Earthflow
An earthflow (earth flow) is a downslope viscous flow of fine-grained materials that have been saturated with water and moves under the pull of gravity. It is an intermediate type of mass wasting that is between downhill creep and mudflow. The types of materials that are susceptible to earthflows are clay, fine sand and silt, and fine-grained pyroclastic material. When the ground materials become saturated with enough water, they will start flowing (soil liquefaction). Its speed can range from being barely noticeable to rapid movement. The velocity of the flow is dictated by water content: the higher the water content is, the higher the velocity will be. Because of the dependency on water content for the velocity of the flow, it can take minutes or years for the materials to move down the slope. Features and behavior Earthflows are just one type of mass movement that can occur on a hill slope. It has been recognized as its own type of movement since the early 20th century. Earthflows are one of the most fluid types of mass movements. Earthflows occur on heavily saturated slopes like mudflows or a debris flow. Though earthflows are a lot like mudflows, overall they are slower and are covered with solid material carried along by flow from within. Earthflows are often made up of fine-grained materials so slopes consisting of clay and silt materials are more likely to create an earthflow. As earthflows are usually water-dependent, the risk of one occurring is much higher in humid areas especially after a period of heavy rainfall or snowmelt. The high level of precipitation, which saturates the ground and adds water to the slope content, increases the pore-water pressure and reduces the shearing strength of the material. As the slope becomes wet, the earthflow may start as a creep downslope due to the clay or silt having less friction. As the material is increasingly more saturated, the slope will fail, which depends on slope stability. In earthflows, the slop
https://en.wikipedia.org/wiki/Specific%20dynamic%20action
Specific dynamic action (SDA), also known as thermic effect of food (TEF) or dietary induced thermogenesis (DIT), is the amount of energy expenditure above the basal metabolic rate due to the cost of processing food for use and storage. Heat production by brown adipose tissue which is activated after consumption of a meal is an additional component of dietary induced thermogenesis. The thermic effect of food is one of the components of metabolism along with resting metabolic rate and the exercise component. A commonly used estimate of the thermic effect of food is about 10% of one's caloric intake, though the effect varies substantially for different food components. For example, dietary fat is very easy to process and has very little thermic effect, while protein is hard to process and has a much larger thermic effect. Factors that affect the thermic effect of food The thermic effect of food is increased by both aerobic training of sufficient duration and intensity or by anaerobic weight training. However, the increase is marginal, amounting to 7-8 calories per hour. The primary determinants of daily TEF are the total caloric content of the meals and the macronutrient composition of the meals ingested. Meal frequency has little to no effect on TEF; assuming total calorie intake for the days are equivalent. Although some believe that TEF is reduced in obesity, discrepant results and inconsistent research methods have failed to validate such claims. The mechanism of TEF is unknown. TEF has been described as the energy used in the distribution of nutrients and metabolic processes in the liver, but a hepatectomized animal shows no signs of TEF and intravenous injection of amino acids results in an effect equal to that of oral ingestion of the same amino acids. Types of foods The thermic effect of food is the energy required for digestion, absorption, and disposal of ingested nutrients. Its magnitude depends on the composition of the food consumed: Carbohydrates
https://en.wikipedia.org/wiki/Distributed%20data%20processing
Distributed data processing (DDP) was the term that IBM used for the IBM 3790 (1975) and its successor, the IBM 8100 (1979). Datamation described the 3790 in March 1979 as "less than successful." Distributed data processing was used by IBM to refer to two environments: IMS DB/DC CICS/DL/I Each pair included a Telecommunications Monitor and a Database system. The layering involved a message, containing information to form a transaction, which was then processed by an application program. Development tools such as program validation services were released by IBM to facilitate expansion. Use of "a number of small computers linked to a central computer" permitted local and central processing, each optimized at what it could best do. Terminals, including those described as intelligent, typically were attached locally, to a "satellite processor." Central systems, sometimes multi-processors, grew to handle the load. Some of this extra capacity, of necessity, is used to enhance data security. Years before open systems made its presence felt, the goal of some hardware suppliers was "to replace the big, central mainframe computer with an array of smaller computers that are tied together." Lower case distributed data processing Hadoop adds another term to the mix: File System. Tools added for this use of distributed data processing include new programming languages. TSI/DPF Flexicom In 1976 Turnkey Systems Inc (TSI)/DPF Inc. introduced a hardware/software telecommunications front-end to off-load some processing that handled distributed data processing. Named Flexicom, The CPU was IBM-manufactured, and it ran (mainframe) DOS Rel. 26, with Flexicom's additions. Of four models available, the smallest had the CPU of a 360/30. See also HPCC
https://en.wikipedia.org/wiki/The%20Second%20Forty%20Years
The Second Forty Years is a 1946 nonfiction self-help book about aging, by Edward Stieglitz. Stieglitz graduated from Rush Medical College in 1921 and was on the faculty of the University of Chicago from 1923 to 1938. In 1940 a grant was given to support a gerontologist at the United States Public Health Service, and Stieglitz was given the post. His research was also supported by the National Institute of Health. Stieglitz left the post after a year, for personal reasons. The Second Forty Years provides Stieglitz's professional advice for the layman. It describes what can be expected during the aging process and what can be done about it, discussing chronic progressive disorders which require preventative care, sex, the importance of leisure and rest, and so forth. According to Stieglitz, mere longevity is not goal enough, but rather constructive health practices should be undertaken to enhance the quality of life. Regarding menopause, Stieglitz characterized it as a "truly normal phase of living" and decried the "distorted descriptions" of old wives' tales. Newspaper advertisements for The Second Forty Years are probably the source for the popular quotation "It's not the years in your life that count, it's the life in your years", which is commonly misattributed to Abraham Lincoln.
https://en.wikipedia.org/wiki/Riemann%E2%80%93Roch%20theorem%20for%20smooth%20manifolds
In mathematics, a Riemann–Roch theorem for smooth manifolds is a version of results such as the Hirzebruch–Riemann–Roch theorem or Grothendieck–Riemann–Roch theorem (GRR) without a hypothesis making the smooth manifolds involved carry a complex structure. Results of this kind were obtained by Michael Atiyah and Friedrich Hirzebruch in 1959, reducing the requirements to something like a spin structure. Formulation Let X and Y be oriented smooth closed manifolds, and f: X → Y a continuous map. Let vf=f*(TY) − TX in the K-group K(X). If dim(X) ≡ dim(Y) mod 2, then where ch is the Chern character, d(vf) an element of the integral cohomology group H2(Y, Z) satisfying d(vf) ≡ f* w2(TY)-w2(TX) mod 2, fK* the Gysin homomorphism for K-theory, and fH* the Gysin homomorphism for cohomology . This theorem was first proven by Atiyah and Hirzebruch. The theorem is proven by considering several special cases. If Y is the Thom space of a vector bundle V over X, then the Gysin maps are just the Thom isomorphism. Then, using the splitting principle, it suffices to check the theorem via explicit computation for line bundles. If f: X → Y is an embedding, then the Thom space of the normal bundle of X in Y can be viewed as a tubular neighborhood of X in Y, and excision gives a map and . The Gysin map for K-theory/cohomology is defined to be the composition of the Thom isomorphism with these maps. Since the theorem holds for the map from X to the Thom space of N, and since the Chern character commutes with u and v, the theorem is also true for embeddings. f: X → Y. Finally, we can factor a general map f: X → Y into an embedding and the projection The theorem is true for the embedding. The Gysin map for the projection is the Bott-periodicity isomorphism, which commutes with the Chern character, so the theorem holds in this general case also. Corollaries Atiyah and Hirzebruch then specialised and refined in the case X = a point, where the condition becomes the existence of a
https://en.wikipedia.org/wiki/Ulam%20matrix
In mathematical set theory, an Ulam matrix is an array of subsets of a cardinal number with certain properties. Ulam matrices were introduced by Stanislaw Ulam in his 1930 work on measurable cardinals: they may be used, for example, to show that a real-valued measurable cardinal is weakly inaccessible. Definition Suppose that κ and λ are cardinal numbers, and let be a -complete filter on . An Ulam matrix is a collection of subsets of indexed by such that If then and are disjoint. For each , the union over of the sets , is in the filter .
https://en.wikipedia.org/wiki/Hydroid%20%28botany%29
A hydroid is a type of vascular cell that occurs in certain bryophytes. In some mosses such as members of the Polytrichaceae family, hydroids form the innermost layer of cells in the stem. At maturity they are long, colourless, thin walled cells of small diameter, containing water but no living protoplasm. Collectively, hydroids function as a conducting tissue, known as the hydrome, transporting water and minerals drawn from the soil. They are surrounded by bundles of living cells known as leptoids which carry sugars and other nutrients in solution. The hydroids are analogous to the tracheids of vascular plants but there is no lignin present in the cell walls to provide structural support. Hydroids have been found in some fossilised plants from the Rhynie chert, including Aglaophyton, where they were initially mistaken for xylem tracheids. See also Leptoid, a related sucrose-transporting vessel analogous to the phloem of vascular plants
https://en.wikipedia.org/wiki/TorChat
TorChat was a peer-to-peer anonymous instant messenger that used Tor onion services as its underlying network. It provided cryptographically secure text messaging and file transfers. The characteristics of Tor's onion services ensure that all traffic between the clients is encrypted and that it is very difficult to tell who is communicating with whom and where a given client is physically located. TorChat is free software licensed under the terms of the GNU General Public License (GPL). Features In TorChat every user has a unique alphanumeric ID consisting of 16 characters. This ID will be randomly created by Tor when the client is started the first time, it is basically the .onion address of an onion service. TorChat clients communicate with each other by using Tor to contact the other's onion service (derived from their ID) and exchanging status information, chat messages and other data over this connection. Since onion services can receive incoming connections even if they are behind a router doing network address translation (NAT), TorChat does not need any port forwarding to work. History The first public version of TorChat was released in November 2007 by Bernd Kreuss (prof7bit). It is written in Python and used the cross-platform widget toolkit wxPython which made it possible to support a wide range of platforms and operating systems. The older Windows versions of TorChat were built with py2exe (since 0.9.9.292 replaced with pyinstaller) and came bundled with a copy of Tor readily configured so that it could be run as a portable application right off a USB flash drive without any installation, configuration or account creation. Between 2008 and 2010 weren't any updated packages, resulting in the bundled version of Tor becoming obsolete and unable to connect to the Tor network, which was the reason for the appearance of forks that basically just replaced the bundled Tor.exe with a current one. In December 2010, an official update finally became availabl
https://en.wikipedia.org/wiki/Dubins%E2%80%93Schwarz%20theorem
In the theory of martingales, the Dubins-Schwarz theorem (or Dambis-Dubins-Schwarz theorem) is a theorem that says all continuous local martingales and martingales are time-changed Brownian motions. The theorem was proven in 1965 by Lester Dubins and Gideon E. Schwarz and independently in the same year by K. E. Dambis, a doctorial student of Eugene Dynkin. Dubins-Schwarz theorem Let be the space of -adapted continuous local martingales with . be the quadratic variation. Statement Let and and define for all the time-changes (i.e. stopping times) Then is a -Brownian motion and . Remarks The condition guarantees that the underlying probability space is rich enough so that the Brownian motion exists. If one removes this conditions one might has to use enlargement of the filitered probability space. is not a -Brownian motion. are almost surely finite since .
https://en.wikipedia.org/wiki/Inferior%20ligament%20of%20epididymis
The inferior ligament of the epididymis is a strand of fibrous tissue which is covered by a reflection of the tunica vaginalis and connects the lower aspect of the epididymis with the testis. Sexual anatomy Ligaments
https://en.wikipedia.org/wiki/Monsanto%20Co.%20v.%20Rohm%20and%20Haas%20Co.
Monsanto Co. v. Rohm and Haas Co., 456 F.2d 592 (3d Cir. 1972), is a 1972 decision of the United States Court of Appeals for the Third Circuit interpreting what conduct amounts to fraudulent procurement of a patent. This case is one of the early decisions following the US Supreme Court's 1964 decision in Walker Process v. Food Machinery holding fraud on the US Patent Office as potentially violating the Sherman Antitrust Act, and one of the first (if not the first) to hold that failure to disclose material information to the Patent Office was fraudulent. Background Monsanto procured U.S. Patent No. No. 3,382,280, issued May 7, 1968, having the title 3',4'-dichloropropionanilide (known as 3,4-DCPA or propanil), a herbicide that selectively killed weeds without killing crop plants such as rice. In November 1969 Monsanto sued Rohm and Haas for patent infringement. The only substantial issue was the validity of the patent, and that ultimately turned on whether Monsanto had committed fraud on the Patent Office in procuring the patent. The application that resulted in the '280 patent was the third of three successive applications, the first two of which were unsuccessful. In the first application, filed in 1957, Monsanto sought a patent on some 100 "compounds, including 3,4-DCPA and 3,4-DCAA (3,4-dichloroacetanilide), a chemical with some similar properties and a similar physical structure. Monsanto claimed that all the members of the class possessed "unusual and valuable herbicidal activity," while related compounds possessed "little or no herbicidal efficiency." Unpersuaded by Monsanto's arguments, the Patent Office rejected the application as unpatentable over the prior art. In 1961, Monsanto filed a new application claiming another large class of compounds, again including 3,4-DCPA and 3,4-DCAA and again asserting that the class possessed "unusual and valuable herbicidal activity." Again, the Patent Office rejected the application as unpatentable over the prior ar
https://en.wikipedia.org/wiki/Neurotrophin-4
Neurotrophin-4 (NT-4), also known as neurotrophin-5 (NT-5), is a protein that in humans is encoded by the NTF4 gene. It is a neurotrophic factor that signals predominantly through the TrkB receptor tyrosine kinase. See also Tropomyosin receptor kinase B § Agonists
https://en.wikipedia.org/wiki/Multilinear%20multiplication
In multilinear algebra, applying a map that is the tensor product of linear maps to a tensor is called a multilinear multiplication. Abstract definition Let be a field of characteristic zero, such as or . Let be a finite-dimensional vector space over , and let be an order-d simple tensor, i.e., there exist some vectors such that . If we are given a collection of linear maps , then the multilinear multiplication of with is defined as the action on of the tensor product of these linear maps, namely Since the tensor product of linear maps is itself a linear map, and because every tensor admits a tensor rank decomposition, the above expression extends linearly to all tensors. That is, for a general tensor , the multilinear multiplication is where with is one of 's tensor rank decompositions. The validity of the above expression is not limited to a tensor rank decomposition; in fact, it is valid for any expression of as a linear combination of pure tensors, which follows from the universal property of the tensor product. It is standard to use the following shorthand notations in the literature for multilinear multiplications:andwhere is the identity operator. Definition in coordinates In computational multilinear algebra it is conventional to work in coordinates. Assume that an inner product is fixed on and let denote the dual vector space of . Let be a basis for , let be the dual basis, and let be a basis for . The linear map is then represented by the matrix . Likewise, with respect to the standard tensor product basis , the abstract tensoris represented by the multidimensional array . Observe that where is the jth standard basis vector of and the tensor product of vectors is the affine Segre map . It follows from the above choices of bases that the multilinear multiplication becomes The resulting tensor lives in . Element-wise definition From the above expression, an element-wise definition of the multilinear multiplication is obtai
https://en.wikipedia.org/wiki/Biomedicum%20Helsinki
The Biomedicum Helsinki is a research and teaching center located in the Meilahti district of Helsinki, Finland. It contains two buildings (Biomedicum Helsinki 1 and 2), and features research and teaching facilities of the University of Helsinki's Faculty of Medicine, the Helsinki University Central Hospital, the EMBL Nordic partner Finnish Institute for Molecular Medicine, the Folkhälsan Research Center, the Minerva Foundation Institute for Medical Research, the Wihuri Research Institute, and several companies in the life science field. Building 1 was completed in 2001 and Building 2 in 2008. A bridge named in honor of Konrad ReijoWaara (1853-1936), a prominent figure in the Medical Association Duodecim and the first editor-in-chief of the magazine of the same name, connects the buildings. The use of the buildings are coordinated by the Biomedicum Helsinki Foundation, which also funds research activities in the field of medicine.
https://en.wikipedia.org/wiki/Protein%20subcellular%20localization%20prediction
Protein subcellular localization prediction (or just protein localization prediction) involves the prediction of where a protein resides in a cell, its subcellular localization. In general, prediction tools take as input information about a protein, such as a protein sequence of amino acids, and produce a predicted location within the cell as output, such as the nucleus, Endoplasmic reticulum, Golgi apparatus, extracellular space, or other organelles. The aim is to build tools that can accurately predict the outcome of protein targeting in cells. Prediction of protein subcellular localization is an important component of bioinformatics based prediction of protein function and genome annotation, and it can aid the identification of drug targets. Background Experimentally determining the subcellular localization of a protein can be a laborious and time consuming task. Immunolabeling or tagging (such as with a green fluorescent protein) to view localization using fluorescence microscope are often used. A high throughput alternative is to use prediction. Through the development of new approaches in computer science, coupled with an increased dataset of proteins of known localization, computational tools can now provide fast and accurate localization predictions for many organisms. This has resulted in subcellular localization prediction becoming one of the challenges being successfully aided by bioinformatics, and machine learning. Many prediction methods now exceed the accuracy of some high-throughput laboratory methods for the identification of protein subcellular localization. Particularly, some predictors have been developed that can be used to deal with proteins that may simultaneously exist, or move between, two or more different subcellular locations. Experimental validation is typically required to confirm the predicted localizations. Tools In 1999 PSORT was the first published program to predict subcellular localization. Subsequent tools and websites
https://en.wikipedia.org/wiki/MIMO
In radio, multiple-input and multiple-output (MIMO) () is a method for multiplying the capacity of a radio link using multiple transmission and receiving antennas to exploit multipath propagation. MIMO has become an essential element of wireless communication standards including IEEE 802.11n (Wi-Fi 4), IEEE 802.11ac (Wi-Fi 5), HSPA+ (3G), WiMAX, and Long Term Evolution (LTE). More recently, MIMO has been applied to power-line communication for three-wire installations as part of the ITU G.hn standard and of the HomePlug AV2 specification. At one time, in wireless the term "MIMO" referred to the use of multiple antennas at the transmitter and the receiver. In modern usage, "MIMO" specifically refers to a class of techniques for sending and receiving more than one data signal simultaneously over the same radio channel by exploiting the difference in signal propagation between different antennas (e.g. due to multipath propagation). Additionally, modern MIMO usage often refers to multiple data signals sent to different receivers (with one or more receive antennas) though this is more accurately termed multi-user multiple-input single-output (MU-MISO). History Early research MIMO is often traced back to 1970s research papers concerning multi-channel digital transmission systems and interference (crosstalk) between wire pairs in a cable bundle: AR Kaye and DA George (1970), Branderburg and Wyner (1974), and W. van Etten (1975, 1976). Although these are not examples of exploiting multipath propagation to send multiple information streams, some of the mathematical techniques for dealing with mutual interference proved useful to MIMO development. In the mid-1980s Jack Salz at Bell Laboratories took this research a step further, investigating multi-user systems operating over "mutually cross-coupled linear networks with additive noise sources" such as time-division multiplexing and dually-polarized radio systems. Methods were developed to improve the performance of cell
https://en.wikipedia.org/wiki/The%20Blob%20%28Pacific%20Ocean%29
The Blob is a large mass of relatively warm water in the Pacific Ocean off the coast of North America that was first detected in late 2013 and continued to spread throughout 2014 and 2015. It is an example of a marine heatwave. Sea surface temperatures indicated that the Blob persisted into 2016, but it was initially thought to have dissipated later that year. By September 2016, the Blob resurfaced and made itself known to meteorologists. The warm water mass was unusual for open ocean conditions and was considered to have played a role in the formation of the unusual weather conditions experienced along the Pacific coast of North America during the same time period. The warm waters of the Blob were nutrient-poor and adversely affected marine life. In 2019 another scare was caused by a weaker form of the effect referred as "The Blob 2.0" and in 2021 the appearance of "The Southern Blob" at south of the equator near New Zealand has caused a major effect in South America, particularly Chile and Argentina. Origin The Blob was first detected in October 2013 and early 2014 by Nicholas Bond and his colleagues at the Joint Institute for the Study of the Atmosphere and Ocean of the University of Washington. It was detected when a large circular body of seawater did not cool as expected and remained much warmer than the average normal temperatures for that location and season. Bond, then the state climatologist for Washington, coined the term the Blob, with the term first appearing in an article in the monthly newsletter of the Office of the Washington State Climatologist for June 2014. Description Initially the Blob was reported as being wide and deep. It later expanded and reached a size of long, wide, and deep in June 2014 when the term the Blob was coined. The Blob now hugged the coast of North America from Mexico to Alaska and beyond, stretching more than , and formed three distinct patches: the first, off the coast of Canada, Washington, Oregon, and northern
https://en.wikipedia.org/wiki/Polydivisible%20number
In mathematics a polydivisible number (or magic number) is a number in a given number base with digits abcde... that has the following properties: Its first digit a is not 0. The number formed by its first two digits ab is a multiple of 2. The number formed by its first three digits abc is a multiple of 3. The number formed by its first four digits abcd is a multiple of 4. etc. Definition Let be a positive integer, and let be the number of digits in n written in base b. The number n is a polydivisible number if for all , . Example For example, 10801 is a seven-digit polydivisible number in base 4, as Enumeration For any given base , there are only a finite number of polydivisible numbers. Maximum polydivisible number The following table lists maximum polydivisible numbers for some bases b, where represent digit values 10 to 35. Estimate for Fb(n) and Σ(b) Let be the number of digits. The function determines the number of polydivisible numbers that has digits in base , and the function is the total number of polydivisible numbers in base . If is a polydivisible number in base with digits, then it can be extended to create a polydivisible number with digits if there is a number between and that is divisible by . If is less or equal to , then it is always possible to extend an digit polydivisible number to an -digit polydivisible number in this way, and indeed there may be more than one possible extension. If is greater than , it is not always possible to extend a polydivisible number in this way, and as becomes larger, the chances of being able to extend a given polydivisible number become smaller. On average, each polydivisible number with digits can be extended to a polydivisible number with digits in different ways. This leads to the following estimate for : Summing over all values of n, this estimate suggests that the total number of polydivisible numbers will be approximately Specific bases All numbers are represented in bas
https://en.wikipedia.org/wiki/Tactile%20discrimination
Tactile discrimination is the ability to differentiate information through the sense of touch. The somatosensory system is the nervous system pathway that is responsible for this essential survival ability used in adaptation. There are various types of tactile discrimination. One of the most well known and most researched is two-point discrimination, the ability to differentiate between two different tactile stimuli which are relatively close together. Other types of discrimination like graphesthesia and spatial discrimination also exist but are not as extensively researched. Tactile discrimination is something that can be stronger or weaker in different people and two major conditions, chronic pain and blindness, can affect it greatly. Blindness increases tactile discrimination abilities which is extremely helpful for tasks like reading braille. In contrast, chronic pain conditions, like arthritis, decrease a person's tactile discrimination. One other major application of tactile discrimination is in new prosthetics and robotics which attempt to mimic the abilities of the human hand. In this case tactile sensors function similarly to mechanoreceptors in a human hand to differentiate tactile stimuli. Pathways Somatosensory system The somatosensory system includes multiple types of sensations from the body. This includes light, touch, pain, pressure, temperature, and joint /muscle sense. Each of these are categorized in three different areas: discriminative touch, pain and temperature, and proprioception. Discriminative touch includes touch, pressure, being able to recognize vibrations, etc. Pain and temperature includes the perception of pain/ amounts of pain and the severity of temperatures. The pain and temperature category of sensations also includes itching and tickling. Proprioception includes receptors for everything that occurs below the surface of the skin. This includes sensations on various muscles, joints, and tendons. Each of these three categories ha
https://en.wikipedia.org/wiki/Serre%27s%20modularity%20conjecture
In mathematics, Serre's modularity conjecture, introduced by , states that an odd, irreducible, two-dimensional Galois representation over a finite field arises from a modular form. A stronger version of this conjecture specifies the weight and level of the modular form. The conjecture in the level 1 case was proved by Chandrashekhar Khare in 2005, and a proof of the full conjecture was completed jointly by Khare and Jean-Pierre Wintenberger in 2008. Formulation The conjecture concerns the absolute Galois group of the rational number field . Let be an absolutely irreducible, continuous, two-dimensional representation of over a finite field . Additionally, assume is odd, meaning the image of complex conjugation has determinant -1. To any normalized modular eigenform of level , weight , and some Nebentype character , a theorem due to Shimura, Deligne, and Serre-Deligne attaches to a representation where is the ring of integers in a finite extension of . This representation is characterized by the condition that for all prime numbers , coprime to we have and Reducing this representation modulo the maximal ideal of gives a mod representation of . Serre's conjecture asserts that for any representation as above, there is a modular eigenform such that . The level and weight of the conjectural form are explicitly conjectured in Serre's article. In addition, he derives a number of results from this conjecture, among them Fermat's Last Theorem and the now-proven Taniyama–Weil (or Taniyama–Shimura) conjecture, now known as the modularity theorem (although this implies Fermat's Last Theorem, Serre proves it directly from his conjecture). Optimal level and weight The strong form of Serre's conjecture describes the level and weight of the modular form. The optimal level is the Artin conductor of the representation, with the power of removed. Proof A proof of the level 1 and small weight cases of the conjecture was obtained in 2004 by Chandrashekha
https://en.wikipedia.org/wiki/Ken%20Stroud
Kenneth Arthur Stroud (; Richmond, Surrey, December, 1908 – Hertfordshire township, February 3, 2000) was a mathematician and Principal Lecturer in Mathematics at Lanchester Polytechnic in Coventry, England. He is most widely known as the author of several mathematics textbooks, especially the very popular Engineering Mathematics. Education Stroud held a B.Sc. and a DipEd. Work Stroud was an innovator in programmed learning and the identification of precise learning outcomes, and Nigel Steele calls his textbook Engineering Mathematics, based on the programmed learning approach, "one of the most successful mathematics textbooks ever published." He died in February 2000, aged 91. Bibliography Laplace Transforms: Programmes and Problems, Stanley Thornes Ltd, 1973, and 1978, . Fourier Series and Harmonic Analysis, Nelson Thornes Ltd, 1983, and Stanley Thornes Ltd, 1986, . Engineering Mathematics, Macmillan, 1970, . 6th ed., (with Dexter J. Booth), Industrial Press, 2007, . Advanced Engineering Mathematics (with Dexter J. Booth), 5th ed., Industrial Press, 2011, , 4th ed., Palgrave Macmillan, 2003, . Differential Equations (with Dexter J. Booth), Industrial Press, 2004, . Vector Analysis (with Dexter J. Booth), Industrial Press, 2005, . Complex Variables (with Dexter J. Booth), Industrial Press, 2007, . Linear Algebra (with Dexter J. Booth), Industrial Press, 2008, . Essential Mathematics for Science and Technology: A Self-Learning Guide (with Dexter J. Booth), Industrial Press, 2009, . Further Engineering Mathematics : Programmes and Problems, Palgrave Macmillan, 3 October 1986, . 2nd ed., Springer-Verlag, 1 November 1989, . 2nd ed., Palgrave Macmillan, June 1990, . 3rd Revised Edition, Palgrave Macmillan, 27 March 1996, . Mathematics for engineering technicians, Stanley Thornes Ltd., 1978, . Book 2A : Practical applications, Stanley Thornes Ltd., 1981, .
https://en.wikipedia.org/wiki/SCAPER
SCAPER (S-phase CyclinA Associated Protein residing in the Endoplasmic Reticulum) is a gene located on the long arm of chromosome 15 (15q24.3). It was first identified in 2007. Gene This gene lies on the Crick strand and has 30 exons. Protein The gene encodes a 1399-amino acid protein with a predicted weight of 158 kiloDaltons. It has a C2H2-type zinc finger motif, a putative transmembrane domain, an ER retrieval signal at the C terminus, 4 coiled-coil domains, 6 potential RXL motifs and 6 consensus Cdk phosphorylation sites. Biochemistry The encoded protein is found in the nucleus and endoplasmic reticulum. It is found in all tissues tested. It appears to have a role in the cell cycle. Clinical significance Mutations in this gene have been associated with intellectual disability and retinitis pigmentosa.
https://en.wikipedia.org/wiki/GLI1
Zinc finger protein GLI1 also known as glioma-associated oncogene is a protein that in humans is encoded by the GLI1 gene. It was originally isolated from human glioblastoma cells. Function The Gli proteins are the effectors of Hedgehog (Hh) signaling and have been shown to be involved in cell fate determination, proliferation and patterning in many cell types and most organs during embryo development. In the developing spinal cord the target genes of Gli proteins, that are themselves transcription factors, are arranged into a complex gene regulatory network that translates the extracellular concentration gradient of Sonic hedgehog into different cell fates along the dorsoventral axis. The Gli transcription factors activate/inhibit transcription by binding to Gli responsive genes and by interacting with the transcription complex. The Gli transcription factors have DNA binding zinc finger domains which bind to consensus sequences on their target genes to initiate or suppress transcription. Yoon showed that mutating the Gli zinc finger domain inhibited the proteins effect proving its role as a transcription factor. Gli proteins have an 18-amino acid region highly similar to the α-helical herpes simplex viral protein 16 activation domain. This domain contains a consensus recognition element for the human TFIID TATA box-binding protein associated factor TAFII31. Other proteins such as Missing in Metastasis (MIM/BEG4) have been shown to potentiate the effects of the Gli transcription factors on target gene transcription. Gli and MIM have been shown to act synergistically to induce epidermal growth and MIM + Gli1 overexpressing grafts show similar growth patterns to Shh grafts. Gli family There are three members of the family; Gli1, Gli2 and Gli3 which are all transcription factors mediating the Hh pathway. The GLI1, GLI2, and GLI3 genes encode transcription factors which all contain conserved tandem C2-H2 zinc finger domains and a consensus histidine/cysteine
https://en.wikipedia.org/wiki/Greater-than%20sign
The greater-than sign is a mathematical symbol that denotes an inequality between two values. The widely adopted form of two equal-length strokes connecting in an acute angle at the right, , has been found in documents dated as far back as 1631. In mathematical writing, the greater-than sign is typically placed between two values being compared and signifies that the first number is greater than the second number. Examples of typical usage include and . The less-than sign and greater-than sign always "point" to the smaller number. Since the development of computer programming languages, the greater-than sign and the less-than sign have been repurposed for a range of uses and operations. History The earliest known use of the symbols and is found in (The Analytical Arts Applied to Solving Algebraic Equations) by Thomas Harriot, published posthumously in 1631. The text states " a > b a b (The sign of majority a > b indicates that a is greater than b)" and " a < b a b (The sign of minority a < b indicates that a is less than b)." According to historian Art Johnson, while Harriot was surveying North America, he saw a Native American with a symbol that resembled the greater-than sign, in both backwards and forwards forms. Johnson says it is likely Harriot developed the two symbols from this symbol. Usage in text markup Angle brackets The greater-than sign is sometimes used for an approximation of the closing angle bracket, . The proper Unicode character is . ASCII does not have angular brackets. HTML In HTML (and SGML and XML), the greater-than sign is used at the end of tags. The greater-than sign may be included with , while produces the greater-than or equal to sign. E-mail and Markdown In some early e-mail systems, the greater-than sign was used to denote quotations. The sign is also used to denote quotations in Markdown. Usage in programming The 'greater-than sign' is encoded in ASCII as character hex 3E, decimal 62. The Unicode code point is , i
https://en.wikipedia.org/wiki/Coping
Coping refers to conscious strategies used to reduce unpleasant emotions. Coping strategies can be cognitions or behaviors and can be individual or social. Coping is to deal with and overcome struggles and difficulties in life. It is a way for us to maintain our mental and emotional well-being. Everybody has a way of handling the hard events that occur in our life and that is what it means to cope. Coping can be healthy and productive, or destructive and unhealthy for you or others. It is recommended that an individual copes in ways that will be beneficial and healthy. "Managing your stress well can help you feel better physically and psychologically and it can impact your ability to perform your best." Theories of coping Hundreds of coping strategies have been proposed in an attempt to understand how people cope. Classification of these strategies into a broader architecture has not been agreed upon. Researchers try to group coping responses rationally, empirically by factor analysis, or through a blend of both techniques. In the early days, Folkman and Lazarus split the coping strategies into four groups, namely problem-focused, emotion-focused, support-seeking, and meaning-making coping. Weiten has identified four types of coping strategies: appraisal-focused (adaptive cognitive), problem-focused (adaptive behavioral), emotion-focused, and occupation-focused coping. Billings and Moos added avoidance coping as one of the emotion-focused coping. Some scholars have questioned the psychometric validity of forced categorization as those strategies are not independent to each other. Besides, in reality, people can adopt multiple coping strategies simultaneously. Typically, people use a mixture of several types of coping strategies, which may change over time. All these strategies can prove useful, but some claim that those using problem-focused coping strategies will adjust better to life. Problem-focused coping mechanisms may allow an individual greater perceived co
https://en.wikipedia.org/wiki/The%20Hidden%20Wiki
The Hidden Wiki was a dark web MediaWiki wiki operating as Tor hidden services that could be anonymously edited after registering on the site. The main page served as a directory of links to other .onion sites. History The first Hidden Wiki was operated through the .onion pseudo top-level domain which can be accessed only by using Tor or a Tor gateway. Its main page provided a community-maintained link directory to other hidden services, including links claiming to offer money laundering, contract killing, cyber-attacks for hire, contraband chemicals, and bomb making. The rest of the wiki was essentially uncensored as well and also offered links to sites hosting child pornography and abuse images. The earliest mention of the hidden wiki is from 2007 when it was located at 6sxoyfb3h2nvok2d.onion. A well known iteration of the Hidden Wiki was founded some time before October 2011, coming to prominence with its associations with illegal content. At some point prior to August 2013, the site was hosted on Freedom Hosting. In March 2014 the site and its kpvz7ki2v5agwt35.onion domain was hacked and redirected to Doxbin. Following this event, the content began to be mirrored to more locations. During Operation Onymous in November 2014, after its Bulgarian hosting was compromised, the site served a message from law enforcement. Successors There are several .onion websites hosting successors based on mirrors of the Hidden Wiki; as such, there is no longer one single official Hidden Wiki. Many are hosted for accessibility reasons, due to frequent downtime and instability of the main wiki, while others were launched in order to filter links to child pornography. See also
https://en.wikipedia.org/wiki/Fighters%20against%20Nazis%20Medal
Fighters against Nazis Medal ( Ot Halohem BaNatzim) is an Israeli decoration that is awarded to World War II veterans. First instituted as a ribbon bar in 1967, it was first awarded to World War II veterans at Yom HaShoah (7 May) the same year by the Prime Minister of Israel, Levi Eshkol. In 1985 to mark 40 years to victory in World War II, an Israeli flag with the ribbon bar was awarded to a number of museums dedicated to the war. In 2000, a medal was created from the ribbon bar. Award criteria An Israeli citizen or permanent resident of Israel who during World War II was engaged in the military campaign against the Nazi oppressors and their supporters and fought them in one of the Allies armed forces, including the British brigade, as a partisan or as an underground movement member during the set period of time which is between 1/9/1939 and the 1/9/1945, as specified by Yad Vashem regulation of 1968, and also based upon the "status of the World War II veterans" Act of 2000. If the person eligible for this award has died, a family member (a spouse, a son, a daughter, a father, a mother, a brother, a sister, a grandson, a granddaughter) is entitled to submit an application requesting the ribbon, or an equivalent to the ribbon in event of loss or wear and tear. Design The Fighters against Nazis Medal is a disc. The non-swivelling straight bar suspender is attached to the medal at the top of the medal. The medals were struck on a gold coloured metal. Obverse The obverse shows two Stars of David; one of them is shaped as a Yellow badge and has a bayonet on it while the other has an olive branch on it. Reverse The reverse shows the Emblem of the State of Israel, a menorah surrounded by an olive branch on each side, and the writing "ישראל" (Hebrew for Israel) below it, in the centre; around it the writing "לוחם בנאצים - ותיק מלחמת העולם השנייה" (Hebrew for Fighter against the Nazis - veteran of World War II). Ribbon The ribbon is red in the centre with a str
https://en.wikipedia.org/wiki/Mosaicity
In crystallography, mosaicity is a measure of the spread of crystal plane orientations. A mosaic crystal is an idealized model of an imperfect crystal, imagined to consist of numerous small perfect crystals (crystallites) that are to some extent randomly misoriented. Empirically, mosaicities can be determined by measuring rocking curves. Diffraction by mosaics is described by the Darwin–Hamilton equations. The mosaic crystal model goes back to a theoretical analysis of X-ray diffraction by C. G. Darwin (1922). Currently, most studies follow Darwin in assuming a Gaussian distribution of crystallite orientations centered on some reference orientation. The mosaicity is commonly equated with the standard deviation of this distribution. Applications and notable materials An important application of mosaic crystals is in monochromators for x-ray and neutron radiation. The mosaicity enhances the reflected flux, and allows for some phase-space transformation. Pyrolitic graphite (PG) can be produced in form of mosaic crystals (HOPG: highly ordered PG) with controlled mosaicity of up to a few degrees. Diffraction by mosaic crystals: the Darwin–Hamilton equations To describe diffraction by a thick mosaic crystal, it is usually assumed that the constituent crystallites are so thin that each of them reflects at most a small fraction of the incident beam. Primary extinction and other dynamical diffraction effects can then be neglected. Reflections by different crystallites add incoherently, and can therefore be treated by classical transport theory. When only beams within the scattering plane are considered, then they obey the Darwin–Hamilton equations (Darwin 1922, Hamilton 1957), where are the directions of the incident and diffracted beam, are the corresponding currents, μ is the Bragg reflectivity, and σ accounts for losses by absorption and by thermal and elastic diffuse scattering. A generic analytical solution has been obtained remarkably late (Sears 1997; for t
https://en.wikipedia.org/wiki/Annual%20Review%20of%20Phytopathology
The Annual Review of Phytopathology is a peer-reviewed academic journal that publishes review articles about phytopathology, the study of diseases that affect plants. It was first published in 1963 as the result of a collaboration between the American Phytopathological Society and the nonprofit publisher Annual Reviews. As of 2023, Journal Citation Reports lists the journal's 2022 impact factor as 10.2, ranking it ninth of 238 journal titles in the category "Plant Sciences". Its current editors are Jan E. Leach and Steven E. Lindow. History In the 1950s, the American Phytopathological Society had intended to publish its own journal to cover significant developments in the field of phytopathology, or plant diseases. However, the nonprofit publisher Annual Reviews offered to publish the journal for them, and they agreed due to their publishing experience. In 1961, the American Phytopathological Society compiled the editorial board of the journal at their annual meeting. The first volume was published in 1963. It was the twelfth journal title published by Annual Reviews. In its first ten volumes, it published a total of 188 reviews from authors from twenty-one countries. It defines its scope as covering significant developments related to plant pathology, including diagnosis, plant pathogens, host–pathogen interactions, epidemiology of plant disease, breeding plants for disease resistance, and plant disease management. It is abstracted and indexed in Scopus, Science Citation Index Expanded, Civil Engineering Abstracts, Inspec, and Academic Search, among others. Editorial processes The Annual Review of Phytopathology is helmed by the editor or the co-editors. The editor is assisted by the editorial committee, which includes associate editors, regular members, and occasionally guest editors. Guest members participate at the invitation of the editor, and serve terms of one year. All other members of the editorial committee are appointed by the Annual Reviews board of
https://en.wikipedia.org/wiki/Clinical%20pharmaceutical%20chemistry
Clinical pharmaceutical chemistry is a specialty branch of chemical sciences, which consists of medicinal chemistry with additional training in clinical aspects of translational sciences and medicine. Typically this involves similar principal training as in general medicine, where inspection of and interaction with the patients are a vital part of the training. Typically students in clinical pharmaceutical chemistry use the same curriculum as medical students, but specialize in medicinal and organic chemistry after and during the theoretical/early clinical studies. In clinical pharmaceutical chemistry the aim is to understand biological transformations and processes associated with chemical entities inside the human body, and how those processes can be influenced with changes in chemical structures. The aim of clinical pharmaceutical chemistry is in addition to manage and manipulate clinical effects of different chemical structures, as well as to manage phenomena recognized in first-in-human studies. Typically clinical pharmaceutical chemistry has an important role in discovery, design and manipulation of new drug entities, and is vital especially in early clinical studies (such as Phase I studies). See also Medicinal chemistry
https://en.wikipedia.org/wiki/Radius%20of%20curvature
In differential geometry, the radius of curvature, , is the reciprocal of the curvature. For a curve, it equals the radius of the circular arc which best approximates the curve at that point. For surfaces, the radius of curvature is the radius of a circle that best fits a normal section or combinations thereof. Definition In the case of a space curve, the radius of curvature is the length of the curvature vector. In the case of a plane curve, then is the absolute value of where is the arc length from a fixed point on the curve, is the tangential angle and is the curvature. Formula In two dimensions If the curve is given in Cartesian coordinates as , i.e., as the graph of a function, then the radius of curvature is (assuming the curve is differentiable up to order 2) where and denotes the absolute value of . If the curve is given parametrically by functions and , then the radius of curvature is where and Heuristically, this result can be interpreted as where In dimensions If is a parametrized curve in then the radius of curvature at each point of the curve, , is given by As a special case, if is a function from to , then the radius of curvature of its graph, , is Derivation Let be as above, and fix . We want to find the radius of a parametrized circle which matches in its zeroth, first, and second derivatives at . Clearly the radius will not depend on the position , only on the velocity and acceleration . There are only three independent scalars that can be obtained from two vectors and , namely , , and . Thus the radius of curvature must be a function of the three scalars , and . The general equation for a parametrized circle in is where is the center of the circle (irrelevant since it disappears in the derivatives), are perpendicular vectors of length (that is, and ), and is an arbitrary function which is twice differentiable at . The relevant derivatives of work out to be If we now equate these derivatives of to
https://en.wikipedia.org/wiki/Graviphoton
In theoretical physics and quantum physics, a graviphoton or gravivector is a hypothetical particle which emerges as an excitation of the metric tensor (i.e. gravitational field) in spacetime dimensions higher than four, as described in Kaluza–Klein theory. However, its crucial physical properties are analogous to a (massive) photon: it induces a "vector force", sometimes dubbed a "fifth force". The electromagnetic potential emerges from an extra component of the metric tensor , where the figure 5 labels an additional, fifth dimension. In gravity theories with extended supersymmetry (extended supergravities), a graviphoton is normally a superpartner of the graviton that behaves like a photon, and is prone to couple with gravitational strength, as was appreciated in the late 1970s. Unlike the graviton, it may provide a repulsive (as well as an attractive) force, and thus, in some technical sense, a type of anti-gravity. Under special circumstances, in several natural models, often descending from five-dimensional theories mentioned, it may actually cancel the gravitational attraction in the static limit. Joël Scherk investigated semirealistic aspects of this phenomenon, stimulating searches for physical manifestations of this mechanism. See also Graviscalar (a.k.a. radion) Supergravity
https://en.wikipedia.org/wiki/Ballooning%20instability
The ballooning instability (a.k.a. ballooning mode instability) is a type of internal pressure-driven plasma instability usually seen in tokamak fusion power reactors or in space plasmas. It is important in fusion research as it determines a set of criteria for the maximum achievable plasma beta. The name refers to the shape and action of the instability, which acts like the elongations formed in a long balloon when it is squeezed. In literature, the structure of these elongations are commonly referred to as 'fingers'. The narrow fingers of plasma produced by the instability are capable of accelerating and pushing aside the surrounding magnetic field in order to cause a sudden, explosive release of energy. Thus, the instability is also known as the explosive instability. Relation to interchange instability The interchange instability can be derived from the equations of the ballooning instability as a special case in which the ballooning mode does not perturb the equilibrium magnetic field. This special limit is known as the Mercier criterion.
https://en.wikipedia.org/wiki/Fleet%20Radio%20Unit
Fleet Radio Units (FRU) were the major centers for Allied cryptological and signals intelligence during the Pacific Campaign of World War II. Initially two FRUs were established in the Pacific, one at Pearl Harbor, Hawaii, called Station HYPO or FRUPAC (Fleet Radio Unit, Pacific), and the other, called Station CAST or Belconnen, at Cavite Naval Yard, then Corregidor, Philippines. With the fall of the Philippines to Imperial Japanese forces in April and May 1942, CAST personnel were evacuated to a newly established FRU at Melbourne, Australia, called FRUMEL (Fleet Radio Unit, Melbourne). HYPO and FRUMEL supervised detached field units scattered at various locations and aboard ships throughout the south, central, and north Pacific areas until the end of World War II. A third center, NEGAT, was based at OP-20-G headquarters in Washington DC. The entire cryptanalysis effort conducted by the units was called Operation Magic. Liaison between the units and with the Far East Combined Bureau (FECB; the British signals intelligence center at Colombo, Ceylon in 1942) was titled Copek.
https://en.wikipedia.org/wiki/Hugo%20Steinhaus
Hugo Dyonizy Steinhaus ( ; ; January 14, 1887 – February 25, 1972) was a Polish mathematician and educator. Steinhaus obtained his PhD under David Hilbert at Göttingen University in 1911 and later became a professor at the Jan Kazimierz University in Lwów (now Lviv, Ukraine), where he helped establish what later became known as the Lwów School of Mathematics. He is credited with "discovering" mathematician Stefan Banach, with whom he gave a notable contribution to functional analysis through the Banach–Steinhaus theorem. After World War II Steinhaus played an important part in the establishment of the mathematics department at Wrocław University and in the revival of Polish mathematics from the destruction of the war. Author of around 170 scientific articles and books, Steinhaus has left his legacy and contribution in many branches of mathematics, such as functional analysis, geometry, mathematical logic, and trigonometry. Notably he is regarded as one of the early founders of game theory and probability theory which led to later development of more comprehensive approaches by other scholars. Early life and studies Steinhaus was born on January 14, 1887, in Jasło, Austria-Hungary to a family with Jewish roots. His father, Bogusław, was a local industrialist, owner of a brick factory and a merchant. His mother was Ewelina, née Lipschitz. Hugo's uncle, Ignacy Steinhaus, was a lawyer and an activist in the Koło Polskie (Polish Circle), and a deputy to the Galician Diet, the regional assembly of the Kingdom of Galicia and Lodomeria. Hugo finished his studies at the gymnasium in Jasło in 1905. His family wanted him to become an engineer but he was drawn to abstract mathematics and began to study the works of famous contemporary mathematicians on his own. In the same year he began studying philosophy and mathematics at the University of Lemberg. In 1906 he transferred to Göttingen University. At that University he received his Ph.D. in 1911, having written his doctora
https://en.wikipedia.org/wiki/Synnema
A synnema (plural synnemata, also coremia; derivation: "Threads together") is a large, erect reproductive structure borne by some fungi, bearing compact conidiophores, which fuse together to form a strand resembling a stalk of wheat, with conidia at the end or on the edges. Fungal genera which bear synnemata include Doratomyces.
https://en.wikipedia.org/wiki/Necroptosis
Necroptosis is a programmed form of necrosis, or inflammatory cell death. Conventionally, necrosis is associated with unprogrammed cell death resulting from cellular damage or infiltration by pathogens, in contrast to orderly, programmed cell death via apoptosis. The discovery of necroptosis showed that cells can execute necrosis in a programmed fashion and that apoptosis is not always the preferred form of cell death. Furthermore, the immunogenic nature of necroptosis favors its participation in certain circumstances, such as aiding in defence against pathogens by the immune system. Necroptosis is well defined as a viral defense mechanism, allowing the cell to undergo "cellular suicide" in a caspase-independent fashion in the presence of viral caspase inhibitors to restrict virus replication. In addition to being a response to disease, necroptosis has also been characterized as a component of inflammatory diseases such as Crohn's disease, pancreatitis, and myocardial infarction. The signaling pathway responsible for carrying out necroptosis is generally understood. TNFα leads to stimulation of its receptor TNFR1. TNFR1 binding protein TNFR-associated death protein TRADD and TNF receptor-associated factor 2 TRAF2 signals to RIPK1 which recruits RIPK3 forming the necrosome also named ripoptosome. Phosphorylation of MLKL by the ripoptosome drives oligomerization of MLKL, allowing MLKL to insert into and permeabilize plasma membranes and organelles. Integration of MLKL leads to the inflammatory phenotype and release of damage-associated molecular patterns (DAMPs), which elicit immune responses. Function Necroptosis is specific to vertebrates and may have originated as an additional defense to pathogens. Necroptosis also acts as an alternative "fail-safe" cell death pathway in cases where cells are unable to undergo apoptosis, such as during viral infection in which apoptosis signaling proteins are blocked by the virus. In innate immunity Cell suicide is an effec
https://en.wikipedia.org/wiki/Myth%20of%20the%20flat%20Earth
The myth of the flat Earth, or the flat-Earth error, is a modern historical misconception that European scholars and educated people during the Middle Ages believed the Earth to be flat. The earliest clear documentation of the idea of a spherical Earth comes from the ancient Greeks (5th century BC). The belief was widespread in the Greek world when Eratosthenes calculated the circumference of Earth around 240 BC. This knowledge spread with Greek influence such that during the Early Middle Ages (~600–1000 AD), most European and Middle Eastern scholars espoused Earth's sphericity. Belief in a flat Earth among educated Europeans was almost nonexistent from the Late Middle Ages onward, though fanciful depictions appear in art, such as the exterior panels of Hieronymus Bosch's famous triptych The Garden of Earthly Delights, in which a disc-shaped Earth is shown floating inside a transparent sphere. According to Stephen Jay Gould, "there never was a period of 'flat Earth darkness' among scholars, regardless of how the public at large may have conceptualized our planet both then and now. Greek knowledge of sphericity never faded, and all major medieval scholars accepted the Earth's roundness as an established fact of cosmology." Historians of science David Lindberg and Ronald Numbers point out that "there was scarcely a Christian scholar of the Middle Ages who did not acknowledge [Earth's] sphericity and even know its approximate circumference". Historian Jeffrey Burton Russell says the flat-Earth error flourished most between 1870 and 1920, and had to do with the ideological setting created by struggles over biological evolution. Russell claims "with extraordinary few exceptions no educated person in the history of Western Civilization from the third century B.C. onward believed that the Earth was flat", and ascribes popularization of the flat-Earth myth to histories by John William Draper, Andrew Dickson White, and Washington Irving. History In Inventing the Flat Ea
https://en.wikipedia.org/wiki/Leonida%20Tonelli
Leonida Tonelli (19 April 1885 – 12 March 1946) was an Italian mathematician, noted for proving Tonelli's theorem, a variation of Fubini's theorem, and for introducing semicontinuity methods as a common tool for the direct method in the calculus of variations. Education Tonelli graduated from the University of Bologna in 1907; his Ph.D. thesis was written under the direction of Cesare Arzelà. Work Selected publications , 1900 . Zanichelli, Bologna, vol. 1: 1922, vol. 2: 1923 . Zanichelli, Bologna 1928 See also Calculus of variations Fourier series Lebesgue integral Mathematical analysis Notes
https://en.wikipedia.org/wiki/Karl%20Landsteiner%20Memorial%20Award
The Karl Landsteiner Memorial Award is a scientific award given by the American Association of Blood Banks (AABB) to scientists with "an international reputation in transfusion medicine or cellular therapies" "whose original research resulted in an important contribution to the body of scientific knowledge". Recipients give a lecture at the AABB Annual Meeting and receive a $7,500 honorarium. The prize was initiated in 1954 to honor Karl Landsteiner, whose research laid the foundation for modern blood transfusion therapy. Recipients 1954 Reuben Ottenberg 1955 Richard Lewisohn 1956 Philip Levine, Alexander Solomon Wiener 1957 Ruth Sanger, Robert Russell Race 1958 Oswald Hope Robertson, Francis Peyton Rous, J. R. Turner 1959 Ernest Witebsky 1960 Patrick L. Mollison 1961 Robert R. A. Coombs 1962 William C. Boyd 1963 Fred H. Allen Jr., Louis K. Diamond 1964 J. J. van Loghem 1965 Ruggero Ceppellini 1966 Elvin A. Kabat 1967 Walter Morgan, Winifred Watkins 1968 Rodney R. Porter 1969 Vincent J. Freda, John G. Gorman, William Pollack 1970 Jean Dausset 1971 Bruce Chown, Marion Lewis 1972 Richard E. Rosenfield 1973 Arthur E. Mourant 1974 Manfred M. Mayer, Hans J. Müller-Eberhard 1975 Baruch S. Blumberg, Alfred M. Prince 1976 Marie Cutbush Crookston, Eloise R. Giblett 1977 Rose Payne, Jon van Rood 1978 Fred Stratton 1979 Nevin C. Hughes-Jones, Serafeim P. Masouredis 1980 Donald M. Marcus, James M. Stavely 1981 James F. Danielli, S. Jonathan Singer 1982 Georges J. F. Köhler, César Milstein 1983 Vincent T. Marchesi 1984 Oliver Smithies 1985 Saul Krugman 1986 Claes F. Högman, Grant R. Bartlett 1987 E. Donnall Thomas 1988 Charles P. Salmon 1989 George W. Bird 1990 Robert Gallo, Luc Montagnier 1991 Paul I. Terasaki 1992 Harvey J. Alter, Daniel W. Bradley, Qui-Lim Choo, Michael Houghton, George Kuo, Lacy Overby 1993 C. Paul Engelfriet 1994 Kenneth Brinkhous, Harold Roberts, Robert Wagner, Robert Langdell 1995 W. Laurence Marsh 1996 Eugene G
https://en.wikipedia.org/wiki/Diphasiastrum%20%C3%97%20sabinifolium
Diphasiastrum × sabinifolium, the savinleaf groundpine or savin leaf club moss, is hybrid of D. sitchense and D. tristachyum. It can be found in North America from Labrador and Newfoundland to Ontario, and south to Pennsylvania and Michigan. Erect stems can reach 20 centimeters high, and branch dichotomously. The sterile branches are flattened, and the leaves are 4-ranked. Peduncles are 1-8 centimeters long. In many disturbed sites, it can be found growing alongside D. sitchense, and can be distinguished by flattened branchlets split into four ranks, as opposed to those of D. sitchense, which generally are rounded and split into five ranks.
https://en.wikipedia.org/wiki/Imaging%20spectroscopy
In imaging spectroscopy (also hyperspectral imaging or spectral imaging) each pixel of an image acquires many bands of light intensity data from the spectrum, instead of just the three bands of the RGB color model. More precisely, it is the simultaneous acquisition of spatially coregistered images in many spectrally contiguous bands. Some spectral images contain only a few image planes of a spectral data cube, while others are better thought of as full spectra at every location in the image. For example, solar physicists use the spectroheliograph to make images of the Sun built up by scanning the slit of a spectrograph, to study the behavior of surface features on the Sun; such a spectroheliogram may have a spectral resolution of over 100,000 () and be used to measure local motion (via the Doppler shift) and even the magnetic field (via the Zeeman splitting or Hanle effect) at each location in the image plane. The multispectral images collected by the Opportunity rover, in contrast, have only four wavelength bands and hence are only a little more than 3-color images. One application is spectral geophysical imaging, which allows quantitative and qualitative characterization of the surface and of the atmosphere, using radiometric measurements. These measurements can then be used for unambiguous direct and indirect identification of surface materials and atmospheric trace gases, the measurement of their relative concentrations, subsequently the assignment of the proportional contribution of mixed pixel signals (e.g., the spectral unmixing problem), the derivation of their spatial distribution (mapping problem), and finally their study over time (multi-temporal analysis). The Moon Mineralogy Mapper on Chandrayaan-1 was a geophysical imaging spectrometer. Background In 1704, Sir Isaac Newton demonstrated that white light could be split up into component colours. The subsequent history of spectroscopy led to precise measurements and provided the empirical foundatio
https://en.wikipedia.org/wiki/Valentino%27s%20syndrome
Valentino's syndrome is pain presenting in the right lower quadrant of the abdomen caused by a duodenal ulcer with perforation through the retroperitoneum. It is named after Rudolph Valentino, an Italian actor, who presented with right lower quadrant pain in New York, which turned out to be a perforated peptic ulcer. He subsequently died from an infection and organ dysfunction in spite of surgery to repair the perforation. Due to his popularity, his case received much attention at the time and is still considered a rare medical condition. However, the degree of peritoneal findings is strongly influenced by a number of factors, including the size of perforation, amount of bacterial and gastric contents contaminating the abdominal cavity, time between perforation and presentation, and spontaneous sealing of perforation. Signs and symptoms Patients with perforated Valentino's syndrome usually present with a sudden onset of severe, sharp abdominal pain in the right lower quadrant (RLQ), that is similar to acute appendicitis. Most patients describe generalized pain; a few present with severe epigastric pain, located in the upper abdominal area. As even slight movement can tremendously worsen their pain, these patients assume a fetal position. These patients may also demonstrate signs and symptoms of septic shock, such as tachycardia (increased heart rate), hypotension (low blood pressure), and anuria (when no urine is produced from the kidneys). Not surprisingly, these indicators of shock may be absent in elderly, immunocompromised patients or in those with diabetes. Patients also experience nausea, vomiting, decreased appetite, and sweating. Cause The cause for Valentino's syndrome is due to a perforated ulcer located in the duodenum. This occurs when ulcers that have gone untreated for long periods of time, and as a result has burned through the stomach wall. Risk factors for a perforated ulcers include bacterial infection, such as H. pylori, and routine use of n
https://en.wikipedia.org/wiki/Security%20convergence
Security convergence refers to the convergence of two historically distinct security functions – physical security and information security – within enterprises; both are integral parts of a coherent risk management program. Security convergence is motivated by the recognition that corporate assets are increasingly information-based. In the past, physical assets demanded the bulk of protection efforts, whereas information assets are demanding increasing attention. Although generally used in relation to cyber-physical convergence, security convergence can also refer to the convergence of security with related risk and resilience disciplines, including business continuity planning and emergency management. Security convergence is often referred to as 'converged security'. Definitions According to the United States Cybersecurity and Infrastructure Security Agency, security convergence is the "formal collaboration between previously disjointed security functions." Survey participants in an ASIS Foundation study The State of Security Convergence in the United States, Europe, and India define security convergence as "getting security/risk management functions to work together seamlessly, closing the gaps and vulnerabilities that exist in the space between functions." In his book Security Convergence: Managing Enterprise Security Risk, Dave Tyson defines security convergence as "the integration of the cumulative security resources of an organization in order to deliver enterprise-wide benefits through enhanced risk mitigation, increased operational effectiveness and efficiency, and cost savings." Background The concept of security convergence has gained currency within the context of the Fourth Industrial Revolution, which, according to founder and Executive Chairman of the World Economic Forum (WEF) Klaus Schwab, "is characterised by a fusion of technologies that is blurring the lines between the physical, digital, and biological spheres." Key results of this fusion
https://en.wikipedia.org/wiki/Appeal%20to%20loyalty
The appeal to loyalty is a logical fallacy committed when the premise of an argument uses a perceived need for loyalty of some sort to distract from the issue being discussed. Example B questions A's statement of x. Anyone who questions A is disloyal. Therefore, B is wrong. Problem: Even if B is disloyal, that doesn't mean that B is wrong, as A is not necessarily always right. See also Groupthink Blood is thicker than water
https://en.wikipedia.org/wiki/Spore%20Hero%20Arena
Spore Hero Arena is a science fiction action-adventure game for the Nintendo DS and is a spin-off of Spore Hero which was released in North America on October 6, 2009, and worldwide on October 8. This game is the second Spore game for the Nintendo DS, and focuses more on battling than its predecessor, Spore Creatures. Creatures are now rendered in full 3D and the creature design system has been overhauled. The online functions/multiplayer were closed on 30 June 2014. Gameplay Spore Hero Arena allows the player to move around an overworld with elements of exploration and puzzle solving. On each planet there are four blue shards to collect by doing certain missions or jobs assigned to the player by the creatures that inhabit that world. These missions can vary from being a bodyguard, racing, and arena battles. Once the player gets all four a battle can be started with specific creature at the front of the champion's arena. The player's job is to enter and fight the champion to obtain the evil red fragments. The main battle mode takes place on a top down arena. Players control their creature in both the overworld mode and battle mode via touchscreen and attack by using the directional pad or the A, B, X, Y buttons; though there is an option to have the Directional pad control movement. The player can set three different types of bio powers (special attacks) to be used in a battle, which can be used by pressing the L or R button and tapping one of three options using the stylus when the bio-power gauge is filled. The three types of bio-power focus on eliminating one enemy, damaging multiple enemies, and affecting the field respectively, and there are multiple kinds for each type. There are multiple modes of battles, including the basic Battle Royale, where players fight to knock their opponents off of the arena which is reminiscent of sumo, and a capture the flag mode in which players attempt to obtain an egg. While the player does have a health bar, it is similar to
https://en.wikipedia.org/wiki/Cyclin%20D
Cyclin D is a member of the cyclin protein family that is involved in regulating cell cycle progression. The synthesis of cyclin D is initiated during G1 and drives the G1/S phase transition. Cyclin D protein is anywhere from 155 (in zebra mussel) to 477 (in Drosophila) amino acids in length. Once cells reach a critical cell size (and if no mating partner is present in yeast) and if growth factors and mitogens (for multicellular organism) or nutrients (for unicellular organism) are present, cells enter the cell cycle. In general, all stages of the cell cycle are chronologically separated in humans and are triggered by cyclin-Cdk complexes which are periodically expressed and partially redundant in function. Cyclins are eukaryotic proteins that form holoenzymes with cyclin-dependent protein kinases (Cdk), which they activate. The abundance of cyclins is generally regulated by protein synthesis and degradation through APC/C- and CRL-dependent pathways. Cyclin D is one of the major cyclins produced in terms of its functional importance. It interacts with four Cdks: Cdk2, 4, 5, and 6. In proliferating cells, cyclin D-Cdk4/6 complex accumulation is of great importance for cell cycle progression. Namely, cyclin D-Cdk4/6 complex partially phosphorylates retinoblastoma tumor suppressor protein (Rb), whose inhibition can induce expression of some genes (for example: cyclin E) important for S phase progression. Drosophila and many other organisms only have one cyclin D protein. In mice and humans, two more cyclin D proteins have been identified. The three homologues, called cyclin D1, cyclin D2, and cyclin D3 are expressed in most proliferating cells and the relative amounts expressed differ in various cell types. Homologues The most studied homologues of cyclin D are found in yeast and viruses. The yeast homologue of cyclin D, referred to as CLN3, interacts with Cdc28 (cell division control protein) during G1. In viruses, like Saimiriine herpesvirus 2 (Herpesvirus sa
https://en.wikipedia.org/wiki/Serial%20port
On computers, a serial port is a serial communication interface through which information transfers in or out sequentially one bit at a time. This is in contrast to a parallel port, which communicates multiple bits simultaneously in parallel. Throughout most of the history of personal computers, data has been transferred through serial ports to devices such as modems, terminals, various peripherals, and directly between computers. While interfaces such as Ethernet, FireWire, and USB also send data as a serial stream, the term serial port usually denotes hardware compliant with RS-232 or a related standard, such as RS-485 or RS-422. Modern consumer personal computers (PCs) have largely replaced serial ports with higher-speed standards, primarily USB. However, serial ports are still frequently used in applications demanding simple, low-speed interfaces, such as industrial automation systems, scientific instruments, point of sale systems and some industrial and consumer products. Server computers may use a serial port as a control console for diagnostics, while networking hardware (such as routers and switches) commonly use serial console ports for configuration, diagnostics, and emergency maintenance access. To interface with these and other devices, USB-to-serial converters can quickly and easily add a serial port to a modern PC. Hardware Modern devices use an integrated circuit called a UART to implement a serial port. This IC converts characters to and from asynchronous serial form, implementing the timing and framing of data specified by the serial protocol in hardware. The IBM PC implements its serial ports, when present, with one or more UARTs. Very low-cost systems, such as some early home computers, would instead use the CPU to send the data through an output pin, using the bit banging technique. These early home computers often had proprietary serial ports with pinouts and voltage levels incompatible with RS-232. Before large-scale integration (LSI) m
https://en.wikipedia.org/wiki/Digital%20down%20converter
In digital signal processing, a digital down-converter (DDC) converts a digitized, band-limited signal to a lower frequency signal at a lower sampling rate in order to simplify the subsequent radio stages. The process can preserve all the information in the frequency band of interest of the original signal. The input and output signals can be real or complex samples. Often the DDC converts from the raw radio frequency or intermediate frequency down to a complex baseband signal. Architecture A DDC consists of three subcomponents: a direct digital synthesizer (DDS), a low-pass filter (LPF), and a downsampler (which may be integrated into the low-pass filter). The DDS generates a complex sinusoid at the intermediate frequency (IF). Multiplication of the intermediate frequency with the input signal creates images centered at the sum and difference frequency (which follows from the frequency shifting properties of the Fourier transform). The lowpass filters pass the difference (i.e. baseband) frequency while rejecting the sum frequency image, resulting in a complex baseband representation of the original signal. Assuming judicious choice of IF and LPF bandwidth, the complex baseband signal is mathematically equivalent to the original signal. In its new form, it can readily be downsampled and is more convenient to many DSP algorithms. Any suitable low-pass filter can be used including FIR, IIR and CIC filters. The most common choice is a FIR filter for low amounts of decimation (less than ten) or a CIC filter followed by a FIR filter for larger downsampling ratios. Variations on the DDC Several variations on the DDC are useful, including many that input a feedback signal into the DDS. These include: Decision directed carrier recovery phase locked loops in which the I and Q are compared to the nearest ideal constellation point of a PSK signal, and the resulting error signal is filtered and fed back into the DDS A Costas loop in which the I and Q are multiplied
https://en.wikipedia.org/wiki/Zombie%20%28computing%29
In computing, a zombie is a computer connected to the Internet that has been compromised by a hacker via a computer virus, computer worm, or trojan horse program and can be used to perform malicious tasks under the remote direction of the hacker. Zombie computers often coordinate together in a botnet controlled by the hacker, and are used for activities such as spreading e-mail spam and launching distributed denial-of-service attacks (DDoS attacks) against web servers. Most victims are unaware that their computers have become zombies. The concept is similar to the zombie of Haitian Voodoo folklore, which refers to a corpse resurrected by a sorcerer via magic and enslaved to the sorcerer's commands, having no free will of its own. A coordinated DDoS attack by multiple botnet machines also resembles a "zombie horde attack", as depicted in fictional zombie films. Advertising Zombie computers have been used extensively to send e-mail spam; as of 2005, an estimated 50–80% of all spam worldwide was sent by zombie computers. This allows spammers to avoid detection and presumably reduces their bandwidth costs, since the owners of zombies pay for their own bandwidth. This spam also greatly increases the spread of Trojan horses, as Trojans are not self-replicating. They rely on the movement of e-mails or spam to grow, whereas worms can spread by other means. For similar reasons, zombies are also used to commit click fraud against sites displaying pay-per-click advertising. Others can host phishing or money mule recruiting websites. Distributed denial-of-service attacks Zombies can be used to conduct distributed denial-of-service (DDoS) attacks, a term which refers to the orchestrated flooding of target websites by large numbers of computers at once. The large number of Internet users making simultaneous requests of a website's server is intended to result in crashing and the prevention of legitimate users from accessing the site. A variant of this type of flooding is kn
https://en.wikipedia.org/wiki/Food%20orange%207
Food orange 7, the ethyl ester of beta-apo-8'-carotenic acid, is a carotenoid with an orange-red color. It is found in small quantities in some plants, but is often produced commercially from apocarotenal (E160e). It is used as a food coloring under the E number E160f and is approved for use in the EU and Australia and New Zealand where it is listed as food additive 160f; it is banned in the United States.
https://en.wikipedia.org/wiki/Glycoside%20hydrolase%20family%2079
In molecular biology, glycoside hydrolase family 79 is a family of glycoside hydrolases. Glycoside hydrolases are a widespread group of enzymes that hydrolyse the glycosidic bond between two or more carbohydrates, or between a carbohydrate and a non-carbohydrate moiety. A classification system for glycoside hydrolases, based on sequence similarity, has led to the definition of >100 different families. This classification is available on the CAZy web site, and also discussed at CAZypedia, an online encyclopedia of carbohydrate active enzymes. Glycoside hydrolase family 79 includes endo-beta-N-glucuronidase and heparanase (CAZY GH_79). Heparan sulphate proteoglycans (HSPGs) play a key role in the self- assembly, insolubility and barrier properties of basement membranes and extracellular matrices. Hence, cleavage of heparan sulphate (HS) affects the integrity and functional state of tissues and thereby fundamental normal and pathological phenomena involving cell migration and response to changes in the extracellular microenvironment. Heparanase degrades HS at specific intrachain sites. The enzyme is synthesized as a latent approximately 65 kDa protein that is processed at the N-terminus into a highly active approximately 50 kDa form. Experimental evidence suggests that heparanase may facilitate both tumour cell invasion and neovascularisation, both critical steps in cancer progression. The enzyme is also involved in cell migration associated with inflammation and autoimmunity.
https://en.wikipedia.org/wiki/Tobacco%20mosaic%20virus
Tobacco mosaic virus (TMV) is a positive-sense single-stranded RNA virus species in the genus Tobamovirus that infects a wide range of plants, especially tobacco and other members of the family Solanaceae. The infection causes characteristic patterns, such as "mosaic"-like mottling and discoloration on the leaves (hence the name). TMV was the first virus to be discovered. Although it was known from the late 19th century that a non-bacterial infectious disease was damaging tobacco crops, it was not until 1930 that the infectious agent was determined to be a virus. It is the first pathogen identified as a virus. The virus was crystallised by W.M. Stanley. It has a similar size to the largest synthetic molecule, known as PG5. History In 1886, Adolf Mayer first described the tobacco mosaic disease that could be transferred between plants, similar to bacterial infections. In 1892, Dmitri Ivanovsky gave the first concrete evidence for the existence of a non-bacterial infectious agent, showing that infected sap remained infectious even after filtering through the finest Chamberland filters. Later, in 1903, Ivanovsky published a paper describing abnormal crystal intracellular inclusions in the host cells of the affected tobacco plants and argued the connection between these inclusions and the infectious agent. However, Ivanovsky remained rather convinced, despite repeated failures to produce evidence, that the causal agent was an unculturable bacterium, too small to be retained on the employed Chamberland filters and to be detected in the light microscope. In 1898, Martinus Beijerinck independently replicated Ivanovsky's filtration experiments and then showed that the infectious agent was able to reproduce and multiply in the host cells of the tobacco plant. Beijerinck adopted the term of "virus" to indicate that the causal agent of tobacco mosaic disease was of non-bacterial nature. Tobacco mosaic virus was the first virus to be crystallized. It was achieved by Wendell
https://en.wikipedia.org/wiki/F-1%20Spirit%20%28series%29
is a series of Formula One-based racing video games developed and published by Konami starting on the MSX in 1987. F-1 Spirit: The Way to Formula-1 F-1 Spirit: The Way to Formula-1 is a top down Formula One Racing game, developed and published by Konami, which was released for the MSX in Japan and Europe in 1987. The game engine is very similar to Konami's Road Fighter. It also features Konami's custom sound chip called Konami SCC (a five-channel chip that complements the three-channel PSG chip of the MSX computer system, or in other words, a sound custom chip that brings five voices more to the three voices of the PSG sound chip on the system), and great MSX1 graphics to go with it. It was one of the first ROM on MSX with this sound feature. Together with its "3D" spinoff (F-1 Spirit: 3D Special), F-1 Spirit: The Way to Formula-1 was the most extended racing game Konami released for the MSX. This third-person racing game features many different types of cars. Everything starts with stock cars, moving up to rally cars and Formula 3. The main goal is to finish at first place with Formula 1, the king class of racing. There are six types of races: Stock race, Rally, F3 race, F3000 race, Endurance race, and finally F1 races (16 tracks). Initially, the player can only race in the stock, rally and F3 races. As the player win races, he will accumulate points that will allow him to play new races. If the player finish a race at first place he will receive nine points. He gets eight points if he finishes second, etc. If the player finish 10th or later, he will not score any points. There are 16 different tracks for F1 cars. As the player wins races, he will be able to play more tracks in the F1 car category. To complete the game, you have to win all of the 16 F1 tracks. There's a grand total of 21 tracks. The first races are the easiest ones: the cars are slow and the enemies do not drive very well. But as the player classify for new tracks the difficulty will increase: F
https://en.wikipedia.org/wiki/North%20Icelandic%20Jet
The North Icelandic Jet is a deep-reaching current that flows along the continental slope of Iceland. The North Icelandic Jet advects overflow water into the Denmark Strait and constitutes a pathway that is distinct from the East Greenland Current. It is a cold current that runs west across the top of Iceland, then southwest between Greenland and Iceland at a depth of about 600 metres (almost 2,000 feet). The North Icelandic Jet is deep and narrow (about 12 mile wide) and can carry more than a million cubic meters of water per second. It was not discovered until 2004.
https://en.wikipedia.org/wiki/LysM%20domain
In molecular biology the LysM domain is a protein domain found in a wide variety of extracellular proteins and receptors. The LysM domain is named after the Lysin Motif which was the original name given to the sequence motif identified in bacterial proteins. The region was originally identified as a C-terminal repeat found in the Enterococcus hirae muramidase. The LysM domain is found in a wide range of microbial extracellular proteins, where the LysM domain is thought to provide an anchoring to extracellular polysaccharides such as peptidoglycan and chitin. LysM domains are also found in plant receptors, including NFP, the receptor for Nod factor which is necessary for the root nodule symbiosis between legumes and symbiotic bacteria. The LysM domain is typically between 44 and 65 amino acid residues in length. The structure of the LysM domain showed that it is composed of a pair of antiparallel beta strands separated by a pair of short alpha helices. See also Nod factor
https://en.wikipedia.org/wiki/Promissum
Promissum is an extinct genus of conodonts, primitive chordates, that lived during the Upper Ordovician period. A conodont, Promissum had a primitive mouth under its eyes with mineralized teeth, which are both typical for conodonts. It had a primitive backbone and probably looked like a small eel or large worm, lacking any kind of fins except for perhaps a small one on the tail. It was relatively large for a conodont, reaching about 40 cm (16 inch) in length. Well-preserved specimens were discovered in the Soom shale of South Africa in 1994. Promissum was probably capable of maintaining a cruising speed, but not of bursts of speed.
https://en.wikipedia.org/wiki/TcoF-DB
The Dragon Database for Human Transcription Co-Factors and Transcription Factor Interacting Proteins (TcoF-DB) is a database that facilitates the exploration of proteins involved in the regulation of transcription in humans by binding to regulatory DNA regions (transcription factors) and proteins involved in the regulation of transcription in humans by interacting with transcription factors and not binding to regulatory DNA regions (transcription co-factors). The database describes a total of 529 (potential) human transcription co-factors interacting with a total of 1365 human transcription factors. See also Transcription factor Transcription coregulator
https://en.wikipedia.org/wiki/New%20Media%20Corporation
New Media Corporation, also known as New Media Technology Corporation, was an American computer company active from 1992 to the early 2000s. The company focused on the design and manufacture of PC Cards, a type of expansion card bus for laptops that had their heyday from the early 1990s to the mid-2000s. New Media was privately held and based out of Irvine, California. History New Media Corporation was founded by Carl Perkins, Rod Corder, and Eric McAfee, in Irvine, California, in March 1992. Perkins and Corder had previously worked together in the early 1980s at Rockwell Semiconductor Systems (which later changed their name to Conexant) of Newport Beach, California, where both led a team of chip designers within the company. In 1989, they founded Togai InfraLogic with another business partner. Togai InfraLogic was founded to capitalize on their ideas on designs for math coprocessor chips that each had developed in their spare time. The company later sold a Japanese speech recognition ASIC to Canon. In 1990, Perkins and Corder left Togai InfraLogic over differences of expectations for the company's future with the third business partner. Perkins and Corder both ended up at ITT Inc., designing ASICs for use in expansion card products for personal computers. It was in this capacity that the two got the inspiration to found New Media Corporation. The duo had been observing the fledgling PC Card expansion bus standard for laptops in the early 1990s and believed they could design PC Card products that were both plug-and-play and intuitive to configure. In 1991, they discovered Eric McAfee, an investor of Silicon Beach technology companies, through a mutual friend and decided to hire him as chief financial officer (CFO). New Media was the first company to manufacture PC Cards in the United States, according to the Orange County Business Journal. Within a year of its existence it gained large laptop manufacturers such as Toshiba and Compaq as clients, manufacturing cards
https://en.wikipedia.org/wiki/Genetic%20monitoring
Genetic monitoring is the use of molecular markers to (i) identify individuals, species or populations, or (ii) to quantify changes in population genetic metrics (such as effective population size, genetic diversity and population size) over time. Genetic monitoring can thus be used to detect changes in species abundance and/or diversity, and has become an important tool in both conservation and livestock management. The types of molecular markers used to monitor populations are most commonly mitochondrial, microsatellites or single-nucleotide polymorphisms (SNPs), while earlier studies also used allozyme data. Species gene diversity is also recognized as an important biodiversity metric for implementation of the Convention on Biological Diversity. Types Types of population changes that can be detected by genetic monitoring include population growth and decline, spread of pathogens, adaptation to environmental change, hybridization, introgression and habitat fragmentation events. Most of these changes are monitored using ‘neutral’ genetic markers (markers for which mutational changes do not change their adaptive fitness within a population). However markers showing adaptive responses to environmental change can be ‘non-neutral’ (e.g. mutational changes affect their relative fitness within a population). Two broad categories of genetic monitoring have been defined: Category I encompasses the use of genetic markers as identifiers of individuals (Category Ia), populations and species (Category Ib) for traditional population monitoring. Category II represents the use of genetic markers to monitor changes of population genetic parameters, which include estimators of effective population size (Ne), genetic variation, population inter-mixing, structure and migration. Examples Estimating abundance and life history parameters – Category Ia At the individual level, genetic identification can enable estimation of population abundance and population increase rates within t
https://en.wikipedia.org/wiki/Clam%20liquor
Description Clam liquor, also called clam extract, is a liquid extracted during cooking and opening of clams. Undiluted it is called clam broth, and when concentrated by evaporation is called clam nectar. It may be canned in all these forms or used to fill up canned clam meat. It is an important component of many seafood recipes and is also used as a natural seasoning. Clam liquor can be used to enhance the flavor of soups, stews, sauces, and other dishes. It is also a popular ingredient in Asian cuisine. Popular Use Clam liquor is one of the main ingredients in the traditional New England Clam Chowder. This drink contains 20 ingredients: cream, flour, garlic, leaves, milk, and oil. A cup of clam liquor should be reserved for each quart of clams, which should be cleaned and separated. Clam liquor, potatoes, water, parsley, and thyme must be combined with 1 cup of milk to make the clams. Melt butter in heavy chicken broth and a cup of clam liquid and cook until it has evaporated. Finally, add parsley, thyme, and salt and pepper. The potatoes should be ready to eat within 15 minutes. See also List of clam dishes
https://en.wikipedia.org/wiki/Lentisphaera%20araneosa
Lentisphera araneosa is a marine bacteria strain in the bacterial phylum Lentisphaerota. They are able to produce viscous transparent exopolymers and grow attached to each other by the polymer in a three-dimensional configuration. They are part of the natural surface bacterial population in the Atlantic and Pacific oceans. They are less than 1% of the total bacterial community. This species is gram negative, non-motile, non-pigmented, aerobic, chemoheterotrophic, and facultatively oligotrophic sphere-shaped. Its genome has been sequenced.
https://en.wikipedia.org/wiki/Minimum%20Fisher%20information
In information theory, the principle of minimum Fisher information (MFI) is a variational principle which, when applied with the proper constraints needed to reproduce empirically known expectation values, determines the best probability distribution that characterizes the system. (See also Fisher information.) Measures of information Information measures (IM) are the most important tools of information theory. They measure either the amount of positive information or of "missing" information an observer possesses with regards to any system of interest. The most famous IM is the so-called Shannon-entropy (1948), which determines how much additional information the observer still requires in order to have all the available knowledge regarding a given system S, when all he/she has is a probability density function (PDF) defined on appropriate elements of such system. This is then a "missing" information measure. The IM is a function of the PDF only. If the observer does not have such a PDF, but only a finite set of empirically determined mean values of the system, then a fundamental scientific principle called the Maximum Entropy one (MaxEnt) asserts that the "best" PDF is the one that, reproducing the known expectation values, maximizes otherwise Shannon's IM. Fisher's information measure Fisher's information (FIM), named after Ronald Fisher, (1925) is another kind of measure, in two respects, namely, 1) it reflects the amount of (positive) information of the observer, 2) it depends not only on the PD but also on its first derivatives, a property that makes it a local quantity (Shannon's is instead a global one). The corresponding counterpart of MaxEnt is now the FIM-minimization, since Fisher's measure grows when Shannon's diminishes, and vice versa. The minimization here referred to (MFI) is an important theoretical tool in a manifold of disciplines, beginning with physics. In a sense it is clearly superior to MaxEnt because the later procedure yields alwa
https://en.wikipedia.org/wiki/C7%20protein
C7 protein is an engineered zinc finger protein based on the murine ZFP, Zif268 and discovered by Wu et al. in 1994 (published in 1995). It shares the same zinc finger 2 and zinc finger 3 of Zif268, but differs in the sequence of finger 1. It also shares the same DNA target, 5'-GCGTGGGCG-3'. The shared sequences in single letter amino acid codes of fingers 2 and 3 are RSD-H-LTT and RAD-E-RKR (positions -1 through 6 in the alpha helix). Zinc finger 1 has the sequence KSA-D-LKR which provides a 13-fold increase in affinity to the target sequence of the entire ZFP over that of Zif268. It is used in zinc finger investigations in which the amino acid sequence of finger 2 is changed in order to determine the appropriate sequence to target a given three-nucleotide target site. A variation of C7, C7.GAT is preferred since it lacks the aspartic acid residue present in finger 3 of C7 and known to cause a phenomenon called 'target site overlap'. In this case the target site overlap is a result of the aspartic acid residue forming a hydrogen bond with the N4 of the cytosine (in the opposite strand) base-paired to the guanine in the finger 2 subsite. It can also form the same hydrogen bond with an adenine base paired to a thymine. This target site overlap would dictate that either a cytosine or adenine residue be present as the 3' nucleotide in the finger 2 subsite which is unacceptable when looking to target sequences containing another nucleotide at this position.
https://en.wikipedia.org/wiki/NeuroIntegration%20Therapy
NeuroIntegration Therapy (NIT) is a non-invasive combination therapy that integrates quantitative electroencephalography (qEEG or QEEG) brain mapping with additional therapies such as neurofeedback, vibroacoustic therapy, pulsed electromagnetic field therapy (PEMFT, or PEMF therapy) and photic stimulation (light therapy.) Neurointegration therapy begins with a brain mapping session using qEEG to help visualize areas of dysregulation within the brain. The supporting therapies are then used to retrain the problem areas of the brain by rewarding the brainwaves when they move in a desired pattern. Follow up qEEG sessions demonstrate changes in brainwave patterns and signify if the therapeutic treatments require adjustments. Application NeuroIntegration therapy is being used as a treatment for brain-related conditions such as chronic pain, addiction, and obesity. The field of neurointegration therapy is relatively new and protocols have not been widely standardized. However, global initiatives are being taken to create a common language of information, to promote consistency in education and certification, and to form collaborative alliances with other mental and physical health disciplines. Criticism Though neurointegration therapy itself has not received any published criticism, the individual components of the treatment have.
https://en.wikipedia.org/wiki/Information%20privacy%20law
Information privacy, data privacy or data protection laws provide a legal framework on how to obtain, use and store data of natural persons. The various laws around the world describe the rights of natural persons to control who is using its data. This includes usually the right to get details on which data is stored, for what purpose and to request the deletion in case the purpose is not given anymore. Over 80 countries and independent territories, including nearly every country in Europe and many in Latin America and the Caribbean, Asia, and Africa, have now adopted comprehensive data protection laws. The European Union has the General Data Protection Regulation (GDPR), in force since May 25, 2018. The United States is notable for not having adopted a comprehensive information privacy law, but rather having adopted limited sectoral laws in some areas like the California Consumer Privacy Act (CCPA). By Jurisdiction The German state of Hessia enacted the World's first data privacy law on 30SEP1970. In Germany the term informational self-determination was first used in the context of a German constitutional ruling relating to personal information collected during the 1983 census. Asia India India passed its Digital Personal Data Protection Act, 2023 in August 2023. China China passed its Personal Information Protection Law (PIPL) in mid-2021, and was effective from November 1, 2021. It focuses heavily on consent, rights of the individual, and transparency of data processing. PIPL has been compared to the EU GDPR as it has similar scope and many similar provisions. Philippines In the Philippines, The Data Privacy Act of 2012 mandated the creation of the National Privacy Commission that would monitor and maintain policies that involve information privacy and personal data protection in the country. Modeled after the EU Data Protection Directive and the Asia-Pacific Economic Cooperation (APEC) Privacy Framework, the independent body would ensure compliance of t
https://en.wikipedia.org/wiki/Organ%20futures
Organ futures is the short term used in academic proposals for futures contracts on organs from human cadavers. They are not legal anywhere at this time. Organ futures would be used as an economic means to encourage organ donation by compensating transplant organ donors. Financial futures contracts are essentially agreements to pay a specified sum at a specified time. The four key academic papers describing proposals for organ futures were published between the mid-1980s and mid-1990s. Proposals The explanations below focus on donation of cadaveric organs and ignore living donation. Proposal by Schwindt, Vining 1986 Schwindt & Vining (1986) suggest that the organ donor is paid at the time they agree to enter the life-time futures contract. The agreement is mutually revocable. They propose a single government broker as the buyer. Organ recipients would pay the supply price plus a load factor to the broker. Proposal by Hansmann 1989 Hansmann (1989) also suggests payment at the time of contract. Instead of direct payment, he proposes reductions to health insurance premiums as indirect incentive. The hospital where the organ donor dies is expected to verify a seller registry and determine the buyer. Buyers may be health insurance providers or specialist traders. Proposal by Cohen 1989 Cohen (1989) introduces a significant change to previous proposals by making payment conditional on organ extraction. Thus, the donor is not directly compensated during their lifetime. However, the payment is allocated to their estate or a designee. Hospitals are expected to notify buyers and preserve cadavers. They can be made liable for consequences of negligence. Buyers may be public or private organizations. Proposal by Crespi 1994 Crespi (1994) aims to integrate what he deems the most useful aspects of previous models into his own. Payment would be either guaranteed upon death or dependent on organ extraction. The money would go to the seller's estate; rights would not be
https://en.wikipedia.org/wiki/CLCN5
The CLCN5 gene encodes the chloride channel Cl-/H+ exchanger ClC-5. ClC-5 is mainly expressed in the kidney, in particular in proximal tubules where it participates to the uptake of albumin and low-molecular-weight proteins, which is one of the principal physiological role of proximal tubular cells. Mutations in the CLCN5 gene cause an X-linked recessive nephropathy named Dent disease (Dent disease 1 MIM#300009) characterized by excessive urinary loss of low-molecular-weight proteins and of calcium (hypercalciuria), nephrocalcinosis (presence of calcium phosphate aggregates in the tubular lumen and/or interstitium) and nephrolithiasis (kidney stones). The CLCN5 gene Structure The human CLCN5 gene (MIM#300008, reference sequence NG_007159.2) is localized in the pericentromeric region on chromosome Xp11.23. It extends over about 170 Kb of genomic DNA, has a coding region of 2,238 bp and consists of 17 exons including 11 coding exons (from 2 to 12). The CLCN5 gene has 8 paralogues (CLCN1, CLCN2, CLCN3, CLCN4, CLCN6, CLCN7, CLCNKA, CLCNKB) and 201 orthologues among jawed vertebrates (Gnathostomata). Five different CLCN5 gene transcripts have been discovered, two of which (transcript variants 3 [NM_000084.5] and 4 [NM_001282163.1]) encode for the canonical 746 amino acid protein, two (transcript variants 1 [NM_001127899.3] and 2 [NM_001127898.3]) for the NH2-terminal extended 816 amino acid protein and one does not encode for any protein (Transcript variant 5, [NM_001272102.2]). The 5’ untranslated region (5’UTR) of CLCN5 is complex and not entirely clarified. Two strong and one weak promoters were predicted to be present in the CLCN5 gene. Several different 5’ alternatively used exons have been recognized in the human kidney. The three promoters drive with varying degree of efficiency 11 different mRNAs, with transcription initiating from at least three different start sites. The chloride channel H+/Cl− exchanger ClC-5 Like all ClC channels, ClC-5 needs to dimerize
https://en.wikipedia.org/wiki/Journal%20of%20Clinical%20Pathology
The Journal of Clinical Pathology (JCP) is a peer-reviewed medical journal covering all aspects of pathology, published by the BMJ Group and co-owned by the Association of Clinical Pathologists. Diagnostic and research areas covered include histopathology, virology, haematology, microbiology, cytopathology, chemical pathology, molecular pathology, forensic pathology, dermatopathology, neuropathology, and immunopathology. Each issue contains reviews, original articles, short reports, case reports, correspondence, and book reviews. In 2005 the Journal of Clinical Pathology incorporated Molecular Pathology, which was published from 1995 to 2004. Indexing and impact The journal is abstracted and indexed in Science Citation Index, Index Medicus (Medline), Scopus, Google Scholar and EMBASE/Excerpta Medica. According to the Journal Citation Reports, its 2022 impact factor is 3.4 ranking it 28th out of 76 journals in the category "Pathology" The journal has been cited most often by the following journals Journal of Clinical Pathology, Histopathology, Archives of Pathology & Laboratory Medicine, Human Pathology, and the World Journal of Gastroenterology. The journals that have been cited most often by the Journal of Clinical Pathology are Journal of Clinical Pathology, American Journal of Surgical Pathology, Cancer, Cancer Research, and Modern Pathology.
https://en.wikipedia.org/wiki/DIVA-GIS
DIVA-GIS is a free geographic information system software program used for the analysis of geographic data, in particular point data on biodiversity. The software was first designed for application to the study of wild potatoes in South America. Development DIVA-GIS was developed as a joint project by the International Potato Center in Peru, the International Plant Genetic Resources Institute, the Museum of Vertebrate Zoology at the University of California at Berkeley, the Secretariat of the Pacific Community, and the FAO. DIVA-GIS has a wide range of tools for evaluation of mapping and modeling of habitats. There is a command-line version of the program that has been developed, AVID-GIS. Formats DIVA-GIS can process data in all standard GIS formats, including data from ESRI's ArcGIS programs. The program runs on Windows and OS X. DIVA raster files generated may be imported and exported into R or the modeling program Maxent.