source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Kaup%E2%80%93Kupershmidt%20equation | The Kaup–Kupershmidt equation (named after David J. Kaup and Boris Abram Kupershmidt) is the nonlinear fifth-order partial differential equation
It is the first equation in a hierarchy of integrable equations with the Lax operator
.
It has properties similar (but not identical) to those of the better-known KdV hierarchy in which the Lax operator has order 2. |
https://en.wikipedia.org/wiki/Vibriocin | Vibriocins are a group of bacteriocins produced by, and active against, gram-negative bacteria in the genus Vibrio. They were first discovered in 1962, considerably after the original bacteriocins, the colicins, which were discovered in 1925.
Like other bacteriocins, vibriocins are protein toxins. They can kill bacteria beyond the genus Vibrio, including other proteobacteria. They have been used for abortive classification schemes of the vibrio, particularly to type various kinds of cholera, against which they were thought to have potential as antibiotics. Their mode of action,
genetics and regulation have all been studied, for at least one example. In all likelihood, however, they are as common and as diverse as the colicins, making it very unlikely that these initial experiments have fully explored the range of mechanisms and forms that the vibriocins take.
In the 1970s, they were investigated, along with some colicins, as potential chemotherapeutic agents. The mode of action appears to be nuclease activity resulting in the induction of apoptosis. The research itself was the result of observing unexpected interactions between the vibriocins and eukaryotic cells. |
https://en.wikipedia.org/wiki/NCP%20Engineering | NCP engineering is a Nuremberg-based company producing software for remote access, industrial internet of things security and information security. NCP's products use virtual private network (VPN) and other technologies like encryption, personal firewalls and electronic certificates in a public key infrastructure (PKI) to secure data communication.
NCP has made its IPsec VPN client compatible with the Windows 10, Windows 11, iOS, macOS, Linux and Android operating systems.
Name
NCP is the abbreviation of "Network Communications Products". The supplement "engineering" describes that the company produces software for secure communications and remote access.
History
NCP engineering was founded in Nuremberg, Germany in 1986. The company produces software for secure data communication through the Internet, networks via 4G/5G and wireless LANs. At the core of NCP's business is provisioning secure communication connections between stationary and mobile end-devices as well as affiliate and branch networks to a company's headquarters.
In 2007, NCP partnered with WatchGuard Technologies.
In January 2010, NCP established a North American affiliate, NCP engineering, Inc.
In February 2010, NCP engineering was awarded US Patent 8811397 B2 for a "System and method for data communication between a user terminal and a gateway via a network node".
NCP engineering has been involved in the ESUKOM project for the development of a real-time security solution that protects corporate networks using integrated security solutions based on a unified metadata format since 2010.
As a SIMU project partner (security information and event management for small SMEs), NCP engineering is focused on optimizing IT security in corporate networks.
After many years of working together, in 2017 NCP engineering and Juniper Networks have intensified their collaboration in a technology partnership.
Juniper Networks and NCP engineering began a partnership in 2011 and intensified their collaborati |
https://en.wikipedia.org/wiki/Comparative%20Toxicogenomics%20Database | The Comparative Toxicogenomics Database (CTD) is a public website and research tool launched in November 2004 that curates scientific data describing relationships between chemicals/drugs, genes/proteins, diseases, taxa, phenotypes, GO annotations, pathways, and interaction modules.
The database is maintained by the Department of Biological Sciences at North Carolina State University.
Background
The Comparative Toxicogenomics Database (CTD) is a public website and research tool that curates scientific data describing relationships between chemicals, genes/proteins, diseases, taxa, phenotypes, GO annotations, pathways, and interaction modules, launched on November 12, 2004.
The database is maintained by the Department of Biological Sciences at North Carolina State University.
Goals and objectives
One of the primary goals of CTD is to advance the understanding of the effects of environmental chemicals on human health on the genetic level, a field called toxicogenomics.
The etiology of many chronic diseases involves interactions between environmental factors and genes that modulate important physiological processes. Chemicals are an important component of the environment. Conditions such as asthma, cancer, diabetes, hypertension, immunodeficiency, and Parkinson's disease are known to be influenced by the environment; however, the molecular mechanisms underlying these correlations are not well understood. CTD may help resolve these mechanisms. The most up-to-date extensive list of peer-reviewed scientific articles about CTD is available at their publications page
Core data
CTD is a unique resource where biocurators read the scientific literature and manually curate four types of core data:
Chemical-gene interactions
Chemical-disease associations
Gene-disease associations
Chemical-phenotype associations
Data integration
By integrating the above four data sets, CTD automatically constructs putative chemical-gene-phenotype-disease networks to illuminate molecula |
https://en.wikipedia.org/wiki/Critical%20heat%20flux | In the study of heat transfer, critical heat flux (CHF) is the heat flux at which boiling ceases to be an effective form of transferring heat from a solid surface to a liquid.
Description
Boiling systems are those in which liquid coolant absorbs energy from a heated solid surface and undergoes a change in phase. In flow boiling systems, the saturated fluid progresses through a series of flow regimes as vapor quality is increased. In systems that utilize boiling, the heat transfer rate is significantly higher than if the fluid were a single phase (i.e. all liquid or all vapor). The more efficient heat transfer from the heated surface is due to heat of vaporization and sensible heat. Therefore, boiling heat transfer has played an important role in industrial heat transfer processes such as macroscopic heat transfer exchangers in nuclear and fossil power plants, and in microscopic heat transfer devices such as heat pipes and microchannels for cooling electronic chips.
The use of boiling as a means of heat removal is limited by a condition called critical heat flux (CHF). The most serious problem that can occur around CHF is that the temperature of the heated surface may increase dramatically due to significant reduction in heat transfer. In industrial applications such as electronics cooling or instrumentation in space, the sudden increase in temperature may possibly compromise the integrity of the device.
Two-phase heat transfer
The convective heat transfer between a uniformly heated wall and the working fluid is described by Newton's law of cooling:
where represents the heat flux, represents the proportionally constant called the heat transfer coefficient, represents the wall temperature and represents the fluid temperature. If decreases significantly due to the occurrence of the CHF condition, will increase for fixed and while will decrease for fixed .
Modes of CHF
The understanding of CHF phenomenon and an accurate prediction of the CHF condition |
https://en.wikipedia.org/wiki/Biasing | In electronics, biasing is the setting of DC (direct current) operating conditions (current and voltage) of an electronic component that processes time-varying signals. Many electronic devices, such as diodes, transistors and vacuum tubes, whose function is processing time-varying (AC) signals, also require a steady (DC) current or voltage at their terminals to operate correctly. This current or voltage is called bias. The AC signal applied to them is superposed on this DC bias current or voltage.
The operating point of a device, also known as bias point, quiescent point, or Q-point, is the DC voltage or current at a specified terminal of an active device (a transistor or vacuum tube) with no input signal applied. A bias circuit is a portion of the device's circuit that supplies this steady current or voltage.
Overview
In electronics, 'biasing' usually refers to a fixed DC voltage or current applied to a terminal of an electronic component such as a diode, transistor or vacuum tube in a circuit in which AC signals are also present, in order to establish proper operating conditions for the component. For example, a bias voltage is applied to a transistor in an electronic amplifier to allow the transistor to operate in a particular region of its transconductance curve. For vacuum tubes, a grid bias voltage is often applied to the grid electrodes for the same reason.
In magnetic tape recording, the term bias is also used for a high-frequency signal added to the audio signal and applied to the recording head, to improve the quality of the recording on the tape. This is called tape bias.
Importance in linear circuits
Linear circuits involving transistors typically require specific DC voltages and currents for correct operation, which can be achieved using a biasing circuit. As an example of the need for careful biasing, consider a transistor amplifier. In linear amplifiers, a small input signal gives a larger output signal without any change in shape (low distortion |
https://en.wikipedia.org/wiki/Principal%20geodesic%20analysis | In geometric data analysis and statistical shape analysis, principal geodesic analysis is a generalization of principal component analysis to a non-Euclidean, non-linear setting of manifolds suitable for use with shape descriptors such as medial representations. |
https://en.wikipedia.org/wiki/Vehicle%20rescheduling%20problem | The vehicle rescheduling problem (VRSP) is a combinatorial optimization and integer programming problem seeking to service customers on a trip after change of schedule such as vehicle break down or major delay. Proposed by Li, Mirchandani and Borenstein in 2007, the VRSP is an important problem in the fields of transportation and logistics.
Determining the optimal solution is an NP-complete problem in combinatorial optimization, so in practice heuristic and deterministic methods are used to find acceptably good solutions for the VRSP.
Overview
Several variations and specializations of the vehicle rescheduling problem exist:
Single Depot Vehicle Rescheduling Problem (SDVRSP): A number of trips need to be rescheduled due to delay, vehicle break down or for any other reason. The goal is to find optimal rescheduling of the existing fleet, using possibly extra vehicles from the depot, in order to minimise the delay and the operating costs. In the Single Depot variation, there is only one depot which contains all extra vehicles, and in which every vehicle starts and ends its schedule.
Multi Depot Vehicle Rescheduling Problem (MDVRSP): Similar to SDVRSP, except additional depots are introduced. Each depot has capacity constraints, as well as variable extra vehicles. Usually vehicle schedules have an additional constraint which requires that each vehicle returns to the depot where it started its schedule.
Open Vehicle Rescheduling Problem (OVRSP): Vehicles are not required to return to the depot.
Although VRSP is related to the Single Depot Vehicle Scheduling Problem and the Multi Depot Vehicle Scheduling Problem, there is a significant difference in runtime requirements, as VRSP need to be solved in near real-time to allow rescheduling during operations, while SDVSP and MDVSP are typically solved using long running linear programming methods.
Another field where VRSP is used is in transportation of goods in order to reschedule the routes when demand substantial |
https://en.wikipedia.org/wiki/Supersymmetric%20theory%20of%20stochastic%20dynamics | Supersymmetric theory of stochastic dynamics or stochastics (STS) is an exact theory of stochastic (partial) differential equations (SDEs), the class of mathematical models with the widest applicability covering, in particular, all continuous time dynamical systems, with and without noise. The main utility of the theory from the physical point of view is a rigorous theoretical explanation of the ubiquitous spontaneous long-range dynamical behavior that manifests itself across disciplines via such phenomena as 1/f, flicker, and crackling noises and the power-law statistics, or Zipf's law, of instantonic processes like earthquakes and neuroavalanches. From the mathematical point of view, STS is interesting because it bridges the two major parts of mathematical physics – the dynamical systems theory and topological field theories. Besides these and related disciplines such as algebraic topology and supersymmetric field theories, STS is also connected with the traditional theory of stochastic differential equations and the theory of pseudo-Hermitian operators.
The theory began with the application of BRST gauge fixing procedure to Langevin SDEs, that was later adapted to classical mechanics and its stochastic generalization, higher-order Langevin SDEs, and, more recently, to SDEs of arbitrary form, which allowed to link BRST formalism to the concept of transfer operators and recognize spontaneous breakdown of BRST supersymmetry as a stochastic generalization of dynamical chaos.
The main idea of the theory is to study, instead of trajectories, the SDE-defined temporal evolution of differential forms. This evolution has an intrinsic BRST or topological supersymmetry representing the preservation of topology and/or the concept of proximity in the phase space by continuous time dynamics. The theory identifies a model as chaotic, in the generalized, stochastic sense, if its ground state is not supersymmetric, i.e., if the supersymmetry is broken spontaneously. Accordingly, |
https://en.wikipedia.org/wiki/Intracellular%20delivery | Intracellular delivery is the process of introducing external materials into living cells. Materials that are delivered into cells include nucleic acids (DNA and RNA), proteins, peptides, impermeable small molecules, synthetic nanomaterials, organelles, and micron-scale tracers, devices and objects.
Such molecules and materials can be used to investigate cellular behavior, engineer cell operations or correct a pathological function.
Medical applications of intracellular delivery range from in vitro fertilisation (IVF)
and mRNA vaccines
to gene therapy
and preparation of CAR-T cells
.
Industrial applications include protein production
, biomanufacture
, and genetic engineering of plants and animals
.
Intracellular delivery is a fundamental technique in the study of biology and genetics, such as the use of DNA plasmid transfection to investigate protein function in living cells
.
A wide range of approaches exist for performing intracellular delivery including biological, chemical and physical techniques that work through either membrane disruption or packaging the delivery material in carriers
.
Intracellular delivery is at the intersection of cell biology and technology, and is related to many fields across science and medicine including genetics, biotechnology, bioengineering and drug delivery.
Applications
Analogous to the way computers operate through electronic signals, cells process and transmit information through molecules. Depending on the molecules and materials that are loaded into cells, different outcomes or applications can be achieved (see Figure "Applications of Intracellular Delivery" for examples). Below are some of the main classifications of cargo materials used to investigate and engineer cells through intracellular delivery.
Cargo Types
Nucleic Acids
Transfection refers to the intracellular delivery of nucleic acids: DNA, RNA and their analogues. Nucleic acids materials that are commonly transfected into cells are plasmid DNA, mRNA, si |
https://en.wikipedia.org/wiki/Argo%20%28oceanography%29 | Argo is an international programme for researching the ocean. It uses profiling floats to observe temperature, salinity and currents. Recently it has observed bio-optical properties in the Earth's oceans. It has been operating since the early 2000s. The real-time data it provides support climate and oceanographic research. A special research interest is to quantify the ocean heat content (OHC). The Argo fleet consists of almost 4000 drifting "Argo floats" (as profiling floats used by the Argo program are often called) deployed worldwide. Each float weighs 20–30 kg. In most cases probes drift at a depth of 1000 metres. Experts call this the parking depth. Every 10 days, by changing their buoyancy, they dive to a depth of 2000 metres and then move to the sea-surface. As they move they measure conductivity and temperature profiles as well as pressure. Scientists calculate salinity and density from these measurements. Seawater density is important in determining large-scale motions in the ocean.
Average current velocities at 1000 metres are directly measured by the distance and direction a float drifts while parked at that depth, which is determined by GPS or Argos system positions at the surface. The data is transmitted to shore via satellite, and is freely available to everyone, without restrictions.
The Argo program is named after the Greek mythical ship Argo to emphasize the complementary relationship of Argo with the Jason satellite altimeters. Both the standard Argo floats and the 4 satellites launched so far to monitor changing sea-level all operate on a 10-day duty cycle.
International collaboration
The Argo program is a collaborative partnership of more than 30 nations from all continents (most shown on the graphic map in this article) that maintains a global array and provides a dataset anyone can use to explore the ocean environment. Argo is a component of the Global Ocean Observing System (GOOS), and is coordinated by the Argo Steering Team, an internat |
https://en.wikipedia.org/wiki/Brachychiton%20sp.%20Ormeau | Brachychiton sp. Ormeau is a rare and endangered rainforest tree found in Queensland, Australia.
Description
A species of tree belonging to the genus Brachychiton, it reaches up to 25 metres in height. The leaves are dropped during the dry season, a time of year the species favours for reproduction, and return as pale to coppery coloured new growth. The flowering period is during September, the profuse display of green to white bell-shaped flowers appearing at the terminus of the branches;the width of each flower is around 10 mm. Fruiting pods appear from January to February, these are 3 cm long, brown, and boat-shaped. During the later stages of growth the trunk begins to form an exaggerated bottle-shape, and the leaves alter from a deeply lobed shape, divided from five to nine times, to a glossy and often elliptical leaf 12 to 20 centimetres long.
The tree is capable of attaining a great age, over 120 years being possible. Sexual maturity is reached after around twenty years.
Distribution and range
The Ormeau bottle trees are noticeably restricted in range, extending over a range of only 6.5 km2 and occurring in very low population densities per square kilometre. The largest stand, regarded as the most viable population, is reported to consist of 131 plants. Another two reproductive populations have been found in separate locations nearby but each contains fewer than ten trees; other individuals occur as non-seeding outliers within this total population of 161 trees.
Conservation
The main population occurs within an 'environmental park', the Wongawallan Conservation Area in the rural suburb of Wongawallan, Queensland, where it is afforded some protection from threatening factors. The small groups outside this area are located on a lease for proposed quarries. The federal government has named this tree as one of thirty plant species to be the given the highest priority for protection from extinction, and that its status be improved by the year 2020. The majo |
https://en.wikipedia.org/wiki/Frobenius%20determinant%20theorem | In mathematics, the Frobenius determinant theorem was a conjecture made in 1896 by the mathematician Richard Dedekind, who wrote a letter to F. G. Frobenius about it (reproduced in , with an English translation in ).
If one takes the multiplication table of a finite group G and replaces each entry g with the variable xg, and subsequently takes the determinant, then the determinant factors as a product of n irreducible polynomials, where n is the number of conjugacy classes. Moreover, each polynomial is raised to a power equal to its degree. Frobenius proved this surprising conjecture, and it became known as the Frobenius determinant theorem.
Formal statement
Let a finite group have elements , and let be associated with each element of . Define the matrix with entries . Then
where the 's are pairwise non-proportional irreducible polynomials and is the number of conjugacy classes of G. |
https://en.wikipedia.org/wiki/Journal%20of%20the%20International%20Association%20of%20Providers%20of%20AIDS%20Care | The Journal of the International Association of Providers of AIDS Care is a bimonthly peer-reviewed medical journal that covers research on AIDS. The editor-in-chief is Jose M. Zuniga (International Association of Providers of AIDS Care). The journal was established in 2002 and is published by SAGE Publications.
Abstracting and indexing
The journal is abstracted and indexed in:
CAB Abstracts
CINAHL
EBSCO databases
Global Health
MEDLINE
Rural Development Abstracts
Tropical Diseases Bulletin
External links
SAGE Publishing academic journals
English-language journals
HIV/AIDS journals
Academic journals established in 2002
Bimonthly journals
Academic journals associated with learned and professional societies |
https://en.wikipedia.org/wiki/Closure%20temperature | In radiometric dating, closure temperature or blocking temperature refers to the temperature of a system, such as a mineral, at the time given by its radiometric date. In physical terms, the closure temperature is the temperature at which a system has cooled so that there is no longer any significant diffusion of the parent or daughter isotopes out of the system and into the external environment. The concept's initial mathematical formulation was presented in a seminal paper by Martin H. Dodson,
"Closure temperature in cooling geochronological and petrological systems" in the journal Contributions to Mineralogy and Petrology, 1973, with refinements to a usable experimental formulation by other scientists in later years. This temperature varies broadly among different minerals and also differs depending on the parent and daughter atoms being considered. It is specific to a particular material and isotopic system.
The closure temperature of a system can be experimentally determined in the lab by artificially resetting sample minerals using a high-temperature furnace. As the mineral cools, the crystal structure begins to form and diffusion of isotopes slows. At a certain temperature, the crystal structure has formed sufficiently to prevent diffusion of isotopes. This temperature is what is known as blocking temperature and represents the temperature below which the mineral is a closed system to measurable diffusion of isotopes. The age that can be calculated by radiometric dating is thus the time at which the rock or mineral cooled to blocking temperature.
These temperatures can also be determined in the field by comparing them to the dates of other minerals with well-known closure temperatures.
Closure temperatures are used in geochronology and thermochronology to date events and determine rates of processes in the geologic past.
Table of values
The following table represents the closure temperatures of some materials. These values are the approximate value |
https://en.wikipedia.org/wiki/Ring-enhancing%20lesion | A ring-enhancing lesion is an abnormal radiologic sign on MRI or CT scans obtained using radiocontrast. On the image, there is an area of decreased density (see radiodensity) surrounded by a bright rim from concentration of the enhancing contrast dye. This enhancement may represent breakdown of the blood-brain barrier and the development of an inflammatory capsule. This can be a finding in numerous disease states. In the brain, it can occur with an early brain abscess as well as in Nocardia infections associated with lung cavitary lesions. In patients with HIV, the major differential is between CNS lymphoma and CNS toxoplasmosis, with CT imaging being the appropriate next step to differentiate between the two conditions. |
https://en.wikipedia.org/wiki/Corymb | Corymb is a botanical term for an inflorescence with the flowers growing in such a fashion that the outermost are borne on longer pedicels than the inner, bringing all flowers up to a common level. A corymb has a flattish top with a superficial resemblance towards an umbel, and may have a branching structure similar to a panicle. Flowers in a corymb structure can either be parallel, or alternate, and form in either a convex, or flat form.
Many species in the Maloideae, such as hawthorns and rowans, produce their flowers in corymbs. The Norway maple and yerba maté are also examples of corymbs.
The word corymb is derived from the Ancient Greek word meaning "bunch of flowers or fruit". |
https://en.wikipedia.org/wiki/Mitotoxin | A mitotoxin is a cytotoxic molecule targeted to specific cells by a mitogen. Generally found in snake venom. Mitotoxins are responsible for mediating cell death by interfering with protein or DNA synthesis. Some mechanisms by which mitotoxins can interfere with DNA or protein synthesis include the inactivation of ribosomes or the inhibition of complexes in the mitochondrial electron transport chain. These toxins have a very high affinity and level of specificity for the receptors that they bind to. Mitotoxins bind to receptors on cell surfaces and are then internalized into cells via receptor-mediated endocytosis. Once in the endosome, the receptor releases its ligand and a mitotoxin can mediate cell death.
There are different classes of mitotoxins, each acting on a different type of cell or system. The mitotoxin classes that have been identified thus far include: interleukin-based, transferrin based, epidermal growth factor-based, nerve growth factor-based, insulin-like growth factor-I-based, and fibroblast growth factor-based mitotoxins. Because of the high affinity and specificity of mitotoxin binding, they present the possibility of creating precise therapeutic agents. A major one of these possibilities is the potential usage of growth factor-based mitotoxins as anti-neoplastic agents that can modulate the growth of melanomas. |
https://en.wikipedia.org/wiki/Axillary%20joints | The axillary joints are two joints in the axillary region of the body, and include the shoulder joint and the acromioclavicular joint.
Shoulder joint
The shoulder joint also known as the glenohumeral joint is a synovial ball and socket joint. The shoulder joint involves articulation between the glenoid cavity of the scapula (shoulder blade) and the head of the upper arm bone (humerus) and functions as a diarthrosis and multiaxial joint.
Due to the very loose joint capsule that gives a limited interface of the humerus and scapula, it is the most mobile joint of the human body.
Acromioclavicular joint
The acromioclavicular joint, is the joint at the top of the shoulder. It is the junction between the acromion (part of the scapula that forms the highest point of the shoulder) and the clavicle. It is a plane synovial joint.
The acromioclavicular joint allows the arm to be raised above the head. This joint functions as a pivot point (although technically it is a gliding synovial joint), acting like a strut to help with movement of the scapula resulting in a greater degree of arm rotation. |
https://en.wikipedia.org/wiki/Specification%20language | A specification language is a formal language in computer science used during systems analysis, requirements analysis, and systems design to describe a system at a much higher level than a programming language, which is used to produce the executable code for a system.
Overview
Specification languages are generally not directly executed. They are meant to describe the what, not the how. It is considered an error if a requirement specification is cluttered with unnecessary implementation detail.
A common fundamental assumption of many specification approaches is that programs are modelled as algebraic or model-theoretic structures that include a collection of sets of data values together with functions over those sets. This level of abstraction coincides with the view that the correctness of the input/output behaviour of a program takes precedence over all its other properties.
In the property-oriented approach to specification (taken e.g. by CASL), specifications of programs consist mainly of logical axioms, usually in a logical system in which equality has a prominent role, describing the properties that the functions are required to satisfy—often just by their interrelationship.
This is in contrast to so-called model-oriented specification in frameworks like VDM and Z, which consist of a simple realization of the required behaviour.
Specifications must be subject to a process of refinement (the filling-in of implementation detail) before they can actually be implemented. The result of such a refinement process is an executable algorithm, which is either formulated in a programming language, or in an executable subset of the specification language at hand. For example, Hartmann pipelines, when properly applied, may be considered a dataflow specification which is directly executable. Another example is the actor model which has no specific application content and must be specialized to be executable.
An important use of specification languages is enabling the |
https://en.wikipedia.org/wiki/Call%E2%80%93Exner%20bodies | Call–Exner bodies, giving a follicle-like appearance, are small eosinophilic fluid-filled punched out spaces between granulosa cells. The granulosa cells are usually arranged haphazardly around the space.
They are pathognomonic for granulosa cell tumors.
Histologically, these tumors consists of monotonous islands of granulosa cells with "coffee-bean" nuclei. That same nuclear groove appearance noted in Brenner tumour, an epithelial-stromal ovarian tumor distinguishable by nests of transitional epithelial cells (urothelial) with longitudinal nuclear grooves (coffee bean nuclei) in abundant fibrous stroma.
They are composed of membrane-packaged secretion of granulosa cells and have relations to the formation of liquor folliculi which are seen among closely arranged granulosa cells.
They are named for Emma Louise Call (1847–1937), an American physician, and Sigmund Exner (1846–1926), an Austrian physiologist. |
https://en.wikipedia.org/wiki/North%20Vietnamese%20%C4%91%E1%BB%93ng | The đồng (Chữ Nôm: 銅; Chữ Hán: 元, nguyên) (; ) was the currency of North Vietnam from 3 November 1946 to 2 May 1978. It was subdivided into 10 hào, each itself divided into 10 xu.
History
The first đồng issued by the communists controlling northern Vietnam was introduced on January 31, 1946, and replaced the French Indochinese piastre at par. Two revaluations followed. In 1951, the second đồng was introduced at a rate of 1 1951 đồng = 100 1946 đồng. However, some sources say there were two consecutive revaluations in 1951 and 1953, each with factor of 10. In 1954, this became the currency of the newly recognized state of North Vietnam, with an exchange rate to the still circulating piastre and South Vietnamese đồng of 32 northern đồng = 1 piastre or southern đồng. In 1956, the đồng was pegged to the Chinese renminbi yuan at a rate of 1.47 đồng = 1 yuan.
On 28 February 1959, another đồng replaced the second at a rate of 1 1959 đồng = 1000 1951 đồng. An exchange rate with the Soviet rouble was established in 1961, with 3.27 đồng = 1 rouble. On May 3, 1978, following the unification of Vietnam, the đồng was also unified. 1 new đồng = 1 northern đồng = 0.8 southern "liberation" đồng.
Coins
1946 đồng
In 1946, aluminium 20 xu, 5 hào and 1 đồng and bronze 2 đồng were issued, with the 20 xu coins dated 1945. These were the only coins issued for this currency, with no coins at all issued for the 1951 đồng.
1958 đồng
In 1958, holed, aluminium coins in denominations of 1, 2 and 5 xu were introduced. These were the only coins issued in this currency.
Banknotes
1946 đồng
The government (Việt Nam Dân Chủ Cộng Hòa) issued two forms of paper money for this currency, "Vietnamese banknotes" (Giấy Bạc Việt Nam) and "Credit notes" (Tín Phiếu). In 1946, banknotes were introduced in denominations of 20 and 50 xu, 1, 5, 20, 50, 100 đồng, together with credit notes for 1 đồng. These were followed in 1948 by banknotes for 10 đồng and credit notes for 20 đồng, in 1949 by 500 đồng b |
https://en.wikipedia.org/wiki/Mirza%20melon | The Mirza melon, also known as the torpedo melon, Mirzachul melon, or gulabi melon, is a cultivar of sweet melon in the genus Cucumis native to Uzbekistan and Central Asia and introduced to California.
Description
They are large, weighing up to and measuring up to in length. They have an elongated shape and the rind is creamy yellow in color with beige streaking and the interior flesh is white and quite juicy. The flavor has been described as sweet and savory. In Uzbekistan, disfigured melons are typically harvested overripe and then sliced and sun-dried. The plant demands light and heat, and may be affected by powdery mildew. The fruit bursts easily.
Naming
The name Mirza has Persian roots and means "prince" or "high nobleman".
Availability
It is sometimes sold at various farmers' markets throughout California and the seeds are available through several online sources. |
https://en.wikipedia.org/wiki/Vasculum | A vasculum or a botanical box is a stiff container used by botanists to keep field samples viable for transportation. The main purpose of the vasculum is to transport plants without crushing them and by maintaining a cool, humid environment.
Construction
Vascula are cylinders typically made from tinned and sometimes lacquered iron, though wooden examples are known. The box was carried horizontally on a strap so that plant specimens lie flat and lined with moistened cloth. Traditionally, British and American vascula were somewhat flat and valise-like with a single room, while continental examples were more cylindrical and often longer, sometimes with two separate compartments. Access to the interior is through one (sometimes two) large lids in the side, allowing plants to be put in and taken out without bending or distorting them unnecessarily. This is particularly important with wildflowers, which are often fragile.
Some early 20th century specimen are made from sheet aluminium rather than tin, but otherwise follow the 19th century pattern. The exterior is usually left rough, or lacquered green.
History
The roots of the vasculum are lost in time, but may have evolved from the 17th century tin candle-box of similar construction. Linnaeus called it a vasculum dillenianum, from Latin vasculum – small container and dillenianum, referring to J.J. Dillenius, Linnaeus' friend and colleague at Oxford Botanic Garden. With rise of botany as a scientific field the mid 18th century, the vasculum became an indispensable part of the botanist's equipment.
Together with the screw-down plant press, the vasculum was popularized in Britain by naturalist William Withering around 1770. The shortened term "vasculum" appears to have become the common name applied to them around 1830. Being a hallmark of field botany, vascula were in common use until World War II . With post-war emphasis on systematics rather than alpha taxonomy and new species often collected in far-away places, field |
https://en.wikipedia.org/wiki/RADOM-7 | RADOM is a Bulgarian Liulin-type instruments-type spectrometry-dosimetry instrument, designed to precisely measure cosmic radiation around the Moon. It is installed on the Indian satellite Chandrayaan-1. Another three instruments were deployed on the International Space Station. All Liulin-type instruments are designed and build by the Solar-Terrestrial Influences Laboratory at the Bulgarian Academy of Sciences.
Further reading
Radiation
Space program of Bulgaria |
https://en.wikipedia.org/wiki/Flag%20of%20Fiji | The national flag of Fiji (Fijian: kuila ni Viti) was adopted on 10 October 1970. The state arms have been slightly modified but the flag has remained the same as during Fiji's colonial period. It is a defaced cyan "Blue Ensign" (the actual Blue Ensign version of the flag is the Government ensign), with the shield from the national coat of arms. It has remained unchanged since Fiji was declared a republic in 1987, despite calls from some politicians for changes.
A plan to change the flag, announced by Prime Minister Frank Bainimarama in 2013, was abandoned in August 2016.
Design
The flag's bright blue background symbolizes the Pacific Ocean, which plays an important part in the lives of the islanders, both in terms of the fishing industry, and the huge tourist trade. The Union Jack reflects the country's links with the United Kingdom. The shield is derived from the country's coat of arms, which was granted by Royal Warrant in 1908. It is a white shield with a red cross and a red chief (upper third of a shield). The images depicted on the shield represent agricultural activities on the islands, and the historical associations with the United Kingdom. At the top of the shield, a British lion holds a cocoa pod between its paws. The first quarter is sugar cane, the second quarter is a coconut palm, the third quarter is a dove of peace, and the fourth quarter is a bunch of bananas.
The flag is very similar to the colonial ensign used prior to independence, the main differences being the latter used a darker shade of blue and displayed the entire Fijian coat of arms as opposed to just the shield. While some reformists have called for the removal of the Union Flag, seeing it a British colonial emblem, others support its retention for the sake of historical continuity. The flags of five other independent countries (Australia, Cook Islands, New Zealand, Niue, and Tuvalu) retain the Union Flag in their national flags. But of these, only Fiji is a republic. The Union flag |
https://en.wikipedia.org/wiki/All%20Tomorrows | All Tomorrows: A Billion Year Chronicle of the Myriad Species and Mixed Fortunes of Man is a 2006 work of science fiction and speculative evolution written and illustrated by the Turkish artist C. M. Kosemen under the pen name Nemo Ramjet. It explores a hypothetical future path of human evolution set from the near future to a billion years from the present. Several future human species evolve through natural means and through genetic engineering, conducted by both humans themselves and by a mysterious and superior alien species called the Qu.
Inspired by the science fiction works of Olaf Stapledon and Edward Gibbon's The History of the Decline and Fall of the Roman Empire, Kosemen worked on All Tomorrows from 2003 to the publication of the book as a free PDF file online in 2006. The book has never been physically published, but as per Kosemen himself "had a life of its own" on the internet. Kosemen intends to eventually publish All Tomorrows in physical form, with new text and updated illustrations.
Summary
Centuries following the Terraforming and colonization of Mars, a brief but catastrophic Interplanetary War takes place between Mars and Earth. After both planets make peace with each other, a large-scale colonization initiative is carried out by genetically engineered humans called Star People throughout the galaxy.
Humans then encounter a malevolent and superior alien species called the Qu. The Qu's religion motivates them to remake the universe through genetic engineering. A short war follows in which humanity is defeated. The Qu bioengineer the surviving humans as punishment into a range of exotic forms, many of them unintelligent. After forty million years of domination, the Qu leave the galaxy, leaving the altered humans to evolve on their own. The bioengineered humans range from worm-like humans to insectivores and modular and cell-based species. The book follows the progress of these new humans as they either go extinct or regain sapience in wildly dif |
https://en.wikipedia.org/wiki/Alexander%20Aitken | Alexander Craig "Alec" Aitken (1 April 1895 – 3 November 1967) was one of New Zealand's most eminent mathematicians. In a 1935 paper he introduced the concept of generalized least squares, along with now standard vector/matrix notation for the linear regression model. Another influential paper co-authored with his student Harold Silverstone established the lower bound on the variance of an estimator, now known as Cramér–Rao bound. He was elected to the Royal Society of Literature for his World War I memoir, Gallipoli to the Somme.
Life and work
Aitken was born on 1 April 1895 in Dunedin, the eldest of the seven children of Elizabeth Towers and William Aitken. He was of Scottish descent, his grandfather having emigrated from Lanarkshire in 1868. His mother was from Wolverhampton.
He was educated at Otago Boys' High School in Dunedin (1908–13) where he was school dux and won the Thomas Baker Calculus Scholarship in his last year at school. He saw active service during World War I enlisting in April 1915 with the New Zealand Expeditionary Force, and serving in Gallipoli from November 1915, in Egypt, and at the Western Front. He was seriously wounded at the Somme. He spent several months in hospital in Chelsea before being invalided out of the army and shipped home to New Zealand in March 1917.
Resuming his studies Aitken graduated with an MA degree from the University of Otago in 1920, then worked as a schoolmaster at Otago Boys' High School from 1920 to 1923.
Aitken studied for a doctorate (PhD) at the University of Edinburgh, in Scotland, under Edmund Taylor Whittaker where his dissertation, "Smoothing of Data", was considered so impressive that he was awarded a DSc degree in 1925. Aitken's impact at the university had been so great that he had been elected a Fellow of the Royal Society of Edinburgh (FRSE) the year before the award of his degree, upon the proposal of Sir Edmund Whittaker, Sir Charles Galton Darwin, Edward Copson and David Gibb. Aitken was awar |
https://en.wikipedia.org/wiki/JStik | The JStik is a microcontroller based on the line of embedded Java processors. It is novel in that it uses Java byte code as the native machine language. This makes it very fast at executing Java code while still maintaining the benefits of programming in a high-level language like Java.
External links
Jstik - JStik Home
Systronix Inc. — Makers of the JStik system.
aJile Systems Inc. - Makers of the Silicon
Microcontrollers |
https://en.wikipedia.org/wiki/IBM%20Blue%20Gene | Blue Gene was an IBM project aimed at designing supercomputers that can reach operating speeds in the petaFLOPS (PFLOPS) range, with low power consumption.
The project created three generations of supercomputers, Blue Gene/L, Blue Gene/P, and Blue Gene/Q. During their deployment, Blue Gene systems often led the TOP500 and Green500 rankings of the most powerful and most power-efficient supercomputers, respectively. Blue Gene systems have also consistently scored top positions in the Graph500 list. The project was awarded the 2009 National Medal of Technology and Innovation.
As of 2015, IBM appears to have ended development of the Blue Gene family, though no formal announcement has been made. IBM has since focused its supercomputer efforts on the OpenPower platform, using accelerators such as FPGAs and GPUs to address the diminishing returns of Moore's law.
History
In December 1999, IBM announced a US$100 million research initiative for a five-year effort to build a massively parallel computer, to be applied to the study of biomolecular phenomena such as protein folding. The project had two main goals: to advance our understanding of the mechanisms behind protein folding via large-scale simulation, and to explore novel ideas in massively parallel machine architecture and software. Major areas of investigation included: how to use this novel platform to effectively meet its scientific goals, how to make such massively parallel machines more usable, and how to achieve performance targets at a reasonable cost, through novel machine architectures. The initial design for Blue Gene was based on an early version of the Cyclops64 architecture, designed by Monty Denneau. The initial research and development work was pursued at IBM T. J. Watson Research Center and led by William R. Pulleyblank.
At IBM, Alan Gara started working on an extension of the QCDOC architecture into a more general-purpose supercomputer: The 4D nearest-neighbor interconnection network was replaced by |
https://en.wikipedia.org/wiki/Symmetry%20of%20second%20derivatives | In mathematics, the symmetry of second derivatives (also called the equality of mixed partials) refers to the possibility of interchanging the order of taking partial derivatives of a function
of variables without changing the result under certain conditions (see below). The symmetry is the assertion that the second-order partial derivatives satisfy the identity
so that they form an symmetric matrix, known as the function's Hessian matrix. Sufficient conditions for the above symmetry to hold are established by a result known as Schwarz's theorem, Clairaut's theorem, or Young's theorem.
In the context of partial differential equations it is called the Schwarz integrability condition.
Formal expressions of symmetry
In symbols, the symmetry may be expressed as:
Another notation is:
In terms of composition of the differential operator which takes the partial derivative with respect to :
.
From this relation it follows that the ring of differential operators with constant coefficients, generated by the , is commutative; but this is only true as operators over a domain of sufficiently differentiable functions. It is easy to check the symmetry as applied to monomials, so that one can take polynomials in the as a domain. In fact smooth functions are another valid domain.
History
The result on the equality of mixed partial derivatives under certain conditions has a long history. The list of unsuccessful proposed proofs started with Euler's, published in 1740, although already in 1721 Bernoulli had implicitly assumed the result with no formal justification. Clairaut also published a proposed proof in 1740, with no other attempts until the end of the 18th century. Starting then, for a period of 70 years, a number of incomplete proofs were proposed. The proof of Lagrange (1797) was improved by Cauchy (1823), but assumed the existence and continuity of the partial derivatives and . Other attempts were made by P. Blanchet (1841), Duhamel (1856), Sturm (1857), Schl |
https://en.wikipedia.org/wiki/Continuous%20testing | Continuous testing is the process of executing automated tests as part of the software delivery pipeline to obtain immediate feedback on the business risks associated with a software release candidate. Continuous testing was originally proposed as a way of reducing waiting time for feedback to developers by introducing development environment-triggered tests as well as more traditional developer/tester-triggered tests.
For Continuous testing, the scope of testing extends from validating bottom-up requirements or user stories to assessing the system requirements associated with overarching business goals.
Adoption drivers
In the 2010s, software has become a key business differentiator. As a result, organizations now expect software development teams to deliver more, and more innovative, software within shorter delivery cycles. To meet these demands, teams have turned to lean approaches, such as Agile, DevOps, and Continuous Delivery, to try to speed up the systems development life cycle (SDLC). After accelerating other aspects of the delivery pipeline, teams typically find that their testing process is preventing them from achieving the expected benefits of their SDLC acceleration initiative. Testing and the overall quality process remain problematic for several key reasons.
Traditional testing processes are too slow. Iteration length has changed from months to weeks or days with the rising popularity of Agile, DevOps, and Continuous Delivery. Traditional methods of testing, which rely heavily on manual testing and automated GUI tests that require frequent updating, cannot keep pace. At this point, organizations tend to recognize the need to extend their test automation efforts.
Even after more automation is added to the existing test process, managers still lack adequate insight into the level of risk associated with an application at any given point in time. Understanding these risks is critical for making the rapid go/no go decisions involved in Continuous |
https://en.wikipedia.org/wiki/Cowboy%20coding | Cowboy coding is software development where programmers have autonomy over the development process. This includes control of the project's schedule, languages, algorithms, tools, frameworks and coding style. Typically, little to no coordination exists with other developers or stakeholders.
A cowboy coder can be a lone developer or part of a group of developers working with minimal process or discipline. Usually it occurs when there is little participation by business users, or fanned by management that controls only non-development aspects of the project, such as the broad targets, timelines, scope, and visuals (the "what", but not the "how").
"Cowboy coding" commonly sees usage as a derogatory term when contrasted with more structured software development methodologies.
Disadvantages
In cowboy coding, the lack of formal software project management methodologies may be indicative (though not necessarily) of a project's small size or experimental nature. Software projects with these attributes may exhibit:
Lack of release structure
Lack of estimation or implementation planning might cause a project to be delayed. Sudden deadlines or pushes to release software may encourage the use of "quick and dirty" techniques that will require further attention later.
Inexperienced developers
Cowboy coding can be common at the hobbyist or student level where developers might initially be unfamiliar with the technologies, such as testing, version control and/or build tools, usually more than just the basic coding a software project requires.
This can result in underestimating time required for learning, causing delays in the development process. Inexperience might also lead to disregard of accepted standards, making the project source difficult to read or causing conflicts between the semantics of the language constructs and the result of their output.
Uncertain design requirements
Custom software applications, even when using a proven development cycle, can experience prob |
https://en.wikipedia.org/wiki/Double%20hyphen | In Latin script, the double hyphen is a punctuation mark that consists of two parallel hyphens. It was a development of the earlier , which developed from a Central European variant of the virgule slash, originally a form of scratch comma. Similar marks (see below) are used in other scripts.
In order to avoid it being confused with the equals sign , the double hyphen is often shown as a double oblique hyphen in modern typography. The double hyphen is also not to be confused with two consecutive hyphens (--), which are often used to represent an em dash or en dash due to the limitations of typewriters and keyboards that do not have distinct hyphen and dash keys.
Usage
The double hyphen is used for several different purposes throughout the world:
Some typefaces, such as Fraktur faces, use the double hyphen as a glyphic variant of the single hyphen. (With Fraktur faces, such a double hyphen is usually oblique.)
It may be also used for artistic or commercial purposes to achieve a distinctive visual effect. For example, the name of The Waldorf⹀Astoria hotel was officially written with a double hyphen from 1949 to 2009.
In Merriam-Webster dictionaries if a word is divided at the end of the line, and the division point happens to be a hyphen, it is replaced with a double hyphen to graphically indicate that the divided word is normally hyphenated, for example cross⸗country.
In several dictionaries published in the late nineteenth and early twentieth centuries, all such compound words are linked with double hyphens, whether at the end of the line or not, and the normal use of the single hyphen for non-compound words is retained. An example from the first or second page of such dictionaries is Aaron's⸗rod. Examples include the Century Dictionary and Funk & Wagnalls New Standard Dictionary of the English Language.
It is used by Coptic language scholars to denote the form of the verb used before pronominal suffixes, e.g. ⲕⲟⲧ⸗ kot⹀ 'to build'.
It is used by scholar |
https://en.wikipedia.org/wiki/Hilbert%27s%20theorem%20%28differential%20geometry%29 | In differential geometry, Hilbert's theorem (1901) states that there exists no complete regular surface of constant negative gaussian curvature immersed in . This theorem answers the question for the negative case of which surfaces in can be obtained by isometrically immersing complete manifolds with constant curvature.
History
Hilbert's theorem was first treated by David Hilbert in "Über Flächen von konstanter Krümmung" (Trans. Amer. Math. Soc. 2 (1901), 87–99).
A different proof was given shortly after by E. Holmgren in "Sur les surfaces à courbure constante négative" (1902).
A far-leading generalization was obtained by Nikolai Efimov in 1975.
Proof
The proof of Hilbert's theorem is elaborate and requires several lemmas. The idea is to show the nonexistence of an isometric immersion
of a plane to the real space . This proof is basically the same as in Hilbert's paper, although based in the books of Do Carmo and Spivak.
Observations: In order to have a more manageable treatment, but without loss of generality, the curvature may be considered equal to minus one, . There is no loss of generality, since it is being dealt with constant curvatures, and similarities of multiply by a constant. The exponential map is a local diffeomorphism (in fact a covering map, by Cartan-Hadamard theorem), therefore, it induces an inner product in the tangent space of at : . Furthermore, denotes the geometric surface with this inner product. If is an isometric immersion, the same holds for
.
The first lemma is independent from the other ones, and will be used at the end as the counter statement to reject the results from the other lemmas.
Lemma 1: The area of is infinite.
Proof's Sketch:
The idea of the proof is to create a global isometry between and . Then, since has an infinite area, will have it too.
The fact that the hyperbolic plane has an infinite area comes by computing the surface integral with the corresponding coefficients of the First fundamen |
https://en.wikipedia.org/wiki/Virus%20%28novel%29 | , literally Day of Resurrection is a 1964 post-apocalyptic science fiction novel written by Sakyo Komatsu. The novel was adapted into a movie of the same name in 1980.
An English translation by Daniel Huddleston was released in November 2012 by Viz Media's Haikasoru imprint.
Plot summary
A US space mission in 1964 gathers a group of microbes in Earth orbit and are later recovered by American biowarfare researchers. Two microbes are found to be coccus-shaped supergerms capable of surviving in absolute zero and have the potential to grow exponentially in terrestrial conditions. One of the researchers, Dr Meyer, discovers the germ's regenerative ability and tries to stop work on the project. However, a sample of the germs is stolen and sent to Britain's Germ Warfare Research Laboratory in Porton Down for further development.
In February 19 (actually implied as 1970, later to be known as the Year of the Calamity), a scientist at Porton Down smuggles a sample of the developed virus, codenamed MM88, and gives it to a group of men who have been tasked to bring it to Dr. Leisener, a Czech molecular biologist who is skilled at developing antidotes for it. The scientist, Dr. Karlsky, insists that the germ must never be exposed to warm temperatures. However, the men double-cross him and an assistant later kills him under the guise of a suicide. The group flies out of England bound for a rendezvous with US intelligence agents in Turkey. They never make it; their plane crashes into the Alps in the middle of a snowstorm and the vial holding the virus breaks open. The crash site is investigated but investigators could not link it with the theft of the virus.
The virus, which is designed to piggyback with other existing illnesses and make them more lethal, begins to reproduce with the advent of spring in the Alps. Although some local communities have been reporting small animals suddenly dropping dead in their area, an Italian matinee idol's death in an auto crash in early Marc |
https://en.wikipedia.org/wiki/Enrollment%20over%20Secure%20Transport | The Enrollment over Secure Transport, or EST is a cryptographic protocol that describes an X.509 certificate management protocol targeting public key infrastructure (PKI) clients that need to acquire client certificates and associated certificate authority (CA) certificates. EST is described in . EST has been put forward as a replacement for SCEP, being easier to implement on devices already having an HTTPS stack. EST uses HTTPS as transport and leverages TLS for many of its security attributes. EST has described standardized URLs and uses the well-known Uniform Resource Identifiers (URIs) definition codified in .
Operations
EST has a following set of operations:
Usage example
The basic functions of EST were designed to be easy to use and although not a REST API, it can be used in a REST-like manner using simple tools such as OpenSSL and cURL.
A simple command to make initial enrollment with a pre-generated PKCS#10 Certificate Signing Request (stored as device.b64), using one of the authentication mechanisms (username:password) specified in EST is:
The issued certificate, returned as a Base64 encoded PKCS#7 message, is stored as device-p7.b64.
See also
Certificate Management Protocol (CMP)
Certificate Management over CMS (CMC)
Simple Certificate Enrollment Protocol (SCEP)
Automated Certificate Management Environment (ACME) |
https://en.wikipedia.org/wiki/TeslaCrypt | TeslaCrypt was a ransomware trojan. It is now defunct, and its master key was released by the developers.
In its early forms, TeslaCrypt targeted game-play data for specific computer games. Newer variants of the malware also affect other file types.
In its original, game-player campaign, upon infection the malware searched for 185 file extensions related to 40 different games, which include the Call of Duty series, World of Warcraft, Minecraft and World of Tanks, and encrypted such files. The files targeted involve the save data, player profiles, custom maps and game mods stored on the victim's hard drives. Newer variants of TeslaCrypt were not focused on computer games alone but also encrypted Word, PDF, JPEG and other files. In all cases, the victim would then be prompted to pay a ransom of $500 worth of bitcoins in order to obtain the key to decrypt the files.
Although resembling CryptoLocker in form and function, Teslacrypt shares no code with CryptoLocker and was developed independently. The malware infected computers via the Angler Adobe Flash exploit.
Even though the ransomware claimed TeslaCrypt used asymmetric encryption, researchers from Cisco's Talos Group found that symmetric encryption was used and developed a decryption tool for it. This "deficiency" was changed in version 2.0, rendering it impossible to decrypt files affected by TeslaCrypt-2.0.
By November 2015, security researchers from Kaspersky had been quietly circulating that there was a new weakness in version 2.0, but carefully keeping that knowledge away from the malware developer so that they could not fix the flaw. As of January 2016, a new version 3.0 was discovered that had fixed the flaw.
A full behavior report, which shows BehaviorGraphs and ExecutionGraphs was published by JoeSecurity.
Shut down
In May 2016, the developers of TeslaCrypt shut down the ransomware and released the master decryption key, thus bringing an end to the ransomware. After a few days, ESET released a pub |
https://en.wikipedia.org/wiki/Fossil%20preparation | Fossil preparation is a complex of tasks that can include excavating, revealing, conserving, and replicating the ancient remains and traces of organisms. It is an integral part of the science of paleontology, of museum exhibition, and the preservation of fossils held in the public trust. It involves a wide variety of techniques, from the mechanical to the chemical, depending upon the qualities of the specimen being prepared and the goals of the effort. Fossil preparation may be executed by scientists, students or collections personnel, but is often undertaken by professional fossil preparators.
Techniques
Acid maceration
Acid maceration is a technique to extract organic microfossils from a surrounding rock matrix using acid. Hydrochloric acid or acetic acid may be used to extract phosphatic fossils, such as the small shelly fossils, from a carbonate matrix. Hydrofluoric acid is also used in acid macerations to extract organic fossils from silicate rocks. Fossiliferous rock may be immersed directly into the acid, or a cellulose nitrate film may be applied (dissolved in amyl acetate), which adheres to the organic component and allows the rock to be dissolved around it.
Film pull
The film pull technique is a means of recovering carbonaceous compression fossils for study under transmitted light microscopy. An acid is applied to the surface of the rock to etch away the matrix from the surface, leaving carbonaceous tissue protruding. (Surfaces not to be etched can be coated in a wax (e.g. Vaseline or grease). This is usually accomplished by placing the rock upside-down in a weak, continually stirred acid, so that any debris can be washed away. Nitrocellulose is then painted on to the fossil-bearing surface, and once dry may be peeled from the rock, or the rock dissolved in hydrofluoric acid.
The method was pioneered by John Walton, in collaboration with Reitze Gerben Koopmans, in 1928 as a method to derive serial thin-sections without the time, expense and lost ma |
https://en.wikipedia.org/wiki/Relative%20fluorescence%20units | The terms "relative fluorescence units" (RFU) and "RFU peak" refer to measurements in electrophoresis methods, such as for DNA analysis. A "relative fluorescence unit" is a unit of measurement used in analysis which employs fluorescence detection. Fluorescence is detected using a charged coupled device (CCD) array, when the labeled fragments, which are separated within a capillary by using electrophoresis, are energized by laser light and travel across the detection window. A computer program measures the results, determining the quantity or size of the fragments, at each data point, from the level of fluorescence intensity. Samples which contain higher quantities of amplified DNA will have higher corresponding RFU values.
An "RFU peak" is a relative maximum point along a graph of the analyzed data. The data can be normalized to DNA input or additional normalizing genes. The RFU heights can range from 0 to several thousands.
DNA PCR analysis
The RFU measurements are used, for DNA profiling, in a real-time polymerase chain reaction (PCR). Two common methods for detection of products in real-time PCR are: (1) non-specific fluorescent dyes that intercalate with any double-stranded DNA, and (2) sequence-specific DNA probes consisting of oligonucleotides that are labeled with a fluorescent reporter which permits detection only after hybridization of the probe with its complementary DNA target. Frequently, real-time PCR is combined with reverse transcription to quantify messenger RNA and Non-coding RNA in cells or tissues.
RFU peak thresholds
The RFU peak height depends on the amount of DNA being analyzed. When the amount of DNA is very low, then
it can be difficult to separate a true low-level RFU peak from signal noise or other technical artifacts. As a result, many forensic DNA laboratories have set minimum RFU peak-height levels in "scoring" the analysis of alleles.
There are no firm industry-wide rules for establishing minimum RFU threshold values. Each laborator |
https://en.wikipedia.org/wiki/Brownian%20motor | Brownian motors are nanoscale or molecular machines that use chemical reactions to generate directed motion in space. The theory behind Brownian motors relies on the phenomenon of Brownian motion, random motion of particles suspended in a fluid (a liquid or a gas) resulting from their collision with the fast-moving molecules in the fluid.
On the nanoscale (1-100 nm), viscosity dominates inertia, and the extremely high degree of thermal noise in the environment makes conventional directed motion all but impossible, because the forces impelling these motors in the desired direction are minuscule when compared to the random forces exerted by the environment. Brownian motors operate specifically to utilise this high level of random noise to achieve directed motion, and as such are only viable on the nanoscale.
The concept of Brownian motors is a recent one, having only been coined in 1995 by Peter Hänggi, but the existence of such motors in nature may have existed for a very long time and help to explain crucial cellular processes that require movement at the nanoscale, such as protein synthesis and muscular contraction. If this is the case, Brownian motors may have implications for the foundations of life itself.
In more recent times, humans have attempted to apply this knowledge of natural Brownian motors to solve human problems. The applications of Brownian motors are most obvious in nanorobotics due to its inherent reliance on directed motion.
History
20th century
The term “Brownian motor” was originally invented by Swiss theoretical physicist Peter Hänggi in 1995. The Brownian motor, like the phenomenon of Brownian motion that underpinned its underlying theory, was also named after 19th century Scottish botanist Robert Brown, who, while looking through a microscope at pollen of the plant Clarkia pulchella immersed in water, famously described the random motion of pollen particles in water in 1827. In 1905, almost eighty years later, theoretical physicist Alb |
https://en.wikipedia.org/wiki/Summer%20Rain%20%28Belinda%20Carlisle%20song%29 | "Summer Rain" is a song written by Robbie Seidman and Maria Vidal, produced by Rick Nowels for Belinda Carlisle's third album Runaway Horses (1989). The power ballad is about a man who goes away to war and leaves his wife, saying that nothing will change—they will be together forever and always. Although the conflict is unidentified, images in the video of a transport aircraft on an airfield and troops parachuting from transport aircraft suggest the man is an airborne soldier. The song is set in the present as his widow sings it, remembering the last time she saw him. It was released around the world in 1990 and was issued as a CD single, 7-inch single and a 12-inch single.
The song had reasonable success on the charts reaching the top 30 in the United States and the United Kingdom. In Australia, it reached the top 10, peaking at number six. It was later covered by artists including Slinkee Minx, whose 2004 cover version charted higher than the original in Australia, reaching number five. Carlisle stated in May 2013 that "Summer Rain" was her favorite song from her recording career.
Release and commercial performance
"Summer Rain" achieved moderate commercial success in North America, where it was released as the second single from Runaway Horses (1989). In January 1990, the song entered the Billboard Hot 100, the main US chart, at number 86. Within seven weeks of its release, it peaked at number 30 and stayed at that position for two weeks. The single spent 13 weeks on the Billboard Hot 100, six of which were in the top 50. The single was Carlisle's second to last song to chart on the Billboard Hot 100. It also peaked at number 29 on the US Billboard Adult Contemporary chart, where it spent eight weeks on the chart.
In Australia, the song had the most commercial success, where it was released as the album's third single on March 19, 1990. In early April 1990, it debuted at number 42. Within two months of its release, it reached the top 10, where it stayed for f |
https://en.wikipedia.org/wiki/Neural%20decoding | Neural decoding is a neuroscience field concerned with the hypothetical reconstruction of sensory and other stimuli from information that has already been encoded and represented in the brain by networks of neurons. Reconstruction refers to the ability of the researcher to predict what sensory stimuli the subject is receiving based purely on neuron action potentials. Therefore, the main goal of neural decoding is to characterize how the electrical activity of neurons elicit activity and responses in the brain.
This article specifically refers to neural decoding as it pertains to the mammalian neocortex.
Overview
When looking at a picture, people's brains are constantly making decisions about what object they are looking at, where they need to move their eyes next, and what they find to be the most salient aspects of the input stimulus. As these images hit the back of the retina, these stimuli are converted from varying wavelengths to a series of neural spikes called action potentials. These pattern of action potentials are different for different objects and different colors; we therefore say that the neurons are encoding objects and colors by varying their spike rates or temporal pattern. Now, if someone were to probe the brain by placing electrodes in the primary visual cortex, they may find what appears to be random electrical activity. These neurons are actually firing in response to the lower level features of visual input, possibly the edges of a picture frame. This highlights the crux of the neural decoding hypothesis: that it is possible to reconstruct a stimulus from the response of the ensemble of neurons that represent it. In other words, it is possible to look at spike train data and say that the person or animal being recorded is looking at a red ball.
With the recent breakthrough in large-scale neural recording and decoding technologies, researchers have begun to crack the neural code and already provided the first glimpse into the real-time neural |
https://en.wikipedia.org/wiki/ZX-calculus | The ZX-calculus is a rigorous graphical language for reasoning about linear maps between qubits, which are represented as string diagrams called ZX-diagrams. A ZX-diagram consists of a set of generators called spiders that represent specific tensors. These are connected together to form a tensor network similar to Penrose graphical notation. Due to the symmetries of the spiders and the properties of the underlying category, topologically deforming a ZX-diagram (i.e. moving the generators without changing their connections) does not affect the linear map it represents. In addition to the equalities between ZX-diagrams that are generated by topological deformations, the calculus also has a set of graphical rewrite rules for transforming diagrams into one another. The ZX-calculus is universal in the sense that any linear map between qubits can be represented as a diagram, and different sets of graphical rewrite rules are complete for different families of linear maps. ZX-diagrams can be seen as a generalisation of quantum circuit notation.
History
The ZX-calculus was first introduced by Bob Coecke and Ross Duncan in 2008 as an extension of the categorical quantum mechanics school of reasoning. They introduced the fundamental concepts of spiders, strong complementarity and most of the standard rewrite rules.
In 2009 Duncan and Perdrix found the additional Euler decomposition rule for the Hadamard gate, which was used by Backens in 2013 to establish the first completeness result for the ZX-calculus. Namely that there exists a set of rewrite rules that suffice to prove all equalities between stabilizer ZX-diagrams, where phases are multiples of , up to global scalars. This result was later refined to completeness including scalar factors.
Following an incompleteness result, in 2017, a completion of the ZX-calculus for the approximately universal fragment was found, in addition to two different completeness results for the universal ZX-calculus (where phases are allow |
https://en.wikipedia.org/wiki/Odour%20activity%20value | Odour activity value (OAV) is a measure of importance of a specific compound to the odour of a sample (e.g. food). It is calculated as the ratio between the concentration of individual substance in a sample and the threshold concentration of this substance (odour threshold value, the minimal concentration that can be detected by human nose). |
https://en.wikipedia.org/wiki/Ship%20model | Ship models or model ships are scale models of ships. They can range in size from 1/6000 scale wargaming miniatures to large vessels capable of holding people.
Ship modeling is a craft as old as shipbuilding itself, stretching back to ancient times when water transport was first developed.
History
Ancient Mediterranean
Ancient ship and boat models have been discovered throughout the Mediterranean, especially from ancient Greece, Egypt, and Phoenicia. These models provide archaeologists with valuable information regarding seafaring technology and the sociological and economic importance of seafaring. In spite of how helpful ancient boat and ship models are to archaeologists, they are not always easily or correctly interpreted due to artists’ mistakes, ambiguity in the model design, and wear and tear over the centuries.
Ships "were among the most technologically complex mechanisms of the ancient world." Ships made far-flung travel and trade more comfortable and economical, and they added a whole new facet to warfare. Thus, ships carried a great deal of significance to the people of the ancient world, and this is expressed partly through the creation of boat and ship models. Ancient boat and ship models are made of a variety of materials and are intended for different purposes. The most common purposes for boat and ship models include burial votives, house hold articles, art, and toys. While archaeologists have found ship and boat models from societies all around the Mediterranean, the three of the most prolific ship model building cultures were the Greeks, Phoenicians, and Egyptians.
Archaeologists have determined that Ancient Greek ship models were used as burial or votive offerings and as household articles such as lamps or drinking vessels. The kinds of ships depicted in Ancient Greek models can be classified broadly as small craft, merchant vessels, and warships. Models were cast in different materials, including wood, bronze, lead, and clay.
Greek wars |
https://en.wikipedia.org/wiki/Life%20Extension%20Institute | The Life Extension Institute was an organization formed in the United States in 1913 with the philanthropic goal of prolonging human life through hygiene and disease prevention. Its organizational officers included many celebrity-philanthropists such as William Howard Taft, Alexander Graham Bell, and Mabel Thorp Boardman but also genuine medical experts including William James Mayo, Russell Henry Chittenden, and J. H. Kellogg and a "Hygiene Reference Board" of dozens of nationally recognized physicians of that era such as Mazÿck Porcher Ravenel and Major General William Crawford Gorgas.
A major project of the institute which fulfilled its mission to disseminate knowledge was publication of the book How to Live, Rules for Healthful Living Based on Modern Science, now in the public domain.
The institute was a proponent of eugenics including sterilization of grossly "unfit" individuals:
See also
Life extension |
https://en.wikipedia.org/wiki/Medical%20photography | Medical photography is a specialized area of photography that concerns itself with the documentation of the clinical presentation of patients, medical and surgical procedures, medical devices and specimens from autopsy. The practice requires a high level of technical skill to present the photograph free from misleading information that may cause misinterpretation. The photographs are used in clinical documentation, research, publication in scientific journals and teaching.
The profession
Employment
Medical photographers document patients at various stages of an illness, injuries and before and after surgical procedures. They record the work of healthcare professionals to assist in the planning of treatment and education of the public and other healthcare professionals. The nature of the work requires a respect for and sensitivity to people, an awareness of sterile procedures and an adherence to privacy legislation and policies.
The BioCommunications Association, in a survey commissioned in 2008 of individuals working in medical photography, found that most medical photographers are employed by university affiliated hospitals and research centers. Ten percent were freelancers working in specialty clinics such as dermatology, ophthalmology and plastic surgery. A few of these provided services to the medical-legal profession. Medical photographers photograph patients in clinics, wards and in operating rooms. They may also be called to photograph an autopsy and gross specimens in the pathology department. Specialized photography techniques using photomacrography and ultra-violet and fluorescence photography may also be used. The role of the medical photographer has changed over the years from being exclusively medical to incorporating more general photography of a commercial or editorial nature to support public relations and education. Video production is playing an increased role; medical photographers are often responsible for video conferencing from operating roo |
https://en.wikipedia.org/wiki/NAPSA | Napsin-A is an aspartic proteinase that is encoded in humans by the NAPSA gene. The name napsin comes from novel aspartic proteinase of the pepsin family.
The activation peptide of an aspartic proteinase acts as an inhibitor of the active site. These peptide segments, or pro-parts, are deemed important for correct folding, targeting, and control of the activation of aspartic proteinase zymogens. The pronapsin A gene is expressed predominantly in lung and kidney. Its translation product is predicted to be a fully functional, glycosylated aspartic proteinase precursor containing an RGD motif and an additional 18 residues at its C-terminus.
Utility
Detection of NAPSA gene expression can be used to distinguish adenocarcinomas from other forms of lung cancer. |
https://en.wikipedia.org/wiki/Cylindric%20numbering | In computability theory a cylindric numbering is a special kind of numbering first introduced by Yuri L. Ershov in 1973.
If a numbering is reducible to then there exists a computable function with . Usually is not injective, but if is a cylindric numbering we can always find an injective .
Definition
A numbering is called cylindric if
That is if it is one-equivalent to its cylindrification
A set is called cylindric if its indicator function
is a cylindric numbering.
Examples
Every Gödel numbering is cylindric
Properties
Cylindric numberings are idempotent: |
https://en.wikipedia.org/wiki/Material%20nonimplication | Material nonimplication or abjunction (Latin ab = "away", junctio= "to join") is the negation of material implication. That is to say that for any two propositions and , the material nonimplication from to is true if and only if the negation of the material implication from to is true. This is more naturally stated as that the material nonimplication from to is true only if is true and is false.
It may be written using logical notation as , , or "Lpq" (in Bocheński notation), and is logically equivalent to , and .
Definition
Truth table
Logical Equivalences
Material nonimplication may be defined as the negation of material implication.
In classical logic, it is also equivalent to the negation of the disjunction of and , and also the conjunction of and
Properties
falsehood-preserving: The interpretation under which all variables are assigned a truth value of "false" produces a truth value of "false" as a result of material nonimplication.
Symbol
The symbol for material nonimplication is simply a crossed-out material implication symbol. Its Unicode symbol is 219B16 (8603 decimal): ↛.
Natural language
Grammatical
"p minus q."
"p without q."
Rhetorical
"p but not q."
"q is false, in spite of p."
Computer science
Bitwise operation: A&(~B)
Logical operation: A&&(!B)
See also
Implication
Boolean algebra
Set difference |
https://en.wikipedia.org/wiki/CCDC116 | CCDC116, also called coiled-coil domain containing 116, is a gene that is patented for experimentation on the possibility of being a cancer marker for prostate cancer.
Gene
CCDC116 has no aliases. This gene is 2252 base pairs long. CCDC116 has 5 exons and is located on chromosome 22q11.21 on the plus strand.
mRNA
CCDC116 has two isoforms. The isoforms differ in the 3’ UTR and coding sequence.
Protein
CCDC116 is 612 amino acids long. CCDC116 is 67.9 kdal. This protein has an isoelectric point of 9.24. This protein is a part of the domain of unknown function 4702.
It is predicted that CCDC116 is located in the nucleus. It is predicted that the membrane typology of this protein is a type 3a rolled up beta sheet. This protein has mostly alpha helices with a few coils.
Expression
CCDC116 is primarily found in the testis, although this gene is found in very small amounts in the brain and connective tissue. There are two SNPs that are upstream variants, one that is a downstream variant, one that is an intron variant.
Regulation of Expression
CCDC116 has four promoter regions and eight primary transcripts. Two promoter regions located on the + strand, and two located on the - strand.
Function Protein Partners
CCDC116 has two predicted function partners. NGFRAP is a nerve growth factor receptor associated protein. This gene is believed to play a role in the pathogenesis of neurogenetic diseases, although that claim has not been proven.
Clinical Significance
There has yet to be any disease association with CCDC116, but currently there is a patent for this gene with the potential for it to be related to cancer. A study was done in 2012 on observing how this gene is expressed in human pancreatic islets and in endocrine pancreatic tumors. |
https://en.wikipedia.org/wiki/FM%20Global | FM Global is an American mutual insurance company based in Johnston, Rhode Island, United States, with offices worldwide, that specializes in loss prevention services primarily to large corporations throughout the world in the Highly Protected Risk (HPR) property insurance market sector. "FM Global" is the communicative name of the company, whereas the legal name is "Factory Mutual Insurance Company". FM Global has been named the "Best Property Insurer in the World” by Euromoney Magazine.
The company employs a non-traditional business model whereby risk and premiums are determined by engineering analysis as opposed to historically based actuarial calculations. This business approach is centered on the belief that property losses can be prevented or mitigated. FM Global engineering personnel regularly visit insured locations to evaluate hazards and recommend improvements to their property or work practices to reduce physical and financial risks if a loss occurs.
History
During the depression of 1835, Zachariah Allen, a prominent textile mill owner, attempted to reduce the insurance premium on his Rhode Island, USA, mill by making property improvements that he believed would minimize the damage in case of fire. At that time, insurance premium increases for losses were shared among all insureds, regardless of individual loss history. The concept of loss prevention and control was virtually unheard of at the time. To Allen, a proactive approach to preventing losses made good economic sense.
After making considerable improvements to his mill, Allen requested a reduction in his premium, but was denied. He called upon other local textile mill owners who shared his loss prevention philosophy to create a mutual insurance company that would insure only factories with lower risks. This approach should result in fewer losses and smaller premium payments. Whatever premium remained at the end of the year would be returned to policyholders in the form of dividends. The group a |
https://en.wikipedia.org/wiki/Parallel%20computing | Parallel computing is a type of computation in which many calculations or processes are carried out simultaneously. Large problems can often be divided into smaller ones, which can then be solved at the same time. There are several different forms of parallel computing: bit-level, instruction-level, data, and task parallelism. Parallelism has long been employed in high-performance computing, but has gained broader interest due to the physical constraints preventing frequency scaling. As power consumption (and consequently heat generation) by computers has become a concern in recent years, parallel computing has become the dominant paradigm in computer architecture, mainly in the form of multi-core processors.
Parallel computing is closely related to concurrent computing—they are frequently used together, and often conflated, though the two are distinct: it is possible to have parallelism without concurrency, and concurrency without parallelism (such as multitasking by time-sharing on a single-core CPU). In parallel computing, a computational task is typically broken down into several, often many, very similar sub-tasks that can be processed independently and whose results are combined afterwards, upon completion. In contrast, in concurrent computing, the various processes often do not address related tasks; when they do, as is typical in distributed computing, the separate tasks may have a varied nature and often require some inter-process communication during execution.
Parallel computers can be roughly classified according to the level at which the hardware supports parallelism, with multi-core and multi-processor computers having multiple processing elements within a single machine, while clusters, MPPs, and grids use multiple computers to work on the same task. Specialized parallel computer architectures are sometimes used alongside traditional processors, for accelerating specific tasks.
In some cases parallelism is transparent to the programmer, such as in |
https://en.wikipedia.org/wiki/Spherical%20model | The spherical model is a model of ferromagnetism similar to the Ising model, which was solved in 1952 by T. H. Berlin and M. Kac. It has the remarkable property that for linear dimension d greater than four, the critical exponents that govern the behaviour of the system near the critical point are independent of d and the geometry of the system. It is one of the few models of ferromagnetism that can be solved exactly in the presence of an external field.
Formulation
The model describes a set of particles on a lattice containing N sites. Each site j of contains a spin which interacts only with its nearest neighbours and an external field H. It differs from the Ising model in that the are no longer restricted to , but can take all real values, subject to the constraint that
which in a homogeneous system ensures that the average of the square of any spin is one, as in the usual Ising model.
The partition function generalizes from that of the Ising model to
where is the Dirac delta function, are the edges of the lattice, and and , where T is the temperature of the system, k is Boltzmann's constant and J the coupling constant of the nearest-neighbour interactions.
Berlin and Kac saw this as an approximation to the usual Ising model, arguing that the -summation in the Ising model can be viewed as a sum over all corners of an N-dimensional hypercube in -space. The becomes an integration over the surface of a hypersphere passing through all such corners.
It was rigorously proved by Kac and C. J. Thompson that the spherical model is a limiting case of the N-vector model.
Equation of state
Solving the partition function and using a calculation of the free energy yields an equation describing the magnetization M of the system
for the function g defined as
The internal energy per site is given by
an exact relation relating internal energy and magnetization.
Critical behaviour
For the critical temperature occurs at absolute zero, resulting in no phase transi |
https://en.wikipedia.org/wiki/Spotter%20Network | The Spotter Network (SN) is a system that utilizes storm spotter and chaser reports of location and severe weather in a centralized framework for use by coordinators such as emergency managers, Skywarn and related spotter organizations, and the National Weather Service. It uses GPS to provide accurate and automated position data of storm spotters and chasers for coordination and reporting, which in turn provides ground truth to public servants engaged in the protection of life and property. The network is a combination of locally installed software for position and status reporting and web-based processing, mapping, and reporting.
The original Spotter Network was developed by Tyler Allison. The current president of the organization is John Wetter. It became operational in April 2006 and quickly grew to over 100 spotters. Several National Weather Service (NWS) employees and other officials soon took an interest in the capabilities it brings to them to integrate ground truth provided by spotters into their operational responsibilities. Subsequent versions of the network expanded the coordinator and reporting capabilities, and NWS eSpotter integration was completed in early September 2006.
Spotters must pass an online test of storm structure and basic meteorology in order to use the system. All reports are also reviewed for quality control purposes. Contact information is provided by users and can be controlled to reach the all users (the general public) or selectively to reach emergency managers and NWS officials. SN features GIS capabilities for use with external websites and apps.
Several papers have been written on the use of the Spotter Network in meteorological research and operations such as:
Emerging Technologies in the Field to Improve Information in Support of Operations and Research
The Digital Revolution of Storm Spotting Modernizations of Training, Tracking, and Reporting
Enriching the Modern Day Storm Spotter Through Technology & Education Enhance |
https://en.wikipedia.org/wiki/Dual%20quaternion | In mathematics, the dual quaternions are an 8-dimensional real algebra isomorphic to the tensor product of the quaternions and the dual numbers. Thus, they may be constructed in the same way as the quaternions, except using dual numbers instead of real numbers as coefficients. A dual quaternion can be represented in the form , where A and B are ordinary quaternions and ε is the dual unit, which satisfies and commutes with every element of the algebra.
Unlike quaternions, the dual quaternions do not form a division algebra.
In mechanics, the dual quaternions are applied as a number system to represent rigid transformations in three dimensions. Since the space of dual quaternions is 8-dimensional and a rigid transformation has six real degrees of freedom, three for translations and three for rotations, dual quaternions obeying two algebraic constraints are used in this application. Since unit quaternions are subject to two algebraic constraints, unit quaternions are standard to represent rigid transformations.
Similar to the way that rotations in 3D space can be represented by quaternions of unit length, rigid motions in 3D space can be represented by dual quaternions of unit length. This fact is used in theoretical kinematics (see McCarthy), and in applications to 3D computer graphics, robotics and computer vision. Polynomials with coefficients given by (non-zero real norm) dual quaternions have also been used in the context of mechanical linkages design.
History
W. R. Hamilton introduced quaternions in 1843, and by 1873 W. K. Clifford obtained a broad generalization of these numbers that he called biquaternions, which is an example of what is now called a Clifford algebra.
In 1898 Alexander McAulay used Ω with Ω2 = 0 to generate the dual quaternion algebra. However, his terminology of "octonions" did not stick as today's octonions are another algebra.
In Russia, Aleksandr Kotelnikov developed dual vectors and dual quaternions for use in the study of mec |
https://en.wikipedia.org/wiki/Fault%20current%20limiter | A fault current limiter (FCL), also known as fault current controller (FCC), is a device which limits the prospective fault current when a fault occurs (e.g. in a power transmission network) without complete disconnection. The term includes superconducting, solid-state and inductive devices.
Applications
Electric power distribution systems include circuit breakers to disconnect power in case of a fault, but to maximize reliability, they wish to disconnect the smallest possible portion of the network. This means that even the smallest circuit breakers, as well as all wiring to them, must be able to disconnect large fault currents.
A problem arises if the electricity supply is upgraded, by adding new generation capacity or by adding cross-connections. Because these increase the amount of power that can be supplied, all of the branch circuits must have their bus bars and circuit breakers upgraded to handle the new higher fault current limit.
This poses a particular problem when distributed generation, such as wind farms and rooftop solar power, is added to an existing electric grid. It is desirable to be able to add additional power sources without large system-wide upgrades.
A simple solution is to add electrical impedance to the circuit. This limits the rate at which current can increase, which limits the level the fault current can rise to before the breaker is opened. However, this also limits the ability of the circuit to satisfy rapidly changing demand, so the addition or removal of large loads causes unstable power.
A fault current limiter is a nonlinear element which has a low impedance at normal current levels, but presents a higher impedance at fault current levels. Further, this change is extremely rapid, before a circuit breaker can trip a few milliseconds later. (High-power circuit breakers are synchronized to the alternating current zero crossing to minimize arcing.)
While the power is unstable during the fault, it is not completely disconne |
https://en.wikipedia.org/wiki/H3S10P | H3S10P is an epigenetic modification to the DNA packaging protein histone H3. It is a mark that indicates the phosphorylation the 10th serine residue of the histone H3 protein.
Depending on the environment in which it happens, the same phosphorylated residue might have drastically different consequences on chromatin structure. H3S10 and H3S28 phosphorylation is an excellent illustration of this duality: both phosphorylated residues are involved in chromatin compaction during mitosis and meiosis, as well as chromatin relaxation after transcription activation. Because of the diverse roles that these phosphorylation events play, individual residue phosphorylation can't be examined in isolation. The impact is determined by the cellular environment and interplay with other cis or trans changes.
The S10 phosphorylation is involved in mitosis, transcription, chromatin condensation, and UVB response. H3S10p causes chromosome condensation and segregation during cell mitosis. H3S10p temporarily increases during mitosis while H3K9me3 decareses and H3K9me3 recovers upon mitotic exit.
R loops are linked to H3S10P and chromatin condensation.
Nomenclature
The name of this modification indicates the protein phosphorylation of serine 10 on histone H3 protein subunit:
Serine/threonine/tyrosine phosphorylation
The addition of a negatively charged phosphate group can lead to major changes in protein structure, leading to the well-characterized role of phosphorylation in controlling protein function. It is not clear what structural implications histone phosphorylation has, but histone phosphorylation has clear functions as a post-translational modification.
Histone modifications
The genomic DNA of eukaryotic cells is wrapped around special protein molecules known as histones. The complexes formed by the looping of the DNA are known as chromatin.
Post-translational modification of histones such as histone phosphorylation has been shown to modify the chromatin structure by chan |
https://en.wikipedia.org/wiki/VADS%20%28organisation%29 | VADS (formerly an initialism for Visual Arts Data Service) is a service of the Library at the University for the Creative Arts (UCA) in the UK that provides digital images and other visual arts resources free and copyright cleared for use in UK higher education and further education.
It has provided services to the academic community since 3 March 1997, and has built up a portfolio of visual art collections comprising over 140,000 images. |
https://en.wikipedia.org/wiki/Aggregate%20modulus | In relation to biomechanics, the aggregate modulus (Ha) is a measurement of the stiffness of a material at equilibrium when fluid has ceased flowing through it. The aggregate modulus can be calculated from Young's modulus (E) and the Poisson ratio (v).
The aggregate modulus of a similar specimen is determined from a unidirectional deformational testing configuration, i.e., the only non-zero strain component is E11. This configuration is opposed to the Young's modulus, which is determined from a unidirectional loading testing configuration, i.e., the only non-zero stress component is, say, in the e1 direction. In this test, the only non-zero component of the stress tensor is T11. |
https://en.wikipedia.org/wiki/Anti-twister%20mechanism | The anti-twister or antitwister mechanism is a method of connecting a flexible link between two objects, one of which is rotating with respect to the other, in a way that prevents the link from becoming twisted. The link could be an electrical cable or a flexible conduit.
This mechanism is intended as an alternative to the usual method of supplying electric power to a rotating device, the use of slip rings. The slip rings are attached to one part of the machine, and a set of fine metal brushes are attached to the other part. The brushes are kept in sliding contact with the slip rings, providing an electrical path between the two parts while allowing the parts to rotate about each other.
However, this presents problems with smaller devices. Whereas with large devices minor fluctuations in the power provided through the brush mechanism are inconsequential, in the case of tiny electronic components, the brushing introduces unacceptable levels of noise in the stream of power supplied. Therefore, a smoother means of power delivery is needed.
A device designed and patented in 1971 by Dale A. Adams and reported in The Amateur Scientist in December 1975, solves this problem with a rotating disk above a base from which a cable extends up, over, and onto the top of the disk. As the disk rotates the plane of this cable is rotated at exactly half the rate of the disk so the cable experiences no net twisting.
What makes the device possible is the peculiar connectivity of the space of 3D rotations, as discovered by P. A. M. Dirac and illustrated in his Plate trick (also known as the string trick or belt trick). Its covering Spin(3) group can be represented by unit quaternions, also known as versors.
See also
Quaternions and spatial rotation
Candle dance |
https://en.wikipedia.org/wiki/Comstock%E2%80%93Needham%20system | The Comstock–Needham system is a naming system for insect wing veins, devised by John Comstock and George Needham in 1898. It was an important step in showing the homology of all insect wings. This system was based on Needham's pretracheation theory that was later discredited by Frederic Charles Fraser in 1938.
Vein terminology
Longitudinal veins
The Comstock and Needham system attributes different names to the veins on an insect's wing. From the anterior (leading) edge of the wing towards the posterior (rear), the major longitudinal veins are named:
costa C, meaning rib
subcosta Sc, meaning below the rib
radius R, in analogy with a bone in the forearm, the radius
media M, meaning middle
cubitus Cu, meaning elbow
anal veins A, in reference to its posterior location
Apart from the costal and the anal veins, each vein can be branched, in which case the branches are numbered from anterior to posterior. For example, the two branches of the subcostal vein will be called Sc1 and Sc2.
The radius typically branches once near the base, producing anteriorly the R1 and posteriorly the radial sector Rs. The radial sector may fork twice.
The media may also fork twice, therefore having four branches reaching the wing margin.
According to the Comstock–Needham system, the cubitus forks once, producing the cubital veins Cu1 and Cu2.
According to some other authorities, Cu1 may fork again, producing the Cu1a and Cu1b.
As there are several anal veins, they are called A1, A2, and so on. They are usually unforked.
Crossveins
Crossveins link the longitudinal veins, and are named accordingly (for example, the medio-cubital crossvein is termed m-cu). Some crossveins have their own name, like the humeral crossvein h and the sectoral crossvein s.
Cell terminology
The cells are named after the vein on the anterior side; for instance, the cell between Sc2 and R1 is called Sc2.
In the case where two cells are separated by a crossvein but have the same anterior longitudinal vein, they |
https://en.wikipedia.org/wiki/Propanil | Propanil is a widely used contact herbicide. With an estimated use of about 8 million pounds in 2001, it is one of the more widely used herbicides in the United States. Propanil is said to be in use in approximately 400,000 acres of rice production each year.
Mode of action
The principal mode of propanil's herbicidal action against weeds is inhibition of their photosynthesis and CO2 fixation. Plants photosynthesize in two stages. In stage I photosynthetic reactions capture sunlight energy and yield molecules with high energy content. In stage II these molecules react to capture CO2, yielding carbohydrate precursors. In the stage I reaction a chlorophyll molecule absorbs one photon (light) and loses one electron, starting an electron transport chain reaction leading to the stage II reactions. Propanil inhibits the electron transport chain reaction and its conversion of CO2 to carbohydrate precursors. That inhibits further development of the weed.
Rice is relatively immune to propanil but most weeds are susceptible to it. The reason for the selectivity is that rice contains a high level of the enzyme aryl acylamidase (AAA), which rapidly metabolizes propanil to relatively nontoxic 3,4-dichloroaniline. Susceptible weeds lack the gene(s) coding for the AAA enzyme and thus succumb to propanil. However, intensive use of propanil and natural selection have caused some weeds to become resistant to propanil.
Synthesis
Propanil is made industrially by nitration of 1,2-dichlorobenzene (1) to give 1,2-dichloro-4-nitrobenzene (2), followed by hydrogenation of the nitro group with Raney nickel to give 3,4-dichloroaniline (3). Acylation of the amine with propanoyl chloride yields propanil (4). The resulting product is white or brown crystals.
Patent litigation
Propanil was the subject of several patent infringement suits. In one, Monsanto Co. v. Rohm and Haas Co., the United States Court of Appeals for the Third Circuit held that Monsanto committed fraud on the Patent Of |
https://en.wikipedia.org/wiki/Fiber%20functor | In category theory, a branch of mathematics, a fiber functor is a faithful k-linear tensor functor from a tensor category to the category of finite-dimensional k-vector spaces.
Definition
A fiber functor (or fibre functor) is a loose concept which has multiple definitions depending on the formalism considered. One of the main initial motivations for fiber functors comes from Topos theory. Recall a topos is the category of sheaves over a site. If a site is just a single object, as with a point, then the topos of the point is equivalent to the category of sets, . If we have the topos of sheaves on a topological space , denoted , then to give a point in is equivalent to defining adjoint functorsThe functor sends a sheaf on to its fiber over the point ; that is, its stalk.
From covering spaces
Consider the category of covering spaces over a topological space , denoted . Then, from a point there is a fiber functorsending a covering space to the fiber . This functor has automorphisms coming from since the fundamental group acts on covering spaces on a topological space . In particular, it acts on the set . In fact, the only automorphisms of come from .
With étale topologies
There is an algebraic analogue of covering spaces coming from the étale topology on a connected scheme . The underlying site consists of finite étale covers, which are finite flat surjective morphisms such that the fiber over every geometric point is the spectrum of a finite étale -algebra. For a fixed geometric point , consider the geometric fiber and let be the underlying set of -points. Then,is a fiber functor where is the topos from the finite étale topology on . In fact, it is a theorem of Grothendieck the automorphisms of form a profinite group, denoted , and induce a continuous group action on these finite fiber sets, giving an equivalence between covers and the finite sets with such actions.
From Tannakian categories
Another class of fiber functors come from cohomological |
https://en.wikipedia.org/wiki/Ars%20Conjectandi | (Latin for "The Art of Conjecturing") is a book on combinatorics and mathematical probability written by Jacob Bernoulli and published in 1713, eight years after his death, by his nephew, Niklaus Bernoulli. The seminal work consolidated, apart from many combinatorial topics, many central ideas in probability theory, such as the very first version of the law of large numbers: indeed, it is widely regarded as the founding work of that subject. It also addressed problems that today are classified in the twelvefold way and added to the subjects; consequently, it has been dubbed an important historical landmark in not only probability but all combinatorics by a plethora of mathematical historians. The importance of this early work had a large impact on both contemporary and later mathematicians; for example, Abraham de Moivre.
Bernoulli wrote the text between 1684 and 1689, including the work of mathematicians such as Christiaan Huygens, Gerolamo Cardano, Pierre de Fermat, and Blaise Pascal. He incorporated fundamental combinatorial topics such as his theory of permutations and combinations (the aforementioned problems from the twelvefold way) as well as those more distantly connected to the burgeoning subject: the derivation and properties of the eponymous Bernoulli numbers, for instance. Core topics from probability, such as expected value, were also a significant portion of this important work.
Background
In Europe, the subject of probability was first formally developed in the 16th century with the work of Gerolamo Cardano, whose interest in the branch of mathematics was largely due to his habit of gambling. He formalized what is now called the classical definition of probability: if an event has a possible outcomes and we select any b of those such that b ≤ a, the probability of any of the b occurring is . However, his actual influence on mathematical scene was not great; he wrote only one light tome on the subject in 1525 titled Liber de ludo aleae (Book on Game |
https://en.wikipedia.org/wiki/Federation%20of%20Analytical%20Chemistry%20and%20Spectroscopy%20Societies | The Federation of Analytical Chemistry and Spectroscopy Societies or FACSS is a scientific society incorporated on June 28, 1972, with the goal of promoting research and education in analytical chemistry. The organization combined the many smaller meetings of the individual societies into an annual meeting that includes all of analytical chemistry. The meetings are intended to provide a forum for scientists to address the development of analytical chemistry, chromatography, and spectroscopy.
The society's main activity is its annual conference held every fall. These conference offer plenary sessions, workshops, job fairs, oral presentations, poster presentations, and conference networking events. The conference was held internationally for the first time in 1999 when it was hosted in Vancouver, BC. The annual conference is often discussed in the journal Applied Spectroscopy, Spectroscopy Magazine, and American Pharmaceutical Reviews.
At the 2011 FACSS Conference in Reno, NV, the FACSS organization changed the name of the annual conference to SciX. The first SciX Conference presented by FACSS was held in Kansas City, MO in 2012. The name change was discussed in Spectroscopy in fall 2011:
. More information about the new name can be found at scixconference.org
Awards
FACSS presents several awards to both students and professionals. These awards honor scientists who have made significant contributions to the field of Analytical Chemistry.
FACSS Student Award and Tomas A. Hirshfeld Award [3]
SAS Student Poster Awards and FACSS Student Poster Awards [4]
FACSS Distinguished Service Award [5]
FACSS Innovation Award [6]
Charles Mann Award for Applied Raman Spectroscopy [7]
Anachem Award [8]
Lester W. Strock Award [9]
Applied Spectroscopy William F. Meggers Award [10]
Ellis R. Lippincott Award [11]
William G. Fateley Student Award
Coblentz Society Craver Award [12]
ACS Div of Analytical Chem Arthur F. Findeis Award for Achievements by a Young Analytical Scienti |
https://en.wikipedia.org/wiki/Orbital%20state%20vectors | In astrodynamics and celestial dynamics, the orbital state vectors (sometimes state vectors) of an orbit are
Cartesian vectors of position () and velocity () that together with their time (epoch) () uniquely determine the trajectory of the orbiting body in space.
Orbital state vectors come in many forms including the traditional Position-Velocity vectors, Two-line element set (TLE), and Vector Covariance Matrix (VCM).
Frame of reference
State vectors are defined with respect to some frame of reference, usually but not always an inertial reference frame. One of the more popular reference frames for the state vectors of bodies moving near Earth is the Earth-centered inertial (ECI) system defined as follows:
The origin is Earth's center of mass;
The Z axis is coincident with Earth's rotational axis, positive northward;
The X/Y plane coincides with Earth's equatorial plane, with the +X axis pointing toward the vernal equinox and the Y axis completing a right-handed set.
The ECI reference frame is not truly inertial because of the slow, 26,000 year precession of Earth's axis, so the reference frames defined by Earth's orientation at a standard astronomical epoch such as B1950 or J2000 are also commonly used.
Many other reference frames can be used to meet various application requirements, including those centered on the Sun or on other planets or moons, the one defined by the barycenter and total angular momentum of the solar system (in particular the ICRF), or even a spacecraft's own orbital plane and angular momentum.
Position and velocity vectors
The position vector describes the position of the body in the chosen frame of reference, while the velocity vector describes its velocity in the same frame at the same time. Together, these two vectors and the time at which they are valid uniquely describe the body's trajectory as detailed in Orbit determination. The principal reasoning is that Newton's law of gravitation yields an acceleration ; if the product of gr |
https://en.wikipedia.org/wiki/Stable%20isotope%20composition%20of%20amino%20acids | The stable isotope composition of amino acids refers to the abundance of heavy and light non-radioactive isotopes of carbon (13C and 12C), nitrogen (15N and 14N), and other elements within these molecules. Amino acids are the building blocks of proteins. They are synthesized from alpha-keto acid precursors that are in turn intermediates of several different pathways in central metabolism. Carbon skeletons from these diverse sources are further modified before transamination, the addition of an amino group that completes amino acid biosynthesis. Bonds to heavy isotopes are stronger than bonds to light isotopes, making reactions involving heavier isotopes proceed slightly slower in most cases. This phenomenon, known as a kinetic isotope effect, gives rise to isotopic differences between reactants and products that can be detected using isotope ratio mass spectrometry. Amino acids are synthesized via a variety of pathways with reactions containing different, unknown isotope effects. Because of this, the 13C content of amino acid carbon skeletons varies considerably between the amino acids. There is also an isotope effect associated with transamination, which is apparent from the abundance of 15N in some amino acids.
Because of these properties, amino acid isotopes record useful information about the organisms that produce them. Variations in metabolism between different taxonomical groups give rise to characteristic patterns of 13C enrichment in their amino acids. This allows the sources of carbon in food webs to be identified. The isotope effect associated with transamination also makes amino acid nitrogen isotopes a useful tool to study the structure of food webs. Repeated transamination by consumers results in a predictable increase in the abundance of 15N as amino acids are transferred up food chains. Together, these application, among others in ecology, demonstrate the utility of stable isotopes as tracers of environmental processes that are difficult to measure |
https://en.wikipedia.org/wiki/Word2vec | Word2vec is a technique for natural language processing (NLP) published in 2013. The word2vec algorithm uses a neural network model to learn word associations from a large corpus of text. Once trained, such a model can detect synonymous words or suggest additional words for a partial sentence. As the name implies, word2vec represents each distinct word with a particular list of numbers called a vector. The vectors are chosen carefully such that they capture the semantic and syntactic qualities of words; as such, a simple mathematical function (cosine similarity) can indicate the level of semantic similarity between the words represented by those vectors.
Approach
Word2vec is a group of related models that are used to produce word embeddings. These models are shallow, two-layer neural networks that are trained to reconstruct linguistic contexts of words. Word2vec takes as its input a large corpus of text and produces a vector space, typically of several hundred dimensions, with each unique word in the corpus being assigned a corresponding vector in the space.
Word2vec can utilize either of two model architectures to produce these distributed representations of words: continuously sliding bag-of-words (CBOW) or continuously sliding skip-gram. In both architectures, word2vec considers both individual words and a sliding context window as it iterates over the corpus.
The CBOW can be viewed as a ‘fill in the blank’ task, where the word embedding represents the way the word influences the relative probabilities of other words in the context window. Words which are semantically similar should influence these probabilities in similar ways, because semantically similar words should be used in similar contexts. The order of context words does not influence prediction (bag-of-words assumption).
In the continuous skip-gram architecture, the model uses the current word to predict the surrounding window of context words. The skip-gram architecture weighs nearby context word |
https://en.wikipedia.org/wiki/Tendinous%20arch%20of%20pelvic%20fascia | At the level of a line extending from the lower part of the pubic symphysis to the spine of the ischium is a thickened whitish band in this upper layer of the diaphragmatic part of the pelvic fascia. It is termed the tendinous arch or white line of the pelvic fascia, and marks the line of attachment of the special fascia (pars endopelvina fasciae pelvis) which is associated with the pelvic viscera.
It joins the fascia of the pubocervical fascia that covers the anterior wall of the vagina. If this fascia falls, the ipsilateral side of the vagina falls, carrying with it the bladder and the urethra, and thus contributing to urinary incontinence. |
https://en.wikipedia.org/wiki/REM%20rebound | REM rebound is the lengthening and increasing frequency and depth of rapid eye movement (REM) sleep which occurs after periods of sleep deprivation. When people have been prevented from experiencing REM, they take less time than usual to attain the REM state. When people are unable to obtain an adequate amount of REM sleep, the pressure to obtain REM sleep builds up. When the subject is able to sleep, they will spend a higher percentage of the night in REM sleep.
After early research connected rapid eye movement with dreaming and established that it made up about 20% of normal human sleep, experimenters started depriving test subjects of only REM sleep, to test its unique importance. Every time a subject's electroencephalogram and eye movements indicated the beginning of REM sleep, the experimenter would thoroughly wake them for several minutes. As this "dream deprivation" continued, tendency to initiate REM increased, and the subjects were woken up more and more times each night. The subjects became irritable, anxious, and hungry, and several left the study early. After five nights, the remaining subjects were allowed to sleep undisturbed, and showed a significant increase in percentage of sleep devoted to REM: from an average of 19.4% to an average of 26.6%. These effects were significant in comparison with a control group woken up on an equal number of occasions each night, at arbitrary times.
The fact that REM rebound exists shows that sleep and achievement of specific sleep stages are needed by the brain. In some marine mammals, such as dolphins and fur seals, when one brain hemisphere is deprived of REM sleep, only the deprived hemisphere will go into REM rebound. The other hemisphere will be unaffected.
REM rebound is common to those who take certain sleeping aids and it is also often seen in the first few nights after patients with sleep apnea are placed on CPAP. Alcohol can also affect REM sleep; it suppresses it during the first half of the night, lea |
https://en.wikipedia.org/wiki/Kelvin | The kelvin, symbol K, is a unit of measurement for temperature. The Kelvin scale is an absolute scale, which is defined such that 0 K is absolute zero and a change of thermodynamic temperature by 1 kelvin corresponds to a change of thermal energy by . The Boltzmann constant was exactly defined in the 2019 redefinition of the SI base units such that the triple point of water is . The kelvin is the base unit of temperature in the International System of Units (SI), used alongside its prefixed forms. It is named after the Belfast-born and University of Glasgow-based engineer and physicist William Thomson, 1st Baron Kelvin (1824–1907).
Historically, the Kelvin scale was developed from the Celsius scale, such that 273.15 K was 0 °C (the approximate melting point of ice) and a change of one kelvin was exactly equal to a change of one degree Celsius. This relationship remains accurate, but the Celsius, Fahrenheit, and Rankine scales are now defined in terms of the Kelvin scale. The kelvin is the primary unit of temperature for engineering and the physical sciences, while in most countries the Celsius scale remains the dominant scale outside of these fields. In the United States, outside of the physical sciences, the Fahrenheit scale predominates, with the kelvin or Rankine scale employed for absolute temperature.
History
Precursors
During the 18th century, multiple temperature scales were developed, notably Fahrenheit and centigrade (later Celsius). These scales predated much of the modern science of thermodynamics, including atomic theory and the kinetic theory of gases which underpin the concept of absolute zero. Instead, they chose defining points within the range of human experience that could be reproduced easily and with reasonable accuracy, but lacked any deep significance in thermal physics. In the case of the Celsius scale (and the long since defunct Newton scale and Réaumur scale) the melting point of water served as such a starting point, with Celsius be |
https://en.wikipedia.org/wiki/HElib | Homomorphic Encryption library or HElib is a free and open-source cross platform software developed by IBM that implements various forms of homomorphic encryption.
History
HElib was primarily developed by Shai Halevi and Victor Shoup, shortly after Craig Gentry was a researcher at IBM, with the initial release being on May 5, 2013.
Features
The library implements the Brakerski-Gentry-Vaikuntanathan (BGV) fully homomorphic encryption scheme, as well as optimizations such as Smart-Vercauteren ciphertext packing techniques.
HElib is written in C++ and uses the NTL mathematical library. |
https://en.wikipedia.org/wiki/Mammaglobin | Mammaglobin is a gene that encodes a 10-kilodalton glycoprotein. In humans expression of the gene is limited to the adult mammary gland, a correlation between increased expression of the gene and breast cancer has been reported. |
https://en.wikipedia.org/wiki/Topographic%20steering | In fluid mechanics, topographic steering is the effect of potential vorticity conservation on the motion of a fluid parcel. This means that the fluid parcels will not only react to physical obstacles in their path, but also to changes in topography or latitude. The two types of 'fluids' where topographic steering is mainly observed in daily life are air (air can be considered a compressible fluid in fluid mechanics) and water in respectively the atmosphere and the oceans. Examples of topographic steering can be found in, among other things, paths of low pressure systems and oceanic currents.
In 1869, Kelvin published his circulation theorem, which states that a barotropic, ideal fluid with conservative body forces conserves the circulation around a closed loop. To generalise this, Bjerknes published his own circulation theorem in 1898. Bjerknes extended the concept to inviscid, geostrophic and baroclinic fluids, resulting in addition of terms in the equation.
Mathematical description
Circulation
The exact mathematical description of the different potential vorticities can all be obtained from the circulation theorem of Bjerkness, which is stated as
.
Here is the circulation, the line integral of the velocity along a closed contour. Also, is the material derivative, is the density, is the pressure, is the angular velocity of the frame of reference and is the area projection of the closed contour onto the equatorial plane. This means the bigger the angle between the contour and the equatorial plane, the smaller this projection becomes.
The formula states that the change of the circulation along a fluid's path is affected by the variation of density in pressure coordinates and by the change in equatorial projection of the contour. Kelvin assumed both a barotropic fluid and a constant projection. Under these assumptions the right hand side of the equation is zero and Kelvin's theorem is found.
Shallow water
When considering a relatively thin layer of flui |
https://en.wikipedia.org/wiki/A-Train | is a series of business simulation video games developed and published by Japanese game developer Artdink in Japan. The first game in the series was published in 1985. The first release in the United States was Take the A-Train II, published in 1988 by the Seika Corporation under the title Railroad Empire. However, the most well known U.S. release is Take the A-Train III, published in 1992 by Maxis as simply A-Train. There is also the spin-off title C.E.O..
A-Train
The first iteration of the A-Train series was released in December 1985 for the FM-7, NEC PC-8801, NEC PC-9801, X1 Turbo, MZ-2500, Famicom, and MSX2. A Microsoft Windows port followed in April 2000.
A-Train II
The second iteration of the A-Train series was released in July 1988 for the NEC PC-9801 and X68000. In the US, it was released under the name Railroad Empire.
A-Train III
Take the A-Train III (known internationally as A-Train) is the third game in the A-Train series. It was originally developed and published by Japanese game developer Artdink for Japan, and was later published by Maxis for the United States. It was originally released in December 1990 for the NEC PC-9801, FM Towns Marty, Sharp X68000, and PC Engine (TurboGrafx-16). The US version was released in October 1992 on DOS and Amiga. Later, the Japanese version was re-released in March 2000 for Windows 95 and 98. Artdink ported the A-Train III along with the editor to Windows 95, and published both titles as a package as the 3rd ARTDINK BEST CHOICE title in Japan.
The game puts players in command of a railway company. There are no rival companies; the player controls the only one in the city and the game is resultingly fairly open-ended. A-Train III is the first game in the series to make use of near-isometric dimetric projection to present the city, similar to Maxis's later SimCity 2000. There are two types of transport that the player's company can take: passengers or building materials. The former is more likely to be profitable |
https://en.wikipedia.org/wiki/Voice%20chat%20in%20online%20gaming | Voice chat is telecommunication via voice over IP (VoIP) technologies—especially when those technologies are used as intercoms among players in multiplayer online games. The VoIP functionality can be built into some games, be a system-wide communication system, or a third-party chat software.
History
Voice chat in video games began in the sixth generation with the Sega Dreamcast (circa 1999). Some games, including Seaman and Alien Front Online included built in voice chat functionality, though it required an active subscription to the Dreamcast's online service, SegaNet.
In 2001, Sony released the Network adapter for their PlayStation 2 video game console, which allowed voice chatting with a headset. In 2002, Microsoft launched the Xbox Live service, including support for voice chat. Later, Microsoft required all Xbox Live console game developers to integrate voice chat capability into their games and bundled a microphone and headset with the Xbox Live retail unit. In 2005, Nintendo launched the Nintendo Wi-Fi Connection, an online multiplayer service for both the Nintendo DS and for the Wii. Metroid Prime Hunters, which was released in March 2006, was the first game that allowed voice chatting through the Nintendo DS's microphone. Nintendo also released a Nintendo DS headset for voice chat alongside the release of Pokémon Diamond and Pearl (2006).
2010s
Starting in the 2010s, third-party software have become very popular among gamers, even when in game VoIP services are available. Notable software includes Discord, Ventrilo, TeamSpeak, SONIX and Mumble. Support for Discord was added to the Xbox Series X|S and Xbox One consoles in 2022, with support coming to PlayStation 5 in 2023.
Impact
While voice chat has become a big hit in console games, also leads to problems such as griefing, cyberbullying, harassment, and scams.
See also
Audio headset
Comparison of VoIP software
Gamergate (harassment campaign)
Glossary of video game terms
Griefer
Massively m |
https://en.wikipedia.org/wiki/Murid%20betaherpesvirus%202 | Murid betaherpesvirus 2 (MuHV-2) is a species of virus in the genus Muromegalovirus, subfamily Betaherpesvirinae, family Herpesviridae, and order Herpesvirales. |
https://en.wikipedia.org/wiki/RegisterFly | RegisterFly was a New Jersey (U.S.) based internet hosting and domain name registrar that had their ICANN-accredited status terminated in March 2007.
History
RegisterFly formerly acted as a reseller of the services of eNom, but became an accredited registrar in its own right in 2006 through the acquisition of "Top Class Names" sold to the company by Directi with Bhavin Turakhia completing the asset transfer. By February 2007, the company was registrar for approximately 2,000,000 domain names held by about 900,000 customers. Notable clients of RegisterFly included the government of Thailand, the Easter Seals charity, and pop star Michael Jackson. In 2007, ICANN launched an investigation of RegisterFly amid allegations of fraud. No lawsuits were initially filed between the governing domain body and the company; however, there was during this time a lawsuit between the company's two owners, CEO John Naruszewicz and Kevin Medina. RegisterFly's website went offline for a time, causing serious concern amongst registrant customers of RegisterFly.
The incidents and lawsuit that then followed were the result of a feud between RegisterFly co-owners Kevin Medina and John Naruszewicz, who were at once both business and intimate partners. As problems at RegisterFly gained momentum, their ten-year-long business and romantic relationship abruptly came to an end. The lawsuit between the former partners alleged, among other matters, that Medina misappropriated corporate funds for personal use. However, the court ruled on March 8, 2007, in favor of Medina, stating that Naruszewicz had no ownership over RegisterFly. Medina resumed control over RegisterFly, but not before Naruszewicz published a public apology to customers on the company's web page.
On March 28, 2007, U.S. District Court Judge William Osteen unsealed a class action lawsuit filed by Attorney E. Clarke Dummit against RegisterFly, eNom, and ICANN. The lawsuit alleges that RegisterFly systematically defrauded custome |
https://en.wikipedia.org/wiki/Variant%20object | Variant objects in the context of HTTP are objects served by an Origin Content Server in a type of transmitted data variation (i.e. uncompressed, compressed, different languages, etc.).
HTTP/1.1 (1997–1999) introduces Content/Accept headers. These are used in HTTP requests and responses to state which variant the data is presented in.
Example Scenario
Client:
GET /encoded_data.html HTTP/1.1
Host: www.example.com
Accept-Encoding: gzip
Server:
HTTP/1.1 200 OK
Server: http-example-server
Content-Length: 23
Content-Encoding: gzip
<23 bytes of gzip compressed data>
See also
HTTP
HTTP compression
List of HTTP headers
Web cache |
https://en.wikipedia.org/wiki/Marcia%20Groszek | Marcia Jean Groszek is an American mathematician whose research concerns mathematical logic, set theory, forcing, and recursion theory. She is a professor of mathematics at Dartmouth College.
Education
As a high school student, Groszek felt isolated for her interest in mathematics,
but she found a sense of community through her participation in the Hampshire College Summer Mathematics Program, and she went on to earn her bachelor's degree at Hampshire College. She completed her Ph.D. in 1981 at Harvard University. Her dissertation, Iterated Perfect Set Forcing and Degrees of Constructibility, was supervised by Akihiro Kanamori.
Research
With Theodore Slaman, Groszek showed that (if they exist at all) non-constructible real numbers must be widespread, in the sense that every perfect set contains one of them, and they asked analogous questions of the non-computable real numbers. With Slaman, she has also shown that the existence of a maximally independent set of Turing degrees, of cardinality less than the cardinality of the continuum, is independent of ZFC.
In the theory of ordinal definable sets, an unordered pair of sets is said to be a Groszek–Laver pair if the pair is ordinal definable but neither of its two elements is; this concept is named for Groszek and Richard Laver, who observed the existence of such pairs in certain models of set theory.
Service and outreach
Groszek was program chair of the 2014 North American annual meeting of the Association for Symbolic Logic. Her interest in logic extends to education as well as to research; she has participated in the Association for Symbolic Logic Committee on Logic Education, and in 2011 she was co-organizer of an Association for Symbolic Logic special session on "Logic in the Undergraduate Mathematics Curriculum".
With mathematics colleague Dorothy Wallace and performance artist Josh Kornbluth, Groszek has also helped write and produce a sequence of educational videos about mathematics.
Selected publications |
https://en.wikipedia.org/wiki/Cluster%20state | In quantum information and quantum computing, a cluster state is a type of highly entangled state of multiple qubits. Cluster states are generated in lattices of qubits with Ising type interactions. A cluster C is a connected subset of a d-dimensional lattice, and a cluster state is a pure state of the qubits located on C. They are different from other types of entangled states such as GHZ states or W states in that it is more difficult to eliminate quantum entanglement (via projective measurements) in the case of cluster states. Another way of thinking of cluster states is as a particular instance of graph states, where the underlying graph is a connected subset of a d-dimensional lattice. Cluster states are especially useful in the context of the one-way quantum computer. For a comprehensible introduction to the topic see.
Formally, cluster states are states which obey the set eigenvalue equations:
where are the correlation operators
with and being Pauli matrices, denoting the neighbourhood of and being a set of binary parameters specifying the particular instance of a cluster state.
Examples with qubits
Here are some examples of one-dimensional cluster states (d=1), for , where is the number of qubits. We take for all , which means the cluster state is the unique simultaneous eigenstate that has corresponding eigenvalue 1 under all correlation operators. In each example the set of correlation operators and the corresponding cluster state is listed.
This is an EPR-pair (up to local transformations).
This is the GHZ-state (up to local transformations).
.
This is not a GHZ-state and can not be converted to a GHZ-state with local operations.
In all examples is the identity operator, and tensor products are omitted. The states above can be obtained from the all zero state by first applying a Hadamard gate to every qubit, and then a controlled-Z gate between all qubits that are adjacent to each other.
Experimental creation of clus |
https://en.wikipedia.org/wiki/Bullseye%21%20%281990%20film%29 | Bullseye! is a 1990 British–American action comedy film starring Michael Caine and Roger Moore. It was directed by Michael Winner. It was released on 2 November 1990, to mixed reviews, and was a box office disappointment. It has since developed a small cult following.
Plot
Moore and Caine play dual roles—a pair of small-time con-men and a pair of inept nuclear physicists who believe they have invented a limitless supply of energy. The con men use their resemblance to the scientists to con their way into the scientists' safe deposit boxes and steal the formula, but in so doing, they become entangled in a shady world of spies and international intrigue. The film includes a number of cameo appearances, including Jenny Seagrove (Winner's partner at the time) playing two different roles, John Cleese, Patsy Kensit, Alexandra Pigg and Nicholas Courtney. The film also features Roger Moore's daughter, Deborah Moore, in a supporting role.
Cast
Michael Caine – Sidney Lipton / Dr Daniel Hicklar
Roger Moore – Gerald Bradley-Smith / Sir John Bavistock
Sally Kirkland – Willie
Lee Patterson – Darrell Hyde
Deborah Moore – Flo Fleming (as Deborah Barrymore)
Mark Burns – Nigel Holden
Derren Nesbitt – Inspector Grosse
Deborah Leng – Francesca
Christopher Adamson – Death's Head
Steffanie Pitt – Donna Dutch
Angus MacKay – Reverend Simkin
Nicholas Courtney – Sir Hugh
Robert McBain – Lawyer
John Woodnutt – Bank Manager
Mildred Shay – Jolene (Tourist Wife)
Helen Horton – Tourist on Coach
Jeff Harding – Agent Merrow
Gordon Honeycombe – TV Announcer
Pamela Armstrong – TV Newsreader
Kiran Shah – Little Boss at Auction
John Cleese – Man on the Beach in Barbados Who Looks Like John Cleese
Jenny Seagrove – Health Club Receptionist / Girl with John Cleese
Patsy Kensit – Sick Lady on Train
Alexandra Pigg – Car Hire Girl
Deborah Bishop – Mr Moore's left eyebrow
Jim Bowen – Himself
Tony Green – Himself
Cliff Lazarenko - Celebrity Dart Thrower
John Scott Martin – Old Jeweller
John Lyons – Train Guar |
https://en.wikipedia.org/wiki/Blockbusting%20%28game%29 | Blockbusting is a solved combinatorial game introduced in 1987 by Elwyn Berlekamp illustrating a generalisation of overheating.
The analysis of Blockbusting may be used as the basis of a strategy for the combinatorial game of Domineering.
Blockbusting is a partisan game for two players known as Red and Blue (or Right and Left) played on an strip of squares called "parcels".
Each player, in turn, claims and colors one previously unclaimed parcel until all parcels have been claimed.
At the end, Left's score is the number of pairs of neighboring parcels both of which he has claimed.
Left therefore tries to maximize that number while Right tries to minimize it.
Adjacent Right-Right pairs do not affect the score.
Although the purpose of the game is to further the study of combinatorial game theory,
Berlekamp provides an interpretation alluding to the practice of blockbusting by real estate agents:
the players may be seen as rival agents buying up all the parcels on a street,
where Left is a segregationist trying to place his clients as neighbors of one another
while Right is an integrationist trying to break them up.
The operation of overheating introduced to analyze Blockbusting was later adapted by Berlekamp and David Wolfe
to warming to analyze the end-game of Go. |
https://en.wikipedia.org/wiki/Trans-Golgi%20network%20vesicle%20protein%2023%20A | Trans-Golgi network vesicle protein 23 A (TVP23A) is a protein coded for the TVP23A gene, formerly known as FAM18A. TVP23A is located on chromosome 16. It is known to have human paralogs, TVP23B and TVP23C, as well as orthologs in many different species, notably yeast, mice, and chickens. The general consensus on the TVP23A protein indicate that it has some function in the late Golgi apparatus and is involved in retrograde transport from endosomes back into the Golgi apparatus. The nature of this transport is still unknown.
Gene
Locus
TVP23A is located at cytogenic band 16p13.13, on the negative strand of Chromosome 16.
Alternative Names
TVP23A stands for Trans-Golgi network Vesicle Protein 23A TVP23A, is the current name for the protein. Aliases of TVP23A include FAM18A, and rarely YDR084C.
mRNA
Isoforms
There are two known isoforms of TVP23A, variant one and variant two, with variant one being the more common variant in humans.
Protein
Structure
TVP23A is a member of the pfam superfamily containing the domain of unknown function 846 (DUF846). TVP23A has a predicted molecular weight of 24.1 kilodaltons, an isoelectric point of 6.5, and relatively high amounts of tryptophan and phenylalanine. The secondary structure of TVP23A consists primarily of alpha helices composing 4 transmembrane domains. There is not much information on the tertiary structure of TVP23A or its homologs. iTASSER was used to generate a prediction for the folding pattern of TVP23A, which supports the presence of multiple helix structures.
Expression
TVP23A is ubiquitously expressed in all human tissues. There is evidence of higher expression in the brain tissue of mice. The promoter for TVP23A is GXP_91266, spanning 1403 base pairs located on the negative strand of chromosome 16.
Function
The hypothesized function of TVP23A is a transmembrane protein involved in retrograde transport of vesicles from early endosomes into the late Golgi apparatus. TVP23A interactions with SNARE |
https://en.wikipedia.org/wiki/Conversations%20on%20the%20Plurality%20of%20Worlds | Conversations on the Plurality of Worlds () is a popular science book by French author Bernard le Bovier de Fontenelle, published in 1686.
Content
The work consists of six lessons popularizing the knowledge of René Descartes and Nicolas Copernicus, given to a Marquise, spread over six evenings and preceded of a preface and a dispatch To Monsieur L*** .
First evening. That the Earth is a Planet which turns on itself, & around the Sun.
Second evening. That the Moon is an inhabited Earth.
Third night. Peculiarities of the Moon World. That the other Planets are also inhabited.
Fourth evening. Peculiarities of the Worlds of Venus, Mercury, Mars, Jupiter, & Saturn .
Fifth night. That the Fixed Stars are so many Suns, each of which illuminates a World.
Sixth evening. New thoughts that confirm those of previous Interviews. Latest discoveries that have been made in Heaven.
Analysis
Unlike many scientific works of its time, Conversations on the Plurality of Worlds was not written in Latin but in French, making it one of the first books to attempt an explanation of scientific theories in a popular language. A precursor to it could be seen in Giordano Bruno's 1584 book .
It is an early exposition of cosmic pluralism, the idea that the stars are distant suns which might have their own planetary systems, including the possibility of extraterrestrial life.
In the preface, Fontenelle suggests that the offered explanation should be easily understood even by those without scientific knowledge, and he specifically addresses female readers.
The book itself is presented as a series of conversations between a gallant philosopher and a marquise, who walk in the latter's garden at night and gaze at stars. The philosopher explains the heliocentric model and also muses on the possibility of extraterrestrial life.
It is the first work introducing the trope that sentient Venusians are gentle, ethereal, and beautiful.
Reception
The book was very well received both in France and |
https://en.wikipedia.org/wiki/Vehicle%20weight | Vehicle weight is a measurement of wheeled motor vehicles; either an actual measured weight of the vehicle under defined conditions or a gross weight rating for its weight carrying capacity.
Curb or kerb weight
Curb weight (American English) or kerb weight (English) is the total mass of a vehicle with standard equipment and all necessary operating consumables such as motor oil, transmission oil, brake fluid, coolant, air conditioning refrigerant, and sometimes a full tank of fuel, while not loaded with either passengers or cargo. The gross vehicle weight is larger and includes the maximum payload of passengers and cargo.
This definition may differ from definitions used by governmental regulatory agencies or other organizations. For example, many European Union manufacturers include the weight of a driver and luggage to follow European Directive 95/48/EC. Organizations may also define curb weight with fixed levels of fuel and other variables to equalize the value for the comparison of different vehicles.
The United States Environmental Protection Agency regulations define curb weight as follows: Curb weight means the actual or the manufacturer's estimated weight of the vehicle in operational status with all standard equipment, and weight of fuel at nominal tank capacity, and the weight of optional equipment computed in accordance with §86.1832–01; incomplete light-duty trucks shall have the curb weight specified by the manufacturer.
For a motorcycle, wet weight is the equivalent term.
Dry weight
Dry weight is the weight of a vehicle without any consumables, passengers, or cargo. It is significantly less than the weight of a vehicle in a drivable condition and therefore rarely used. Quoting a dry weight can make a car's weight and power-to-weight figures appear far more favorable than those of rival cars using curb weight.
The difference between dry weight and curb weight depends on many variables such as the capacity of the fuel tank. There is no standard |
https://en.wikipedia.org/wiki/Craterellus%20tubaeformis | Craterellus tubaeformis (formerly Cantharellus tubaeformis) is an edible fungus, also known as yellowfoot, winter mushroom, or funnel chanterelle. It is mycorrhizal, forming symbiotic associations with plants, making it very challenging to cultivate. It is smaller than the golden chanterelle (Cantharellus cibarius) and has a dark brown cap with paler gills and a hollow yellow stem. C. tubaeformis tastes stronger but less fruity than the golden chanterelle. It has a very distinctive smokey, peppery taste when raw. It grows in temperate and cold parts of Northern America and Europe, including Scandinavia, Finland, Russia, and the British Isles, as well as in the Himalayas in Asia, including Assam, in the central parts of the Indian subcontinent, and in Thailand.
C. tubaeformis is a yellowish-brown and trumpet-shaped mushroom found in great numbers late in the mushroom season, thus earning the common name winter mushroom. The cap is convex and sometimes hollow down the middle. The gills are widely separated, and of lighter color than the cap. It grows on moss or rotten wood, and in Northern America it is found mostly in conifer bogs. It is an excellent food mushroom, especially fried or in soups, and is easily dried for preservation.
Molecular phylogenetics has shown that C. tubaeformis deserves its reclassification from Cantharellus to Craterellus. Additionally, it appears that there are two distinct genetic populations that have traditionally been called tubaeformis: one in Europe and eastern North America, and another in western North America. If these two groups are defined as separate species, the "eastern" yellowfoot would retain the scientific epithet tubaeformis due to the origin of the type specimens in Sweden.
The western North American C. tubaeformis has been shown to make ectomycorrhizal relationships with western hemlock (Tsuga heterophylla) and Douglas-fir (Pseudotsuga menziesii). It is also most common in forests with a large amount of well-rotte |
https://en.wikipedia.org/wiki/Chlorophacinone | Chlorophacinone is an anticoagulant used as a rodenticide. It is classified as an extremely hazardous substance in the United States as defined in Section 302 of the U.S. Emergency Planning and Community Right-to-Know Act (42 U.S.C. 11002) and is subject to strict reporting requirements by facilities which produce, store, or use it in significant quantities.
See also
1,3-Indandione |
https://en.wikipedia.org/wiki/PTS-DOS | PTS-DOS (aka PTS/DOS) is a disk operating system, a DOS clone, developed in Russia by PhysTechSoft and Paragon Technology Systems.
History and versions
PhysTechSoft was formed in 1991 in Moscow, Russia by graduates and members of MIPT, informally known as PhysTech. At the end of 1993, PhysTechSoft released the first commercially available PTS-DOS as PTS-DOS v6.4. The version numbering followed MS-DOS version numbers, as Microsoft released MS-DOS 6.2 in November 1993.
In 1995, some programmers left PhysTechSoft and founded Paragon Technology Systems. They took source code with them and released their own version named PTS/DOS 6.51CD as well as S/DOS 1.0 ("Source DOS"), a stripped down open-source version. According to official PhysTechSoft announcements, these programmers violated both copyright laws and Russian military laws, as PTS-DOS was developed in close relationship with Russia's military and thus may be subject to military secrets law.
PhysTechSoft continued development on their own and released PTS-DOS v6.6 somewhere between and presented PTS-DOS v6.65 at the CeBIT exhibition in 1997. The next version from PhysTechSoft, formally PTS/DOS Extended Version 6.70 was labeled PTS-DOS 2000 and is still being distributed as a last 16-bit PTS-DOS system, .
Paragon continued their PTS-DOS line and released Paragon DOS Pro 2000 (also known and labeled in some places as PTS/DOS Pro 2000). According to Paragon, this was the last version and all development since then ceased. Moreover, this release contained bundled source code of older PTS-DOS v6.51.
Later, PhysTechSoft continued developing PTS-DOS and finally released PTS-DOS 32, formally known as PTS-DOS v7.0, which added support for the FAT32 file system.
PTS-DOS is certified by the Russian Ministry of Defense.
Commands
The following list of commands are supported by PTS-DOS 2000 Pro.
APPEND
ASK
ASSIGN
ATTR
BEEP
BREAK
CALL
CD
CHDIR
CHKDSK
CHOICE
CLS
COMMAND
COPY
CTTY
DATE
DEBUG
DEL
DIR
DI |
https://en.wikipedia.org/wiki/Tyndall%27s%20bar%20breaker | Tyndall's bar breaker is a physical demonstration experiment to demonstrate the forces created by thermal expansion and shrinkage. It was demonstrated 1867 by the Irish scientist John Tyndall in his Christmas lectures for a "juvenile auditory".
Setup
The bar breaker experiment comprises a very rigid frame (d) and a massive connecting rod (b). The rod is held on one side by a cast iron bar (c) that is going to be broken in the experiment and, at the other end, by a nut (a) that is used to compensate the thermal expansion.
Procedure
During the experiment the steel rod (b) is heated with a flame (e) up to red heat temperature. During the heating phase the thermal expansion of the rod (b) is compensated by tightly fastening the nut (a). Taking away the flame starts the cooling phase. Typically the bar (c) breaks within a few minutes with a loud bang or it is at least deformed significantly. |
https://en.wikipedia.org/wiki/Ampullary%20cupula | The ampullary cupula, or cupula, is a structure in the vestibular system, providing the sense of spatial orientation.
The cupula is located within the ampullae of each of the three semicircular canals. Part of the crista ampullaris, the cupula has embedded within it hair cells that have several stereocilia associated with each kinocilium. The cupula itself is the gelatinous component of the crista ampullaris that extends from the crista to the roof of the ampullae. When the head rotates, the endolymph filling the semicircular ducts initially lags behind due to inertia. As a result, the cupula is deflected opposite the direction of head movement. As the endolymph pushes the cupula, the stereocilia is bent as well, stimulating the hair cells within the crista ampullaris. After a short time of continual rotation however, the endolymph's acceleration normalizes with the rate of rotation of the semicircular ducts. As a result, the cupula returns to its resting position and the hair cells cease to be stimulated. This continues until the head stops rotating which simultaneously halts semicircular duct rotation. Due to inertia, however, the endolymph continues on. As the endolymph continues to move, the cupula is once again deflected resulting in the compensatory movements of the body when spun. In only the first situation, as fluid rushes by the cupula, the hair cells stimulated transmit the corresponding signal to the brain through the vestibulocochlear nerve (CN VIII). In the second one, there is no stimulation as the kinocilium can only be bent in one direction.
In their natural orientation within the head, the cupulae are located on the medial aspect of the semicircular canals. In this orientation, the kinocilia rest on the posterior aspect of the cupula.
Effects of alcohol
The Buoyancy Hypothesis posits that alcohol causes vertigo by affecting the neutral buoyancy of the cupula within the surrounding fluid called the endolymph. Linear accelerations (such as that |
https://en.wikipedia.org/wiki/Schweizerisches%20Volksliedarchiv | The Schweizerische Volksliedarchiv is an institution established at the University of Basel.
It was founded in 1906 as a department of the . Under the direction of John Meier, associations and the press were called upon to send in the song collection known in Switzerland so that it could be scientifically and critically reviewed and published in a suitable manner. The appeal was sent to all parts of Switzerland.
As a result of this first call, a collection of 31,000 German-language songs, 3,200 songs from Romandy, 1,400 from Ticino and 1,200 Rhaeto-Romanic has been secured.
The collection was subsequently expanded through bequests. For example, the holdings of Hanns In der Gand, the folk song collection of Arthur Rossat or the estate of Armin Breu with around 400 records and tapes with historical field recordings were added to the archive.
See also
Deutsches Volksliedarchiv
Music archives
Swiss folklore
Sound archives
Music organisations based in Switzerland
1906 establishments in Switzerland
Organisations based in Basel |
https://en.wikipedia.org/wiki/Qaboos%20bin%20Said | Qaboos bin Said Al Said (, ; 18 November 1940 – 10 January 2020) was Sultan of Oman from 23 July 1970 until his death in 2020. A fifteenth-generation descendant of the founder of the House of Al Said, he was the longest-serving leader in the Middle East and Arab world at the time of his death, having ruled for almost half a century.
The only son of Said bin Taimur, Sultan of Muscat and Oman, Qaboos was educated in Suffolk, England. After graduating from the Royal Military Academy Sandhurst, he served briefly in the British Army. He returned to Oman in 1966 and was the subject of considerable restrictions from his father. In 1970, Qaboos ascended to the Omani throne after overthrowing his father in a coup d'état, with British support. The country was subsequently renamed the Sultanate of Oman.
As sultan, Qaboos implemented a policy of modernization and ended Oman's international isolation. His reign saw a rise in living standards and development in the country, the abolition of slavery, the end of the Dhofar Rebellion, and the promulgation of Oman's constitution. Suffering from poor health in later life, Qaboos died in 2020. He had no children, so he entailed the royal court to reach consensus on a successor upon his death. As a precaution he hid a letter which named his successor in case an agreement was not achieved. After his death the royal court decided to view Qaboos's letter and named his intended successor, his cousin Haitham bin Tariq, as sultan.
Early life and education
Sayyid Qaboos bin Said was born in the southern city of Salalah in Dhofar on 18 November 1940 as an only son of Sultan Said bin Taimur and Mazoon al-Mashani. He received his primary and secondary education at Salalah, and was sent to a private educational establishment at Bury St Edmunds in England at age 16. At 20, he entered the Royal Military Academy Sandhurst. After graduating from Sandhurst in September 1962, he joined the British Army and was posted to the 1st Battalion The Cameron |
https://en.wikipedia.org/wiki/71st%20meridian%20east | The meridian 71° east of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Asia, the Indian Ocean, the Southern Ocean, and Antarctica to the South Pole.
The 71st meridian east forms a great circle with the 109th meridian west.
From Pole to Pole
Starting at the North Pole and heading south to the South Pole, the 71st meridian east passes through:
{| class="wikitable plainrowheaders"
! scope="col" width="120" | Co-ordinates
! scope="col" | Country, territory or sea
! scope="col" | Notes
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Arctic Ocean
| style="background:#b0e0e6;" |
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Kara Sea
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Bely Island
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Malygina Strait
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Yamal Peninsula
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Gulf of Ob
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|Passing just east of Kokand
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|Khyber Pakhtunkhwa - for about 9 km
|-
|
! scope="row" |
|
|-valign="top"
|
! scope="row" |
|Khyber Pakhtunkhwa Punjab
|-
|
! scope="row" |
| Rajasthan
|-
|
! scope="row" |
| Sindh
|-
|
! scope="row" |
| Gujarat
|-
|
! scope="row" |
| Sindh - for about 10 km
|-valign="top"
|
! scope="row" |
| Gujarat Dadra and Nagar Haveli and Daman and Diu - for about 1 km
|-valign="top"
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Indian Ocean
| style="background:#b0e0e6;" | Passing just west of Danger Island, Passing just east of Grande Terre,
|-
| style="background:#b0e0e6;" |
https://en.wikipedia.org/wiki/Faron%20Moller | Faron George Moller (born February 25, 1962 in Trail, British Columbia) is a Canadian-born British computer scientist and expert on theoretical computer science, particularly infinite-state automata theory and temporal logic. His work has focussed on structural decomposition techniques for analysing abstract models of computing systems. He is founding Director of the Swansea Railway Verification Group; Director of Technocamps; and Head of the Institute of Coding in Wales. In 2023, he as elected General Secretary of the Learned Society of Wales.
Biography
Moller studied mathematics and computer science as an undergraduate at the University of British Columbia, and then as a Masters student at the University of Waterloo, before going on to do a PhD supervised by Robin Milner in the Laboratory for Foundations of Computer Science at the University of Edinburgh. He has held posts at the universities of Strathclyde and Edinburgh, The Swedish Institute for Computer Science, The Royal Institute of Technology in Stockholm, and Uppsala University before moving to Wales as Professor of Computer Science at Swansea University in 2000.
Appointments and honours
Moller is a Fellow of the Learned Society of Wales, a Fellow of the British Computer Society and Fellow of the Institute of Mathematics and its Applications, and served as President of the British Colloquium for Theoretical Computer Science for 15 years (2004-2019). He is a Chartered Mathematician, a Chartered Scientist, and a Chartered IT Professional. His full nomenclature with post-nominal letters is Professor Faron Moller BSc, MMath, PhD, CITP, CMath, CSci, FLSW, FBCS, FIMA.
He is also Director of Technocamps, a pan-Wales schools outreach programme aimed at introducing and reinforcing Computer Science and Digital Competency within all Welsh schools and inspiring young people to study computing-based topics; and Head of the Institute of Coding in Wales.
See also
List of University of Waterloo people |
https://en.wikipedia.org/wiki/Morinosaurus | Morinosaurus (meaning "Morini lizard", for an ancient people of northern France) was a genus of sauropod dinosaur from an unnamed formation of Kimmeridgian-age Upper Jurassic rocks from Boulogne-sur-Mer, Département du Pas-de-Calais, France. It is an obscure tooth genus sometimes referred to the Lower Cretaceous English wastebasket taxon Pelorosaurus.
History and taxonomy
The French paleontologist H. E. Sauvage based this genus on a single worn tooth, apparently now lost, which he compared to those of Hypselosaurus. Oddly, despite illustrations of the tooth, and the implications of comparing it to a titanosaur with narrow-crowned teeth, it was included as a synonym of Pelorosaurus in two major reviews. Pelorosaurus, being a putative brachiosaurid, is assumed to have had broad-crowned teeth.
Age, however, was not an issue, because it was referred to the possible Pelorosaurus species P. manseli (="Ischyrosaurus"), which was also from the Upper Jurassic (the question of whether "Ischyrosaurus" or any Jurassic species should be included in Pelorosaurus at all is another issue). The most recent review considers it to be a nomen dubium without further comment.
Sauvage also suggested that a partial right humerus belonged to the type individual.
Paleobiology
Morinosaurus would have been a large, quadrupedal herbivore. Having titanosaur-like teeth may suggest more titanosaurian- or diplodocoid-like feeding habits, but this is speculative. The tooth crown was 50 mm (1.97 im) tall and had a cross-section of . |
https://en.wikipedia.org/wiki/Searle%27s%20bar%20method | Searle's bar method (named after George Frederick Charles Searle) is an experimental procedure to measure thermal conductivity of material. A bar of material is being heated by steam on one side and the other side cooled down by water while the length of the bar is thermally insulated. Then the heat ΔQ propagating through the bar in a time interval of Δt is given by
where
ΔQ is the heat supplied to the bar in time Δt
k is the coefficient of thermal conductivity of the bar.
A is the cross-sectional area of the bar,
ΔTbar is the temperature difference between each end of the bar
L is the length of the bar
and the heat ΔQ absorbed by water in a time interval of Δt is:
where
Cw is the specific heat of water,
Δm is the mass of water collected during time Δt,
ΔTwater is difference in the temperature of water before and after it has gone through the bar.
Assuming perfect insulation and no energy loss, then
which leads to |
https://en.wikipedia.org/wiki/List%20of%20NHL%20mascots | This is a list of current and former National Hockey League (NHL) mascots, sorted alphabetically by team. As of 2023, the New York Rangers are the only team to not have a mascot.
Current mascots
Al the Octopus
Al the Octopus is the octopus mascot of the Detroit Red Wings. It is also the only mascot that is not costumed. In 1952, when east side fish merchants Pete and Jerry Cusimano threw a real octopus onto the Olympia arena ice, the eight legs represented the eight victories needed to secure a Stanley Cup in those six-team days. Since then, fans throw an octopus onto the ice for good luck. In the 1995 Playoffs, fans threw fifty-four onto the ice. Arena Manager and Zamboni driver Al Sobotka ceremoniously scoops them up and whirls them over his head, and play continues. NHL Commissioner Gary Bettman forbade Sobotka from doing so during the 2008 playoffs, claiming that debris flew off the octopuses and onto the ice. Sobotka and the Red Wings have denied that this occurs, but even so Sobotka acquiesced and now twirls the octopuses once he departs the ice. In 2011, the NHL forbade fans from throwing any octopuses on the ice, penalizing all violators with a $500 fine. This has led to local outcry at the seemingly intentional destruction of a classic tradition. Red Wings' forward Johan Franzen has pledged to pay any and all fines as an attempt to continue the tradition.
Two identical large purple prop octopuses (Al), named after ice manager Al Sobotka, used to be positioned in or on top of Joe Louis Arena for the duration of the playoffs. After closing down the arena after the 2016-2017 season, one was sold for $7,700.
Bailey
Bailey, the mascot of the Los Angeles Kings, is a 6-foot lion (6 foot 4 inches with mane included) who wears No. 72 because it is the average temperature in Los Angeles. He debuted during the 2007-2008 season and was named in honor of Garnet Bailey, who served as the Kings' Director of Pro Scouting from 1994 until his death in the September 11 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.