source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Common%20Intermediate%20Language | Common Intermediate Language (CIL), formerly called Microsoft Intermediate Language (MSIL) or Intermediate Language (IL), is the intermediate language binary instruction set defined within the Common Language Infrastructure (CLI) specification. CIL instructions are executed by a CLI-compatible runtime environment such as the Common Language Runtime. Languages which target the CLI compile to CIL. CIL is object-oriented, stack-based bytecode. Runtimes typically just-in-time compile CIL instructions into native code.
CIL was originally known as Microsoft Intermediate Language (MSIL) during the beta releases of the .NET languages. Due to standardization of C# and the CLI, the bytecode is now officially known as CIL. Windows Defender virus definitions continue to refer to binaries compiled with it as MSIL.
General information
During compilation of CLI programming languages, the source code is translated into CIL code rather than into platform- or processor-specific object code. CIL is a CPU- and platform-independent instruction set that can be executed in any environment supporting the Common Language Infrastructure, such as the .NET runtime on Windows, or the cross-platform Mono runtime. In theory, this eliminates the need to distribute different executable files for different platforms and CPU types. CIL code is verified for safety during runtime, providing better security and reliability than natively compiled executable files.
The execution process looks like this:
Source code is converted to CIL bytecode and a CLI assembly is created.
Upon execution of a CIL assembly, its code is passed through the runtime's JIT compiler to generate native code. Ahead-of-time compilation may also be used, which eliminates this step, but at the cost of executable-file portability.
The computer's processor executes the native code.
Instructions
CIL bytecode has instructions for the following groups of tasks:
Load and store
Arithmetic
Type conversion
Object creation and manipulat |
https://en.wikipedia.org/wiki/Topoisomer | Topoisomers or topological isomers are molecules with the same chemical formula and stereochemical bond connectivities but different topologies. Examples of molecules for which there exist topoisomers include DNA, which can form knots, and catenanes. Each topoisomer of a given DNA molecule possesses a different linking number associated with it. DNA topoisomers can be interchanged by enzymes called topoisomerases. Using a topoisomerase along with an intercalator, topoisomers with different linking number may be separated on an agarose gel via gel electrophoresis.
See also
Mechanically-interlocked molecular architectures
Catenane
Rotaxanes
Molecular knot
Molecular Borromean rings |
https://en.wikipedia.org/wiki/PMOS%20logic | PMOS or pMOS logic (from p-channel metal–oxide–semiconductor) is a family of digital circuits based on p-channel, enhancement mode metal–oxide–semiconductor field-effect transistors (MOSFETs). In the late 1960s and early 1970s, PMOS logic was the dominant semiconductor technology for large-scale integrated circuits before being superseded by NMOS and CMOS devices.
History and application
Mohamed Atalla and Dawon Kahng manufactured the first working MOSFET at Bell Labs in 1959. They fabricated both PMOS and NMOS devices but only the PMOS devices were working. It would be more than a decade before contaminants in the manufacturing process (particularly sodium) could be managed well enough to manufacture practical NMOS devices.
Compared to the bipolar junction transistor, the only other device available at the time for use in an integrated circuit, the MOSFET offers a number of advantages:
Given semiconductor device fabrication processes of similar precision, a MOSFET requires only 10% of the area of a bipolar junction transistor. The main reason is that the MOSFET is self-insulating and does not require p–n junction isolation from neighboring components on the chip.
A MOSFET requires fewer process steps and is therefore simpler and cheaper to manufacture (one diffusion doping step compared to four for a bipolar process).
Since there is no static gate current for a MOSFET, the power consumption of an integrated circuit based on MOSFETs can be lower.
Disadvantages relative to bipolar integrated circuits were:
The switching speed was considerably lower, due to large gate capacitances.
The high threshold voltage of early MOSFETs led to a higher minimum power-supply voltage (-24 V to -28 V).
General Microelectronics introduced the first commercial PMOS circuit in 1964, a 20-bit shift register with 120 MOSFETs – at the time an incredible level of integration. The attempt by General Microelectronics in 1965 to develop a set of 23 custom integrated circuits for an elect |
https://en.wikipedia.org/wiki/Turtle%20Taxonomy%20Working%20Group | The Turtle Taxonomy Working Group (TTWG) is an informal working group of the IUCN/SSC Tortoise and Freshwater Turtle Specialist Group (TFTSG). It is composed of a number of leading turtle taxonomists, with varying participation by individual participants over the years, some dropping out and others joining.
Works
The TTWG has produced an annual checklist of living and recently extinct turtles since 2007, deliberates on proposed changes to turtle taxonomy, and describes its consideration whether to accept, reject, or suspend adoption of proposed changes in a series of annotations to the checklist. Recent versions of the checklist have included full primary synonymies and citations to all original descriptions of recent turtle taxa, as well as CITES and IUCN Red List status of each species as applicable. |
https://en.wikipedia.org/wiki/Ping-pong%20scheme | Algorithms said to employ a Ping-Pong scheme exist in different fields of software engineering. They are characterized by an alternation between two entities. In the examples described below, these entities are communication partners, network paths or file blocks.
Databases
In most database management systems durable database transactions are supported through a log file. However, multiple writes to the same page of that file can produce a slim chance of data loss. Assuming for simplicity that the log file is organized in pages whose size matches the block size of its underlying medium, the following problem can occur:
If the very last page of the log file is only partially filled with data and has to be written to permanent storage in this state, the very same page will have to be overwritten during the next write operation. If a crash happens during that later write operation, previously stored log data may be lost.
The Ping-Pong scheme described in Transaction Processing eliminates this problem by alternately writing the contents of said (logical) last page to two different physical pages inside the log file (the actual last page i and its empty successor i+1). Once said logical log page is no longer the last page (i.e. it is completely filled with log data), it is written one last time to the regular physical position (i) inside the log file.
This scheme requires the usage of time stamps for each page in order to distinguish the most recent version of the logical last page one from its predecessor.
Networking
Internet
A functionality which lets a computer A find out whether a computer B is reachable and responding is built into the Internet Control Message Protocol (ICMP). Through an "Echo Request" Computer A asks B to send back an "Echo Reply". These two messages are also sometimes erroneously called "ping" and "pong".
Routing
In Routing, a Ping-Pong scheme is a simple algorithm for distributing data packets across
two paths. If you had two paths A and B |
https://en.wikipedia.org/wiki/Fortuna%20%28PRNG%29 | Fortuna is a cryptographically secure pseudorandom number generator (PRNG) devised by Bruce Schneier and Niels Ferguson and published in 2003. It is named after Fortuna, the Roman goddess of chance. FreeBSD uses Fortuna for /dev/random and /dev/urandom is symbolically linked to it since FreeBSD 11. Apple OSes have switched to Fortuna since 2020 Q1.
Design
Fortuna is a family of secure PRNGs; its design
leaves some choices open to implementors. It is composed of the following pieces:
The generator itself, which once seeded will produce an indefinite quantity of pseudo-random data.
The entropy accumulator, which collects genuinely random data from various sources and uses it to reseed the generator when enough new randomness has arrived.
The seed file, which stores enough state to enable the computer to start generating random numbers as soon as it has booted.
Generator
The generator is based on any good block cipher. Practical Cryptography suggests AES, Serpent or Twofish. The basic idea is to run the cipher in counter mode, encrypting successive values of an incrementing counter.
With a 128-bit block cipher, this would produce statistically identifiable deviations from randomness; for instance, generating 264 genuinely random 128-bit blocks would produce on average about one pair of identical blocks, but there are no repeated blocks at all among the first 2128 produced by a 128-bit cipher in counter mode. Therefore, the key is changed periodically: no more than 1 MiB of data (216 128-bit blocks) is generated without a key change. The book points out that block ciphers with a 256-bit (or greater) block size, which did not enjoy much popularity at the time, do not have this statistical problem.
The key is also changed after every data request (however small), so that a future key compromise doesn't endanger previous generator outputs. This property is sometimes described as "Fast Key Erasure" or Forward secrecy.
Entropy accumulator
The entropy accumulator |
https://en.wikipedia.org/wiki/Viremia | Viremia is a medical condition where viruses enter the bloodstream and hence have access to the rest of the body. It is similar to bacteremia, a condition where bacteria enter the bloodstream. The name comes from combining the word "virus" with the Greek word for "blood" (haima). It usually lasts for 4 to 5 days in the primary condition.
Primary versus secondary
Primary viremia refers to the initial spread of virus in the blood from the first site of infection.
Secondary viremia occurs when primary viremia has resulted in infection of additional tissues via bloodstream, in which the virus has replicated and once more entered the circulation.
Usually secondary viremia results in higher viral shedding and viral loads within the bloodstream due to the possibility that the virus is able to reach its natural host cell from the bloodstream and replicate more efficiently than the initial site. An excellent example to profile this distinction is the rabies virus. Usually the virus will replicate briefly within the first site of infection, within the muscle tissues. Viral replication then leads to viremia and the virus spreads to its secondary site of infection, the central nervous system (CNS). Upon infection of the CNS, secondary viremia results and symptoms usually begin. Vaccination at this point is useless, as the spread to the brain is unstoppable. Vaccination must be done before secondary viremia takes place for the individual to avoid brain damage or death.
Active versus passive
Active viremia is caused by the replication of viruses which results in viruses being introduced into the bloodstream. Examples include the measles, in which primary viremia occurs in the epithelial lining of the respiratory tract before replicating and budding out of the cell basal layer (viral shedding), resulting in viruses budding into capillaries and blood vessels.
Passive viremia is the introduction of viruses in the bloodstream without the need of active viral replication. Exampl |
https://en.wikipedia.org/wiki/Cable%20theory | Classical cable theory uses mathematical models to calculate the electric current (and accompanying voltage) along passive neurites, particularly the dendrites that receive synaptic inputs at different sites and times. Estimates are made by modeling dendrites and axons as cylinders composed of segments with capacitances and resistances combined in parallel (see Fig. 1). The capacitance of a neuronal fiber comes about because electrostatic forces are acting through the very thin lipid bilayer (see Figure 2). The resistance in series along the fiber is due to the axoplasm's significant resistance to movement of electric charge.
History
Cable theory in computational neuroscience has roots leading back to the 1850s, when Professor William Thomson (later known as Lord Kelvin) began developing mathematical models of signal decay in submarine (underwater) telegraphic cables. The models resembled the partial differential equations used by Fourier to describe heat conduction in a wire.
The 1870s saw the first attempts by Hermann to model neuronal electrotonic potentials also by focusing on analogies with heat conduction. However, it was Hoorweg who first discovered the analogies with Kelvin's undersea cables in 1898 and then Hermann and Cremer who independently developed the cable theory for neuronal fibers in the early 20th century. Further mathematical theories of nerve fiber conduction based on cable theory were developed by Cole and Hodgkin (1920s–1930s), Offner et al. (1940), and Rushton (1951).
Experimental evidence for the importance of cable theory in modelling the behavior of axons began surfacing in the 1930s from work done by Cole, Curtis, Hodgkin, Sir Bernard Katz, Rushton, Tasaki and others. Two key papers from this era are those of Davis and Lorente de Nó (1947) and Hodgkin and Rushton (1946).
The 1950s saw improvements in techniques for measuring the electric activity of individual neurons. Thus cable theory became important for analyzing data collect |
https://en.wikipedia.org/wiki/Property%20P%20conjecture | In mathematics, the Property P conjecture is a statement about 3-manifolds obtained by Dehn surgery on a knot in the 3-sphere. A knot in the 3-sphere is said to have Property P if every 3-manifold obtained by performing (non-trivial) Dehn surgery on the knot is not simply-connected. The conjecture states that all knots, except the unknot, have Property P.
Research on Property P was started by R. H. Bing, who popularized the name and conjecture.
This conjecture can be thought of as a first step to resolving the Poincaré conjecture, since the Lickorish–Wallace theorem says any closed, orientable 3-manifold results from Dehn surgery on a link.
If a knot has Property P, then one cannot construct a counterexample to the Poincaré conjecture by surgery along .
A proof was announced in 2004, as the combined result of efforts of mathematicians working in several different fields.
Algebraic Formulation
Let denote elements corresponding to a preferred longitude and meridian of a tubular neighborhood of .
has Property P if and only if its Knot group is never trivialised by adjoining a relation of the form for some .
See also
Property R conjecture |
https://en.wikipedia.org/wiki/Rhs%20toxins | Rhs toxins belong to the polymorphic toxin category of bacterial exotoxins. Rhs proteins are widespread and can be produced by both Gram-negative and Gram-positive bacteria. Rhs toxins are very large proteins of usually more than 1,500 aminoacids with variable C-terminal toxic domains. Their toxic activity can either target eukaryotes or other bacteria.
Domain architecture
In their large N-terminal region, Rhs toxins comprise RHS/YD repeats in various number (PF05593) (RHS meaning Rearrangement Hot Spot) and another "RHS-repeats associated core" domain (PF03527). In contrast, their C-terminal regions are shorter and harbor highly variable C-terminal domains including many domains with a predicted nuclease activity.
Function
Anti-eukaryotic activity
These toxins encompass Rhs toxins of insect pathogens with an activity against insects. This group also include Rhs toxins with an activity against human phagocytic cells that contribute to pathogenesis of Pseudomonas aeruginosa.
Anti-bacterial activity
A role in inter-bacterial competition has been demonstrated for the plant pathogen Dickeya dadantii and for the human pathogen Escherichia coli.
When a polymorphic toxin with anti-bacterial activity is produced by a bacterial strain, this strain is protected by a specific immunity protein encoded by a gene immediately downstream of the toxin gene.
Delivery
Some Rhs toxins such as the previously mentioned system in Dickeya dadantii appear to be dependent on the type VI secretion system for delivery into neighbouring cells. PAAR domain toxins such as Rhs appear to form the sharp tip of the type VI secretion system being attached to the VgrG of the secretion apparatus. The C-terminal toxins of Rhs may vary to diversify the antimicrobial activity of the type VI secretion system. |
https://en.wikipedia.org/wiki/Long%20double | In C and related programming languages, long double refers to a floating-point data type that is often more precise than double precision though the language standard only requires it to be at least as precise as double. As with C's other floating-point types, it may not necessarily map to an IEEE format.
long double in C
History
The long double type was present in the original 1989 C standard, but support was improved by the 1999 revision of the C standard, or C99, which extended the standard library to include functions operating on long double such as sinl() and strtold().
Long double constants are floating-point constants suffixed with "L" or "l" (lower-case L), e.g., 0.3333333333333333333333333333333333L or 3.1415926535897932384626433832795029L for quadruple precision. Without a suffix, the evaluation depends on FLT_EVAL_METHOD.
Implementations
On the x86 architecture, most C compilers implement long double as the 80-bit extended precision type supported by x86 hardware (generally stored as 12 or 16 bytes to maintain data structure alignment), as specified in the C99 / C11 standards (IEC 60559 floating-point arithmetic (Annex F)). An exception is Microsoft Visual C++ for x86, which makes long double a synonym for double. The Intel C++ compiler on Microsoft Windows supports extended precision, but requires the /Qlong‑double switch for long double to correspond to the hardware's extended precision format.
Compilers may also use long double for the IEEE 754 quadruple-precision binary floating-point format (binary128). This is the case on HP-UX, Solaris/SPARC, MIPS with the 64-bit or n32 ABI, 64-bit ARM (AArch64) (on operating systems using the standard AAPCS calling conventions, such as Linux), and z/OS with FLOAT(IEEE). Most implementations are in software, but some processors have hardware support.
On some PowerPC systems, long double is implemented as a double-double arithmetic, where a long double value is regarded as the exact sum of two double-precis |
https://en.wikipedia.org/wiki/Kloosterman%20sum | In mathematics, a Kloosterman sum is a particular kind of exponential sum. They are named for the Dutch mathematician Hendrik Kloosterman, who introduced them in 1926 when he adapted the Hardy–Littlewood circle method to tackle a problem involving positive definite diagonal quadratic forms in four as opposed to five or more variables, which he had dealt with in his dissertation in 1924.
Let be natural numbers. Then
Here x* is the inverse of modulo .
Context
The Kloosterman sums are a finite ring analogue of Bessel functions. They occur (for example) in the Fourier expansion of modular forms.
There are applications to mean values involving the Riemann zeta function, primes in short intervals, primes in arithmetic progressions, the spectral theory of automorphic functions and related topics.
Properties of the Kloosterman sums
If or then the Kloosterman sum reduces to the Ramanujan sum.
depends only on the residue class of and modulo . Furthermore and if .
Let with and coprime. Choose and such that and . Then
This reduces the evaluation of Kloosterman sums to the case where for a prime number and an integer .
The value of is always an algebraic real number. In fact is an element of the subfield which is the compositum of the fields
where ranges over all odd primes such that and
for with .
The Selberg identity:
was stated by Atle Selberg and first proved by Kuznetsov using the spectral theory of modular forms. Nowadays elementary proofs of this identity are known.
For an odd prime, there are no known simple formula for , and the Sato–Tate conjecture suggests that none exist. The lifting formulas below, however, are often as good as an explicit evaluation. If one also has the important transformation:
where denotes the Jacobi symbol.
Let with prime and assume . Then:
where is chosen so that and is defined as follows (note that is odd):
This formula was first found by Hans Salie and there are many simple proofs in the |
https://en.wikipedia.org/wiki/Film%20plane | A film plane is the surface of an image recording device such as a camera, upon which the lens creates the focused image. In cameras from different manufacturers, the film plane varies in distance from the lens. Thus each lens used has to be chosen carefully to assure that the image is focused on the exact place where the individual frame of film or digital sensor is positioned during exposure. It is sometimes marked on a camera body with the 'Φ' symbol where the vertical bar represents the exact location.
Movie cameras often also have small focus hooks where the focus puller can attach one side of a tape measure to quickly gauge the distance to objects that he intends to bring into focus. The measurement is taken from the film plane to the subject.
Due to Petzval field curvature, the film plane upon which a lens focuses may not be a literal plane. Cameras may bend the film stock or even plate stock slightly to compensate, improving the area of critical focus and sharpness. Nevertheless, the general concept of a focal plane is understood to refer to this position in the camera sensor relative to the lens.
See also
Cardinal point (optics)
Flange focal distance
Photography equipment
Planes (geometry)
Basically Film Panel is an area inside any cameras or image taking device that has a lens,film and a digital sensor. |
https://en.wikipedia.org/wiki/CHRNA9 | Neuronal acetylcholine receptor subunit alpha-9, also known as nAChRα9, is a protein that in humans is encoded by the CHRNA9 gene. The protein encoded by this gene is a subunit of certain nicotinic acetylcholine receptors (nAchR).
α9 subunit-containing receptors are notably blocked by nicotine. The role of this antagonism in the effects of tobacco are unknown.
This gene is a member of the ligand-gated ionic channel family and nicotinic acetylcholine receptor gene superfamily. It encodes a plasma membrane protein that forms homo- or hetero-oligomeric divalent cation channels. This protein is involved in cochlea hair cell function and is expressed in both the inner and outer hair cells (OHCs) of the adult cochlea, although expression levels in adult inner hair cells is low. The activation of the alpha9/10 nAChR is via olivocochlear activity, represented by cholinergic efferent synaptic terminals originating from the superior olive region of the brainstem. The protein is additionally expressed in keratinocytes, the pituitary gland, B-cells and T-cells.
Selective block of α9α10 nicotinic acetylcholine receptors by the conotoxin RgIA has been shown to be analgesic in an animal model of nerve injury pain.
See also
Nicotinic acetylcholine receptor |
https://en.wikipedia.org/wiki/Heterothallism | Heterothallic species have sexes that reside in different individuals. The term is applied particularly to distinguish heterothallic fungi, which require two compatible partners to produce sexual spores, from homothallic ones, which are capable of sexual reproduction from a single organism.
In heterothallic fungi, two different individuals contribute nuclei to form a zygote. Examples of heterothallism are included for Saccharomyces cerevisiae, Aspergillus fumigatus, Aspergillus flavus, Penicillium marneffei and Neurospora crassa. The heterothallic life cycle of N. crassa is given in some detail, since similar life cycles are present in other heterothallic fungi.
Life cycle of Saccharomyces cerevisiae
The yeast Saccharomyces cerevisiae is heterothallic. This means that each yeast cell is of a certain mating type and can only mate with a cell of the other mating type. During vegetative growth that ordinarily occurs when nutrients are abundant, S. cerevisiae reproduces by mitosis as either haploid or diploid cells. However, when starved, diploid cells undergo meiosis to form haploid spores. Mating occurs when haploid cells of opposite mating type, MATa and MATα, come into contact. Ruderfer et al. pointed out that such contacts are frequent between closely related yeast cells for two reasons. The first is that cells of opposite mating type are present together in the same ascus, the sac that contains the tetrad of cells directly produced by a single meiosis, and these cells can mate with each other. The second reason is that haploid cells of one mating type, upon cell division, often produce cells of the opposite mating type with which they may mate.
Katz Ezov et al. presented evidence that in natural S. cerevisiae populations clonal reproduction and a type of “self-fertilization” (in the form of intratetrad mating) predominate. Ruderfer et al. analyzed the ancestry of natural S. cerevisiae strains and concluded that outcrossing occurs only about once every |
https://en.wikipedia.org/wiki/Central%20Board%20of%20Film%20Certification | The Central Board of Film Certification (CBFC) is a statutory film-certification body in the Ministry of Information and Broadcasting of the Government of India. It is tasked with "regulating the public exhibition of films under the provisions of the Cinematograph Act 1952." The Cinematograph Act 1952 outlines a strict certification process for commercial films shown in public venues. Films screened in cinemas and on television may only be publicly exhibited in India after certification by the board and edited.
Certificates and guidelines
The board currently issues four certificates. Originally, there were two: U (unrestricted public exhibition with family-friendly movies) and A (restricted to adult audiences but any kind of nudity not allowed). Two more were added in June U/A (unrestricted public exhibition, with parental guidance for children under 12) and S (restricted to specialised audiences, such as doctors or scientists). The board may refuse to certify a film. Additionally, V/U, V/UA, V/A are used for video releases with U, U/A and A carrying the same meaning as above.
U certificate
Films with the U certification are fit for unrestricted public exhibition and are family-friendly. These films can contain universal themes like education, family, drama, romance, sci-fi, action etc. These films can also contain some mild violence, but it cannot be prolonged. It may also contain very mild sexual scenes (without any traces of nudity or sexual detail).
U/A certificate
Films with the U/A certification can contain moderate adult themes that are not strong in nature and are not considered appropriate to be watched by a child below 12 years of age with parental guidance. These films may contain moderate to strong violence, moderate sexual scenes (traces of nudity and moderate sexual detail can be found), frightening scenes, blood flow, or muted abusive language. Sometimes such films are re-certified with V/U for video viewing.
A certificate
Films with the A ce |
https://en.wikipedia.org/wiki/Roguing | In agriculture, roguing refers to the act of identifying and removing plants with undesirable characteristics from agricultural fields. Rogues are removed from the fields to preserve the quality of the crop being grown. Plants being removed may be diseased, be of an unwanted variety, or undesirable for other reasons. For example, to ensure that the crop retains its integrity as regards certain physical attributes, such as color and shape, individual plants that exhibit differing traits are removed. Roguing is particularly important when growing seed crops, to prevent plants with undesirable characteristics from propagating into subsequent generations. |
https://en.wikipedia.org/wiki/Logical%20NOR | In Boolean logic, logical NOR or non-disjunction or joint denial is a truth-functional operator which produces a result that is the negation of logical or. That is, a sentence of the form (p NOR q) is true precisely when neither p nor q is true—i.e. when both of p and q are false. It is logically equivalent to and , where the symbol signifies logical negation, signifies OR, and signifies AND.
Non-disjunction is usually denoted as or or (prefix) or .
As with its dual, the NAND operator (also known as the Sheffer stroke—symbolized as either , or ), NOR can be used by itself, without any other logical operator, to constitute a logical formal system (making NOR functionally complete).
The computer used in the spacecraft that first carried humans to the moon, the Apollo Guidance Computer, was constructed entirely using NOR gates with three inputs.
Definition
The NOR operation is a logical operation on two logical values, typically the values of two propositions, that produces a value of true if and only if both operands are false. In other words, it produces a value of false if and only if at least one operand is true.
Truth table
The truth table of is as follows:
Logical equivalences
The logical NOR is the negation of the disjunction:
Alternative notations and names
Peirce is the first to show the functional completeness of non-disjunction while he doesn't publish his result. Peirce used for non-conjunction and for non-disjunction (in fact, what Peirce himself used is and he didn't introduce while Peirce's editors made such disambiguated use). Peirce called as (from Ancient Greek , , "cutting both ways").
In 1911, was the first to publish a description of both non-conjunction (using , the Stamm hook), and non-disjunction (using , the Stamm star), and showed their functional completeness.</ref> Note that most uses in logical notation of use this for negation.
In 1913, Sheffer described non-disjunction and showed its functional completeness. S |
https://en.wikipedia.org/wiki/Electrotropism | In biology, electrotropism, also known as galvanotropism, is a kind of tropism which results in growth or migration of an organism, usually a cell, in response to an exogenous electric field. Several types of cells such as nerve cells, muscle cells, fibroblasts, epithelial cells, green algae, spores, and pollen tubes, among others, have been already reported to respond by either growing or migrating in a preferential direction when exposed to an electric field.
Electrotropism in Pollen Tubes
Electrotropism is known to play a role in the control of growth in cells and the development of tissues. By imposing an exogenous electric field, or modifying an endogenous one, a cell or a group of cells can greatly redirect their growth. Pollen tubes, for instance, align their polar growth with respect to an exogenous electric field. It has been observed that cells respond to electric fields as small as 0.1 mV/cell diameter (Note that the average radius of a large cell is in the order of a few micrometers). Electric fields have also been shown to act as directional signals in the repair and regeneration of wounded tissue.
The pollen tube is an excellent model for the understanding of electrotropism and plant cell behavior in general. They are easily cultivated in vitro and have a very dynamic cytoskeleton that polymerizes at very high rates, providing the pollen tube with interesting growth properties. For instance, the pollen tube has an unusual kind of growth; it extends exclusively at its apex. Pollen tubes, as most biological systems, are influenced by electrical stimulus.
Introduction to Electrotropism Experiment in Pollen Tubes
Electrical fields have been shown to influence a gamut of cellular processes and responses. Animals, plants, and bacteria have a range of responses to electrical structures. The electrophysiology in humans consists of the nervous system regulating our actions and behaviors through controlled responses. Action potentials in our nerves and our h |
https://en.wikipedia.org/wiki/Thermomonosporaceae | Thermomonosporaceae represents a Family of bacteria that share similar genotypic and phenotypic characteristics. The Family Thermomonosporaceae includes aerobic, Gram-positive, non-acid-fast, chemo-organotrophic Actinomycetota. They produce a branched substrate mycelium bearing aerial hyphae that undergo differentiation into single or short chains of arthrospores. All species of Thermomonosporaceae share the same cell wall type (type III; meso-diaminopimelic acid), a similar menaquinone profile in which MK-9(H6)is predominant, and fatty acid profile type 3a. The presence of the diagnostic sugar madurose is variable, but can be found in most species of this family. The polar lipid profiles are characterized as phospholipid type PI for most species of Thermomonospora, Actinomadura and Spirillospora. The members of Actinocorallia are characterized by phospholipid type PII.
The G+C content of the DNA lies within the range 66±72 mol%. The pattern of 16S rRNA signatures consists of nucleotides at positions 440 : 497 (C–G), 501 : 544 (C–G), 502 : 543 (G–C), 831 : 855 (G–G), 843 (U), 844 (A) and 1355 : 1367 (A–U).
The type genus is Thermomonospora |
https://en.wikipedia.org/wiki/Natural%20Standard | Natural Standard is an international research collaboration that systematically reviews scientific evidence on complementary and alternative medicine. Together with the faculty of Harvard Medical School, Natural Standard provides consumer information on complementary and alternative medicine for Harvard Health Publications and Susan G. Komen for the Cure. Natural Standard also provides information on herbal medicine and dietary supplements to MedlinePlus, which is produced and maintained by the U.S. National Library of Medicine (NLM) and the National Institutes of Health (NIH).
History
Natural Standard Research Collaboration was founded in 2000 to serve as a clearing house for information on evidence-based medicine covering numerous healthcare disciplines. This international effort involves authors, editors, and peer reviewers from multiple academic and research institutions. It used an A thru F grading system.
Natural Standard was taken over by Therapeutic Research Center in 2013.
Natural Standard thru the new Therapeutic Research Center is a subscription-only database that delivers information about complementary and alternative medicine and dietary supplements and produces the Natural Medicines Comprehensive Database (NMCD). A new combined database will be called "Natural Medicines" (http://info.therapeuticresearch.com/natural-medicine-comprehensive-database). Natural Standard provides information that is broad in scope, and focused on both medications and disease states. A standout feature of this database is its graded, evidence-based evaluation of alternative therapies.
Personnel
The founders were:
Catherine Ulbricht, PharmD, MBA[c]
Ethan Basch, MD, MSc, MPhil
Members of the Senior Editorial Board include:
Edzard Ernst, a well-known figure in complementary and alternative medicine research and co-author of Trick or Treatment
See also
Examine.com
ConsumerLab.com |
https://en.wikipedia.org/wiki/Multi-attribute%20global%20inference%20of%20quality | Multi-attribute global inference of quality (MAGIQ) is a multi-criteria decision analysis technique. MAGIQ is based on a hierarchical decomposition of comparison attributes and rating assignment using rank order centroids.
Description
The MAGIQ technique is used to assign a single, overall measure of quality to each member of a set of systems where each system has an arbitrary number of comparison attributes. The MAGIQ technique has features similar to the analytic hierarchy process and the simple multi-attribute rating technique exploiting ranks (SMARTER) technique. The MAGIQ technique was first published by James D. McCaffrey. The MAGIQ process begins with an evaluator determining which system attributes are to be used as the basis for system comparison. These attributes are ranked by importance to the particular problem domain, and the ranks are converted to ratings using rank order centroids. Each system under analysis is ranked against each comparison attribute and the ranks are transformed into rank order centroids. The final overall quality metric for each system is the weighted (by comparison attribute importance) sum of each attribute rating. The references provide specific examples of the process. There is little direct research on the theoretical soundness and effectiveness of the MAGIQ technique as a whole, however the use of hierarchical decomposition and the use of rank order centroids in multi-criteria decision analyses have been studied, with generally positive results. Anecdotal evidence suggests that the MAGIQ technique is both practical and useful.
See also
Multi-attribute utility
Multi-attribute auction |
https://en.wikipedia.org/wiki/Crucifixion%20%28Corpus%20Hypercubus%29 | Crucifixion (Corpus Hypercubus) is a 1954 oil-on-canvas painting by Salvador Dalí. A nontraditional, surrealist portrayal of the Crucifixion, it depicts Christ on a polyhedron net of a tesseract (hypercube). It is one of his best-known paintings from the later period of his career.
Background
During the 1940s and 1950s Dalí's interest in traditional surrealism diminished and he became fascinated with nuclear science, feeling that "thenceforth, the atom was [his] favorite food for thought". The atomic bombing at the end of World War II left a lasting impression; his 1951 essay "Mystical Manifesto" introduced an art theory he called "nuclear mysticism" that combined his interests in Catholicism, mathematics, science, and Catalan culture in an effort to reestablish classical values and techniques, which he extensively utilized in Corpus Hypercubus.
That same year, to promote nuclear mysticism and explain the "return to spiritual classicism movement" in modern art, he traveled throughout the United States giving lectures. Before painting Corpus Hypercubus, Dalí announced his intention to portray an exploding Christ using both classical painting techniques along with the motif of the cube, and he declared that "this painting will be the great metaphysical work of [his] summer". Juan de Herrera's Treatise on Cubic Forms was particularly influential to Dalí.
Composition and meaning
Corpus Hypercubus is painted in oil on canvas, and its dimensions are 194.3 cm × 123.8 cm (76.5 in × 48.75 in). Consistent with his theory of nuclear mysticism, Dalí uses classical elements along with ideas inspired by mathematics and science. Some noticeably classic features are the drapery of the clothing and the Caravaggesque lighting that theatrically envelops Christ, though like his 1951 painting Christ of Saint John of the Cross, Corpus Hypercubus takes the traditional biblical scene of Christ's Crucifixion and almost completely reinvents it. The union of Christ and the tesseract refle |
https://en.wikipedia.org/wiki/Center%20for%20Veterinary%20Medicine | The Center for Veterinary Medicine (CVM) is a branch of the U.S. Food and Drug Administration (FDA) that regulates the manufacture and distribution of food, food additives, and drugs that will be given to animals. These include animals from which human foods are derived, as well as food additives and drugs for pets or companion animals. CVM is responsible for regulating drugs, devices, and food additives given to, or used on, over one hundred million companion animals, plus millions of poultry, cattle, swine, and minor animal species. Minor animal species include animals other than cattle, swine, chickens, turkeys, horses, dogs, and cats.
CVM monitors the safety of animal foods and medications. Much of the center's work focuses on animal medications used in food animals to ensure that significant drug residues are not present in the meat or other products from these animals.
CVM does not regulate vaccines for animals; these are handled by the United States Department of Agriculture
History
In 1953, a Veterinary Medical Branch of the FDA was created within the Bureau of Medicine. A separate Bureau of Veterinary Medicine (BVM) was established in 1965. At this time, the BVM included a Division of Veterinary Medical Review, Division of Veterinary New Drugs, and a Division of Veterinary Research. In 1970, the Division of Compliance and Division of Nutritional Sciences were added. The Bureau underwent reorganization in 1976 and in 1984, it was renamed the Center for Veterinary Medicine.
Dr. Steven Solomon, DVM became the Director of the Center in 2017. He received his Doctor of Veterinary Medicine (DVM) degree from The Ohio State University and a Master's in Public Health from Johns Hopkins University. He succeeded Tracey Forfa, who had been acting director for a few months. The previous director was Dr. Bernadette Dunham; she served as Director from 2008 to 2016.
Mission and vision
The mission of the center is "protecting human and animal health" and the vision |
https://en.wikipedia.org/wiki/Charopinesta%20sema | Charopinesta sema, also known as the Blackburn Island pinhead snail, is a species of land snail that is endemic to Australia's Lord Howe Island group in the Tasman Sea.
Description
The depressedly turbinate to discoidal shell of the mature snail is 1.1 mm in height, with a diameter of 1.8 mm, and a low, stepped spire. It is pale golden in colour. The whorls are rounded, with deeply impressed sutures and moderately spaced radial ribs. It has a roundedly lunate aperture and widely open umbilicus.
Distribution and habitat
This extremely rare snail is known from a single empty shell from Blackburn Island. It may be extinct. |
https://en.wikipedia.org/wiki/Lion%20and%20Tusk | The Lion and Tusk was the main logo of the British South Africa Company (BSAC) and later as a state symbol of Rhodesia. The logo was used following the Company being set up during the scramble for Africa and was used as they governed Rhodesia. Following the company relinquishing control of Northern and Southern Rhodesia, the symbol fell out of favour with the Rhodesian public. However, following the Rhodesian republic being declared in 1970, the Lion and Tusk symbol was adopted as a state symbol to replace the British Empire's Royal crown until the establishment of Zimbabwe in 1980.
Company history
The Lion and Tusk was adopted as a corporate insignia of the British South Africa Company following it receiving a Royal Charter allowing it to fly its own flag. The Lion and Tusk featured in the centre of the flag of the British South Africa Company. The Lion and Tusk was referred to in heraldic terms as "a lion passant guardant Or, supporting with its right forepaw an ivory tusk proper", though it was also irreverently referred to as "the lion with the tooth-pick". A 1902 Admiralty warrant later entitled the BSAC to use the Lion and Tusk on Blue and Red Ensigns.
During the early days of Company rule in Rhodesia, it was used as the badges of the British South Africa Police and the Rhodesia Horse cavalry unit. Following the 1922 Southern Rhodesian government referendum where Southern Rhodesia voted for responsible government independent of the BSAC, the Lion and Tusk fell into disuse in the Rhodesias, especially in Southern Rhodesia where their people wanted new symbols independent of BSAC influence. It continued to be used as the BSAC's corporate logo until 1965 when the company merged with the Anglo-American Corporation.
Rhodesia
Following Rhodesia's Unilateral Declaration of Independence in 1965, their insignia remained the same due to loyalty to Queen Elizabeth II. But the Lion and Tusk started to regain popularity as a separate Rhodesian symbol with it being use |
https://en.wikipedia.org/wiki/Quantum%20mind | The quantum mind or quantum consciousness is a group of hypotheses proposing that classical mechanics alone cannot explain consciousness, positing instead that quantum-mechanical phenomena, such as entanglement and superposition, may play an important part in the brain's function and could explain critical aspects of consciousness. These scientific hypotheses are as yet untested, and can overlap with quantum mysticism.
History
Eugene Wigner developed the idea that quantum mechanics has something to do with the workings of the mind. He proposed that the wave function collapses due to its interaction with consciousness. Freeman Dyson argued that "mind, as manifested by the capacity to make choices, is to some extent inherent in every electron".
Other contemporary physicists and philosophers considered these arguments unconvincing. Victor Stenger characterized quantum consciousness as a "myth" having "no scientific basis" that "should take its place along with gods, unicorns and dragons".
David Chalmers argues against quantum consciousness. He instead discusses how quantum mechanics may relate to dualistic consciousness. Chalmers is skeptical that any new physics can resolve the hard problem of consciousness. He argues that quantum theories of consciousness suffer from the same weakness as more conventional theories. Just as he argues that there is no particular reason why particular macroscopic physical features in the brain should give rise to consciousness, he also thinks that there is no particular reason why a particular quantum feature, such as the EM field in the brain, should give rise to consciousness either.
Approaches
Bohm
David Bohm viewed quantum theory and relativity as contradictory, which implied a more fundamental level in the universe. He claimed that both quantum theory and relativity pointed to this deeper theory, which he formulated as a quantum field theory. This more fundamental level was proposed to represent an undivided wholeness and a |
https://en.wikipedia.org/wiki/Cloud%20%28operating%20system%29 | Cloud was a browser-based operating system created by Good OS LLC, a Los Angelesbased corporation. The company initially launched a Linux distribution called gOS which is heavily based on Ubuntu, now in its third incarnation.
Overview
The Cloud was a simplified operating system that ran just a web browser, providing access to a variety of web-based applications that allowed the user to perform many simple tasks without booting a full-scale operating system. Because of its simplicity, Cloud could boot in just a few seconds. The operating system is designed for Netbooks, Mobile Internet Devices, and PCs that are mainly used to browse the Internet. From Cloud the user can quickly boot into the main OS, because Cloud continues booting the main OS in the background.
Combining a browser with a basic operating system allows the use of cloud computing, in which applications and data "live and run" on the Internet instead of the hard drive.
Cloud can be installed and used together with other operating systems, or act as a standalone operating system. When used as a standalone operating system, hardware requirements are relatively low.
In 2009, Cloud was only officially available built into the GIGABYTE M912 Touch Screen Netbook.
Early reviews compared the operating system's user interface to OS X and noted the similarity of its browser to Google Chrome, although it is actually based on a modified Mozilla Firefox browser
See also
ChromeOS
Mozilla Firefox
EasyPeasy
Joli OS
EyeOS |
https://en.wikipedia.org/wiki/Lucio%20Lombardo-Radice | Lucio Lombardo-Radice (Catania, 10 July 1916; Brussels, 21 November 1982) was an Italian mathematician. A student of Gaetano Scorza, Lombardo-Radice contributed to finite geometry and geometric combinatorics together with Guido Zappa and Beniamino Segre, and wrote important works concerning the Non-Desarguesian plane. He was also a leading member of the Italian Communist Party and a member of its central committee.
Lombardo-Radice's parents were Giuseppe Lombardo Radice and Gemma Harasim. His children included the writer and actor Giovanni Lombardo Radice.
The Istituto Tecnico Statale Commerciale "Lucio Lombardo Radice" per Programmatori, a school in Rome, Italy, founded in 1982 as the XXV Istituto Tecnico Commerciale per Programmatori, was in 1992 renamed after Lombardo-Radice. It is now named Istituto di Istruzione Superiore Lombardo Radice |
https://en.wikipedia.org/wiki/2009%20ConAgra%20Foods%20plant%20explosion | On June 9, 2009, a natural gas explosion occurred at the ConAgra Foods plant in Garner, North Carolina, United States.
Background
After World War II, the Jesse Jones Sausage Company constructed a factory in Wake County, North Carolina. General Mills acquired the facility in 1968 and soon thereafter began producing Slim Jims there. The plant came under the control of ConAgra Foods in 1998. By June 2009, the factory was the only one in the world which produced Slim Jims and employed 900 people, making it the largest employer in the town of Garner.
Explosion
On June 9, 2009 a contractor from Energy Systems Analysts (ESA) was installing an industrial water heater at the ConAgra plant. The ESA worker attempted to purge the air out of a three-inch pipe meant to supply natural gas to the heater, using natural gas. As per ESA's typical practices, the worker vented the purged gases into the utility room, which was ventilated by a fan. Workers struggled to light the heater, so they continued to purge the line over a two-and-a-half-hour period. Plant personnel were aware of the indoor purging, and were using their sense of smell to detect odors to gauge possible gas buildups. Personnel who entered the utility room and some in the packing room noticed the smell of gas, but were not concerned by it. The gas concentration ultimately surpassed its explosive limit and was ignited at about 11:25 am.
The explosion damaged of the facility, concentrated in the southern part of the plant where packing operations were conducted, and triggered a leak from the plant's refrigeration system, causing the release of about of ammonia. Emergency responders calculated the ammonia levels at 20 parts per million outside the plant. It also set small fires around the facility grounds. Over 200 workers were inside the plant at the time of the explosion. One member of ConAgra's safety team ran back into the plant to rescue a co-worker and was killed by falling debris.
Plant workers were brou |
https://en.wikipedia.org/wiki/Timeline%20of%20numerals%20and%20arithmetic | A timeline of numerals and arithmetic.
Before 2000 BC
c. 20,000 BC — Nile Valley, Ishango Bone: suggested, though disputed, as the earliest reference to prime numbers as also a common number.
c. 3400 BC — the Sumerians invent the first so-known numeral system, and a system of weights and measures.
c. 3100 BC — Egypt, earliest known decimal system allows indefinite counting by way of introducing new symbols, .
c. 2800 BC — Indus Valley civilization on the Indian subcontinent, earliest use of decimal ratios in a uniform system of ancient weights and measures, the smallest unit of measurement used is 1.704 millimetres and the smallest unit of mass used is 28 grams.
c. 2000 BC — Mesopotamia, the Babylonians use a base-60 decimal system, and compute the first known approximate value of π at 3.125.
1st millennium BC
c. 1000 BC — Vulgar fractions used by the Egyptians.
second half of 1st millennium BC — The Lo Shu Square, the unique normal magic square of order three, was discovered in China.
c. 400 BC — Jaina mathematicians in India write the “Surya Prajinapti”, a mathematical text which classifies all numbers into three sets: enumerable, innumerable and infinite. It also recognises five different types of infinity: infinite in one and two directions, infinite in area, infinite everywhere, and infinite perpetually.
c. 300 BC — Brahmi numerals are conceived in India.
300 BC — Mesopotamia, the Babylonians invent the earliest calculator, the abacus.
c. 300 BC — Indian mathematician Pingala writes the “Chhandah-shastra”, which contains the first Indian use of zero as a digit (indicated by a dot) and also presents a description of a binary numeral system, along with the first use of Fibonacci numbers and Pascal's triangle.
c. 250 BC — late Olmecs had already begun to use a true zero (a shell glyph) several centuries before Ptolemy in the New World. See 0 (number).
150 BC — Jain mathematicians in India write the “Sthananga Sutra”, which contains work on the theo |
https://en.wikipedia.org/wiki/Leibniz%20Institute%20for%20Solar%20Physics | The Leibniz Institute for Solar Physics (aka: KIS; ), formerly known as Kiepenheuer Institute for Solar Physics (KIS) is a research institute located in Freiburg, Germany. As a member of the Leibniz Association, the institute conducts basic research in astronomy and astrophysics with a particular focus on solar physics. The institute's structure and operation is based on three strategic pillars: 1) fundamental research, 2) operation of the German solar telescope infrastructure on Tenerife, and 3) applied research in data science and operation of the Science Data Center. Institute's Professors appointed and habilitated at the University of Freiburg offer lectures at various university degree levels and train young scientists.
History
The institute was founded in 1943 as the 'Fraunhofer Institute' by Karl-Otto Kiepenheuer. Kiepenheuer was director of the institute from 1943 until his death in 1975. The institute was renamed as the 'Kiepenheuer Institute for Solar Physics' to honour the founder of the institute and to enable the Fraunhofer Society to call their own institutes (the first of which was founded in 1954), 'Fraunhofer Institutes'. Both institutions had been named independently after the physicist Joseph von Fraunhofer, and they had no other connection besides the name. In November 2018, name of the institute was renamed to Leibniz Institute for Solar Physics (KIS – Leibniz-Institut für Sonnenphysik (KIS) – to highlight the institute's membership in the Leibniz Association.
Organisation
The Leibniz Institute for Solar Physics (KIS) is a foundation under public law of the state of Baden-Württemberg and a member of the Leibniz Association. The organs of the Leibniz Institute for Solar Physics (KIS) are the Foundation Council, the Scientific Advisory Council and the Board of Directors. The institute is divided into two scientific departments: "Solar and Stellar Astrophysics" and "Observatory and Instrumentation." A third administrative-technical department c |
https://en.wikipedia.org/wiki/Calonarius%20splendens | Calonarius splendens is a species of fungus in the family Cortinariaceae. It is commonly known as the splendid webcap. The species is native to Europe where it has been implicated in poisonings resulting in kidney failure, though with milder symptoms than other deadly webcaps.
Taxonomy
The species was described in 1939 by the mycologist Robert Henry who classified it as Cortinarius splendens.
In 2022 the species was transferred from Cortinarius and reclassified as Calonarius splendens based on genomic data.
Habitat and distribution
It has been classed as conspecific with Cortinarius meinhardii, although the two species have different tree hosts—the former with the European beech (Fagus sylvatica) and the latter with Norway spruce (Picea abies).
See also
List of Cortinarius species |
https://en.wikipedia.org/wiki/Non-numerical%20words%20for%20quantities | The English language has a number of words that denote specific or approximate quantities that are themselves not numbers. Along with numerals, and special-purpose words like some, any, much, more, every, and all, they are Quantifiers. Quantifiers are a kind of determiner and occur in many constructions with other determiners, like articles: e.g., two dozen or more than a score. Scientific non-numerical quantities are represented as SI units.
List of non-numerical quantities |
https://en.wikipedia.org/wiki/Semigroup%20with%20involution | In mathematics, particularly in abstract algebra, a semigroup with involution or a *-semigroup is a semigroup equipped with an involutive anti-automorphism, which—roughly speaking—brings it closer to a group because this involution, considered as unary operator, exhibits certain fundamental properties of the operation of taking the inverse in a group: uniqueness, double application "cancelling itself out", and the same interaction law with the binary operation as in the case of the group inverse. It is thus not a surprise that any group is a semigroup with involution. However, there are significant natural examples of semigroups with involution that are not groups.
An example from linear algebra is the multiplicative monoid of real square matrices of order n (called the full linear monoid). The map which sends a matrix to its transpose is an involution because the transpose is well defined for any matrix and obeys the law , which has the same form of interaction with multiplication as taking inverses has in the general linear group (which is a subgroup of the full linear monoid). However, for an arbitrary matrix, AAT does not equal the identity element (namely the diagonal matrix). Another example, coming from formal language theory, is the free semigroup generated by a nonempty set (an alphabet), with string concatenation as the binary operation, and the involution being the map which reverses the linear order of the letters in a string. A third example, from basic set theory, is the set of all binary relations between a set and itself, with the involution being the converse relation, and the multiplication given by the usual composition of relations.
Semigroups with involution appeared explicitly named in a 1953 paper of Viktor Wagner (in Russian) as result of his attempt to bridge the theory of semigroups with that of semiheaps.
Formal definition
Let S be a semigroup with its binary operation written multiplicatively. An involution in S is a unary operation * |
https://en.wikipedia.org/wiki/The%20Lightning%20Process | The Lightning Process (LP) is a three-day personal training programme developed and trademarked by British osteopath Phil Parker. It claims to be beneficial for various conditions, including chronic fatigue syndrome, depression and chronic pain.
Developed in the late 1990s, it aims to teach techniques for managing the acute stress response that the body experiences under threat. The course aims to help recognise the stress response, calm it and manage it in the long term. It also applies some ideas drawn from neurolinguistic programming (a pseudoscience), as well as elements of life coaching.
A clinical trial in 2017 found that Lightning Process was effective when added to treatment for chronic fatigue syndrome, but it is not recommended by the NHS. Two corrections of this article were published due to methodological weaknesses.
The approach has raised some controversy due to using psychological techniques to cure a physical illness. The website was amended after the Advertising Standards Authority ruled that it was misleading. In 2021, after a review of the available evidence, the National Institute for Health and Care Excellence explicitly advised against the use of Lighting Process among patients with chronic fatigue syndrome.
Description
The Lightning Process comprises three group sessions conducted on three consecutive days, lasting about 12 hours altogether, conducted by trained practitioners.
According to its developer, Phil Parker, the programme aims to teach participants about the acute stress response the body experiences under threat. It aims to help trainees spot when this response is happening and learn how to calm it. Techniques based on movement, postural awareness and personal coaching are intended to modify the production of stress hormones. Participants practise a learnt series of steps to habituate the calming method.
The Lightning Process is based on the theory that the body can get stuck in a persistent stress response. The initial stresso |
https://en.wikipedia.org/wiki/Giulio%20Ascoli | Giulio Ascoli (20 January 1843, Trieste – 12 July 1896, Milan) was a Jewish-Italian mathematician. He was a student of the Scuola Normale di Pisa, where he graduated in 1868.
In 1872 he became Professor of Algebra and Calculus of the Politecnico di Milano University. From 1879 he was professor of mathematics at the Reale Istituto Tecnico Superiore, where, in 1901, was affixed a plaque that remembers him.
He was also a corresponding member of Istituto Lombardo.
He made contributions to the theory of functions of a real variable and to Fourier series. For example, Ascoli introduced equicontinuity in 1884, a topic regarded as one of the fundamental concepts in the theory of real functions. In 1889, Italian mathematician Cesare Arzelà generalized Ascoli's Theorem into the Arzelà–Ascoli theorem, a practical sequential compactness criterion of functions.
See also
Measure (mathematics)
Oscillation (mathematics)
Riemann Integral
Notes
Biographical references
.
(in Italian). Available from the website of the. |
https://en.wikipedia.org/wiki/UNdata | UNdata is an Internet search engine, retrieving data series from statistical databases provided by the UN System. UNdata was launched in February 2008. It is a product of the United Nations Statistics Division (UNSD) developed in partnership with Statistics Sweden and the Swedish International Development Cooperation Agency (SIDA).
UNdata allows searching and downloading a variety of statistical resources covering the following areas: Education, Employment, Energy, Environment, Food and Agriculture, Health, Human Development, Industry, Information and Communication Technology, National Accounts, Population, Refugees, Trade and Tourism.
UNdata has been featured in CNET TV and listed as Best Of The Internet in PC Magazine.
UNdata is listed in the Registry of Research Data Repositories re3data.org.
PET Lab
UNSD is the lead body for the Privacy-Enhancing Technologies Lab (PET Lab). The PET Lab together with the ITU AI for Good programme jointly organize the Trustworthy AI standardization PET programme of work. |
https://en.wikipedia.org/wiki/GSTT1 | Glutathione S-transferase theta-1 is an enzyme that in humans is encoded by the GSTT1 gene.
Glutathione S-transferase (GST) theta 1 (GSTT1) is a member of a superfamily of proteins that catalyze the conjugation of reduced glutathione to a variety of electrophilic and hydrophobic compounds. Human GSTs can be divided into five main classes: alpha, mu, pi, theta, and zeta. The theta class includes GSTT1 and GSTT2. The GSTT1 and GSTT2 share 55% amino acid sequence identity and both of them were claimed to have an important role in human carcinogenesis. The GSTT1 gene is located approximately 50kb away from the GSTT2 gene. The GSTT1 and GSTT2 genes have a similar structure, being composed of five exons with identical exon/intron boundaries. |
https://en.wikipedia.org/wiki/Lipopolysaccharide%20kinase%20%28Kdo/WaaP%29%20family | In molecular biology, the lipopolysaccharide kinase (Kdo/WaaP) family is a family of lipopolysaccharide kinases that includes lipopolysaccharide core heptose(I) kinase rfaP (encoded by the waaP (rfaP) gene). Lipopolysaccharide core heptose(I) kinase rfaP is required for the addition of phosphate to O-4 of the first heptose residue of the lipopolysaccharide (LPS) inner core region. It has previously been shown that it is necessary for resistance to hydrophobic and polycationic antimicrobials in E. coli and that it is required for virulence in invasive strains of Salmonella enterica. The family also includes 3-deoxy-D-manno-octulosonic acid kinase (KDO kinase) from Haemophilus influenzae, which phosphorylates Kdo-lipid IV(A), a lipopolysaccharide precursor, and is involved in virulence. |
https://en.wikipedia.org/wiki/Tomsrtbt | tomsrtbt (pronounced: Tom's Root Boot) is a very small Linux distribution. It is short for "Tom's floppy which has a root filesystem and is also bootable." Its author, Tom Oehser, touts it as "The most GNU/Linux on one floppy disk", containing many common Linux command-line tools useful for system recovery (Linux and other operating systems.) It also features drivers for many types of hardware, and network connectivity.
It could be created from within Linux or earlier versions of Windows running in MS-DOS mode, either by formatting a standard 1.44MB floppy disk as a higher density 1.722MB disk and writing the tomsrtbt image to the disk, or by burning it as a bootable CD. It is capable of reading and writing the filesystems of many operating systems of its era, including ext2/ext3 (used in Linux), FAT (used by DOS and Windows), NTFS (used in Windows NT, 2000, and XP) and Minix (used by the Minix operating system).
Windows NT and most later versions of Windows, including 2000, XP, and Vista, cannot create a tomsrtbt floppy as their floppy driver does not allow the extended format.
A few of the utilities on tomsrtbt are written in the Lua programming language, and many more use BusyBox. Space saving compiler options were used throughout, the kernel was patched to support loading an image compressed with Bzip2, and in many cases older or alternate versions of programs were selected due to their smaller size.
See also
Lightweight Linux distribution |
https://en.wikipedia.org/wiki/Malwarebytes | Malwarebytes Inc. is an American Internet security company that specializes in protecting home computers, smartphones, and companies from malware and other threats. It has offices in Santa Clara, California; Clearwater, Florida; Tallinn, Estonia; Bastia Umbra, Italy; and Cork, Ireland.
Marcin Kleczynski has been the CEO of Malwarebytes since 2008.
History
Early history and background
Malwarebytes Inc. was informally established in 2004. CEO and founder Marcin Kleczynski, originally from Poland, was still a teenager attending high school in Bensenville, Illinois, and was working as a technician in a computer repair shop in Chicago. He noticed that whenever infected computers arrived, the shop would typically reformat the computer entirely, rather than combat the virus, even if the infection was only minor. Kleczynski later discovered that, when his mother's computer became infected, neither McAfee nor Symantec would remove the malware from his system. He later recalled "I've never been as angry as when I got my computer infected", and professed that his mother told him to fix it "under penalty of death".
It was only after Kleczynski posted on the forum SpywareInfo that he was able to learn how to remove the virus, which took three days. The company was unofficially founded after this, when Kleczynski conversed and became friends with several of the editors of the forum, who tempted him to buy an unused domain from them.
With one of the site's regulars, Bruce Harrison, Kleczynski wrote the inaugural version of the company's software. In 2006, Kleczynski worked with a college roommate to produce a freely available program called "RogueRemover", a utility which specialized in fighting against a type of infection known as "rogues", which scam computer users into giving away their credit card information through fake anti-virus software. RogueRemover proved instrumental in developing Malwarebytes Anti-Malware, and Kleczynski was able to set up a forum which enabled |
https://en.wikipedia.org/wiki/Niko%20Bellic | Niko Bellic is a fictional character and the playable protagonist of Rockstar North's 2008 video game Grand Theft Auto IV, the sixth main instalment in Rockstar Games' Grand Theft Auto series. He also makes non-playable appearances in the game's episodic content The Lost and Damned and The Ballad of Gay Tony, both released in 2009. Michael Hollick provided the character's voice and motion capture.
Within the game's storyline, Niko is an ex-soldier from Eastern Europe, who was shaped by his experiences in an unidentified war he fought in, developing a very cynical view on life. After becoming involved with a Russian crime syndicate, and discovering that his unit was sold out to enemy forces, he decides to move to Liberty City to pursue the American Dream, inspired by his cousin Roman’s personal tales of luxury and riches that he had experienced while living there for the past decade. However, upon his arrival, he discovers that those stories were greatly exaggerated, and attempts to improve his and Roman's financial situation by becoming involved with the local criminal underworld. As the game's story progresses, Niko works for various prominent fictional crime figures, in the hopes of finding the traitor who betrayed his unit during the war, while slowly learning to let go of his past and quest for revenge, and attempting to leave the criminal life that comes with major risks.
Niko's character received critical acclaim for his maturity, moral ambiguity, and personal growth, and has been called one of the best protagonists in the series. For his role, Hollick won Best Performance by a Human Male at the 2008 Spike Video Game Awards.
Nationality
Niko's nationality is unspecified in the game and is subject to debate. It was believed by some that he was Russian, Serbian, or Croatian. Executive producer Sam Houser spoke on the matter, saying that Niko is "from that grey part of broken-down Eastern Europe", suggesting that Niko's nationality was left intentionally vague |
https://en.wikipedia.org/wiki/Ethyl%20eicosapentaenoic%20acid | Ethyl eicosapentaenoic acid (E-EPA, icosapent ethyl), sold under the brand name Vascepa among others, is a medication used to treat dyslipidemia and hypertriglyceridemia. It is used in combination with changes in diet in adults with hypertriglyceridemia ≥ 150 mg/dL. Further, it is often required to be used with a statin (maximally-tolerated dose).
The most common side effects are musculoskeletal pain, peripheral edema (swelling of legs and hands), atrial fibrillation, and arthralgia (joint pain). Other common side effects include bleeding, constipation, gout, and rash.
It is made from the omega-3 fatty acid eicosapentaenoic acid (EPA). The US Food and Drug Administration (FDA) granted the approval of icosapent ethyl in 2012 to Amarin Corporation, and it became the second fish oil-based medication after omega-3-acid ethyl esters (brand named Lovaza, itself approved in 2004). On 13 December 2019, the FDA also approved Vascepa as the first drug specifically "to reduce cardiovascular risk among people with elevated triglyceride levels". It is available as a generic medication. In 2020, it was the 285th most commonly prescribed medication in the United States, with more than 1million prescriptions.
Medical uses
In the European Union, icosapent ethyl is indicated to reduce cardiovascular risk as an adjunct to statin therapy.
In the United States, icosapent ethyl is indicated as an adjunct to maximally tolerated statin therapy to reduce the risk of myocardial infarction, stroke, coronary revascularization, and unstable angina requiring hospitalization in adults with elevated triglyceride levels (≥ 150 mg/dL) and established cardiovascular disease or diabetes and two or more additional risk factors for cardiovascular disease. It is also indicated as an adjunct to diet to reduce triglyceride levels in adults with severe (≥ 500 mg/dL) hypertriglyceridemia.
Intake of large doses (2.0 to 4.0 g/day) of long-chain omega-3 fatty acids as prescription drugs or dietary suppleme |
https://en.wikipedia.org/wiki/Straw%20chamber | A straw chamber is a type of Gaseous ionization detector. It is a long tube with a wire down the center and a gas which becomes ionized when a particle passes through. A potential difference is maintained between the wire and the walls of the tube, so that once the gas is ionized electrons move in one direction and ions in the other. This produces a current which indicates that a particle has passed through the chamber.
Many straws together can be used to track particles in a straw tracker. A straw tracker is a type of particle detector which uses many straw chambers to track the path of a particle. The path of a particle is determined by the best fit to all the straws with hits. Since the time for a particular straw to produce a signal is proportional to the distance of the particle's closest approach to that chamber's wire, if a particle on a predictable path (e.g. a helix in a magnetic field) passes through many straws, the path of the particle can be determined more precisely than the size of any particular straw.
Specific uses
There are about 298,000 drift tubes (straws) in the Transition Radiation Tracker (TRT) of the ATLAS_experiment at the Large Hadron Collider. |
https://en.wikipedia.org/wiki/Reinhard%20Oehme | Reinhard Oehme (; born 26 January 1928, Wiesbaden; died sometime between 29 September and 4 October 2010, Hyde Park) was a German-American physicist known for the discovery of C (charge conjugation) non-conservation in the presence of P (parity) violation, the formulation and proof of hadron dispersion relations, the "Edge of the Wedge Theorem" in the function theory of several complex variables, the Goldberger-Miyazawa-Oehme sum rule, reduction of quantum field theories, Oehme-Zimmermann superconvergence relations for gauge field correlation functions, and many other contributions.
Oehme was born in Wiesbaden, Germany as the son of Dr. Reinhold Oehme and Katharina Kraus. In 1952, in São Paulo, Brazil, he married Mafalda Pisani, who was born in Berlin as the daughter of Giacopo Pisani and Wanda d'Alfonso. Mafalda died in Chicago in August of the year 2004.
Education and career
Completing the Abitur at the Rheingau Gymnasium in Geisenheim near Wiesbaden, Oehme started to study physics and mathematics at the Johann Wolfgang Goethe University Frankfurt am Main, receiving the Diploma in 1948 as student of Erwin Madelung.
Then he moved to Göttingen, joining the Max Planck Institute for Physics as a doctoral student of Werner Heisenberg, who was also a professor at the University of Göttingen. Early in 1951, Oehme completed the requirements for his Dr.rer.nat at Göttingen Universität. The translation of the title of his thesis is: "Creation of Photons in Collisions of Nucleons” Later this year, Heisenberg asked him to join Carl Friedrich von Weizsäcker on a trip to Brazil for the start-up of the Instituto de Física Teórica in São Paulo, considered also as a possible escape in view of the tense situation in Europe. In 1953, he returned to his assistant position at the Max Planck Institute in Göttingen. During the early fifties, the institute was a most interesting place. Oehme was there among an exceptional group of people around Heisenberg, including Vladimir Glaser |
https://en.wikipedia.org/wiki/Gosset%20graph | The Gosset graph, named after Thorold Gosset, is a specific regular graph (1-skeleton of the 7-dimensional 321 polytope) with 56 vertices and valency 27.
Construction
The Gosset graph can be explicitly constructed as follows: the 56 vertices are the vectors in R8 obtained by permuting the coordinates and possibly taking the opposite of the vector (3, 3, −1, −1, −1, −1, −1, −1). Two such vectors are adjacent when their inner product is 8, or equivalently when their distance is .
An alternative construction is based on the 8-vertex complete graph K8. The vertices of the Gosset graph can be identified with two copies of the set of edges of K8.
Two vertices of the Gosset graph that come from the same copy are adjacent if they correspond to disjoint edges of K8; two vertices that come from different copies are adjacent if they correspond to edges that share a single vertex.
Properties
In the vector representation of the Gosset graph, two vertices are at distance two when their inner product is −8 and at distance three when their inner product is −24 (which is only possible if the vectors are each other's opposite). In the representation based on the edges of K8, two vertices of the Gosset graph are at distance three if and only if they correspond to different copies of the same edge of K8.
The Gosset graph is distance-regular with diameter three.
The induced subgraph of the neighborhood of any vertex in the Gosset graph is isomorphic to the Schläfli graph.
The automorphism group of the Gosset graph is isomorphic to the Coxeter group E7 and hence has order 2903040. The Gosset 321 polytope is a semiregular polytope. Therefore, the automorphism group of the Gosset graph, E7, acts transitively upon its vertices, making it a vertex-transitive graph.
The characteristic polynomial of the Gosset graph is
Therefore, this graph is an integral graph. |
https://en.wikipedia.org/wiki/Auricular%20hypertrichosis | Auricular hypertrichosis (hypertrichosis lanuginosa acquisita, hypertrichosis pinnae auris) is a genetic condition expressed as long and strong hairs growing from the helix of the pinna.
Presentation
Ear hair generally refers to the terminal hair arising from follicles inside the external auditory meatus in humans. In its broader sense, ear hair may also include the fine vellus hair covering much of the ear, particularly at the prominent parts of the anterior ear, or even the abnormal hair growth as seen in hypertrichosis and hirsutism. Medical research on the function of ear hair is currently very scarce.
Hair growth within the ear canal is often observed to increase in older men, together with increased growth of nose hair. Excessive hair growth within or on the ear is known medically as auricular hypertrichosis. Some men, particularly in the male population of India, have coarse hair growth along the lower portion of the helix, a condition referred to as "having hairy pinnae" (hypertrichosis lanuginosa acquisita).
Genetics
The genetic basis of auricular hypertrichosis has not been settled. Some researchers have proposed a Y-linked pattern of inheritance and others have suggested an autosomal gene is responsible. A third hypothesis predicts the phenotype results from the interaction of two loci, one on the homologous part of the X and Y and one on the nonhomologous sequence of the Y. These hypotheses are not mutually exclusive, and there may be a variety of genetic mechanisms underlying this phenotype.
Lee et al. (2004), using Y-chromosomal DNA binary-marker haplotyping, suggested that a cohort of southern Indian hairy-eared males carried Y chromosomes from many haplogroup. The hypothesis of Y- linkage would require multiple independent mutations within a single population. No significant difference between the Y-haplogroup frequencies of hairy-eared males and those of a geographically matched control sample of unaffected males was established. The study conc |
https://en.wikipedia.org/wiki/American%20Society%20of%20Plant%20Biologists | The American Society of Plant Biologists (ASPB) is a non-profit professional society for research and education in plant science with over 4,000 members world-wide. It was founded in 1924, as the American Society of Plant Physiologists (ASPP). The name was changed to the American Society of Plant Biologists (ASPB) as of 2001. Membership in the society is open to any person from any country who deals with physiology, molecular biology, environmental biology, cell biology and plant biophysics or related issues.
The society publishes the peer-reviewed journals Plant Physiology (1926-) and The Plant Cell (1989-) as well as ASPB News. The American Society of Plant Biologists also has partnered with the Society for Experimental Biology, and Wiley to publish an online-only science journal Plant Direct. In 2000, it published the first edition of the textbook Biochemistry & Molecular Biology of Plants.
The society has given the Charles Reid Barnes Life Membership Award since 1925. It established the Stephen Hales Prize in 1927. As of 2007, the society began to designate Fellows of the ASPB for "long-term contributions to plant biology". ASPB Fellows are distinct from ASPB's "Plantae Fellows", who are selected from a variety of countries and backgrounds for their work as science communicators.
The first President of the Society was Charles Albert Shull (1924-1925), with founder R. B. Harvey as Secretary-Treasurer.
Other presidents of the Society include Harry Beevers (1961-1962) and Aubrey Naylor (1960-1961). The first woman to be president of the society was Elisabeth Gantt (1988-1989). |
https://en.wikipedia.org/wiki/Chaos%3A%20Making%20a%20New%20Science | Chaos: Making a New Science is a debut non-fiction book by James Gleick that initially introduced the principles and early development of the chaos theory to the public. It was a finalist for the National Book Award and the Pulitzer Prize in 1987, and was shortlisted for the Science Book Prize in 1989. The book was published on October 29, 1987 by Viking Books.
Overview
Chaos: Making a New Science was the first popular book about chaos theory. It describes the Mandelbrot set, Julia sets, and Lorenz attractors without using complicated mathematics. It portrays the efforts of dozens of scientists whose separate work contributed to the developing field. The text remains in print and is widely used as an introduction to the topic for the mathematical layperson. The book approaches the history of chaos theory chronologically, starting with Edward Norton Lorenz and the butterfly effect, through Mitchell Feigenbaum, and ending with more modern applications.
The book covers chaos theory under the lens of four themes: sensitive dependence on initial conditions, self-similarity, universality, and nonlinearity.
An enhanced ebook edition was released by Open Road Media in 2011, adding embedded video and hyperlinked notes.
Reception
Robert Sapolsky said, "Chaos is the first book since Baby Beluga where I've gotten to the last page and immediately started reading it over again from the front: I've found this to be the most influential book in my thinking about science since college."
Freeman Dyson praised the book for its popular account but critiqued the omitting of the earlier work of Dame Mary L. Cartwright and J. E. Littlewood in forming the foundation of chaos theory. |
https://en.wikipedia.org/wiki/Kokichi%20Sugihara | Kōkichi Sugihara (, born June 29, 1948, in Gifu Prefecture) is a Japanese mathematician and artist known for his three-dimensional optical illusions that appear to make marbles roll uphill, pull objects to the highest point of a building's roof, and make circular pipes look rectangular. His illusions, which often involve videos of three-dimensional objects shown from carefully chosen perspectives, won first place at the Best Illusion of the Year Contest in 2010, 2013, 2018,and 2020
and second place in 2015 and 2016.
Education and career
Sugihara earned bachelor's, master's, and doctoral degrees in mathematical engineering from the University of Tokyo in 1971, 1973, and 1980 respectively. From 1973 to 1981 he worked as a researcher at the Ministry of International Trade and Industry. He then became an associate professor in the Department of Information and Computer Engineering at Nagoya University in 1981, and moved back to the Department of Mathematical Engineering and Information Physics at the University of Tokyo in 1986. Since 2009 he has been a professor at Meiji University.
Illusions
Five of Sugihara's illusions have won awards at the annual Best Illusion of the Year Contest:
In 2010, his illusion "Impossible Motion: Magnet Slopes" won first place in the contest. This illusion uses forced perspective to show marbles seemingly rolling up ramps.
In 2013 he won first place again (with Jun Ono and Akiyasu Tomoeda) for "Rotation Generated by Translation", an illusion that uses Moiré patterns to create the appearance of rotation from objects moving only by translation.
In 2015 his "Ambiguous Garage Roof" won second place. The illusion appears to show a convex roof surface that is reflected in a mirror to a corrugated zig-zag shape. Neither of these appearances accurately describes the true shape of the roof.
In 2016 he won second place again for "Ambiguous Cylinder Illusion", which shows a stack of cylinders that from one point of view appear to have a circular c |
https://en.wikipedia.org/wiki/Data%20localization | Data localization or data residency law requires data about a nation's citizens or residents to be collected, processed, and/or stored inside the country, often before being transferred internationally. Such data is usually transferred only after meeting local privacy or data protection laws, such as giving the user notice of how the information will be used, and obtaining their consent.
Data localization builds upon the concept of data sovereignty that regulates certain data types by the laws applicable to the data subjects or processors. While data sovereignty may require that records about a nation's citizens or residents follow its personal or financial data processing laws, data localization goes a step further in requiring that initial collection, processing, and storage first occur within the national boundaries. In some cases, data about a nation's citizens or residents must also be deleted from foreign systems before being removed from systems in the data subject's nation.
Motivations and concerns
One of the first moves towards data localization occurred in 2005 when the Government of Kazakhstan passed a law for all ".kz" domains to be run domestically (with later exceptions for Google). However, the push for data localization greatly increased after revelations by Edward Snowden regarding United States counter-terrorism surveillance programs in 2013. Since then, various governments in Europe and around the world have expressed the desire to be able to control the flow of residents' data through technology. Some governments are accused of and some openly admit to using data localization laws as a way to surveil their own populaces or to boost local economic activity.
Technology companies and multinational organizations often oppose data localization laws because they impact efficiencies gained by regional aggregation of data centers and unification of services across national boundaries. Some vendors, such as Microsoft, have used data storage locale cont |
https://en.wikipedia.org/wiki/PRQ | PRQ is a Swedish Internet service provider and web hosting company created in 2004.
Ownership
Based in Stockholm, PRQ was created by Gottfrid Svartholm and Fredrik Neij, two founders of The Pirate Bay.
Business model
Part of PRQ's business model is to host any customers, regardless of how odd or controversial they may be. The New York Times wrote in 2008 that "The Pirate Bay guys have made a sport out of taunting all forms of authority, including the Swedish police, and PRQ has gone out of its way to be a host to sites that other companies would not touch." The PRQ service has been described as "highly secure, no-questions-asked hosting services". The company holds almost no information about its clientele and is maintaining few if any of its own logs, according to a 2008 news report. Fredrik Neij and Gottfrid Svartholm are said to have amassed "considerable expertise in withstanding legal attacks". Svartholm is quoted to have said, "We do employ our own legal staff. We are used to this sort of situation" in a telephone interview. Due to hosting The Pirate Bay, PRQ was the target of a police raid.
Criticism
The co-founders have been criticized for hosting controversial websites, including web pages that promote paedophilia, such as the North American Man/Boy Love Association (NAMBLA), a paedophile and pederasty advocacy organization. Local authorities and anti-paedophilia activists in Sweden have failed to persuade PRQ to close the sites. The pair defended their decision, citing freedom of speech.
The co-owners were also criticized for creating and hosting AMERICASDUMBESTSOLDIERS.COM, a website identifying deceased military personnel that invited visitors to rank how "dumb" the soldiers were based on the manner in which they died.
Other criticism originates from the hosting of BitTorrent website The Pirate Bay, WikiLeaks, and the French far-right blog Fdesouche.
Legal issues
On 1 October 2012, PRQ was raided and a number of sites which they provided hosting |
https://en.wikipedia.org/wiki/Thermosphaera | Thermosphaera is a genus of the Desulfurococcaceae. They are a group of prokaryotic organisms which have been discovered in extremely hot environments such as sulfur springs, volcanoes, and magma pools. Isolates of Thermosphaera were first identified in 1998 from the Obsidian Pool in Yellowstone National Park.
Cell structure and metabolism
Cells of Thermosphaera are cocci (spherical) and form grape-like aggregates during the exponential growth phase. In the late exponential and stationary growth phases, smaller groups, including some single cells, were visible. Aggregates were shown to have several flagella; single cells could have as many as eight. The cell envelope is an amorphous layer covering a cytoplasmic membrane. Temperatures exceeding 92 °C inhibits growth, as does sulfur and hydrogen. Thermosphaera cells are heterotrophic, processing energy from yeast.
Ecology
Thermosphaera are found mainly in sulfuric pools, where they thrive on the extreme temperatures. In terms of research and economic significance, learning more about these organisms and their properties may help advancements in biotechnology.
Genome structure
Sequencing the 16S rRNA of Thermosphaera showed that this isolate was a member of the group Crenarchaeota and closely related to Staphylothermus and Desulfurococcus.
See also
List of Archaea genera |
https://en.wikipedia.org/wiki/KIF2C | Kinesin-like protein KIF2C is a protein that in humans is encoded by the KIF2C gene.
The protein encoded by this gene is a member of kinesin-like protein family. Most proteins of this family are microtubule-dependent molecular motors that transport organelles within cells and move chromosomes during cell division. This protein acts to regulate microtubule dynamics in cells and is important for anaphase chromosome segregation and may be required to coordinate the onset of sister centromere separation. |
https://en.wikipedia.org/wiki/World%20Network%20of%20Biosphere%20Reserves | The UNESCO World Network of Biosphere Reserves (WNBR) covers internationally designated protected areas, known as biosphere reserves, which are meant to demonstrate a balanced relationship between people and nature (e.g. encourage sustainable development). They are created under the Man and the Biosphere Programme (MAB).
Aug
Mission
The World Network of Biosphere Reserves (WNBR) of the MAB Programme consists of a dynamic and interactive network of sites. It works to foster the harmonious integration of people and nature for sustainable development through participatory dialogue, knowledge sharing, poverty reduction, human well-being improvements, respect for cultural values and by improving society's ability to cope with climate change. It promotes north–south and South-South collaboration and represents a unique tool for international cooperation through the exchange of experiences and know-how, capacity-building and the promotion of best practices.
The network
total membership had reached 738 biosphere reserves in 134 countries (including 22 transboundary sites) occurring in all regions of the world. Myanmar had its first biosphere reserve inscribed in 2015. This already takes into account some biosphere reserves that have been withdrawn or revised through the years, as the program's focus has shifted from simple protection of nature to areas displaying close interaction between man and environment.
Criteria and periodic review process
Article 4 of the defines general criteria for an area to be qualified for designation as a biosphere reserve as follows:
It should encompass a mosaic of ecological systems representative of major biogeographic regions, including a gradation of human interventions.
It should be of significance for biological diversity conservation.
It should provide an opportunity to explore and demonstrate approaches to sustainable development on a regional scale.
It should have an appropriate size to serve the three functions of bios |
https://en.wikipedia.org/wiki/Bjorn%20Bjorholm | Bjorn Bjorholm (; born 1986) is an American professional bonsai artist and educator. He is the founder and owner of Eisei-en Bonsai Garden near Nashville, TN.
Biography
Bjorholm was born in 1986 and grew up in Knoxville, Tennessee. At the early age of 13, he received his first bonsai tree inspired by the movie "The Karate Kid." Although the tree was dead within a few months, Bjorholm was hooked on the Japanese artform, which would prove to be his life's calling. In 2001, he and his father, Tom Bjorholm, founded the Knoxville Bonsai Society.
At age 16, Bjorholm visited Japan as part of a student group. There, he met bonsai master Keiichi Fujikawa who will later become his teacher and mentor. Before beginning a formal apprenticeship, Bjorholm studied the artform with several bonsai professionals in the United States.
After graduating from college in 2008 at age 22, Bjorholm applied and became an apprentice under Keiichi Fujikawa at Kouka-en Bonsai Nursery in Osaka, Japan. He apprenticed for six years before becoming certified as a bonsai professional by the Nippon Bonsai Association. Thereafter, he worked as an artist-in-residence at Kouka-en, making him Japan's first foreign-born working bonsai artist. As part of his job at Kouka-en, Bjorholm styled and maintained many trees registered as "masterpiece" by Japan's Nippon Bonsai Association, and other important trees that would subsequently enter the prestigious Kokufu-ten bonsai exhibition, Japan's premier showcase of the artform.
Upon his return to the United States, in 2018, Bjorholm established Eisei-en () Bonsai Garden in Mount Juliet, Tennessee, where he teaches bonsai, styles and cares for privately-owned trees and for his own collection, and sells and brokers bonsai for a discerning clientele.
As part of his work spreading bonsai art, Bjorholm has an active YouTube channel with a base of over 200,000 subscribers around the world, as well as a subscription-based online learning platform, Bonsai-U.
In 2019 |
https://en.wikipedia.org/wiki/Reachability%20problem | Reachability is a fundamental problem that appears in several different contexts: finite- and infinite-state concurrent systems, computational models like cellular automata and Petri nets, program analysis, discrete and continuous systems, time critical systems, hybrid systems, rewriting systems, probabilistic and parametric systems, and open systems modelled as games.
In general the reachability problem can be formulated as follows: Given a computational (potentially infinite state) system with a set of allowed rules or transformations, decide whether a certain state of a system is reachable from a given initial state of the system.
Variants of the reachability problem may result from additional constraints on the initial or final states, specific requirement for reachability paths as well as for iterative reachability or changing the questions into analysis of winning strategies in infinite games or unavoidability of some dynamics.
Typically, for a fixed system description given in some form (reduction rules, systems of equations, logical formulas, etc.) a reachability problem consists of checking whether a given set of target states can be reached starting from a fixed set of initial states. The set of target states can be represented explicitly or via some implicit representation (e.g., a system of equations, a set of minimal elements with respect to some ordering on the states). Sophisticated quantitative and qualitative properties can often be reduced to basic reachability questions. Decidability and complexity boundaries, algorithmic solutions, and efficient heuristics are all important aspects to be considered in this context. Algorithmic solutions are often based on different combinations of exploration strategies, symbolic manipulations of sets of states, decomposition properties, or reduction to linear programming problems, and they often benefit from approximations, abstractions, accelerations and extrapolation heuristics. Ad hoc solutions as well as |
https://en.wikipedia.org/wiki/Target%20selection | Target selection is the process by which axons (nerve fibres) selectively target other cells for synapse formation. Synapses are structures which enable electrical or chemical signals to pass between nerves. While the mechanisms governing target specificity remain incompletely understood, it has been shown in many organisms that a combination of genetic and activity-based mechanisms govern initial target selection and refinement. The process of target selection has multiple steps that include Axon pathfinding when neurons extend processes to specific regions, cellular target selection when neurons choose appropriate partners in a target region from a multitude of potential partners, and subcellular target selection where axons often target particular regions of a partner neuron.
Description
As bundled axons finish navigating through various neural circuits during neural development, the growth cones must selectively target with which cells it will synapse. This can be particularly well observed in the visual and olfactory systems of organisms. In order to develop into a properly functioning nervous system, there must be an extremely high degree of accuracy in which cell the growth cone forms neural connections. Although the target cell selection must be highly accurate, the degree of specificity that the neural connectivity achieves varies based on the neuronal circuitry system. The target selection process of an axon to develop synaptic connections with specific cells can be broken down into multiple stages that are not necessarily confined to exact chronological order.
The stages of targeting include:
region specification
target cell specification
subcellular specification
synaptic refinement
Region specification
The first stage in target selection is specification of target region, a process known as Axon pathfinding. Growing neurites follow gradients of cell surface molecules that serve as chemoattractants and repellents to the growth cone. This perspect |
https://en.wikipedia.org/wiki/Scalar%20chromodynamics | In quantum field theory, scalar chromodynamics, also known as scalar quantum chromodynamics or scalar QCD, is a gauge theory consisting of a gauge field coupled to a scalar field. This theory is used experimentally to model the Higgs sector of the Standard Model.
It arises from a coupling of a scalar field to gauge fields. Scalar fields are used to model certain particles in particle physics; the most important example is the Higgs boson. Gauge fields are used to model forces in particle physics: they are force carriers. When applied to the Higgs sector, these are the gauge fields appearing in electroweak theory, described by Glashow–Weinberg–Salam theory.
Matter content and Lagrangian
Matter content
This article discusses the theory on flat spacetime , commonly known as Minkowski space.
The model consists of a complex vector valued scalar field minimally coupled to a gauge field .
The gauge group of the theory is a Lie group . Commonly, this is for some , though many details hold even when we don't concretely fix .
The scalar field can be treated as a function , where is the data of a representation of . Then is a vector space. The 'scalar' refers to how transforms (trivially) under the action of the Lorentz group, despite being vector valued. For concreteness, the representation is often chosen to be the fundamental representation. For , this fundamental representation is . Another common representation is the adjoint representation. In this representation, varying the Lagrangian below to find the equations of motion gives the Yang–Mills–Higgs equation.
Each component of the gauge field is a function where is the Lie algebra of from the Lie group–Lie algebra correspondence. From a geometric point of view, are the components of a principal connection under a global choice of trivialization (which can be made due to the theory being on flat spacetime).
Lagrangian
The Lagrangian density arises from minimally coupling the Klein–Gordon Lagrangian |
https://en.wikipedia.org/wiki/First%20passage%20percolation | First passage percolation is a mathematical method used to describe the paths reachable in a random medium within a given amount of time.
Introduction
First passage percolation is one of the most classical areas of probability theory. It was first introduced by John Hammersley and Dominic Welsh in 1965 as a model of fluid flow in a porous media. It is part of percolation theory, and classical Bernoulli percolation can be viewed as a subset of first passage percolation.
Most of the beauty of the model lies in its simple definition (as a random metric space) and the property that several of its fascinating conjectures do not require much effort to be stated. Most times, the goal of first passage percolation is to understand a random distance on a graph, where weights are assigned to edges. Most questions are tied to either find the path with the least weight between two points, known as a geodesic, or to understand how the random geometry behaves in large scales.
Mathematics
As is the case in percolation theory in general, many of the problems related to first passage percolation involve finding optimal routes or optimal times.
The model is defined as follows.
Let be a graph. We place a non-negative random variable , called the passage time of the edge , at each nearest-neighbor edge of the graph . The collection is usually assumed to be independent, identically distributed but there are variants of the model.
The random variable is interpreted as the time or the cost needed to traverse edge .
Since each edge in first passage percolation has its own individual weight (or time) we can write the total time of a path as the summation of weights of each edge in the path.
Given two vertices of one then sets
where the infimum is over all finite paths that start at and end at .
The function induces a random pseudo-metric on .
The most famous model of first passage percolation is on the lattice . One of its most notorious questions is "What does a ball o |
https://en.wikipedia.org/wiki/Trace%20fossil | A trace fossil, also known as an ichnofossil (; from ikhnos "trace, track"), is a fossil record of biological activity by lifeforms but not the preserved remains of the organism itself. Trace fossils contrast with body fossils, which are the fossilized remains of parts of organisms' bodies, usually altered by later chemical activity or mineralization. The study of such trace fossils is ichnology and is the work of ichnologists.
Trace fossils may consist of physical impressions made on or in the substrate by an organism. For example, burrows, borings (bioerosion), urolites (erosion caused by evacuation of liquid wastes), footprints and feeding marks and root cavities may all be trace fossils.
The term in its broadest sense also includes the remains of other organic material produced by an organism; for example coprolites (fossilized droppings) or chemical markers (sedimentological structures produced by biological means; for example, the formation of stromatolites). However, most sedimentary structures (for example those produced by empty shells rolling along the sea floor) are not produced through the behaviour of an organism and thus are not considered trace fossils.
The study of traces – ichnology – divides into paleoichnology, or the study of trace fossils, and neoichnology, the study of modern traces. Ichnological science offers many challenges, as most traces reflect the behaviour – not the biological affinity – of their makers. Accordingly, researchers classify trace fossils into form genera, based on their appearance and on the implied behaviour, or ethology, of their makers.
Occurrence
Traces are better known in their fossilized form than in modern sediments. This makes it difficult to interpret some fossils by comparing them with modern traces, even though they may be extant or even common. The main difficulties in accessing extant burrows stem from finding them in consolidated sediment, and being able to access those formed in deeper water.
Trace fo |
https://en.wikipedia.org/wiki/Migration%20%28ecology%29 | Migration, in ecology, is the large-scale movement of members of a species to a different environment. Migration is a natural behavior and component of the life cycle of many species of mobile organisms, not limited to animals, though animal migration is the best known type. Migration is often cyclical, frequently occurring on a seasonal basis, and in some cases on a daily basis. Species migrate to take advantage of more favorable conditions with respect to food availability, safety from predation, mating opportunity, or other environmental factors.
Migration is most commonly seen as animal migration, the physical movement by animals from one area to another. That includes bird, fish, and insect migration. However, plants can be said to migrate, as seed dispersal enables plants to grow in new areas, under environmental constraints such as temperature and rainfall, resulting in changes such as forest migration.
Mechanisms
While members of some species learn a migratory route on their first journey with older members of their group, other species genetically pass on information regarding their migratory paths. Despite many differences in organisms’ migratory cues and behaviors, “considerable similarities appear to exist in the cues involved in the different phases of migration.” Migratory organisms use environmental cues like photoperiod and weather conditions as well as internal cues like hormone levels to determine when it is time to begin a migration. Migratory species use senses such as magnetoreception or olfaction to orient themselves or navigate their route, respectively.
Factors
The factors that determine migration methods are variable due to the inconsistency of major seasonal changes and events. When an organism migrates from one location to another, its energy use and rate of migration are directly related to each other and to the safety of the organism. If an ecological barrier presents itself along a migrant's route, the migrant can either choose t |
https://en.wikipedia.org/wiki/Single-chip%20Cloud%20Computer | The Single-Chip Cloud Computer (SCC) is a computer processor (CPU) created by Intel Corporation in 2009 that has 48 distinct physical cores that communicate through an architecture similar to that of a cloud computer data center. Cores are a part of the processor that carry out instructions of code that allow the computer to run. The SCC was a product of a project started by Intel to research multi-core processors and parallel processing (doing multiple calculations at once). Additionally, Intel wanted to experiment with incorporating the designs and architecture of huge cloud computer data centers (cloud computing) into a single processing chip. They took the aspect of cloud computing, in which there are many remote servers that communicate with each other, and applied it to a microprocessor. It was a new concept that Intel wanted to experiment with. The name "Single-chip Cloud Computer" originated from this concept.
Uses
The SCC is currently still being used for research purposes. It currently can run the GNU operating system on the chip, but cannot boot Windows. Some applications of the SCC are web servers, data informatics, bioinformatics, and financial analytics.
Technical details
Intel developed this new chip architecture based on huge cloud data centers, the cores are separated across the chip but are able to directly communicate with each other. The chip contains 48 P54C Pentium cores connected with a 4×6 2D-mesh. This mesh is a group of 24 tiles set up in four rows and six columns. Each tile contained two cores and a 16 KB (8 per core) message passing buffer (MPB) shared by the two cores, essentially a router. This router allows each core to communicate with each other. Previously cores had to send information back to the main memory and there it would be re-routed to other cores. The SCC contains 1.3 billion 45 nm transistors that can amplify signals or act as a switch and turn core pairs on and off. These transistors use anywhere from 25 to 125 watts |
https://en.wikipedia.org/wiki/Torsion%20conjecture | In algebraic geometry and number theory, the torsion conjecture or uniform boundedness conjecture for torsion points for abelian varieties states that the order of the torsion group of an abelian variety over a number field can be bounded in terms of the dimension of the variety and the number field. A stronger version of the conjecture is that the torsion is bounded in terms of the dimension of the variety and the degree of the number field. The torsion conjecture has been completely resolved in the case of elliptic curves.
Elliptic curves
From 1906 to 1911, Beppo Levi published a series of papers investigating the possible finite orders of points on elliptic curves over the rationals. He showed that there are infinitely many elliptic curves over the rationals with the following torsion groups:
Cn with 1 ≤ n ≤ 10, where Cn denotes the cyclic group of order n;
C12;
C2n × C2 with 1 ≤ n ≤ 4, where × denotes the direct sum.
At the 1908 International Mathematical Congress in Rome, Levi conjectured that this is a complete list of torsion groups for elliptic curves over the rationals. The torsion conjecture for elliptic curves over the rationals was independently reformulated by and again by , with the conjecture becoming commonly known as Ogg's conjecture.
drew the connection between the torsion conjecture for elliptic curves over the rationals and the theory of classical modular curves. In the early 1970s, the work of Gérard Ligozat, Daniel Kubert, Barry Mazur, and John Tate showed that several small values of n do not occur as orders of torsion points on elliptic curves over the rationals. proved the full torsion conjecture for elliptic curves over the rationals. His techniques were generalized by and , who obtained uniform boundedness for quadratic fields and number fields of degree at most 8 respectively. Finally, proved the conjecture for elliptic curves over any number field.
An effective bound for the size of the torsion group in terms of the degree of |
https://en.wikipedia.org/wiki/MusicDNA%20%28company%29 | MusicDNA, formerly Bach Technology, is a Norwegian company that develops and licenses digital music technology, notably MusicDNA to provide custom, Internet updated multi-media content - like videos, song lyrics, or social media - while audio is played. It has partnered with the company that invented MP3, Fraunhofer Institute for Digital Media Technology (IDMT), for technical expertise. Two of its key investors are former Chief Executive Officer (CEO) at Sony Music Entertainment, Shigeo Maruyama, and MP3's inventor, Karlheinz Brandenburg.
History
Bach Technology A.S., founded by Dagfinn Bach and Karlheinz Brandenburg, is headquartered in Bergen, Norway. Bach was instrumental in the development of digital music technology and MP3 technology in 1993. He is the company's president. Brandenburg's experience is in the media; He worked for Bertelsmann in Germany. The company has partnered with the company that invented MP3, Fraunhofer Institute for Digital Media Technology (IDMT) in Germany, for their technical expertise. Investors include the former Chief Executive Officer (CEO) at Sony Music Entertainment, Shigeo Maruyama, and MP3's inventor Karlheinz Brandenburg.
Bach Technology is a company that licenses an Internet-accessible database containing information about the contents of audiofiles, and other software related to music recognition. Bach Technology uses MPEG-7 technology to create custom selections based upon the customer's preferences for multi-media content. Dagfinn Bach and Sebastian Schmidt are named on the United States patent for the technology to manage the communication of multiple types of data.
MusicDNA
Bach Technology has developed MusicDNA technology, the next evolutionary step for playing music after MP3. Introduced at the January 2010 MIDEM music industry trade fair in Cannes, MusicDNA is an MP3 file format that produces data and video content while music plays on MP3 players or iPods. This could include social media, like Twitter, videos or i |
https://en.wikipedia.org/wiki/Wratten%20number | Wratten numbers are a labeling system for optical filters, usually for photographic use comprising a number sometimes followed by a letter. The number denotes the color of the filter, but is arbitrary and does not encode any information (the 80A–80D are blue, the next filters in numerical order, 81A–81EF, are orange); letters almost always increase with increasing strength (the exception being 2B, 2A, 2C, 2E).
They are named for the founder of the first photography company, British inventor Frederick Wratten. Wratten and partner C.E.K. Mees sold their company to Eastman Kodak in 1912, and Kodak started manufacturing Wratten filters. They remain in production, and are sold under license through the Tiffen corporation.
Wratten filters are often used in observational astronomy by amateur astronomers. Color filters for visual observing made by GSO, Baader, Lumicon, or other companies are actually Wratten filters mounted in standard or 2 in (nominal, 48 mm actual) filter threads. For imaging interference filters are used. Wratten filters are also used in photomicrography.
Filters made by various manufacturers may be identified by Wratten numbers but not precisely match the spectral definition for that number. This is especially true for filters used for aesthetic (as opposed to technical) reasons. For example, an 81B warming filter is a filter used to slightly "warm" the colors in a color photo, making the scene a bit less blue and more red. Many manufacturers make filters labeled as 81B with transmission curves which are similar, but not identical, to the Kodak Wratten 81B. This is according to that manufacturer's idea of how best to warm a scene, and depending on the dyes and layering techniques used in manufacturing. Some manufacturers use their own designations to avoid this confusion, for example Singh-Ray has a warming filter which they designate A‑13, which is not a Wratten number. Filters used where precisely specified and repeatable characteristics are requi |
https://en.wikipedia.org/wiki/Strengthen%20the%20Arm%20of%20Liberty%20Monument%20%28Pine%20Bluff%2C%20Arkansas%29 | The Strengthen the Arm of Liberty Monument is a replica of the Statue of Liberty (Liberty Enlightening the World) in Pine Bluff Memorial Gardens, on the south side of 10th Avenue between Georgia and State Street in Pine Bluff, Arkansas. It was placed by the Boy Scouts of America (BSA) as part of its 1950s era campaign, "Strengthen the Arm of Liberty." The statue is in height, made of copper, and is mounted on concrete base tall. The statue faces north, toward the Pine Bluff Civic Center, and there is a bronze commemorative plaque on the north face of the base. It is one of two BSA-placed statues in the state; the other is in Fayetteville.
The statue was listed on the National Register of Historic Places in 2000.
See also
National Register of Historic Places listings in Jefferson County, Arkansas
Replicas of the Statue of Liberty
Scouting memorials
Scouting museums |
https://en.wikipedia.org/wiki/United%20Nations%20Audiovisual%20Library%20of%20International%20Law | The United Nations Audiovisual Library of International Law is a free online international law research and training tool. It was created and is maintained by the Codification Division of the United Nations Office of Legal Affairs as a part of its mandate under the United Nations Programme of Assistance in the Teaching, Study, Dissemination and Wider Appreciation of International Law.
Background
The United Nations Audiovisual Library of International Law was established in 2008 under the United Nations Programme of Assistance in the Teaching, Study, Dissemination and Wider Appreciation of International Law as a tool for promoting knowledge of international law.
Faculty
Over 240 international law experts from different regions, legal systems and sectors of the legal profession have recorded lectures for the Lecture Series, prepared introductory notes for the Historic Archives and contributed their scholarly writings to the Research Library.
Notable faculty members as of 1 July 2016:
Abi-Saab, Georges
Abraham, Ronny
Al-Khasawneh, Awn S.
Annan, Kofi A.
Bassiouni, M. Cherif
Bennouna, Mohamed
Brilmayer, Lea
Buergenthal, Thomas
Cançado Trindade, Antônio Augusto
Cassese, Antonio
Caron, David
Charlesworth, Hilary
Corell, Hans
Couvreur, Philippe
Crane, David M.
Crawford, James
Donoghue, Joan E.
Dugard, John
Ferencz, Benjamin B.
Gaillard, Emmanuel
Gaja, Giorgio
Goldstone, Richard
Goodwin-Gill, Guy S.
Greenwood, Christopher
Heyns, Christof
Higgins, Rosalyn
Hossain, Kamal
Jallow, Hassan Bubacar
Kälin, Walter
Keith, Kenneth
Kingsbury, Benedict
Klabbers, Jan
Koh, Harold Hongju
Koh, Tommy
Laborde, Santiago Oñate
Lacarte Muró, Julio
Lamy, Pascal
Lauterpacht, Elihu
Mautner, Menachem
Mayr-Harting, Thomas
McDougal, Myres S.
McWhinney, Edward
Meron, Theodor
Michel, Nicolas
Momtaz, Djamchid
Odio Benito, Elizabeth
Owada, Hisashi
Palmer, Geoffrey |
https://en.wikipedia.org/wiki/Microsatellite%20enrichment | Microsatellite enrichment is a method in molecular biology used for enriching the amount of microsatellite sequences in a DNA sample. This can be achieved by designing oligonucleotide probes that hybridize with the repeats in the microsatellites and then pull out the probe/microsatellite complexes from the solution. This has been shown to be a cost-effective method to sample the genetic diversity in non-model organisms. |
https://en.wikipedia.org/wiki/Genome%20instability | Genome instability (also genetic instability or genomic instability) refers to a high frequency of mutations within the genome of a cellular lineage. These mutations can include changes in nucleic acid sequences, chromosomal rearrangements or aneuploidy. Genome instability does occur in bacteria. In multicellular organisms genome instability is central to carcinogenesis, and in humans it is also a factor in some neurodegenerative diseases such as amyotrophic lateral sclerosis or the neuromuscular disease myotonic dystrophy.
The sources of genome instability have only recently begun to be elucidated. A high frequency of externally caused DNA damage can be one source of genome instability since DNA damage can cause inaccurate translesion DNA synthesis past the damage or errors in repair, leading to mutation. Another source of genome instability may be epigenetic or mutational reductions in expression of DNA repair genes. Because endogenous (metabolically-caused) DNA damage is very frequent, occurring on average more than 60,000 times a day in the genomes of human cells, any reduced DNA repair is likely an important source of genome instability.
The usual genome situation
Usually, all cells in an individual in a given species (plant or animal) show a constant number of chromosomes, which constitute what is known as the karyotype defining this species (see also List of number of chromosomes of various organisms), although some species present a very high karyotypic variability. In humans, mutations that would change an amino acid within the protein coding region of the genome occur at an average of only 0.35 per generation (less than one mutated protein per generation).
Sometimes, in a species with a stable karyotype, random variations that modify the normal number of chromosomes may be observed. In other cases, there are structural alterations (e.g., chromosomal translocations, deletions) that modify the standard chromosomal complement. In these cases, it is indica |
https://en.wikipedia.org/wiki/Kinetic%20energy%20weapon | A kinetic energy weapon (also known as kinetic weapon, kinetic energy warhead, kinetic warhead, kinetic projectile, kinetic kill vehicle) is a weapon based solely on a projectile's kinetic energy instead of an explosive or any other kind of payload.
The term Hit-to-kill, or kinetic kill, is also used in the military aerospace field to describe kinetic energy weapons. It has been used primarily in the anti-ballistic missiles (ABM) and anti-satellite weapons (ASAT) area, but some modern anti-aircraft missiles are also hit-to-kill. Hit-to-kill systems are part of the wider class of kinetic projectiles, a class that has widespread use in the anti-tank field.
Typical kinetic energy weapons are blunt projectiles such as rocks and round shots, pointed ones such as arrows, and somewhat pointed ones such as bullets. Among projectiles that do not contain explosives are those launched from railguns, coilguns, and mass drivers, as well as kinetic energy penetrators. All of these weapons work by attaining a high muzzle velocity, or initial velocity, generally up to hypervelocity, and collide with their targets, converting the kinetic energy associated with the relative velocity between the two objects into destructive shock waves and heat. Other types of kinetic weapons are accelerated over time by a rocket engine, or by gravity. In either case, it is this kinetic energy that destroys its target.
Basic concept
Kinetic energy is a function of mass and the velocity of an object. For a kinetic energy weapon in the aerospace field, both objects are moving and it is the relative velocity that is important. In the case of the interception of a reentry vehicle (RV) from an intercontinental ballistic missile (ICBM) during the terminal phase of the approach, the RV will be traveling at approximately while the interceptor will be on the order of . Because the interceptor may not be approaching head-on, a lower bound on the relative velocity on the order of can be assumed, or converti |
https://en.wikipedia.org/wiki/Mabel%20Gweneth%20Humphreys | Mabel Gweneth Humphreys was a Canadian-American mathematician and Professor of Mathematics at Randolph-Macon Woman's College.
The M. Gweneth Humphreys Award of the Association for Women in Mathematics was established in her honor.
Education
Humphreys attended North Vancouver High School from 1925 to 1928.
She received her Bachelor of Arts with honors in mathematics from the University of British Columbia in 1932, where she held scholarships for all four years.
She studied at Smith College under Neil McCoy, Susan Miller Rambo, and Ruth G. Wood, and she received a master's degree in mathematics in 1933.
She received her Ph.D. in mathematics from the University of Chicago in 1935.
Her dissertation was entitled On the Waring Problem with Polynomial Summands and her advisor was Leonard Eugene Dickson.
Career
In 1981, Humphreys described her first attempts to find a job after completing her Ph.D.:
From 1935 to 1936, Humphreys was an instructor of mathematics and physics at Mount St. Scholastica College.
She began teaching at H. Sophie Newcomb Memorial College in 1936 and was promoted to assistant professor in 1941.
She was also an assistant professor at Barnard College in the summer of 1944, and an assistant professor at Tulane University in the summer of 1946.
In 1949, Humphreys became an associate professor at Randolph-Macon Woman's College.
After one year at Randolph-Macon, she was named Gillie A. Larew Professor and head of the mathematics department.
She was head of the department until 1979.
For the 1955-1956 academic year, Humphreys went on sabbatical leave to the University of British Columbia (UBC).
During this time, she visited undergraduate mathematics programs at several colleges and universities to examine their methods.
From 1962 to 1963 she was a visiting professor at UBC as a National Science Foundation (NSF) faculty fellow.
In the summers, Humphreys taught high school teachers at NSF summer institutes.
From 1965 to 1969, Humphreys worked for the E |
https://en.wikipedia.org/wiki/Multiple%20scattering%20theory | Multiple scattering theory (MST) is the mathematical formalism that is used to describe the propagation of a wave through a collection of scatterers. Examples are acoustical waves traveling through porous media, light scattering from water droplets in a cloud, or x-rays scattering from a crystal. A more recent application is to the propagation of quantum matter waves like electrons or neutrons through a solid.
As pointed out by Jan Korringa, the origin of this theory can be traced back to an 1892 paper by Lord Rayleigh. An important mathematical formulation of the theory was made by Paul Peter Ewald. Korringa and Ewald acknowledged the influence on their work of the 1903 doctoral dissertation of Nikolai Kasterin, portions of which were published in German in the Proceedings of the Royal Academy of Sciences in Amsterdam under the sponsorship of Heike Kamerlingh Onnes. The MST formalism is widely used for electronic structure calculations as well as diffraction theory, and is the subject of many books.
The multiple-scattering approach is the best way to derive one-electron Green's functions. These functions differ from the Green's functions used to treat the many-body problem, but they are the best starting point for calculations of the electronic structure of condensed matter systems that cannot be treated with band theory.
The terms "multiple scattering" and "multiple scattering theory" are often used in other contexts. For example, Molière's theory of the scattering of fast charged particles in matter is described in that way.
Mathematical formulation
The MST equations can be derived with different wave equations, but one of the simplest and most useful ones is the Schrödinger equation for an electron moving in a solid. With the help of density functional theory, this problem can be reduced to the solution of a one-electron equation
where the effective one-electron potential, , is a functional of the density of the electrons in the system.
In the Dirac notat |
https://en.wikipedia.org/wiki/Animal%20studies | Animal studies is a recently recognised field in which animals are studied in a variety of cross-disciplinary ways. Scholars who engage in animal studies may be formally trained in a number of diverse fields, including art history, anthropology, biology, film studies, geography, history, psychology, literary studies, museology, philosophy, communication, and sociology. They engage with questions about notions of "animality," "animalization," or "becoming animal," to understand human-made representations of and cultural ideas about "the animal" and what it is to be human by employing various theoretical perspectives. Using these perspectives, those who engage in animal studies seek to understand both human-animal relations now and in the past as defined by our knowledge of them. Because the field is still developing, scholars and others have some freedom to define their own criteria about what issues may structure the field.
History
Animal studies became popular in the 1970s as an interdisciplinary subject, animal studies exists at the intersection of a number of different fields of study such as journals and books series, etc. Different fields began to turn to animals as an important topic at different times and for various reasons, and these separate disciplinary histories shape how scholars approach animal studies. Historically, the field of environmental history has encouraged attention to animals.
Throughout Western history, humankind has put itself above the "nonhuman species." In part, animal studies developed out of the animal liberation movement and was grounded in ethical questions about co-existence with other species: whether it is moral to eat animals, to do scientific research on animals for human benefit, and so on. Take rats, for example, with a history of being used as “an experimental subject, feeder, and “pest.” Animal studies scholars who explore the field from an ethical perspective frequently cite Australian philosopher Peter Singer's 1975 wor |
https://en.wikipedia.org/wiki/Business%20continuity%20planning | Business continuity may be defined as "the capability of an organization to continue the delivery of products or services at pre-defined acceptable levels following a disruptive incident", and business continuity planning (or business continuity and resiliency planning) is the process of creating systems of prevention and recovery to deal with potential threats to a company. In addition to prevention, the goal is to enable ongoing operations before and during execution of disaster recovery. Business continuity is the intended outcome of proper execution of both business continuity planning and disaster recovery.
Several business continuity standards have been published by various standards bodies to assist in checklisting ongoing planning tasks.
An organization's resistance to failure is "the ability ... to withstand changes in its environment and still function". Often called resilience, it is a capability that enables organizations to either endure environmental changes without having to permanently adapt, or the organization is forced to adapt a new way of working that better suits the new environmental conditions.
Overview
Any event that could negatively impact operations should be included in the plan, such as supply chain interruption, loss of or damage to critical infrastructure (major machinery or computing/network resource). As such, BCP is a subset of risk management. In the U.S., government entities refer to the process as continuity of operations planning (COOP). A business continuity plan outlines a range of disaster scenarios and the steps the business will take in any particular scenario to return to regular trade. BCP's are written ahead of time and can also include precautions to be put in place. Usually created with the input of key staff as well as stakeholders, a BCP is a set of contingencies to minimize potential harm to businesses during adverse scenarios.
Resilience
A 2005 analysis of how disruptions can adversely affect the operations of |
https://en.wikipedia.org/wiki/Lane%20W.%20Martin | Lane Wyatt Martin is an American chemical engineer. He is a professor in the department of materials science and engineering at the University of California, Berkeley.
Early life and education
Martin was born and raised in rural western Pennsylvania. He chose to enroll at Carnegie Mellon University for his undergraduate degree in business but eventually switched to material science. Following this, he completed his master's degree in 2006 and PhD in 2008 from the University of California, Berkeley.
Career
Following his PhD, Martin served as a postdoctoral fellow in the quantum materials program at Lawrence Berkeley National Laboratory before accepting a faculty position at the University of Illinois Urbana-Champaign. As an assistant professor of materials science and engineering, Martin received a National Science Foundation CAREER Award for his proposal, "Enhanced Pyroelectric and Electrocaloric Effects in Complex Oxide Thin Film Heterostructures." He also helped devise a method to make thin films of ferroelectric material with twice the strain of traditional methods, giving the films exceptional electric properties. In 2013, Martin was nominated for a Presidential Early Career Award for Scientists and Engineers by the United States Department of Defense "for his research accomplishments in the synthesis and study of multifunctional materials that have enabled the development and understanding of fundamentally new materials phenomena and potential for advanced devices."
In 2014, Martin returned to his alma mater, the University of California, Berkeley, as a faculty member in their department of materials science and engineering. As an associate professor of materials science and engineering, Martin oversaw a research team that found a way to control the movement and placement of electrons in graphene. While serving in this role, he received the 2015 American Associate for Crystal Growth Young Author Award for his "outstanding accomplishments in the heteroepitaxi |
https://en.wikipedia.org/wiki/Delphi%20%28online%20service%29 | Delphi Forums is a U.S. online service provider and since the mid-1990s has been a community internet forum site. It started as a nationwide dialup service in 1983. Delphi Forums remains active as of 2023.
History
The company that became Delphi was founded by Wes Kussmaul as Kussmaul Encyclopedia in 1981 and featured an encyclopedia, e-mail, and a primitive chat. Newswires, bulletin boards and better chat were added in early 1982.
Kussmaul recalled:
Delphi was actually launched in October 1981, at Jerry Milden's Northeast Computer Show, as the Kussmaul Encyclopedia--the world's first commercially available computerized encyclopedia. (Frank Greenagle's Arête Encyclopedia was announced at about the same time, but you couldn't buy it until much later.) The Kussmaul Encyclopedia was actually a complete home computer system (your choice of Tandy Color Computer or Apple II) with a 300-bps modem that dialed up to a VAX computer hosting our online encyclopedia database. We sold the system for about the same price and terms as Britannica. People wandered around in it and were impressed with the ease with which they could find information. We had a wonderful cross-referencing system that turned every occurrence of a word that was the name of an entry in the encyclopedia into a hypertext link—in 1981...
In November 1982, Wes hired Glenn McIntyre as a software engineer primarily doing internal systems. Glenn brought in colleagues Kip Bryan and Dan Bruns. Kip wrote the software that became Delphi Conference and Delphi Forums. Dan upon finishing his MBA at Harvard, become President and subsequently CEO when Wes moved on to form Global Villages.
On March 15, 1983, the Delphi name was first used by General Videotex Corporation. Forums were text-based, and accessed via Telenet, Sprintnet, Tymnet, Uninet, and Datapac. In 1984, it had 4 million members.
Delphi was extended to Argentina in 1985, through a partnership with the Argentine IT company Siscotel S.A.
Delphi partnered |
https://en.wikipedia.org/wiki/Micropia | Micropia is the name of a family of LTR retrotransposons widespread in the genomes of fruitflies of the genus Drosophila. Micropia retrotransposons in some species of Drosophila express a male germline-specific and meiotic-specific antisense transcript complementary to the reverse transcriptase (RT) and ribonuclease A (RNaseA) genes of the proviral retrotransposon. No active transposition of micropia has been registered so far. micropia is likely part of a selfish driver system responsible for the Drosophila Y chromosomal lampbrushloop evolution in some species.
Micropia was discovered after micro-cloning experiments carried out on Y-chromosomal lampbrush loops by Prof. Wolfgang Hennig. Similar loops can be found in lampbrush chromosomes (see Lampbrush chromosome) that are characteristic for the female germ cells of most animals, except mammals. The name micropia is an artificial word, i.e. a concoction of "microcloning" and "copia-like element". |
https://en.wikipedia.org/wiki/Leucopelargonidin | Leucopelargonidin is a colorless chemical compound related to leucoanthocyanins. It can be found in Albizia lebbeck (East Indian walnut), in the fruit of Anacardium occidentale (Cashew), in the fruit of Areca catechu (Areca nut), in the fruit of Hydnocarpus wightiana (Hindi Chaulmoogra), in the rhizome of Rumex hymenosepalus (Arizona dock), in Zea mays (Corn) and in Ziziphus jujuba (Chinese date).
(+)-Leucopelargonidin can be synthesized from (+)-aromadendrin by sodium borohydride reduction.
Metabolism
Dihydrokaempferol 4-reductase uses cis-3,4-leucopelargonidin and NADP+ to produce (+)-aromadendrin, NADPH, and H+.
Leucoanthocyanidin reductase transforms cis-3,4-leucopelargonidin into afzelechin. |
https://en.wikipedia.org/wiki/Microsoft%20Security%20Essentials | Microsoft Security Essentials (MSE) is an antivirus software (AV) product that provides protection against different types of malicious software, such as computer viruses, spyware, rootkits, and Trojan horses. Prior to version 4.5, MSE ran on , Windows Vista, and Windows 7, but not on Windows 8 and later versions, which have built-in AV components known as Windows Defender. MSE 4.5 and later versions do not run on Windows XP. The license agreement allows home users and small businesses to install and use the product free of charge. It replaces Windows Live OneCare, a discontinued commercial subscription-based AV service, and the free Windows Defender, which only protected users from spyware until Windows8.
Built upon the same scanning engine and virus definitions as other Microsoft antivirus products, it provides real-time protection, constantly monitoring activities on the computer, scanning new files as they are created or downloaded, and disabling detected threats. It lacks the OneCare personal firewall and the Forefront Endpoint Protection centralized management features.
Microsoft's announcement of its own AV software on 18 November 2008, was met with mixed reactions from the AV industry. Symantec, McAfee, and Kaspersky Lab—three competing independent software vendors—dismissed it as an unworthy competitor, but AVG Technologies and Avast Software appreciated its potential to expand consumers' choices of AV software. AVG, McAfee, Sophos, and Trend Micro claimed that the integration of the product into Microsoft Windows would be a violation of competition law.
The product received generally positive reviews, praising its user interface, low resource usage, and freeware license. It secured AV-TEST certification in October 2009, having demonstrated its ability to eliminate all widely encountered malware. It lost that certification in October 2012; in June 2013, MSE achieved the lowest possible protection score, zero. However, Microsoft significantly improved thi |
https://en.wikipedia.org/wiki/Microcin | Microcins are very small bacteriocins, composed of relatively few amino acids. For this reason, they are distinct from their larger protein cousins. The classic example is microcin V, of Escherichia coli. Subtilosin A is another bacteriocin from Bacillus subtilis. The peptide has a cyclized backbone and forms three cross-links between the sulphurs of Cys13, Cys7 and Cys4 and the alpha-positions of Phe22, Thr28 and Phe31.
Microcins produced by commensal E. coli strains target and eliminate enteric pathogens such as Salmonella enterica by mimicking the siderophores the pathogens use for iron scavenging. Microcins also help commensal strains of E. coli outcompete pathogenic strains.
BACTIBASE database is an open-access database for bacteriocins including microcins. |
https://en.wikipedia.org/wiki/German%20Parliamentary%20Committee%20investigation%20of%20the%20NSA%20spying%20scandal | The German Parliamentary Committee investigation of the NSA spying scandal (official title: 1. Untersuchungsausschuss „NSA“) was started on March 20, 2014, by the German Parliament in order to investigate the extent and background of foreign secret services spying in Germany in the light of the Global surveillance disclosures (2013–present).
The Committee is also in search of strategies on how to protect telecommunication with technical means.
Members
The committee is formed by eight members of the German Parliament. The parliamentarian of the Christian Democratic Union (CDU) Clemens Binninger was head of the committee but stepped down after six days. In a statement, Binninger clarified that the other committee members had insisted on inviting Edward Snowden to testify; Binninger objected to this and resigned in protest. Patrick Sensburg (CDU) succeeded him.
Witnesses
Discussion about the witness Snowden
On May 8, 2014, the committee unanimously decided to let US whistleblower Edward Snowden testify as a witness.
On September 23, 2014, the Green Party and The Left filed a constitutional complaint against the Christian Democratic Union, the Social Democrats and the NSA Parliamentary Committee because of the Christian Democrats' and the Social Democrats' refusal to let the witness Edward Snowden testify in Berlin. The accused proposed a video conference from Moscow which Snowden had refused.
On September 28, 2014, the Green Party and The Left filed a constitutional complaint against German chancellor Merkel. According to them, she refuses to comply with her duty according to Chapter 44 of the German constitution to ensure a real investigation; especially by refusing to ensure the legal requirements to allow the witness Edward Snowden to testify.
Testimony of Binney and Drake
On July 3, 2014, the former Technical Director of the NSA, William Binney, who had become a whistleblower after the terrorist attacks of September 11, 2001, testified to the committee. H |
https://en.wikipedia.org/wiki/Census%20of%20Marine%20Zooplankton | The Census of Marine Zooplankton is a field project of the Census of Marine Life that has aimed to produce a global assessment of the species diversity, biomass, biogeographic distribution, and genetic diversity of more than 7,000 described species of zooplankton that drift the ocean currents throughout their lives. CMarZ focuses on the deep sea, under-sampled regions, and biodiversity hotspots. From 2004 until 2011, Ann Bucklin was the lead scientist for the project.
Technology plays a great role in CMarZ's research, including the use of integrated morphological and molecular sampling through DNA Barcoding. CMarZ makes its datasets available via the CMarZ Database. |
https://en.wikipedia.org/wiki/Halorubrum%20salsolis | "Halorubrum salsolis" is an undescribed species of halobacteria which is known to live in the Great Salt Lake in the United States.
The microbe was named by two children who took part in a naming contest held by the discoverers of the organism in 2006; the children independently suggested salsolis as the species name for the microbe.
This halophilic extremophile lives in water 10 times saltier than the ocean. It contains carotenoids that make it resistant to ultraviolet rays. |
https://en.wikipedia.org/wiki/Annihilation%20radiation | Annihilation radiation is a term used in Gamma spectroscopy for the photon radiation produced when a particle and its antiparticle collide and annihilate. Most commonly, this refers to 511-keV photons produced by an electron interacting with a positron. These photons are frequently referred to as gamma rays, despite having their origin outside the nucleus, due to unclear distinctions between types of photon radiation. Positively charged electrons (Positrons) are emitted from the nucleus as it undergoes β+ decay. The positron travels a short distance (a few millimeters), depositing any excess energy before it combines with a free electron. The mass of the e- and e+ is completely converted into two photons with an energy of 511 keV each. These annihilation photons are emitted in opposite directions, 180˚ apart. This is the basis for PET scanners in a process called coincidence counting.
Annihilation radiation is not monoenergetic, unlike gamma rays produced by radioactive decay. The production mechanism of annihilation radiation introduces Doppler broadening. The annihilation peak produced in a photon spectrum by annihilation radiation therefore has a higher full width at half maximum (FWHM) than decay-generated gamma rays in spectrum. The difference is more apparent with high resolution detectors, such as Germanium detectors, than with low resolution detectors such as Sodium iodide detectors.
Because of their well-defined energy (511 keV) and characteristic, Doppler-broadened shape, annihilation radiation can often be useful in defining the energy calibration of a gamma ray spectrum. |
https://en.wikipedia.org/wiki/GARUDA | GARUDA(Global Access to Resource Using Distributed Architecture) is India's Grid Computing initiative connecting 17 cities across the country. The 45 participating institutes in this nationwide project include all the IITs and C-DAC centers and other major institutes in India.
GARUDA is a collaboration of science researchers and experimenters on a nationwide grid of computational nodes, mass storage and scientific instruments that aims to provide the technological advances required to enable data and compute intensive science for the 21st century. One of GARUDA's most important challenges is to strike the right balance between research and the daunting task of deploying that innovation into some of the most complex scientific and engineering endeavors being undertaken today.
The Department of Information Technology (DIT), Government of India has funded the Centre for Development of Advanced Computing (C-DAC) to deploy the nationwide computational grid GARUDA. In Proof of Concept (PoC) phase which ended in March 2008, 17 cities across the country were connected with an aim to bring “Grid” networked computing to research labs and industry. From April 2008 the Foundation phase is in progress with an aim to include more users’ applications, providing Service Oriented architecture, improving network stability and upgrading grid resources. GARUDA will assist to accelerate India's drive to turn its substantial research investment into tangible economic benefits.
The Main Monitoring Centre also called the Garuda Monitoring and Management Centre is set up at C-DAC Knowledge Park, Bangalore. From this point, the whole grid which has now extended even into Europe is Monitored and Managed by C-DAC's young scientists. In India, GARUDA uses National Knowledge Network as network backbone.
Grid Middleware
GARUDA has adopted a pragmatic approach for using existing Grid infrastructure and Web Services technologies. The deployment of grid tools and services for GARUDA will be base |
https://en.wikipedia.org/wiki/GNU%20Libtool | In computer programming, GNU Libtool is a software development tool, part of the GNU build system, consisting of a shell script created to address the software portability problem when compiling shared libraries from source code.
It hides the differences between computing platforms for the commands which compile shared libraries.
It provides a command-line interface that is identical across platforms and it executes the platform's native commands.
Rationale
Different operating systems handle shared libraries differently.
Some platforms do not use shared libraries at all.
It can be difficult to make a software program portable: the C compiler differs from system to system; certain library functions are missing on some systems; header files may have different names.
Libtool helps manage the creation of static and dynamic libraries on various Unix-like operating systems.
Libtool accomplishes this by abstracting the library-creation process, hiding differences between various systems (e.g. Linux systems vs. Solaris).
GNU Libtool is designed to simplify the process of compiling a computer program on a new system, by "encapsulating both the platform-specific dependencies, and the user interface, in a single script".
When porting a program to a new system, Libtool is designed so the porter need not read low-level documentation for the shared libraries to be built, rather just run a configure script (or equivalent).
Use
Libtool is used by Autoconf and Automake, two other portability tools in the GNU build system.
It can also be used directly.
Clones and derivatives
Since GNU Libtool was released, other free software projects have created drop-in replacements under different software licenses.
See also
GNU Compiler Collection
GNU build system
pkg-config |
https://en.wikipedia.org/wiki/Reactances%20of%20synchronous%20machines | The reactances of synchronous machines comprise a set of characteristic constants used in the theory of synchronous machines. Technically, these constants are specified in units of the electrical reactance (ohms), although they are typically expressed in the per-unit system and thus dimensionless. Since for practically all (except for the tiniest) machines the resistance of the coils is negligibly small in comparison to the reactance, the latter can be used instead of (complex) electrical impedance, simplifying the calculations.
Two reactions theory
The air gap of the machines with a salient pole rotor is quite different along the pole axis (so called direct axis) and in the orthogonal direction (so called quadrature axis). Andre Blondel in 1899 proposed in his paper "Empirical Theory of Synchronous Generators" the two reactions theory that divided the armature magnetomotive force (MMF) into two components: the direct axis component and the quadrature axis component. The direct axis component is aligned with the magnetic axis of the rotor, while the quadrature (or transverse) axis component is perpendicular to the direct axis. The relative strengths of these two components depend on the design of the machine and the operating conditions. Since the equations naturally split into direct and quadrature components, many reactances come in pairs, one for the direct axis (with the index d), one for the quadrature axis (with the index q). In the machines with a cylindrical rotor the air gap is uniform, the reactances along the d and q axes are equal, and d/q indices are frequently dropped.
States of the generator
The flux linkages of the generator vary with its state. Three states are considered:
the steady-state is the normal operating condition with the armature magnetic flux going through the rotor;
the sub-transient state is the one the generator enters immediately after the fault (short circuit). In this state the armature flux is pushed completely out of the r |
https://en.wikipedia.org/wiki/Fenamidone | Fenamidone is a foliar fungicide used on grapes, ornamentals, potatoes, tobacco, and vegetables such as tomatoes. It exerts its fungicidal effects by acting as a Qo inhibitor. |
https://en.wikipedia.org/wiki/Crystallinity | Crystallinity refers to the degree of structural order in a solid. In a crystal, the atoms or molecules are arranged in a regular, periodic manner. The degree of crystallinity has a big influence on hardness, density, transparency and diffusion. In an ideal gas, the relative positions of the atoms or molecules are completely random. Amorphous materials, such as liquids and glasses, represent an intermediate case, having order over short distances (a few atomic or molecular spacings) but not over longer distances.
Many materials, such as glass-ceramics and some polymers, can be prepared in such a way as to produce a mixture of crystalline and amorphous regions. In such cases, crystallinity is usually specified as a percentage of the volume of the material that is crystalline. Even within materials that are completely crystalline, however, the degree of structural perfection can vary. For instance, most metallic alloys are crystalline, but they usually comprise many independent crystalline regions (grains or crystallites) in various orientations separated by grain boundaries; furthermore, they contain other crystallographic defects (notably dislocations) that reduce the degree of structural perfection. The most highly perfect crystals are silicon boules produced for semiconductor electronics; these are large single crystals (so they have no grain boundaries), are nearly free of dislocations, and have precisely controlled concentrations of defect atoms.
Crystallinity can be measured using x-ray crystallography, but calorimetric techniques are also commonly used.
Rock crystallinity
Geologists describe four qualitative levels of crystallinity:
holocrystalline rocks are completely crystalline;
hypocrystalline rocks are partially crystalline, with crystals embedded in an amorphous or glassy matrix;
hypohyaline rocks are partially glassy;
holohyaline rocks (such as obsidian) are completely glassy. |
https://en.wikipedia.org/wiki/Aromat | Aromat is a food seasoning, invented in Switzerland by Walter Obrist for Knorr Thayngen, the swiss branch of the German food company Knorr, in 1952. Aromat was originally called "", which means plant extract in German. Knorr dropped the name in 1953 and altered its form, from cubes to a powdered seasoning. Aromat was marketed internationally and is described by Knorr as an "all purpose savoury seasoning".
The ingredients in Aromat vary by market, but include the flavour enhancer monosodium glutamate, and may also comprise yeast extract, wheat or corn flour, trans fat (partially hydrogenated vegetable oil), and various herbs, spices, vegetable extracts and other flavourings.
See also
Bouillon cube
List of brand name condiments |
https://en.wikipedia.org/wiki/International%20System%20of%20Electrical%20and%20Magnetic%20Units | The International System of Electrical and Magnetic Units is an obsolete system of units used for measuring electrical and magnetic quantities. It was proposed as a system of practical international units (e.g., the international ampere, the international ohm, the international volt) by unanimous recommendation at the International Electrical Congress (Chicago, 1893), discussed at other Congresses, and finally adopted at the International Conference on Electric Units and Standards in London in 1908. It was rendered obsolete by the inclusion of electromagnetic units in the International System of Units (SI) at the 9th General Conference on Weights and Measures in 1948.
Earlier systems
The link between electromagnetic units and the more familiar units of length, mass and time was first demonstrated by Carl Friedrich Gauss in 1832 with his measurement of the Earth's magnetic field, and the principle was extended to electrical measurements by Franz Ernst Neumann in 1845. A complete system of metric electrical and magnetic units was proposed by Wilhelm Eduard Weber in 1851, based on the idea that electrical units could be defined solely in relation to absolute units of length, mass, and time. Weber's original proposal was based on a millimetre–milligram–second system of units.
The development of the electric telegraph (an invention of Gauss and Weber) demonstrated the need for accurate electrical measurements. At the behest of William Thomson, the British Association for the Advancement of Science (B.A.) set up a committee in 1861, initially to examine standards for electrical resistance, which was expanded in 1862 to include other electrical standards. After two years of discussion, experiment and considerable differences of opinion, the committee decided to adapt Weber's approach to the CGS system of units, but used metre, gramme and second as their absolute units. However these units were both difficult to realize and (often) impractically small. To overcome these h |
https://en.wikipedia.org/wiki/MT-TS2 | Mitochondrially encoded tRNA serine 2 (AGU/C) also known as MT-TS2 is a transfer RNA which in humans is encoded by the mitochondrial MT-TS2 gene.
MT-TS2 is a small 59 nucleotide RNA (human mitochondrial map position 12207-12265) that transfers the amino acid serine to a growing polypeptide chain at the ribosome site of protein synthesis during translation. |
https://en.wikipedia.org/wiki/Runs%20of%20homozygosity | Runs of Homozygosity (ROH) are contiguous lengths of homozygous genotypes that are present in an individual due to parents transmitting identical haplotypes to their offspring.
The potential of predicting or estimating individual autozygosity for a subpopulation is the proportion of the autosomal genome above a specified length, termed Froh.
This technique can be used to identify the genomic footprint of inbreeding in conservation programs, as organisms that have undergone recent inbreeding will exhibit long runs of homozygosity. For example, the step-wise reintroduction strategy of the Alpine Ibex in the Swiss Alps created several strong population bottlenecks that reduced the genetic diversity of the newly introduced individuals. The effect of inbreeding in the resulting sub-populations could be studied by measuring the runs of homozygosity in different individuals |
https://en.wikipedia.org/wiki/Designated%20verifier%20signature | A designated verifier signature is a signature scheme in which signatures can only be verified by a single, designated verifier, designated as part of the signature creation. Designated verifier signatures were first proposed in 1996 by Jakobsson Markus, Kazue Sako, and Russell Impagliazzo. Proposed as a way to combine authentication and off-the-record messages, designated verifier signatures allow authenticated, private conversations to take place.
Unlike in undeniable signature scheme the protocol of verifying is non-interactive; i.e., the signer chooses the designated verifier (or the set of designated verifiers) in advance and does not take part in the verification process.
See also
Non-repudiation
Undeniable signature |
https://en.wikipedia.org/wiki/Packet%20switching | In telecommunications, packet switching is a method of grouping data into packets that are transmitted over a digital network. Packets are made of a header and a payload. Data in the header is used by networking hardware to direct the packet to its destination, where the payload is extracted and used by an operating system, application software, or higher layer protocols. Packet switching is the primary basis for data communications in computer networks worldwide.
During the early 1960s, Polish-American engineer Paul Baran developed a concept he called "distributed adaptive message block switching", with the goal of providing a fault-tolerant, efficient routing method for telecommunication messages as part of a research program at the RAND Corporation, funded by the United States Department of Defense. His ideas contradicted then-established principles of pre-allocation of network bandwidth, exemplified by the development of telecommunications in the Bell System. The new concept found little resonance among network implementers until the independent work of British computer scientist Donald Davies at the National Physical Laboratory in 1965. Davies coined the modern term packet switching and inspired numerous packet switching networks in the decade following, including the incorporation of the concept into the design of the ARPANET in the United States and the CYCLADES network in France. The ARPANET and CYCLADES were the primary precursor networks of the modern Internet.
Concept
A simple definition of packet switching is:
Packet switching allows delivery of variable bit rate data streams, realized as sequences of packets, over a computer network which allocates transmission resources as needed using statistical multiplexing or dynamic bandwidth allocation techniques. As they traverse networking hardware, such as switches and routers, packets are received, buffered, queued, and retransmitted (stored and forwarded), resulting in variable latency and throughput de |
https://en.wikipedia.org/wiki/Blue%20Oak%20Ranch%20Reserve | The Blue Oak Ranch Reserve, a unit of the University of California Natural Reserve System, is an ecological reserve and biological field station in Santa Clara County, California. It is located on in the Diablo Range, northwest of Mount Hamilton, at elevation.
The land, part of the 19th century Mexican land grant of Rancho Cañada de Pala, was donated to the University of California on December 1, 2007 by the Blue Oak Ranch Trust, an anonymous benefactor.
Overnight accommodations for academic researchers and educational groups may be made by permission only.
Flora
Flora of Blue Oak Ranch Reserve includes:
purple needle grass (Nassella pulchra)
barley (Hordeum)
bluegrass (Poa spp.)
three-awn (Aristida spp.)
melic (Melica sspspp
wildrye (Elymus and Leymus spp.)
Invasive introduced species
yellow starthistle (Centaurea solstitialis)
medusahead grass (Taeniatherum caput-medusae)
Italian thistle (Carduus pycnocephalus)
tocalote (Centaurea melitensis)
Canada thistle (Cirsium arvense)
bull thistle (Cirsium vulgare)
Plant communities
valley oak woodland
black oak woodland
coast live oak woodland
riparian forest
chamise chaparral
Diablan sage scrub
non-native annual grassland
wildflower field
native perennial grassland
Fauna
Fauna of Blue Oak Ranch Reserve includes:
Western toad
Pacific tree frog
Pacific chorus frog
red-legged frog
California newt
California tiger salamander
Western pond turtle
red-winged blackbird
pied-billed grebe
Canada goose
American coot
wood duck
Invasive fish
sunfish
largemouth bass
mosquitofish
See also
Quercus douglasii — Blue oak
California oak woodlands — a plant community within the ranch
California interior chaparral and woodlands — the plant community that the ranch is within
California chaparral and woodlands ecoregion — ecoregion of the Mediterranean forests, woodlands, and scrub Biome, that the ranch is within
List of California native plants
Restoration ecology |
https://en.wikipedia.org/wiki/Color%20reproduction | Color reproduction is an aspect of color science concerned with producing light spectra that evoke a desired color, either through additive (light emitting) or subtractive (surface color) models. It converts physical correlates of color perception (CIE 1931 XYZ color space tristimulus values and related quantities) into light spectra that can be experienced by observers. In this way, it is the opposite of colorimetry.
It is concerned with the faithful reproduction of a color in one medium, with a color in another, so it is a central concept in color management and relies heavily on color calibration. For example, food packaging must be able to faithfully reproduce the colors of the foods therein in order to appeal to a customer. This involves proper color calibration of at least four devices:
Lighting, which must have a high color rendering index and not give a color cast to the object.
Camera, which measures the reflected spectrum of the object and converts to a trichromatic color space (e.g. RGB).
Screen, which reproduces color so a designer can proof the captured image and make color corrections as necessary.
Printer, which reproduces the final color on paper. |
https://en.wikipedia.org/wiki/Super%2035 | Super 35 (originally known as Superscope 235) is a motion picture film format that uses exactly the same film stock as standard 35 mm film, but puts a larger image frame on that stock by using the space normally reserved for the optical analog sound track.
History
Super 35 was revived from a similar Superscope variant known as Superscope 235, which was originally developed by the Tushinsky Brothers (who founded Superscope Inc. in 1954) for RKO in 1954. The first film to be shot in Superscope was Vera Cruz, a western film produced by Hecht-Lancaster Productions and distributed through United Artists.
When cameraman Joe Dunton was preparing to shoot Dance Craze in 1980, he chose to revive the Superscope format by using a full silent-standard gate and slightly optically recentering the lens port (to adjust for the inclusion of the area of the optic soundtrack -the gray track on left side of the illustration). These two characteristics are central to the format.
It was adopted by Hollywood starting with Greystoke in 1984, under the format name Super Techniscope. It also received much early publicity for making the cockpit shots in Top Gun possible, since it was otherwise impossible to fit 35 mm cameras with large anamorphic lenses into the small free space in the cockpit. Later, as other camera rental houses and labs started to embrace the format, Super 35 became popular in the mid-1990s, and is now considered a ubiquitous production process, with usage on well over a thousand feature films. It is often the standard production format for television shows, music videos, and commercials. Since none of these require a release print, it is unnecessary to reserve space for an optical soundtrack.
When composing for 1.85:1, it was known as Super 1.85, since it was larger than standard 1.85.
When composing for 2.39:1, it was often typical to employ either a "common center", which keeps the 2.39 extraction area at the center of the film that results in extra headroom if o |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.