source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/It%27s%20a%20Bird... | It's a Bird... is a 2004 graphic novel released by DC Comics under its mature-readers Vertigo imprint. Written by Steven T. Seagle, with art by Teddy Kristiansen, it is an autobiographical book that chronicles Seagle's thoughts as he tries to work out a new approach in one of the world's most popular characters, Superman while dealing with the "family secret" which has come back to the forefront.
Plot
In a flashback, five year old Steven and his little brother, Dave, are in a hospital where their grandmother soon passes away. Adult Steven narrates how he always thinks about that time, what was said about his grandmother, and about what was muttered concerning the "family secret" between his parents and the doctors. He also mentions how puzzled he was when he saw the Superman logo as one of the letters on a doctor's examination sheet. This was when he first read a Superman comic to his little brother; this would set up his path into being a comic book writer, but needless to say, he is not a fan of the Man of Steel.
In the present, as Steven finishes his current comic book project, his agent calls offering him a job to write Superman. Steven refuses, because he doesn't get the character, but his agent asks him to think about it. Steven narrates one thing he doesn't get about Superman, starting with the costume: he doesn't get why would anyone would wear that, and it makes him think of a kid from his school who wore to escape his outcast life for Halloween, only to wear it in the days after and get bullied for it. He meets with his girlfriend, Lisa, who mentions his mother called her looking for him fearfully. He doesn't tell her the details after he calls her, because he is afraid to talk about the "family secret" that claimed his grandmother: Huntington's disease. It's also what has made him afraid of having children, despite being told his father doesn't have it, so he and his brother won't. Steven discusses the idea of the outsider, someone in his words people w |
https://en.wikipedia.org/wiki/TRPV1 | The transient receptor potential cation channel subfamily V member 1 (TRPV1), also known as the capsaicin receptor and the vanilloid receptor 1, is a protein that, in humans, is encoded by the TRPV1 gene. It was the first isolated member of the transient receptor potential vanilloid receptor proteins that in turn are a sub-family of the transient receptor potential protein group. This protein is a member of the TRPV group of transient receptor potential family of ion channels. Fatty acid metabolites with affinity for this receptor are produced by cyanobacteria, which diverged from eukaryotes at least 2000 million years ago (MYA).
The function of TRPV1 is detection and regulation of body temperature. In addition, TRPV1 provides a sensation of scalding heat and pain (nociception). In primary afferent sensory neurons, it cooperates with TRPA1 (a chemical irritant receptor) to mediate the detection of noxious environmental stimuli.
Function
TRPV1 is an element of or mechanism used by the mammalian somatosensory system. It is a nonselective cation channel that may be activated by a wide variety of exogenous and endogenous physical and chemical stimuli. The best-known activators of TRPV1 are: temperature greater than ; acidic conditions; capsaicin (the irritating compound in hot chili peppers); and allyl isothiocyanate, the pungent compound in mustard and wasabi. The activation of TRPV1 leads to a painful, burning sensation. Its endogenous activators include: low pH (acidic conditions), the endocannabinoid anandamide, N-oleyl-dopamine, and N-arachidonoyl-dopamine. TRPV1 receptors are found mainly in the nociceptive neurons of the peripheral nervous system, but they have also been described in many other tissues, including the central nervous system. TRPV1 is involved in the transmission and modulation of pain (nociception), as well as the integration of diverse painful stimuli.
Sensitization
The sensitivity of TRPV1 to noxious stimuli, such as high temperatures, i |
https://en.wikipedia.org/wiki/The%20Physical%20Impossibility%20of%20Death%20in%20the%20Mind%20of%20Someone%20Living | The Physical Impossibility of Death in the Mind of Someone Living is an artwork created in 1991 by Damien Hirst, an English artist and a leading member of the "Young British Artists" (or YBA). It consists of a preserved tiger shark submerged in formaldehyde in a glass-panel display case. It was originally commissioned in 1991 by Charles Saatchi, who sold it in 2004 to Steven A. Cohen for an undisclosed amount, widely reported to have been at least $8 million. However, the title of Don Thompson's book, The $12 Million Stuffed Shark: The Curious Economics of Contemporary Art, suggests a higher figure.
Owing to deterioration of the original tiger shark, it was replaced with a new specimen in 2006. It was on loan to the Metropolitan Museum of Art in New York City from 2007 to 2010.
It is considered an iconic work of British art in the 1990s, and has become a symbol of Britart worldwide.
Background and concept
The work was funded by Charles Saatchi, who in 1991 had offered to pay for whatever artwork Hirst wanted to create. The shark itself cost Hirst £6,000 and the total cost of the work was £50,000. Hirst asked Doris Lockhart for a loan to cover the cost of shipping the shark from Australia, but she gifted him the required amount; in return, Hirst invited Lockhart to choose anything she liked from his studio, and she selected a piece called The Only Way is Up. The shark was caught off Hervey Bay in Queensland, Australia, by a fisherman commissioned to do so. Hirst wanted something "big enough to eat you".
It was first exhibited in 1992 in the first of a series of Young British Artists shows at the Saatchi Gallery, then at its premises in St John's Wood, north London. The British tabloid newspaper The Sun ran a story titled "£50,000 for fish without chips." The show also included Hirst's artwork A Thousand Years. He was then nominated for the Turner Prize, but it was awarded to Grenville Davey. Saatchi sold the work in 2004 to Steven A. Cohen for an estimated $8 m |
https://en.wikipedia.org/wiki/Convection%20%28heat%20transfer%29 | Convection (or convective heat transfer) is the transfer of heat from one place to another due to the movement of fluid. Although often discussed as a distinct method of heat transfer, convective heat transfer involves the combined processes of conduction (heat diffusion) and advection (heat transfer by bulk fluid flow). Convection is usually the dominant form of heat transfer in liquids and gases.
Note that this definition of convection is only applicable in Heat transfer and thermodynamic contexts. It should not be confused with the dynamic fluid phenomenon of convection, which is typically referred to as Natural Convection in thermodynamic contexts in order to distinguish the two.
Overview
Convection can be "forced" by movement of a fluid by means other than buoyancy forces (for example, a water pump in an automobile engine). Thermal expansion of fluids may also force convection. In other cases, natural buoyancy forces alone are entirely responsible for fluid motion when the fluid is heated, and this process is called "natural convection". An example is the draft in a chimney or around any fire. In natural convection, an increase in temperature produces a reduction in density, which in turn causes fluid motion due to pressures and forces when fluids of different densities are affected by gravity (or any g-force). For example, when water is heated on a stove, hot water from the bottom of the pan is displaced (or forced up) by the colder denser liquid, which falls. After heating has stopped, mixing and conduction from this natural convection eventually result in a nearly homogeneous density, and even temperature. Without the presence of gravity (or conditions that cause a g-force of any type), natural convection does not occur, and only forced-convection modes operate.
The convection heat transfer mode comprises one mechanism. In addition to energy transfer due to specific molecular motion (diffusion), energy is transferred by bulk, or macroscopic, motion of the |
https://en.wikipedia.org/wiki/Shq1 | Shq1p is a protein involved in the rRNA processing pathway. It was discovered by Pok Yang in the Chanfreau laboratory at UCLA. Depletion of Shq1p has led to decreased level of various H/ACA box snoRNAs (H/ACA box snoRNAs are responsible for pseuduridylation of pre-rRNA) and certain pre-rRNA intermediates.
Background
During the synthesis of eukaryotic ribosomes, four mature ribosomal RNAs (the 5S, 5.8S, 18S, and 25S) must be synthesized. Three of these rRNAs (5.8S, 18S, and 25S) come from a single pre-rRNA known as the 35S. Although many of the intermediates in this rRNA processing pathway have been identified in the last thirty years, there are still a number of proteins involved in this process whose specific function is unknown.
Function
Shq1, a protein thought to play a role in the stabilization and/or production of box H/ACA snoRNA, is still uncharacterized. It has been proposed that Shq1, along with Naf1p, is involved in the initial steps of the biogenesis of H/ACA box snoRNPs (box H/ACA snoRNAs form complexes with proteins, thereby forming snoRNPs) because of its association with certain snoRNP proteins during the snoRNP’s maturation, while showing very little association with the mature snoRNP complex. Despite the known involvement of Shq1 in H/ACA box snoRNP's production, the exact function of this protein in the overall rRNA processing pathway is still unknown.
See also
rRNA
snoRNA
Ribosomes
Eukaryotic translation
Proteins |
https://en.wikipedia.org/wiki/TOMNET | The TOMNET optimization Environment is a platform for solving applied optimization problems in Microsoft .NET. It makes it possible to use solvers like SNOPT, MINOS and CPLEX with one single model formulation. The solvers handle everything from linear programming and integer programming to global optimization.
External links
(home page)
Numerical software
Mathematical optimization software |
https://en.wikipedia.org/wiki/Amorphism | An amorphism, in chemistry, crystallography and, by extension, to other areas of the natural sciences is a substance or feature that lacks an ordered form. In the specific case of crystallography, an amorphic material is one that lacks long range (significant) crystalline order at the molecular level. In the history of chemistry, amorphism was recognised even before the discovery of the nature of the exact atomic crystalline lattice structure. The concept of amorphism can also be found in the fields of art, biology, archaeology and philosophy as a characterisation of objects without form, or with random or unstructured form.
Amorphous and Crystalline solid
In the context of solids, amorphous and crystalline are terms used to describe the structure of materials. Amorphous solids are the opposite of crystalline. The atoms or molecules in amorphous substances are arranged randomly without any long-range order. As a result, they do not have a sharp melting point. The phase transition from solid to liquid occurs over a range of temperatures. Some examples include glass, rubber and some plastics.
See also
Glass
Obsidian |
https://en.wikipedia.org/wiki/Tanaka%27s%20formula | In the stochastic calculus, Tanaka's formula for the Brownian motion states that
where Bt is the standard Brownian motion, sgn denotes the sign function
and Lt is its local time at 0 (the local time spent by B at 0 before time t) given by the L2-limit
One can also extend the formula to semimartingales.
Properties
Tanaka's formula is the explicit Doob–Meyer decomposition of the submartingale |Bt| into the martingale part (the integral on the right-hand side, which is a Brownian motion), and a continuous increasing process (local time). It can also be seen as the analogue of Itō's lemma for the (nonsmooth) absolute value function , with and ; see local time for a formal explanation of the Itō term.
Outline of proof
The function |x| is not C2 in x at x = 0, so we cannot apply Itō's formula directly. But if we approximate it near zero (i.e. in [−ε, ε]) by parabolas
and use Itō's formula, we can then take the limit as ε → 0, leading to Tanaka's formula. |
https://en.wikipedia.org/wiki/Minimal%20polynomial%20%28field%20theory%29 | In field theory, a branch of mathematics, the minimal polynomial of an element of a field extension is, roughly speaking, the polynomial of lowest degree having coefficients in the field, such that is a root of the polynomial. If the minimal polynomial of exists, it is unique. The coefficient of the highest-degree term in the polynomial is required to be 1.
More formally, a minimal polynomial is defined relative to a field extension and an element of the extension field . The minimal polynomial of an element, if it exists, is a member of , the ring of polynomials in the variable with coefficients in . Given an element of , let be the set of all polynomials in such that . The element is called a root or zero of each polynomial in
More specifically, Jα is the kernel of the ring homomorphism from F[x] to E which sends polynomials g to their value g(α) at the element α. Because it is the kernel of a ring homomorphism, Jα is an ideal of the polynomial ring F[x]: it is closed under polynomial addition and subtraction (hence containing the zero polynomial), as well as under multiplication by elements of F (which is scalar multiplication if F[x] is regarded as a vector space over F).
The zero polynomial, all of whose coefficients are 0, is in every since for all and . This makes the zero polynomial useless for classifying different values of into types, so it is excepted. If there are any non-zero polynomials in , i.e. if the latter is not the zero ideal, then is called an algebraic element over , and there exists a monic polynomial of least degree in . This is the minimal polynomial of with respect to . It is unique and irreducible over . If the zero polynomial is the only member of , then is called a transcendental element over and has no minimal polynomial with respect to .
Minimal polynomials are useful for constructing and analyzing field extensions. When is algebraic with minimal polynomial , the smallest field that contains both and is isom |
https://en.wikipedia.org/wiki/Minimal%20polynomial%20%28linear%20algebra%29 | In linear algebra, the minimal polynomial of an matrix over a field is the monic polynomial over of least degree such that . Any other polynomial with is a (polynomial) multiple of .
The following three statements are equivalent:
is a root of ,
is a root of the characteristic polynomial of ,
is an eigenvalue of matrix .
The multiplicity of a root of is the largest power such that strictly contains . In other words, increasing the exponent up to will give ever larger kernels, but further increasing the exponent beyond will just give the same kernel. Formally, is the nilpotent index of .
If the field is not algebraically closed, then the minimal and characteristic polynomials need not factor according to their roots (in ) alone, in other words they may have irreducible polynomial factors of degree greater than . For irreducible polynomials one has similar equivalences:
divides ,
divides ,
the kernel of has dimension at least .
the kernel of has dimension at least .
Like the characteristic polynomial, the minimal polynomial does not depend on the base field. In other words, considering the matrix as one with coefficients in a larger field does not change the minimal polynomial. The reason for this differs from the case with the characteristic polynomial (where it is immediate from the definition of determinants), namely by the fact that the minimal polynomial is determined by the relations of linear dependence between the powers of : extending the base field will not introduce any new such relations (nor of course will it remove existing ones).
The minimal polynomial is often the same as the characteristic polynomial, but not always. For example, if is a multiple of the identity matrix, then its minimal polynomial is since the kernel of is already the entire space; on the other hand its characteristic polynomial is (the only eigenvalue is , and the degree of the characteristic polynomial is always equal to the dimension of the |
https://en.wikipedia.org/wiki/Energy%20drift | In computer simulations of mechanical systems, energy drift is the gradual change in the total energy of a closed system over time. According to the laws of mechanics, the energy should be a constant of motion and should not change. However, in simulations the energy might fluctuate on a short time scale and increase or decrease on a very long time scale due to numerical integration artifacts that arise with the use of a finite time step Δt. This is somewhat similar to the flying ice cube problem, whereby numerical errors in handling equipartition of energy can change vibrational energy into translational energy.
More specifically, the energy tends to increase exponentially; its increase can be understood intuitively because each step introduces a small perturbation δv to the true velocity vtrue, which (if uncorrelated with v, which will be true for simple integration methods) results in a second-order increase in the energy
(The cross term in v · δv is zero because of no correlation.)
Energy drift - usually damping - is substantial for numerical integration schemes that are not symplectic, such as the Runge-Kutta family. Symplectic integrators usually used in molecular dynamics, such as the Verlet integrator family, exhibit increases in energy over very long time scales, though the error remains roughly constant. These integrators do not in fact reproduce the actual Hamiltonian mechanics of the system; instead, they reproduce a closely related "shadow" Hamiltonian whose value they conserve many orders of magnitude more closely. The accuracy of the energy conservation for the true Hamiltonian is dependent on the time step. The energy computed from the modified Hamiltonian of a symplectic integrator is from the true Hamiltonian.
Energy drift is similar to parametric resonance in that a finite, discrete timestepping scheme will result in nonphysical, limited sampling of motions with frequencies close to the frequency of velocity updates. Thus the restriction on |
https://en.wikipedia.org/wiki/Periodic%20boundary%20conditions | Periodic boundary conditions (PBCs) are a set of boundary conditions which are often chosen for approximating a large (infinite) system by using a small part called a unit cell. PBCs are often used in computer simulations and mathematical models. The topology of two-dimensional PBC is equal to that of a world map of some video games; the geometry of the unit cell satisfies perfect two-dimensional tiling, and when an object passes through one side of the unit cell, it re-appears on the opposite side with the same velocity. In topological terms, the space made by two-dimensional PBCs can be thought of as being mapped onto a torus (compactification). The large systems approximated by PBCs consist of an infinite number of unit cells. In computer simulations, one of these is the original simulation box, and others are copies called images. During the simulation, only the properties of the original simulation box need to be recorded and propagated. The minimum-image convention is a common form of PBC particle bookkeeping in which each individual particle in the simulation interacts with the closest image of the remaining particles in the system.
One example of periodic boundary conditions can be defined according to smooth real functions by
for all m = 0, 1, 2, ... and for constants and .
In molecular dynamics simulations and Monte Carlo molecular modeling, PBCs are usually applied to calculate properties of bulk gasses, liquids, crystals or mixtures. A common application uses PBC to simulate solvated macromolecules in a bath of explicit solvent. Born–von Karman boundary conditions are periodic boundary conditions for a special system.
In electromagnetics, PBC can be applied for different mesh types to analyze the electromagnetic properties of periodical structures.
Requirements and artifacts
Three-dimensional PBCs are useful for approximating the behavior of macro-scale systems of gases, liquids, and solids. Three-dimensional PBCs can also be used to simulate plan |
https://en.wikipedia.org/wiki/Legion%20%28taxonomy%29 | The legion, in biological classification, is a non-obligatory taxonomic rank within the Linnaean hierarchy sometimes used in zoology.
Taxonomic rank
In zoological taxonomy, the legion is:
subordinate to the class
superordinate to the cohort.
consists of a group of related orders
Legions may be grouped into superlegions or subdivided into sublegions, and these again into infralegions.
Use in zoology
Legions and their super/sub/infra groups have been employed in some classifications of birds and mammals. Full use is made of all of these (along with cohorts and supercohorts) in, for example, McKenna and Bell's classification of mammals.
See also
Linnaean taxonomy
Mammal classification |
https://en.wikipedia.org/wiki/International%20Journal%20of%20Theoretical%20Physics | The International Journal of Theoretical Physics is a peer-reviewed scientific journal of physics published by Springer Science+Business Media since 1968. According to the Journal Citation Reports, the journal has a June 2023 real-time impact factor of 2.6 and publishes both original research and review articles. The editor-in-chief is Andreas Wipf (Friedrich-Schiller-University Jena).
Scope and indexing
The journal covers the following areas: general relativity, quantum theory with relativistic quantum field theory, quantum measurement theory, quantum geometry and quantum logic. Services abstracting and indexing this journal include Chemical Abstracts Service, Mathematical Reviews, Science Citation Index, Scopus, and Zentralblatt MATH. |
https://en.wikipedia.org/wiki/Kinetic%20logic | Kinetic logic, developed by René Thomas, is a Qualitative Modeling approach feasible to model impact, feedback, and the temporal evolution of the variables. It uses symbolic descriptions and avoids continuous descriptions e.g. differential equations.The derivation of the dynamics from the interaction graphs of systems is not easy. A lot of parameters have to be inferred, for differential description, even if the type of each interaction is known in the graph. Even small modifications in parameters can lead to a strong change in the dynamics. Kinetic Logic is used to build discrete models, in which such details of the systems are not required. The information required can be derived directly from the graph of interactions or from a sufficiently explicit verbal description. It only considers the thresholds of the elements and uses logical equations to construct state tables. Through this procedure, it is a straightforward matter to determine the behavior of the system.
Formalism
Following is René Thomas’s formalism for Kinetic Logic :
In a directed graph G = (V, A), we note G− (v) and G+ (v) the set of predecessors and successors of a node v ∈ V respectively.
Definition 1: A biological regulatory network (BRN) is a tuple G = (V, A, l, s, t, K) where
(V, A) is a directed graph denoted by G,
l is a function from V to N,
s is a function from A to {+, −},
t is a function from A to N such that, for all u ∈ V, if G+(u) is not empty then {t(u, v) | v ∈ G+(u)} = { 1, . . . , l(u)}.
K = {Kv | v ∈ V} is a set of maps: for each v ∈ V, Kv is a function from 2G− (v) to {0, . . . , l(v)} such that Kv(ω) ≤ Kv(ω_) for all ω ⊆ ω_ ⊆ G−(v).
The map l describes the domain of each variable v: if l (v) = k, the abstract concentration on v holds its value in {0, 1, . . . , k}. Similarly, the map s represents the sign of the regulation (+ for an activation, − for an inhibition). t (u, v) is the threshold of the regulation from u to v: this regulation takes place iff the abstract conce |
https://en.wikipedia.org/wiki/Self-discharge | Self-discharge is a phenomenon in batteries in which internal chemical reactions reduce the stored charge of the battery without any connection between the electrodes or any external circuit. Self-discharge decreases the shelf life of batteries and causes them to have less than a full charge when actually put to use.
How fast self-discharge in a battery occurs is dependent on the type of battery, state of charge, charging current, ambient temperature and other factors. Primary batteries are not designed for recharging between manufacturing and use, thus have battery chemistry that has to have a much lower self-discharge rate than older types of secondary cells, but have lost that advantage with the development of rechargeable secondary cells with very low self discharge rates like NiMH cells.
Self-discharge is a chemical reaction, just as closed-circuit discharge is, and tends to occur more quickly at higher temperatures. Storing batteries at lower temperatures thus reduces the rate of self-discharge and preserves the initial energy stored in the battery. Self-discharge is also thought to be reduced as a passivation layer develops on the electrodes over time.
Typical self-discharge by battery type |
https://en.wikipedia.org/wiki/Beat%20detection | In signal analysis, beat detection is using computer software or computer hardware to detect the beat of a musical score. There are many methods available and beat detection is always a tradeoff between accuracy and speed. Beat detectors are common in music visualization software such as some media player plugins. The algorithms used may utilize simple statistical models based on sound energy or may involve sophisticated comb filter networks or other means. They may be fast enough to run in real time or may be so slow as to only be able to analyze short sections of songs.
See also
Pitch detection
External links
Beat This > Beat Detection Algorithm
Audio Analysis using the Discrete Wavelet Transform
Signal processing |
https://en.wikipedia.org/wiki/Jordan%20Transverse%20Mercator | Jordan Transverse Mercator (JTM) (Arabic: نظام تربيع ميركاتور الأردني المستعرض) is a grid system created by the Royal Jordan Geographic Center (RJGC). This system is based on 6° belts with a Central Meridian of 37° East and a Scale Factor at Origin (mo) = 0.9998. The JTM is based on the Hayford ellipsoid adopted by the IUGG in 1924. No transformation parameters are presently offered by the government. However, Prof. Stephen H. Savage of Arizona State University provides the following parameters for the projection:
Jordan Transverse Mercator
Geographic Coordinate System: GCS_International_1924
Datum: D:International_1924
Spheroid: International_1924
Axis: 6378388
Flattening: 297
Prime Meridian: Greenwich
Prime Meridian Longitude: 0
Units: Degree
Unit Scale Factor: 0.017453292519943295
Projection: Transverse Mercator
False Easting: 500,000
False Northing: -3,000,000
Central Meridian: 37
Scale Factor: 0.9998
Central Parallel: 0
Units: Meter
Scale Factor 1
Three-parameter transformation to WGS84 is:
ΔX = –86 meters
ΔY = –98 meters
ΔZ = –119 meters
Prof. Savage also offers software, ReprojectME!, which will convert coordinates between JTM and other systems. (See http://daahl.ucsd.edu/gaialab/# for more information.)
The central meridian of 37° East is roughly midway between the extremes of Jordan: the Karameh Border Crossing with Iraq is close to 39° East, while the city of Aqaba on the Red Sea is close to 35° East.
See also
Jordan
Mercator projection |
https://en.wikipedia.org/wiki/Split%20tunneling | Split tunneling is a computer networking concept which allows a user to access dissimilar security domains like a public network (e.g., the Internet) and a local area network or wide area network at the same time, using the same or different network connections. This connection state is usually facilitated through the simultaneous use of a LAN network interface controller (NIC), radio NIC, Wireless LAN (WLAN) NIC, and VPN client software application without the benefit of an access control.
For example, suppose a user utilizes a remote access VPN software client connecting to a campus network using a hotel wireless network. The user with split tunneling enabled is able to connect to file servers, database servers, mail servers and other servers on the corporate network through the VPN connection. When the user connects to Internet resources (websites, FTP sites, etc.), the connection request goes directly out the gateway provided by the hotel network. However, not every VPN allows split tunneling. Some VPNs with split tunneling include Private Internet Access (PIA), ExpressVPN, and Surfshark.
Split tunneling is sometimes categorized based on how it is configured. A split tunnel configured to only tunnel traffic destined to a specific set of destinations is called a split-include tunnel. When configured to accept all traffic except traffic destined to a specific set of destinations, it is called a split-exclude tunnel.
Advantages
One advantage of using split tunneling is that it alleviates bottlenecks and conserves bandwidth as Internet traffic does not have to pass through the VPN server.
Another advantage is in the case where a user works at a supplier or partner site and needs access to network resources on both networks. Split tunneling prevents the user from having to continually connect and disconnect.
Disadvantages
A disadvantage is that when split tunneling is enabled, users bypass gateway level security that might be in place within the company infrast |
https://en.wikipedia.org/wiki/Comparison%20of%20open-source%20configuration%20management%20software | This is a comparison of notable free and open-source configuration management software, suitable for tasks like server configuration, orchestration and infrastructure as code typically performed by a system administrator.
Basic properties
"Verify mode" (also called dry run) refers to having an ability to determine whether a node is conformant with a guarantee of not modifying it, and typically involves the exclusive use of an internal language supporting read-only mode for all potentially system-modifying operations. "Mutual auth" refers to the client verifying the server and vice versa.
"Agent" describes whether additional software daemons are required. Depending on the management software these agents are usually deployed on the target system or on one or many central "controller" servers. Although "Agent-less" = "No" is colored red and might seem to be a negative, in fact having an agent can be considered quite advantageous to many. Consider the impact if an agent-less tool loses connectivity to a node while making critical changes—leaving the node in an indeterminate state that compromises its (production?) functionality.
Platform support
Note: This means platforms on which a recent version of the tool has actually been used successfully, not platforms where it should theoretically work since it is written in good portable C/C++ or an interpreted language. It should also be listed as a supported platform on the project's web site.
Short descriptions
Not all tools have the same goal and the same feature set. To help distinguish between all of these software packages, here is a short description of each one.
Ansible
Combines multi-node deployment, ad-hoc task execution, and configuration management in one package. Manages nodes over SSH and requires python (2.6+ or 3.5+) to be installed on them. Modules work over JSON and standard output and can be written in any language. Uses YAML to express reusable descriptions of systems.
Bcfg2
Software to manage the |
https://en.wikipedia.org/wiki/Ada%20Dietz | Ada K. Dietz (October 7, 1888 – January 12, 1981) was an American weaver best known for her 1949 monograph Algebraic Expressions in Handwoven Textiles, which defines a novel method for generating weaving patterns based on algebraic patterns. Her method employs the expansion of multivariate polynomials to devise a weaving scheme. Dietz' work is still well-regarded today, by both weavers and mathematicians. Along with the references listed below, Griswold (2001) cites several additional articles on her work.
Algebraic weaving
Ada Dietz developed her algebraic method in 1946 while living in Long Beach, California. An avid weaver, Dietz drew upon her experience as a former math teacher to devise a threading pattern based on a cubic binomial expansion. She describes her idea as follows:
"Taking the cube of a binomial [ (x + y)3], I approached [the pattern] in the way applied algebraic problems are approached - by letting x equal one unknown and y equal the other unknown.
"In this case, x equaled the first and second harnesses, and y equaled the third and fourth harnesses. Then it was simply a matter of expanding the cube of the binomial and substituting the values of x and y to write the threading draft." (Dietz, 1949)
A piece based on the formula (a + b + c + d + e + f)2, submitted to the Little Loomhouse Country Fair in Louisville, Kentucky received such a positive response, which prompted a collaboration between Dietz and Little Loomhouse's founder, Lou Tate. The fruits of the collaboration included the booklet Algebraic Expressions in Handwoven Textiles and a traveling exhibit which continued throughout the 1950s.
History and development
Dietz was a high school biology and math teacher when she met Ruth E. Foster, a professional weaver with the Hewson Studios in Los Angeles. Foster's work inspired Dietz to begin studying weaving at Wayne University in Detroit under Nellie Sargent Johnson. Her experiments in writing weaving drafts began in Johnson's classes |
https://en.wikipedia.org/wiki/Lake%20Pohjalampi | Lake Pohjalampi is a lake in eastern Finland. The lake has a surface area of 61 hectares, and the mean and maximum depth are 3.2 meters and 5.3 meters, respectively.
The lake was used to study the effects of biomanipulation on a small, mesotrophic lake. From 1993 to 1997, over 200 kilograms per hectare of fish, mainly roach (Rutilus rutilus) and bream (Abramis brama) were caught and removed. This resulted in a reduction of nearly 80% of fish biomass. The effects on the benthic invertebrates were investigated using biomass and density, during the years of fish removal and for the following three years. The decrease in benthivorous fish stock led to higher biomass and density in all the major benthic invertebrate groups.
Three years after fish removal had been discontinued, the fish biomass had recovered to almost its initial value; the community structure, however, was different because of much greater abundance of perch. The biomasses of the benthic invertebrates in many groups was seen to drop in the years 1999-2000, apparently due to the increased predation. The relationship between the two trophic levels is thought to be as a consequence of the open and sparsely vegetated bottom, which offers minimal shelter to invertebrate prey.
An additional factor in the reduced invertebrate biomasses was a fall in both phosphorus and chlorophyll a concentrations, which led to a drop in primary production. The phosphorus and chlorophyll a decreases occurred after the three years of fishing and have remained low since. This could be related to the loss of waste products and detritus the fish would have produced over the years. |
https://en.wikipedia.org/wiki/Isotonic%20contraction | In an isotonic contraction, tension remains the same, whilst the muscle's length changes. Isotonic contractions differ from isokinetic contractions in that in isokinetic contractions the muscle speed remains constant. While superficially identical, as the muscle's force changes via the length-tension relationship during a contraction, an isotonic contraction will keep force constant while velocity changes, but an isokinetic contraction will keep velocity constant while force changes. A near isotonic contraction is known as Auxotonic contraction.
There are two types of isotonic contractions: (1) concentric and (2) eccentric. In a concentric contraction, the muscle tension rises to meet the resistance, then remains the same as the muscle shortens. In eccentric, the muscle lengthens due to the resistance being greater than the force the muscle is producing.
Concentric
This type is typical of most exercise. The external force on the muscle is less than the force the muscle is generating - a shortening contraction. The effect is not visible during the classic biceps curl, which is in fact auxotonic because the resistance (torque due to the weight being lifted) does not remain the same through the exercise. Tension is highest at a parallel to the floor level, and eases off above and below this point. Therefore, tension changes as well as muscle length.
Eccentric
There are two main features to note regarding eccentric contractions. First, the absolute tensions achieved can be very high relative to the muscle's maximum tetanic tension generating capacity (you can set down a much heavier object than you can lift). Second, the absolute tension is relatively independent of lengthening velocity.
Muscle injury and soreness are selectively associated with eccentric contraction. Muscle strengthening using exercises that involve eccentric contractions is lower than using concentric exercises. However because higher levels of tension are easier to attain during exercises th |
https://en.wikipedia.org/wiki/Sakurai%27s%20Object | Sakurai's Object (V4334 Sagittarii) is a star in the constellation of Sagittarius. It is thought to have previously been a white dwarf that, as a result of a very late thermal pulse, swelled and became a red giant. It is located at the center of a planetary nebula and is believed to currently be in thermal instability and within its final shell helium flash phase.
At the time of its discovery, astronomers believed Sakurai's Object to be a slow nova. Later spectroscopic analysis suggested that the star was not a nova, but had instead undergone a very late thermal pulse similar to that of V605 Aquilae, causing it to vastly expand. V605 Aquilae, which was discovered in 1919, is the only other star known to have been observed during the high luminosity phase of a very late thermal pulse, and models predict that Sakurai's Object, over the next few decades, will follow a similar life cycle.
Sakurai's Object and other similar stars are expected to end up as helium-rich white dwarfs after retracing their evolution track from the "born-again" giant phase back to the white dwarf cooling track. There are few other suspected "born-again" objects, one example being FG Sagittae. Having erupted in 1995, it is expected that Sakurai's Object's final helium flash will be the first well-observed one.
Observation history
An International Astronomical Union Circular sent on 23 February 1996 announced the discovery of a "possible 'slow' nova" of magnitude 11.4 by Yukio Sakurai, an amateur astronomer. Japanese astronomer Syuichi Nakano reported the discovery, drawing attention to the fact that the object had not been visible in images from 1993 nor in Center for Astrophysics Harvard & Smithsonian records for the years 1930–1951, despite it appearing to slowly brighten over the previous years. Nakano wrote that "While the outburst [suggests] a slow or symbiotic nova, the lack of obvious emission lines one year after brightening is very unusual."
Following the initial announcement, Hi |
https://en.wikipedia.org/wiki/Pharmacokinetics | Pharmacokinetics (from Ancient Greek pharmakon "drug" and kinetikos "moving, putting in motion"; see chemical kinetics), sometimes abbreviated as PK, is a branch of pharmacology dedicated to describing how the body affects a specific substance after administration. The substances of interest include any chemical xenobiotic such as pharmaceutical drugs, pesticides, food additives, cosmetics, etc. It attempts to analyze chemical metabolism and to discover the fate of a chemical from the moment that it is administered up to the point at which it is completely eliminated from the body. Pharmacokinetics is based on mathematical modeling that places great emphasis on the relationship between drug plasma concentration and the time elapsed since the drug's administration. Pharmacokinetics is the study of how an organism is effected by drug, whereas pharmacodynamics (PD) is the study of how the drug affects the organism. Both together influence dosing, benefit, and adverse effects, as seen in PK/PD models.
ADME
A number of phases occur once the drug enters into contact with the organism, these are described using the acronym ADME (or LADME if liberation is included as a separate step from absorption):
Liberation – the process of the active ingredient separating from its pharmaceutical formulation. See also IVIVC.
Absorption – the process of a drug entering into systemic circulation from the site of administration
Distribution – the dispersion or dissemination of substances throughout the fluids and tissues of the body.
Metabolism (or biotransformation, or inactivation) – the chemical reactions of the drug and irreversible breakdown into metabolites (e.g. by metabolic enzymes such as cytochrome P450 or glucuronosyltransferase enzymes)
Excretion – the removal of the substance or metabolites from the body. In rare cases, some drugs irreversibly accumulate in body tissue.
Some textbooks combine the first two phases as the drug is often administered in an active form, |
https://en.wikipedia.org/wiki/Internet%20resource%20locator | Internet resource locators, described in RFC 1736, convey location and access information for resources. Typical examples of resources include network accessible documents, WAIS databases, FTP servers, and Telnet destinations.
Locators may apply to resources that are not always or not ever network accessible. Examples of the latter include human beings and physical objects that have no electronic instantiation (that is, objects without an existence completely defined by digital objects such as disk files).
A resource locator is a kind of resource identifier. Other kinds of resource identifiers allow names and descriptions to be associated with resources. A resource name is intended to provide a stable handle to refer to a resource long after the resource itself has moved or perhaps gone out of existence. A resource description comprises a body of meta-information to assist resource search and selection.
An Internet resource locator is a locator defined by an Internet resource location standard. A resource location standard in conjunction with resource description and resource naming standards specifies a comprehensive infrastructure for network-based information dissemination. Mechanisms for mapping between locators, names, and descriptive identifiers.
Network-based information resource providers require a method of describing the location of and access to their resources. Information systems users require a method whereby client software can interpret resource access and location descriptions on their behalf in a relatively transparent way. Without such a method, transparent and widely distributed, open information access on the Internet would be difficult if not impossible.
See also
Uniform Resource Identifier |
https://en.wikipedia.org/wiki/Fast%20neutron%20therapy | Fast neutron therapy utilizes high energy neutrons typically between 50 and 70 MeV to treat cancer. Most fast neutron therapy beams are produced by reactors, cyclotrons (d+Be) and linear accelerators. Neutron therapy is currently available in Germany, Russia, South Africa and the United States. In the United States, one treatment center is operational, in Seattle, Washington. The Seattle center uses a cyclotron which produces a proton beam impinging upon a beryllium target.
Advantages
Radiation therapy kills cancer cells in two ways depending on the effective energy of the radiative source. The amount of energy deposited as the particles traverse a section of tissue is referred to as the linear energy transfer (LET). X-rays produce low LET radiation, and protons and neutrons produce high LET radiation. Low LET radiation damages cells predominantly through the generation of reactive oxygen species, see free radicals. The neutron is uncharged and damages cells by direct effect on nuclear structures. Malignant tumors tend to have low oxygen levels and thus can be resistant to low LET radiation. This gives an advantage to neutrons in certain situations. One advantage is a generally shorter treatment cycle. To kill the same number of cancerous cells, neutrons require one third the effective dose as protons. Another advantage is the established ability of neutrons to better treat some cancers, such as salivary gland, adenoid cystic carcinomas and certain types of brain tumors, especially high-grade gliomas
LET
When therapeutic energy X-rays (1 to 25 MeV) interact with cells in human tissue, they do so mainly by Compton interactions, and produce relatively high energy secondary electrons. These high energy electrons deposit their energy at about 1 keV/µm. By comparison, the charged particles produced at a site of a neutron interaction may deliver their energy at a rate of 30–80 keV/µm. The amount of energy deposited as the particles traverse a section of tissue is refe |
https://en.wikipedia.org/wiki/Phylogenetic%20footprinting | Phylogenetic footprinting is a technique used to identify transcription factor binding sites (TFBS) within a non-coding region of DNA of interest by comparing it to the orthologous sequence in different species. When this technique is used with a large number of closely related species, this is called phylogenetic shadowing.
Researchers have found that non-coding pieces of DNA contain binding sites for regulatory proteins that govern the spatiotemporal expression of genes. These transcription factor binding sites (TFBS), or regulatory motifs, have proven hard to identify, primarily because they are short in length, and can show sequence variation. The importance of understanding transcriptional regulation to many fields of biology has led researchers to develop strategies for predicting the presence of TFBS, many of which have led to publicly available databases. One such technique is Phylogenetic Footprinting.
Phylogenetic footprinting relies upon two major concepts:
The function and DNA binding preferences of transcription factors are well-conserved between diverse species.
Important non-coding DNA sequences that are essential for regulating gene expression will show differential selective pressure. A slower rate of change occurs in TFBS than in other, less critical, parts of the non-coding genome.
History
Phylogenetic footprinting was first used and published by Tagle et al. in 1988, which allowed researchers to predict evolutionary conserved cis-regulatory elements responsible for embryonic ε and γ globulin gene expression in primates.
Before phylogenetic footprinting, DNase footprinting was used, where protein would be bound to DNA transcription factor binding sites (TFBS) protecting it from DNase digestion. One of the problems with this technique was the amount of time and labor it would take. Unlike DNase footprinting, phylogenetic footprinting relies on evolutionary constraints within the genome, with the "important" parts of the sequence being con |
https://en.wikipedia.org/wiki/Rotary%20feeder | Rotary feeders, also known as rotary airlocks or rotary valves, are commonly used in industrial and agricultural applications as a component in a bulk or specialty material handling system. Rotary feeders are primarily used for discharge of bulk solid material from hoppers/bins, receivers, and cyclones into a pressure or vacuum-driven pneumatic conveying system. Components of a rotary feeder include a rotor shaft, housing, head plates, and packing seals and bearings.
Rotors have large vanes cast or welded on and are typically driven by small internal combustion engines or electric motors.
Use
Rotary airlock feeders have wide application in industry wherever dry free-flowing powders, granules, crystals, or pellets are used. Typical materials include: cement, ore, sugar, minerals, grains, plastics, dust, fly ash, flour, gypsum, lime, coffee, cereals, pharmaceuticals, etc.
Industries requiring this type include cement, asphalt, chemical, mining, plastics, food, etc.
Rotary feeders are ideal for pollution control applications in wood, grain, food, textile, paper, tobacco, rubber, and paint industries, the Standard Series works beneath dust collectors and cyclone separators even with high temperatures and different pressure differentials.
Rotary valves are available with square or round inlet and outlet flanges. Housing can be fabricated out of sheet material or cast. Common materials are cast iron, carbon steel, 304 SS, 316 SS, and other materials.
Rotary airlock feeders are often available in standard and heavy duty models, the difference being the head plate and bearing configuration. Heavy duty models use an outboard bearing in which the bearings are moved out away from the head plate.
Housing inlet and discharge configurations are termed drop-thru or side entry.
Different wear protections are available such as hard chrome or ceramic plating on the inner housing surfaces.
Grease and air purge fittings are often provided to prevent contaminants from entering th |
https://en.wikipedia.org/wiki/Pituitary-specific%20positive%20transcription%20factor%201 | POU domain, class 1, transcription factor 1 (Pit1, growth hormone factor 1), also known as POU1F1, is a transcription factor for growth hormone.
Function
PIT1 is a pituitary-specific transcription factor responsible for pituitary development and hormone expression in mammals and is a member of the POU family of transcription factors that regulate mammalian development. The POU family is so named because the first 3 members identified were PIT1 and OCT1 (MIM 164175) of mammals, and Unc-86 of C. elegans (Herr et al., 1988). PIT1 contains 2 protein domains, termed POU-specific and POU-homeo, which are both necessary for high affinity DNA binding on genes encoding growth hormone (GH; MIM 139250) and prolactin (PRL; MIM 176760). PIT1 is also important for regulation of the genes encoding prolactin and thyroid-stimulating hormone beta subunit (TSHB; MIM 188540) by thyrotropin-releasing hormone (TRH; MIM 257120) and cyclic AMP.[supplied by OMIM]
Interactions
Pituitary-specific positive transcription factor 1 has been shown to interact with GATA2 and PITX1. |
https://en.wikipedia.org/wiki/Octamer%20transcription%20factor | Octamer transcription factors are a family of transcription factors which binds to the "ATTTGCAT" DNA sequence. Their DNA-binding domain is a POU domain.
There are eight Octamer proteins in humans (Oct1–11), which have been renamed according to the different classes of POU domain. Octamer-3/4, also known as POU5F1, is one of the Yamanaka factors, which are critical for the maintenance and self-renewal of embryonic stem cells. On the other hand, Oct-1 and Oct-2 are widely expressed in adult tissues. Oct-7, 8 and 9, also known as "brain factors", are predominantly expressed in the central nervous system during embryonic development. Oct-6 expression is confined to embryonic stem cells and the developing nervous system and skin, while Oct-11 is also involved in skin differentiation.
Human Oct proteins
Oct-1 -
Oct-2 -
Oct-3/4 –
Oct-6 –
Oct-7 –
Oct-8 –
Oct-9 –
Oct-11 – |
https://en.wikipedia.org/wiki/Oct-2 | Oct-2 (octamer-binding protein 2) also known as POU domain, class 2, transcription factor 2 is a protein that in humans is encoded by the POU2F2 gene.
Oct-2 is an octamer transcription factor which is a member of the POU family. |
https://en.wikipedia.org/wiki/Pseudomedian | In statistics, the pseudomedian is a measure of centrality for data-sets and populations. It agrees with the median for symmetric data-sets or populations. In mathematical statistics, the pseudomedian is also a location parameter for probability distributions.
Description
The pseudomedian of a distribution is defined to be a median of the distribution of , where and are independent, each with the same distribution .
When is a symmetric distribution, the pseudomedian coincides with the median; otherwise this is not generally the case.
The Hodges–Lehmann statistic, defined as the median of all of the midpoints of pairs of observations, is a consistent estimator of the pseudomedian.
Like the set of medians, the pseudomedian is well defined for all probability distributions, even for the many distributions that lack modes or means.
Pseudomedian filter in signal processing
In signal processing there is another definition of pseudomedian filter for discrete signals.
For a time series of length 2N + 1, the pseudomedian is defined as follows. Construct N + 1 sliding windows each of length N + 1. For each window, compute the minimum and maximum. Across all N + 1 windows, find the maximum minimum and the minimum maximum. The pseudomedian is the average of these two quantities.
See also
Hodges–Lehmann estimator
Median filter
Lulu smoothing |
https://en.wikipedia.org/wiki/Benjamin%20Kagan | Veniamin Fyodorovich Kagan (; 10 March 1869 – 8 May 1953) was a Russian and Soviet mathematician and expert in geometry. He is the maternal grandfather of mathematicians Yakov Sinai and Grigory Barenblatt.
Biography
Kagan was born in Shavli, in the Kovno Governorate of the Russian Empire (now Šiauliai, Lithuania) in 1869, to a poor Lithuanian Jewish family. In 1871 his family moved to Yekaterinoslav (now Dnipro), where he grew up. Kagan entered the Imperial Novorossiya University in Odesa in 1887, but was expelled for revolutionary activities in 1889. He was put on probation and sent back to Yekaterinoslav. He studied mathematics on his own and in 1892 passed the state exam at Kyiv University.
In 1894 Kagan moved to Saint Petersburg where he continued his studies with Andrey Markov and Konstantin Posse. They tried to help him to obtain an academic position, but Kagan's Jewish background was an obstacle. Only in 1897 was he allowed to become a dozent at the Imperial Novorossiya University, where he continued to work until 1923. His students in the theory of relativity class he taught in 1921-22 included Nikolai Papaleksi, Alexander Frumkin and Igor Tamm.
Kagan worked at Moscow State University where he held the Geometry Chair from 1923 till 1952.
In 1924 he joined Otto Schmidt in drawing up plans for the Great Soviet Encyclopedia.
Mathematical work
He published over 100 mathematical papers in different parts of geometry, particularly on hyperbolic geometry and on Riemannian geometry. He received the Stalin Prize in 1943. He founded the science publisher Mathesis in Odesa. He was a director of the mathematics and natural sciences department of the Great Soviet Encyclopedia. He wrote a definitive biography of Nikolai Lobachevsky and edited his collected works (5 volumes, 1946–1951).
Kagan's doctoral students include Viktor Wagner and Isaak Yaglom.
Trivia
He's a minor character in The Fourth Prose (1930) by Osip Mandelstam.
External links
Biography – in t |
https://en.wikipedia.org/wiki/Kasami%20code | Kasami sequences are binary sequences of length where is an even integer. Kasami sequences have good cross-correlation values approaching the Welch lower bound. There are two classes of Kasami sequences—the small set and the large set.
Kasami Set
The process of generating a Kasami sequence is initiated by generating a maximum length sequence , where . Maximum length sequences are periodic sequences with a period of exactly . Next, a secondary sequence is derived from the initial sequence via cyclic decimation sampling as , where . Modified sequences are then formed by adding and cyclically time shifted versions of using modulo-two arithmetic, which is also termed the exclusive or (xor) operation. Computing modified sequences from all unique time shifts of forms the Kasami set of code sequences.
See also
Gold sequence (aka Gold code)
JPL sequence (aka JPL code) |
https://en.wikipedia.org/wiki/Winged-helix%20transcription%20factors | Consisting of about 110 amino acids, the domain in winged-helix transcription factors (see Regulation of gene expression) has four helices and a two-strand beta-sheet.
These proteins are classified into 19 families called FoxA-FoxS.
Mutations in FoxP proteins are implicated in human autoimmune diseases.
See also
FOX proteins
External links
Transcription factors |
https://en.wikipedia.org/wiki/Effective%20medium%20approximations | In materials science, effective medium approximations (EMA) or effective medium theory (EMT) pertain to analytical or theoretical modeling that describes the macroscopic properties of composite materials. EMAs or EMTs are developed from averaging the multiple values of the constituents that directly make up the composite material. At the constituent level, the values of the materials vary and are inhomogeneous. Precise calculation of the many constituent values is nearly impossible. However, theories have been developed that can produce acceptable approximations which in turn describe useful parameters including the effective permittivity and permeability of the materials as a whole. In this sense, effective medium approximations are descriptions of a medium (composite material) based on the properties and the relative fractions of its components and are derived from calculations, and effective medium theory. There are two widely used formulae.
Effective permittivity and permeability are averaged dielectric and magnetic characteristics of a microinhomogeneous medium. They both were derived in quasi-static approximation when the electric field inside a mixture particle may be considered as homogeneous. So, these formulae can not describe the particle size effect. Many attempts were undertaken to improve these formulae.
Applications
There are many different effective medium approximations, each of them being more or less accurate in distinct conditions. Nevertheless, they all assume that the macroscopic system is homogeneous and, typical of all mean field theories, they fail to predict the properties of a multiphase medium close to the percolation threshold due to the absence of long-range correlations or critical fluctuations in the theory.
The properties under consideration are usually the conductivity or the dielectric constant of the medium. These parameters are interchangeable in the formulas in a whole range of models due to the wide applicability of the La |
https://en.wikipedia.org/wiki/FUDI | FUDI (Fast Universal Digital Interface) is a networking protocol used by the Pure Data patching language invented by Miller Puckette. It is a string based protocol in which messages are separated by semicolons. Messages are made up of tokens separated by whitespaces, and numerical tokens are represented as strings.
Format
FUDI is a packet oriented protocol.
Each message consists of one or more atoms, separated by one or more whitespace characters, and it's terminated by a semicolon character.
An atom is a sequence of one or more characters; whitespaces inside atoms can be escaped by the backslash (ascii 92) character (see Examples below).
A whitespace is either a space (ascii 32), a tab (ascii 9) or a newline (ascii 10).
A semicolon (ascii 59) is mandatory to terminate (and send) a message.
A newline is just treated as whitespace and not needed for message termination.
Implementations
pdsend / pdreceive
Those command-line tools are distributed with the software Pure Data. They are meant to be used with their counterparts, the classes [netsend] / [netreceive] of Pd.
[netsend] / [netreceive]
Those classes can be used to transport Pd-messages over a TCP or UDP socket. Both are part of Pd-vanilla.
[netserver] / [netclient]
Those are part of maxlib and allow bidirectional connections of multiple clients with one server.
Example messages
test/blah 123.45314;
my-slider 12;
hello this is a message;
this message continues
in the following
line;
you; can; send; multiple messages; in a line;
this\ is\ one\ whole\ atom;
this_atom_contains_a\
newline_character_in_it; |
https://en.wikipedia.org/wiki/Sexy%20son%20hypothesis | The sexy son hypothesis in evolutionary biology and sexual selection, proposed by Patrick J. Weatherhead and Raleigh J. Robertson of Queen's University in Kingston, Ontario in 1979, states that a female's ideal mate choice among potential mates is one whose genes will produce males with the best chance of reproductive success. This implies that other benefits the father can offer the mother or offspring are less relevant than they may appear, including his capacity as a parental caregiver, territory and any nuptial gifts. Fisher's principle means that the sex ratio (except in certain eusocial insects) is always near 1:1 between males and females, yet what matters most are her "sexy sons'" future breeding successes, more likely if they have a promiscuous father, in creating large numbers of offspring carrying copies of her genes. This sexual selection hypothesis has been researched in species such as the European pied flycatcher.
Context
Female mating preferences are widely recognized as being responsible for the rapid and divergent evolution of male secondary sex characteristics. In 1915, Ronald Fisher wrote:
Granted that while this taste and preference prevails among the females of the species, the males will grow more and more elaborate and beautiful tail feathers, the question must be answered "Why have the females this taste? Of what use is it to the species that they should select this seemingly useless ornament?"
The first step to a solution lies in the fact that the success of an animal in the struggle for existence is not measured only by the number of offspring which it produces and rears, but also by the probable success of these offspring. So that in selecting a mate from a number of different competitors, it is important to select that one which is most likely to produce successful children.
In 1976, prior to Weatherhead and Robertson's paper, Richard Dawkins had written in his book The Selfish Gene:
Ronald Fisher's principle, as published in his bo |
https://en.wikipedia.org/wiki/Inotuzumab%20ozogamicin | Inotuzumab ozogamicin, sold under the brand name Besponsa, is an antibody-drug conjugate medication used to treat relapsed or refractory B-cell precursor acute lymphoblastic leukemia (ALL).
The medication consists of a humanized monoclonal antibody against CD22 (inotuzumab), linked to a cytotoxic agent from the class of calicheamicins called ozogamicin.
This drug was discovered by scientists collaborating at Celltech and Wyeth, and it was developed by Pfizer which had acquired Wyeth. The U.S. Food and Drug Administration (FDA) considers it to be a first-in-class medication.
Medical use
Inotuzumab ozogamicin is used to treat relapsed or refractory B-cell precursor acute lymphoblastic leukemia.
It is administered by intravenous infusion in a doctor's office or clinic.
In studies in pregnant animals, the drug caused harm to the fetus at doses less than those used clinically, and so the drug has not been tested in pregnant women. Pregnant women should not take inotuzumab ozogamicin and must not become pregnant while taking it. It is unknown if the drug or its metabolites are secreted in breast milk, but women should not breastfeed while taking it, and should wait two months after the last dose to start breastfeeding.
The drug prolongs the QT interval in some people, so it should be used with caution in people with heart arrhythmias.
Adverse effects
The US label for the use of inotuzumab ozagamicin carries an FDA black box warning concerning the risk of liver toxicity, in particular hepatic veno-occlusive disease (VOD), which has been fatal in some people. The risk of this is higher in people who take the drug before having hematopoietic stem cell transplantation (HSCT) and more people die who have HSCT following treatment with this drug, than people who have HSCT, taking other chemotherapies. The risk gets higher as more rounds of treatment with inotuzumab ozogamicin are administered.
The most common serious adverse reactions in people taking the drug in the cl |
https://en.wikipedia.org/wiki/Yttrium%20%2890Y%29%20tacatuzumab%20tetraxetan | {{DISPLAYTITLE:Yttrium (90Y) tacatuzumab tetraxetan}}
Yttrium (90Y) tacatuzumab tetraxetan (trade name AFP-Cide) is a humanized monoclonal antibody intended for the treatment of cancer. The antibody itself, tacatuzumab, is conjugated with tetraxetan, a chelator for yttrium-90, a radioisotope which destroys the tumour cells. |
https://en.wikipedia.org/wiki/Tucotuzumab%20celmoleukin | Tucotuzumab celmoleukin is an anti-cancer drug. It is a fusion protein of a humanized monoclonal antibody (tucotuzumab) and an interleukin-2 (celmoleukin).
This drug was developed by EMD Pharmaceuticals. |
https://en.wikipedia.org/wiki/Finnan%20haddie | Finnan haddie (also known as Finnan haddock, Finnan, Finny haddock, Finny haddie or Findrum speldings) is cold-smoked haddock, representative of a regional method of smoking with green wood and peat in north-east Scotland.
Origin
The origin of finnan haddie is the subject of debate, as some sources attribute the origin to the hamlet of Findon, Aberdeenshire, (also sometimes called Finnan) near Aberdeen, while others insist that the name is a corruption of the village name of Findhorn at the mouth of the River Findhorn in Moray. The "dispute" goes back to the eighteenth century, although it is hard to trace, as adherents fail to acknowledge even the possibility of the alternative view (except for the etymology note in the Oxford English Dictionary). It may have been a popular dish in Aberdeenshire since at least as early as the 1640s.
Popularization
Although known and admired in Scotland for a long time, Finnan haddie became a popular food item in London only in the 1830s. In earlier times, because of the light smoking that the fish received, it did not have a long shelf life—by most contemporary estimates, at most three days (although some suggested no more than one day). Thus, although the fish was often available in Aberdeen "within twelve hours of being [caught]", the distance to London was at that time nearly insurmountable if spoilage was to be avoided. The fish started making its first appearances in London when shipped by established mail coach, but became widely available with the construction of the railway link connecting Aberdeen to London in the 1840s. The association with Findon became strong because of the Aberdeen connection. Occasionally, confusion was so deep that Findon was referred to as Findhorn.
Use in modern cuisine
Finnan has a long association with the traditional Scottish fish soup Cullen skink, and most old Scottish recipe books cite Finnan haddie as the smoked haddock to be used for this dish.
The traditional preparation is to roast |
https://en.wikipedia.org/wiki/Disk%20aggregation | Disk aggregation is the abstraction of two or more hard disks, disk partitions, or other logical volumes into a single logical disk.
This is done to:
create a single logical disk with a capacity larger than any of the available physical disks
provide a simple way to increase disk performance
provide a simple way to implement LUN-level storage virtualization
See also
RAID
Storage virtualization |
https://en.wikipedia.org/wiki/Logical%20disk | A logical disk, logical volume or virtual disk (VD or vdisk for short) is a virtual device that provides an area of usable storage capacity on one or more physical disk drive(s) in a computer system. The disk is described as logical or virtual because it does not actually exist as a single physical entity in its own right. The goal of the logical disk is to provide computer software with what seems a contiguous storage area, sparing them the burden of dealing with the intricacies of storing files on multiple physical units. Most modern operating systems provide some form of logical volume management.
Levels
Logical disks can be defined at various levels in the storage infrastructure.
Operating system
An operating system may define volumes or logical disks and assign each to one physical disk, more than one physical disk or part of the storage area of a physical disk. For example, Windows NT can create several partitions on a hard disk drive, each of which a separate volume with its own file system. Each floppy disk drive, optical disc drive or USB flash drive in Windows NT becomes one volume. Windows NT can also create partitions that span multiple hard disks drives. Each volume is identified with a drive letter.
Storage area network
Storage area networks (SANs) consolidate inhomogeneous storage devices. As such logical disks or vdisks allow computer programs to access files stored on a SAN.
Storage subsystem
A hardware-level redundant array of independent disks (RAID) exposes itself to the operating system as one logical disk while the array itself consists of several disks. The operating system either does not know that the hardware with which it is interfacing is a RAID, or knows but still does not concern itself with intricate details of storage. In case of the latter, specialized management, maintenance and diagnostics software dedicated to that specific RAID may run on the operating system.
Motivation
When IBM first released the magnetic disk drive in |
https://en.wikipedia.org/wiki/Kapustinskii%20equation | The Kapustinskii equation calculates the lattice energy UL for an ionic crystal, which is experimentally difficult to determine. It is named after Anatoli Fedorovich Kapustinskii who published the formula in 1956.
{|
|-
|where ||K = 1.20200 J·m·mol−1
|-
| ||d = 3.45 m
|-
| ||ν is the number of ions in the empirical formula,
|-
| || z+ and z− are the numbers of elementary charge on the cation and anion, respectively, and
|-
| || r+ and r− are the radii of the cation and anion, respectively, in meters.
|}
The calculated lattice energy gives a good estimation for the Born–Landé equation; the real value differs in most cases by less than 5%.
Furthermore, one is able to determine the ionic radii (or more properly, the thermochemical radius) using the Kapustinskii equation when the lattice energy is known. This is useful for rather complex ions like sulfate (SO) or phosphate (PO).
Derivation from the Born–Landé equation
Kapustinskii originally proposed the following simpler form, which he faulted as "associated with antiquated concepts of the character of repulsion forces".
Here, K' = 1.079 J·m·mol−1. This form of the Kapustinskii equation may be derived as an approximation of the Born–Landé equation, below.
Kapustinskii replaced r0, the measured distance between ions, with the sum of the corresponding ionic radii. In addition, the Born exponent, n, was assumed to have a mean value of 9. Finally, Kapustinskii noted that the Madelung constant, M, was approximately 0.88 times the number of ions in the empirical formula. The derivation of the later form of the Kapustinskii equation followed similar logic, starting from the quantum chemical treatment in which the final term is where d is as defined above. Replacing r0 as before yields the full Kapustinskii equation.
See also
Born–Haber cycle |
https://en.wikipedia.org/wiki/Sleep%20medicine | Sleep medicine is a medical specialty or subspecialty devoted to the diagnosis and therapy of sleep disturbances and disorders. From the middle of the 20th century, research has provided increasing knowledge of, and answered many questions about, sleep–wake functioning. The rapidly evolving field has become a recognized medical subspecialty in some countries. Dental sleep medicine also qualifies for board certification in some countries. Properly organized, minimum 12-month, postgraduate training programs are still being defined in the United States. In some countries, the sleep researchers and the physicians who treat patients may be the same people.
The first sleep clinics in the United States were established in the 1970s by interested physicians and technicians; the study, diagnosis and treatment of obstructive sleep apnea were their first tasks. As late as 1999, virtually any American physician, with no specific training in sleep medicine, could open a sleep laboratory.
Disorders and disturbances of sleep are widespread and can have significant consequences for affected individuals as well as economic and other consequences for society. The US National Transportation Safety Board has, according to Charles Czeisler, member of the Institute of Medicine and Director of the Harvard University Medical School Division of Sleep Medicine at Brigham and Women's Hospital, discovered that the leading cause (31%) of fatal-to-the-driver heavy truck crashes is fatigue related (though rarely associated directly with sleep disorders, such as sleep apnea), with drugs and alcohol as the number two cause (29%). Sleep deprivation has also been a significant factor in dramatic accidents, such as the Exxon Valdez oil spill, the nuclear incidents at Chernobyl and Three Mile Island and the explosion of the space shuttle Challenger.
Scope and classification
Competence in sleep medicine requires an understanding of a plethora of very diverse disorders, many of which present with si |
https://en.wikipedia.org/wiki/Pandemonium%20architecture | Pandemonium architecture is a theory in cognitive science that describes how visual images are processed by the brain. It has applications in artificial intelligence and pattern recognition. The theory was developed by the artificial intelligence pioneer Oliver Selfridge in 1959. It describes the process of object recognition as a hierarchical system of detection and association by a metaphorical set of "demons" sending signals to each other. This model is now recognized as the basis of visual perception in cognitive science.
Pandemonium architecture arose in response to the inability of template matching theories to offer a biologically plausible explanation of the image constancy phenomenon. Contemporary researchers praise this architecture for its elegancy and creativity; that the idea of having multiple independent systems (e.g., feature detectors) working in parallel to address the image constancy phenomena of pattern recognition is powerful yet simple. The basic idea of the pandemonium architecture is that a pattern is first perceived in its parts before the "whole".
Pandemonium architecture was one of the first computational models in pattern recognition. Although not perfect, the pandemonium architecture influenced the development of modern connectionist, artificial intelligence, and word recognition models.
History
Most research in perception has been focused on the visual system, investigating the mechanisms of how we see and understand objects. A critical function of our visual system is its ability to recognize patterns, but the mechanism by which this is achieved is unclear.
The earliest theory that attempted to explain how we recognize patterns is the template matching model. According to this model, we compare all external stimuli against an internal mental representation. If there is "sufficient" overlap between the perceived stimulus and the internal representation, we will "recognize" the stimulus. Although some machines follow a template m |
https://en.wikipedia.org/wiki/N-Acetylglucosamine%20receptor | The N-Acetylglucosamine receptor is a receptor which binds N-Acetylglucosamine.
Studies
The N-Acetylglucosamine (GlcNAc) receptor has been recently found to interact and bind with vimentins at the cell surface. Research indicates that the GlcNAc receptor can therefore be used to target vimentin-expressing cells for gene delivery via receptor-mediated endocytosis. |
https://en.wikipedia.org/wiki/Galectin | Galectins are a class of proteins that bind specifically to β-galactoside sugars, such as N-acetyllactosamine (Galβ1-3GlcNAc or Galβ1-4GlcNAc), which can be bound to proteins by either N-linked or O-linked glycosylation. They are also termed S-type lectins due to their dependency on disulphide bonds for stability and carbohydrate binding. There have been about 15 galectins discovered in mammals, encoded by the LGALS genes, which are numbered in a consecutive manner. Only galectin-1, -2, -3, -4, -7, -7B, -8, -9, -9B, 9C, -10, -12, -13, -14, and -16 have been identified in humans. Galectin-5 and -6 are found in rodents, whereas galectin-11 and -15 are uniquely found in sheep and goats. Members of the galectin family have also been discovered in other mammals, birds, amphibians, fish, nematodes, sponges, and some fungi. Unlike the majority of lectins they are not membrane bound, but soluble proteins with both intra- and extracellular functions. They have distinct but overlapping distributions but found primarily in the cytosol, nucleus, extracellular matrix or in circulation. Although many galectins must be secreted, they do not have a typical signal peptide required for classical secretion. The mechanism and reason for this non-classical secretion pathway is unknown.
Structure
There are three different forms of galectin structure: dimeric, tandem or chimera. Dimeric galectins, also called prototypical galectins, are homodimers, consisting of two identical galectin subunits that have associated with one another. The galectins that fall under this category are galectin-1, -2, -5, -7, -10, -11, -14 and -15. Tandem galectins contain at least two distinct carbohydrate recognition domains (CRD) within one polypeptide, thus are considered intrinsically divalent. The CRDs are linked with a small peptide domain. Tandem galectins include galectin-4, -6, -8, -9 and -12. The final galectin is galectin-3 which is the only galectin found in the chimera category in vertebrates. G |
https://en.wikipedia.org/wiki/LeanCMMI | LeanCMMI is an approach to software engineering process improvement that integrates agile computing methods with process design and deployment for organization's wishing to improve software engineering capability and achieve a maturity level two or three rating based upon the Software Engineering Institute's Capability Maturity Model Integration (CMMI).
Developed by Broadsword, LeanCMMI is based on Lean Engineering and the same concepts that spawned Extreme Programming (XP), Scrum, and Crystal, and traces its roots back to Edward Demings' "Theory of Profound Knowledge." Based upon the principle of "Just Enough Not Too Much," LeanCMMI maps the software process improvement journey across three major releases, each with seven iterations based on three "dimensions" of software process improvement each with equal weight and with equal importance to the success of the program.
Three Dimensions of Process Deployment
Dimension 1: Design - Design and build the process along with all of its artifacts
Virtual teams develop a set of common, standard process deliverables that meet the requirements of their process area
Dimension 2: Communicate - Determine who received what message when
Virtual teams develop a strategy for communicating their work including identifying what message is delivered to each stakeholder group
Dimension 3: Educate - Determine who receives what training, at what level, at which time
Virtual teams develop training strategy and materials including identifying what level or training is delivered to each stakeholder group
Acceptance Test
LeanCMMI uses a series of CMMI SCAMPI B and C Appraisals to be performed at the end of each Release in order to validate the contents of the Release and identify any gaps in the processes as it relates to the CMMI model. This additive approach reduces risk and helps to ensure that a successful SCAMPI A Appraisal can be performed at the conclusion of the program.
Virtual Enduring Teams
LeanCMMI requires an end |
https://en.wikipedia.org/wiki/Couch%20surfing | Couch surfing is a term that is generally defined as when an individual moves from house to house, sleeping in whatever spare space available, oftentimes a couch or floor, generally staying a few days before moving on to the next house. People sometimes couch-surf when they are travelling or because they are homeless.
Couch surfing in travel
Couch surfing's cultural significance grew when the website CouchSurfing was launched in 2004. Upon the release of the app, what previously used to be a cheap alternative for budget travelers became recognized as a hip, new way to travel. Couch surfing became not only a way to save money, but a way to meet new people and have new experiences. Its attraction was in the way it allowed people to have a more immersive and authentic travel experience. Besides CouchSurfing, many other platforms were created and groups were formed in order to help people who are looking to couch-surf connect with potential hosts and other travelers. While couch surfing may not be considered the most popular or mainstream way to travel, in 2018 around 15 million people had identified using couch surfing accommodations to travel. However, couch surfing comes with the issue of safety. It can be less regulated than traditional forms of travel accommodations, making it a more risky choice for vulnerable travelers.
Couch surfing as homelessness
Couch surfing is also considered a form of homelessness. It is the most common type of homelessness amongst youth. It can be a result of substance abuse, conflict in home relationships, or aftermath of leaving abusive situations. The individual may turn to couch surfing as a temporary solution, staying with friends or family members while they search for permanent housing or a way to get back on their feet. It is different from sleeping on the streets or in a shelter, but it still has significant challenges, including the lack of stability and the strain on an individual. Couch surfing homelessness can be a short |
https://en.wikipedia.org/wiki/IKBKG | NF-kappa-B essential modulator (NEMO) also known as inhibitor of nuclear factor kappa-B kinase subunit gamma (IKK-γ) is a protein that in humans is encoded by the IKBKG gene. NEMO is a subunit of the IκB kinase complex that activates NF-κB. The human gene for IKBKG is located on the chromosome band Xq28. Multiple transcript variants encoding different isoforms have been found for this gene.
Function
NEMO (IKK-γ) is the regulatory subunit of the inhibitor of IκB kinase (IKK) complex, which activates NF-κB resulting in activation of genes involved in inflammation, immunity, cell survival, and other pathways.
Clinical significance
Mutations in the IKBKG gene results in incontinentia pigmenti, hypohidrotic ectodermal dysplasia, and several other types of immunodeficiencies.
Incontinentia Pigmenti (IP) is an X-linked dominant disease caused by a mutation in the IKBKG gene. Since IKBKG helps activate NF-κB, which protects cells against TNF-alpha induced apoptosis, a lack of IKBKG (and hence a lack of active NF-κB) makes cells more prone to apoptosis.
Moreover, NEMO has been shown to play a role in preeclampsia and may offer insights into the genetic etiology of this condition. An increased level of NEMO gene expression was found in the blood of pregnant women with preeclampsia and their children. However, a decrease of the mRNA levels of total NEMO and the transcripts 1A, 1B, and 1C in placentas derived from preeclamptic women may be the main reason for intensified apoptosis. Sanger sequencing has indicated two distinct variations in the 3’ UTR region of the NEMO gene in preeclamptic women (IKBKG:c.*368C>A and IKBKG:c.*402C>T). The occurrence of a maternal TT genotype and either a TT genotype in the daughter or T allele in the son increases the risk of preeclampsia by 2.59 fold. The configuration of those maternal and fetal genotypes (TT mother/TT daughter or TT mother/T son) is also associated with the level of NEMO gene expression.
NEMO deficiency syndrome i |
https://en.wikipedia.org/wiki/Contact%20mechanics | Contact mechanics is the study of the deformation of solids that touch each other at one or more points. A central distinction in contact mechanics is between stresses acting perpendicular to the contacting bodies' surfaces (known as normal stress) and frictional stresses acting tangentially between the surfaces (shear stress). Normal contact mechanics or frictionless contact mechanics focuses on normal stresses caused by applied normal forces and by the adhesion present on surfaces in close contact, even if they are clean and dry.
Frictional contact mechanics emphasizes the effect of friction forces.
Contact mechanics is part of mechanical engineering. The physical and mathematical formulation of the subject is built upon the mechanics of materials and continuum mechanics and focuses on computations involving elastic, viscoelastic, and plastic bodies in static or dynamic contact. Contact mechanics provides necessary information for the safe and energy efficient design of technical systems and for the study of tribology, contact stiffness, electrical contact resistance and indentation hardness. Principles of contacts mechanics are implemented towards applications such as locomotive wheel-rail contact, coupling devices, braking systems, tires, bearings, combustion engines, mechanical linkages, gasket seals, metalworking, metal forming, ultrasonic welding, electrical contacts, and many others. Current challenges faced in the field may include stress analysis of contact and coupling members and the influence of lubrication and material design on friction and wear. Applications of contact mechanics further extend into the micro- and nanotechnological realm.
The original work in contact mechanics dates back to 1881 with the publication of the paper "On the contact of elastic solids" ("Ueber die Berührung fester elastischer Körper") by Heinrich Hertz. Hertz was attempting to understand how the optical properties of multiple, stacked lenses might change with the force h |
https://en.wikipedia.org/wiki/Lymphocyte%20cytosolic%20protein%202 | Lymphocyte cytosolic protein 2 (SH2 domain containing leukocyte protein of 76kDa), also known as LCP2 or SLP-76, is a signal-transducing adaptor protein expressed in T cells and myeloid cells and is important in the signaling of T-cell receptors (TCRs). As an adaptor protein, SLP-76 does not have catalytic functions, primarily binding other signaling proteins to form larger signaling complexes. It is a key component of the signaling pathways of receptors with immunoreceptor tyrosine-based activation motifs (ITAMs) such as T-cell receptors, its precursors, and receptors for the Fc regions of certain antibodies. SLP-76 is expressed in T-cells and related lymphocytes like natural killer cells.
Structure and function
The amino acid sequence of the protein has a central domain with a high concentration of prolines, as well as domains at the amino-terminal and carboxy-terminal of the amino acid sequence. The PDB file 1H3H depicts the SH3 domain of GRAP2 in complex with an RSTK-containing peptide representing residues 226–235 of SLP-76. The human and murine cDNAs both encode 533 amino acid proteins that are 72% identical and composed of three modular domains. The central domain binds SRC-Homology 3 (SH3) domains of other adaptor molecules such as Grb2 and Gads. The N-terminus has an acidic region with sections for SH2-domain binding and tyrosine residues that bind the proteins Vav and Nck when phosphorylated. The C-terminus region is itself a SH2 domain and binds FYB among other proteins. SLP-76 is triggered when the TCR binds its ligand by the phosphorylation of tyrosines on the N-terminus by ZAP-70, a tyrosine kinase. Along with the LAT (linker for activation of T cells) adaptor protein, SLP-76 is essential to nearly all downstream effects from T-cell receptor signals. SLP-76, LAT, and Gads together combine into protein complexes, typically with LAT at the center and SLP-76 proteins on the outside. These complexes associate into larger microclusters that activate a mul |
https://en.wikipedia.org/wiki/Unbundled%20access | Unbundled access is an often practiced form of regulation during liberalization, where new entrants of the market (challengers) are offered access to facilities of the incumbent that are hard to duplicate (e.g. for technical or business case reasons). Its applications are mostly found in network-oriented industries (like telecommunication, mail and energy) and often concerns the last mile.
Unbundled access is similar to Bit-stream access, where the incumbent provider gives competitive access not to the actual copper wire of the local loop, but to a high-speed ADSL data connection. Both setups ensure competition for the backhaul but leave "last mile" infrastructure the responsibility of the incumbent carrier.
Telecommunications
United States
In the United States, the Telecommunications Act of 1996 added a number of provisions designed to increase competition. Incumbent local exchange carriers (ILEC), under this law, are required to interconnect with competing telecommunications carriers, allowing access to individual elements of the ILEC's own network on an unbundled basis. The Telecommunications Act of 1996 provided three ways for companies to enter the new competitive telecommunications market: facilities systems, unbundled access, and resale networks.
Unbundled access is defined as "The duty to provide, to any requesting telecommunications carrier for the provision of a telecommunications service, nondiscriminatory access to network elements on an unbundled basis at any technically feasible point on rates, terms, and conditions that are just, reasonable, and nondiscriminatory in accordance with the terms and conditions of the agreement and the requirements of this section and section 252. An incumbent local exchange carrier shall provide such unbundled network elements in a manner that allows requesting carriers to combine such elements in order to provide such telecommunications service.
In 2005, after much litigation concerning its original unbundling rule |
https://en.wikipedia.org/wiki/Association%20for%20Politics%20and%20the%20Life%20Sciences | The Association for Politics and the Life Sciences (APLS) was formed in 1981 and exists to study the field of biopolitics as a subfield of political science. APLS owns an academic peer-reviewed journal, Politics and the Life Sciences (PLS), which is published semi-annually by Cambridge University Press.
External links
Politics and the Life Sciences
Organizations established in 1981
Biology organizations
Political science organizations |
https://en.wikipedia.org/wiki/Harary%27s%20generalized%20tic-tac-toe | Harary's generalized tic-tac-toe or animal tic-tac-toe is a generalization of the game tic-tac-toe, defining the game as a race to complete a particular polyomino on a square grid of varying size, rather than being limited to "in a row" constructions. It was devised by Frank Harary in March 1977, and is a broader definition than that of an m,n,k-game.
Harary's generalization does not include tic-tac-toe itself, as diagonal constructions are not considered a win.
Like many other two-player games, strategy stealing means that the second player can never win. All that is left to study is to determine whether the first player can win, on what board sizes he may do so, and in how many moves it will take.
Results
Square boards
Let b be the smallest size square board on which the first player can win, and let m be the smallest number of moves in which the first player can force a win, assuming perfect play by both sides.
monomino: b = 1, m = 1
domino: b = 2, m = 2
I-tromino: b = 4, m = 3
V-tromino: b = 3, m = 3
I-tetromino: b = 7, m = 8
L-tetromino: b = 4, m = 4
O-tetromino: The first player cannot win
T-tetromino: b = 5, m = 4
Z-tetromino: b = 3, m = 5
F-pentomino: The first player cannot win
I-pentomino: The first player cannot win
L-pentomino: b = 7, m = 10
N-pentomino: b = 6, m = 6
P-pentomino: The first player cannot win
T-pentomino: The first player cannot win
U-pentomino: The first player cannot win
V-pentomino: The first player cannot win
W-pentomino: The first player cannot win
X-pentomino: The first player cannot win
Y-pentomino: b = 7, m = 9
Z-pentomino: The first player cannot win
All hexominoes (with a possible exception of the N-hexomino, which is still currently unsolved, may have b = 15 and m = 13): The first player cannot win
All heptominoes and above: The first player cannot win |
https://en.wikipedia.org/wiki/Malu | is a word in the Samoan language for a female-specific tattoo of cultural significance. The covers the legs from just below the knee to the upper thighs just below the buttocks, and is typically finer and delicate in design compared to the , the equivalent tattoo for males. The takes its name from a particular motif of the same name, usually tattooed in the popliteal fossa (sometimes referred to as the kneepit, or poplit) behind the knee. It is one of the key motifs not seen on men. According to Samoan scholar Albert Wendt and tattooist Su'a Suluape Paulo II, in tattooing, the term refers to notions of sheltering and protection. Samoan women were also tattooed on the hands and sometimes the lower abdomen. These practices have undergone a resurgence since the late 1990s.
Changing significance
In the late 19th and early 20th centuries, only the district Paramount Chief's daughter was eligible to wear the , which was applied to these young women in the years following puberty. Women with the were expected to perform key ceremony tasks and represent their families and villages on ceremonial occasions. However, in 1930, anthropologist Te Rangi Hiroa observed that "...the tattooing of a girl is often used as an opportunity for a student to try his prentice hand. This is also rendered possible by the fact that there is no fusita (fine mat) passed or any of the ceremony that marks the tattooing of the male. It is often sufficient reward for the novice to have the opportunity of practice and to be well fed during the period occupied by the operation...For the daughter of a high chief, who is to become the village taupou, it can be readily understood that an expert artist would be requisitioned and his reward greater."
Ceremonial roles are still important in Samoan society and are restricted in similar ways to particular people with the correct qualifications and cultural knowledge, but the significance of the has shifted. From at least the 1990s, there has been les |
https://en.wikipedia.org/wiki/Salt%20fingering | Salt fingering is a mixing process, example of double diffusive instability, that occurs when relatively warm, salty water overlies relatively colder, fresher water. It is driven by the fact that heated water diffuses more readily than salty water. A small parcel of warm, salty water sinking downwards into a colder, fresher region will lose its heat before losing its salt, making the parcel of water increasingly denser than the water around it and sinking further. Likewise, a small parcel of colder, fresher water will be displaced upwards and gain heat by diffusion from surrounding water, which will then make it lighter than the surrounding waters, and cause it to rise further. Paradoxically, the fact that salinity diffuses less readily than temperature means that salinity mixes more efficiently than temperature due to the turbulence caused by salt fingers.
Salt fingering was first described mathematically by Professor Melvin Stern of Florida State University in 1960 and important field measurements of the process have been made by Raymond Schmitt of the Woods Hole Oceanographic Institution and Mike Gregg and Eric Kunze of the University of Washington, Seattle. A particularly interesting area for salt fingering is found in the Caribbean Sea, where it is responsible for producing a "staircase" of well-mixed layers a few metres in thickness that extend for hundreds of kilometres.
Pre-dating the work of Stern, a paper by the American oceanographer Henry Stommel discussed the creation of a large-scale salt finger in which a column of water would be surrounded by a membrane that would allow diffusion of temperature but not salinity. Once primed by the upward movement of the colder and fresher intermediate water, the resultant "perpetual salt fountain" would be able to draw energy (heat) from the local ocean water stratification. |
https://en.wikipedia.org/wiki/1%20%E2%88%92%202%20%2B%203%20%E2%88%92%204%20%2B%20%E2%8B%AF | In mathematics, 1 − 2 + 3 − 4 + ··· is an infinite series whose terms are the successive positive integers, given alternating signs. Using sigma summation notation the sum of the first m terms of the series can be expressed as
The infinite series diverges, meaning that its sequence of partial sums, , does not tend towards any finite limit. Nonetheless, in the mid-18th century, Leonhard Euler wrote what he admitted to be a paradoxical equation:
A rigorous explanation of this equation would not arrive until much later. Starting in 1890, Ernesto Cesàro, Émile Borel and others investigated well-defined methods to assign generalized sums to divergent series—including new interpretations of Euler's attempts. Many of these summability methods easily assign to a "value" of . Cesàro summation is one of the few methods that do not sum , so the series is an example where a slightly stronger method, such as Abel summation, is required.
The series 1 − 2 + 3 − 4 + ... is closely related to Grandi's series . Euler treated these two as special cases of the more general sequence , where and respectively. This line of research extended his work on the Basel problem and leading towards the functional equations of what are now known as the Dirichlet eta function and the Riemann zeta function.
Divergence
The series' terms do not approach 0; therefore diverges by the term test. Divergence can also be shown directly from the definition: an infinite series converges if and only if the sequence of partial sums converges to limit, in which case that limit is the value of the infinite series. The partial sums of are:
The sequence of partial sums shows that the series does not converge to a particular number: for any proposed limit x, there exists a point beyond which the subsequent partial sums are all outside the interval , so diverges.
The partial sums include every integer exactly once—even 0 if one counts the empty partial sum—and thereby establishes the countability of the s |
https://en.wikipedia.org/wiki/Motion%20ratio | The motion ratio of a mechanism is the ratio of the displacement of the point of interest to that of another point.
The most common example is in a vehicle's suspension, where it is used to describe the displacement and forces in the springs and shock absorbers. The force in the spring is (roughly) the vertical force at the contact patch divided by the motion ratio, and the spring rate is the wheel rate divided by the motion ratio squared.
This is described as the Installation Ratio in the reference. Motion ratio is the more common term in the industry, but sometimes is used to mean the inverse of the above definition.
Motion ratio in suspension of a vehicle describes the amount of shock travel for a given amount of wheel travel. Mathematically, it is the ratio of shock travel and wheel travel. The amount of force transmitted to the vehicle chassis reduces with increase in motion ratio. A motion ratio close to one is desired in the vehicle for better ride and comfort. One should know the desired wheel travel of the vehicle before calculating motion ratio, which depends much on the type of track the vehicle will run upon.
Selecting the appropriate ratio depends on multiple factors:
Bending moment: To reduce the bending moment the strut point should be close to the wheel.
Suspension stiffness: Suspensions tends to stiffen when the inclination of the shock absorber to horizontal tends to 90 deg.
Half-shafts: In rear suspensions, wheel travel is constrained by the universal joints of the half shafts. Design the motion ratio such that at maximum bounce and rebound shocks are the first components that bottom out by hitting bump stops. |
https://en.wikipedia.org/wiki/Indium%20%28111In%29%20capromab%20pendetide | {{DISPLAYTITLE:Indium (111In) capromab pendetide}}
Indium (111In) capromab pendetide (trade name Prostascint) is used to image the extent of prostate cancer. Capromab is a mouse monoclonal antibody which recognizes a protein found on both prostate cancer cells and normal prostate tissue. It is linked to pendetide, a derivative of DTPA. Pendetide acts as a chelating agent for the radionuclide indium-111. Following an intravenous injection of Prostascint, imaging is performed using single-photon emission computed tomography (SPECT).
Early trials with yttrium (90Y) capromab pendetide were also conducted. |
https://en.wikipedia.org/wiki/Nacolomab%20tafenatox | Nacolomab tafenatox is a mouse monoclonal antibody. The antibody itself, nacolomab, is fused with enterotoxin A from Staphylococcus aureus (which is reflected by 'tafenatox' in the drug's name).
See also
Naptumomab estafenatox, a drug with a similar chemical structure and mechanism |
https://en.wikipedia.org/wiki/P25%20ISSI | The Project 25 Inter RF Subsystem Interface (P25 ISSI) is a non-proprietary interface that enables RF subsystems (RFSSs) built by different manufacturers to be connected together into wide area networks so that users on different networks can talk with each other. The wide area network connections using the ISSI provide an extended coverage area for subscriber units (SUs) that are roaming. The extended coverage area is important for public safety first responders that provide assistance in other jurisdictions during an emergency.
The ISSI supports the messaging, and procedures necessary to enable RFSSs to track and locate SUs, set up and teardown calls and transfer voice information to the SUs. The ISSI uses SIP and RTP protocols (standardized protocols) to provide the messaging between RFSSs.
The modern Land Mobile Radio (LMR) system includes many features in addition to voice communication. Many features will not work across systems connected using the ISSI. Whether a particular feature will work is determined by the systems and the particular ISSI implementation.
Overview of P25 ISSI Documentation Suite
The documentation suite which defines Scope One of the P25 ISSI consists of five standards.
ISSI Messages and Procedures for Voice Services
ISSI Measurement Methods for Voice Services
ISSI Performance Recommendations for Voice Services
ISSI Conformance Testing for Voice Services
ISSI Interoperability Testing for Voice Operations in Trunked Systems
A brief overview of each of these standards is provided in the subsections that follow.
ISSI Messages and Procedures for Voice Services
Scope One of the ISSI Messages and Procedures for Voice Services specifies the functional services of mobility management, call control and transmission control to provide trunked voice services for SU-to-SU and Group PTT calls involving multiple RFSSs.
Mobility management uses the SIP protocol and describes the messages and procedures necessary for RFSSs to perform |
https://en.wikipedia.org/wiki/Indium%20%28111In%29%20satumomab%20pendetide | {{DISPLAYTITLE:Indium (111In) satumomab pendetide}}
Indium (111In) satumomab pendetide (trade name OncoScint CR103) is a mouse monoclonal antibody which is used for cancer diagnosis. The antibody, satumomab, is linked to pendetide, a derivative of DTPA. Pendetide acts as a chelating agent for the radionuclide indium-111. |
https://en.wikipedia.org/wiki/Taplitumomab%20paptox | Taplitumomab paptox is a mouse monoclonal antibody. The antibody itself, taplitumomab, is linked to the protein PAP, an antiviral from Phytolacca americana, a species of pokeweed. This is reflected by the 'paptox' in the drug's name. |
https://en.wikipedia.org/wiki/Decision-to-decision%20path | A decision-to-decision path, or DD-path, is a path of execution (usually through a flow graph representing a program, such as a flow chart) between two decisions. More recent versions of the concept also include the decisions themselves in their own DD-paths.
Definition
In Huang's 1975 paper, a decision-to-decision path is defined as path in a program's flowchart such that all the following hold (quoting from the paper):
its first constituent edge emanates either from an entry node or a decision box;
its last constituent edge terminates either at a decision box or at an exit node; and
there are no decision boxes on the path except those at both ends
Jorgensen's more recent textbooks restate it in terms of a program's flow graph (called a "program graph" in that textbook). First define some preliminary notions: chain and a maximal chain. A chain is defined as a path in which:
initial and terminal nodes are distinct, and
all interior nodes have in-degree = 1 and out-degree = 1.
A maximal chain is a chain that is not part of a bigger chain.
A DD-path is a set of nodes in a program graph such that one of the following holds (quoting and keeping Jorgensen's numbering, with comments added in parentheses):
It consists of a single node with in-degree = 0 (initial node)
It consists of a single node with out-degree = 0 (terminal node)
It consists of a single node with in-degree ≥ 2 or out-degree ≥ 2 (decision/merge points)
It consists of a single node with in-degree = 1 and out-degree = 1
It is a maximal chain of length ≥ 1.
According to Jorgensen (2013), in Great Britain and ISTQB literature, the same notion is called linear code sequence and jump (LCSAJ).
Properties
From the latter definition (of Jorgensen) we can conclude the following:
Every node on a flow graph of a program belongs to one DD-path.
If the first node on a DD-path is traversed, then all other nodes on that path will also be traversed.
The DD path graph is used to find independent path f |
https://en.wikipedia.org/wiki/John%20Friedlander | John Friedlander is a Canadian mathematician specializing in analytic number theory. He received his B.Sc. from the University of Toronto in 1965, an M.A. from the University of Waterloo in 1966, and a Ph.D. from Pennsylvania State University in 1972. He was a lecturer at M.I.T. in 1974–76, and has been on the faculty of the University of Toronto since 1977, where he served as Chair during 1987–91. He has also spent several years at the Institute for Advanced Study. In addition to his individual work, he has been notable for his collaborations with other well-known number theorists, including Enrico Bombieri, William Duke, Andrew Granville, and especially Henryk Iwaniec.
In 1997, in joint work with Henryk Iwaniec, Friedlander proved that infinitely many prime numbers can be obtained as the sum of a square and fourth power: . Friedlander and Iwaniec improved Enrico Bombieri's "asymptotic sieve" technique to construct their proof.
Awards and honors
In 1999, Friedlander received the Jeffery–Williams Prize.
In 1988, Friedlander became a fellow of the Royal Society of Canada.
In 2002, CRM-Fields-PIMS prize
In 2012, he became a fellow of the American Mathematical Society.
In 2017, he received the Joseph L. Doob prize, jointly with Henryk Iwaniec, for their book Opera de Cribro.
Selected publications
See also
List of University of Waterloo people |
https://en.wikipedia.org/wiki/Technical%20features%20new%20to%20Windows%20Vista | Windows Vista (formerly codenamed Windows "Longhorn") has many significant new features compared with previous Microsoft Windows versions, covering most aspects of the operating system.
In addition to the new user interface, security capabilities, and developer technologies, several major components of the core operating system were redesigned, most notably the audio, print, display, and networking subsystems; while the results of this work will be visible to software developers, end-users will only see what appear to be evolutionary changes in the user interface.
As part of the redesign of the networking architecture, IPv6 has been incorporated into the operating system, and a number of performance improvements have been introduced, such as TCP window scaling. Prior versions of Windows typically needed third-party wireless networking software to work properly; this is no longer the case with Windows Vista, as it includes comprehensive wireless networking support.
For graphics, Windows Vista introduces a new as well as major revisions to Direct3D. The new display driver model facilitates the new Desktop Window Manager, which provides the tearing-free desktop and special effects that are the cornerstones of the Windows Aero graphical user interface. The new display driver model is also able to offload rudimentary tasks to the GPU, allow users to install drivers without requiring a system reboot, and seamlessly recover from rare driver errors due to illegal application behavior.
At the core of the operating system, many improvements have been made to the memory manager, process scheduler, heap manager, and I/O scheduler. A Kernel Transaction Manager has been implemented that can be used by data persistence services to enable atomic transactions. The service is being used to give applications the ability to work with the file system and registry using atomic transaction operations.
Audio
Windows Vista features a completely re-written audio stack designed to provi |
https://en.wikipedia.org/wiki/Gating%20signal | Signal gating is a concept commonly used in the field of electronics and signal processing. It refers to the process of controlling the flow of signals based on certain conditions or criteria. The goal of signal gating is to selectively allow or block the transmission of signals through a circuit or system.
In signal gating, a gating signal is used to modulate the passage of the main signal. The gating signal acts as a control mechanism, determining when the main signal can pass through the gate and when it is blocked. The gating signal can be generated by various means, such as an external trigger, a specific voltage level, or a specific frequency range.
Signal gating is often employed in applications where precise control over the transmission of signals is required. Here are a few examples of how signal gating is used in different fields:
1. Telecommunications: In telecommunications systems, signal gating is used to regulate the flow of data packets. By opening and closing the gate based on specific criteria, such as error detection or network congestion, signal gating helps ensure that the data is transmitted efficiently and reliably.
2. Audio processing: In audio applications, signal gating is used to reduce background noise or eliminate unwanted sounds. For example, in live sound reinforcement, a noise gate is often employed to mute or attenuate the microphone signal when the sound level falls below a certain threshold. This helps minimize the pickup of ambient noise and unwanted signals.
3. Radar systems: Signal gating plays a crucial role in radar systems, particularly in pulse-Doppler radar. Gating is used to control the transmission and reception of radar pulses, allowing the system to focus on specific ranges or angles of interest while ignoring other signals. This helps improve target detection and reduces interference from unwanted reflections.
4. Medical imaging: Signal gating is utilized in medical imaging techniques like computed tomography (CT |
https://en.wikipedia.org/wiki/Hay%20fever%20in%20Japan | is most commonly caused by pollen from Cryptomeria japonica (known as sugi in Japanese and often translated as "cedar" though it is not a member of the Cedrus genus) and Japanese cypress (known as hinoki), two native Japanese tree species.
According to the Ministry of the Environment, as of 2019, 42.5% of Japanese suffer from some form of hay fever, and 38.8% suffer from cedar pollinosis.
Cause
Hay fever was relatively uncommon in Japan until the early 1960s. Shortly after World War II, reforestation policies resulted in large forests of cryptomeria and Japanese cypress trees, which were an important resource for the construction industry. As these trees matured, they started to produce large amounts of pollen. Peak production of pollen occurs in trees of 30 years and older. As the Japanese economy developed in the 1970s and 1980s, cheaper imported building materials decreased the demand for cryptomeria and Japanese cypress materials. This resulted in increasing forest density and aging trees, further contributing to pollen production and thus, hay fever. In 1970, about 50% of cryptomeria were more than 10 years old, and just 25% were more than 20 years old. By 2000, almost 85% of cryptomeria were over 20 years old, and more than 60% of trees were over 30 years old. This cryptomeria aging trend has continued since then, and though cryptomeria forest acreage has hardly increased since 1980, pollen production has continued to increase. Furthermore, urbanization of land in Japan led to increasing coverage of soft soil and grass land by concrete and asphalt. Pollen settling on such hard surfaces can easily be swept up again by winds to recirculate and contribute to hay fever.
Pollen season
Cryptomeria pollen dispersal starts when average daily temperatures reach 10 degrees Celsius, partly depending on wind and terrain. Like the cherry blossom season, the pollen season moves from south to north across Japan, and from lower to higher elevations as spring progresses. Fo |
https://en.wikipedia.org/wiki/Riemann%E2%80%93Hilbert%20problem | In mathematics, Riemann–Hilbert problems, named after Bernhard Riemann and David Hilbert, are a class of problems that arise in the study of differential equations in the complex plane. Several existence theorems for Riemann–Hilbert problems have been produced by Mark Krein, Israel Gohberg and others (see the book by Clancey and Gohberg (1981)).
The Riemann problem
Suppose that is a closed simple contour in the complex plane dividing the plane into two parts denoted by (the inside) and (the outside), determined by the index of the contour with respect to a point. The classical problem, considered in Riemann's PhD dissertation (see ), was that of finding a function
analytic inside such that the boundary values of M+ along satisfy the equation
for all , where a, b, and c are given real-valued functions .
By the Riemann mapping theorem, it suffices to consider the case when is the unit circle . In this case, one may seek M+(z) along with its Schwarz reflection:
On the unit circle Σ, one has , and so
Hence the problem reduces to finding a pair of functions M+(z) and M−(z) analytic, respectively, on the inside and the outside of the unit disc, so that on the unit circle
and, moreover, so that the condition at infinity holds:
The Hilbert problem
Hilbert's generalization was to consider the problem of attempting to find M+ and M− analytic, respectively, on the inside and outside of the curve Σ, such that on one has
where α, β, and c are arbitrary given complex-valued functions (no longer just complex conjugates).
Riemann–Hilbert problems
In the Riemann problem as well as Hilbert's generalization, the contour was simple. A full Riemann–Hilbert problem allows that the contour may be composed of a union of several oriented smooth curves, with no intersections. The + and − sides of the "contour" may then be determined according to the index of a point with respect to . The Riemann–Hilbert problem is to find a pair of functions, M+ and M− analytic, respectivel |
https://en.wikipedia.org/wiki/Filled%20Julia%20set | The filled-in Julia set of a polynomial is a Julia set and its interior, non-escaping set
Formal definition
The filled-in Julia set of a polynomial is defined as the set of all points of the dynamical plane that have bounded orbit with respect to
where:
is the set of complex numbers
is the -fold composition of with itself = iteration of function
Relation to the Fatou set
The filled-in Julia set is the (absolute) complement of the attractive basin of infinity.
The attractive basin of infinity is one of the components of the Fatou set.
In other words, the filled-in Julia set is the complement of the unbounded Fatou component:
Relation between Julia, filled-in Julia set and attractive basin of infinity
The Julia set is the common boundary of the filled-in Julia set and the attractive basin of infinity
where: denotes the attractive basin of infinity = exterior of filled-in Julia set = set of escaping points for
If the filled-in Julia set has no interior then the Julia set coincides with the filled-in Julia set. This happens when all the critical points of are pre-periodic. Such critical points are often called Misiurewicz points.
Spine
The most studied polynomials are probably those of the form , which are often denoted by , where is any complex number. In this case, the spine of the filled Julia set is defined as arc between -fixed point and ,
with such properties:
spine lies inside . This makes sense when is connected and full
spine is invariant under 180 degree rotation,
spine is a finite topological tree,
Critical point always belongs to the spine.
-fixed point is a landing point of external ray of angle zero ,
is landing point of external ray .
Algorithms for constructing the spine:
detailed version is described by A. Douady
Simplified version of algorithm:
connect and within by an arc,
when has empty interior then arc is unique,
otherwise take the shortest way that contains .
Curve :
divides dynamical plane into two com |
https://en.wikipedia.org/wiki/Bing%20metrization%20theorem | In topology, the Bing metrization theorem, named after R. H. Bing, characterizes when a topological space is metrizable.
Formal statement
The theorem states that a topological space is metrizable if and only if it is regular and T0 and has a σ-discrete basis. A family of sets is called σ-discrete when it is a union of countably many discrete collections, where a family of subsets of a space is called discrete, when every point of has a neighborhood that intersects at most one member of
History
The theorem was proven by Bing in 1951 and was an independent discovery with the Nagata–Smirnov metrization theorem that was proved independently by both Nagata (1950) and Smirnov (1951). Both theorems are often merged in the Bing-Nagata-Smirnov metrization theorem. It is a common tool to prove other metrization theorems, e.g. the Moore metrization theorem – a collectionwise normal, Moore space is metrizable – is a direct consequence.
Comparison with other metrization theorems
Unlike the Urysohn's metrization theorem which provides a sufficient condition for metrization, this theorem provides both a necessary and sufficient condition for a topological space to be metrizable.
See also |
https://en.wikipedia.org/wiki/PCAF | P300/CBP-associated factor (PCAF), also known as K(lysine) acetyltransferase 2B (KAT2B), is a human gene and transcriptional coactivator associated with p53.
Structure
Several domains of PCAF can act independently or in unison to enable its functions. PCAF has separate acetyltransferase and E3 ubiquitin ligase domains as well as a bromodomain for interaction with other proteins. PCAF also possesses sites for its own acetylation and ubiquitination.
Function
CBP and p300 are large nuclear proteins that bind to many sequence-specific factors involved in cell growth and/or differentiation, including c-jun and the adenoviral oncoprotein E1A. The protein encoded by the PCAF gene associates with p300/CBP. It has in vitro and in vivo binding activity with CBP and p300, and competes with E1A for binding sites in p300/CBP. It has histone acetyl transferase activity with core histones and nucleosome core particles, indicating that this protein plays a direct role in transcriptional regulation.
Regulation
The acetyltransferase activity and cellular location of PCAF are regulated through acetylation of PCAF itself. PCAF may be autoacetylated (acetylated by itself) or by p300. Acetylation leads to migration to the nucleus and enhances its acetyltransferase activity. PCAF interacts with and is deacetylated by HDAC3, leading to a reduction in PCAF acetyltransferase activity and cytoplasmic localisation.
Protein interactions
PCAF forms complexes with numerous proteins that guide its activity. For example PCAF is recruited by ATF to acetylate histones and promote transcription of ATF4 target genes.
Targets
There are various protein targets of PCAF's acetyltransferase activity including transcription factors such as Fli1, p53 and numerous histone residues. Hdm2, itself a ubiquitin ligase that targets p53, has also been demonstrated to be a target of the ubiquitin-ligase activity of PCAF.
Interactions
PCAF has been shown to interact with:
BRCA2,
CTNNB1,
CREBBP,
EVI1,
|
https://en.wikipedia.org/wiki/MAD1L1 | Mitotic spindle assembly checkpoint protein MAD1 is a protein that in humans is encoded by the MAD1L1 gene.
MAD1L1 is also known as Human Accelerated Region 3. It may have played a key role in the evolution of humans from apes.
Function
MAD1L1 is a component of the mitotic spindle-assembly checkpoint that prevents the onset of anaphase until all chromosome are properly aligned at the metaphase plate. MAD1L1 functions as a homodimer and interacts with MAD2L1. MAD1L1 may play a role in cell cycle control and tumor suppression. Some studies indicate associations of MAD1L1 with psychiatric disorders, including schizophrenia, bipolar disorder, and depression. Three transcript variants encoding the same protein have been found for this gene.
Interactions
MAD1L1 has been shown to interact with:
HDAC1,
Histone deacetylase 2, and
MAD2L1,
See also
MAD1
MAD2
Hyperphosphorylation |
https://en.wikipedia.org/wiki/Uridine%20diphosphate%20glucose | Uridine diphosphate glucose (uracil-diphosphate glucose, UDP-glucose) is a nucleotide sugar. It is involved in glycosyltransferase reactions in metabolism.
Functions
UDP-glucose is used in nucleotide sugar metabolism as an activated form of glucose, a substrate for enzymes called glucosyltransferases.
UDP-glucose is a precursor of glycogen and can be converted into UDP-galactose and UDP-glucuronic acid, which can then be used as substrates by the enzymes that make polysaccharides containing galactose and glucuronic acid.
UDP-glucose can also be used as a precursor of sucrose, lipopolysaccharides and glycosphingolipids.
Components
UDP-glucose consists of the pyrophosphate group, ribose, glucose, and uracil.
See also
DNA
Nucleoside
Nucleotide
Oligonucleotide
RNA
TDP-glucose
Uracil
Uridine diphosphate |
https://en.wikipedia.org/wiki/List%20of%20features%20removed%20in%20Windows%20Vista | While Windows Vista contains many new features, a number of capabilities and certain programs that were a part of previous Windows versions up to Windows XP were removed or changed – some of which were later re-introduced in Windows 7 and later versions.
The following is a list of features that were present in Windows XP and earlier versions but were removed in Windows Vista.
Windows Explorer
Windows Briefcase no longer allows synchronizing items across multiple computers and a removable media device.
Windows Briefcase cannot sync files or folders in locations protected by User Account Control. This removes the ability to sync many locations.
Grouping items by name in Explorer no longer groups them under each individual letter of the alphabet (A, B, C... Z) like in Windows XP. When using Group By Name, items are always combined into just a few groups (A-H, I-P, Q-Z). This removes the ability to locate items by their first letter.
If hidden files are not allowed to be shown in Windows Explorer, the Status bar does not report how many hidden files are present. In addition, if all items within a folder are selected at once (by pressing or Select all), the user is not alerted to hidden files being selected.
Even after setting the ForceCopyAclwithFile and MoveSecurityAttributes values as documented in KB310316, permissions are not retained/copied when Windows Explorer is used to copy or move objects across volumes or in the same volume. A hotfix is available (KB2617058) to restore the MoveSecurityAttributes value but not ForceCopyAclwithFile.
Thumbnails can no longer be forced to regenerate by right-clicking the image and selecting Refresh thumbnail.
Thumbnail support for .HTM, .HTML, .MHT and .URL files has been removed in Windows Vista.
The Explorer thumbnail handler and metadata property handler for .AVI and .WAV files (Shmedia.dll) has been removed.
Ctrl+Enter on the selected folder no longer opens it in a new Explorer window.
Tiles view only shows the name, type |
https://en.wikipedia.org/wiki/EAN-8 | An EAN-8 is an EAN/UPC symbology barcode and is derived from the longer International Article Number (EAN-13) code. It was introduced for use on small packages where an EAN-13 barcode would be too large; for example on cigarettes, pencils, and chewing gum packets. It is encoded identically to the 12 digits of the UPC-A barcode, except that it has 4 (rather than 6) digits in each of the left and right halves.
EAN-8 barcodes may be used to encode GTIN-8 (8-digit Global Trade Identification Numbers) which are product identifiers from the GS1 System. A GTIN-8 begins with a 2- or 3-digit GS1 prefix (which is assigned to each national GS1 authority) followed by a 5- or 4-digit item reference element depending on the length of the GS1 prefix), and a checksum digit.
EAN-8 codes are common throughout the world, and companies may also use them to encode RCN-8 (8-digit Restricted Circulation Numbers), and use them to identify own-brand products sold only in their stores. RCN-8 are a subset of GTIN-8 which begin with a first digit of 0 or 2. |
https://en.wikipedia.org/wiki/Occipital%20emissary%20vein | The occipital emissary vein is a small emissary vein which passes through the condylar canal. |
https://en.wikipedia.org/wiki/Archaeocyte | Archaeocytes (from Greek archaios "beginning" and kytos "hollow vessel") or amoebocytes are amoeboid cells found in sponges. They are totipotent and have varied functions depending on the species.
The structure of these cells match to that of the stem cells as of containing high cytoplasmic content that helps the cells to morph according to their function.
Location
Archaeocytes are along with other specialized sponge cells including collencytes and structural elements called spicules. They move about within the mesohyl with amoeba-like movements performing a number of important functions.
Functions
Cellular differentiation is an essential function of the archaeocyte. All specialized cells within the sponge have its origins with the archaeocyte. This is especially important in reproduction as the sex cells of the sponge in sexual reproduction are formed from these amoeboid cells. Similarly in asexual reproduction amoebocytes result in the formation of gemmules which are cyst-like spheres containing more amoebocytes as well as other sponge cells including the phylum specific choanocyte. These cells move within the walls of a sponge and form spicules. |
https://en.wikipedia.org/wiki/Supraclavicular%20lymph%20nodes | Supraclavicular lymph nodes are lymph nodes found above the clavicle, that can be felt in the supraclavicular fossa. The supraclavicular lymph nodes on the left side are called Virchow's nodes. It leads to an appreciable mass that can be recognized clinically, called Troisier sign.
Structure
A Virchow's node is a left-sided supraclavicular lymph node.
Clinical significance
Malignancies of the internal organs can reach an advanced stage before giving symptoms. Stomach cancer, for example, can remain asymptomatic while metastasizing. One of the first visible spots where these tumors metastasize is one of the left supraclavicular lymph node.
Virchow's nodes take their supply from lymph vessels in the abdominal cavity, and are therefore sentinel lymph nodes of cancer in the abdomen, particularly gastric cancer, ovarian cancer, testicular cancer and kidney cancer, that has spread through the lymph vessels, and Hodgkin's lymphoma. Such spread typically results in Troisier's sign, which is the finding of an enlarged, hard Virchow's node.
The left supraclavicular nodes are the classical Virchow's node because they receive lymphatic drainage of most of the body (from the thoracic duct) and enters the venous circulation via the left subclavian vein. The metastasis may block the thoracic duct leading to regurgitation into the surrounding Virchow's nodes. Another concept is that one of the supraclavicular nodes corresponds to the end node along the thoracic duct and hence the enlargement.
Differential diagnosis of an enlarged Virchow's node includes lymphoma, various intra-abdominal malignancies, breast cancer, and infection (e.g. of the arm). Similarly, an enlarged right supraclavicular lymph node tends to drain thoracic malignancies such as lung and esophageal cancer, as well as Hodgkin's lymphoma.
History
Virchow's nodes are named after Rudolf Virchow (1821–1902), the German pathologist who first described the nodes and their association with gastric cancer in 1848. |
https://en.wikipedia.org/wiki/Superficial%20cervical%20lymph%20nodes | The superficial cervical lymph nodes are lymph nodes that lie near the surface of the neck.
Some sources state simply that they lie along the external jugular vein, while other sources state that they are only adjacent to the external jugular vein in the posterior triangle, and they are adjacent to the anterior jugular vein in the anterior triangle.
They can be broken down into:
superficial anterior cervical lymph nodes
superficial lateral cervical lymph nodes |
https://en.wikipedia.org/wiki/Retropharyngeal%20lymph%20nodes | The retropharyngeal lymph nodes, from one to three in number, lie in the buccopharyngeal fascia, behind the upper part of the pharynx and in front of the arch of the atlas, being separated, however, from the latter by the Longus capitis.
Their afferents drain the nasal cavities, the nasal part of the pharynx, and the auditory tubes.
Their efferents pass to the superior deep cervical lymph nodes.
They are in the retropharyngeal space.
They frequently disappear by age 4-5. (This is why retropharyngeal abscess is rare in older children.)
See also
Rouvière node |
https://en.wikipedia.org/wiki/Viacheslav%20Belavkin | Viacheslav Pavlovich Belavkin (; 20 May 1946 – 27 November 2012) was a Russian-British professor in applied mathematics at the University of Nottingham. An active researcher, he was one of the pioneers of quantum probability. His research spanned areas such as quantum filtering, quantum information and quantum chaos.
Biography
He was born in Lviv, and graduated from Moscow State University in 1970 where his teachers include Evgeny Lifshitz, Victor Pavlovich Maslov, Andrey Kolmogorov and Ruslan L. Stratonovich. In the 1980s Belavkin held visiting professorship in the Dublin Institute for Advanced Studies, and the Volterra Centre in Rome before taking up an appointment at the University of Nottingham in 1992. He was promoted to a Chair in Mathematical Physics in 1996. He and Ruslan L. Stratonovich were awarded the State Prize of the Russian Federation (formerly the Lenin Prize) for outstanding achievements in science and technology, in part due to his work on the measurement problem. He is survived by his wife Nadezda Belavkin and son Roman Belavkin. |
https://en.wikipedia.org/wiki/Buccinator%20lymph%20node | The buccinator lymph node or nodes are one or more lymph nodes placed on the Buccinator opposite the angle of the mouth. |
https://en.wikipedia.org/wiki/Myomesin | Myomesin is a protein family found in the M-line of the sarcomere structure. Myomesin has various forms throughout the body in striated muscles with specialized functions. This includes both slow and fast muscle fibers. Myomesin are made of 13 domains including a unique N-terminal followed by two immunoglobulin-like (Ig) domains, five fibronectin type III (Fn) domains, five more Ig domains. These domains all promote binding which indicates that myomesin is regulated through binding.
Functions
Sarcomere structure
Myomesin plays an important role in the structure of sarcomeres. They are found in the M-band region of the sarcomere, between the thick filaments (myosin). Its main purpose in this setting is to provide structural integrity by linking the antiparallel myosin fibers and titin filaments which are connected to the Z-discs. These myosin filaments form a hexagonal lattice with titin and myomesin. This shape allows the M-band to withstand large conformational changes during muscle contraction and return to their original shape upon relaxation. Since the Z-disc region of the sarcomere is very stiff and unable to bend for contraction, the elastic activity of myomesin in the M-band is what makes muscle contraction possible as it acts as a molecular spring.
Sarcomere assembly
In addition to sarcomere activity, it has been shown that myomesin also plays a role in the assembly of the sarcomere. In order for myomesin to be implemented into the sarcomere, myosin and titin must be present, indicating that myomesin is the last component to be added during assembly of the lattice. It is believed that this postponed addition is due to the role of myomesin to act as an "integrity check" to ensure the sarcomere has been formed correctly and monitor its integrity. This is extremely important as if even one piece of the M-line is missing, the A-band of the sarcomere will collapse and the muscle will be paralyzed.
Response to injury
Myomesin has also been shown to play a |
https://en.wikipedia.org/wiki/Submandibular%20lymph%20nodes | The submandibular lymph nodes (submaxillary glands in older texts), are some 3-6 lymph nodes situated at the inferior border of the ramus of mandible.
Anatomy
They are situated just superficial to the submandibular salivary gland, and posterolateral to the anterior belly of either digastric muscle.
One gland, the middle gland of Stahr, which lies on the facial artery as it turns over the mandible, is the most constant of the series; small lymph glands are sometimes found on the deep surface of the submandibular gland.
Afferents
They drain the upper lip, body of tongue, cheeks, anterior portion of the hard palate, and most teeth with their associated periodontium and gingiva (except for the mandibular incisor teeth and third molar teeth).
The facial and submental lymph nodes may also drain into the submandibular glands.
Efferents
They drain to the superior deep cervical lymph nodes.
Clinical significance
The most common causes of enlargement of the submandibular lymph nodes are infections of the head, neck, ears, eyes, nasal sinuses, pharynx, and scalp.
The lymph glands may be affected by metastatic spread of cancers of the oral cavity, anterior portion of the nasal cavity, soft tissues of the mid-face, and submandibular salivary gland.
Additional images |
https://en.wikipedia.org/wiki/Deep%20parotid%20lymph%20nodes | The deep parotid lymph nodes are lymph nodes found below the parotid gland.
The afferents of the subparotid glands drain the nasal part of the pharynx and the posterior parts of the nasal cavities.
Their efferents pass to the superior deep cervical glands.
Additional images |
https://en.wikipedia.org/wiki/Quantum%20state%20space | In physics, a quantum state space is an abstract space in which different "positions" represent, not literal locations, but rather quantum states of some physical system. It is the quantum analog of the phase space of classical mechanics.
Relative to Hilbert space
In quantum mechanics a state space is a complex Hilbert space in which each unit vector represents a different state that could come out of a measurement. The number of dimensions in this Hilbert space depends on the system we choose to describe. Any state vectors in this space can be written as a linear combination of unit vectors. Having an nonzero component along multiple dimensions is called a superposition. In the formalism of quantum mechanics these state vectors are often written using Dirac's compact bra–ket notation.
Examples
The spin (physics) state of a silver atom in the Stern-Gerlach experiment can be represented in a two state space. The spin can be aligned with a measuring apparatus (arbitrarily called 'up') or oppositely ('down'). In Dirac's notation these two states can be written as . The space of a two spin system has four states, .
The spin state is a discrete degree of freedom; quantum state spaces can have continuous degrees of freedom. For example, a particle in one space dimension has one degree of freedom ranging from to . In Dirac notation, the states in this space might be written as or .
Relative to 3D space
Even in the early days of quantum mechanics, the state space (or configurations as they were called at first) was understood to be essential for understanding simple QM problems. In 1929, Nevill Mott showed that "tendency to picture the wave as existing in ordinary three dimensional space, whereas we are really dealing with wave functions in multispace" makes analysis of simple interaction problems more difficult. Mott analyzes -particle emission in a cloud chamber. The emission process is isotropic, a spherical wave in QM, but the tracks observed are linear.
|
https://en.wikipedia.org/wiki/Eilenberg%27s%20inequality | Eilenberg's inequality, also known as the coarea inequality is a mathematical inequality for Lipschitz-continuous functions between metric spaces. Informally, it gives an upper bound on the average size of the fibers of a Lipschitz map in terms of the Lipschitz constant of the function and the measure of the domain.
The Eilenberg's inequality has applications in geometric measure theory and manifold theory. It is also a key ingredient in the proof of the coarea formula.
Formal statement
Let ƒ : X → Y be a Lipschitz-continuous function between metric spaces whose Lipschitz constant is denoted by Lip ƒ. Let s and t be nonnegative real numbers. Then, Eilenberg's inequality states that
for any A ⊂ X.
the asterisk denotes the upper integral,
vt are universal constants. If v=n, then vt equals the volume of the unit ball in Rn,
Ht is the t-dimensional Hausdorff measure.
The use of upper integral is necessary because in general the function
may fail to be Ht measurable.
History
The inequality was first proved by Eilenberg in 1938 for the case when the function was the distance to a fixed point in the metric space. Then it was generalized in 1943 by Eilenberg and Harold to the case of any real-valued Lipschitz function on a metric space.
The inequality in the form above was proved by Federer in 1954, except that he could prove it only under additional assumptions that he conjectured were unnecessary. Years later, Davies proved some deep results about Hausdorff contents and this conjecture was proved as a consequence. But recently a new proof, independent of Davies's result, has been found as well.
About the proof
In many texts the inequality is proved for the case where the target space is a Euclidean space or a manifold. This is because the isodiametric inequality is available (locally in the case of manifolds), which allows for a straightforward proof. The isodiametric inequality is not available in general metric spaces. The proof of Eilenberg's inequality |
https://en.wikipedia.org/wiki/Bonnesen%27s%20inequality | Bonnesen's inequality is an inequality relating the length, the area, the radius of the incircle and the radius of the circumcircle of a Jordan curve. It is a strengthening of the classical isoperimetric inequality.
More precisely, consider a planar simple closed curve of length bounding a domain of area . Let and denote the radii of the incircle and the circumcircle. Bonnesen proved the inequality
The term in the right hand side is known as the isoperimetric defect.
Loewner's torus inequality with isosystolic defect is a systolic analogue of Bonnesen's inequality. |
https://en.wikipedia.org/wiki/Reduction%20%28computability%20theory%29 | In computability theory, many reducibility relations (also called reductions, reducibilities, and notions of reducibility) are studied. They are motivated by the question: given sets and of natural numbers, is it possible to effectively convert a method for deciding membership in into a method for deciding membership in ? If the answer to this question is affirmative then is said to be reducible to .
The study of reducibility notions is motivated by the study of decision problems. For many notions of reducibility, if any noncomputable set is reducible to a set then must also be noncomputable. This gives a powerful technique for proving that many sets are noncomputable.
Reducibility relations
A reducibility relation is a binary relation on sets of natural numbers that is
Reflexive: Every set is reducible to itself.
Transitive: If a set is reducible to a set and is reducible to a set then is reducible to .
These two properties imply that reducibility is a preorder on the powerset of the natural numbers. Not all preorders are studied as reducibility notions, however. The notions studied in computability theory have the informal property that is reducible to if and only if any (possibly noneffective) decision procedure for can be effectively converted to a decision procedure for . The different reducibility relations vary in the methods they permit such a conversion process to use.
Degrees of a reducibility relation
Every reducibility relation (in fact, every preorder) induces an equivalence relation on the powerset of the natural numbers in which two sets are equivalent if and only if each one is reducible to the other. In computability theory, these equivalence classes are called the degrees of the reducibility relation. For example, the Turing degrees are the equivalence classes of sets of naturals induced by Turing reducibility.
The degrees of any reducibility relation are partially ordered by the relation in the following manner. L |
https://en.wikipedia.org/wiki/Goldbach%27s%20comet | Goldbach's comet is the name given to a plot of the function , the so-called Goldbach function . The function, studied in relation to Goldbach's conjecture, is defined for all even integers to be the number of different ways in which E can be expressed as the sum of two primes. For example, since 22 can be expressed as the sum of two primes in three different ways ().
The coloring of points in the above image is based on the value of modulo 3 with red points corresponding to 0 mod 3, blue points corresponding to 1 mod 3 and green points corresponding to 2 mod 3. In other words, the red points are multiples of 6, the blue points are multiples of 6 plus 2, and the green points are multiples of 6 plus 4.
Anatomy of the Goldbach Comet
An illuminating way of presenting the comet data is as a histogram. The function can be normalized by dividing by the locally averaged value of g, gav, taken over perhaps 1000 neighboring values of the even number E. The histogram can then be accumulated over a range of up to about 10% either side of a central E.
Such a histogram appears on the right. A series of well-defined peaks is evident. Each of these peaks can be identified as being formed by a set of values of which have certain smallest factors. The major peaks correspond to lowest factors of 3, 5, 7 ... as labeled. As the lowest factors become higher the peaks move left and eventually merge to give the lowest value primary peak.
There is in fact a hierarchy of peaks; the main peaks are composed of subsidiary peaks, with a succession of second smallest factors of . This hierarchy continues until all factors are exhausted.
The magnified section shows the succession of subsidiary peaks in more detail.
The relative location of the peaks follows from the form developed by Hardy and Littlewood:
where the product is taken over all primes p that are factors of . The factor on the right is Hardy–Littlewood's twin prime constant
Here the product is taken over |
https://en.wikipedia.org/wiki/Totivirus | Totivirus is a genus of double-stranded RNA viruses in the family Totiviridae. Fungi serve as natural hosts. The name of the group derives from Latin toti which means undivided or whole. There are seven species in this genus.
Structure
Viruses in the genus Totivirus are non-enveloped, with icosahedral symmetry, and T=2 architecture. The diameter is around 40 nm.
Genome
Totiviruses have a genome of 4700–6700 nucleotides in length and only a single copy of the genome is present in the particle. The nucleic acid content of a totivirus capsid is usually of one segment but can also contain three or four segments of linear double stranded RNA. The genome contains two large overlapping open reading frames (ORFs). These open reading frames (ORFs) code for a capsid protein (CP) and an RNA-dependent RNA polymerase (RdRp). The 5' end of the positive strand of the dsRNA genome has no cap and is very structured. Totiviruses contain a long 5' untranslated region (5' UTR) which functions as an internal ribosome entry site (IRES). Totiviruses can have satellite RNAs encoding a toxin.
Life cycle
Viral replication is cytoplasmic. Entry into the host cell is achieved by virus remains intracellular. Replication follows the double-stranded RNA virus replication model. Double-stranded RNA virus transcription is the method of transcription. Translation takes place by -1 ribosomal frameshifting. The virus exits the host cell by cell-to-cell movement. Fungi Saccharomyces cerevisiae and smut serve as the natural host. The virus is transmitted during cell division, sporogenesis, and cell fusion.
Taxonomy
The genus Totivirus contains the member species:
Saccharomyces cerevisiae virus L-A
Saccharomyces cerevisiae virus LBCLa
Scheffersomyces segobiensis virus L
Tuber aestivum virus 1
Ustilago maydis virus H1
Xanthophyllomyces dendrorhous virus L1A
Xanthophyllomyces dendrorhous virus L1B |
https://en.wikipedia.org/wiki/Shop%20drawing | A shop drawing is a drawing or set of drawings produced by the contractor, supplier, manufacturer, subcontractor, consultants, or fabricator. Shop drawings are typically required for prefabricated components. Examples of these include: elevators, structural steel, trusses, pre-cast concrete, windows, appliances, cabinets, air handling units, and millwork. Also critical are the installation and coordination shop drawings of the MEP trades such as sheet metal ductwork, piping, plumbing, fire protection, and electrical. Shop drawings are produced by contractors and suppliers under their contract with the owner. The shop drawing is the manufacturer’s or the contractor’s drawn version of information shown in the construction documents. The shop drawing normally shows more detail than the construction documents. It is drawn to explain the fabrication and/or installation of the items to the manufacturer’s production crew or contractor's installation crews. The style of the shop drawing is usually very different from that of the architect’s drawing. The shop drawing’s primary emphasis is on the particular product or installation and excludes notation concerning other products and installations, unless integration with the subject product is necessary.
Information required to be included in Shop Drawings
Comparison information for the architect and engineer
The shop drawings should include information for the architect and engineer to compare to the specifications and drawings. The shop drawing should address the appearance, performance, and prescriptive descriptions in the specifications and construction drawings.
The shop drawing often is more detailed than the information shown in the construction documents to give the architect and engineer the opportunity to review the fabricator’s version of the product, prior to fabrication. References to the construction documents, drawings, and specifications assist the architect and engineer in their review of the shop drawings. |
https://en.wikipedia.org/wiki/Kalmanson%20combinatorial%20conditions | In mathematics, the Kalmanson combinatorial conditions are a set of conditions on the distance matrix used in determining the solvability of the traveling salesman problem. These conditions apply to a special kind of cost matrix, the Kalmanson matrix, and are named after Kenneth Kalmanson. |
https://en.wikipedia.org/wiki/Worms%3F | Worms? is a software toy written by David Maynard for the Atari 8-bit family and ported to the Commodore 64. Published by Electronic Arts in 1983, it was one of initial batch of releases from the company. Worms? is an interactive version of Paterson's Worms.
Maynard later worked on the 8-bit word processor from Electronic Arts, Cut & Paste. On March 17, 2021, the source code to Worms? was made available under the MIT License.
Gameplay
The game is abstract, like Conway's Game of Life, but the player's ostensible goal is to optimally program one or more worms (each a sort of cellular automaton) to grow and survive as long as possible. The game area is divided up into hexagonal cells, and the worms are programmed to move in a particular direction for each combination of filled-in and empty frame segments in their immediate vicinity. Over the course of a game, the player needs to give the worm less and less input, and more and more moves by the worm results in the encountering of a familiar situation for which the worm has already been trained. As the worms move, they generate aleatoric music.
Development
When David Maynard finished developing the game for his Atari 800 in the fall of 1982, it was titled Sumo Worms. It was written in the Forth programming language.
Reception
Allen Doum reviewed the game for Computer Gaming World, and stated that "Worms? can be played competitively, either by teams or individuals or can be used solitaire as a pattern drawing puzzle. Its sound and graphics are excellent, and some of the graphics elements and the speed of the game can be varied." Orson Scott Card in Compute! gave complimentary reviews to the EA games Worms?, M.U.L.E., and Archon: The Light and the Dark. He said of the trio that "they are original; they do what they set out to do very, very well; they allow the player to take part in the creativity; they do things that only computers can do". William Michael Brown for Electronic Fun with Computers & Games praised its o |
https://en.wikipedia.org/wiki/Tiny%20Ionospheric%20Photometer | The tiny ionospheric photometer (TIP) is a small space-based photometer that observes the Earth's ionosphere at 135.6 nm. The TIP instruments were designed and built by the US Naval Research Laboratory (NRL) and are a part of the COSMIC program.
Operation
Although each TIP instrument is fairly simple in design and operation, the value of this instrument is that six of them were launched at once, and they observe the Earth simultaneously from three orbital planes spaced equally apart around the Earth. The data of this instrument when combined with the data from the other COSMIC payloads allows a 3D tomographic analysis of the Earth's ionosphere to be performed.
See also
Constellation Observing System for Meteorology, Ionosphere, and Climate (COSMIC) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.