source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Don%27t%20repeat%20yourself
"Don't repeat yourself" (DRY) is a principle of software development aimed at reducing repetition of information which is likely to change, replacing it with abstractions that are less likely to change, or using data normalization which avoids redundancy in the first place. The DRY principle is stated as "Every piece of knowledge must have a single, unambiguous, authoritative representation within a system". The principle has been formulated by Andy Hunt and Dave Thomas in their book The Pragmatic Programmer. They apply it quite broadly to include database schemas, test plans, the build system, even documentation. When the DRY principle is applied successfully, a modification of any single element of a system does not require a change in other logically unrelated elements. Additionally, elements that are logically related all change predictably and uniformly, and are thus kept in sync. Besides using methods and subroutines in their code, Thomas and Hunt rely on code generators, automatic build systems, and scripting languages to observe the DRY principle across layers. Single choice principle A particular case of DRY is the single choice principle. It was defined by Bertrand Meyer as: "Whenever a software system must support a set of alternatives, one and only one module in the system should know their exhaustive list." It was applied when designing Eiffel. Alternatives WET The opposing view to DRY is called WET, a backronym commonly taken to stand for write everything twice (alternatively write every time, we enjoy typing or waste everyone's time). WET solutions are common in multi-tiered architectures where a developer may be tasked with, for example, adding a comment field on a form in a web application. The text string "comment" might be repeated in the label, the HTML tag, in a read function name, a private variable, database DDL, queries, and so on. A DRY approach eliminates that redundancy by using frameworks that reduce or eliminate all those editing t
https://en.wikipedia.org/wiki/Ligand-gated%20ion%20channel
Ligand-gated ion channels (LICs, LGIC), also commonly referred to as ionotropic receptors, are a group of transmembrane ion-channel proteins which open to allow ions such as Na+, K+, Ca2+, and/or Cl− to pass through the membrane in response to the binding of a chemical messenger (i.e. a ligand), such as a neurotransmitter. When a presynaptic neuron is excited, it releases a neurotransmitter from vesicles into the synaptic cleft. The neurotransmitter then binds to receptors located on the postsynaptic neuron. If these receptors are ligand-gated ion channels, a resulting conformational change opens the ion channels, which leads to a flow of ions across the cell membrane. This, in turn, results in either a depolarization, for an excitatory receptor response, or a hyperpolarization, for an inhibitory response. These receptor proteins are typically composed of at least two different domains: a transmembrane domain which includes the ion pore, and an extracellular domain which includes the ligand binding location (an allosteric binding site). This modularity has enabled a 'divide and conquer' approach to finding the structure of the proteins (crystallising each domain separately). The function of such receptors located at synapses is to convert the chemical signal of presynaptically released neurotransmitter directly and very quickly into a postsynaptic electrical signal. Many LICs are additionally modulated by allosteric ligands, by channel blockers, ions, or the membrane potential. LICs are classified into three superfamilies which lack evolutionary relationship: cys-loop receptors, ionotropic glutamate receptors and ATP-gated channels. Cys-loop receptors The cys-loop receptors are named after a characteristic loop formed by a disulfide bond between two cysteine residues in the N terminal extracellular domain. They are part of a larger family of pentameric ligand-gated ion channels that usually lack this disulfide bond, hence the tentative name "Pro-loop receptors"
https://en.wikipedia.org/wiki/IBM%20hexadecimal%20floating-point
Hexadecimal floating point (now called HFP by IBM) is a format for encoding floating-point numbers first introduced on the IBM System/360 computers, and supported on subsequent machines based on that architecture, as well as machines which were intended to be application-compatible with System/360. In comparison to IEEE 754 floating point, the HFP format has a longer significand, and a shorter exponent. All HFP formats have 7 bits of exponent with a bias of 64. The normalized range of representable numbers is from 16−65 to 1663 (approx. 5.39761 × 10−79 to 7.237005 × 1075). The number is represented as the following formula: (−1)sign × 0.significand × 16exponent−64. Single-precision 32-bit A single-precision HFP number (called "short" by IBM) is stored in a 32-bit word: {| |- style="text-align:center" |style="width:20px"|1 |style="width:20px"| |style="width:50px"|7 |style="width:20px"| |style="width:20px"| |style="width:210px"|24 |style="width:20px"| |style="text-align:left"|(width in bits) |- style="text-align:center" |colspan="1" style="text-align:center;background-color:#FC9"|S |colspan="3" style="text-align:center;background-color:#99F"|Exp |colspan="3" style="text-align:center;background-color:#9F9"|Fraction |colspan="1" style="text-align:center;background-color:#FFF"|  |- style="text-align:center" |31 |30 |... |24 |23 |... |0 |align="left"|(bit index)* |- |colspan="8"| * IBM documentation numbers the bits from left to right, so that the most significant bit is designated as bit number 0. |} In this format the initial bit is not suppressed, and the radix (hexadecimal) point is set to the left of the significand (fraction in IBM documentation and the figures). Since the base is 16, the exponent in this form is about twice as large as the equivalent in IEEE 754, in order to have similar exponent range in binary, 9 exponent bits would be required. Example Consider encoding the value −118.625 as an HFP single-precision floating-point value. The value is ne
https://en.wikipedia.org/wiki/History%20of%20quantum%20mechanics
The history of quantum mechanics is a fundamental part of the history of modern physics. The major chapters of this history begin with the emergence of quantum ideas to explain individual phenomena -- blackbody radiation, the photoelectric effect, solar emission spectra -- an era called the Old or Older quantum theories. The invention of wave mechanics by Schrodinger and expanded by many others triggers the "modern" era beginning around 1925. Dirac's relativistic quantum theory work lead him to explore quantum theories of radiation, culminating in quantum electrodynamics, the first quantum field theory. The history of quantum mechanics continues in the history of quantum field theory. The history of quantum chemistry, theoretical basis of chemical structure, reactivity, and bonding, interlaces with the events discussed in this article. The phrase "quantum mechanics" was coined (in German, Quantenmechanik) by the group of physicists including Max Born, Werner Heisenberg, and Wolfgang Pauli, at the University of Göttingen in the early 1920s, and was first used in Born's 1925 paper "Zur Quantenmechanik". The word quantum comes from the Latin word for "how much" (as does quantity). Something that is quantized, as the energy of Planck's harmonic oscillators, can only take specific values. For example, in most countries, money is effectively quantized, with the quantum of money being the lowest-value coin in circulation. Mechanics is the branch of science that deals with the action of forces on objects. So, quantum mechanics is the part of mechanics that deals with objects for which particular properties are quantized. Triumph and trouble at the end of the classical era The discoveries of the 19th century, both the successes and failures, set the stage for the emergence of quantum mechanics. Wave theory of light Beginning in 1670 and progressing over three decades, Isaac Newton developed and championed his corpuscular theory, arguing that the perfectly straight li
https://en.wikipedia.org/wiki/Shenzhen%20Kangtai%20Biological%20Products
Shenzhen Kangtai Biological Products (, ), also known as BioKangtai (), is a Chinese biopharmaceutical company that develops, manufactures and markets vaccines and other vaccine products. Kangtai develops the Minhai COVID-19 vaccine and has the exclusive right to manufacture the Oxford–AstraZeneca COVID-19 vaccine in mainland China.
https://en.wikipedia.org/wiki/Relative%20canonical%20model
In the mathematical field of algebraic geometry, the relative canonical model of a singular variety of a mathematical object where is a particular canonical variety that maps to , which simplifies the structure. Description The precise definition is: If is a resolution define the adjunction sequence to be the sequence of subsheaves if is invertible where is the higher adjunction ideal. Problem. Is finitely generated? If this is true then is called the relative canonical model of , or the canonical blow-up of . Some basic properties were as follows: The relative canonical model was independent of the choice of resolution. Some integer multiple of the canonical divisor of the relative canonical model was Cartier and the number of exceptional components where this agrees with the same multiple of the canonical divisor of Y is also independent of the choice of Y. When it equals the number of components of Y it was called crepant. It was not known whether relative canonical models were Cohen–Macaulay. Because the relative canonical model is independent of , most authors simplify the terminology, referring to it as the relative canonical model of rather than either the relative canonical model of or the canonical blow-up of . The class of varieties that are relative canonical models have canonical singularities. Since that time in the 1970s other mathematicians solved affirmatively the problem of whether they are Cohen–Macaulay. The minimal model program started by Shigefumi Mori proved that the sheaf in the definition always is finitely generated and therefore that relative canonical models always exist.
https://en.wikipedia.org/wiki/Sound%20Retrieval%20System
Sound Retrieval System (SRS) is a patented psychoacoustic 3D audio processing technology originally invented by Arnold Klayman in the early 1980s. The SRS technology applies head-related transfer functions (HRTFs) to create an immersive 3D soundfield using only two speakers, widening the "sweet spot", creating a more spacious sense of ambience, and producing strong localization cues for discrete instruments within an audio mix. SRS is not a Dolby matrix surround decoder but works with normal stereo recordings. History and technology Initially Hughes Aircraft, for whom Klayman was doing acoustic consulting at the time, offered a standalone SRS audio processor, as well as licensing the technology to Sony and Thomson (RCA) for inclusion in their products. In the early 1990s, Hughes sold off its non-aerospace-related holdings, and a group of entrepreneurs formed SRS Labs to acquire the SRS technology. Many TV sets employ built-in SRS to make their built-in audio systems sound "bigger". An article in the November 1994 issue of Consumers Digest magazine tested several SRS-equipped sets from Sony and other manufacturers and concluded that the circuit was essentially a gimmick in these products due to their small, close-set speakers and low-wattage amplifiers. SRS is not a panacea for audio systems that are marginal to begin with; it works best with full-range, high-fidelity sound reproduction. The company (publicly traded on NASDAQ under the symbol SRSL after a 1996 IPO) since developed or acquired several additional audio technologies, including SRS Headphone, TruSurround XT, TruBass ("psychoacoustic bass enhancement to enable deeper, natural bass of audio source material to be perceived over small speaker drivers"), FOCUS ("sound-image elevation used in combination with SRS to create a large sound image", originally conceived for in-car listening), SRS Virtual Surround, Circle Surround, SRS Wow (an audio-enhancement suite made up of SRS 3D "wide stereo imaging and no
https://en.wikipedia.org/wiki/Flyback%20diode
A flyback diode is any diode connected across an inductor used to eliminate flyback, which is the sudden voltage spike seen across an inductive load when its supply current is suddenly reduced or interrupted. It is used in circuits in which inductive loads are controlled by switches, and in switching power supplies and inverters. Flyback circuits have been used since 1930 and were refined starting in 1950 for use in television receivers. The word flyback comes from the horizontal movement of the electron beam in a cathode ray tube, because the beam flew back to begin the next horizontal line. This diode is known by many other names, such as snubber diode, commutating diode, freewheeling diode, suppressor diode, clamp diode, or catch diode. Operation Fig. 1 shows an inductor connected to a battery - a constant voltage source. The resistor represents the small residual resistance of the inductor's wire windings. When the switch is closed, the voltage from the battery is applied to the inductor, causing current from the battery's positive terminal to flow down through the inductor and resistor. The increase in current causes a back EMF (voltage) across the inductor due to Faraday's law of induction which opposes the change in current. Since the voltage across the inductor is limited to the battery's voltage of 24 volts, the rate of increase of the current is limited to an initial value of so the current through the inductor increases slowly as energy from the battery is stored in the inductor's magnetic field. As the current rises, more voltage is dropped across the resistor and less across the inductor, until the current reaches a steady value of with all the battery voltage across the resistance and none across the inductance. However, the current drops rapidly when the switch is opened in Fig. 2. The inductor resists the drop in current by developing a very large induced voltage of polarity in the opposite direction of the battery, positive at the lower end
https://en.wikipedia.org/wiki/Animal%20hoarding
Animal hoarding, sometimes called Noah syndrome, is keeping a higher-than-usual number of animals as domestic pets without the ability to properly house or care for them, while at the same time denying this inability. Compulsive hoarding can be characterized as a symptom of a mental disorder rather than deliberate cruelty towards animals. Hoarders are deeply attached to their pets and find it extremely difficult to let the pets go. They typically cannot comprehend that they are harming their pets by failing to provide them with proper care. Hoarders tend to believe that they provide the right amount of care for them. The American Society for the Prevention of Cruelty to Animals provides a "Hoarding Prevention Team", which works with hoarders to help them attain a manageable and healthy number of pets. Characteristics of a hoarder An animal hoarder keeps an unusually large number of pets for their premises, and fails to care for them properly. A hoarder is distinguished from an animal breeder, who would have numerous animals as the central component of their business; this distinction can be problematic, however, as some hoarders are former breeders who have ceased selling and caring for their animals, while others will claim to be breeders as a psychological defense mechanism, or in hopes of forestalling intervention. Gary Patronek, director of the Center for Animals and Public Policy at Tufts University, defines hoarding as the "pathological human behavior that involves a compulsive need to obtain and control animals, coupled with a failure to recognize their suffering". According to another study, the distinguishing feature is that a hoarder "fails to provide the animals with adequate food, water, sanitation, and veterinary care, and... is in denial about this inability to provide adequate care." Along with other compulsive hoarding behaviors, it is linked in the DSM-IV to obsessive–compulsive disorder and obsessive–compulsive personality disorder. The DSM-5 incl
https://en.wikipedia.org/wiki/%C3%96kologisches%20Wirtschaften
Ökologisches Wirtschaften is an academic journal for socioeconomics and ecological economics. The journal was introduced in 1986 by (IÖW) and (VÖW). Since 1996 it has been published four times a year with a focus on a specific topic by , Munich. The journal relates new research approaches to practical experience in politics and business. Discussions of the conflict between economy, ecology and society, and new ideas for a future-oriented, sustainable economy are presented. In the archive, all articles published since 1986 are available online. See also journal Ecological Economics External links Ökologisches Wirtschaften online – Open Access Portal Academic journals established in 1986 Ecology journals German economics journals Academic journals of Germany German-language journals Quarterly journals
https://en.wikipedia.org/wiki/Standard%20flowgram%20format
Standard flowgram format (SFF) is a binary file format used to encode results of pyrosequencing from the 454 Life Sciences platform for high-throughput sequencing. SFF files can be viewed, edited and converted with DNA Baser SFF Workbench (graphic tool), or converted to FASTQ format with sff2fastq or seq_crumbs. Further reading NCBI reference for SFF format Biotechnology companies of the United States Computer file formats DNA sequencing
https://en.wikipedia.org/wiki/Open%20Bioinformatics%20Foundation
The Open Bioinformatics Foundation is a non-profit, volunteer-run organization focused on supporting open source programming in bioinformatics. The mission of the foundation is to support the development of open source toolkits for bioinformatics, organise developer-centric hackathon events and generally assist in the development and promotion of open source software development in the life sciences. The foundation also organises and runs the annual Bioinformatics Open Source Conference, a satellite meeting of the Intelligent Systems for Molecular Biology conference. The foundation participates in the Google Summer of Code, acting as an umbrella organisation for individual bioinformatics-related projects. The Open Bioinformatics Foundation was started in 2001, arising from the BioJava, BioPerl and BioPython projects. A formal membership for the foundation was created in 2005. In October 2012, the foundation began an association with Software in the Public Interest (SPI), a US-based non-profit which aids other organizations in the creation and distribution of free and open-source software. The association with SPI allows financial donations to the foundation (these are 501(c)3 tax-exempt in the US). The foundation is governed by a board of directors, representing various Bio* projects. As of 2019, the OBF President is Peter Cock (BioPython). Previous OBF presidents include Ewan Birney and Hilmar Lapp (NESCent), previous Board members include Steven E. Brenner. Projects The foundation hosts servers for mailing lists, websites, and code repositories for a number of bioinformatics-related open source projects, including: BioJava – Java toolkit BioMOBY – Data and application execution through web services BioPerl – Perl toolkit BioPython – Python toolkit BioRuby – Ruby toolkit BioPHP EMBOSS – Sequence analysis toolkit. See also List of open-source bioinformatics software Generic Model Organism Database
https://en.wikipedia.org/wiki/Pleistocene%20Park
Pleistocene Park () is a nature reserve on the Kolyma River south of Chersky in the Sakha Republic, Russia, in northeastern Siberia, where an attempt is being made to re-create the northern subarctic steppe grassland ecosystem that flourished in the area during the last glacial period. The project is being led by Russian scientists Sergey Zimov and Nikita Zimov, testing the hypothesis that repopulating with large herbivores (and predators) can restore rich grasslands ecosystems, as expected if overhunting, and not climate change, was primarily responsible for the extinction of wildlife and the disappearance of the grasslands at the end of the Pleistocene epoch. The aim of the project is to research the climatic effects of the expected changes in the ecosystem. Here the hypothesis is that the change from tundra to grassland will result in a raised ratio of energy emission to energy absorption of the area, leading to less thawing of permafrost and thereby less emission of greenhouse gases. It is also thought that removal of snow by large herbivores will further reduce the permafrost's insulation. To study this, large herbivores have been released, and their effect on the local fauna is being monitored. Preliminary results point at the ecologically low-grade tundra biome being converted into a productive grassland biome and at the energy emission of the area being raised. Research goals Effects of large herbivores on the arctic tundra/grasslands ecosystem The primary aim of Pleistocene Park is to recreate the mammoth steppe (ancient taiga/tundra grasslands that were widespread in the region during the last ice age). The key concept is that animals, rather than climate, maintained that ecosystem. Reintroducing large herbivores to Siberia would then initiate a positive feedback loop promoting the reestablishment of grassland ecosystems. This argument is the basis for rewilding Pleistocene Park's landscape with megafauna that were previously abundant in the area, as
https://en.wikipedia.org/wiki/Rodos%20%28operating%20system%29
Rodos (Realtime Onboard Dependable Operating System) is a real-time operating system for embedded systems and was designed for application domains demanding high dependability. History Rodos was developed at the German Aerospace Center and has its roots in the operating system BOSS. It is used for the current micro satellite program of the German Aerospace Center. The system runs on the operational satellite TET-1 and will be used for the currently developed satellite BiROS. Rodos is further enhanced and extended at the German Aerospace Center as well as the department for aerospace information technology at the University of Würzburg. Features An important aspect of Rodos is its integrated real time middleware. Developing the control and payload software on the top of a middleware provides the maximum of modularity today. Applications/modules can be developed independently and it is very simple to interchange modules later without worrying about side effects, because all modules are encapsulated as Building Blocks (BB) and can be accessed and they can access other resources only by well defined interfaces. Rodos was implemented as a software framework in C++ with an object oriented application interface (API). it is organized in layers: The lowest layer (1) is responsible for control of the embedded system hardware (HAL: Hardware abstraction layer). The next layer (2) kernel: administrates the local resources, threads and time. On top of the kernel we have the middleware (layer 3) which enables communication between BBs using a publisher subscriber multicast protocol. And on the top of the middleware the user may implement his applications (layer 4) as a distributed software network of simple BBs. The Building Blocks API on the top of the middleware is a service oriented interface. BBs interact by providing services to other BBs and using services from other BBs. As mentioned before, the original purpose of Rodos was to control satellites. It was designe
https://en.wikipedia.org/wiki/Quadrupole
A quadrupole or quadrapole is one of a sequence of configurations of things like electric charge or current, or gravitational mass that can exist in ideal form, but it is usually just part of a multipole expansion of a more complex structure reflecting various orders of complexity. Mathematical definition The quadrupole moment tensor Q is a rank-two tensor—3×3 matrix. There are several definitions, but it is normally stated in the traceless form (i.e. ). The quadrupole moment tensor has thus nine components, but because of transposition symmetry and zero-trace property, in this form only five of these are independent. For a discrete system of point charges or masses in the case of a gravitational quadrupole, each with charge , or mass , and position relative to the coordinate system origin, the components of the Q matrix are defined by: The indices run over the Cartesian coordinates and is the Kronecker delta. This means that must be equal, up to sign, to distances from the point to mutually perpendicular hyperplanes for the Kronecker delta to equal 1. In the non-traceless form, the quadrupole moment is sometimes stated as: with this form seeing some usage in the literature regarding the fast multipole method. Conversion between these two forms can be easily achieved using a detracing operator. For a continuous system with charge density, or mass density, , the components of Q are defined by integral over the Cartesian space r: As with any multipole moment, if a lower-order moment, monopole or dipole in this case, is non-zero, then the value of the quadrupole moment depends on the choice of the coordinate origin. For example, a dipole of two opposite-sign, same-strength point charges, which has no monopole moment, can have a nonzero quadrupole moment if the origin is shifted away from the center of the configuration exactly between the two charges; or the quadrupole moment can be reduced to zero with the origin at the center. In contrast,
https://en.wikipedia.org/wiki/Penrose%20tiling
A Penrose tiling is an example of an aperiodic tiling. Here, a tiling is a covering of the plane by non-overlapping polygons or other shapes, and a tiling is aperiodic if it does not contain arbitrarily large periodic regions or patches. However, despite their lack of translational symmetry, Penrose tilings may have both reflection symmetry and fivefold rotational symmetry. Penrose tilings are named after mathematician and physicist Roger Penrose, who investigated them in the 1970s. There are several different variations of Penrose tilings with different tile shapes. The original form of Penrose tiling used tiles of four different shapes, but this was later reduced to only two shapes: either two different rhombi, or two different quadrilaterals called kites and darts. The Penrose tilings are obtained by constraining the ways in which these shapes are allowed to fit together in a way that avoids periodic tiling. This may be done in several different ways, including matching rules, substitution tiling or finite subdivision rules, cut and project schemes, and coverings. Even constrained in this manner, each variation yields infinitely many different Penrose tilings. Penrose tilings are self-similar: they may be converted to equivalent Penrose tilings with different sizes of tiles, using processes called inflation and deflation. The pattern represented by every finite patch of tiles in a Penrose tiling occurs infinitely many times throughout the tiling. They are quasicrystals: implemented as a physical structure a Penrose tiling will produce diffraction patterns with Bragg peaks and five-fold symmetry, revealing the repeated patterns and fixed orientations of its tiles. The study of these tilings has been important in the understanding of physical materials that also form quasicrystals. Penrose tilings have also been applied in architecture and decoration, as in the floor tiling shown. Background and history Periodic and aperiodic tilings Covering a flat surface ("
https://en.wikipedia.org/wiki/Nico%20Habermann
Arie Nicolaas Habermann (26 June 1932 – 8 August 1993), often known as Nico Habermann, was a noted Dutch computer scientist. Habermann was born in Groningen, Netherlands, and earned his B.S. in mathematics and physics and M.S. in mathematics from the Free University of Amsterdam in 1953 and 1958. After working as a mathematics teacher, in 1967 he received his Ph.D. in applied mathematics from the Eindhoven University of Technology under advisor Edsger Dijkstra. In 1968, Habermann was invited to join the department of computer science at Carnegie Mellon University as a visiting research scientist. In 1969 he was appointed an associate professor, and was made full professor in 1974, acting department head in 1979, and department head from 1980 to 1988, after which he was named Dean of the new School of Computer Science (established under Allen Newell and Herbert A. Simon). He also cofounded Carnegie Mellon's Software Engineering Institute (SEI) in 1985. Habermann's research included programming languages, operating systems, and development of large software systems. He was known for his work on inter-process communication, process synchronization and deadlock avoidance, and software verification, but particularly for the programming languages ALGOL 60, BLISS, Pascal, and Ada. He also contributed to new operating systems such as Edsger Dijkstra's THE multiprogramming system, the Family of Operating Systems (FAMOS) at Carnegie Mellon, Berlin's Dynamically Adaptable System (DAS), and Unix. Habermann served as visiting professor at the University of Newcastle upon Tyne (1973) and the Technical University of Berlin (1976), and as adjunct professor at Shanghai Jiao Tong University (1986–1993). In 1994, the Computing Research Association began giving the A. Nico Habermann Award to people for work that increases the involvement of underrepresented communities in computer research.
https://en.wikipedia.org/wiki/Calabi%20conjecture
In the mathematical field of differential geometry, the Calabi conjecture was a conjecture about the existence of certain kinds of Riemannian metrics on certain complex manifolds, made by . It was proved by , who received the Fields Medal and Oswald Veblen Prize in part for his proof. His work, principally an analysis of an elliptic partial differential equation known as the complex Monge–Ampère equation, was an influential early result in the field of geometric analysis. More precisely, Calabi's conjecture asserts the resolution of the prescribed Ricci curvature problem within the setting of Kähler metrics on closed complex manifolds. According to Chern–Weil theory, the Ricci form of any such metric is a closed differential 2-form which represents the first Chern class. Calabi conjectured that for any such differential form , there is exactly one Kähler metric in each Kähler class whose Ricci form is . (Some compact complex manifolds admit no Kähler classes, in which case the conjecture is vacuous.) In the special case that the first Chern class vanishes, this implies that each Kähler class contains exactly one Ricci-flat metric. These are often called Calabi–Yau manifolds. However, the term is often used in slightly different ways by various authors — for example, some uses may refer to the complex manifold while others might refer to a complex manifold together with a particular Ricci-flat Kähler metric. This special case can equivalently be regarded as the complete existence and uniqueness theory for Kähler–Einstein metrics of zero scalar curvature on compact complex manifolds. The case of nonzero scalar curvature does not follow as a special case of Calabi's conjecture, since the 'right-hand side' of the Kähler–Einstein problem depends on the 'unknown' metric, thereby placing the Kähler–Einstein problem outside the domain of prescribing Ricci curvature. However, Yau's analysis of the complex Monge–Ampère equation in resolving the Calabi conjecture was suffic
https://en.wikipedia.org/wiki/Scattering%20parameters
Scattering parameters or S-parameters (the elements of a scattering matrix or S-matrix) describe the electrical behavior of linear electrical networks when undergoing various steady state stimuli by electrical signals. The parameters are useful for several branches of electrical engineering, including electronics, communication systems design, and especially for microwave engineering. The S-parameters are members of a family of similar parameters, other examples being: Y-parameters, Z-parameters, H-parameters, T-parameters or ABCD-parameters. They differ from these, in the sense that S-parameters do not use open or short circuit conditions to characterize a linear electrical network; instead, matched loads are used. These terminations are much easier to use at high signal frequencies than open-circuit and short-circuit terminations. Contrary to popular belief, the quantities are not measured in terms of power (except in now-obsolete six-port network analyzers). Modern vector network analyzers measure amplitude and phase of voltage traveling wave phasors using essentially the same circuit as that used for the demodulation of digitally modulated wireless signals. Many electrical properties of networks of components (inductors, capacitors, resistors) may be expressed using S-parameters, such as gain, return loss, voltage standing wave ratio (VSWR), reflection coefficient and amplifier stability. The term 'scattering' is more common to optical engineering than RF engineering, referring to the effect observed when a plane electromagnetic wave is incident on an obstruction or passes across dissimilar dielectric media. In the context of S-parameters, scattering refers to the way in which the traveling currents and voltages in a transmission line are affected when they meet a discontinuity caused by the insertion of a network into the transmission line. This is equivalent to the wave meeting an impedance differing from the line's characteristic impedance. Although appli
https://en.wikipedia.org/wiki/Roussos%20Dimitrakopoulos
Roussos Dimitrakopoulos is a Canadian geoscientist, and a professor at McGill University. He was selected to receive the William Christian Krumbein Medal in 2018 from the International Association for Mathematical Geosciences. He was also the recipient of the Georges Matheron Lectureship Award 2015 from the International Association for Mathematical Geosciences. Since 2007 Dimitrakopoulos has been the Editor-in-Chief of Mathematical Geosciences. Research Selected books
https://en.wikipedia.org/wiki/2345.com
2345.com is a Chinese web directory founded in 2005. The website is the second most used web directory in China. It is ranked 47th place in China and has a worldwide ranking of 419 on Alexa, however, when downloaded, contains adware. It is hosted at Abitcool China Inc. Beijing, China.
https://en.wikipedia.org/wiki/SnapLogic
SnapLogic is a commercial software company that provides integration platform as a service (iPaaS) tools for connecting cloud data sources, SaaS applications and on-premises business software applications. SnapLogic was founded in 2006, and its headquarters are in San Mateo, California. SnapLogic is headed by Ex-CEO and co-founder of Informatica Gaurav Dhillon, and is venture-backed by Andreessen Horowitz, Ignition Partners, Floodgate Fund, Brian McClendon, and Naval Ravikant. On December 10, 2015, SnapLogic announced a $37.5 million funding round led by Microsoft and Silver Lake Waterman along with existing investors Andreessen Horowitz, Ignition Partners, and Triangle Peak Partners. On December 13, 2021, SnapLogic raised $160 milion funding at a valuation of $1 Billion Valuation. Products SnapLogic's Elastic Integration Platform consists of an Integration Cloud, prebuilt connectors called Snaps, and a Snaplex for data processing in the cloud or behind the firewall. The company's products have been referred to as targeting the Internet of Things marketplace for connecting data, applications, and devices. The Integration Cloud approaches big data integration through the following tools: Designer: An HTML5-based user interface for specifying and building integration workflows, called pipelines. Manager: Controls and monitors the performance of SnapLogic orchestrations and administers the lifecycle of data and process flows. Dashboards: Provides visibility into the health of integrations, including performance, reliability, and utilization. The Snaplex is a self-upgrading, elastic execution grid that streams data between applications, databases, files, social and big data sources. The Snaplex can run in the cloud, behind the firewall and on Hadoop. Snaps are modular collections of integration components built for a specific application or data source and are available for analytics and big data sources, identity management, social media, online storage,
https://en.wikipedia.org/wiki/British%20National%20Vegetation%20Classification
The British National Vegetation Classification or NVC is a system of classifying natural habitat types in Great Britain according to the vegetation they contain. A large scientific meeting of ecologists, botanists, and other related professionals in the United Kingdom resulted in the publication of a compendium of five books: British Plant Communities, edited by John S. Rodwell, which detail the incidence of plant species in twelve major habitat types in the British natural environment. They are the first systematic and comprehensive account of the vegetation types of the country. They cover all natural, semi-natural and major artificial habitats in Great Britain (not Northern Ireland) and represent fifteen years of research by leading plant ecologists. From the data collated from the books, commercial software products have been developed to help to classify vegetation identified into one of the many habitat types found in Great Britain – these include MATCH, TABLEFIT and MAVIS. Terminology used in connection with the British National Vegetation Classification The following is a list of terms used in connection with the British National Vegetation Classification, together with their meanings: Communities, subcommunities and variants A community is the fundamental unit of categorisation for vegetation. A subcommunity is a distinct recognisable subdivision of a community. A variant is a further subdivision of a subcommunity. Constant species A constant species in a community is a species that is always present in any given stand of vegetation belonging to that community. For a list of the constant species, and the NVC communities in which they are present, see List of constant species in the British National Vegetation Classification. Rare species A rare species is a species which is associated with a particular community and is rare nationally. The sources used by the authors of British Plant Communities for assessing rarity were as follows. a) fo
https://en.wikipedia.org/wiki/The%20Stone%20%28video%20game%29
The Stone is an online game developed by web company Abject Modernity Internet Creations Inc. in 1995. The mystery game was created in 1996 but launched as a consumer product in 1997. People had to buy a physical stone containing the login credentials to the website, which was unheard of at the time. In 1999, The Stone was profiled by Forbes magazine. "Stoners", a film about The Stone, was released by Rod Bruinooge and Scott Jaworski in September 2004. It covered the activities of the internet/online gaming community that emerged around The Stone. Pink Floyd provided the soundtrack to the film, with all music taken from The Division Bell Album. Gameplay Puzzles of The Stone are located in a place called The Immediate. There are a total of 216 Stone puzzles, grouped into 6 categories, each category having 6 different levels of difficulty. Once all the puzzles are solved, The Stone is said to unlock an ancient mystery called the Enigma. The secret of The Stone is kept by the Stonekeepers. A player of The Stone is often referred to as a stoner. When trying to solve a certain Stone puzzle, a stoner may go to a place called The Commons and ask for a nudge (i.e., a hint) from other stoners who have already solved that particular puzzle. Once a stoner has solved all available Stone puzzles, he or she is allowed into the Sisyphus Lodge. Championship tournament The Stone Championship Tournament, also known as the Final Six Tournament, began September 30, 2005. The first Stone player to solve all of the six final puzzles would be crowned Champion of The Stone and would immediately be granted the status of Stonekeeper. The tournament was won August 11, 2007 by the Stone players "Gary_" and "cinnabar." The solution to the Enigma was discovered August 22, 2007 by Stone player "grissy". Subsequent to the completion of the Final Six Tournament, the tournament puzzles were opened to all the remaining Stone players as the Final Six Redux, featuring slightly different answers
https://en.wikipedia.org/wiki/Tree%20moss
Tree moss is a common name for several organisms and may refer to: Climacium, a genus of mosses which resemble miniature trees Climacium dendroides, a common species of Climacium Evernia, a genus of lichens which grow on trees Usnea, a genus of lichens which grow on trees
https://en.wikipedia.org/wiki/Wall%20stress%20relaxation
The plant cell wall is made up of hydrated polymetric material, allowing it to have viscoelastic properties. The primary cell wall of a plant consists of cellulose fibers, hemicellulose, and xyloglucans. This load bearing network is also surrounded by pectins and glycoproteins. Wall stress relaxation is an important factor in cell wall expansion. Wall stress (measured in force per unit area) is created in response to the plant cell's turgor pressure. Turgor pressure creates tension in the cell walls of plants, fungi, and bacteria, as it opposes the pressure of the cell's primary cell wall; this also allows for stretching of the cell wall. The stretching of the cell wall, or the reduction of stress, occurs as a result of cell expansion and rearrangement. Cell expansion is crucial for the reshaping and rearranging of plant cells. Expansion is the result of "creep", or selective wall loosening, which is driven by turgor pressure. During this "creep", cellulose microfibers move relative to each other creating an irreversible extension Cell expansion Cell expansion begins with the selective loosening of the cell wall, reducing the plant cell's turgor pressure and water potential. This allows for the influx of water, leading to cell enlargement. This enlargement is made possible by the sliding of polymers, increasing the cell wall's surface area. In most plants, cell expansion is anisotropic. Previous experiments have confirmed that the cellulose microfibril orientation in the primary cell wall is the key for determining the direction of anisotropic growth and expansion. Cells tend to grow transversely to the cellulose microfibril orientation. It has been found that cell walls expand faster under acidic conditions, this is called acid growth. Treating living cells with acid induces acidification of the cell wall by activating an ATPase in the cell wall's plasma membrane. In onion epidermal cells, which are used as models to study anisotropy in extension, extension is
https://en.wikipedia.org/wiki/Dennis%20the%20Menace%20and%20Gnasher
Dennis the Menace and Gnasher (originally titled Dennis the Menace and currently titled Dennis and Gnasher) is a long-running comic strip in the British children's comic The Beano, published by DC Thomson, of Dundee, Scotland. The comic stars a boy named Dennis the Menace and his dog, an "Abyssinian wire-haired tripe hound" named Gnasher. The strip first appeared in issue 452, dated 17 March 1951, and on sale from 12 March 1951. It is the longest-running strip in the comic. The idea and name of the character emerged when the comic's editor heard a British music hall song with the chorus "I'm Dennis the Menace from Venice". The creation of Dennis in the 1950s had sales of The Beano soar. From issue 1678 onward (dated 14 September 1974), Dennis the Menace replaced Biffo the Bear on the front cover, and has been there ever since. Coincidentally, on 12 March 1951, another comic strip named Dennis the Menace debuted in the US. As a result of this, the US series has initially been retitled Dennis for UK audiences, while the British character's appearances are often titled Dennis and Gnasher outside the UK. In 2021, Dennis (and Gnasher) featured on a series of UK postage stamps issued by the Royal Mail to mark 70 years of the character. Dennis is the archetypal badly behaved schoolboy. The main recurring storyline throughout the years features his campaign of terror against a gang of "softies" (effeminate, well-behaved boys), particularly Walter the Softy. Walter finds himself in unfavourable circumstances on many occasions, although he sometimes gets the last laugh. Author Michael Rosen states, "In most children's books, a bad child gets made good – but the great thing about Dennis is he never gets better". Dennis the Menace and Gnasher was first drawn by David Law (1951–1970), then David Sutherland (1970–1998). David Parkins took over from 1998 until 2006, although Nigel Parkinson drew most of the strips after 2002, and Jimmy Hansen alternated with him in 2005–06. Tom
https://en.wikipedia.org/wiki/Atomic%20whirl
The atomic whirl is a symbol of science and has come to be used as a symbol of atheism in general. The Atomic Whirl is based on the historical Rutherford model of the atom, which erroneously showed the orbital paths of electrons around the central nucleus, and not on the atomic orbitals. It resembles the authoritative logos and symbols of the United States Atomic Energy Commission and the International Atomic Energy Agency who also based their designs on the Rutherford Model. The symbol is used by the American Atheists organization to symbolize that "only through the use of scientific analysis and free, open inquiry can humankind reach out for a better life". The lower part of the central loop is left open or "broken" to represent the fact that atheists accept that while they rely on the scientific method, they are searching for the answers, and in some cases, further questions. This central loop forms an "A" which represents the word "atheism". American Atheists has a copyright on the symbol. Approved emblem of belief (US) This symbol is one of the permitted "Emblems of Belief" that the United States Department of Veterans Affairs allows on government-furnished headstones and markers. See also Rutherford model Atheism
https://en.wikipedia.org/wiki/Alain%20Lascoux
Alain Lascoux (17 October 1944 – 20 October 2013) was a French mathematician at Université de Paris VII, University of Marne la Vallée and Nankai University. His research was primarily in algebraic combinatorics, particularly Hecke algebras and Young tableaux. Lascoux earned his doctorate in 1977 from the University of Paris. He worked for twenty years with Marcel-Paul Schützenberger on properties of the symmetric group. They wrote many articles together and had a major impact on the development of algebraic combinatorics. They succeeded in giving a combinatorial understanding of various algebraic and geometric questions in representation theory. Thus they introduced many new objects related to both fields like Schubert polynomials and Grothendieck polynomials, as well as novel terminology like the plactic monoid and vexillary permutations. They were also the first to define the crystal graph structure on Young tableaux (though not under this name). Lascoux was an invited speaker at the 1998 International Congress of Mathematicians in Berlin, Germany. See also LLT polynomial
https://en.wikipedia.org/wiki/Natural%20density
In number theory, natural density, also referred to as asymptotic density or arithmetic density, is one method to measure how "large" a subset of the set of natural numbers is. It relies chiefly on the probability of encountering members of the desired subset when combing through the interval as grows large. Intuitively, it is thought that there are more positive integers than perfect squares, since every perfect square is already positive, and many other positive integers exist besides. However, the set of positive integers is not in fact larger than the set of perfect squares: both sets are infinite and countable and can therefore be put in one-to-one correspondence. Nevertheless if one goes through the natural numbers, the squares become increasingly scarce. The notion of natural density makes this intuition precise for many, but not all, subsets of the naturals (see Schnirelmann density, which is similar to natural density but defined for all subsets of ). If an integer is randomly selected from the interval , then the probability that it belongs to is the ratio of the number of elements of in to the total number of elements in . If this probability tends to some limit as tends to infinity, then this limit is referred to as the asymptotic density of . This notion can be understood as a kind of probability of choosing a number from the set . Indeed, the asymptotic density (as well as some other types of densities) is studied in probabilistic number theory. Definition A subset of positive integers has natural density if the proportion of elements of among all natural numbers from 1 to converges to as tends to infinity. More explicitly, if one defines for any natural number the counting function as the number of elements of less than or equal to , then the natural density of being exactly means that It follows from the definition that if a set has natural density then . Upper and lower asymptotic density Let be a subset of the set of nat
https://en.wikipedia.org/wiki/Eisosome
Eisosomes ('eis' meaning into or portal and 'soma', meaning body) are large, heterodimeric, immobile protein complexes at the plasma membrane which mark the site of endocytosis in some eukaryotes, and were discovered in the yeast Saccharomyces cerevisiae in 2006. Currently, seven genes: Pil1, Lsp1 and Sur7, Eis1, Seg1 and Ygr130C, Seg2, are annotated to the formation of the proteins identified in eisosomes. These organelle-like structures have put to rest the idea that sites of endocytosis in cells are chosen at random. Eisosomes have a profound role in regulating plasma membrane architecture and organization in yeast. Microscopic and genetic analyses link these stable, ultrastructural assemblies to the endocytosis of both lipid and protein cargoes in cells. There are approximately 50–100 eisosomes in each mature yeast cell distributed uniformly across the cell surface periphery in a characteristic dotted pattern with each eisosome containing approximately 2000–5000 copies of Pil1 and Lsp1 proteins, as well as, integral membrane protein Sur7. Only a few of the eisosomes present in a cell are active at any one time, suggesting that eisosomes function by using reversible phosphorylation and are regulated portals that govern both location and magnitude of membrane traffic into the cell. Endocytosis in yeast The yeast plasma membrane consists of three compartments: Membrane compartment containing Can1 (MCC) Membrane compartment containing Pma1 (MCP) Membrane compartment containing TORC2 (MCT) The MCC, a furrow in the plasma membrane, is generated by eisosomes, it disappears in a cell lacking Pil1 which is one of the main eisosome components. Structural classification These are large protein complexes composed primarily of subunits of two Bin-Amphiphysin-RVS (BAR) domain containing proteins Pil1 and Lsp1. These two paralogue proteins self-assemble in higher order structure helices and bind preferentially to phosphoinositide-containing membrane. It is also found
https://en.wikipedia.org/wiki/S%20phase%20index
S-phase index (SPI), is a measure of cell growth and viability, especially the capacity of tumor cells to proliferate. It is defined as the number of BrdU-incorporating cells relative to the volume of DNA staining determined from whole mount confocal analyses. Only cells in the S phase will incorporate BrdU into their DNA structure, which assists in determining length of the cell cycle.
https://en.wikipedia.org/wiki/PicoScope%20%28software%29
PicoScope is computer software for real-time signal acquisition of Pico Technology oscilloscopes. PicoScope is supported on Microsoft Windows, Mac OS X, Debian and Ubuntu platforms. PicoScope is primarily used to view and analyze real-time signals from PicoScope oscilloscopes and data loggers. PicoScope software enables analysis using FFT, a spectrum analyser, voltage-based triggers, and the ability to save/load waveforms to disk. PicoScope is compatible with parallel port oscilloscopes and the newer USB oscilloscopes. The software has been described as "very good for laptops" and can be used with desktop or laptop PCs. The Linux version has been described as "lightyears ahead [of] Qpicoscope and other attempts at Linux scope software" and "well capable of replacing a professional benchtop scope". Beta versions of the software also work on the ARM-based BeagleBone Black and Raspberry Pi development hardware. PicoScope software requires a USB or LPT oscilloscope from the PicoScope range developed by Pico Technology. Such oscilloscopes are available with bandwidths up to 1 GHz, up to four input channels, hardware vertical resolutions up to 16 bits, sampling rates up to 5 GS/s, buffer sizes up to 2 GS, and built-in signal generators. Other features available on some models include flexible hardware resolution, switchable bandwidth limiters, switchable high-impedance and 50 ohm inputs, and differential inputs. PicoScope for Linux won the EDN Hot 100 Products of 2014 award, under the Test & Measurement category, for "converting a Linux PC into an oscilloscope, FFT spectrum analyser and measuring device". Features Windows PicoScope for Microsoft Windows is the full-featured oscilloscope application, and was first released in 1992 by Pico Technology. PicoScope software enables real-time scope display with zooming and panning, and buffers captured waveforms on the PC to enable engineers to view previous measurements. PicoScope uses configurable triggers, which are avai
https://en.wikipedia.org/wiki/Power%20bandwidth
The power bandwidth of an amplifier is sometimes taken as the frequency range (or, rarely, the upper frequency limit) for which the rated power output of an amplifier can be maintained (without excessive distortion) to at least half of the full rated power. (Some specifications may mandate 100% of the rated power; sometimes referring to the full-power bandwidth.) It should not be confused with "half-power" bandwidth, only used in conjunction with filter frequency response curves, where it refers to -3dB points in the frequency response of a band-pass filter. Data sheets for operational amplifiers often use the term (full-)power bandwidth to indicate the highest frequency at which the achievable peak-to-peak output voltage swing is still equal to the DC output voltage range. This is also sometimes described as the slew-rate-limited bandwidth. The full-power bandwidth is then related to the slew rate in volts per microsecond and the peak-to-peak voltage swing by where is expressed in hertz. In data sheets for commonly available operational amplifiers, slew rate is usually given in volts per microsecond. Specifying power bandwidth Power bandwidth may be specified as a frequency limit or shown in the form of a graph (e.g. as maximum available voltage swing versus frequency, often with power supply voltage as a parameter).
https://en.wikipedia.org/wiki/Blotto%20%28biology%29
In biology, BLOTTO is a blocking reagent made from nonfat dry milk, phosphate buffered saline, and sodium azide. Its name is an almost-acronym of bovine lacto transfer technique optimizer. It constitutes an inexpensive source of nonspecific protein (milk casein) which blocks protein binding sites in a variety of experimental paradigms, notably Southern blots, Western blots, and ELISA. Its use was first reported in 1984 by Johnson and Elder's lab at Scripps. Prior to 1984, partially purified proteins such as bovine serum albumin, ovalbumin, or gelatin from various species had been used as blocking reagents but had the disadvantage of being expensive.
https://en.wikipedia.org/wiki/AFGROW
AFGROW (Air Force Grow) is a Damage Tolerance Analysis (DTA) computer program that calculates crack initiation, fatigue crack growth, and fracture to predict the life of metallic structures. Originally developed by the Air Force Research Laboratory, AFGROW is mainly used for aerospace applications, but can be applied to any type of metallic structure that experiences fatigue cracking. History AFGROW's history traces back to a crack growth life prediction program (ASDGRO) which was written in BASIC for IBM-PCs by E. Davidson at ASD/ENSF in the early-mid-1980s. In 1985, ASDGRO was used as the basis for crack growth analysis for the Sikorsky H-53 helicopter under contract to Warner-Robins ALC. The program was modified to utilize very large load spectra, approximate stress intensity solutions for cracks in arbitrary stress fields, and use a tabular crack growth rate relationship based on the Walker equation on a point-by-point basis (Harter T-Method). The point loaded crack solution from the Tada, Paris, and Irwin Stress Intensity Factor Handbook was originally used to determine K (for arbitrary stress fields) by integration over the crack length using the unflawed stress distribution independently for each crack dimension. A new method was developed by F. Grimsley (AFWAL/FIBEC) to determine stress intensity, which used a 2-D Gaussian integration scheme with Richardson Extrapolation which was optimized by G. Sendeckyj (AFWAL/FIBEC). The resulting program was named MODGRO since it was a modified version of ASDGRO. Many modifications were made during the late 1980s and early 1990s. The primary modification was changing the coding language from BASIC to Turbo Pascal and C. Numerous small changes/repairs were made based on errors that were discovered. During this time period, NASA/Dryden implemented MODGRO in the analysis for the flight test program for the X-29. In 1993, the Navy was interested in using MODGRO to assist in a program to assess the effect of certain
https://en.wikipedia.org/wiki/Apache%20ORC
Apache ORC (Optimized Row Columnar) is a free and open-source column-oriented data storage format. It is similar to the other columnar-storage file formats available in the Hadoop ecosystem such as RCFile and Parquet. It is used by most of the data processing frameworks Apache Spark, Apache Hive, Apache Flink and Apache Hadoop. In February 2013, the Optimized Row Columnar (ORC) file format was announced by Hortonworks in collaboration with Facebook. A month later, the Apache Parquet format was announced, developed by Cloudera and Twitter. History See also Apache Spark Apache Arrow Apache Hive Apache NiFi Pig (programming tool) Trino (SQL query engine) Presto (SQL query engine)
https://en.wikipedia.org/wiki/Vehicle-to-everything
Vehicle-to-everything (V2X) is communication between a vehicle and any entity that may affect, or may be affected by, the vehicle. It is a vehicular communication system that incorporates other more specific types of communication as V2I (vehicle-to-infrastructure), V2N (vehicle-to-network), V2V (vehicle-to-vehicle), V2P (vehicle-to-pedestrian), V2D (vehicle-to-device). The main motivations for V2X are road safety, traffic efficiency, energy savings, and mass surveillance. The U.S. NHTSA estimates a minimum of 13% reduction in traffic accidents if a V2V system were implemented, resulting in 439,000 fewer crashes per year. There are two types of V2X communication technology depending on the underlying technology being used: (1) WLAN-based, and (2) cellular-based. The V2X contains the following sub categories: Vehicle-to-Everything (V2X) - "communication between a vehicle and any entity that may affect, or may be affected by, the vehicle." Vehicle-to-Device (V2D) - Bluetooth / WiFi-Direct, e.g. Apple’s CarPlay and Google’s Android Auto. Vehicle-to-Grid (V2G) - information exchange with the smart grid to balance loads more efficiently. Vehicle-to-Building (V2B), also known as Vehicle-to-Home (V2H) Vehicle-to-Load (V2L) Vehicle-to-Network (V2N) - communication based on Cellular (3GPP) / 802.11p. Vehicle-to-Cloud (V2C) - e.g. OTA updates, remote vehicle diagnostics (DoIP). Vehicle-to-Infrastructure (V2I) - e.g. traffic lights, lane markers and parking meters. Vehicle-to-Pedestrian (V2P) - e.g. wheelchairs and bicycles. Vehicle-to-Vehicle (V2V) - real-time data exchange with nearby vehicles. History The history of working on vehicle-to-vehicle communication projects to increase safety, reduce accidents and driver assistance can be traced back to the 1970s with projects such as the US Electronic Road Guidance System (ERGS) and Japan's CACS. Most milestones in the history of vehicle networks originate from the United States, Europe, and Japan. Standardizat
https://en.wikipedia.org/wiki/Almeida%E2%80%93Pineda%20recurrent%20backpropagation
Almeida–Pineda recurrent backpropagation is an extension to the backpropagation algorithm that is applicable to recurrent neural networks. It is a type of supervised learning. It was described somewhat cryptically in Richard Feynman's senior thesis, and rediscovered independently in the context of artificial neural networks by both Fernando Pineda and Luis B. Almeida. A recurrent neural network for this algorithm consists of some input units, some output units and eventually some hidden units. For a given set of (input, target) states, the network is trained to settle into a stable activation state with the output units in the target state, based on a given input state clamped on the input units.
https://en.wikipedia.org/wiki/Common%20external%20power%20supply
The common external power supply (common EPS) was a specification for a universal charger, made by the European Commission for smartphones sold within the European Union. Although compliance was voluntary, a majority of the world's largest mobile phone manufacturers agreed to make their applicable mobile phones compatible with the specification. The common EPS initiative specified a universal charger (external power supply) that must effectively conform to the USB Battery Charging V1.2 specification and include a cable with a USB 2.0 micro-B connector. The specification EN 62684:2010 was active from 2009 to 2014, and has since been withdrawn as obsolete. The European Union later passed Radio Equipment Directive 2021/0291, which requires new smartphones to use USB-C by the end of 2024, and laptops by 2026. Purpose According to the European Commission, a common external power supply / "charger" standard is desirable because, History In June 2009, many of the world's largest mobile phone manufacturers signed an EC-sponsored memorandum of understanding (MoU), agreeing to make most new data-enabled mobile phones marketed in the European Union compatible with a to-be-specified common EPS. All signatories agreed to develop a common specification for the EPS "to allow for full compatibility and safety of chargers and mobile phones." 14 mobile phone manufacturers and technology providers signed the MoU – the original 10 signatories, Apple, LG, Motorola, NEC, Nokia, Qualcomm, RIM, Samsung, Sony Ericsson, and Texas Instruments as well as Atmel, Emblaze Mobile, Huawei Technologies and TCT Mobile (Alcatel). To develop and formalize the needed technical standards, the Commission issued a standardisation mandate to CEN, CENELEC and ETSI on a common "charging capability for mobile telephones." In response, CENELEC created a task force to develop the interoperability specifications of a common external power supply. In line with the Dresden agreement signed in 1996 by both CENE
https://en.wikipedia.org/wiki/Urban%20Ecosystems
Urban Ecosystems is an peer-reviewed bimonthly transformative international scientific journal published by Springer. The journal is interdisciplinary, with its articles covering relationships "between socioeconomic and ecological structures and processes in urban environments." Associated with the Society for Urban Ecology, the journal was established in 1997. Additionally, the journal undergoes a hybrid publishing method. See also Landscape and Urban Planning
https://en.wikipedia.org/wiki/Uzbekistan%20cough%20syrup%20scandal
The Uzbekistan cough syrup scandal was a series of poisonings that resulted in the deaths of 18 children in Samarkand and two more children elsewhere in Uzbekistan in December 2022 and January 2023. It was caused by the toxic levels of diethylene glycol and ethylene glycol in cold medicines produced by the Indian company Marion Biotech, such as the Dok-1 Max brand. Subsequently, the Indian government investigated Marion Biotech's manufacturing processes, while Uzbek authorities opened a criminal case against members of the health system that had contributed to the children's deaths, such as regulatory officials and pharmacy administrators. On December 22, the sale and distribution of Dok-1 Max was temporarily suspended in Uzbekistan, followed by the complete ban of Marion-produced cough syrups one week later. Uzbek customs officials prevented the distribution of 60,000 boxes of Dok-1 Max in response to the new regulations. In late December, the Indian government also ordered the company's manufacturing plant in Noida, Uttar Pradesh to shutdown production and on January 10, 2023, it suspended Marion's license to conduct business. Shavkat Mirziyoyev, the current President of Uzbekistan, removed Sardor Kariyev from directing the Ministry of Health's Agency for the Development of the Pharmaceutical Network for failed regulatory oversight. The World Health Organization is continuing to investigate the poisonings. Background Dok-1 Max is a combination drug produced in syrup and tablet forms, used to resolve symptoms of acute respiratory diseases. It is mainly used in cases of whooping cough, runny nose, wet cough, sore throat, angina, headache, body aches, and fever. Despite the packaging indicating that the syrup can be safely used for children aged 2 and over, other companies only recommend use among those over 12 years of age. Each 10 ml dose of the drug contains paracetamol (500 mg), guaifenesin (200 mg), phenylephrine hydrochloride (10 mg), and minor amounts of so
https://en.wikipedia.org/wiki/Sono-Seq
Sono-Seq (Sonication of Cross-linked Chromatin Sequencing) is a method in molecular biology used for determining the sequences of those DNA regions in the genome near regions of open chromatin of expressed genes. It is also known as "Input" in the Chip-Seq protocol, since it follows the same steps except it doesn't require immunoprecipitation.
https://en.wikipedia.org/wiki/Kerr%E2%80%93Newman%20metric
The Kerr–Newman metric is the most general asymptotically flat, stationary solution of the Einstein–Maxwell equations in general relativity that describes the spacetime geometry in the region surrounding an electrically charged, rotating mass. It generalizes the Kerr metric by taking into account the field energy of an electromagnetic field, in addition to describing rotation. It is one of a large number of various different electrovacuum solutions, that is, of solutions to the Einstein–Maxwell equations which account for the field energy of an electromagnetic field. Such solutions do not include any electric charges other than that associated with the gravitational field, and are thus termed vacuum solutions. This solution has not been especially useful for describing astrophysical phenomena, because observed astronomical objects do not possess an appreciable net electric charge, and the magnetic fields of stars arise through other processes. As a model of realistic black holes, it omits any description of infalling baryonic matter, light (null dusts) or dark matter, and thus provides at best an incomplete description of stellar mass black holes and active galactic nuclei. The solution is of theoretical and mathematical interest as it does provide a fairly simple cornerstone for further exploration. The Kerr–Newman solution is a special case of more general exact solutions of the Einstein–Maxwell equations with non-zero cosmological constant. History In Dec 1963 Kerr and Schild found the Kerr–Schild metrics that gave all Einstein spaces that are exact linear perturbations of Minkowski space. In early 1964 Roy Kerr looked for all Einstein–Maxwell spaces with this same property. By Feb 1964 the special case where the Kerr–Schild spaces were charged (this includes the Kerr–Newman solution) was known but the general case where the special directions were not geodesics of the underlying Minkowski space proved very difficult. The problem was given to George Debney to
https://en.wikipedia.org/wiki/International%20Society%20of%20Biometeorology
The International Society of Biometeorology (ISB) is a professional society for scientists interested in biometeorology, specifically environmental and ecological aspects of the interaction of the atmosphere and biosphere. The organization's stated purpose is: "to provide one international organization for the promotion of interdisciplinary collaboration of meteorologists, physicians, physicists, biologists, climatologists, ecologists and other scientists and to promote the development of Biometeorology". The International Society of Biometeorology was founded in 1956 at UNESCO headquarters in Paris, France, by S. W. Tromp, a Dutch geologist, H. Ungeheuer, a German meteorologist, and several human physiologists of which F. Sargent II of the United States became the first President of the society. ISB affiliated organizations include: the International Association for Urban Climate, the International Society for Agricultural Meteorology, the International Union of Biological Sciences, the World Health Organization, and the World Meteorological Organization. ISB affiliate members include: the American Meteorological Society, the Centre for Renewable Energy Sources, the German Meteorological Society, the Society for the Promotion of Medicine-Meteorological Research e.V., International Society of Medical Hydrology and Climatology, and the UK Met Office. Publications ISB publishes the following journals: Bulletin of the Society of Biometeorology International Journal of Biometeorology
https://en.wikipedia.org/wiki/Vitamin%20K-dependent%20protein
A Vitamin K-dependent protein (VKDP) is a protein that can bind calcium ions but only after being carboxylated at a certain glutamic residue. This carboxylation, said to activate the protein, is facilitated by some form of vitamin K1 or vitamin K2. The relevant part of a vitamin K-dependent protein is a Gla domain, and such a protein is informally called a Gla protein. Some Gla proteins have "Gla" in their name, for example Matrix Gla protein, but many don't, such as osteocalcin. Proteins Vitamin K
https://en.wikipedia.org/wiki/Networking%20hardware
Networking hardware, also known as network equipment or computer networking devices, are electronic devices that are required for communication and interaction between devices on a computer network. Specifically, they mediate data transmission in a computer network. Units which are the last receiver or generate data are called hosts, end systems or data terminal equipment. Range Networking devices includes a broad range of equipment which can be classified as core network components which interconnect other network components, hybrid components which can be found in the core or border of a network and hardware or software components which typically sit on the connection point of different networks. The most common kind of networking hardware today is a copper-based Ethernet adapter which is a standard inclusion on most modern computer systems. Wireless networking has become increasingly popular, especially for portable and handheld devices. Other networking hardware used in computers includes data center equipment (such as file servers, database servers and storage areas), network services (such as DNS, DHCP, email, etc.) as well as devices which assure content delivery. Taking a wider view, mobile phones, tablet computers and devices associated with the internet of things may also be considered networking hardware. As technology advances and IP-based networks are integrated into building infrastructure and household utilities, network hardware will become an ambiguous term owing to the vastly increasing number of network-capable endpoints. Specific devices Network hardware can be classified by its location and role in the network. Core Core network components interconnect other network components. Gateway: an interface providing a compatibility between networks by converting transmission speeds, protocols, codes, or security measures. Router: a networking device that forwards data packets between computer networks. Routers perform the "traffic directing" fun
https://en.wikipedia.org/wiki/Mixed%20acid%20fermentation
In biochemistry, mixed acid fermentation is the metabolic process by which a six-carbon sugar (e.g. glucose, ) is converted into a complex and variable mixture of acids. It is an anaerobic (non-oxygen-requiring) fermentation reaction that is common in bacteria. It is characteristic for members of the Enterobacteriaceae, a large family of Gram-negative bacteria that includes E. coli. The mixture of end products produced by mixed acid fermentation includes lactate, acetate, succinate, formate, ethanol and the gases and . The formation of these end products depends on the presence of certain key enzymes in the bacterium. The proportion in which they are formed varies between different bacterial species. The mixed acid fermentation pathway differs from other fermentation pathways, which produce fewer end products in fixed amounts. The end products of mixed acid fermentation can have many useful applications in biotechnology and industry. For instance, ethanol is widely used as a biofuel. Therefore, multiple bacterial strains have been metabolically engineered in the laboratory to increase the individual yields of certain end products. This research has been carried out primarily in E. coli and is ongoing. Variations of mixed acid fermentation occur in a number of bacterial species, including bacterial pathogens such as Haemophilus influenzae where mostly acetate and succinate are produced and lactate can serve as a growth substrate. Mixed acid fermentation in E. coli E. coli use fermentation pathways as a final option for energy metabolism, as they produce very little energy in comparison to respiration. Mixed acid fermentation in E. coli occurs in two stages. These stages are outlined by the biological database for E. coli, EcoCyc. The first of these two stages is a glycolysis reaction. Under anaerobic conditions, a glycolysis reaction takes place where glucose is converted into pyruvate:       glucose → 2 pyruvate There is a net production of 2 ATP and 2 NADH m
https://en.wikipedia.org/wiki/Equisingularity
In algebraic geometry, an equisingularity is, roughly, a family of singularities that are not non-equivalent and is an important notion in singularity theory. There is no universal definition of equisingularity but Zariki's equisingularity is the most famous one. Zariski's equisingualrity, introduced in 1971 under the name " algebro-geometric equisingularity", gives a stratification that is different from the usual Whitney stratification on a real or complex algebraic variety. See also stratified space
https://en.wikipedia.org/wiki/Wide%20and%20narrow%20data
Wide and narrow (sometimes un-stacked and stacked, or wide and tall) are terms used to describe two different presentations for tabular data. Wide Wide, or unstacked data is presented with each different data variable in a separate column. Narrow Narrow, stacked, or long data is presented with one column containing all the values and another column listing the context of the value This is often easier to implement; addition of a new field does not require any changes to the structure of the table, however it can be harder for people to understand. Implementations Many statistical and data processing systems have functions to convert between these two presentations, for instance the R programming language has several packages such as the tidyr package. The pandas package in Python implements this operation as "melt" function which converts a wide table to a narrow one. The process of converting a narrow table to wide table is generally referred to as "pivoting" in the context of data transformations. The "pandas" python package provides a "pivot" method which provides for a narrow to wide transformation. See also Abstract data type Pivot table Table (information) Information graphics Row (database) Table (database) Table (HTML)
https://en.wikipedia.org/wiki/Natrialba
In taxonomy, Natrialba is a genus of the Natrialbaceae. The genus consists of many diverse species that can survive extreme environmental niches, especially they are capable to live in the waters saturated or nearly saturated with salt (halophiles). They have certain adaptations to live within their salty environments. For example, their cellular machinery is adapted to high salt concentrations by having charged amino acids on their surfaces, allowing the cell to keep its water molecules around these components. The osmotic pressure and these amino acids help to control the amount of salt within the cell. For instance, N. magadii is an aerobic chemoorganotrophic, dual extremophile requiring alkaline conditions and hypersalinity for optimal growth. Those harsh conditions resulted in changed composition of charged amino acids in the proteins (average isoelectric point is only 4.64, whereas other organisms average is 6.5) with almost all proteins being highly acidic. The genome of N. magadii consists of four replicons with a total sequence of 4,443,643 bp and encodes 4,212 putative proteins. The genome analysis identified multiple genes coding putative proteins involved in adaptation to hypersalinity, stress response, glycosylation, and polysaccharide biosynthesis. Additionally, proton-driven ATP synthase and a variety of putative cytochromes and other proteins required for aerobic respiration and electron transfer had been found. The genome encodes a number of putative proteases/peptidases. Their resistance to salt allows for the use of some members of the genus in biotechnological processes.
https://en.wikipedia.org/wiki/Osteopathy
Osteopathy () is a pseudoscientific system of alternative medicine that emphasizes physical manipulation of the body's muscle tissue and bones. In most countries, practitioners of osteopathy are not medically trained and are referred to as osteopaths. Osteopathic manipulation is the core set of techniques in osteopathy. Parts of osteopathy, such as craniosacral therapy, have been described by Quackwatch as having no therapeutic value and have been labeled by them as pseudoscience and quackery. The techniques are based on an ideology created by Andrew Taylor Still (1828–1917) which posits the existence of a "myofascial continuity"—a tissue layer that "links every part of the body with every other part". Osteopaths attempt to diagnose and treat what was originally called "the osteopathic lesion", but which is now named "somatic dysfunction", by manipulating a person's bones and muscles. Osteopathic Manipulative Treatment (OMT) techniques are most commonly used to treat back pain and other musculoskeletal issues. Osteopathic manipulation is still included in the curricula of osteopathic physicians or Doctors of Osteopathic Medicine (DO) training in the US. The Doctor of Osteopathic Medicine degree, however, became a medical degree and is no longer a degree of non-medical osteopathy. History The practice of osteopathy began in the United States in 1874. Osteopathy was founded by Andrew Taylor Still, a 19th-century American physician (MD), Civil War surgeon, and Kansas state and territorial legislator. He lived near Baldwin City, Kansas, during the American Civil War and it was there that he founded the practice of osteopathy. Still claimed that human illness was rooted in problems with the musculoskeletal system, and that osteopathic manipulations could solve these problems by harnessing the body's own self-repairing potential. Still's patients were forbidden from treatment by conventional medicine, as well as from other practices such as drinking alcohol. These pra
https://en.wikipedia.org/wiki/Environmental%20epigenetics
Environmental epigenetics is a branch of epigenetics that studies the influence of external environmental factors on the gene expression of a developing embryo. The way that genes are expressed may be passed down from parent to offspring through epigenetic modifications, although environmental influences do not alter the genome itself. During embryonic development, epigenetic modifications determine which genes are expressed, which in turn determines the embryo's phenotype. When the offspring is still developing, genes can be turned on and off depending on exposure to certain environmental factors. While certain genes being turned on or off can increase the risk of developmental diseases or abnormal phenotypes, there is also the possibility that the phenotype will be non-functional. Environmental influence on epigenetics is highly variable, but certain environmental factors can greatly increase the risk of detrimental diseases being expressed at both early and adult life stages. Environmental triggers for epigenetic change The way that genes are expressed is influenced by the environment that the genome is in. These environmental influences are referred to as triggers and can involve anything that influences normal gene expression. How the genome is expressed depends on the environmental factors present during gestation. It is possible for the environmental effects of epigenetics to be deleterious or to be a natural part of the development pathway. When these environmental factors are detrimental it causes the deactivation of some DNA sequences, which can lead to atypical phenotypes. Some of the most common triggers include diet, temperature, exposure to harmful substances, and lifestyle. These triggers can cause low birth weight, neurological disorders, cancers, autoimmune diseases, and many other malformations. These epigenetic triggers can change the way that an organism develops and have lifelong effects. Epigenetic changes can be passed down through offspri
https://en.wikipedia.org/wiki/Cutting%20sequence
In digital geometry, a cutting sequence is a sequence of symbols whose elements correspond to the individual grid lines crossed ("cut") as a curve crosses a square grid. Sturmian words are a special case of cutting sequences where the curves are straight lines of irrational slope.
https://en.wikipedia.org/wiki/CISPE
CISPE (Cloud Infrastructure Services Providers in Europe) is a non-profit trade association for infrastructure as a service (IaaS) cloud providers in Europe. It was started to aid IaaS providers in explaining their business model to policymakers. Registered in early 2017, CISPE has been operating since 2015. The association aims to advocate for an EU-wide cloud-first public procurement policy and engage for a European Digital Single Market including the promotion of high-level security and data protection rules/standards as well as avoiding vendor lock-in. In June 2020, the association became one the 22 founding members of GAIA-X, announced by the German and French Ministers of Economic Affairs Peter Altmaier and Bruno Le Maire. CISPE joined forces with European cloud users and providers like BMW, EDF, Safran, Atos, Siemens, Bosch, OVHcloud, and Deutsche Telekom. The CISPE Data Protection Code of Conduct To help IaaS providers and their customers to comply with the EU General Data Protection Regulation (GDPR), which entered into force from 25 May 2018, CISPE released the CISPE Data Protection Code of Conduct. On top of the required compliance to meet with the GDPR, the code also ensures that IaaS customers can choose to have their data located and processed exclusively in Europe, and that the supplier will not re-use a customer's data. The compliance has to be declared by CISPs/IaaS providers service by service. The CISPE Code of Conduct was launched on 27 September 2016 at the European Parliament, and the first thirty services had been declared by the first CISPs/IaaS providers on 14 February 2017. Announcements received press coverage from Le Monde, InfoDSI, El Pais, La Repubblica, Silicon, Cloud Magazine, Computer Sweden, Tom's Hardware, L'informaticien, Global Security Mag, EU Observer, Politico, Computer Weekly, IAPP, Il corriere della Sicurezza, LeMagIT, Bloomberg Television, ITR Manager, Heise.de, COR.COM, ZDNet, ElEconomista.es, IT Channel, EuropaPre
https://en.wikipedia.org/wiki/Grave%20accent
The grave accent () ( or ) is a diacritical mark used to varying degrees in French, Dutch, Portuguese, Italian and many other western European languages, as well as for a few unusual uses in English. It is also used in other languages using the Latin alphabet, such as Mohawk and Yoruba, and with non-Latin writing systems such as the Greek and Cyrillic alphabets and the Bopomofo or Zhuyin Fuhao semi-syllabary. It has no single meaning, but can indicate pitch, stress, or other features. For the most commonly encountered uses of the accent in the Latin and Greek alphabets, precomposed characters are available. For less-used and compound diacritics, a combining character facility is available. A free-standing version of the symbol (), commonly called a backtick, also exists and has acquired other uses. Uses Pitch The grave accent first appeared in the polytonic orthography of Ancient Greek to mark a lower pitch than the high pitch of the acute accent. In modern practice, it replaces an acute accent in the last syllable of a word when that word is followed immediately by another word. The grave and circumflex have been replaced with an acute accent in the modern monotonic orthography. The accent mark was called , the feminine form of the adjective (), meaning "heavy" or "low in pitch." This was calqued (loan-translated) into Latin as which then became the English word grave. Stress The grave accent marks the stressed vowels of words in Maltese, Catalan, and Italian. A general rule in Italian is that words that end with stressed , , or must be marked with a grave accent. Words that end with stressed or may bear either an acute accent or a grave accent, depending on whether the final e or o sound is closed or open, respectively. Some examples of words with a final grave accent are ("city"), ("so/then/thus"), ("more"/"plus"), ("Moses"), and ("[he/she/it] brought/carried"). Typists who use a keyboard without accented characters and are unfamiliar with input
https://en.wikipedia.org/wiki/76%20%28number%29
76 (seventy-six) is the natural number following 75 and preceding 77. In mathematics 76 is: a composite number; a square-prime, of the form (p2, q) where q is a higher prime. It is the ninth of this general form and the seventh of the form (22.q). a Lucas number. a telephone or involution number, the number of different ways of connecting 6 points with pairwise connections. a nontotient. a 14-gonal number. a centered pentagonal number. an Erdős–Woods number since it is possible to find sequences of 76 consecutive integers such that each inner member shares a factor with either the first or the last member. with an aliquot sum of 64; within an aliquot sequence of two composite numbers (76,64,63,1,0) to the Prime in the 63-aliquot tree. an automorphic number in base 10. It is one of two 2-digit numbers whose square, 5,776, and higher powers, end in the same two digits. The other is . There are 76 unique compact uniform hyperbolic honeycombs in the third dimension that are generated from Wythoff constructions. In science The atomic number of osmium. The Little Dumbbell Nebula in the constellation Pegasus is designated as Messier object 76 (M76). In other fields Seventy-six is also: In colloquial American parlance, reference to 1776, the year of the signing of the United States Declaration of Independence. Seventy-Six, an 1823 novel by American writer John Neal. The Spirit of '76, patriotic painting by Archibald MacNeal Willard. A brand of ConocoPhillips gas stations, 76. The number of trombonists leading the parade in "Seventy-Six Trombones", from Meredith Willson's musical The Music Man. The 76ers, a professional basketball team based in Philadelphia. 76, the debut album of Dutch trance producer and DJ Armin van Buuren. Years like 1876 and 1976 See also List of highways numbered 76
https://en.wikipedia.org/wiki/KASW
KASW (channel 61) is a television station in Phoenix, Arizona, United States, affiliated with The CW. It is owned by the E. W. Scripps Company alongside ABC affiliate KNXV-TV (channel 15). Both stations share studios on North 44th Street on the city's east side, while KASW's primary transmitter is located on South Mountain. KASW went on the air in 1995 as the Phoenix affiliate of The WB. Its first owner contracted with KTVK (channel 3) for programming and support services, and KTVK bought the station in 1999. In addition to being an affiliate of The WB and later The CW, the station also broadcast several secondary local sports teams at various times. KASW was split from KTVK in 2014 as the result of KTVK's sale. Scripps acquired it in 2019 and has added local newscasts from KNXV. KASW is the high-power ATSC 3.0 (NextGen TV) station for the Phoenix area and provides the ATSC 3.0 broadcasts of six major Phoenix commercial stations. History Prior history of UHF channel 61 in Phoenix Prior to KASW's sign-on, the UHF channel 61 frequency in the Phoenix market was originally occupied by low-power station K61CA; that station carried a locally programmed music video format known as "Music Channel" and operated from March 15, 1983, until November 12, 1984, closing due to mounting debts and lack of cash to continue operating. The construction permit for K61CA remained active for several more years; by 1988, it was owned by Channel 61 Development Corporation and was planned as a satellite-fed relay of KSTS, a Telemundo affiliate in San Jose, California. In November 1987, the FCC allocated channel 61 for full-power use in Phoenix. KUSK-TV applied alongside four other groups; the field was narrowed to three, and Brooks Broadcasting, owned by Chandler farmer Gregory R. Brooks, was granted the permit in February 1991 by the FCC review board. WB affiliation Little activity occurred on the permit, with the call sign KAIK; Brooks considered running home shopping on the station
https://en.wikipedia.org/wiki/Torre%20Entel
Torre Entel (Entel Tower) is the name of a high TV and telecommunications tower in Santiago, Chile. Torre Entel has an observation deck open for visitors. Construction began in 1970 during Eduardo Frei Montalva term as president and it was inaugurated in 1974. In 1976 it carried its first television transmissions. For many years it was the tallest building in Chile and today remains a symbol of Santiago. The tower is constructed of concrete, steel, and aluminum. With 128 m high and 18 floors, it was after the end of its construction in 1974, the highest architectural structure in the country, a title it kept until the inauguration of the Telefonica Tower in 1996 with 143 m. Already surpassed in height by other buildings, it continues being the structure of greater prominence in the commune of Santiago, being located next to the Avenida Libertador General Bernardo O'Higgins and to a block of the La Moneda Palace, reason why it has stayed like an icon of the city. Its design represents a torch, an ancient form of telecommunication. History Its construction began during the government of Eduardo Frei Montalva, on July 1, 1970, as part of the National Telecommunications Center. After four years of construction, the Tower reached its current height on August 30, 1974. Its structure was influenced by the Post Office Tower in London, which had been built a few years earlier. Later, on September 8, 1975, two satellite dishes were installed, which were the first telecommunications elements visible from the outside, and finally, on April 12, 1976, the telephone channels came into service. From that moment on, the Entel Tower became the vital nucleus of the country's communications system by allowing the interconnection of Entel's telephone, television, radio and microwave network services with those of the Chilean Telephone Company (currently Movistar) and with the north, center and south of the country and the province of Mendoza, Argentina. In addition, it is connecte
https://en.wikipedia.org/wiki/NFL%20on%20Nickelodeon
The NFL on Nickelodeon is the branding used for broadcasts of National Football League (NFL) games that are produced by CBS Sports, and broadcast on the American pay television channel Nickelodeon. In 2021, Nickelodeon hosted a one-time simulcast in coordination with CBS of the Chicago Bears–New Orleans Saints Wild Card game. This marked the first time that a major live sporting event would be broadcast on the channel. Following positive reception from media and fans, Nickelodeon announced that the simulcast would return for a Wild Card game during the 2021–22 NFL playoffs between the San Francisco 49ers and Dallas Cowboys. In May 2022, Nickelodeon announced that the simulcast would return for a Christmas game during the 2022 NFL season between the Denver Broncos and Los Angeles Rams. In May 2023, Nickelodeon announced that the simulcast would return for a second consecutive Christmas game during the 2023 NFL season between the Las Vegas Raiders and Kansas City Chiefs. In August 2023, CBS announced that it will carry Super Bowl LVIII on Nickelodeon in the same format, marking the first time that a Super Bowl game will feature an alternate broadcast. Background In December 2019, Viacom re-merged with CBS Corporation to form ViacomCBS. To capitalize on the re-merger, CBS announced plans to add content from Nickelodeon to its CBS All Access streaming service. Additionally, the National Football League announced that Nickelodeon would air a youth-specific broadcast of an early 2021 Wild Card playoff game that CBS Sports acquired the rights to, marking the first major live sporting event on the channel. The idea for Nickelodeon to televise an NFL game was that of CBS Sports' chairman Sean McManus. McManus brought the idea up to the league, which was looking to reach younger and diverse audiences. While the network has aired sports-related programming in the past (such as its athletics game show Nickelodeon Guts, and its Kids' Choice Sports awards), this marked the fi
https://en.wikipedia.org/wiki/Prentice%20Hall%20International%20Series%20in%20Computer%20Science
Prentice Hall International Series in Computer Science was a series of books on computer science published by Prentice Hall. The series' founding editor was Tony Hoare. Richard Bird subsequently took over editing the series. Many of the books in the series have been in the area of formal methods in particular. Selected books The following books were published in the series: R. S. Bird, Introduction to Functional Programming using Haskell, 2nd edition, 1998. . R. S. Bird and O. de Moor, Algebra of Programming, 1996. . (100th volume in the series.) O.-J. Dahl, Verifiable Programming, 1992. . D. M. Gabbay, Elementary Logics: A Procedural Perspective, 1998. . I. J. Hayes (ed.), Specification Cases Studies, 2nd edition, 1993. . M. G. Hinchey and J. P. Bowen (eds.), Applications of Formal Methods, 1996. . C. A. R. Hoare, Communicating Sequential Processes, 1985. hardback or paperback. C. A. R. Hoare and M. J. C. Gordon, Mechanized Reasoning and Hardware Design, 1998. . C. A. R. Hoare and He Jifeng, Unifying Theories of Programming, 1998. . INMOS Limited, Occam 2 Reference Manual, 1988. . Cliff Jones, Systematic Software Development Using VDM, 1986. hardback or paperback. M. Joseph (ed.), Real-Time Systems: Specification, Verification and Analysis, 1996. . Bertrand Meyer, Object-Oriented Software Construction (first edition only). Robin Milner, Communication and Concurrency, 1989. (for the paperback). C. C. Morgan, Programming from Specifications, 2nd edition, 1994. . P. N. Nissanke, Realtime Systems, 1997. . B. Potter, J. Sinclair and D. Till, An Introduction to Formal Specification and Z, 2nd edition, 1996. . A. W. Roscoe (ed.), A Classical Mind: Essays in Honour of C. A. R. Hoare, 1994. . A. W. Roscoe, The Theory and Practice of Concurrency, 1997. . J. M. Spivey, The Z Notation: A Reference Manual, 2nd edition, 1992. . J. C. P. Woodcock and J. W. Davies, Using Z: Specification, Refinement and Proof, 1996. .
https://en.wikipedia.org/wiki/Rare%20events
Rare or extreme events are events that occur with low frequency, and often refers to infrequent events that have a widespread effect and which might destabilize systems (for example, stock markets, ocean wave intensity or optical fibers or society). Rare events encompass natural phenomena (major earthquakes, tsunamis, hurricanes, floods, asteroid impacts, solar flares, etc.), anthropogenic hazards (warfare and related forms of violent conflict, acts of terrorism, industrial accidents, financial and commodity market crashes, etc.), as well as phenomena for which natural and anthropogenic factors interact in complex ways (epidemic disease spread, global warming-related changes in climate and weather, etc.). Overview Rare or extreme events are discrete occurrences of infrequently observed events. Despite being statistically improbable, such events are plausible insofar as historical instances of the event (or a similar event) have been documented. Scholarly and popular analyses of rare events often focus on those events that could be reasonably expected to have a substantial negative effect on a society — either economically or in terms of human casualties (typically, both). Examples of such events might include an 8.0+ Richter magnitude earthquake, a nuclear incident that kills thousands of people, or a 10%+ single-day change in the value of a stock market index. Modeling and analysis Rare event modeling (REM) refers to efforts to characterize the statistical distribution parameters, generative processes, or dynamics that govern the occurrence of statistically rare events, including but not limited to highly influential natural or human-made catastrophes. Such “modeling” may include a wide range of approaches, including, most notably, statistical models for analyzing historical event data and computational software models that attempt to simulate rare event processes and dynamics. REM also encompasses efforts to forecast the occurrence of similar events over some f
https://en.wikipedia.org/wiki/List%20of%20plasma%20physics%20articles
This is a list of plasma physics topics. A Ablation Abradable coating Abraham–Lorentz force Absorption band Accretion disk Active galactic nucleus Adiabatic invariant ADITYA (tokamak) Aeronomy Afterglow plasma Airglow Air plasma, Corona treatment, Atmospheric-pressure plasma treatment Ayaks, Novel "Magneto-plasmo-chemical engine" Alcator C-Mod Alfvén wave Ambipolar diffusion Aneutronic fusion Anisothermal plasma Anisotropy Antiproton Decelerator Appleton-Hartree equation Arcing horns Arc lamp Arc suppression ASDEX Upgrade, Axially Symmetric Divertor EXperiment Astron (fusion reactor) Astronomy Astrophysical plasma Astrophysical X-ray source Atmospheric dynamo Atmospheric escape Atmospheric pressure discharge Atmospheric-pressure plasma Atom Atomic emission spectroscopy Atomic physics Atomic-terrace low-angle shadowing Auger electron spectroscopy Aurora (astronomy) B Babcock Model Ball lightning Ball-pen probe Ballooning instability Baryon acoustic oscillations Beam-powered propulsion Beta (plasma physics) Birkeland current Blacklight Power Blazar Bohm diffusion Bohr–van Leeuwen theorem Boltzmann relation Bow shock Bremsstrahlung Bussard ramjet C Capacitively coupled plasma Carbon nanotube metal matrix composites Cassini–Huygens, Cassini Plasma Spectrometer Cathode ray Cathodic arc deposition Ceramic discharge metal-halide lamp Charge carrier Charged-device model Charged particle Chemical plasma Chemical vapor deposition Chemical vapor deposition of diamond Chirikov criterion Chirped pulse amplification Chromatography detector Chromo–Weibel instability Classical-map hypernetted-chain method Cnoidal wave Colored-particle-in-cell Coilgun Cold plasma, Ozone generator Collisionality Colored-particle-in-cell Columbia Non-neutral Torus Comet tail Compact toroid Compressibility Compton–Getting effect Contact lithography Coupling (physics) Convection cell Cooling flow Corona Corona di
https://en.wikipedia.org/wiki/Center%20for%20Biological%20Diversity
The Center for Biological Diversity is a nonprofit membership organization known for its work protecting endangered species through legal action, scientific petitions, creative media and grassroots activism. It was founded in 1989 by Kieran Suckling, Peter Galvin, Todd Schulke and Robin Silver. The center is based in Tucson, Arizona, with its headquarters in the historic Owls club building, and has offices and staff in New Mexico, Nevada, California, Oregon, Illinois, Minnesota, Alaska, Vermont, Florida and Washington, D.C. Background Given a small grant by the Fund For Wild Nature, the organization started in 1989 as a small group by the name of Greater Gila Biodiversity Project, with the objective to protect endangered species and critical habitat in the Southwestern United States. The organization grew and became the Center for Biological Diversity. Kieran Suckling, Peter Galvin, and Todd Schulke founded the organization in response to what they perceived as a failure on the part of the United States Forest Service to protect imperiled species from logging, grazing, and mining. As surveyors in New Mexico, the three men discovered "a rare Mexican spotted owl nest in an old-growth tree", but their discovery was ignored and the Forest Service continued with plans to lease the land to timber companies; Suckling, Galvin, and Schulke believed that it was within the Forest Service's mission to save sensitive species like the Mexican spotted owl from harm, and that the government had not performed its duty in deference to corporate interests. Suckling, Galvin and Schulke went to the media to register their outrage with success: the old-growth tree was protected from harm, and this success led to the founding of the Center for Biological Diversity. Suckling, Galvin and Schulke assert that in 1990 they discovered the Forest Service was allowing commercial logging within the protected habitat of the owl nests. Speaking to the New York Times in 2010 a spokeswoman for the
https://en.wikipedia.org/wiki/Levi%27s%20lemma
In theoretical computer science and mathematics, especially in the area of combinatorics on words, the Levi lemma states that, for all strings u, v, x and y, if uv = xy, then there exists a string w such that either uw = x and v = wy (if |u| ≤ |x|) or u = xw and wv = y (if |u| ≥ |x|) That is, there is a string w that is "in the middle", and can be grouped to one side or the other. Levi's lemma is named after Friedrich Wilhelm Levi, who published it in 1944. Applications Levi's lemma can be applied repeatedly in order to solve word equations; in this context it is sometimes called the Nielsen transformation by analogy with the Nielsen transformation for groups. For example, starting with an equation xα = yβ where x and y are the unknowns, we can transform it (assuming |x| ≥ |y|, so there exists t such that x=yt) to ytα = yβ, thus to tα = β. This approach results in a graph of substitutions generated by repeatedly applying Levi's lemma. If each unknown appears at most twice, then a word equation is called quadratic; in a quadratic word equation the graph obtained by repeatedly applying Levi's lemma is finite, so it is decidable if a quadratic word equation has a solution. A more general method for solving word equations is Makanin's algorithm. Generalizations The above is known as the Levi lemma for strings; the lemma can occur in a more general form in graph theory and in monoid theory; for example, there is a more general Levi lemma for traces originally due to Christine Duboc. Several proofs of Levi's Lemma for traces can be found in The Book of Traces. A monoid in which Levi's lemma holds is said to have the equidivisibility property. The free monoid of strings and string concatenation has this property (by Levi's lemma for strings), but by itself equidivisibility is not enough to guarantee that a monoid is free. However an equidivisible monoid M is free if additionally there exists a homomorphism f from M to the monoid of natural numbers (free monoid on
https://en.wikipedia.org/wiki/Decomposition%20method%20%28queueing%20theory%29
In queueing theory, a discipline within the mathematical theory of probability, the decomposition method is an approximate method for the analysis of queueing networks where the network is broken into subsystems which are independently analyzed. The individual queueing nodes are considered to be independent G/G/1 queues where arrivals are governed by a renewal process and both service time and arrival distributions are parametrised to match the first two moments of data.
https://en.wikipedia.org/wiki/Combinatorics%2C%20Probability%20and%20Computing
Combinatorics, Probability and Computing is a peer-reviewed scientific journal in mathematics published by Cambridge University Press. Its editor-in-chief is Béla Bollobás (DPMMS and University of Memphis). History The journal was established by Bollobás in 1992. Fields Medalist Timothy Gowers calls it "a personal favourite" among combinatorics journals and writes that it "maintains a high standard". Content The journal covers combinatorics, probability theory, and theoretical computer science. Currently, it publishes six issues annually. As with other journals from the same publisher, it follows a hybrid green/gold open access policy, in which authors may either place copies of their papers in an institutional repository after a six-month embargo period, or pay an open access charge to make their papers free to read on the journal's website. Abstracting and indexing The journal is abstracted and indexed in: According to the Journal Citation Reports, the journal has a 2014 impact factor of 0.623. Since 2007, it has been ranked by SCImago Journal Rank as a first-quartile journal in four areas: applied mathematics, computational theory, statistics and probability, and theoretical computer science.
https://en.wikipedia.org/wiki/Vagueness%20and%20Degrees%20of%20Truth
Vagueness and Degrees of Truth is a 2008 book by Nicholas J. Smith, in which the author examines vagueness based on the idea of "degrees of truth". It means that although some sentences are true and some are false, others possess intermediate truth values. In other words, some sentences are truer than the false sentences, but not as true as the true ones. See also Fuzzy logic Half-truth
https://en.wikipedia.org/wiki/Shore%20power
Shore power or shore supply is the provision of shoreside electrical power to a ship at berth while its main and auxiliary engines are shut down. While the term denotes shore as opposed to off-shore, it is sometimes applied to aircraft or land-based vehicles (such as campers, heavy trucks with sleeping compartments and tour buses), which may plug into grid power when parked for idle reduction. The source for land-based power may be grid power from an electric utility company, but also possibly an external remote generator. These generators may be powered by diesel or renewable energy sources such as wind or solar. Shore power saves consumption of fuel that would otherwise be used to power vessels while in port, and eliminates the air pollution associated with consumption of that fuel. A port city may have anti-idling laws that require ships to use shore power. Use of shore power may facilitate maintenance of the ship's engines and generators, and reduces noise. Oceangoing ships "Cold ironing" is specifically a shipping industry term that came into use when all ships had coal-fired engines. When a ship tied up at port, there was no need to continue to feed the fire and the iron engines would cool down, eventually going completely cold – hence the term "cold ironing". Commercial ships can use shore-supplied power for services such as cargo handling, pumping, ventilation and lighting while in port, they need not run their own diesel engines, reducing air pollution emissions. Examples are ferries and cruise ships for hotel electric power, and a salmon feeder ship uses shore power while at the salmon farm. Small craft On small private boats, electrical power supply on board is usually provided by 12 or 24 volt DC batteries whilst at sea unless the vessel has a generator. When the vessel is berthed in a marina or harbourside, mains electricity is often offered via a shore power connection. This allows the vessel to use a battery charger to recharge batteries and also
https://en.wikipedia.org/wiki/Chromosome%20abnormality
A chromosomal abnormality, chromosomal anomaly, chromosomal aberration, chromosomal mutation, or chromosomal disorder, is a missing, extra, or irregular portion of chromosomal DNA. These can occur in the form of numerical abnormalities, where there is an atypical number of chromosomes, or as structural abnormalities, where one or more individual chromosomes are altered. Chromosome mutation was formerly used in a strict sense to mean a change in a chromosomal segment, involving more than one gene. Chromosome anomalies usually occur when there is an error in cell division following meiosis or mitosis. Chromosome abnormalities may be detected or confirmed by comparing an individual's karyotype, or full set of chromosomes, to a typical karyotype for the species via genetic testing. Numerical abnormality An abnormal number of chromosomes is known as aneuploidy, and occurs when an individual is either missing a chromosome from a pair (resulting in monosomy) or has more than two chromosomes of a pair (trisomy, tetrasomy, etc.). Aneuploidy can be full, involving a whole chromosome missing or added, or partial, where only part of a chromosome is missing or added. Aneuploidy can occur with sex chromosomes or autosomes. An example of trisomy in humans is Down syndrome, which is a developmental disorder caused by an extra copy of chromosome 21; the disorder is therefore also called trisomy 21. An example of monosomy in humans is Turner syndrome, where the individual is born with only one sex chromosome, an X. Sperm aneuploidy Exposure of males to certain lifestyle, environmental and/or occupational hazards may increase the risk of aneuploid spermatozoa. In particular, risk of aneuploidy is increased by tobacco smoking, and occupational exposure to benzene, insecticides, and perfluorinated compounds. Increased aneuploidy is often associated with increased DNA damage in spermatozoa. Structural abnormalities When the chromosome's structure is altered, this can take several
https://en.wikipedia.org/wiki/Linux%20Malware%20Detect
Linux Malware Detect, abbreviated as LMD or maldet, is a software package that looks for malware on Linux systems and reports on it. Details To protect Linux systems against vulnerabilities, various software packages are available, such as rootkit detectors such as Rootkit Hunter and chkrootkit, and auditing systems like lynis. Malware detection software such as LMD and ClamAV improve the security of systems by scanning them based on the signatures of thousands of instances of known malware. For malware signatures, LMD uses various sources such as the signatures database of ClamAV and the Malware Hash Registry of Team Cymru. Besides such third party signature databases, it also maintains its own database of signatures. If a ClamAV scanner engine is already available on a system, LMD will use this for its scanner engine. This will normally give better performance than its built-in scanner engine. LMD can quarantine malware, and it can clean software that contains malicious code. The executable command of LMD is maldet. Typical command invocations are maldet -d, to check for later versions; maldet -u, to check for malware signature updates; and maldet -a, to scan the file system of the server on which LMD resides. Checking for malware signature updates is typically done in an automated manner. Besides periodic scans, real-time monitoring is also supported with the --monitor command-line argument. LMD can monitor users, paths and files in such a way. Linux Malware Detect is one of the objectives for the LPI 303 certification.
https://en.wikipedia.org/wiki/Myrtle%20L.%20Richmond
Myrtle Leila Richmond (September 30, 1882 – January 2, 1973) was an American astronomical researcher, a computer who worked at the Mount Wilson Observatory from 1913 to 1947. Early life and education Richmond was born in Vinland, Kansas, the daughter of Frank L. Richmond and Leila Delight Richmond. Her father was construction superintendent in the railroad industry. She graduated from Smith College in 1907, and earned a master's degree in 1908 at the University of Denver. She was active in Smith College alumnae activities in Los Angeles. Career Richmond taught mathematics at the University of Denver, and worked at Chamberlin Observatory in Colorado in 1909. She was a fellow in mathematics and astronomy at Goodsell Observatory in 1912, where she worked on Variable stars and a comet's orbit. Richmond joined the Mount Wilson Observatory computing department in 1913, and retired in 1947, after she "ably assisted in a large number of stellar and solar investigations." She was listed as a member of the observatory's "investigatory staff" in 1917. Her work also helped to establish the location of the planet Pluto, and of the moons of Jupiter. She contributed to several observatory publications, including A photometric study of the pleiades (1931, with Harlow Shapley), Mean distribution of stars according to apparent magnitude and galactic latitude (1925), The mean color-index of stars of different apparent magnitudes. Some relations between magnitude scales (1925), and Mount Wilson catalogue of photographic magnitudes in selected areas 1–139 (1930). She co-authored articles with American astronomer Seth Barnes Nicholson and Danish astronomer Julie Vinter Hansen. Personal life Richmond enjoyed hiking. She died in 1973, aged 90 years, in Pasadena. Her gravesite is in Woodstock, Vermont, her father's hometown.
https://en.wikipedia.org/wiki/IBM%20System/34
The IBM System/34 was an IBM midrange computer introduced in 1977. It was withdrawn from marketing in February 1985. It was a multi-user, multi-tasking successor to the single-user System/32. It included two processors, one based on the System/32 and the second based on the System/3. Like the System/32 and the System/3, the System/34 was primarily programmed in the RPG II language. Hardware The 5340 System Unit contained the processing unit, the disk storage and the diskette drive. It had several access doors on both sides. Inside, were swing-out assemblies where the circuit boards and memory cards were mounted. It weighed and used 220V power. The IBM 5250 series of terminals were the primary interface to the System/34. Processors S/34s had two processors, the Control Storage Processor (CSP), and the Main Storage Processor (MSP). The MSP was the workhorse, based on System/3 architecture; it performed the instructions in the computer programs. The CSP was the governor, a different processor with different RISC-like instruction set, based on System/32 architecture; it performed system functions in the background. The CSP also executed the optional Scientific Macroinstructions, which were a set of emulated floating point operations used by the System/34 Fortran compiler and optionally in assembly code. The clock speed of the CPUs inside a System/34 was fixed at 1 MHz for the MSP and 4 MHz for the CSP. Special utility programs were able to make direct calls to the CSP to perform certain functions; these are usually system programs like $CNFIG which was used to configure the computer system. Memory and storage The smallest S/34 had 48K of RAM and an 8.6 MB hard drive. The largest configured S/34 could support 256K of RAM and 256MB of disk space. S/34 hard drives contained a feature called "the extra cylinder," so that bad spots on the drive were detected and dynamically mapped out to good spots on the extra cylinder. Disk space on the System/34 was organized by bloc
https://en.wikipedia.org/wiki/Vine%20Linux
is a Japanese Linux distribution sponsored by VineCaves. It has been a fork of Red Hat Linux 7.2 since Vine Linux 3.0. Work on Vine Linux was started in 1998. All versions except Vine Seed have been announced to be discontinued from May 4, 2021. Release history
https://en.wikipedia.org/wiki/Link-richness
Link-richness is the quality, possessed by some websites, of having many hyperlinks. Classified advertising sites like Craigslist tend to be very link-rich, sometimes with hundreds of links on their main page. They help users find the links they are looking for by grouping links into clusters. Inadequate link richness has been described as frustrating to readers, as it reduces transparency of site content from the main page. Students new to wiki collaboration were found to need guidance in how to take full advantage of the medium's potential for creating link-rich content. Link-richness in some contexts can be distracting, as when an article is surrounded by extraneous links. Indeed, it is becoming accepted as a best practice for universities to have link-rich home pages that do not rely on user categorisation and exploration of long sequences of links and are not constrained by traditional boundaries between departments. Tools are sometimes needed to make the publishing of link-rich web sites tractable, and many people may lack the technical skills, time, or inclination to engage in hand- crafting new digital document forms. A link-rich site that is low on content is sometimes referred to as a "gateway site." Link-rich portals were popular on the Web in 2000. Yahoo! and other sites featuring categories with many links were heavily used and often required fewer than three clicks to reach the content. Web designers were creating flat sites with content positioned close to the top of pages.
https://en.wikipedia.org/wiki/The%20Geography%20of%20Thought
The Geography of Thought: How Asians and Westerners Think Differently...and Why is a book by social psychologist Richard Nisbett that was published by Free Press in 2003. By analyzing the differences between Asia and the West, it argues that cultural differences affect people's thought processes more significantly than believed. Thesis In the book, Nisbett demonstrates that "people actually think about—and even see—the world differently because of differing ecologies, social structures, philosophies, and educational systems that date back to ancient Greece and China". At its core, the book assumes that human behavior is not “hard-wired” but a function of culture. The book proposes that the passion for strong ontology and scientific rationality based on forward chaining from axioms is essentially a "Western" phenomenon. Ancient Greece's passion for abstract categories into which the entire world can be taxonomically arranged, he claims, is prototypically Western, as is the notion of causality. In other words, he claims that the law of the excluded middle is not applied in Chinese thought, and that a different standard applies. This has been described by other thinkers as being hermeneutic reasonableness. Implications There are several implications to Nisbett's theory. For instance, in law, Eastern and Western cultures assign different priorities to, and roles of, the law in society. The ratio of lawyers to engineers is forty times higher in the US than in Japan. Moreover, the role of US lawyers is, generally, to handle legal confrontations, and the aim is demands for justice with a clear winner and loser based upon universal principles of justice that apply equally to everyone. In contrast, Eastern lawyers are more often used as intermediaries to reduce hostilities, and reach a compromise; the principles they operate by are more flexible and circumstantial. Another aspect where there is great divergence between these two systems of thought concerns human rights.
https://en.wikipedia.org/wiki/Visual%20sensor%20network
A visual sensor network or smart camera network or intelligent camera network is a network of spatially distributed smart camera devices capable of processing, exchanging data and fusing images of a scene from a variety of viewpoints into some form more useful than the individual images. A visual sensor network may be a type of wireless sensor network, and much of the theory and application of the latter applies to the former. The network generally consists of the cameras themselves, which have some local image processing, communication and storage capabilities, and possibly one or more central computers, where image data from multiple cameras is further processed and fused (this processing may, however, simply take place in a distributed fashion across the cameras and their local controllers). Visual sensor networks also provide some high-level services to the user so that the large amount of data can be distilled into information of interest using specific queries. The primary difference between visual sensor networks and other types of sensor networks is the nature and volume of information the individual sensors acquire: unlike most sensors, cameras are directional in their field of view, and they capture a large amount of visual information which may be partially processed independently of data from other cameras in the network. Alternatively, one may say that while most sensors measure some value such as temperature or pressure, visual sensors measure patterns. In light of this, communication in visual sensor networks differs substantially from traditional sensor networks. Applications Visual sensor networks are most useful in applications involving area surveillance, tracking, and environmental monitoring. Of particular use in surveillance applications is the ability to perform a dense 3D reconstruction of a scene and storing data over a period of time, so that operators can view events as they unfold over any period of time (including the current moment) f
https://en.wikipedia.org/wiki/Beta-lactamase%20inhibitor%20protein
Beta-Lactamase Inhibitor Proteins (BLIPs) are a family of proteins produced by bacterial species including Streptomyces. BLIP acts as a potent inhibitor of beta-lactamases such as TEM-1, which is the most widespread resistance enzyme to penicillin antibiotics. BLIP binds competitively the surface of TEM-1 and inserting residues into the active site to make direct contacts with catalytic residues. BLIP is able to inhibit a variety of class A beta-lactamases, possibly through flexibility of its two domains. The two tandemly repeated domains of BLIP have an α2-β4 structure, the β-hairpin loop from domain 1 inserting into the active site of beta-lactamase. BLIP shows no sequence similarity with BLIP-II, even though both bind to and inhibit TEM-1.
https://en.wikipedia.org/wiki/Target%202035
Target 2035 is a global effort or movement to discover open science, pharmacological modulator(s) for every protein in the human proteome by the year 2035. The effort is led by the Structural Genomics Consortium with the intention that this movement evolves organically. Target 2035 has been borne out of the success that chemical probes have had in elevating or de-prioritizing the therapeutic potential of protein targets. The availability of open access pharmacological tools is a largely unmet aspect of drug discovery especially for the dark proteome. The first five years will include building mechanisms (Phase 1 below) which allow researchers to find collaborators with like-minded goals towards discovering a pharmacological tool for a specific protein or protein family, and make it open access (without encumbrances due to intellectual property). One strategic goal is seeding new open science programs on components of the drug discovery pipeline with the goal to bring medicines to the bedside equitably, affordably and rapidly. Phase 1 will also build a framework that welcomes new and (re-)emerging enabling technologies in hit-finding and characterization. An update on the progress was published. Target 2035 will draw on successes from past and current publicly-funded programs including National Institutes of Health (NIH) Illuminating the Druggable Genome initiative for under-explored kinases, GPCR’s and ion channels, Innovative Medicines Initiative's RESOLUTE project on human SLCs, Innovative Medicines Initiative's Enabling and Unlocking Biology in the Open (EUbOPEN), and Innovative Medicines Initiative's Unrestricted Leveraging of Targets for Research Advancement and Drug Discovery. The NIH recently re-iterated their commitment to making their data open to mitigate the tens of billions due to irreproducible data. Target 2035 will collaborate with the Chemical Probes Portal and open science platforms, e.g. Just One Giant Lab, in order to spread awareness and ed
https://en.wikipedia.org/wiki/Character%20variety
In the mathematics of moduli theory, given an algebraic, reductive, Lie group and a finitely generated group , the -character variety of is a space of equivalence classes of group homomorphisms from to : More precisely, acts on by conjugation, and two homomorphisms are defined to be equivalent (denoted ) if and only if their orbit closures intersect. This is the weakest equivalence relation on the set of conjugation orbits, , that yields a Hausdorff space. Formulation Formally, and when the reductive group is defined over the complex numbers , the -character variety is the spectrum of prime ideals of the ring of invariants (i.e., the affine GIT quotient). Here more generally one can consider algebraically closed fields of prime characteristic. In this generality, character varieties are only algebraic sets and are not actual varieties. To avoid technical issues, one often considers the associated reduced space by dividing by the radical of 0 (eliminating nilpotents). However, this does not necessarily yield an irreducible space either. Moreover, if we replace the complex group by a real group we may not even get an algebraic set. In particular, a maximal compact subgroup generally gives a semi-algebraic set. On the other hand, whenever is free we always get an honest variety; it is singular however. Examples An interesting class of examples arise from Riemann surfaces: if is a Riemann surface then the -character variety of , or Betti moduli space, is the character variety of the surface group . For example, if and is the Riemann sphere punctured three times, so is free of rank two, then Henri G. Vogt, Robert Fricke, and Felix Klein proved that the character variety is ; its coordinate ring is isomorphic to the complex polynomial ring in 3 variables, . Restricting to gives a closed real three-dimensional ball (semi-algebraic, but not algebraic). Another example, also studied by Vogt and Fricke–Klein is the case with and is the Riemann
https://en.wikipedia.org/wiki/Aruba%20S.p.A.
Aruba S.p.A is an Italian company mainly active in the Web hosting and domain registration businesses. It is the market leader in Italy, and also has a large market share in the Czech Republic and Slovakia.
https://en.wikipedia.org/wiki/Lucas%20chain
In mathematics, a Lucas chain is a restricted type of addition chain, named for the French mathematician Édouard Lucas. It is a sequence a0, a1, a2, a3, ... that satisfies a0=1, and for each k > 0: ak = ai + aj, and either ai = aj or |ai − aj| = am, for some i, j, m < k. The sequence of powers of 2 (1, 2, 4, 8, 16, ...) and the Fibonacci sequence (with a slight adjustment of the starting point 1, 2, 3, 5, 8, ...) are simple examples of Lucas chains. Lucas chains were introduced by Peter Montgomery in 1983. If L(n) is the length of the shortest Lucas chain for n, then Kutz has shown that most n do not have L < (1-ε) logφ n, where φ is the Golden ratio.
https://en.wikipedia.org/wiki/LOC100272216
Uncharacterized LOC100272216 is a protein present in humans that is encoded by the LOC100272216 gene.
https://en.wikipedia.org/wiki/Three-finger%20toxin
Three-finger toxins (abbreviated 3FTx) are a protein superfamily of small toxin proteins found in the venom of snakes. Three-finger toxins are in turn members of a larger superfamily of three-finger protein domains which includes non-toxic proteins that share a similar protein fold. The group is named for its common structure consisting of three beta strand loops connected to a central core containing four conserved disulfide bonds. The 3FP protein domain has no enzymatic activity and is typically between 60-74 amino acid residues long. Despite their conserved structure, three-finger toxin proteins have a wide range of pharmacological effects. Most members of the family are neurotoxins that act on cholinergic intercellular signaling; the alpha-neurotoxin family interacts with muscle nicotinic acetylcholine receptors (nAChRs), the kappa-bungarotoxin family with neuronal nAChRs, and muscarinic toxins with muscarinic acetylcholine receptors (mAChRs). Structure The three-finger toxin superfamily is defined by a common tertiary structure consisting of three beta strand-containing loops (designated loops I, II, and III) projecting from a small hydrophobic core containing four conserved disulfide bonds. This structure is thought to resemble a hand with three fingers, giving rise to the name. The proteins are typically 60-74 amino acid residues long, though some have additional N- or C-terminal extensions. An additional disulfide bond may be present in either loop I or loop II. The superfamily can be broadly divided into three classes: short-chain toxins have under 66 residues and four core disulfide bonds. long-chain toxins have at least 66 residues, a disulfide bond in loop II, and possibly a C-terminal extension. non-conventional toxins have a disulfide bond in loop I and possibly terminal extensions. Oligomerization Most 3FTx proteins are monomers. However, some 3FTx subgroups form functional non-covalent homodimers. The kappa-bungarotoxin group is the best chara
https://en.wikipedia.org/wiki/Lyman-break%20galaxy
Lyman-break galaxies are star-forming galaxies at high redshift that are selected using the differing appearance of the galaxy in several imaging filters due to the position of the Lyman limit. The technique has primarily been used to select galaxies at redshifts of z = 3–4 using ultraviolet and optical filters, but progress in ultraviolet astronomy and in infrared astronomy has allowed the use of this technique at lower and higher redshifts using ultraviolet and near-infrared filters. The Lyman-break galaxy selection technique relies on the fact that radiation at higher energies than the Lyman limit at 912 Å is almost completely absorbed by neutral gas around star-forming regions of galaxies. In the rest frame of the emitting galaxy, the emitted spectrum is bright at wavelengths longer than 912 Å, but very dim or imperceptible at shorter wavelengths—this is known as a "dropout", or "break", and can be used to find the position of the Lyman limit. Light with a wavelength shorter than 912 Å is in the far-ultraviolet range and is blocked by the Earth's atmosphere, but for very distant galaxies the wavelengths of light are stretched considerably because of the expansion of the Universe. For a galaxy at redshift z = 3, the Lyman break will appear to be at wavelengths of about 3600 Å, which is long enough to be detected by ground- or space-based telescopes. Candidate galaxies around redshift z = 3 can then be selected by looking for galaxies which appear in optical images (which are sensitive to wavelengths greater than 3600 Å), but do not appear in ultraviolet images (which are sensitive to light at wavelengths shorter than 3600 Å). The technique may be adapted to look for galaxies at other redshifts by choosing different sets of filters—the method works as long as images may be taken through at least one filter above and below the wavelength of the redshifted Lyman limit. In order to confirm the redshift estimated by the color selection, follow-up spectroscopy i
https://en.wikipedia.org/wiki/Binary%20code
A binary code represents text, computer processor instructions, or any other data using a two-symbol system. The two-symbol system used is often "0" and "1" from the binary number system. The binary code assigns a pattern of binary digits, also known as bits, to each character, instruction, etc. For example, a binary string of eight bits (which is also called a byte) can represent any of 256 possible values and can, therefore, represent a wide variety of different items. In computing and telecommunications, binary codes are used for various methods of encoding data, such as character strings, into bit strings. Those methods may use fixed-width or variable-width strings. In a fixed-width binary code, each letter, digit, or other character is represented by a bit string of the same length; that bit string, interpreted as a binary number, is usually displayed in code tables in octal, decimal or hexadecimal notation. There are many character sets and many character encodings for them. A bit string, interpreted as a binary number, can be translated into a decimal number. For example, the lower case a, if represented by the bit string 01100001 (as it is in the standard ASCII code), can also be represented as the decimal number "97". History of binary codes The modern binary number system, the basis for binary code, was invented by Gottfried Leibniz in 1689 and appears in his article Explication de l'Arithmétique Binaire. The full title is translated into English as the "Explanation of the binary arithmetic", which uses only the characters 1 and 0, with some remarks on its usefulness, and on the light it throws on the ancient Chinese figures of Fu Xi. Leibniz's system uses 0 and 1, like the modern binary numeral system. Leibniz encountered the I Ching through French Jesuit Joachim Bouvet and noted with fascination how its hexagrams correspond to the binary numbers from 0 to 111111, and concluded that this mapping was evidence of major Chinese accomplishments in the sor
https://en.wikipedia.org/wiki/Peixoto%27s%20theorem
In the theory of dynamical systems, Peixoto's theorem, proved by Maurício Peixoto, states that among all smooth flows on surfaces, i.e. compact two-dimensional manifolds, structurally stable systems may be characterized by the following properties: The set of non-wandering points consists only of periodic orbits and fixed points. The set of fixed points is finite and consists only of hyperbolic equilibrium points. Finiteness of attracting or repelling periodic orbits. Absence of saddle-to-saddle connections. Moreover, they form an open set in the space of all flows endowed with C1 topology. See also Andronov–Pontryagin criterion
https://en.wikipedia.org/wiki/MHC%20multimer
MHC multimers are oligomeric forms of MHC molecules, designed to identify and isolate T-cells with high affinity to specific antigens amid a large group of unrelated T-cells. Multimers generally range in size from dimers to octamers; however, some companies use even higher quantities of MHC per multimer. Multimers may be used to display class 1 MHC, class 2 MHC, or nonclassical molecules (e.g. CD1d) from species such as monkeys, mice, and humans. Background Since T-cell receptors have a low affinity for their MHC counterparts, it was historically problematic to label T cells effectively using single MHC-T-cell interactions. However, in 1996 it was proposed by John Altman to use a complex of multiple MHC molecules to form a more stable bond between corresponding T-cells. Production The most commonly used MHC multimers are tetramers. These are typically produced by biotinylating soluble MHC monomers, which are typically produced recombinantly in eukaryotic or bacterial cells. These monomers then bind to a backbone, such as streptavidin or avidin, creating a tetravalent structure. These backbones are conjugated with fluorochromes to subsequently isolate bound T-cells via flow cytometry. Potential clinical applications MHC multimers allow for a previously unattainable level of specificity in antigen-specific T-cell detection and isolation. This ability gives rise to several clinical applications. MHC multimers allow for ex vivo selection and proliferation of T-cells specific to viral or tumor-related antigens, which can then be reintroduced to augment the immune system. MHC multimers can also be used to eliminate graft-originating T-cells on transplant organs, ex vivo. MHC multimers may also be used to eliminate harmful or unwanted T-cells in vivo, such as those that target self cells and lead to autoimmune disease. Cancer immunotherapy and vaccine development can also be largely influenced by this technology. Sub-types MHC tetramer MHC tetramers consist of fou
https://en.wikipedia.org/wiki/Calcium-dependent%20chloride%20channel
The Calcium-Dependent Chloride Channel (Ca-ClC) proteins (or calcium-activated chloride channels (CaCCs), are heterogeneous groups of ligand-gated ion channels for chloride that have been identified in many epithelial and endothelial cell types as well as in smooth muscle cells. They include proteins from several structurally different families: chloride channel accessory (CLCA), bestrophin (BEST), and calcium-dependent chloride channel anoctamin (ANO or TMEM16) channels ANO1 is highly expressed in human gastrointestinal interstitial cells of Cajal, which are proteins which serve as intestinal pacemakers for peristalsis. In addition to their role as chloride channels some CLCA proteins function as adhesion molecules and may also have roles as tumour suppressors. These eukaryotic proteins are "required for normal electrolyte and fluid secretion, olfactory perception, and neuronal and smooth muscle excitability" in animals. Members of the Ca-CIC family are generally 600 to 1000 amino acyl residues (aas) in length and exhibit 7 to 10 transmembrane segments (TMSs). Function Tmc1 and Tmc2 (TC#s 1.A.17.4.6 and 1.A.17.4.1, respectively) may play a role in hearing and are required for normal function of cochlear hair cells, possibly as Ca2+ channels or Ca2+ channel subunits (see also family TC# 1.A.82). Mice lacking both channels lack hair cell mechanosensory potentials. There are 8 members of this family in humans, 1 in Drosophila and 2 in C. elegans. One of the latter two is expressed in mechanoreceptors. Tmc1 is a sodium-sensitive cation channel required for salt (Na+) chemosensation in C. elegans "where it is required for salt-evoked neuronal activity and behavioural avoidance of high concentrations of NaCl". TMEM16A is over-expressed in several tumor types. The role of TMEM16A in gliomas and the potential underlying mechanisms were analyzed by Liu et al. 2014. Knockdown of TMEM16A suppressed cell proliferation, migration and invasion. The reactions believed to
https://en.wikipedia.org/wiki/Bring%20radical
In algebra, the Bring radical or ultraradical of a real number a is the unique real root of the polynomial The Bring radical of a complex number a is either any of the five roots of the above polynomial (it is thus multi-valued), or a specific root, which is usually chosen such that the Bring radical is real-valued for real a and is an analytic function in a neighborhood of the real line. Because of the existence of four branch points, the Bring radical cannot be defined as a function that is continuous over the whole complex plane, and its domain of continuity must exclude four branch cuts. George Jerrard showed that some quintic equations can be solved in closed form using radicals and Bring radicals, which had been introduced by Erland Bring. In this article, the Bring radical of a is denoted For real argument, it is odd, monotonically decreasing, and unbounded, with asymptotic behavior for large . Normal forms The quintic equation is rather difficult to obtain solutions for directly, with five independent coefficients in its most general form: The various methods for solving the quintic that have been developed generally attempt to simplify the quintic using Tschirnhaus transformations to reduce the number of independent coefficients. Principal quintic form The general quintic may be reduced into what is known as the principal quintic form, with the quartic and cubic terms removed: If the roots of a general quintic and a principal quintic are related by a quadratic Tschirnhaus transformation the coefficients α and β may be determined by using the resultant, or by means of the power sums of the roots and Newton's identities. This leads to a system of equations in α and β consisting of a quadratic and a linear equation, and either of the two sets of solutions may be used to obtain the corresponding three coefficients of the principal quintic form. This form is used by Felix Klein's solution to the quintic. Bring–Jerrard normal form It is possible to
https://en.wikipedia.org/wiki/International%20Association%20for%20Cereal%20Science%20and%20Technology
The International Association for Cereal Science and Technology (ICC) was founded in 1955 and was originally called the International Association for Cereal Chemistry. It was set up to develop international standard testing procedures for cereals and flour. It has currently more than fifty member countries and headquarters in Vienna, Austria. External links ICC home page List of ICC standard methods Standards organisations in Austria Food technology organizations 1955 establishments in Austria Cereals Chemistry organizations
https://en.wikipedia.org/wiki/Lorica%20%28biology%29
In biology, a lorica is a shell-like protective outer covering, often reinforced with sand grains and other particles that some protozoans and loriciferan animals secrete. Usually it is tubular or conical in shape, with a loose case that is closed at one end. An example is the protozoan genus Stentor, in which the lorica is trumpet-shaped. In the tintinnids, the lorica is frequently transparent and is used as domicile. Halofolliculina corallasia has a lorica that is attached as an outer structure, and into which it retracts when disturbed. There are three phases in the formation of lorica: agglomeration in a natural cast; helical extension; and stabilization. The original meaning of the word is: cuirass, a type of chest armor, originally made of leather, afterward of plates of metal or horn sewed on linen or the like. See also Chitin Periostracum Tectin Notes Ciliate biology
https://en.wikipedia.org/wiki/Epigenetic%20code
The epigenetic code is hypothesised to be a defining code in every eukaryotic cell consisting of the specific epigenetic modification in each cell. It consists of histone modifications defined by the histone code and additional epigenetic modifications such as DNA methylation. The base for the epigenetic code is a system above the genetic code of a single cell. While in one individual the genetic code in each cell is the same, the epigenetic code is tissue and cell specific. The epigenetic code can be multidimensional in nature. It could include any of the three major cellular macromolecucles; namely, DNA (code independent), RNA, and/or protein. In some ciliates potential structural codes have also been suggested. See also Histone code Epigenetics DNA methylation
https://en.wikipedia.org/wiki/Integrated%20circuit%20layout
In integrated circuit design, integrated circuit (IC) layout, also known IC mask layout or mask design, is the representation of an integrated circuit in terms of planar geometric shapes which correspond to the patterns of metal, oxide, or semiconductor layers that make up the components of the integrated circuit. Originally the overall process was called tapeout, as historically early ICs used graphical black crepe tape on mylar media for photo imaging (erroneously believed to reference magnetic data—the photo process greatly predated magnetic media). When using a standard process—where the interaction of the many chemical, thermal, and photographic variables is known and carefully controlled—the behaviour of the final integrated circuit depends largely on the positions and interconnections of the geometric shapes. Using a computer-aided layout tool, the layout engineer—or layout technician—places and connects all of the components that make up the chip such that they meet certain criteria—typically: performance, size, density, and manufacturability. This practice is often subdivided between two primary layout disciplines: analog and digital. The generated layout must pass a series of checks in a process known as physical verification. The most common checks in this verification process are Design rule checking (DRC), Layout versus schematic (LVS), parasitic extraction, antenna rule checking, and electrical rule checking (ERC). When all verification is complete, layout post processing is applied where the data is also translated into an industry-standard format, typically GDSII, and sent to a semiconductor foundry. The milestone completion of the layout process of sending this data to the foundry is now colloquially called "tapeout". The foundry converts the data into mask data and uses it to generate the photomasks used in a photolithographic process of semiconductor device fabrication. In the earlier, simpler, days of IC design, layout was done by
https://en.wikipedia.org/wiki/Curry%27s%20paradox
Curry's paradox is a paradox in which an arbitrary claim F is proved from the mere existence of a sentence C that says of itself "If C, then F", requiring only a few apparently innocuous logical deduction rules. Since F is arbitrary, any logic having these rules allows one to prove everything. The paradox may be expressed in natural language and in various logics, including certain forms of set theory, lambda calculus, and combinatory logic. The paradox is named after the logician Haskell Curry. It has also been called Löb's paradox after Martin Hugo Löb, due to its relationship to Löb's theorem. In natural language Claims of the form "if A, then B" are called conditional claims. Curry's paradox uses a particular kind of self-referential conditional sentence, as demonstrated in this example: Even though Germany does not border China, the example sentence certainly is a natural-language sentence, and so the truth of that sentence can be analyzed. The paradox follows from this analysis. The analysis consists of two steps. First, common natural-language proof techniques can be used to prove that the example sentence is true. Second, the truth of the example sentence can be used to prove that Germany borders China. Because Germany does not border China, this suggests that there has been an error in one of the proofs. The claim "Germany borders China" could be replaced by any other claim, and the sentence would still be provable. Thus every sentence appears to be provable. Because the proof uses only well-accepted methods of deduction, and because none of these methods appears to be incorrect, this situation is paradoxical. Informal proof The standard method for proving conditional sentences (sentences of the form "if A, then B") is called "conditional proof". In this method, in order to prove "if A, then B", first A is assumed and then with that assumption B is shown to be true. To produce Curry's paradox, as described in the two steps above, apply this
https://en.wikipedia.org/wiki/Identifier-Locator%20Network%20Protocol
The Identifier-Locator Network Protocol (ILNP) is a network protocol that divides the two functions of network addresses, namely the identification of network endpoints, and assisting routing, by separating topological information from node identity. ILNP is backwards-compatible with existing Internet Protocol functions, and is incrementally deployable. ILNP has an architecture with two different instantiations. ILNPv4 is ILNP engineered to work as a set of IPv4 extensions, while ILNPv6 has a set of IPv6 extensions. At least three independent open-source implementations of ILNPv6 exist. University of St Andrews (Scotland) has a prototype in Linux/x86 and FreeBSD/x86, while Tsinghua U. (China) has a prototype in Linux/x86. The University of St Andrews ILNP group is led by Prof. Saleem Bhatti. Other academics involved in continuing research include Ryo Yanagida, Samuel J. Ivey and Gregor Haywood. In February 2011, the IRTF Routing Research Group (RRG) Chairs recommended that the IETF standardise ILNP () as the preferred evolutionary direction for IPv6. RFC specifications – ILNP Architectural Description – ILNP Engineering Considerations – DNS Resource Records for ILNP – ICMPv6 Locator Update Message for ILNPv6 – IPv6 Nonce Destination Option for ILNPv6 – ICMP Locator Update for IPv4 – IPv4 Options for ILNPv4 – Address Resolution Protocol (ARP) for ILNPv4 – Optional Advanced Deployment Scenarios for ILNP See also Host Identity Protocol Locator/Identifier Separation Protocol Mobile IP Proxy Mobile IPv6
https://en.wikipedia.org/wiki/Noise%20measurement
In acoustics, noise measurement can be for the purpose of measuring environmental noise or measuring noise in the workplace. Applications include monitoring of construction sites, aircraft noise, road traffic noise, entertainment venues and neighborhood noise. One of the definitions of noise covers all "unwanted sounds". When sound levels reach a high enough intensity, the sound, whether it is wanted or unwanted, may be damaging to hearing. Environmental noise monitoring is the measurement of noise in an outdoor environment caused by transport (e.g. motor vehicles, aircraft, and trains), industry (e.g. machines) and recreational activities (e.g. music). The laws and limits governing environmental noise monitoring differ from country to country. At the very least, noise may be annoying or displeasing or may disrupt the activity or balance of human or animal life, increasing levels of aggression, hypertension and stress. In the extreme, excessive levels or periods of noise can have long-term negative health effects such as hearing loss, tinnitus, sleep disturbances, a rise in blood pressure, an increase in stress and vasoconstriction, and an increased incidence of coronary artery disease. In animals, noise can increase the risk of death by altering predator or prey detection and avoidance, interfering with reproduction and navigation, and contributing to permanent tinnitus and hearing loss. Various interventions are available to combat environmental noise. Roadway noise can be reduced by the use of noise barriers, limitation of vehicle speeds, alteration of roadway surface texture, limitation of heavy vehicles, use of traffic controls that smooth vehicle flow to reduce braking and acceleration, and tire design. Aircraft noise can be reduced by using quieter jet engines, altering flight paths and considering the time of day to benefit residents near airports. Industrial noise is addressed by redesign of industrial equipment, shock mounted assemblies and physical bar
https://en.wikipedia.org/wiki/Recombineering
Recombineering (recombination-mediated genetic engineering) is a genetic and molecular biology technique based on homologous recombination systems, as opposed to the older/more common method of using restriction enzymes and ligases to combine DNA sequences in a specified order. Recombineering is widely used for bacterial genetics, in the generation of target vectors for making a conditional mouse knockout, and for modifying DNA of any source often contained on a bacterial artificial chromosome (BAC), among other applications. Development Although developed in bacteria, much of the inspiration for recombineering techniques came from methods first developed in Saccharomyces cerevisiae where a linear plasmid was used to target genes or clone genes off the chromosome. In addition, recombination with single-strand oligonucleotides (oligos) was first shown in Saccharomyces cerevisiae. Recombination was observed to take place with oligonucleotides as short as 20 bases. Recombineering is based on homologous recombination in Escherichia coli mediated by bacteriophage proteins, either RecE/RecT from Rac prophage or Redαβδ from bacteriophage lambda. The lambda Red recombination system is now most commonly used and the first demonstrations of Red in vivo genetic engineering were independently made by Kenan Murphy and Francis Stewart. However, Murphy's experiments required expression of RecA and also employed long homology arms. Consequently, the implications for a new DNA engineering technology were not obvious. The Stewart lab showed that these homologous recombination systems mediate efficient recombination of linear DNA molecules flanked by homology sequences as short as 30 base pairs (40-50 base pairs are more efficient) into target DNA sequences in the absence of RecA. Now the homology could be provided by oligonucleotides made to order, and standard recA cloning hosts could be used, greatly expanding the utility of recombineering. Recombineering with dsDNA Recomb
https://en.wikipedia.org/wiki/Principal%20orbit%20type%20theorem
In mathematics, the principal orbit type theorem states that compact Lie group acting smoothly on a connected differentiable manifold has a principal orbit type. Definitions Suppose G is a compact Lie group acting smoothly on a connected differentiable manifold M. An isotropy group is the subgroup of G fixing some point of M. An isotropy type is a conjugacy class of isotropy groups. The principal orbit type theorem states that there is a unique isotropy type such that the set of points of M with isotropy groups in this isotropy type is open and dense. The principal orbit type is the space G/H, where H is a subgroup in the isotropy type above.
https://en.wikipedia.org/wiki/Berezin%20transform
In mathematics — specifically, in complex analysis — the Berezin transform is an integral operator acting on functions defined on the open unit disk D of the complex plane C. Formally, for a function ƒ : D → C, the Berezin transform of ƒ is a new function Bƒ : D → C defined at a point z ∈ D by where denotes the complex conjugate of w and is the area measure. It is named after Felix Alexandrovich Berezin.