source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Oracle%20unified%20method | The Oracle unified method (OUM), first released by Oracle Corporation in 2006, is a standards-based method with roots in the unified process (UP). OUM is business-process and use-case driven and includes support for the Unified Modeling Language (UML), though the use of UML is not required. OUM combines these standards with aspects of Oracle's legacy methods and Oracle implementation best-practices.
OUM is applicable to any size or type of information technology project. While OUM is a plan-based method – that includes extensive overview material, task and artifact descriptions, and associated templates – the method is intended to be tailored to support the appropriate level of ceremony required for each project. Guidance is provided for identifying the minimum subset of tasks, tailoring the project approach, executing iterative and incremental project planning, and applying agile techniques. Supplemental guidance provides specific support for Oracle products, tools, and technologies.
Supported topics
OUM v6.4.0 provides support for:
Application implementation
Cloud application services implementation
Software upgrade projects
as well as the complete range of technology projects including:
Business intelligence (BI)
Enterprise security
WebCenter
Service-oriented architecture (SOA)
Application Integration Architecture (AIA)
Business process management (BPM)
Enterprise integration
Custom software
Detailed techniques and tool guidance are provided, including a supplemental guide related to Oracle Tutor and UPK.
Availability
OUM is available for use by Oracle employees; for Oracle PartnerNetwork Diamond, Platinum, and Gold Partners; and for customers who participate in the OUM Customer Program.
Legacy method retirement dates:
Oracle Custom Development Method (CDM), February 2010
Oracle CDM Fast Track, February 2010
Oracle Application Implementation Methodology (AIM), January 2011
Oracle AIM for Business Flows, January 2011
Oracle's Siebel Result |
https://en.wikipedia.org/wiki/Birger%20M%C3%B8ller-Pedersen | Birger Møller-Pedersen (born 11 November 1949) is a computer scientist and professor at the University of Oslo, department of informatics. He published numerous works on object-oriented programming and has contributed to the creation of the BETA programming language, which is a descendant of Simula.
Academic work
Møller-Pedersen is a professor at the department of informatics at the University of Oslo, Norway. He teaches courses mainly in compiler design and programming languages. |
https://en.wikipedia.org/wiki/Complex%20quadratic%20polynomial | A complex quadratic polynomial is a quadratic polynomial whose coefficients and variable are complex numbers.
Properties
Quadratic polynomials have the following properties, regardless of the form:
It is a unicritical polynomial, i.e. it has one finite critical point in the complex plane, Dynamical plane consist of maximally 2 basins: basin of infinity and basin of finite critical point ( if finite critical point do not escapes)
It can be postcritically finite, i.e. the orbit of the critical point can be finite, because the critical point is periodic or preperiodic.
It is a unimodal function,
It is a rational function,
It is an entire function.
Forms
When the quadratic polynomial has only one variable (univariate), one can distinguish its four main forms:
The general form: where
The factored form used for the logistic map:
which has an indifferent fixed point with multiplier at the origin
The monic and centered form,
The monic and centered form has been studied extensively, and has the following properties:
It is the simplest form of a nonlinear function with one coefficient (parameter),
It is a centered polynomial (the sum of its critical points is zero).
it is a binomial
The lambda form is:
the simplest non-trivial perturbation of unperturbated system
"the first family of dynamical systems in which explicit necessary and sufficient conditions are known for when a small divisor problem is stable"
Conjugation
Between forms
Since is affine conjugate to the general form of the quadratic polynomial it is often used to study complex dynamics and to create images of Mandelbrot, Julia and Fatou sets.
When one wants change from to :
When one wants change from to , the parameter transformation is
and the transformation between the variables in and is
With doubling map
There is semi-conjugacy between the dyadic transformation (the doubling map) and the quadratic polynomial case of c = –2.
Notation
Iteration
Here denotes the n-th it |
https://en.wikipedia.org/wiki/Distributed%20multipole%20analysis | In computational chemistry, distributed multipole analysis (DMA) is a compact and accurate way of describing the spatial distribution of electric charge within a molecule.
Multipole expansion
The DMA method was devised by Prof. Anthony Stone of Cambridge University to describe the charge distribution of a molecule in terms of a multipole expansion around a number of centers. The idea of using a multi-center multipole expansion was earlier proposed by Robert Rein. Typically, the centers correspond to the atoms constituting the molecule, though this is not a requirement. A multipole series, consisting of a charge, dipole, quadrupole and higher terms is located at each center. Importantly, the radius of convergence of this multipole series is sufficiently small that the relevant series will be convergent when describing two molecules in van der Waals contact.
The DMA series are derived from ab initio or density functional theory calculations using Gaussian basis sets. If the molecular orbitals are written as linear combinations of atomic basis functions the electron density takes the form of a sum of products of the basis functions, called density matrix elements. Boys (1950) showed that the product of two spherical Gaussian functions, centered at different points, can be expressed as a single Gaussian at an intermediate point known as the overlap center.
If a basis of Gaussian functions is used, the product of two s functions is spherically symmetric and can be represented completely just by a point charge at the ‘overlap center’ of the two Gaussian functions. The product of an s orbital and a p orbital has only charge and dipole components, and the product of two p functions has charge, dipole and quadrupole components.
If the overlap center is not at an atom, one can move the origin of the multipole expansion to the nearest distributed multipole site, re-expressing the series to account for the change of origin. The multipole expansion will no longer terminate, |
https://en.wikipedia.org/wiki/Amiga%20Hombre%20chipset | Hombre is a RISC chipset for the Amiga, designed by Commodore, which was intended as the basis of a range of Amiga personal computers and multimedia products, including a successor to the Amiga 1200, a next generation game machine called CD64 and a 3D accelerator PCI card. Hombre was canceled along with the bankruptcy of Commodore International.
History
In 1993, Commodore International ceased the development of the AAA chipset when they concluded conventional PC clones would have similar performance shortly after the AAA machines would be released.
In the place of AAA, Commodore began to design a new 64-bit 3D graphics chipset based on Hewlett-Packard's PA-RISC architecture to serve as the new basis of the Amiga personal computer series. It was codenamed Hombre (pronounced "ómbre" which means man in Spanish) and was developed in conjunction with Hewlett-Packard over an estimated eighteen-month period.
Backward compatibility
Hombre does not support any planar mode, nor any emulation for the legacy Amiga chipset or Motorola 680x0 CPU registers, so it was completely incompatible with former Amiga models. According to Hombre designer Dr. Ed Hepler, Commodore intended to produce an AGA Amiga upon a single chip to solve the backward compatibility issues. This single chip would include Motorola MC680x0 core, plus the AGA chipset. The chip could be integrated in Hombre based computers for backward compatibility with AGA software.
Design
Hombre is based around two chips: Nathaniel, a System Controller chip, and Natalie, a Display Controller chip.
The System Controller chip was designed by Dr. Ed Hepler, well known as the designer of the AAA Andrea chip. The chip is similar in principle to the chip bus controller found in Agnus, Alice, and Andrea of the Amiga chipsets. Nathaniel features the following:
An inhouse designed 100+ MHz 64-bit integer PA-RISC microprocessor with SIMD and additional graphics processing related instructions
An advanced DMA engine and blitter |
https://en.wikipedia.org/wiki/TZSP | TaZmen Sniffer Protocol (TZSP) is an encapsulation protocol used to wrap other protocols. It is commonly used to wrap 802.11 wireless packets to support Intrusion Detection Systems (IDS), wireless tracking, or other wireless applications.
Protocol Summary
A number of 802.11 sensors and Access Points use the TZSP protocol for packet capture. It is an open protocol that was designed to encapsulate other protocols over UDP. The primary use for this protocol has been the capture of wireless traffic and transmission of them over a wired network.
Protocol Header
The protocol specified three parts to each TZSP packet: A 4-byte header followed by one or more tagged fields, the last of which has to be the TAG_END(0x01), and finally the encapsulated data.
The header and tagged fields use big-endian ordering where appropriate.
Version
The TZSP version should always be set to 1 (0x01).
Type
0 (0x00) Received tag list
1 (0x01) Packet for transmit
2 (0x02) Reserved
3 (0x03) Configuration
4 (0x04) Keepalive
5 (0x05) Port opener
Encapsulated Protocol
1 (0x01) Ethernet
18 (0x12) IEEE 802.11
119 (0x77) Prism Header
127 (0x7F) WLAN AVS
Tagged Fields
The tagged fields consist of a one-byte tag type, followed by a one-byte tag length, followed by a variable amount of data. The tag length does not include the tag type or tag length. All TZSP compatible decoders should skip unknown tagged fields.
<1 unsigned byte > Tag Type
<1 unsigned byte > Tag Length
<variable length > Dependent on Tag Length
There are two exceptions to this format. First, when a TAG_END (0x01) is received it is the end of all of the tagged fields. It DOES NOT have a Tag length nor any associated data. The second is the TAG_PADDING (0x00). This can be inserted at any point and should be ignored. It DOES NOT have a tag length nor any tag data.
The following are the valid TZSP tag types:
TAG_PADDING = 0 (0x00)
This special tagged field has neither tag length nor any tag data. The receiver |
https://en.wikipedia.org/wiki/Probabilistic%20bisimulation | In theoretical computer science, probabilistic bisimulation is an extension of the concept of bisimulation for fully probabilistic transition systems first described by K.G. Larsen and A. Skou.
A discrete probabilistic transition system is a triple
where gives the probability of starting in the state s, performing the action a and ending up in the state t. The set of states is assumed to be countable. There is no attempt to assign probabilities to actions. It is assumed that the actions are chosen nondeterministically by an adversary or by the environment. This type of system is fully probabilistic, there is no other indeterminacy.
The definition of a probabilistic bisimulation on a system S is an equivalence relation R on the state space St, such that for every pair s,t in St with sRt and for every action a in Act and for every equivalence class C of R
Two states are said to be probabilistically bisimilar if there is some such R relating them.
When applied to Markov chains, probabilistic bisimulation is the same concept as lumpability.
Probabilistic bisimulation extends naturally to weighted bisimulation. |
https://en.wikipedia.org/wiki/Molecular%20drive | Molecular drive is a term coined by Gabriel Dover in 1982 to describe evolutionary processes that change the genetic composition of a population through DNA turnover mechanisms. Molecular drive operates independently of natural selection and genetic drift.
The best-known such process is the concerted evolution of genes present in many tandem copies, such as those for ribosomal RNAs or silk moth egg shell chorion proteins, in sexually reproducing species. The concept has been proposed to extend to the diversification of multigene families. The mechanisms involved include gene conversion, unequal crossing-over, transposition, slippage replication and RNA-mediated exchanges. Because mutations changing the sequence of one copy are less common than deletions, duplications and replacement of one copy by another, the copies gradually come to resemble each other much more than they would if they had been evolving independently.
Concerted evolution can be unbiased, in which case every version has an equal probability of being the one that replaces the others. However, if the molecular events have any bias favouring one version of the sequence over others, that version will dominate the process and eventually replace the others. The name 'molecular drive' reflects the similarity of the process with what was originally the better-known process of meiotic drive.
Molecular drive can also act in bacteria, where parasexual processes such as natural transformation cause DNA turnover.
TRAM
According to Dover, TRAM is a genetic system that has features of non-mendelian inheritance Turnover, copy number and functional Redundancy And Modulatory. To date all regulatory regions (promoters) and genes that have been examined in detail at the molecular level, have TRAM characteristics. As such, part of their evolutionary history will have been influenced by the molecular drive process.
Adoptation
According to Dover, Adoptation is an evolved feature of an organism that contributes t |
https://en.wikipedia.org/wiki/Spatiotemporal%20Epidemiological%20Modeler | The Spatiotemporal Epidemiological Modeler (STEM) is free software available through the Eclipse Foundation. Originally developed by IBM Research, STEM is a framework and development tool designed to help scientists create and use spatial and temporal models of infectious disease. STEM uses a component software architecture based on the OSGi standard. The Eclipse Equinox platform is a reference implementation of that standard. By using a component software architecture, all of the components or elements required for a disease model, including the code and the data are available as software building blocks that can be independently exchanged, extended, reused, or replaced. These building blocks or plug-ins are called eclipse "plug-ins" or "extensions". STEM plug-ins contain denominator data for administrative regions of interest. The regions are indexed by standard (ISO3166) codes.
STEM currently includes a large number of plug-ins for the 244 countries and dependent areas defined by the Geographic Coding Standard maintained by the International Organization for Standardization. These plug-ins contain global data including geographic data, population data, demographics, and basic models of disease. The disease models distributed with STEM include epidemiological compartment models. Other plug-ins describe relationships between regions including nearest-neighbor or adjacency relationships as well as information about transportation, such as connections by roads and a model of air transportation.
Relationships between regions can then be included in models of how a disease spreads from place to place. To accomplish this, STEM represents the world as a "graph". The nodes in the graph correspond to places or regions, and the edges in the graph describe relationships or connections between regions. Both the nodes and the edges can be labeled or "decorated" with a variety of denominator data and models. This graphical representation is implemented using the Eclipse Mo |
https://en.wikipedia.org/wiki/Randle%20cycle | The Randle cycle, also known as the glucose fatty-acid cycle, is a metabolic process involving the competition of glucose and fatty acids for substrates. It is theorized to play a role in explaining type 2 diabetes and insulin resistance.
It was named for Philip Randle, who described it in 1963.
Cycle
The Randle cycle is a biochemical mechanism involving the competition between glucose and fatty acids for their oxidation and uptake in muscle and adipose tissue. The cycle controls fuel selection and adapts the substrate supply and demand in normal tissues. This cycle adds a nutrient-mediated fine tuning on top of the more coarse hormonal control on fuel metabolism. This adaptation to nutrient availability applies to the interaction between adipose tissue and muscle. Hormones that control adipose tissue lipolysis affect circulating concentrations of fatty acids, these in turn control the fuel selection in muscle. Mechanisms involved in the Randle Cycle include allosteric control, reversible phosphorylation and the expression of key enzymes. The energy balance from meals composed of differing macronutrient composition is identical, but the glucose and fat balances that contribute to the overall energy balance change reciprocally with meal composition.
Glucose is spared and rerouted
Fasted state
When fasting, the activation of lipolysis provides fatty acids as the preferred fuel source for respiration. In the liver β-oxidation of fatty acids fulfills the local energy needs and may lead to ketogenesis (creating ketone bodies out of fatty acids.) The ketone bodies are then used to meet the demands of tissues other than the liver. This inhibition of glucose oxidation at the level of pyruvate dehydrogenase preserves pyruvate and lactate, both of which are gluconeogenic precursors.
Fed state
The glucose fatty acid cycle is also observed in the fed state after a high-fat meal or during exercise. This is when plasma concentrations of fatty acids or ketone bodies are inc |
https://en.wikipedia.org/wiki/OurGrid | OurGrid is an opensource grid middleware based on a peer-to-peer architecture. OurGrid was mainly developed at the Federal University of Campina Grande (Brazil), which has run an OurGrid instance named "OurGrid" since December 2004. Anyone can freely join it to gain access to large amount of computational power and run parallel applications. This computational power is provided by the idle resources of all participants, and is shared in a way that makes those who contribute more get more when they need. Currently, the platform can be used to run any application whose tasks (i.e. parts that run on a single machine) do not communicate among themselves during execution, like most simulations, data mining and searching. |
https://en.wikipedia.org/wiki/Primes%20in%20arithmetic%20progression | In number theory, primes in arithmetic progression are any sequence of at least three prime numbers that are consecutive terms in an arithmetic progression. An example is the sequence of primes (3, 7, 11), which is given by for .
According to the Green–Tao theorem, there exist arbitrarily long sequences of primes in arithmetic progression. Sometimes the phrase may also be used about primes which belong to an arithmetic progression which also contains composite numbers. For example, it can be used about primes in an arithmetic progression of the form , where a and b are coprime which according to Dirichlet's theorem on arithmetic progressions contains infinitely many primes, along with infinitely many composites.
For integer k ≥ 3, an AP-k (also called PAP-k) is any sequence of k primes in arithmetic progression. An AP-k can be written as k primes of the form a·n + b, for fixed integers a (called the common difference) and b, and k consecutive integer values of n. An AP-k is usually expressed with n = 0 to k − 1. This can always be achieved by defining b to be the first prime in the arithmetic progression.
Properties
Any given arithmetic progression of primes has a finite length. In 2004, Ben J. Green and Terence Tao settled an old conjecture by proving the Green–Tao theorem: The primes contain arbitrarily long arithmetic progressions. It follows immediately that there are infinitely many AP-k for any k.
If an AP-k does not begin with the prime k, then the common difference is a multiple of the primorial k# = 2·3·5·...·j, where j is the largest prime ≤ k.
Proof: Let the AP-k be a·n + b for k consecutive values of n. If a prime p does not divide a, then modular arithmetic says that p will divide every pth term of the arithmetic progression. (From H.J. Weber, Cor.10 in ``Exceptional Prime Number Twins, Triplets and Multiplets," arXiv:1102.3075[math.NT]. See also Theor.2.3 in ``Regularities of Twin, Triplet and Multiplet Prime Numbers," arXiv:1103.0447[math.NT], G |
https://en.wikipedia.org/wiki/Cyclin-dependent%20kinase%20inhibitor%20protein | A cyclin-dependent kinase inhibitor protein (also known as CKIs, CDIs, or CDKIs) is a protein which inhibits the enzyme cyclin-dependent kinase (CDK) and Cyclin activity by stopping the cell cycle if there are unfavorable conditions, therefore, acting as tumor suppressors. Cell cycle progression is stopped by Cyclin-dependent kinase inhibitor protein at the G1 phase. CKIs are vital proteins within the control system that point out whether the process of DNA synthesis, mitosis, and cytokines control one another. If a malfunction prevents the successful completion of DNA synthesis during the G1 phase, a signal is sent to delay or stop the progression to the S phase. Cyclin-dependent kinase inhibitor proteins are essential in the regulation of the cell cycle. If cell mutations surpass the cell cycle checkpoints during cell cycle regulation, it can result in various types of cancer.
CKI Inactivation Process
Cyclin-dependent kinase inhibitor proteins work by inactivating the CDKs by degradation. The typical inactivation mechanism of the CDK/ Cyclin complex is based on binding a CDK inhibitor to the CDK cyclin complex and a partial conformational rotation of the CDK. The cyclin is thus forced to release the T loop and detach from the CDK. Then, the CDK inhibitor initiates a small Helix into the cleft blocking the cleft and blocking the active site of the CDK. Eventually, it releases the ATP out of the aperture of the CDK and deactivates it. Cyclin-dependent kinase inhibitor proteins use ATP as a phosphate contributor to phosphorylate serine and threonine residues.
Human cells contain many different cyclins binding to different CDKs. CDKs and cyclins appear and activate at specific cell cycle phases. Seven cyclin-dependent kinase inhibitor proteins have been identified. They are p15, p16, p18, p19, p21, p27, and p57. These cyclin-dependent kinase inhibitor protein emerges only in their specific cell cycle phase. Each Cyclin/CDK complex are specific to the part of the |
https://en.wikipedia.org/wiki/Infinite%20alleles%20model | The infinite alleles model is a mathematical model for calculating genetic mutations. The Japanese geneticist Motoo Kimura and American geneticist James F. Crow (1964) introduced the infinite alleles model, an attempt to determine for a finite diploid population what proportion of loci would be homozygous. This was, in part, motivated by assertions by other geneticists that more than 50 percent of Drosophila loci were heterozygous, a claim they initially doubted. In order to answer this question they assumed first, that there were a large enough number of alleles so that any mutation would lead to a different allele (that is the probability of back mutation to the original allele would be low enough to be negligible); and second, that the mutations would result in a number of different outcomes from neutral to deleterious.
They determined that in the neutral case, the probability that an individual would be homozygous, F, was:
where u is the mutation rate, and Ne is the effective population size. The effective number of alleles n maintained in a population is defined as the inverse of the homozygosity, that is
which is a lower bound for the actual number of alleles in the population.
If the effective population is large, then a large number of alleles can be maintained. However, this result only holds for the neutral case, and is not necessarily true for the case when some alleles are subject to selection, i.e. more or less fit than others, for example when the fittest genotype is a heterozygote (a situation often referred to as overdominance or heterosis).
In the case of overdominance, because Mendel's second law (the law of segregation) necessarily results in the production of homozygotes (which are by definition in this case, less fit), this means that population will always harbor a number of less fit individuals, which leads to a decrease in the average fitness of the population. This is sometimes referred to as genetic load, in this case it is a sp |
https://en.wikipedia.org/wiki/Scattering%20from%20rough%20surfaces | Surface roughness scattering or interface roughness scattering is the elastic scattering of particles against a rough solid surface or imperfect interface between two different materials. This effect has been observed in clasical systems, such as microparticle scattering, as well as quantum systems, where it arises electronic devices, such as field effect transistors and quantum cascade lasers.
Classical description
In the classical mechanics framework, a rough surface, such as a machined metal surface, randomizses the probability distribution function governing the incoming particles, leading to net momentum loss of the particle flux.
Quantum description
In the quantum mechanical framework, this scattering is most noticeable in confined systems, in which the energies for charge carriers are determined by the locations of interfaces. An example of such a system is a quantum well, which may be constructed from a sandwich of different layers of semiconductor. Variations in the thickness of these layers therefore causes the energy of particles to be dependent on their in-plane location in the layer. Classification of the roughness at a given position, , is complex, but as in the classical models, it has been modelled as a Gaussian distribution by some researchers
This assumption may be formulated in terms of the ensemble average for some given characteristic height, , and correlation length, , such that
TYPES OF SCATTERING :
Selective Scattering : In selective Scattering scattering depends upon the wavelength of light.
Mie Scattering : In this size of the molecules is greater than the wavelength of light that results in non-uniform scaterring of light.
Electromagnetic Scattering : Electromagnetic waves are used in this form of Scattering .
Notes
Scattering |
https://en.wikipedia.org/wiki/Trillion%20Tree%20Campaign | The Trillion Tree Campaign is a project which aims to plant one trillion trees worldwide. It seeks to repopulate the world's trees and combat climate change as a nature-based solution. The project was launched at PlantAhead 2018 in Monaco by Plant-for-the-Planet.
In the fall of 2018, the project's official website was published in order to register, monitor, and donate trees to reforestation projects around the world.
The campaign is a continuation of the activities of the earlier Billion Tree Campaign, instigated by Wangari Maathai, who founded the Green Belt Movement in Africa in 1977.
, 164 restoration projects participate in the campaign and 13.96 billion (1.396% of the goal) trees have been planted worldwide.
History
Billion Tree Campaign
The Green Belt Movement began its activity in Africa in 1977, eventually planting more than 30 million trees.
The Billion Tree Campaign was inspired by Nobel Peace Prize laureate Wangari Maathai, founder of the Green Belt Movement. When an executive in the United States told Maathai their corporation was planning to plant a million trees, her response was: "That's great, but what we really need is to plant a billion trees."
The project was launched in 2006 by the United Nations Environment Programme (UNEP) under the patronage of Prince Albert II of Monaco and the World Agroforestry Centre-ICRAF as a response to the challenges of climate change, as well as to a wider array of sustainability challenges from water supply to biodiversity loss, and achieved the initial target of planting a billion trees in 2007. The billionth tree, commonly known as an African olive, was planted in Ethiopia in November 2007.
In 2008, the campaign's objective was raised to 7 billion trees, a goal which was surpassed three months before its target of the climate change conference that was held in Copenhagen, Denmark, in December 2009.
The 2-billionth tree took root as part of the United Nation's World Food Programme agroforestry initiative. The |
https://en.wikipedia.org/wiki/Messenger%20of%20Mathematics | The Messenger of Mathematics is a defunct British mathematics journal. The founding editor-in-chief was William Allen Whitworth with Charles Taylor and volumes 1–58 were published between 1872 and 1929. James Whitbread Lee Glaisher was the editor-in-chief after Whitworth. In the nineteenth century, foreign contributions represented 4.7% of all pages of mathematics in the journal.
History
The journal was originally titled Oxford, Cambridge and Dublin Messenger of Mathematics. It was supported by mathematics students and governed by a board of editors composed of members of the universities of Oxford, Cambridge and Dublin (the last being its sole constituent college, Trinity College Dublin). Volumes 1–5 were published between 1862 and 1871. It merged with The Quarterly Journal of Pure and Applied Mathematics to form the Quarterly Journal of Mathematics. |
https://en.wikipedia.org/wiki/ProSyst | ProSyst Software GmbH was founded in Cologne in 1997 as a company specializing in Java software and middleware. ProSyst's first commercial application was a Java EE application server. In 2000, the company sold this server technology and has since focused completely on OSGi solutions.
In 1999, ProSyst was among the first companies to join the OSGi Alliance and since then has made important contributions to the development of each release of OSGi specifications (Release 1–4). ProSyst is a member of the OSGi Alliance board of directors alongside IBM, Nokia, NTT, Siemens, Oracle Corporation, Samsung, Motorola and Telcordia. Additionally, members of ProSyst staff serve in several positions on the OSGi Alliance.
In recent years ProSyst set its focus exclusively on the development of OSGi related software such as Frameworks, Bundles, Remote Management Systems and OSGi tools for developers including a full SDK available for download. ProSyst's OSGi applications are used by SmartHome devices, mobile phone manufacturers, network equipment providers (in CPEs), white goods manufacturers, car manufacturers and in the eHealth market.
ProSyst employs more than 120 Java and OSGi experts and offers OSGi related training, support (SLAs), technical consulting and development services.
As a member, ProSyst contributes to OSGi, Eclipse, Java Community Process, Nokia Forum Pro and the CVTA Connected Vehicle Trade Association.
Prosyst was acquired by Bosch in February 2015, and was merged into Bosch Group's software and systems unit Bosch Software Innovations GmbH.
Notable products
Commercial off-the-shelf products around OSGi mBS
Reduced-size Java client from 1999 |
https://en.wikipedia.org/wiki/Dialyte%20lens | A dialyte lens (sometimes called a dialyt) is a compound lens design that corrects optical aberrations where the lens elements are widely air-spaced. The design is used to save on the amount of glass used for specific elements or where elements can not be cemented because they have dissimilar curvatures. The word dialyte means "parted", "loose" or "separated".
Design
In its simplest form, a dialyte can be formed by separating the elements in a cemented achromatic doublet of positive and negative lenses, although the powers of the individual elements must be increased to compensate.
Applications
Telescopes
The idea of widely separating the color correcting elements of a lens dates back to W. F. Hamilton's 1814 catadioptric Hamiltonian telescope and Alexander Rogers' 1828 proposals for a dialytic refractor. The goal was to combine a large crown glass objective with a much smaller flint glass downstream to make an achromatic lens since flint glass at that time was very expensive. Dialyte designs were also used in the Schupmann medial telescope designed by in German optician Ludwig Schupmann near the end of the 19th century, in John Wall's 1999 "Zerochromat" retrofocally corrected dialytic refractor and the Russian made "TAL Apolar125" telescope which uses 6 elements arranged in three widely separated groups.
Photography
There are many types of dialyte camera lenses. One popular design is perfectly symmetric, which provides good correction for many aberrations. This consists of two air-spaced achromatic doublets arranged back-to-back around a central stop, or four air spaced lens elements in total: the outer pair is biconvex and the inner pair is biconcave; one example is the Celor. The Swiss mathematician Emil von Höegh, who had designed the popular Dagor anastigmat lens for Goerz in 1892, continued to refine that design, resulting in the Goerz Dagor Type B lens of 1899, later renamed to Celor and Syntor.
The Aviar lens (Taylor Hobson) designed by Arthur Warmish |
https://en.wikipedia.org/wiki/Mechanical%20rectifier | A mechanical rectifier is a device for converting alternating current (AC) to direct current (DC) by means of mechanically operated switches. The best-known type is the commutator, which is an integral part of a DC dynamo, but before solid-state devices became available, independent mechanical rectifiers were used for certain applications. Before the invention of semiconductors, rectification at high currents involved serious losses.
There were various vacuum/gas devices, such as the mercury arc rectifiers, thyratrons, ignitrons, and vacuum diodes. Solid-state technology was in its infancy, represented by copper oxide and selenium rectifiers. All of these gave excessive forward voltage drop at high currents. One answer was mechanically opening and closing contacts, if this could be done quickly and cleanly enough.
Vibrator type
This was the reverse of a vibrator inverter. An electromagnet, powered by DC through contacts it operated (like a buzzer) (or fed with AC), caused a spring to vibrate and the spring-operated change-over contacts which converted the AC to DC. This arrangement was only suitable for low-power applications, e.g. auto radios and was also found in some motorcycle electrical systems, where it was combined with a voltage regulator.
Motor-driven type
This operated on the same principle as the vibrator type but the change-over contacts were operated by a synchronous motor. It was suitable for high-power applications, e.g. electrolysis cells and electrostatic precipitators.
Still rectifier
A mechanical rectifier was patented in 1895 (US patent 547043) by William Joseph Still. The details are obscure but it appears from the diagram to be similar to a third-brush dynamo.
BTH rectifier
The machine shown in the reference was designed by Read and Gimson et al., at British Thomson-Houston (BTH) Rugby, Warwickshire, England, in the early 1950s. It is a three-phase mechanical rectifier working at 220 volts and 15,000 amperes, and its application wa |
https://en.wikipedia.org/wiki/Cheese | Cheese is a dairy product produced in wide ranges of flavors, textures, and forms by coagulation of the milk protein casein. It comprises proteins and fat from milk (usually the milk of cows, buffalo, goats, or sheep). During production, milk is usually acidified and either the enzymes of rennet or bacterial enzymes with similar activity are added to cause the casein to coagulate. The solid curds are then separated from the liquid whey and pressed into finished cheese. Some cheeses have aromatic molds on the rind, the outer layer, or throughout.
Over a thousand types of cheese exist and are produced in various countries. Their styles, textures and flavors depend on the origin of the milk (including the animal's diet), whether they have been pasteurized, the butterfat content, the bacteria and mold, the processing, and how long they have been aged. Herbs, spices, or wood smoke may be used as flavoring agents. The yellow to red color of many cheeses is produced by adding annatto. Other ingredients may be added to some cheeses, such as black pepper, garlic, chives, or cranberries. A cheesemonger, or specialist seller of cheeses, may have expertise with selecting, purchasing, receiving, storing and ripening cheeses.
For a few cheeses, the milk is curdled by adding acids such as vinegar or lemon juice. Most cheeses are acidified to a lesser degree by bacteria, which turn milk sugars into lactic acid, then the addition of rennet completes the curdling. Vegetarian alternatives to rennet are available; most are produced by fermentation of the fungus Mucor miehei, but others have been extracted from various species of the Cynara thistle family. Cheesemakers near a dairy region may benefit from fresher, lower-priced milk, and lower shipping costs.
Cheese is valued for its portability, long shelf life, and high content of fat, protein, calcium, and phosphorus. Cheese is more compact and has a longer shelf life than milk, although how long a cheese will keep depends on the t |
https://en.wikipedia.org/wiki/Binary%20constraint | A binary constraint, in mathematical optimization, is a constraint that involves exactly two variables.
For example, consider the n-queens problem, where the goal is to place n chess queens on an n-by-n chessboard such that none of the queens can attack each other (horizontally, vertically, or diagonally). The formal set of constraints are therefore "Queen 1 can't attack Queen 2", "Queen 1 can't attack Queen 3", and so on between all pairs of queens. Each constraint in this problem is binary, in that it only considers the placement of two individual queens.
Linear programs in which all constraints are binary can be solved in strongly polynomial time, a result that is not known to be true for more general linear programs. |
https://en.wikipedia.org/wiki/Girder%20and%20Panel%20building%20sets | Girder and Panel Building Sets were a series of plastic toy construction kits created by Kenner Toys in the mid-1950s. Since then, the building sets have gone in and out of production several times, under a succession of different owners of the designs.
Overview
The Girder and Panel Building Set construction kits enabled a child to build plastic models of mid-twentieth century style buildings. Vertical plastic columns were placed in the holes of a Masonite base board and horizontal girders were then locked into the vertical columns to create the skeletal structure of a model building. Brightly coloured plastic panels containing translucent "windows" could then be snapped onto the outer girders to create a curtain wall. Square navy-blue roof panels—some with translucent skylight domes molded into them—were laid on the topmost beams to complete the structure.
Bridge and Turnpike sets were later introduced that also employed frameworks of girders but with roadway sections instead of curtain wall panels and the addition of truss bracing and other techniques to construct models of various types of bridges, turnpikes, and interchanges. Still later, Kenner introduced sets with plastic-cased battery-operated motors that could be used to construct buildings with elevators, drawbridges that opened and closed, and other motorized structures.
The Girder and Panel construction style emulated twentieth century construction techniques such as curtain walls of prefabricated panels attached to frameworks of girders, trusses, and cantilevers. Girder and Panel toy sets were an important toy in the transition from the metal-based Gilbert Erector Sets of the 1920-to-1950 era to the plastic toys of the modern age. While Lego is arguably the most popular contemporary construction toy, no other toy has replaced Girder and Panel as a direct reflection of modern building techniques.
Girder and Panel products have been produced by several companies since 1957, and from 2005–2016 they wer |
https://en.wikipedia.org/wiki/John%20Riordan%20%28mathematician%29 | John Francis Riordan (April 22, 1903 – August 27, 1988) was an American mathematician and the author of major early works in combinatorics, particularly Introduction to Combinatorial Analysis and Combinatorial Identities.
Biography
Riordan was a graduate of Yale University. In his early life he wrote a number of poems and essays and a book of short-stories, On the Make, published in 1929, and was Editor-in-Chief of Salient and The Figure in the Carpet, literary magazines published by The New School for Social Research in New York. He married Mavis McIntosh, the well-known poet and literary agent and founder of McIntosh & Otis. The couple had two daughters: Sheila Riordan and Kathleen Riordan Speeth, and were long time residents of Hastings-on-Hudson, New York.
Riordan's long professional career was at Bell Labs, which he joined in 1926 (a year after its foundation) and where he remained, publishing over a hundred scholarly papers on combinatorial analysis, until he retired in 1968. He then joined the faculty at Rockefeller University as professor emeritus. A Festschrift was published in his honor in 1978.
Throughout his life Riordan led an active literary life, with many distinguished friends such as Kenneth Burke, William Carlos Williams, and A. R. Orage.
The Riordan array, created by mathematician Louis W. Shapiro, is named after John Riordan.
Tribute
From the Introduction by Marc Kac to the Special Issue of the JCTA in honor of John Riordan:
Foremost among the keepers of the barely flickering combinatorial flame was John Riordan. John’s work in Combinatorial Theory (or Combinatorial Analysis as he prefers to call it) is uncompromisingly classical in spirit and appearance. Though largely tolerant of modernity he does not let anyone forget that Combinatorial Analysis is the art and science of counting (enumerating is the word he prefers) and that a generating function by any other name or definition is still a generating function.
From an interview wi |
https://en.wikipedia.org/wiki/Millioctave | The millioctave (moct) is a unit of measurement for musical intervals. As is expected from the prefix milli-, a millioctave is defined as 1/1000 of an octave. From this it follows that one millioctave is equal to the ratio 21/1000, the 1000th root of 2, or approximately 1.0006934 ().
Given two frequencies a and b, the measurement of the interval between them in millioctaves can be calculated by
Likewise, if you know a note b and the number n of millioctaves in the interval, then the other note a may be calculated by:
Like the more common cent, the millioctave is a linear measure of intervals, and thus the size of intervals can be calculated by adding their millioctave values, instead of multiplication, which is necessary for calculations of frequencies.
A millioctave is exactly 1.2 cents.
History and use
The millioctave was introduced by the German physicist Arthur von Oettingen in his book Das duale Harmoniesystem (1913). The invention goes back to John Herschel, who proposed a division of the octave into 1000 parts, which was published (with appropriate credit to Herschel) in George Biddell Airy's book on musical acoustics.
Compared to the cent, the millioctave has not been as popular because it is not aligned with just intervals. It is however occasionally used by authors who wish to avoid the close association between the cent and twelve-tone equal temperament. Some considers that the millioctave introduces as well a bias for the less familiar 10-tone equal temperament however this bias is common in the decimal system.
See also
Cent (music)
Savart
Musical tuning
Logarithm
Degree (angle)
Chiliagon
Notes
External links
Logarithmic Interval Measures
Equal temperaments
Intervals (music)
Units of measurement
1913 introductions
1000 (number) |
https://en.wikipedia.org/wiki/Kn%C3%B6del%20number | In number theory, an n-Knödel number for a given positive integer n is a composite number m with the property that each i < m coprime to m satisfies . The concept is named after Walter Knödel.
The set of all n-Knödel numbers is denoted Kn.
The special case K1 is the Carmichael numbers. There are infinitely many n-Knödel numbers for a given n.
Due to Euler's theorem every composite number m is an n-Knödel number for where is Euler's totient function.
Examples |
https://en.wikipedia.org/wiki/Harold%20P.%20Boas | Harold P. Boas (born June 26, 1954) is an American mathematician.
Life
Boas was born in Evanston, Illinois, United States. He is the son of two noted mathematicians, Ralph P. Boas, Jr and Mary L. Boas.
Education
He received his A.B. and S.M. degrees in applied mathematics from Harvard University in 1976 and his Ph.D. in mathematics from the Massachusetts Institute of Technology in 1980 under the direction of Norberto Kerzman.
Teaching
Boas was a J. F. Ritt Assistant Professor at Columbia University (1980–1984) before moving to Texas A&M University, where he advanced to the rank of associate professor in 1987 and full professor in 1992. He has held visiting positions at the University of North Carolina at Chapel Hill and the Mathematical Sciences Research Institute in Berkeley, California.
Publications and awards
He has published over thirty papers, including Reflections on the arbelos (for which he won the Chauvenet Prize in 2009), and has also translated several dozen papers and a book from Russian into English. He is a winner of the Lester R. Ford Award (2007) of the Mathematical Association of America and a co-winner of the Stefan Bergman Prize (with Emil J. Straube, 1995) of the American Mathematical Society. In 2012 he became a fellow of the American Mathematical Society.
He revised and updated his father's book A Primer of Real Functions for the fourth edition. |
https://en.wikipedia.org/wiki/Hybrid%20cryptosystem | In cryptography, a hybrid cryptosystem is one which combines the convenience of a public-key cryptosystem with the efficiency of a symmetric-key cryptosystem. Public-key cryptosystems are convenient in that they do not require the sender and receiver to share a common secret in order to communicate securely. However, they often rely on complicated mathematical computations and are thus generally much more inefficient than comparable symmetric-key cryptosystems. In many applications, the high cost of encrypting long messages in a public-key cryptosystem can be prohibitive. This is addressed by hybrid systems by using a combination of both.
A hybrid cryptosystem can be constructed using any two separate cryptosystems:
a key encapsulation mechanism, which is a public-key cryptosystem
a data encapsulation scheme, which is a symmetric-key cryptosystem
The hybrid cryptosystem is itself a public-key system, whose public and private keys are the same as in the key encapsulation scheme.
Note that for very long messages the bulk of the work in encryption/decryption is done by the more efficient symmetric-key scheme, while the inefficient public-key scheme is used only to encrypt/decrypt a short key value.
All practical implementations of public key cryptography today employ the use of a hybrid system. Examples include the TLS protocol and the SSH protocol, that use a public-key mechanism for key exchange (such as Diffie-Hellman) and a symmetric-key mechanism for data encapsulation (such as AES). The OpenPGP file format and the PKCS#7 file format are other examples.
Hybrid Public Key Encryption (HPKE, published as RFC 9180) is a modern standard for generic hybrid encryption. HPKE is used within multiple IETF protocols, including MLS and TLS Encrypted Hello.
Envelope encryption is an example of a usage of hybrid cryptosystems in cloud computing. In a cloud context, hybrid cryptosystems also enable centralized key management.
Example
To encrypt a message addressed to |
https://en.wikipedia.org/wiki/Yamabe%20problem | The Yamabe problem refers to a conjecture in the mathematical field of differential geometry, which was resolved in the 1980s. It is a statement about the scalar curvature of Riemannian manifolds:
By computing a formula for how the scalar curvature of relates to that of , this statement can be rephrased in the following form:
The mathematician Hidehiko Yamabe, in the paper , gave the above statements as theorems and provided a proof; however, discovered an error in his proof. The problem of understanding whether the above statements are true or false became known as the Yamabe problem. The combined work of Yamabe, Trudinger, Thierry Aubin, and Richard Schoen provided an affirmative resolution to the problem in 1984.
It is now regarded as a classic problem in geometric analysis, with the proof requiring new methods in the fields of differential geometry and partial differential equations. A decisive point in Schoen's ultimate resolution of the problem was an application of the positive energy theorem of general relativity, which is a purely differential-geometric mathematical theorem first proved (in a provisional setting) in 1979 by Schoen and Shing-Tung Yau.
There has been more recent work due to Simon Brendle, Marcus Khuri, Fernando Codá Marques, and Schoen, dealing with the collection of all positive and smooth functions such that, for a given Riemannian manifold , the metric has constant scalar curvature. Additionally, the Yamabe problem as posed in similar settings, such as for complete noncompact Riemannian manifolds, is not yet fully understood.
The Yamabe problem in special cases
Here, we refer to a "solution of the Yamabe problem" on a Riemannian manifold as a Riemannian metric on for which there is a positive smooth function with
On a closed Einstein manifold
Let be a smooth Riemannian manifold. Consider a positive smooth function so that is an arbitrary element of the smooth conformal class of A standard computation shows
Taking the -i |
https://en.wikipedia.org/wiki/NameBase | NameBase is a web-based cross-indexed database of names that focuses on individuals involved in the international intelligence community, U.S. foreign policy, crime, and business. The focus is on the post-World War II era and on left of center, conspiracy theory, and espionage activities up to 2008.
Overview
Founder Daniel Brandt began collecting clippings and citations pertaining to influential people and intelligence agents in the 1960s. He did so especially in the 1970s after becoming a member of Students for a Democratic Society, an organization that opposed US foreign policy. With the advent of personal computing, he developed a database which allowed subscribers to access the names of US intelligence agents.
In the 1980s, through his company Micro Associates, he sold subscriptions to this computerized database under its original name, Public Information Research Inc. (PIR). At PIR's onset, Brandt was President of the newly formed non-profit corporation, and investigative researcher Peggy Adler served as its Vice President. The material was described as "information on all sorts of spooks, military officials, political operators and other cloak-and-dagger types". He told The New York Times at the time that "many of these sources are fairly obscure so it's a very effective way to retrieve information on U.S. intelligence that no one else indexes." One research librarian calls it "a unique part of the 'Deep Web'", equally useful to investigative journalists and students.
By 1992, private citizens, news organizations, and universities were all using NameBase. With the advent of public access to the Internet and World Wide Web in the 1990s these efforts became the basis of the NameBase website starting in 1995. , the database contained "over 100,000 names with over 260,000 citations drawn from books and serials with a few documents obtained under the Freedom of Information Act." The website utilizes hyperlinks to allow users to both visualize relationships in |
https://en.wikipedia.org/wiki/Kontron | Kontron AG is a German-based multinational company which designs and manufactures embedded computer modules, boards and systems.
Kontron AG serves original equipment manufacturers, system integrators and application providers of different market segments (amongst others: Industrial automation, communications, transportation, energy, avionics, medical, infotainment, and military). Kontron develops, manufactures and sells its products worldwide. Kontron is a premier member of the Intel Embedded Alliance.
The corporate group is headquartered in Augsburg and consists of the Kontron Europe GmbH (Ismaning, Augsburg, Deggendorf, and Saarbrücken). Other locations are in San Diego, Fremont, California, Montreal, Plzeň, Toulon, Bangalore, Taipei, Tokyo and Beijing. Kontron acquired Dolch in 2005.
At a trade show called Embedded World in 2007, Kontron introduced a product called the UGM-M72 using a "Universal Graphics Module".
The card used the M72S from ATI Technologies, and was 84 x 95 mm.
A version 1.1 of UGM was published in July 2008,
and Advanced Micro Devices (AMD, the parent company of ATI) announced it would support the UGM in August 2008.
A web site promoting the module was co-sponsored by XGI Technology until the Great Recession in 2009.
In August 2017 Kontron was merged into Austria-based S&T Group.
In September 2023, it was announced Kontron had acquired the Bucharest-headquartered software producer, Altimate. The company specialises in urban and interurban mobility solutions involving urban traffic control, automated fare collection, tolling solutions, and traffic violations. |
https://en.wikipedia.org/wiki/Radial%20immunodiffusion | Radial immunodiffusion (RID), Mancini immunodiffusion or single radial immunodiffusion assay, is an immunodiffusion technique used in immunology to determine the quantity or concentration of an antigen in a sample.
Description
Preparation
A solution containing antibody is added to a heated medium such as agar or agarose dissolved in buffered normal saline. The molten medium is then poured onto a microscope slide or into an open container, such as a Petri dish, and allowed to cool and form a gel. A solution containing the antigen is then placed in a well that is punched into the gel. The slide or container is then covered, closed or placed in a humidity box to prevent evaporation.
The antigen diffuses radially into the medium, forming a circle of precipitin that marks the boundary between the antibody and the antigen. The diameter of the circle increases with time as the antigen diffuses into the medium, reacts with the antibody, and forms insoluble precipitin complexes. The antigen is quantitated by measuring the diameter of the precipitin circle and comparing it with the diameters of precipitin circles formed by known quantities or concentrations of the antigen.
Antigen-antibody complexes are small and soluble when in antigen excess. Therefore, precipitation near the center of the circle is usually less dense than it is near the circle's outer edge, where antigen is less concentrated.
Expansion of the circle reaches an endpoint and stops when free antigen is depleted and when antigen and antibody reach equivalence. However, the clarity and density of the circle's outer edge may continue to increase after the circle stops expanding.
Interpretation
For most antigens, the area and the square of the diameter of the circle at the circle's endpoint are directly proportional to the initial quantity of antigen and are inversely proportional to the concentration of antibody. Therefore, a graph that compares the quantities or concentrations of antigen in the origin |
https://en.wikipedia.org/wiki/Natural%20history%20of%20disease | The natural history of disease is the course a disease takes in individual people from its pathological onset ("inception") until its resolution (either through complete recovery or eventual death). The inception of a disease is not a firmly defined concept. The natural history of a disease is sometimes said to start at the moment of exposure to causal agents. Knowledge of the natural history of disease ranks alongside causal understanding in importance for disease prevention and control. Natural history of disease is one of the major elements of descriptive epidemiology.
As an example, the cartilage of the knee, trapeziometacarpal and other joints deteriorates with age in most humans (osteoarthritis). There are no disease-modifying treatments for osteoarthritis---no way to slow, arrest, or reverse this pathophysiological process. There are only palliative/symptomatic treatments such as analgesics and exercises. In contrast, consider rheumatoid arthritis, a systemic inflammatory disease that damages articular cartilage throughout the body. There are now treatments that can modify that auto-immune inflammatory process (immune modulating drugs) that can slow the progression of the disease. Because these medications can alter the natural history of disease, they are referred to as disease-modifying antirheumatic drug.
The subclinical (pre-symptomatic) and clinical (symptomatic) evolution of disease is the natural progression of a disease without any medical intervention. It constitutes the course of biological events that occurs during the development of a disease from its root causes (etiology) to its outcome, whether that be recovery, chronicity, or death.
In regards to the natural history of disease, the goal of the medical field is to discover all of the different phases and components of each pathological process in order to intervene as early as possible and change the course of the disease before it leads to the deterioration of a patient's health.
There are |
https://en.wikipedia.org/wiki/Georg%20Kreisel | Georg Kreisel FRS (September 15, 1923 – March 1, 2015) was an Austrian-born mathematical logician who studied and worked in the United Kingdom and America.
Biography
Kreisel was born in Graz and came from a Jewish background; his family sent him to the United Kingdom before the Anschluss in 1938. He studied mathematics at Trinity College, Cambridge, and then, during World War II, worked on military subjects. Kreisel never took a Ph.D., though much later, in 1962, he was awarded the Cambridge degree of Sc.D., a `higher doctorate' given on the basis of published research.
He taught at the University of Reading from 1949 until 1954 and then worked at the Institute for Advanced Study from 1955 to 1957. He returned to Reading in 1957, but then taught at Stanford University from 1958-1959. Then back at Reading for the year 1959-1960, and then the University of Paris 1960-1962. Kreisel was appointed a professor at Stanford University in 1962 and remained on the faculty there until he retired in 1985.
Kreisel worked in various areas of logic, and especially in proof theory, where he is known for his so-called "unwinding" program, whose aim was to extract constructive content from superficially non-constructive proofs.
Kreisel was elected to the Royal Society in 1966; Kreisel remained a close friend of Francis Crick whom he had met in the Royal Navy during WWII.
While a student at Cambridge, Kreisel was the student most respected by Ludwig Wittgenstein. Ray Monk writes, "In 1944--when Kreisel was still only twenty-one--Wittgenstein shocked Rush Rhees by declaring Kreisel to be the most able philosopher he had ever met who was also a mathematician."
Kreisel was also a close friend of the Anglo-Irish philosopher and novelist Iris Murdoch. They met at Cambridge in 1947 during Murdoch's year of study there. Peter Conradi reports that Murdoch transcribed Kreisel's letters into her journals over the next fifty years. According to Conradi, "For half a century she none |
https://en.wikipedia.org/wiki/BCODE | A bCODE is an identifier that can be sent to a mobile phone/device and used as a ticket/voucher/identification or other type of token. The bCODE is an SMS message that can be read electronically from the screen of a mobile device. Bcodes can be sent by text message, and as they are just a standard SMS they can be received on over 99% of all devices.
Bcodes have many uses such as advertising, loyalty programs, promotions, ticketing and more.
History
bCODE was developed by an Australian company from 2003 to 2005.
bCODE Technology
A bCODE is a simple SMS text message that looks something like this:
This text message is read from the screen of a mobile phone/device and decoded into a unique token ID. This ID can then be used to supply the consumer with their own unique experience. |
https://en.wikipedia.org/wiki/Ecological%20self | In environmental philosophy, ecological self is central to the school of Experiential Deep Ecology, which, based on the work of Norwegian philosopher Arne Næss, argues that through the process of self-actualisation, one transcends the notions of the individuated "egoic" self and arrives at a position of an ecological self. So long as one is working within the narrower concept of self, Næss argues, environmentally responsible behaviour is a form of altruism, a "doing good for the other", which historically has been a precarious ethical basis, usually involved in exhorting others to "be good". Næss argues that in his Ecosophy, the enlargement of the ego-self to the eco-self results in environmentally responsible behaviour as a form of self-interest.
Warwick Fox argued that Næss's philosophy was based upon a variety of "transpersonal ecology" in which self-interest was firmly embedded within the interest of the ecommunity ecosphere of which the self was eternally embedded
As deep ecologist John Seed has stated, "Deep ecology critiques the idea that we are the crown of creation, the measure of all being: that the world is a pyramid with humanity rightly on top, merely a resource, and that nature has instrumental value only". The concept of the Ecological Self goes beyond anthropocentrism, which, by contrast locates human concerns as the exclusive source of all value. It draws upon the Land Ethic of Aldo Leopold. Leopold argued that within conventional ethics, the land itself was considered only as property, occupying a role analogous to slavery in earlier societies that permitted the ownership of people. By comparison a land ethic enlarges the boundary of moral concern to include "soils, waters, plants, and animals, or collectively: the land". The basis of such a non-anthropocentric ethic, according to Leopold was that "A thing is right when it tends to preserve the integrity, stability, and beauty of the biotic community. It is wrong when it tends otherwise."
|
https://en.wikipedia.org/wiki/Welcome%20sign | A welcome sign (or gateway sign) is a road sign at the border of a jurisdiction or region that introduces or welcomes visitors to the city/county/state/province/prefecture/canton/region. Examples of welcome signs can be found near political borders, such as when entering a state, province, county, city, or town, and they are increasingly found in neighborhoods and private communities. In European countries under the Schengen Agreement, a welcome sign may be found at borders between countries. Its purpose is partly informational, to inform drivers where they are, and partly for tourism, as it affords an opportunity to advertise features within the region to people who are entering it. A welcome sign is a type of town sign—a sign placed at the entrance to and exit from a city, town, or village. In many jurisdictions, the format of town signs is standardized; in some, welcome signs may be distinct from the legally mandated town sign.
A municipality's welcome sign may give its population or date of foundation, list twinned towns or services within the town, or depict the town's crest, typical local products, or the logo of sponsor organizations which maintain the sign (such as the local Lions Club).
Gallery |
https://en.wikipedia.org/wiki/Wellcome%20Centre%20for%20Human%20Neuroimaging | The Wellcome Centre for Human Neuroimaging at University College London is a world-leading interdisciplinary centre for neuroimaging research based in London, United Kingdom. Researchers at the Centre use expertise to investigate how the human brain generates behaviour, thoughts and feelings and how to use this knowledge to help patients with neurological and psychiatric disorders. Human neuroimaging allows scientists to non-invasively investigate the brain structure and functions including Action, Decision Making, Emotion, Hearing, Language, Memory, Navigation, Seeing, Self awareness, Social Behaviour and the Bayesian Brain
The Wellcome Centre for Human Neuroimaging is part of UCL's Department for Imaging Neuroscience, alongside The Max Planck UCL Centre for Computational Psychiatry and Ageing and Affiliated Principal Investigators (PIs).
The current team of researchers and support staff use their diverse and interdisciplinary skills to work collaboratively towards one shared goal: to use neuroimaging to understand and help patients with neurological and psychiatric disorders. Principal investigators working at the Centre include Professors John Ashburner, Dominik Bach, Gareth Barnes, Sven Bestmann, Neil Burgess, Martina Callaghan, Jenny Crinion, Ray Dolan, Stephen Fleming, Karl Friston, Tim Griffiths, Alex Leff, Vladimir Litvak, Eleanor Maguire, Mairead MacSweeney, Tamar Makin, Cathy Price, Geraint Rees, Jon Roiser; and Drs Guillaume Flandin, Tobias Hauser, Quentin Huys, Peter Kok, Christian Lambert, Clare Press, Rimona Weil, Elliot Wimmer and Peter Zeidman.
The centre is located at 12 Queen Square in the Bloomsbury area of Central London, adjacent to the National Hospital for Neurology and Neurosurgery
History
Over 25 years, the Wellcome Centre for Human Neuroimaging has pioneered innovation and applications in imaging neuroscience, addressed fundamental biological questions, and played a leading role in transforming cognitive and systems neuroscie |
https://en.wikipedia.org/wiki/Concatenated%20error%20correction%20code | In coding theory, concatenated codes form a class of error-correcting codes that are derived by combining an inner code and an outer code. They were conceived in 1966 by Dave Forney as a solution to the problem of finding a code that has both exponentially decreasing error probability with increasing block length and polynomial-time decoding complexity.
Concatenated codes became widely used in space communications in the 1970s.
Background
The field of channel coding is concerned with sending a stream of data at the highest possible rate over a given communications channel, and then decoding the original data reliably at the receiver, using encoding and decoding algorithms that are feasible to implement in a given technology.
Shannon's channel coding theorem shows that over many common channels there exist channel coding schemes that are able to transmit data reliably at all rates less than a certain threshold , called the channel capacity of the given channel. In fact, the probability of decoding error can be made to decrease exponentially as the block length of the coding scheme goes to infinity. However, the complexity of a naive optimum decoding scheme that simply computes the likelihood of every possible transmitted codeword increases exponentially with , so such an optimum decoder rapidly becomes infeasible.
In his doctoral thesis, Dave Forney showed that concatenated codes could be used to achieve exponentially decreasing error probabilities at all data rates less than capacity, with decoding complexity that increases only polynomially with the code block length.
Description
Let Cin be a [n, k, d] code, that is, a block code of length n, dimension k, minimum Hamming distance d, and rate r = k/n, over an alphabet A:
Let Cout be a [N, K, D] code over an alphabet B with |B| = |A|k symbols:
The inner code Cin takes one of |A|k = |B| possible inputs, encodes into an n-tuple over A, transmits, and decodes into one of |B| possible outputs. We regard this as |
https://en.wikipedia.org/wiki/Nucleate%20boiling | In fluid thermodynamics, nucleate boiling is a type of boiling that takes place when the surface temperature is hotter than the saturated fluid temperature by a certain amount but where the heat flux is below the critical heat flux. For water, as shown in the graph below, nucleate boiling occurs when the surface temperature is higher than the saturation temperature () by between . The critical heat flux is the peak on the curve between nucleate boiling and transition boiling. The heat transfer from surface to liquid is greater than that in film boiling.
Nucleate boiling is common in electric kettles and is responsible for the noise that occurs before boiling occurs. It also occurs in water boilers where water is rapidly heated.
Mechanism
Two different regimes may be distinguished in the nucleate boiling range. When the temperature difference is between approximately above TS, isolated bubbles form at nucleation sites and separate from the surface. This separation induces considerable fluid mixing near the surface, substantially increasing the convective heat transfer coefficient and the heat flux. In this regime, most of the heat transfer is through direct transfer from the surface to the liquid in motion at the surface and not through the vapor bubbles rising from the surface.
Between above TS, a second flow regime may be observed. As more nucleation sites become active, increased bubble formation causes bubble interference and coalescence. In this region the vapor escapes as jets or columns which subsequently merge into slugs of vapor.
Interference between the densely populated bubbles inhibits the motion of liquid near the surface. This is observed on the graph as a change in the direction of the gradient of the curve or an inflection in the boiling curve. After this point, the heat transfer coefficient starts to reduce as the surface temperature is further increased although the product of the heat transfer coefficient and the temperature difference (t |
https://en.wikipedia.org/wiki/Microvesicle | Microvesicles (ectosomes, or microparticles) are a type of extracellular vesicle (EV) that are released from the cell membrane. In multicellular organisms, microvesicles and other EVs are found both in tissues (in the interstitial space between cells) and in many types of body fluids. Delimited by a phospholipid bilayer, microvesicles can be as small as the smallest EVs (30 nm in diameter) or as large as 1000 nm. They are considered to be larger, on average, than intracellularly-generated EVs known as exosomes. Microvesicles play a role in intercellular communication and can transport molecules such as mRNA, miRNA, and proteins between cells.
Though initially dismissed as cellular debris, microvesicles may reflect the antigenic content of the cell of origin and have a role in cell signaling. Like other EVs, they have been implicated in numerous physiologic processes, including anti-tumor effects, tumor immune suppression, metastasis, tumor-stroma interactions, angiogenesis, and tissue regeneration. Microvesicles may also remove misfolded proteins, cytotoxic agents and metabolic waste from the cell. Changes in microvesicle levels may indicate diseases including cancer.
Formation and contents
Different cells can release microvesicles from the plasma membrane. Sources of microvesicles include megakaryocytes, blood platelets, monocytes, neutrophils, tumor cells and placenta.
Platelets play an important role in maintaining hemostasis: they promote thrombus growth, and thus they prevent loss of blood. Moreover, they enhance immune response, since they express the molecule CD154 (CD40L). Platelets are activated by inflammation, infection, or injury, and after their activation microvesicles containing CD154 are released from platelets. CD154 is a crucial molecule in the development of T cell-dependent humoral immune response. CD154 knockout mice are incapable of producing IgG, IgE, or IgA as a response to antigens. Microvesicles can also transfer prions and molecules CD4 |
https://en.wikipedia.org/wiki/Journal%20of%20Chemometrics | The Journal of Chemometrics is a monthly peer-reviewed scientific journal published since 1987 by John Wiley & Sons. It publishes original scientific papers, reviews, and short communications on fundamental and applied aspects of chemometrics. The current editor-in-chief is Age K. Smilde (University of Amsterdam).
Abstracting and indexing
Journal of Chemometrics is abstracted and indexed in:
Chemical Abstracts Service
Scopus
Web of Science
According to the Journal Citation Reports, the journal has a 2020 impact factor of 2.467, ranking it 27th out of 64 journals in the category "Instruments & Instrumentation", 35th out of 63 journals in the category "Automation & Control Systems", 30th out of 125 journals in the category "Statistics & Probability", 40th out of 108 journals in the category "Mathematics Interdisciplinary Applications", and 54th out of 87 journals in the category "Chemistry Analytical",
Highest cited papers
Selectivity, local rank, three-way data analysis and ambiguity in multivariate curve resolution, Volume 9, Issue 1, Jan-Feb 1995, Pages: 31–58, Tauler R, Smilde A, Kowalski B. Cited 370 times.
Genetic algorithms as a strategy for feature-selection, Volume 6, Issue 5, Sep-Oct 1992, Pages: 267–281, Leardi R, Boggia R, Terrile M. Cited 296 times.
Multiway calibration. Multilinear PLS, Volume 10, Issue 1, Jan-Feb 1996, Pages: 47–61, Bro R. Cited 290 times. |
https://en.wikipedia.org/wiki/Bonnet%27s%20theorem | In classical mechanics, Bonnet's theorem states that if n different force fields each produce the same geometric orbit (say, an ellipse of given dimensions) albeit with different speeds v1, v2,...,vn at a given point P, then the same orbit will be followed if the speed at point P equals
History
This theorem was first derived by Adrien-Marie Legendre in 1817, but it is named after Pierre Ossian Bonnet.
Derivation
The shape of an orbit is determined only by the centripetal forces at each point of the orbit, which are the forces acting perpendicular to the orbit. By contrast, forces along the orbit change only the speed, but not the direction, of the velocity.
Let the instantaneous radius of curvature at a point P on the orbit be denoted as R. For the kth force field that produces that orbit, the force normal to the orbit Fk must provide the centripetal force
Adding all these forces together yields the equation
Hence, the combined force-field produces the same orbit if the speed at a point P is set equal to |
https://en.wikipedia.org/wiki/Penrose%20transform | In theoretical physics, the Penrose transform, introduced by , is a complex analogue of the Radon transform that relates massless fields on spacetime, or more precisely the space of solutions to massless field equations to sheaf cohomology groups on complex projective space. The projective space in question is the twistor space, a geometrical space naturally associated to the original spacetime, and the twistor transform is also geometrically natural in the sense of integral geometry. The Penrose transform is a major component of classical twistor theory.
Overview
Abstractly, the Penrose transform operates on a double fibration of a space Y, over two spaces X and Z
In the classical Penrose transform, Y is the spin bundle, X is a compactified and complexified form of Minkowski space (which as a complex manifold is ) and Z is the twistor space (which is ). More generally examples come from double fibrations of the form
where G is a complex semisimple Lie group and H1 and H2 are parabolic subgroups.
The Penrose transform operates in two stages. First, one pulls back the sheaf cohomology groups Hr(Z,F) to the sheaf cohomology Hr(Y,η−1F) on Y; in many cases where the Penrose transform is of interest, this pullback turns out to be an isomorphism. One then pushes the resulting cohomology classes down to X; that is, one investigates the direct image of a cohomology class by means of the Leray spectral sequence. The resulting direct image is then interpreted in terms of differential equations. In the case of the classical
Penrose transform, the resulting differential equations are precisely the massless field equations for a given spin.
Example
The classical example is given as follows
The "twistor space" Z is complex projective 3-space CP3, which is also the Grassmannian Gr1(C4) of lines in 4-dimensional complex space.
X = Gr2(C4), the Grassmannian of 2-planes in 4-dimensional complex space. This is a compactification of complex Minkowski space.
Y is the flag |
https://en.wikipedia.org/wiki/Rhodolith | Rhodoliths (from Greek for red rocks) are colorful, unattached calcareous nodules, composed of crustose, benthic marine red algae that resemble coral. Rhodolith beds create biogenic habitat for diverse benthic communities. The rhodolithic growth habit has been attained by a number of unrelated coralline red algae, organisms that deposit calcium carbonate within their cell walls to form hard structures or nodules that resemble beds of coral.
Rhodoliths do not attach themselves to the rocky seabed. Rather, they roll like tumbleweeds along the seafloor until they become too large in size to be mobilised by the prevailing wave and current regime. They may then become incorporated into a semi-continuous algal mat or form an algal build-up. While corals are animals that are both autotrophic (photosynthesize via their symbionts) or heterotrophic (feeding on plankton), rhodoliths produce energy solely through photosynthesis (i.e. they can only grow and survive in the photic zone of the ocean).
Scientists believe rhodoliths have been present in the world's oceans since at least the Eocene epoch, some 55 million years ago.
Overview
Rhodoliths (including maërl) have been defined as calcareous nodules composed of more than 50% of coralline red algal material and consisting of one to several coralline species growing together.
Habitat
Rhodolith beds have been found throughout the world's oceans, including in the Arctic near Greenland, in waters off British Columbia, Canada, the Gulf of California, Mexico, the Mediterranean as off New Zealand and eastern Australia. Globally, rhodoliths fill an important niche in the marine ecosystem, serving as a transition habitat between rocky areas and barren, sandy areas. Rhodoliths provide a stable and three-dimensional habitat onto and into which a wide variety of species can attach, including other algae, commercial species such as clams and scallops, and true corals. Rhodoliths are resilient to a variety of environmental disturbanc |
https://en.wikipedia.org/wiki/Intelligent%20vehicular%20ad%20hoc%20network |
Intelligent vehicular ad hoc networks (InVANETs) use WiFi IEEE 802.11p (WAVE standard) and effective communication between vehicles with dynamic mobility. Effective measures such as media communication between vehicles can be enabled as well methods to track automotive vehicles. InVANET is not foreseen to replace current mobile (cellular phone) communication standards.
"Older" designs within the IEEE 802.11 scope may refer just to IEEE 802.11b/g. More recent designs refer to the latest issues of IEEE 802.11p (WAVE, draft status). Due to inherent lag times, only the latter one in the IEEE 802.11 scope is capable of coping with the typical dynamics of vehicle operation.
Automotive vehicular information can be viewed on electronic maps using the Internet or specialized software. The advantage of WiFi based navigation system function is that it can effectively locate a vehicle which is inside big campuses like universities, airports, and tunnels.
InVANET can be used as part of automotive electronics, which has to identify an optimally minimal path for navigation with minimal traffic intensity. The system can also be used as a city guide to locate and identify landmarks in a new city.
Communication capabilities in vehicles are the basis of an envisioned InVANET or intelligent transportation systems (ITS). Vehicles are enabled to communicate among themselves (vehicle-to-vehicle, V2V) and via roadside access points (vehicle-to-roadside, V2R) also called as Road Side Units (RSUs). Vehicular communication is expected to contribute to safer and more efficient roads by providing timely information to drivers, and also to make travel more convenient. The integration of V2V and V2R communication is beneficial because V2R provides better service sparse networks and long-distance communication, whereas V2V enables direct communication for small to medium distances/areas and at locations where roadside access points are not available.
Providing vehicle–vehicle and vehicle–roa |
https://en.wikipedia.org/wiki/Curvature%20of%20a%20measure | In mathematics, the curvature of a measure defined on the Euclidean plane R2 is a quantification of how much the measure's "distribution of mass" is "curved". It is related to notions of curvature in geometry. In the form presented below, the concept was introduced in 1995 by the mathematician Mark S. Melnikov; accordingly, it may be referred to as the Melnikov curvature or Menger-Melnikov curvature. Melnikov and Verdera (1995) established a powerful connection between the curvature of measures and the Cauchy kernel.
Definition
Let μ be a Borel measure on the Euclidean plane R2. Given three (distinct) points x, y and z in R2, let R(x, y, z) be the radius of the Euclidean circle that joins all three of them, or +∞ if they are collinear. The Menger curvature c(x, y, z) is defined to be
with the natural convention that c(x, y, z) = 0 if x, y and z are collinear. It is also conventional to extend this definition by setting c(x, y, z) = 0 if any of the points x, y and z coincide. The Menger-Melnikov curvature c2(μ) of μ is defined to be
More generally, for α ≥ 0, define c2α(μ) by
One may also refer to the curvature of μ at a given point x:
in which case
Examples
The trivial measure has zero curvature.
A Dirac measure δa supported at any point a has zero curvature.
If μ is any measure whose support is contained within a Euclidean line L, then μ has zero curvature. For example, one-dimensional Lebesgue measure on any line (or line segment) has zero curvature.
The Lebesgue measure defined on all of R2 has infinite curvature.
If μ is the uniform one-dimensional Hausdorff measure on a circle Cr or radius r, then μ has curvature 1/r.
Relationship to the Cauchy kernel
In this section, R2 is thought of as the complex plane C. Melnikov and Verdera (1995) showed the precise relation of the boundedness of the Cauchy kernel to the curvature of measures. They proved that if there is some constant C0 such that
for all x in C and all r > 0, then there is another constan |
https://en.wikipedia.org/wiki/Journal%20of%20Experimental%20and%20Theoretical%20Physics | The Journal of Experimental and Theoretical Physics (JETP) [ (ЖЭТФ), or Zhurnal Éksperimental'noĭ i Teoreticheskoĭ Fiziki (ZhÉTF)] is a peer-reviewed Russian bilingual scientific journal covering all areas of experimental and theoretical physics. For example, coverage includes solid-state physics, elementary particles, and cosmology. The journal is published simultaneously in both Russian and English languages.
The editor-in-chief is Alexander F. Andreev. In addition, this journal is a continuation of Soviet physics, JETP (1931–1992), which began English translation in 1955.
Indexing
JETP is indexed in: |
https://en.wikipedia.org/wiki/Integrated%20Management%20of%20Childhood%20Illness | Integrated Management of Childhood Illness (IMCI) is a systematic approach to children's health which focuses on the whole child.This means focusing not only on curative care but also on prevention of disease. The approach was developed by United Nations Children's Fund and the World Health Organization in 1995. It includes both preventive components on the part of families and communities, as well as curative actions to be taken by health workers. It also has an objective to help improve health systems.
Inequalities of child health
Although the annual number of deaths among children less than five years old has decreased by almost a third since the 1970s, this reduction has not been evenly distributed throughout the world. According to the 1999 World Health Report, children in low- to middle-income countries are 10 times more likely to die before reaching age five than children living in the industrialized world. In 1998, more than 50 countries still had childhood mortality rates of over 100 per 1,000 live births.
Every year more than 10 million children in these countries die before they reach their fifth birthday. Seven in 10 of these deaths are due to acute respiratory infections (mostly pneumonia), diarrhea, measles, malaria, or malnutrition – and often to a combination of these conditions
Rationale for an evidence-based syndromic approach to case management
Many well-known prevention and treatment strategies have already proven effective for saving young lives. Childhood vaccinations have successfully reduced deaths due to measles. Oral rehydration therapy has contributed to a major reduction in diarrhea deaths. Effective antibiotics have saved millions of children with pneumonia. Prompt treatment of malaria has allowed more children to recover and lead healthy lives. Even modest improvements in breastfeeding practices have reduced childhood deaths. Many of these interventions have been summed up into 16 key family practices.
While each of these intervent |
https://en.wikipedia.org/wiki/Gauss%27s%20continued%20fraction | In complex analysis, Gauss's continued fraction is a particular class of continued fractions derived from hypergeometric functions. It was one of the first analytic continued fractions known to mathematics, and it can be used to represent several important elementary functions, as well as some of the more complicated transcendental functions.
History
Lambert published several examples of continued fractions in this form in 1768, and both Euler and Lagrange investigated similar constructions, but it was Carl Friedrich Gauss who utilized the algebra described in the next section to deduce the general form of this continued fraction, in 1813.
Although Gauss gave the form of this continued fraction, he did not give a proof of its convergence properties. Bernhard Riemann and L.W. Thomé obtained partial results, but the final word on the region in which this continued fraction converges was not given until 1901, by Edward Burr Van Vleck.
Derivation
Let be a sequence of analytic functions so that
for all , where each is a constant.
Then
Setting
So
Repeating this ad infinitum produces the continued fraction expression
In Gauss's continued fraction, the functions are hypergeometric functions of the form , , and , and the equations arise as identities between functions where the parameters differ by integer amounts. These identities can be proven in several ways, for example by expanding out the series and comparing coefficients, or by taking the derivative in several ways and eliminating it from the equations generated.
The series 0F1
The simplest case involves
Starting with the identity
we may take
giving
or
This expansion converges to the meromorphic function defined by the ratio of the two convergent series (provided, of course, that a is neither zero nor a negative integer).
The series 1F1
The next case involves
for which the two identities
are used alternately.
Let
etc.
This gives where , producing
or
Similarly
or
Since , setting a to 0 an |
https://en.wikipedia.org/wiki/Private%20server | Strictly, a private server is any machine or virtual machine used as a server that is privately administrated. Colloquially the term is nearly-exclusively used to refer to independently operated, unofficial servers for video games.
Private Servers (Broadly Defined)
As servers need adequate internet connection, power and can be noisy, they are often located in a colocation center. Servers are available on the market the same way as laptops or desktops are available and can be purchased by individual already pre-configured. Ordinary desktop computers are not suitable to house in colocation centers as servers have specific form factor that allows them to fit many into a standard rack. This group also includes custom-designed experimental servers, made by hobbyists Virtual server also offer high degree of freedom, superuser access and low-cost service.
Private Servers for Video Games
A private server is a reimplementation in online game servers, typically as clones of proprietary commercial software by a third party of the game community. The private server is often not made or sanctioned by the original company.
Private servers often host MMORPG genre games such as World of Warcraft, Runescape, and MapleStory. These servers can attempt to provide a "stock"/standard experience, or can modify gameplay to change the difficulty, pace of character progress, available content, or available controls over the gameplay environment.
Many private servers encourage players to donate funds for server upkeep. Sometimes these donations are rewarded with in-game benefits. Less commonly, private server operators might sell in-game benefits directly.
Rightsholding companies often attempt to and succeed in shutting down private server versions of their games. Between this and potential for operators to lose interest, private servers can be short-lived. However, there are also cases of games no longer being operated by the rightsholders, in which case a private server may have gr |
https://en.wikipedia.org/wiki/Andr%C3%A9%20Bloch%20%28mathematician%29 | André Bloch (20 November 1893 – 11 October 1948) was a French mathematician who is best remembered for his fundamental contribution to complex analysis.
Bloch killed three of his family members, for which he was institutionalized in a mental asylum for 31 years, during which all of his mathematical output was produced.
Early life
Bloch was born in 1893 in Besançon, France. According to one of his teachers, Georges Valiron, both André Bloch and his younger brother Georges were in the same class in October 1910. Valiron believed Georges to have the better talent, and due to lack of preparation, André finished last in the class. André was spared from failing the class by convincing Ernest Vessiot to give him an oral exam. The exam convinced Vessiot of Andre's talent and both André and Georges entered the École Polytechnique.
Both brothers served for a year in the military prior to World War I. Both André and Georges studied for only one year at the École Polytechnique before the outbreak of the war.
World War I
Upon the outbreak of World War I in 1914, André and Georges Bloch were drafted. André, as a second-lieutenant in the artillery, was assigned to the headquarters of General de Castelnau in Nancy.
Both Bloch brothers were injured: André suffered a fall from an observation post, while Georges sustained a head wound which cost him an eye. Georges was released from service and returned to the École Polytechnique on 7 October 1917. André, however, was allowed to convalesce but not released from duty.
Murder
On 17 November 1917, while on convalescent leave from service in World War I, Bloch killed his brother Georges and his aunt and uncle. Several conjectures about the motives for Bloch's crime exist among mathematicians. However, Cartan and Ferrand quote Henri Baruk, who was the medical head of the asylum where Bloch was confined. Bloch told Baruk that the murders were a eugenic act, in order to eliminate branches of his family affected by mental illn |
https://en.wikipedia.org/wiki/Weibull%20modulus | The Weibull modulus is a dimensionless parameter of the Weibull distribution which is used to describe variability in measured material strength of brittle materials.
For ceramics and other brittle materials, the maximum stress that a sample can be measured to withstand before failure may vary from specimen to specimen, even under identical testing conditions. This is related to the distribution of physical flaws present in the surface or body of the brittle specimen, since brittle failure processes originate at these weak points. When flaws are consistent and evenly distributed, samples will behave more uniformly than when flaws are clustered inconsistently. This must be taken into account when describing the strength of the material, so strength is best represented as a distribution of values rather than as one specific value. The Weibull modulus is a shape parameter for the Weibull distribution model which, in this case, maps the probability of failure of a component at varying stresses.
Consider strength measurements made on many small samples of a brittle ceramic material. If the measurements show little variation from sample to sample, the calculated Weibull modulus will be high and a single strength value would serve as a good description of the sample-to-sample performance. It may be concluded that its physical flaws, whether inherent to the material itself or resulting from the manufacturing process, are distributed uniformly throughout the material. If the measurements show high variation, the calculated Weibull modulus will be low; this reveals that flaws are clustered inconsistently and the measured strength will be generally weak and variable. Products made from components of low Weibull modulus will exhibit low reliability and their strengths will be broadly distributed.
Test procedures for determining the Weibull modulus are specified in DIN EN 843-5 and DIN 51 110-3.
A further method to determine the strength of brittle materials has been descri |
https://en.wikipedia.org/wiki/Dried%20shrimp | Dried shrimp are shrimp that have been sun-dried and shrunk to a thumbnail size. They are used in many East Asian, Southeast Asian and South Asian cuisines, imparting a unique umami taste. A handful of shrimp is generally used for dishes. The flavors of this ingredient are released when allowed to simmer.
Use
East Asia
In Chinese cuisine, dried shrimp are used quite frequently for their sweet and unique flavor that is very different from fresh shrimp. They have the coveted umami flavor (or so-called "fifth taste"). It is an ingredient in the Cantonese XO sauce. Dried shrimp are also used in Chinese (mostly Cantonese) soups and braised dishes. It is also featured in Cantonese cuisine, particularly in some dim sum dishes such as rolled and rice noodle roll and in zongzi. Despite the literal meaning of the name Chinese name xiā mǐ ("shrimp rice"), it has nothing to do with rice other than the fact that the shrimp are shrunk to a tiny size similar to grains of rice.
Dried shrimp are also used in Korean cuisine, where they are soaked briefly to reconstitute them, and are then stir-fried with seasonings—typically garlic, ginger, scallions, soy sauce, sugar, and hot peppers—and served as a side dish. It is called "mareunsaeu bokkeum" (hangul: 마른새우볶음) in Korean. They are also used in some Korean braised dishes (jorim) and used for making broth.
Southeast Asia
Dried shrimp are used in Vietnamese cuisine, where they are called tôm khô, and are used in soups, congee, fried rice, or as a topping on stirfries (Mì Xào) or savoury snack items. They are also commonly eaten as snacks.
The Malays living in Malaysia developed sambal udang kering, which uses dried shrimp. Dried shrimp is a staple ingredient in the cuisine of Malaysia, with it being a base to rempah, a spice paste that forms the body of many Malay curries.
In Indonesia dried shrimp is called ebi, the name was derived from either Chinese Hokkian dialects "hebi" means "shrimp rice" or Japanese word "ebi" means "s |
https://en.wikipedia.org/wiki/IBM%20PS/ValuePoint | The PS/ValuePoint (or just ValuePoint) personal computer was IBM's answer to the PC clone market, where the IBM PS/2 could not compete due to price and proprietary interfaces. Announced in October 1992 and withdrawn in July 1995, it was replaced by the IBM PC Series 300.
Specifications and history
These systems used standard Industry Standard Architecture (ISA) bus, SVGA graphics and IDE hard disks. Later models introduced VESA Local Bus and PCI.
Processors range from the 386SLC-25, 486SX-25, 486DX-33, and 486DX2-66 to the Pentium 60.
IBM PS/ValuePoints were shipped in the following form factors:
Space saving desktop introductory: IBM 6381 model #: /Si (3 expansion card slots & 3 drive bays)
Space saving desktop: IBM 6382 model #: /S (3 expansion card slots & 3 drive bays)
Desktop: IBM 6384 model #: /D (5 expansion card slots & 5 drive bays)
Mini Tower: IBM 8387 model #: /T (8 expansion slots & 6 drive bays)
Predecessor
The IBM PS/ValuePoint series was preceded by these series:
IBM PS/1
IBM PS/2
Internal concurrents
The IBM PS/ValuePoint series was sold concurrently with these series:
IBM Aptiva and official Aptiva clones by AMBRA Computer Corporation
Successor
The IBM PS/ValuePoint series was succeeded by these series:
IBM PC Series
Models
Main line
Performance line
Budget line
Monitor
The PS/ValuePoint was shipped with the following monitors:
PS/2 8511, color (S)VGA (shipped with 325T)
6312, color (S)VGA
6314, color (S)VGA
6317, color (S)VGA
6319, color (S)VGA
Timeline |
https://en.wikipedia.org/wiki/MEO%20%28telecommunication%20company%29 | MEO is a mobile and fixed telecommunications service and brand from Altice Portugal (formerly Portugal Telecom), managed by MEO - Serviços de Comunicações e Multimédia. The service was piloted in Lisbon in 2006 and was later extended to Porto and Castelo Branco.
History
MEO in its current form was founded in 2007 after the separation of PT Comunicações and PT Multimédia (later ZON Multimédia). While PT Multimédia employed coaxial cables, after separation, MEO started making use of copper cables. The television service supplied by MEO within the copper cable network is served on the ADSL line. Telecomunicações Móveis Nacionais (TMN), Portugal's first and largest mobile network operator, was later integrated into the MEO brand in 2014 after two of TMN's shareholders, Telefones de Lisboa e Porto (TLP) and Marconi Comunicações Internacionais (the Portuguese operations of the UK-based Marconi Company) were acquired by Portugal Telecom in 1994 and 2002 respectively.
The commercial launch of the ADSL2+ service took place in June 2007. The satellite service began in April 2008, using the Hispasat satellite, soon followed by the FTTH service. The ADSL2+ and FTTH offers reached across Portugal and included broadband Internet services (at up to 400Mbit/s) as well as a telephone service.
In May 2009, PT Comunicações announced, after digital terrestrial television (TDT) transmissions had started, that the triple play service was also available with fiber optic speeds can achieve 400 Mbit/s.
Another service that rely on TDT, MEO TDT, which is included in the 3G plates service that is captured through the mobile internet signals from TDT. This service included one High Definition (HD) channel and the five main Portuguese channels: RTP1, RTP2, SIC, TVI and ARTV. MEO TDT service also allows some of the advantages found on the ADSL and Fiber Optic service (pause, record...).
In July 2010, Portugal Telecom informed that MEO had surpassed 700 thousand clients.
In November 2011, |
https://en.wikipedia.org/wiki/Baum%E2%80%93Connes%20conjecture | In mathematics, specifically in operator K-theory, the Baum–Connes conjecture suggests a link between the K-theory of the reduced C*-algebra of a group and the K-homology of the classifying space of proper actions of that group. The conjecture sets up a correspondence between different areas of mathematics, with the K-homology of the classifying space being related to geometry, differential operator theory, and homotopy theory, while the K-theory of the group's reduced C*-algebra is a purely analytical object.
The conjecture, if true, would have some older famous conjectures as consequences. For instance, the surjectivity part implies the Kadison–Kaplansky conjecture for discrete torsion-free groups, and the injectivity is closely related to the Novikov conjecture.
The conjecture is also closely related to index theory, as the assembly map is a sort of index, and it plays a major role in Alain Connes' noncommutative geometry program.
The origins of the conjecture go back to Fredholm theory, the Atiyah–Singer index theorem and the interplay of geometry with operator K-theory as expressed in the works of Brown, Douglas and Fillmore, among many other motivating subjects.
Formulation
Let Γ be a second countable locally compact group (for instance a countable discrete group). One can define a morphism
called the assembly map, from the equivariant K-homology with -compact supports of the classifying space of proper actions to the K-theory of the reduced C*-algebra of Γ. The subscript index * can be 0 or 1.
Paul Baum and Alain Connes introduced the following conjecture (1982) about this morphism:
Baum-Connes Conjecture. The assembly map is an isomorphism.
As the left hand side tends to be more easily accessible than the right hand side, because there are hardly any general structure theorems of the -algebra, one usually views the conjecture as an "explanation" of the right hand side.
The original formulation of the conjecture was somewhat different, as the not |
https://en.wikipedia.org/wiki/CBM-CFS3 | CBM-CFS3 (Carbon Budget Model of the Canadian Forest Sector) is a Windows-based software modelling framework for stand- and landscape-level forest ecosystem carbon accounting. It is used to calculate forest carbon stocks and stock changes for the past (monitoring) or into the future (projection). It can be used to create, simulate and compare various forest management scenarios in order to assess impacts on carbon. It is compliant with requirements under the Kyoto Protocol and with the Good Practice Guidance for Land Use, Land-Use Change and Forestry (2003) report published by the Intergovernmental Panel on Climate Change (IPCC).
It is the central model of the Government of Canada's National Forest Carbon Monitoring, Accounting and Reporting System (NFCMARS). The CBM-CFS3 was developed through a collaboration between Natural Resources Canada's Canadian Forest Service (CFS) and the Canadian Model Forest Network, and is currently supported by the CFS. The CBM-CFS3 is distributed at no charge by the Canadian Forest Service through Canada's National Forest Information System web site. Technical support is available by contacting Stephen Kull, Carbon Model Extension Forester, at the CFS.
See also
Carbon accounting
External links
Canadian Forest Service, Forest Carbon Accounting Web Site
Canadian Forest Service CBM-CFS3 Web Site
Natural Resources Canada Web Site
Canadian Forest Service Web Site
The Canadian Model Forest Network
Good Practice Guidance for Land Use, Land-Use Change and Forestry
Forest models
Climate change in Canada |
https://en.wikipedia.org/wiki/IBM%20PC%20Series | The Personal Computer Series, or PC Series, was IBM's follow-up to the Personal System/2 and PS/ValuePoint. Announced in October 1994 and withdrawn in October 2000, it was replaced by the IBM NetVista, apart from the Pentium Pro-based PC360 and PC365, which were replaced by the IBM IntelliStation.
Models
x86-based
PC 100
The PC 100 was a budget model, available only in selected markets.
PC 140
The PC 140 was a budget model, available only in selected markets.
PC Series 300
Industry standard ISA/PCI architecture, first IBM machines with USB. Processors ranged from the 486DX2-50, 486SX-25, 486DX4-100 to the Pentium 200 and in case of the Models 360 and 365 the Pentium Pro. 486 models had a selectable bus architecture (SelectaBus) through a replaceable riser-card, offering the choice of either VESA Local Bus/ISA or PCI/ISA.
Within the 300 series the following models appeared:
PC 330
Its last sub-model used the Pentium P54C processor clocked at 100, 133, 166, or 200 MHz. It had, depending on the sub-model, up to 4 ISA and/or 3 PCI expansion slots and four (2 external 5.25", 1 external and 1 internal 3.5") drive bays. It had in its latest version, the 6577, one DIMM-168 and 4 SIMM-72 memory slots, and featured an IBM SurePath BIOS. This PC has 2 USB 1.0 slots in the back. The latest version of Windows which can be installed on this PC is Windows XP, though Windows 2000 and Windows ME are optimal choices.
The DIMM-168 memory slot takes 5V EDO DRAM and is incompatible with the more commonly used 3.3V SDRAM. The slot looks the same at first glance, but the keying is different. Trying to force a 3.3V SDRAM module into the slot could damage both it and the memory module.
Submodels were:
PC 340
The PC 340, introduced in 1996, was a budget model. It used the Pentium processor clocked at 100, 133 or 166 MHz. It had 4 ISA and 3 PCI expansion slots and four (2 external 5.25 inch, 1 external and 1 internal 3.5 inch) drive bays. It had 4 SIMM-72 RAM slots, and featured an |
https://en.wikipedia.org/wiki/Milton%20Feng | Milton Feng co-created the first transistor laser, working with Nick Holonyak in 2004. The paper discussing their work was voted in 2006 as one of the five most important papers published by the American Institute of Physics since its founding 75 years ago. In addition to the invention of transistor laser, he is also well known for inventions of other "major breakthrough" devices, including the world's fastest transistor and light-emitting transistor (LET). As of May, 2009 he is a professor at the University of Illinois at Urbana–Champaign and holds the Nick Holonyak Jr. Endowed Chair Professorship.
Feng was born and raised in Taiwan.
Inventions
World's fastest transistor
In 2003, Milton Feng and his graduate students Walid Hafez and Jie-Wei Lai broke the record for the world's fastest transistor. Their device, made of indium phosphide and indium gallium arsenide with 25 nm thick base and 75 nm thick collector, marked a frequency of 509 GHz, which was 57 GHz faster than the previous record.
In 2005, they succeeded in fabricating a device at Micro and Nanotechnology Laboratory to break their own record, reaching 604 GHz.
In 2006, Feng and his other graduate student William Snodgrass fabricated an indium phosphide and indium gallium arsenide device with 12.5 nm thick base, operating at 765 GHz at room temperature and 845 GHz at -55 °C.
Light-emitting transistor
Reported in the January 5 issue of the journal Applied Physics Letters in 2004, Milton Feng and Nick Holonyak, the inventor of the first practical light-emitting diode (LED) and the first semiconductor laser to operate in the visible spectrum, made the world's first light-emitting transistor. This hybrid device, fabricated by Feng's graduate student Walid Hafez, had one electrical input and two outputs (electrical output and optical output) and operated at a frequency of 1 MHz. The device was made of indium gallium phosphide, indium gallium arsenide, and gallium arsenide, and emitted infrared photons fro |
https://en.wikipedia.org/wiki/Network%20allocation%20vector | The network allocation vector (NAV) is a virtual carrier-sensing mechanism used with wireless network protocols such as IEEE 802.11 (Wi-Fi) and IEEE 802.16 (WiMax). The virtual carrier-sensing is a logical abstraction which limits the need for physical carrier-sensing at the air interface in order to save power. The MAC layer frame headers contain a duration field that specifies the transmission time required for the frame, in which time the medium will be busy. The stations listening on the wireless medium read the Duration field and set their NAV, which is an indicator for a station on how long it must defer from accessing the medium.
The NAV may be thought of as a counter, which counts down to zero at a uniform rate. When the counter is zero, the virtual carrier-sensing indication is that the medium is idle; when nonzero, the indication is busy. The medium shall be determined to be busy when the station (STA) is transmitting. In IEEE 802.11, the NAV represents the number of microseconds the sending STA intends to hold the medium busy (maximum of 32,767 microseconds). When the sender sends a Request to Send the receiver waits one SIFS before sending Clear to Send. Then the sender will wait again one SIFS before sending all the data. Again the receiver will wait a SIFS before sending ACK. So NAV is the duration from the first SIFS to the ending of ACK. During this time the medium is considered busy.
Wireless stations are often battery-powered, so to conserve power the stations may enter a power-saving mode. A station decrements its NAV counter until it becomes zero, at which time it is awakened to sense the medium again.
The NAV virtual carrier sensing mechanism is a prominent part of the CSMA/CA MAC protocol used with IEEE 802.11 WLANs. NAV is used in DCF, PCF and HCF.
Media access control
Computer networking |
https://en.wikipedia.org/wiki/Ripe%20Digital%20Entertainment | Ripe Digital Entertainment (RDE) was an on-demand digital entertainment company with video on demand components on several platforms. RDE was founded by CEO Ryan Magnussen.
Networks
Ripe's 'networks', which consisted of Ripe TV, Octane TV and Flow TV were delivered over multiple platform. Ripe TV, launched in October 2005, focused on male content such as sports, comedy and content usually seen in the lad mag format of men's magazines such as FHM. Octane TV, which launched in August 2006, mainly involved automotive content. Flow TV launched in April 2007 and focused mainly on hip-hop music. All three services ended in June 2009 with the company's shutdown.
Distribution
The services were mainly distributed through cable video on demand services such as Comcast, Time Warner Cable and FiOS, along with syndication through the AOL, Google and Yahoo ad networks. Sprint and MVNO Helio provided RDE's content via mobile. |
https://en.wikipedia.org/wiki/Salinosporamide%20A | Salinosporamide A (Marizomib) is a potent proteasome inhibitor being studied as a potential anticancer agent. It entered phase I human clinical trials for the treatment of multiple myeloma, only three years after its discovery in 2003. This marine natural product is produced by the obligate marine bacteria Salinispora tropica and Salinispora arenicola, which are found in ocean sediment. Salinosporamide A belongs to a family of compounds, known collectively as salinosporamides, which possess a densely functionalized γ-lactam-β-lactone bicyclic core.
History
Salinosporamide A was discovered by William Fenical and Paul Jensen from Scripps Institution of Oceanography in La Jolla, CA. In preliminary screening, a high percentage of the organic extracts of cultured Salinispora strains possessed antibiotic and anticancer activities, which suggests that these bacteria are an excellent resource for drug discovery. Salinispora strain CNB-392 was isolated from a heat-treated marine sediment sample and cytotoxicity-guided fractionation of the crude extract led to the isolation of salinosporamide A. Although salinosporamide A shares an identical bicyclic ring structure with omuralide, it is uniquely functionalized. Salinosporamide A displayed potent in vitro cytotoxicity against HCT-116 human colon carcinoma with an IC50 value of 11 ng mL-1. This compound also displayed potent and highly selective activity in the NCI's 60-cell-line panel with a mean GI50 value (the concentration required to achieve 50% growth inhibition) of less than 10 nM and a greater than 4 log LC50 differential between resistant and susceptible cell lines. The greatest potency was observed against NCI-H226 non-small cell lung cancer, SF-539 brain tumor, SK-MEL-28 melanoma, and MDA-MB-435 melanoma (formerly misclassified as breast cancer), all with LC50 values less than 10 nM. Salinosporamide A was tested for its effects on proteasome function because of its structural relationship to omuralide. When teste |
https://en.wikipedia.org/wiki/OLPC%20XO | The OLPC XO (formerly known as $100 Laptop, Children's Machine, 2B1) is a low cost laptop computer intended to be distributed to children in developing countries around the world, to provide them with access to knowledge, and opportunities to "explore, experiment and express themselves" (constructionist learning). The XO was developed by Nicholas Negroponte, a co-founder of MIT's Media Lab, and designed by Yves Behar's Fuseproject company. The laptop is manufactured by Quanta Computer and developed by One Laptop per Child (OLPC), a non-profit 501(c)(3) organization.
The subnotebooks were designed for sale to government-education systems which then would give each primary school child their own laptop. Pricing was set to start at US$188 in 2006, with a stated goal to reach the $100 mark in 2008 and the 50-dollar mark by 2010. When offered for sale in the Give One Get One campaigns of Q4 2006 and Q4 2007, the laptop was sold at $199.
The rugged, low-power computers use flash memory instead of a hard disk drive (HDD), and come with a pre-installed operating system derived from Fedora Linux, with the Sugar graphical user interface (GUI). Mobile ad hoc networking via 802.11s Wi-Fi mesh networking, to allow many machines to share Internet access as long as at least one of them could connect to an access point, was initially announced, but quickly abandoned after proving unreliable.
The latest version of the OLPC XO is the XO-4 Touch, introduced in 2012.
History
The first early prototype was unveiled by the project's founder Nicholas Negroponte and then-United Nations Secretary-General Kofi Annan on November 16, 2005, at the World Summit on the Information Society (WSIS) in Tunis, Tunisia. The device shown was a rough prototype using a standard development board. Negroponte estimated that the screen alone required three more months of development. The first working prototype was demonstrated at the project's Country Task Force Meeting on May 23, 2006.
In 2006, there |
https://en.wikipedia.org/wiki/Complex%20response | A complex response refers to an environmental reaction to change that occurs at multiple levels to multiple objects, and can induce a chain reaction of responses to a single initial change. It is akin to the butterfly effect: one small event (change) can cascade through a given system creating new agents of change, and operating at several levels. The term is most commonly used in fluvial geomorphology, or the study of river systems and changes within those systems.
See also
complex system
Complex systems theory
Geomorphology |
https://en.wikipedia.org/wiki/Etest | Etest (previously known as the Epsilometer test) is a way of determining antimicrobial sensitivity by placing a strip impregnated with antimicrobials onto an agar plate. A strain of bacterium or fungus will not grow near a concentration of antibiotic or antifungal if it is sensitive. For some microbial and antimicrobial combinations, the results can be used to determine a minimum inhibitory concentration (MIC). Etest is a proprietary system manufactured by bioMérieux. It is a laboratory test used in healthcare settings to help guide physicians by indicating what concentration of antimicrobial could successfully be used to treat patients' infections.
Use
Etest is a quantitative technique for determining the antibiotic sensitivity and minimum inhibitory concentration (in µg/mL) of some bacteria including Gram-negative and Gram-positive aerobic bacteria such as Enterobacteriaceae, Pseudomonas, Burkholderia, Staphylococcus, and Enterococcus species and fastidious bacteria, such as anaerobes, N. gonorrhoeae, S. pneumoniae, Streptococcus and Haemophilius species. It can also be used to determine MICs against certain fungi.
Procedure
Etest is a pre-prepared non-porous plastic reagent strip with a predefined gradient of antibiotic, covering a continuous concentration range. It is applied to the surface of an agar plate inoculated with the test strain, where there is release of the antimicrobial gradient from the plastic carrier to the agar to form a stable and continuous gradient beneath and in nearby to the strip.
The time taken for a plate to be ready depends on the bacterium that is being tested, and the conditions of the agar plate. The predefined Etest gradient remains stable for at least 18 to 24 hours; that is, a period that covers the critical times of many species of fastidious and non-fastidious organisms.
After the test, the bacterial growth becomes visible after incubation and a symmetrical inhibition ellipse centered along the strip is seen. The MIC value |
https://en.wikipedia.org/wiki/Marine%20sediment | Marine sediment, or ocean sediment, or seafloor sediment, are deposits of insoluble particles that have accumulated on the seafloor. These particles have their origins in soil and rocks and have been transported from the land to the sea, mainly by rivers but also by dust carried by wind and by the flow of glaciers into the sea. Additional deposits come from marine organisms and chemical precipitation in seawater, as well as from underwater volcanoes and meteorite debris.
Except within a few kilometres of a mid-ocean ridge, where the volcanic rock is still relatively young, most parts of the seafloor are covered in sediment. This material comes from several different sources and is highly variable in composition. Seafloor sediment can range in thickness from a few millimetres to several tens of kilometres. Near the surface seafloor sediment remains unconsolidated, but at depths of hundreds to thousands of metres the sediment becomes lithified (turned to rock).
Rates of sediment accumulation are relatively slow throughout most of the ocean, in many cases taking thousands of years for any significant deposits to form. Sediment transported from the land accumulates the fastest, on the order of one metre or more per thousand years for coarser particles. However, sedimentation rates near the mouths of large rivers with high discharge can be orders of magnitude higher. Biogenous oozes accumulate at a rate of about one centimetre per thousand years, while small clay particles are deposited in the deep ocean at around one millimetre per thousand years.
Sediments from the land are deposited on the continental margins by surface runoff, river discharge, and other processes. Turbidity currents can transport this sediment down the continental slope to the deep ocean floor. The deep ocean floor undergoes its own process of spreading out from the mid-ocean ridge, and then slowly subducts accumulated sediment on the deep floor into the molten interior of the earth. In turn, molt |
https://en.wikipedia.org/wiki/Urban%20forests%20of%20Perm | The total area of urban forests of Perm in 2006 was 33,890 ha (338.90 km2). They are controlled by the Perm urban forestry commission (). On September 5, 2005 by Order #400-P of the governor of Perm Oblast, control over the local urban forests was transferred from federal forestries to the city.
In 2005 the Perm municipal urban forestry commission reported taking the following measures:
Gathering of wind-fallen trees and deadfalls in Chernyayevsky Forest over an area of 11 ha.
Gathering of rubbish over an area of 54 ha.
Maintenance of forest over an area of 6 ha.
Some 3.5 ha of pine-trees were planted in Chernyayevsky Forest
Measures were taken in biological forest pestscontrol; artificial nests were placed:
in Chernyayevsky Forest (80 nests);
in SPNT (Special Protected Natural Territory) "Sosnovy Bor" (70 nests);
in the forests of Ordzhonikidzevsky City District (60 nests);
in the forests of Kirovsky and Dzerzhinsky city districts (74 nests).
18 violations were noted, files were transferred to the MVD and Rosprirodnadzor (Russian nature inspection).
18 barriers were installed to prevent cars from entering Chernyayevsky Forest.
16 forest fires were extinguished.
See also
Trees and shrubs of Perm |
https://en.wikipedia.org/wiki/RespOrg | A RespOrg, or responsible organization, is a company that maintains the registration for individual toll-free telephone numbers In the North American Numbering Plan by means of the distributed Service Management System/800 database.
RespOrgs were established in 1993 as part of a Federal Communications Commission order instituting toll-free number portability. A RespOrg (pronounced as though it were a single word, something like "ressporg") can be a long-distance company, reseller, end user or an independent that offers an outsourced service.
Operation
The initial implementation of toll-free calling was primitive. In the 1950s, a collect call or a call to a Zenith number had to be completed manually by a telephone operator. By 1967, a direct-dial 1-800-number could be provided using Wide Area Telephone Service (WATS), but each prefix was tied to a specific geographic destination and each number was installed with special fixed-rate trunks which were priced beyond the reach of most small businesses. There was no means to select between rival carriers and little room for vanity numbering; a subscriber would need three separate numbers to be reachable from Canada, US interstate and US intrastate.
A "data base communication call processing method" patented by Roy P. Weber of Bell Labs, and implemented by AT&T in 1982, broke the link between individual telephone numbers and a specific trunk, city, or carrier. A toll-free number was merely an index into a large, distributed database; any number could be reassigned geographically anywhere by changing its database record. A call could be routed to one of multiple locations based on the call origin, load balancing between multiple call centers, times, or days. While this data was originally maintained by telephone companies, the breakup of the Bell System in the 1980s and the introduction of toll-free number portability in 1993 required an independent operator to maintain the SMS/800 database.
If the Service Management Sy |
https://en.wikipedia.org/wiki/ITerating | ITerating was a Wiki-based software guide, where users could find, compare and give reviews to software products. As of January 2021 the domain is listed as being for sale and the website no longer on-line. Founded in October 2005, and based in New York, ITerating was created by CEO Nicolas Vandenberghe, who saw that there was an industry need for a comprehensive resource to help evaluate software solutions.
The site aims to be a reference guide for the IT industry and includes reviews, ratings, articles, and detailed product feature comparisons. ITerating uses Semantic Web tools (including RDF - Resource Description Framework) to combine user edits with Web service feeds from other sites.
Designed for use by developers and industry consultants, ITerating allows users to contribute to categories such as Software Engineering Tools; Website Design & Tools; Website Software Tools; Website & Communication Applications & Social Networking; or to create their own category if does not exist yet.
Wiki Matrix
Iterating announced the addition of a Feature Matrix in June 2007, which allows users to dynamically create customized, side-by-side feature comparisons of software solutions. |
https://en.wikipedia.org/wiki/Copypasta | A copypasta is a block of text copied and pasted to the internet and social media. Copypasta containing controversial ideas or lengthy rants are often posted for humorous purposes, to provoke reactions from those unaware that the posted text is a meme.
History
The term copypasta is derived from the computer interface term "copy and paste", the act of selecting a piece of text and copying it elsewhere.
Usage of the word can be traced back to an anonymous 4chan thread from 2006, and Merriam-Webster record it appearing on Usenet and Urban Dictionary for the first time that year.
Examples
Navy Seal
The Navy Seal copypasta, also sometimes known as Gorilla Warfare per a misspelling of "guerrilla warfare" in its contents, is an aggressive but humorous attack paragraph supposedly written by an extremely well-trained member of the United States Navy SEALs (hence its name) to an unidentified "kiddo", ostensibly whoever the copypasta is directed to. Written in a manner similar to a non-serious death threat, the copypasta has the author threaten the recipient while boasting of their own increasingly absurd or unfeasible accomplishments, such as having "over 300 confirmed kills" or being able to kill someone "in over seven hundred ways, and thats just with my bare hands". This copypasta is often reposted as a humorous overreaction to an insult and is thought to have originated on the military-themed imageboard OperatorChan, in a post dated 11 November 2010.
In 2019, the copypasta appeared in the manifesto of the perpetrator of the Christchurch mosque shootings, causing some news sources to report the claims at face value.
Bee Movie
The Bee Movie copypasta, often called the Bee Movie script per its contents, is the entire screenplay of the 2007 animated film Bee Movie, though this is sometimes shortened to just the introductory monologue ("According to all known laws of aviation, there is no way a bee should be able to fly. Its wings are too small to get its fat little |
https://en.wikipedia.org/wiki/Concordance%20correlation%20coefficient | In statistics, the concordance correlation coefficient measures the agreement between two variables, e.g., to evaluate reproducibility or for inter-rater reliability.
Definition
The form of the concordance correlation coefficient as
where and are the means for the two variables and and are the corresponding variances. is the correlation coefficient between the two variables.
This follows from its definition as
When the concordance correlation coefficient is computed on a -length data set (i.e., paired data values , for ), the form is
where the mean is computed as
and the variance
and the covariance
Whereas the ordinary correlation coefficient (Pearson's) is immune to whether the biased or unbiased versions for estimation of the variance is used, the concordance correlation coefficient is not. In the original article Lin suggested the 1/N normalization,
while in another article Nickerson appears to have used the 1/(N-1),
i.e., the concordance correlation coefficient may be computed slightly differently between implementations.
Relation to other measures of correlation
The concordance correlation coefficient is nearly identical to some of the measures called intra-class correlations. Comparisons of the concordance correlation coefficient with an "ordinary" intraclass correlation on different data sets found only small differences between the two correlations, in one case on the third decimal. It has also been stated that the ideas for concordance correlation coefficient "are quite similar to results already published by Krippendorff
in 1970".
In the original article Lin suggested a form for multiple classes (not just 2). Over ten years later a correction to this form was issued.
One example of the use of the concordance correlation coefficient is in a comparison of analysis method for functional magnetic resonance imaging brain scans.
External links
Statistical Calculator. Provided by NIWA, it is an online version of Lin’s concordance used to asses |
https://en.wikipedia.org/wiki/Isostatic%20depression | Isostatic depression is the sinking of large parts of the Earth's crust into the asthenosphere caused by a heavy weight placed on the Earth's surface, often glacial ice during continental glaciation. Isostatic depression and isostatic rebound occur at rates of centimeters per year. Greenland is an example of an isostatically depressed region.
Glacial isostatic depression
Isostatic depression is a phase of glacial isostasy, along with isostatic rebound. Glacial isostasy is the Earth's response to changing surface loads of ice and water during the expansion and contraction of large ice sheets. The Earth's asthenosphere acts elastically, flowing when exposed to loads and non-hydrostatic stress, such as ice sheets, for an extended period of time. The Earth's crust is depressed by the product of thickness of ice and the ratio of ice and mantle densities. This large ice load results in elastic deformation of the entire lithospheric mantle over the span of 10,000-100,000 years, with the load eventually supported by the lithosphere after the limit of local isostatic depression has been attained. Historically, isostatic depression has been used to estimate global ice volume by relating the magnitude of depression to the density of ice and upper mantle material.
Glacial megalakes can form in regional depressions under the influence of glacial load.
Isostatic depression in Greenland
Greenland is isostatically depressed by the Greenland ice sheet. However, due to deglaciation induced by climate change, the regions near the shrinking ice sheet have begun to uplift, a process known as post-glacial rebound. Modeling these glacial isostatic adjustments has been an area of interest for some time now as the entire topography of Greenland is affected by these movements. These movements are unique in that they can be observed on a human time scale unlike other geological processes. Models have been created to assess what future equilibrium states of the Greenland ice sheet will l |
https://en.wikipedia.org/wiki/GABA%20receptor%20antagonist | GABA receptor antagonists are drugs that inhibit the action of GABA. In general these drugs produce stimulant and convulsant effects, and are mainly used for counteracting overdoses of sedative drugs.
Examples include bicuculline, securinine and metrazol, and the benzodiazepine GABAA receptor antagonist flumazenil.
Other agents which may have GABAA receptor antagonism include the antibiotic ciprofloxacin, tranexamic acid, thujone, ginkgo biloba, and kudzu.
See also
GABAA receptor negative allosteric modulators
External links |
https://en.wikipedia.org/wiki/Hidehiko%20Yamabe | was a Japanese mathematician. Above all, he is famous for discovering that every conformal class on a smooth compact manifold is represented by a Riemannian metric of constant scalar curvature. Other notable contributions include his definitive solution of Hilbert's fifth problem.
Life
Hidehiko Yamabe was born on August 22, 1923, in the city of Ashiya, belonging to the Hyōgo Prefecture, the sixth son of Takehiko and Rei Yamabe. After completing the Senior High School in September 1944, he joined Tokyo University as a student of the Department of Mathematics and graduated in September 1947: his doctoral advisor was Shokichi Iyanaga. He was then associated with the Department of Mathematics at Osaka University until June 1956, even while employed by the Department of Mathematics at Princeton University in Princeton, New Jersey. Shortly before coming to the United States of America, Yamabe married his wife Etsuko, and by 1956 they had two daughters. Yamabe died suddenly of a stroke in November 1960, just months after accepting a full professorship at Northwestern University.
Academic career
After graduating from the University of Tokyo in 1947, Yamabe became an assistant at Osaka University. From 1952 until 1954 he was an assistant at Princeton University, receiving his Ph.D. from Osaka University while at Princeton. He left Princeton in 1954 to become assistant professor at the University of Minnesota. Except for one year as a professor at Osaka University, he stayed in Minnesota until 1960. Yamabe died suddenly of a stroke in November 1960, just months after accepting a full professorship at Northwestern University.
The Yamabe Memorial Lecture and the Yamabe Symposium
After coming back to Japan, Etsuko Yamabe and her daughters lived with the benefits of Hidehiko's social security and of funds raised privately by her and her husband's friends in the United States of America. When she had achieved some financial stability, it was her wish to return the kindness sh |
https://en.wikipedia.org/wiki/MK-886 | MK-886, or L-663536, is a leukotriene antagonist. It may perform this by blocking the 5-lipoxygenase activating protein (FLAP), thus inhibiting 5-lipoxygenase (5-LOX), and may help in treating atherosclerosis. |
https://en.wikipedia.org/wiki/Linphone |
Linphone (contraction of Linux phone) is a free voice over IP softphone, SIP client and service. It may be used for audio and video direct calls and calls through any VoIP softswitch or IP-PBX. Linphone also provides the possibility to exchange instant messages. It has a simple multilanguage interface based on Qt for GUI and can also be run as a console-mode application on Linux.
The softphone is currently developed by Belledonne Communications in France. Linphone was initially developed for Linux but now supports many additional platforms including Microsoft Windows, macOS, and mobile phones running Windows Phone, iOS or Android. It supports ZRTP for end-to-end encrypted voice and video communication.
Linphone is licensed under the GNU GPL-3.0-or-later and supports IPv6. Linphone can also be used behind network address translator (NAT), meaning it can run behind home routers. It is compatible with telephony by using an Internet telephony service provider (ITSP).
Features
Linphone hosts a free SIP service on its website.
The Linphone client provides access to following functionalities:
Multi-account work
Registration on any SIP-service and line status management
Contact list with status of other users
Conference call initiation
Combination of message history and call details
DTMF signals sending (SIP INFO / RFC 2833)
File sharing
Additional plugins
Open standards support
Protocols
SIP according to RFC 3261 (UDP, TCP and TLS)
SIP SIMPLE
NAT traversal by TURN and ICE
RTP/RTCP
Media-security: SRTP and ZRTP
Audio codecs
Audio codec support: Speex (narrow band and wideband), G.711 (μ-law, A-law), GSM, Opus, and iLBC (through an optional plugin)
Video codecs
Video codec support: MPEG-4, Theora, VP8 and H.264 (with a plugin based on x264), with resolutions from QCIF (176×144) to SVGA (800×600) provided that network bandwidth and CPU power are sufficient.
Gallery
See also
Comparison of VoIP software
List of SIP software
Opportunistic encryption |
https://en.wikipedia.org/wiki/Dung%20midden | Dung middens, also known as dung hills, are piles of dung that mammals periodically return to and build up. They are used as a form of territorial marker. A range of animals are known to use them including steenbok, hyrax, and rhinoceros. Other animals are attracted to middens for a variety of purposes, including finding food and locating mates. Some species, such as the dung beetle genus Dicranocara of the Richtersveld in South western Africa spend their whole lifecycle in close association with dung middens. Dung middens are also used in the field of Paleobotany, which relies on the fact that each ecosystem is characterized by certain plants, which in turn act as a proxy for climate. Dung middens are useful as they often contain pollen which means fossilized dung middens can be used in Paleobotany to learn about past climates.
Examples of dung midden production in wild
Hippopotamus
The common hippopotamus has been known to use dung middens as a social tool. The middens are created and maintained by bulls to mark territorial boundaries. To mark their scent upon a midden, the bull will approach the midden in reverse and simultaneously defecate and urinate on the mound, using its tail to disperse, or paddle, the excrement. This action is called dung showering and thought to assert dominance. The middens, usually several feet across, are constantly maintained during the bulls' travels in the night and day.
Rhinoceros
Dung-midden production is also observed in the White and Black rhinoceroses. The middens are shown to provide cues as to the age, sex, and reproductive health of the producer. Some of the middens can be 65 feet across. Dung beetles are frequently found in these middens and lay their eggs within the mounds. Their presence and activity in the middens also aid in pest and parasite control. Unlike the hippopotamus, rhino dung middens are shared between individuals that are not necessarily related.
White rhino middens are distinguished by a black color |
https://en.wikipedia.org/wiki/Fortepiano%20%28musical%20dynamic%29 | The expression fortepiano (sometimes called forte piano) is a sudden dynamic change used in a musical score, usually with the abbreviation , to designate a section of music in which the music should be played loudly (forte), then immediately softly (piano). It is not unusual for it to be followed by a crescendo, a gradual increase in dynamics. The word is of Italian etymology literally translated as 'loudsoft'.
Examples
Beethoven's Piano Sonata No. 8 begins with a fortepiano: |
https://en.wikipedia.org/wiki/Apache%20CXF | Apache CXF is an open source software project developing a Web services framework. It originated as the combination of Celtix developed by IONA Technologies and XFire developed by a team hosted at Codehaus in 2006. These two projects were combined at the Apache Software Foundation. The name "CXF" was derived by combining "Celtix" and "XFire".
Description
CXF is often used with Apache ServiceMix, Apache Camel and Apache ActiveMQ in service-oriented architecture (SOA) infrastructure projects.
Apache CXF supports the Java programming interfaces JAX-WS, JAX-RS, JBI, JCA, JMX, JMS over SOAP, Spring, and the XML data binding frameworks JAXB, Aegis, Apache XMLBeans, SDO.
CXF includes the following:
Web Services Standards Support:
SOAP
WS-Addressing
WS-Policy
WS-ReliableMessaging
WS-SecureConversation
WS-Security
WS-SecurityPolicy
JAX-WS API for Web service development
Java-first support
WSDL-first tooling
JAX-RS (JSR 339 2.0) API for RESTful Web service development
JavaScript programming model for service and client development
Maven tooling
CORBA support
HTTP, JMS and WebSocket transport layers
Embeddable Deployment:
ServiceMix or other JBI containers
Geronimo or other Java EE containers
Tomcat or other servlet containers
OSGi
Reference OSGi Remote Services implementation
IONA Technologies distributes a commercial Enterprise version of Apache CXF under the name FUSE Services Framework.
See also
The Axis Web Services framework
Apache Wink, a project in incubation with JAX-RS support
List of web service frameworks
Citations |
https://en.wikipedia.org/wiki/Daly%20detector | A Daly detector is a gas-phase ion detector that consists of a metal "doorknob", a scintillator (phosphor screen) and a photomultiplier. It was named after its inventor Norman Richard Daly. Daly detectors are typically used in mass spectrometers.
Principle of operation
Ions that hit the doorknob release secondary electrons. A high voltage (about ) between the doorknob and the scintillator accelerates the electrons onto the phosphor screen, where they are converted to photons. These photons are detected by the photomultiplier.
The advantage of the Daly detector is that the photomultiplier can be separated by a window, which lets the photons through from the high vacuum of the mass spectrometer, thus preventing an otherwise possible contamination and extending life span of the detector. The Daly detector also allows a higher acceleration after the field-free region of a time-of-flight mass spectrometer flight tube, which can improve the sensitivity for heavy ions.
Norman Richard Daly
Norman Daly was awarded 6 patents in the years 1962–1973 relating to ion detection and mass spectrometers, from his work at the United Kingdom Atomic Energy Authority. |
https://en.wikipedia.org/wiki/Software%20industry%20in%20Karnataka | The software industry in Karnataka state in India has become one of the main pillars of the state's economy. Karnataka stands first among all the states of India in terms of revenue generated from software exports, accounting for 41.6% of all software exports from the country during the financial year of 2018–19. Karnataka's capital city Bangalore has the sobriquet of Silicon Valley of India, with total IT exports worth US$ 53 billion during the financial year 2021–22, employing 1 000 000 people directly and 3 000 000 lakh indirectly. Though most software companies are located in Bangalore, some have settled in other cities like Mysore, Mangalore, Belgaum and Hubli in Karnataka. The infrastructure required for setting up software industries in Karnataka is provided by STPI. The software industry in Karnataka includes companies dealing with various fields like telecommunication, banking software, avionics, database, automotive, networking, semiconductors, mobile handsets, internet applications and business process outsourcing. Currently, out of total IT exports, 95% is from Bengaluru alone and the other Karnataka cities contribute just 5%.
The Nandi Hills area on the outskirts of Devanahalli is the site of the upcoming $22 billion, BIAL IT Investment Region, one of the largest infrastructure projects in the history of Karnataka. This project is expected to create over four million jobs by the year 2030.
Origin
Starting in the 1980s, Karnataka emerged as the information technology capital of the country. A total of 1973 companies in Karnataka are involved in Information Technology related business including big firms like Infosys and Wipro who have their headquarters in Bangalore. The origin of the growth of the software industry in Karnataka seems to have been the entry of Texas Instruments which was the first multinational to set up base in Sona Tower, Millers Road, Bangalore in 1985.Texas Instruments
was searching for a location to set up their overseas develo |
https://en.wikipedia.org/wiki/Bioaerosol | Bioaerosols (short for biological aerosols) are a subcategory of particles released from terrestrial and marine ecosystems into the atmosphere. They consist of both living and non-living components, such as fungi, pollen, bacteria and viruses. Common sources of bioaerosols include soil, water, and sewage.
Bioaerosols are typically introduced into the air via wind turbulence over a surface. Once in the atmosphere, they can be transported locally or globally: common wind patterns/strengths are responsible for local dispersal, while tropical storms and dust plumes can move bioaerosols between continents. Over ocean surfaces, bioaerosols are generated via sea spray and bubbles
Bioaerosols can transmit microbial pathogens, endotoxins, and allergens to which humans are sensitive. A well-known case was the meningococcal meningitis outbreak in sub-Saharan Africa, which was linked to dust storms during dry seasons. Other outbreaks linked to dust events including Mycoplasma pneumonia and tuberculosis.
Another instance was an increase in human respiratory problems in the Caribbean that may have been caused by traces of heavy metals, microorganism bioaerosols, and pesticides transported via dust clouds passing over the Atlantic Ocean.
Background
Charles Darwin was the first to observe the transport of dust particles but Louis Pasteur was the first to research microbes and their activity within the air. Prior to Pasteur’s work, laboratory cultures were used to grow and isolate different bioaerosols.
Since not all microbes can be cultured, many were undetected before the development of DNA-based tools. Pasteur also developed experimental procedures for sampling bioaerosols and showed that more microbial activity occurred at lower altitudes and decreased at higher altitudes.
Types of bioaerosols
Bioaerosols include fungi, bacteria, viruses, and pollen. Their concentrations are greatest in the planetary boundary layer (PBL) and decrease with altitude. Survival rate of bioae |
https://en.wikipedia.org/wiki/IBM%20NetVista | NetVista is an umbrella name for a variety of products manufactured by IBM.
Software suite
The Software Suite was introduced in April 1996 as a client–server software suite, with the server software running on OS/2, and the client software on Windows 3.1 and Windows 95. Meant to provide Internet access to K-12 users, it included such things as a web browser, nanny software and other internet utilities, including a TCP/IP stack.
Starting with version 1.1, the server side was also supported on Windows NT. The software suite was withdrawn without replacement in January 2000.
Products:
NetVista V1.0
NetVista V1.1
NetVista V2.0
Network station
In April 2000, the IBM Network Station product line was renamed to IBM NetVista, as were the associated software tools. The NetVista computers were thin client systems. The line was withdrawn in April 2002 with no replacement.
Hardware products:
NetVista N2200 (Cyrix MediaGX at 233 MHz, 32-288 MB RAM, CompactFlash, Ethernet, USB 1.1, VGA, Audio I/O)
NetVista N2200e
NetVista N2200l
NetVista N2200w
NetVista N2800
NetVista N2800e
NetVista N70
Software products:
NetVista Thin Client Manager V2R1
Kiosk
Hardware products:
NetVista Kiosk Model 120
NetVista Kiosk Model 150
Appliance
This appliance is meant to allow internet access on a TV. It was not sold directly to end-users, but offered as an OEM product to internet providers.
Hardware products:
NetVista Internet Appliance i30
Personal computer
The IBM NetVista personal computer was the follow-on to the IBM PC Series. It was announced in May 2000, and withdrawn in May 2004. It was replaced by the IBM ThinkCentre (now Lenovo ThinkCentre since 2005). Initially offered in the typical white/beige cases of the 1990s the NetVista was sold in black later on.
Products:
A Series
IBM NetVista A10
IBM NetVista A20 (Pentium III)
IBM NetVista A20i (Pentium III)
IBM NetVista A21 (Celeron)
IBM NetVista A21i (Pentium III)
IBM NetVista A22 (Celeron)
IBM NetVista A22p (P |
https://en.wikipedia.org/wiki/MALDI%20imaging | MALDI mass spectrometry imaging (MALDI-MSI) is the use of matrix-assisted laser desorption ionization as a mass spectrometry imaging technique in which the sample, often a thin tissue section, is moved in two dimensions while the mass spectrum is recorded. Advantages, like measuring the distribution of a large amount of analytes at one time without destroying the sample, make it a useful method in tissue-based study.
Sample preparation
Sample preparation is a critical step in imaging spectroscopy. Scientists take thin tissue slices mounted on conductive microscope slides and apply a suitable MALDI matrix to the tissue, either manually or automatically. Next, the microscope slide is inserted into a MALDI mass spectrometer. The mass spectrometer records the spatial distribution of molecular species such as peptides, proteins or small molecules. Suitable image processing software can be used to import data from the mass spectrometer to allow visualization and comparison with the optical image of the sample. Recent work has also demonstrated the capacity to create three-dimensional molecular images using MALDI imaging technology and comparison of these image volumes to other imaging modalities such as magnetic resonance imaging (MRI).
Tissue preparation
The tissue samples must be preserved quickly in order to reduce molecular degradation. The first step is to freeze the sample by wrapping the sample then submerging it in a cryogenic solution. Once frozen, the samples can be stored below -80 °C for up to a year.
When ready to be analyzed, the tissue is embedded in a gelatin media which supports the tissue while it is being cut, while reducing contamination that is seen in optimal cutting temperature compound (OCT) techniques. The mounted tissue section thickness varies depending on the tissue.
Tissue sections can then be thaw-mounted by placing the sample on the surface of a conductive slide that is of the same temperature, and then slowly warmed from below. The s |
https://en.wikipedia.org/wiki/Percolation%20threshold | The percolation threshold is a mathematical concept in percolation theory that describes the formation of long-range connectivity in random systems. Below the threshold a giant connected component does not exist; while above it, there exists a giant component of the order of system size. In engineering and coffee making, percolation represents the flow of fluids through porous media, but in the mathematics and physics worlds it generally refers to simplified lattice models of random systems or networks (graphs), and the nature of the connectivity in them. The percolation threshold is the critical value of the occupation probability p, or more generally a critical surface for a group of parameters p1, p2, ..., such that infinite connectivity (percolation) first occurs.
Percolation models
The most common percolation model is to take a regular lattice, like a square lattice, and make it into a random network by randomly "occupying" sites (vertices) or bonds (edges) with a statistically independent probability p. At a critical threshold pc, large clusters and long-range connectivity first appears, and this is called the percolation threshold. Depending on the method for obtaining the random network, one distinguishes between the site percolation threshold and the bond percolation threshold. More general systems have several probabilities p1, p2, etc., and the transition is characterized by a critical surface or manifold. One can also consider continuum systems, such as overlapping disks and spheres placed randomly, or the negative space (Swiss-cheese models).
To understand the threshold, you can consider a quantity such as the probability that there is a continuous path from one boundary to another along occupied sites or bonds—that is, within a single cluster. For example, one can consider a square system, and ask for the probability P that there is path from the top boundary to the bottom boundary. As a function of the occupation probability p, one finds a sigmoid |
https://en.wikipedia.org/wiki/Bioclimatology | Bioclimatology is the interdisciplinary field of science that studies the interactions between the biosphere and the Earth's atmosphere on time scales of the order of seasons or longer (in contrast to biometeorology).
Examples of relevant processes
Climate processes largely control the distribution, size, shape and properties of living organisms on Earth. For instance, the general circulation of the atmosphere on a planetary scale broadly determines the location of large deserts or the regions subject to frequent precipitation, which, in turn, greatly determine which organisms can naturally survive in these environments. Furthermore, changes in climates, whether due to natural processes or to human interferences, may progressively modify these habitats and cause overpopulation or extinction of indigenous species.
The biosphere, for its part, and in particular continental vegetation, which constitutes over 99% of the total biomass, has played a critical role in establishing and maintaining the chemical composition of the Earth's atmosphere, especially during the early evolution of the planet (See History of Earth for more details on this topic). Currently, the terrestrial vegetation exchanges some 60 billion tons of carbon with the atmosphere on an annual basis (through processes of carbon fixation and carbon respiration), thereby playing a critical role in the carbon cycle. On a global and annual basis, small imbalances between these two major fluxes, as do occur through changes in land cover and land use, contribute to the current increase in atmospheric carbon dioxide. |
https://en.wikipedia.org/wiki/3-Mercaptopropane-1%2C2-diol | 3-Mercaptopropane-1,2-diol, also known as thioglycerol, is a chemical compound and thiol that is used as a matrix in fast atom bombardment mass spectrometry and liquid secondary ion mass spectrometry.
See also
Glycerol
Mercaptoethanol |
https://en.wikipedia.org/wiki/Tobu%20World%20Square | is a theme park in Kinugawa Onsen, Nikkō, Tochigi, Japan. The theme park contains over a hundred 1:25 scale models of famous buildings, including UNESCO-designated World Cultural and Heritage Sites, complete with 140,000 1:25 miniature people.
On 24 April 2010, a 1:25 scale model of the Tokyo Skytree was unveiled at the park. This is 26 metres tall, taller than the 19.95-metre replica of the New York World Trade Center.
Night time illumination
Between November and March the park has extended opening hours. During this time approximately 1.4 million red and blue lights illuminate the park. Buildings such as the Tokyo Sky Tree, the Eiffel Tower and the Duomo di Milano are lit up with LEDs and spotlights. There is also a 150-meter tunnel of lights that leads up to the Alpine Roses Park.
List of exhibits
Bold Exhibits are UNESCO Cultural heritage sites.
Modern Japan zone
America Zone
Egypt Zone
Europe zone
Asia zone
Japan Zone |
https://en.wikipedia.org/wiki/Asia%20Sentinel | The Asia Sentinel is an online blog focused on news, business, arts and culture in Asia. The site was launched in August 2006 in Hong Kong, and its assets were transferred to a U.S. registered company in 2017.
History
The Asia Sentinel was founded in Hong Kong in August 2006 by four journalists from the United Kingdom and the United States who were based in Asia. The editor-in-chief, John Berthelsen, was formerly a correspondent with The Wall Street Journal Asia, as well as the Newsweek correspondent in Vietnam and managing editor of the Hong Kong Standard. Co-founder Philip Bowring was formerly the editor of the Far Eastern Economic Review and a former columnist for the International Herald Tribune. Executive Editor A. Lin Neumann, a reporter, was formerly executive editor of The Standard and also represented the Committee to Protect Journalists in Asia. Neumann left the Asia Sentinel in 2012. The fourth founder, Anthony Spaeth, was previously a Time Asia regional correspondent, and left the Asia Sentinel shortly after its founding to work for Bloomberg.
In 2017, the publication's parent company in Hong Kong ceased operations, and assets were transferred to a new company, registered in California, that owns all newly written stories. In 2023, access to the Asia Sentinel was blocked in Singapore after it failed to comply with a POFMA order issued by the Singapore Government. |
https://en.wikipedia.org/wiki/Amazon%20Simple%20Queue%20Service | Amazon Simple Queue Service (Amazon SQS) is a distributed message queuing service introduced by Amazon.com as a beta in late 2004, and generally available in mid 2006. It supports programmatic sending of messages via web service applications as a way to communicate over the Internet. SQS is intended to provide a highly scalable hosted message queue that resolves issues arising from the common producer–consumer problem or connectivity between producer and consumer.
Amazon SQS can be described as commoditization of the messaging service. Well-known examples of messaging service technologies include IBM WebSphere MQ and Microsoft Message Queuing. Unlike these technologies, users do not need to maintain their own server. Amazon does it for them and sells the SQS service at a per-use rate.
API
Amazon provides SDKs in several programming languages including Java, Ruby, Python, .NET, PHP, Go and JavaScript. A Java Message Service (JMS) 1.1 client for Amazon SQS was released in December 2014.
Authentication
Amazon SQS provides authentication procedures to allow for secure handling of data. Amazon uses its Amazon Web Services (AWS) identification to do this, requiring users to have an AWS enabled account with Amazon.com. AWS assigns a pair of related identifiers, your AWS access keys, to an AWS enabled account to perform identification. The first identifier is a public 20-character Access Key. This key is included in an AWS service request to identify the user. If the user is not using SOAP with WS-Security, a digital signature is calculated using the Secret Access Key. The Secret Access Key is a 40-character private identifier. AWS uses the Access Key ID provided in a service request to look up an account's Secret Access Key. Amazon.com then calculates a digital signature with the key. If they match then the user is considered authentic, if not then the authentication fails and the request is not processed.
Message delivery
Amazon SQS guarantees at-least-once d |
https://en.wikipedia.org/wiki/Fourier%20algebra | Fourier and related algebras occur naturally in the harmonic analysis of locally compact groups. They play an important role in the duality theories of these groups. The Fourier–Stieltjes algebra and the Fourier–Stieltjes transform on the Fourier algebra of a locally compact group were introduced by Pierre Eymard in 1964.
Definition
Informal
Let G be a locally compact abelian group, and Ĝ the dual group of G. Then is the space of all functions on Ĝ which are integrable with respect to the Haar measure on Ĝ, and it has a Banach algebra structure where the product of two functions is convolution. We define to be the set of Fourier transforms of functions in , and it is a closed sub-algebra of , the space of bounded continuous complex-valued functions on G with pointwise multiplication. We call the Fourier algebra of G.
Similarly, we write for the measure algebra on Ĝ, meaning the space of all finite regular Borel measures on Ĝ. We define to be the set of Fourier-Stieltjes transforms of measures in . It is a closed sub-algebra of , the space of bounded continuous complex-valued functions on G with pointwise multiplication. We call the Fourier-Stieltjes algebra of G. Equivalently, can be defined as the linear span of the set of continuous positive-definite functions on G.
Since is naturally included in , and since the Fourier-Stieltjes transform of an function is just the Fourier transform of that function, we have that . In fact, is a closed ideal in .
Formal
Let be a Fourier–Stieltjes algebra and be a Fourier algebra such that the locally compact group is abelian. Let be the measure algebra of finite measures on and let be the convolution algebra of integrable functions on , where is the character group of the Abelian group .
The Fourier–Stieltjes transform of a finite measure on is the function on defined by
The space of these functions is an algebra under pointwise multiplication is isomorphic to the measure algebra . Restricted to |
https://en.wikipedia.org/wiki/Blackle | Blackle is an internet search engine powered by Google Programmable Search Engine. Blackle was created by Tony Heap of Heap Media Australia, which aims to save energy by displaying a black background with a grayish-white text color on search results. Blackle claims to have saved over 10.07 MWh of electrical energy as of July 2023.
Concept
The concept behind Blackle is that computer monitors can be made to use less energy by displaying much darker colors. Blackle is based on a study which tested a variety of CRT and LCD monitors. However, these claims are disputed over whether there are any energy saving effects, especially for users of LCD screens, where there is a constant backlight.
This concept was first brought to the attention of Heap Media by a blog post, which estimated that Google could save 750 megawatt hours a year by utilizing it for CRT screens. The homepage of Blackle provides a count of the number of watt hours claimed to have been saved by enabling this concept.
History
Blackle launched in January 2007. During this time, Blackle gained popularity and was featured in multiple mainstream media outlets.
Blackle international, which translated Blackle into Portuguese, French, Czech, Italian, and Dutch was retired in 2019. While the International page is still up, every link listed has experienced link rot. As of 2021, the site is only available in English.
See also
Light-on-dark color scheme
Performance per watt
Comparison of web search engines
List of search engines
List of search engines by popularity |
https://en.wikipedia.org/wiki/I%20Wish%20It%20Would%20Rain%20Down | "I Wish It Would Rain Down" is a song by English musician Phil Collins from his fourth solo studio album, ...But Seriously (1989). The song was a chart success in early 1990, peaking at 7 on the UK Singles Chart, No. 3 on the US Billboard Hot 100 and No. 1 on the Canadian RPM 100 Singles chart; in the latter country, it was the most successful song of 1990. Collins felt that it was as close as he had ever got, at the time, to writing a blues song.
Eric Clapton plays lead guitar throughout the song, which also features a large gospel choir. Regarding Clapton's contribution, Collins recalls, "I said 'Eric, have I never asked you to play? Come on, I've got a song right up your street.'" Collins also felt that it was a Clapton song.
Chart performance
The song was a significant chart hit in 1990, peaking at No. 3 on the Billboard Hot 100 in the United States and No. 1 on the RPM 100 Singles chart in Canada. In Canada, it was the longest-running number one single of 1990, spending six weeks atop the charts, and ranked as the top single of the year on RPMs year-end chart. It also reached number seven on the UK Singles Chart and number three in Belgium and the Netherlands.
Music video
The 8:30 minute-long black-and-white music video, produced by Paul Flattery and directed by Jim Yukich for FYI, contains 2:30 minutes of acting prior to the start of the music. The setting is a theatre in the 1930s. Actor Jeffrey Tambor plays a hyper-critical, unhappy theatre director. He is rehearsing some dancers (who are dancing to the guitar/bass guitar riff from the song "Sunshine of Your Love" by the band Cream, in which Eric Clapton played guitar.) The director complains that the girls can neither dance nor sing, and then discovers that his star has appendicitis.
Eric Clapton, seated on a stool, says that Billy (played by Collins) used to be the drummer in a good band and assumed singing responsibilities when the original singer departed—an in-joke referencing Collins' tenure with |
https://en.wikipedia.org/wiki/Electrochemical%20equivalent | In chemistry, the electrochemical equivalent (Eq or Z) of a chemical element is the mass of that element (in grams) transported by a specific quantity of electricity, usually expressed in grams per coulomb of electric charge. The electrochemical equivalent of an element is measured with a voltameter.
Definition
The electrochemical equivalent of a substance is the mass of the substance deposited to one of the electrodes when a current of 1 ampere is passed for 1 second, i.e. a quantity of electricity of one coulomb is passed.
The formula for finding electrochemical equivalent is as follows:
where is the mass of substance and is the charge passed. Since , where is the current applied and is time, we also have
Eq values of some elements in kg/C |
https://en.wikipedia.org/wiki/Density%20dependence | In population ecology, density-dependent processes occur when population growth rates are regulated by the density of a population. This article will focus on density dependence in the context of macroparasite life cycles.
Positive density-dependence
Positive density-dependence, density-dependent facilitation, or the Allee effect describes a situation in which population growth is facilitated by increased population density.
Examples
For dioecious (separate sex) obligatory parasites, mated female worms are required to complete a transmission cycle. At low parasite densities, the probability of a female worm encountering a male worm and forming a mating pair can become so low that reproduction is restricted due to single sex infections. At higher parasite densities, the probability of mating pairs forming and successful reproduction increases. This has been observed in the population dynamics of Schistosomes.
Positive density-dependence processes occur in macroparasite life cycles that rely on vectors with a cibarial armature, such as Anopheles or Culex mosquitoes. For Wuchereria bancrofti, a filarial nematode, well-developed cibarial armatures in vectors can damage ingested microfilariae and impede the development of infective L3 larvae. At low microfilariae densities, most microfilariae can be ruptured by teeth, preventing successful development of infective L3 larvae. As more larvae are ingested, the ones that become entangled in the teeth may protect the remaining larvae, which are then left undamaged during ingestion.
Positive density-dependence processes may also occur in macroparasite infections that lead to immunosuppression. Onchocerca volvulus infection promotes immunosuppressive processes within the human host that suppress immunity against incoming infective L3 larvae. This suppression of anti-parasite immunity causes parasite establishment rates to increase with higher parasite burden.
Negative density-dependence
Negative density-dependence, or dens |
https://en.wikipedia.org/wiki/Mixed%20ammonium%20salts%20of%20phosphorylated%20glycerides | The mix of ammonium salts of phosphorylated glycerides can be either made synthetically or from mixture of glycerol and partially hardened plant (most often used: rapeseed oil) oils.
Applications
It is most often used in chocolate industry as an emulsifier, often as alternative to lecithin.
Properties
At room temperature it is liquid.
Synonyms
Ammonium phosphatide
Emulsifier YN
E number E442
See also
Polyglycerol polyricinoleate (PGPR)
Food additives
E-number additives |
https://en.wikipedia.org/wiki/Sucrose%20acetate%20isobutyrate | Sucrose acetoisobutyrate (SAIB) is an emulsifier and has E number E444. In the United States, SAIB is categorized as generally recognized as safe (GRAS) as a food additive in cocktail mixers, beer, malt beverages, or wine coolers and is a potential replacement for brominated vegetable oil.
Chemistry
SAIB can be prepared by esterification of sucrose with acetic and isobutyric anhydride.
Uses
Beverage emulsions - weighting agent
Color cosmetics and skin care
Flavorings (orange flavor)
Fragrance fixative
Hair care
Horse styling products |
https://en.wikipedia.org/wiki/Priam%20Corporation | Priam Corporation was a company located in San Jose, California, founded in 1978 by William Schroeder and Al Wilson, two former Memorex executives, as a manufacturer of hard disk drives. Originally, they made high-capacity 14-inch drives, developed for mainframe computers, available for minicomputers and high-end workstations, but switched to 8-inch disk drives in 1980.
Their 8-inch hard disks could be found in a wide variety of add-on products like the huge Mator Shark box with IEEE-488 interface for Commodore PET/CBM computers or the Priam DataTower series, external storage solutions, combining high-capacity hard disks and streamers in a single case, which could interface to various computers including IBM PCs.
While Priam was considered a leader in certain technology segments at one time, they were late catching up in the transition to the 5.25-inch form factor and were ultimately one of the many hard drive manufacturers in the 1980s and 1990s that went out of business, merged, or closed their hard drive divisions; as a result of capacities and demand for products increased, and profits became hard to find.
In 1985, Priam dropped their own 5.25-inch drive in-house development and merged with Vertex Peripherals, a company competing in the market for 5.25-inch drives.
In the mid-1980s Priam's InnerSpace series offered fully integrated disk drive solutions to overcome disk capacity limitations imposed by MS-DOS prior to the release of version 3.31/4.0. This included custom hard disk controllers (based on Western Digital designs) as well as dedicated software for formatting (PFMT.EXE) and partitioning (EDISK.EXE). PFMT.EXE placed a driver EVDR.SYS into the root directory of the boot volume. This was used by DOS to retrieve drive and partitioning information located in the last 12 KB of the disk. Priam's EDISK also used dedicated MBR partition IDs 0x45 and 0x5C. Solutions were offered for DOS, NetWare, Unix and Pick.
Priam opened a manufacturing facility in Taiwa |
https://en.wikipedia.org/wiki/Wang%20Institute%20of%20Graduate%20Studies | The Wang Institute of Graduate Studies was an independent educational institution founded in 1979 by computer entrepreneur An Wang. Its purpose was to provide professional and continuing studies in the nascent field of software engineering. It was accredited by the New England Association of Schools and Colleges in 1983. Faculty members were recruited from industry and students were required to have a minimum of three years prior experience in industry as a condition of acceptance.
The Institute acquired its campus from the Marist Brothers who had operated a seminary on the site since 1924. Located in Tyngsborough, Massachusetts, it housed two divisions: The School of Information Technology and a fellowship program in East Asian studies.
The Institute never grew beyond a dozen or so faculty. As a result of declining business fortunes Dr. Wang closed the Institute, graduating the last class on August 27, 1988.McKeeman, William, "Graduation Talk at Wang Institute," Computer, vol. 22, no. 5, pp. 78-80 (1989) The campus was transferred to Boston University where it served as a corporate education center. Today, it is the location of the Innovation Academy Charter School.
Software engineering curriculum
The Institute graduated seven classes between 1982 and 1988 in its Master of Software Engineering program, requiring study in eleven three-credit courses. Two project courses involved students in team-based analysis, specification, design, implementation, testing, and integration of software products.Fairley, Richard and Martin, Nancy. "Software engineering programs at the Wang Institute of Graduate Studies," Proceedings of the 1983 annual conference on Computers (1983)
The original six core courses were:
The curriculum was later modified to include an optional operating systems course instead of the architecture course.Ardis, Mark. "The Evolution of Wang Institute's Master of Software Engineering Program," IEEE Transactions on Software Engineering 13(11), |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.