source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Fractional%20ideal | In mathematics, in particular commutative algebra, the concept of fractional ideal is introduced in the context of integral domains and is particularly fruitful in the study of Dedekind domains. In some sense, fractional ideals of an integral domain are like ideals where denominators are allowed. In contexts where fractional ideals and ordinary ring ideals are both under discussion, the latter are sometimes termed integral ideals for clarity.
Definition and basic results
Let be an integral domain, and let be its field of fractions.
A fractional ideal of is an -submodule of such that there exists a non-zero such that . The element can be thought of as clearing out the denominators in , hence the name fractional ideal.
The principal fractional ideals are those -submodules of generated by a single nonzero element of . A fractional ideal is contained in if and only if it is an (integral) ideal of .
A fractional ideal is called invertible if there is another fractional ideal such that
where
is the product of the two fractional ideals.
In this case, the fractional ideal is uniquely determined and equal to the generalized ideal quotient
The set of invertible fractional ideals form an abelian group with respect to the above product, where the identity is the unit ideal itself. This group is called the group of fractional ideals of . The principal fractional ideals form a subgroup. A (nonzero) fractional ideal is invertible if and only if it is projective as an -module. Geometrically, this means an invertible fractional ideal can be interpreted as rank 1 vector bundle over the affine scheme .
Every finitely generated R-submodule of K is a fractional ideal and if is noetherian these are all the fractional ideals of .
Dedekind domains
In Dedekind domains, the situation is much simpler. In particular, every non-zero fractional ideal is invertible. In fact, this property characterizes Dedekind domains:
An integral domain is a Dedekind domain if and on |
https://en.wikipedia.org/wiki/Sprung%20mass | Sprung mass (or sprung weight), in a vehicle with a suspension, such as an automobile, motorcycle, or a tank, is the portion of the vehicle's total mass that is supported by the suspension, including in most applications approximately half of the weight of the suspension itself. The sprung mass typically includes the body, frame, the internal components, passengers, and cargo, but does not include the mass of the components at the other end of the suspension components (including the wheels, wheel bearings, brake rotors, calipers, and/or continuous tracks (also called caterpillar tracks), if any), which are part of the vehicle's unsprung mass.
The larger the ratio of sprung mass to unsprung mass, the less the body and vehicle occupants are affected by bumps, dips, and other surface imperfections such as small bridges. However, a large sprung mass to unsprung mass ratio can also be deleterious to vehicle control.
See also
Unsprung mass |
https://en.wikipedia.org/wiki/Arctic%20moss | Arctic moss is a common name for several plants and may refer to:
Calliergon giganteum, an aquatic moss
Cladonia, a genus of lichens |
https://en.wikipedia.org/wiki/Mass%20diffusivity | Diffusivity, mass diffusivity or diffusion coefficient is usually written as the proportionality constant between the molar flux due to molecular diffusion and the negative value of the gradient in the concentration of the species. More accurately, the diffusion coefficient times the local concentration is the proportionality constant between the negative value of the mole fraction gradient and the molar flux. This distinction is especially significant in gaseous systems with strong temperature gradients. Diffusivity derives its definition from Fick's law and plays a role in numerous other equations of physical chemistry.
The diffusivity is generally prescribed for a given pair of species and pairwise for a multi-species system. The higher the diffusivity (of one substance with respect to another), the faster they diffuse into each other. Typically, a compound's diffusion coefficient is ~10,000× as great in air as in water. Carbon dioxide in air has a diffusion coefficient of 16 mm2/s, and in water its diffusion coefficient is 0.0016 mm2/s.
Diffusivity has dimensions of length2 / time, or m2/s in SI units and cm2/s in CGS units.
Temperature dependence of the diffusion coefficient
Solids
The diffusion coefficient in solids at different temperatures is generally found to be well predicted by the Arrhenius equation:
where
D is the diffusion coefficient (in m2/s),
D0 is the maximal diffusion coefficient (at infinite temperature; in m2/s),
EA is the activation energy for diffusion (in J/mol),
T is the absolute temperature (in K),
R ≈ 8.31446J/(mol⋅K) is the universal gas constant.
Liquids
An approximate dependence of the diffusion coefficient on temperature in liquids can often be found using Stokes–Einstein equation, which predicts that
where
D is the diffusion coefficient,
T1 and T2 are the corresponding absolute temperatures,
μ is the dynamic viscosity of the solvent.
Gases
The dependence of the diffusion coefficient on temperature for gases can be expre |
https://en.wikipedia.org/wiki/Transversality%20theorem | In differential topology, the transversality theorem, also known as the Thom transversality theorem after French mathematician René Thom, is a major result that describes the transverse intersection properties of a smooth family of smooth maps. It says that transversality is a generic property: any smooth map , may be deformed by an arbitrary small amount into a map that is transverse to a given submanifold . Together with the Pontryagin–Thom construction, it is the technical heart of cobordism theory, and the starting point for surgery theory. The finite-dimensional version of the transversality theorem is also a very useful tool for establishing the genericity of a property which is dependent on a finite number of real parameters and which is expressible using a system of nonlinear equations. This can be extended to an infinite-dimensional parametrization using the infinite-dimensional version of the transversality theorem.
Finite-dimensional version
Previous definitions
Let be a smooth map between smooth manifolds, and let be a submanifold of . We say that is transverse to , denoted as , if and only if for every we have that
.
An important result about transversality states that if a smooth map is transverse to , then is a regular submanifold of .
If is a manifold with boundary, then we can define the restriction of the map to the boundary, as . The map is smooth, and it allows us to state an extension of the previous result: if both and , then is a regular submanifold of with boundary, and
.
Parametric transversality theorem
Consider the map and define . This generates a family of mappings . We require that the family vary smoothly by assuming to be a (smooth) manifold and to be smooth.
The statement of the parametric transversality theorem is:
Suppose that is a smooth map of manifolds, where only has boundary, and let be any submanifold of without boundary. If both and are transverse to , then for almost every , both and ar |
https://en.wikipedia.org/wiki/Comparison%20of%20screencasting%20software | This page provides a comparison of notable screencasting software, used to record activities on the computer screen. This software is commonly used for desktop recording, gameplay recording and video editing. Screencasting software is typically limited to streaming and recording desktop activity alone, in contrast with a software vision mixer, which has the capacity to mix and switch the output between various input streams.
Comparison by specification
Comparison by features
The following table compares features of screencasting software. The table has seven fields, as follows:
Product name: Product's name; sometime includes edition if a certain edition is targeted
Audio: Specifies whether the product supports recording audio commentary on the video
Entire desktop: Specifies whether product supports recording the entire desktop
OpenGL: Specifies whether the product supports recording from video games and software that employ OpenGL to render digital image
Direct3D: Specifies whether the product supports recording from video games or software that employ Direct3D to render digital image
Editing: Specifies whether the product supports editing recorded video at least to some small extent, such as cropping, trimming or splitting
Output: Specifies the file format in which the software saves the final video (audio output types are omitted)
See also
Comparison of webcam software |
https://en.wikipedia.org/wiki/Influence%20of%20the%20IBM%20PC%20on%20the%20personal%20computer%20market | Following the introduction of the IBM Personal Computer, or IBM PC, many other personal computer architectures became extinct within just a few years. It led to a wave of IBM PC compatible systems being released.
Before the IBM PC's introduction
Before the IBM PC was introduced, the personal computer market was dominated by systems using the 6502 and Z80 8-bit microprocessors, such as the TRS 80, Commodore PET, and Apple II series, which used proprietary operating systems, and by computers running CP/M. After IBM introduced the IBM PC, it was not until 1984 that IBM PC and clones became the dominant computers. In 1983, Byte forecast that by 1990, IBM would command only 11% of business computer sales. Commodore was predicted to hold a slim lead in a highly competitive market, at 11.9%.
Around 1978, several 16-bit CPUs became available. Examples included the Data General mN601, the Fairchild 9440, the Ferranti F100-L, the General Instrument CP1600 and CP1610, the National Semiconductor INS8900, Panafacom's MN1610, Texas Instruments' TMS9900, and, most notably, the Intel 8086. These new processors were expensive to incorporate in personal computers, as they used a 16-bit data bus and needed rare (and thus expensive) 16-bit peripheral and support chips.
More than 50 new business-oriented personal computer systems came on the market in the year before IBM released the IBM PC. Very few of them used a 16- or 32-bit microprocessor, as 8-bit systems were generally believed by the vendors to be perfectly adequate, and the Intel 8086 was too expensive to use.
Some of the main manufacturers selling 8-bit business systems during this period were:
Acorn Computers
Apple Computer Inc.
Atari Inc.
Commodore International
Cromemco
Digital Equipment Corporation
Durango Systems Inc.
Hewlett-Packard
InterSystems
Morrow Designs
North Star Computers
Ohio Scientific
Olivetti
Processor Technology
Sharp
South West Technical Products Corporation
Tandy Corporation
Zeni |
https://en.wikipedia.org/wiki/Relativistic%20chaos | In physics, relativistic chaos is the application of chaos theory to dynamical systems described primarily by general relativity, and also special relativity and also by Newton's law of segments and motion.
One of the earlier references on the topic is (Barrow 1982) and a particularly relevant result is that relativistic chaos is coordinate invariant (Motter 2003).
See also
Quantum chaos |
https://en.wikipedia.org/wiki/List%20of%20mathematical%20functions | In mathematics, some functions or groups of functions are important enough to deserve their own names. This is a listing of articles which explain some of these functions in more detail. There is a large theory of special functions which developed out of statistics and mathematical physics. A modern, abstract point of view contrasts large function spaces, which are infinite-dimensional and within which most functions are 'anonymous', with special functions picked out by properties such as symmetry, or relationship to harmonic analysis and group representations.
See also List of types of functions
Elementary functions
Elementary functions are functions built from basic operations (e.g. addition, exponentials, logarithms...)
Algebraic functions
Algebraic functions are functions that can be expressed as the solution of a polynomial equation with integer coefficients.
Polynomials: Can be generated solely by addition, multiplication, and raising to the power of a positive integer.
Constant function: polynomial of degree zero, graph is a horizontal straight line
Linear function: First degree polynomial, graph is a straight line.
Quadratic function: Second degree polynomial, graph is a parabola.
Cubic function: Third degree polynomial.
Quartic function: Fourth degree polynomial.
Quintic function: Fifth degree polynomial.
Sextic function: Sixth degree polynomial.
Rational functions: A ratio of two polynomials.
nth root
Square root: Yields a number whose square is the given one.
Cube root: Yields a number whose cube is the given one.
Elementary transcendental functions
Transcendental functions are functions that are not algebraic.
Exponential function: raises a fixed number to a variable power.
Hyperbolic functions: formally similar to the trigonometric functions.
Logarithms: the inverses of exponential functions; useful to solve equations involving exponentials.
Natural logarithm
Common logarithm
Binary logarithm
Power functions: raise a variable numb |
https://en.wikipedia.org/wiki/Split%20gene%20theory | The split gene theory is a theory of the origin of introns, long non-coding sequences in eukaryotic genes between the exons. The theory holds that the randomness of primordial DNA sequences would only permit small (< 600bp) open reading frames (ORFs), and that important intron structures and regulatory sequences are derived from stop codons. In this introns-first framework, the spliceosomal machinery and the nucleus evolved due to the necessity to join these ORFs (now "exons") into larger proteins, and that intronless bacterial genes are less ancestral than the split eukaryotic genes. The theory originated with Periannan Senapathy.
The theory provides solutions to key questions concerning the split gene architecture, including split eukaryotic genes, exons, introns, splice junctions, and branch points, based on the origin of split genes from random genetic sequences. It also provides possible solutions to the origin of the spliceosomal machinery, the nuclear boundary and the eukaryotic cell.
This theory led to the Shapiro–Senapathy algorithm, which provides the methodology for detecting the splice sites, exons and split genes in eukaryotic DNA, and which is the main method for detecting splice site mutations in genes that cause hundreds of diseases.
Split gene theory requires a separate origin of all eukaryotic species. It also requires that the simpler prokaryotes evolved from eukaryotes. This completely contradicts the scientific consensus about the formation of eukaryotic cells by endosymbiosis of bacteria. In 1994, Senapathy wrote a book about this aspect of his theory - The Independent Birth of Organisms. It proposed that all eukaryotic genomes were formed separately in a primordial pool. Dutch biologist Gert Korthoff criticized the theory by posing various problems that cannot be explained by a theory of independent origins. He pointed out that various eukaryotes need nurturing and called this the 'boot problem', in that even the initial eukaryote needed pa |
https://en.wikipedia.org/wiki/VESA%20BIOS%20Extensions | VESA BIOS Extensions (VBE) is a VESA standard, currently at version 3, that defines the interface that can be used by software to access compliant video boards at high resolutions and bit depths. This is opposed to the "traditional" INT 10h BIOS calls, which are limited to resolutions of 640×480 pixels with 16 colour (4-bit) depth or less. VBE is made available through the video card's BIOS, which installs during boot up some interrupt vectors that point to itself.
Most newer cards implement the more capable VBE 3.0 standard. Older versions of VBE provide only a real mode interface, which cannot be used without a significant performance penalty from within protected mode operating systems. Consequently, the VBE standard has almost never been used for writing a video card's drivers; each vendor has thus had to invent a proprietary protocol for communicating with its own video card. Despite this, it is common that a driver thunk out to the real mode interrupt in order to initialize screen modes and gain direct access to a card's linear frame buffer, because these tasks would otherwise require handling many hundreds of proprietary variations that exist from card to card.
In EFI 1.x systems, the INT 10H and the VESA BIOS Extensions (VBE) are replaced by the EFI UGA protocol. In widely used UEFI 2.x systems, the INT 10H and the VBE are replaced by the UEFI GOP.
Standards
Early VBE
VBE 1.0 (VS891001) was defined in 1989. VBE 1.1 (VS900602) was defined in 1990. VBE 1.2 (VS911022) was defined in 1991. These versions of VBE require real mode to work.
VBE defines several new functions called through INT 10H. The function numbering start with AX=4F00, or (AH=4F, AL=00), for Function 00h. 00h to 05h were defined in VBE 1.0, 06h and 07h in 1.1, and 08h in 1.2.
VESA BIOS Extensions (VBE core) 2.0 (November 1994)
This standard provides the primary functionality of the VESA BIOS Extensions. It allows applications to determine the capabilities of the graphics card and provide |
https://en.wikipedia.org/wiki/North%20South%20Atlantic%20Training%20Transect | The North South Atlantic Training Transect (NoSoAT) is a program developed by the Alfred Wegener Institute (AWI), the Strategic Marine Alliance for Research and Training (SMART), and the Partnership for Observation of the Global Oceans (POGO) to further the education and practical training of postgraduate students in climate and marine sciences. Each year, about 30 students are selected through a rigorous application process to join a voyage from Bremerhaven, Germany to Cape Town, South Africa aboard the RV Polarstern. The month-long course provides students with relevant lectures and projects, including hands-on training with atmospheric and oceanographic equipment, and instruction on data processing and analysis.
Results from Previous Cruises
Each cruise has a unique focus that varies from year to year as topics become more or less important. While the 2015 cruise had a strong focus on biological oceanography, the 2016 cruise was more focused on physical and chemical oceanography. These different approaches to studying marine and climate sciences allow groups of post-doctorate students to be specifically trained in their respective fields.
NoSoAT 2015 - Expeditions PS95.1 and PS95.2
The cruise track covered a broad range of habitats for various organisms and a broad range of characteristically different waters. Data collected came from equipment that included a Conductivity, Temperature, and Depth (CTD) Sonde; a Bongonet; a Ferrybox; and a Continuous Plankton Recorder (CPR). Samples of phytoplankton and zooplankton were obtained and analyzed in conjunction with satellite data. Preliminary results included the identification of three distinct water masses along the transect; discrepancies in the comparison of ship-board chlorophyll a measurements with satellite-derived chlorophyll a data; latitudinal variations in the depth of the chlorophyll a maximum; and variations in the impact of microzooplankton grazing on phytoplankton communities. |
https://en.wikipedia.org/wiki/Wilson%20prime | In number theory, a Wilson prime is a prime number such that divides , where "" denotes the factorial function; compare this with Wilson's theorem, which states that every prime divides . Both are named for 18th-century English mathematician John Wilson; in 1770, Edward Waring credited the theorem to Wilson, although it had been stated centuries earlier by Ibn al-Haytham.
The only known Wilson primes are 5, 13, and 563 . Costa et al. write that "the case is trivial", and credit the observation that 13 is a Wilson prime to . Early work on these numbers included searches by N. G. W. H. Beeger and Emma Lehmer, but 563 was not discovered until the early 1950s, when computer searches could be applied to the problem. If any others exist, they must be greater than 2 × 1013. It has been conjectured that infinitely many Wilson primes exist, and that the number of Wilson primes in an interval is about .
Several computer searches have been done in the hope of finding new Wilson primes.
The Ibercivis distributed computing project includes a search for Wilson primes. Another search was coordinated at the Great Internet Mersenne Prime Search forum.
Generalizations
Wilson primes of order
Wilson's theorem can be expressed in general as for every integer and prime . Generalized Wilson primes of order are the primes such that divides .
It was conjectured that for every natural number , there are infinitely many Wilson primes of order .
The smallest generalized Wilson primes of order are:
Near-Wilson primes
A prime satisfying the congruence with small can be called a near-Wilson prime. Near-Wilson primes with are bona fide Wilson primes. The table on the right lists all such primes with from up to 4.
Wilson numbers
A Wilson number is a natural number such that , where and where the term is positive if and only if has a primitive root and negative otherwise. For every natural number , is divisible by , and the quotients (called generalized Wilson quotien |
https://en.wikipedia.org/wiki/Methylated%20DNA%20immunoprecipitation | Methylated DNA immunoprecipitation (MeDIP or mDIP) is a large-scale (chromosome- or genome-wide) purification technique in molecular biology that is used to enrich for methylated DNA sequences. It consists of isolating methylated DNA fragments via an antibody raised against 5-methylcytosine (5mC). This technique was first described by Weber M. et al. in 2005 and has helped pave the way for viable methylome-level assessment efforts, as the purified fraction of methylated DNA can be input to high-throughput DNA detection methods such as high-resolution DNA microarrays (MeDIP-chip) or next-generation sequencing (MeDIP-seq). Nonetheless, understanding of the methylome remains rudimentary; its study is complicated by the fact that, like other epigenetic properties, patterns vary from cell-type to cell-type.
Background
DNA methylation, referring to the reversible methylation of the 5 position of cytosine by methyltransferases, is a major epigenetic modification in multicellular organisms. In mammals, this modification primarily occurs at CpG sites, which in turn tend to cluster in regions called CpG islands. There is a small fraction of CpG islands that can overlap or be in close proximity to promoter regions of transcription start sites. The modification may also occur at other sites, but methylation at either of these sites can repress gene expression by either interfering with the binding of transcription factors or modifying chromatin structure to a repressive state.
Disease condition studies have largely fueled the effort in understanding the role of DNA methylation. Currently, the major research interest lies in investigating disease conditions such as cancer to identify regions of the DNA that has undergone extensive methylation changes. The genes contained in these regions are of functional interest as they may offer a mechanistic explanation to the underlying genetic causes of a disease. For instance, the abnormal methylation pattern of cancer cells |
https://en.wikipedia.org/wiki/Food%20processing | Food processing is the transformation of agricultural products into food, or of one form of food into other forms. Food processing takes many forms, from grinding grain into raw flour, home cooking, and complex industrial methods used in the making of convenience foods. Some food processing methods play important roles in reducing food waste and improving food preservation, thus reducing the total environmental impact of agriculture and improving food security.
The Nova classification groups food according to different food processing techniques.
Primary food processing is necessary to make most foods edible while secondary food processing turns ingredients into familiar foods, such as bread. Tertiary food processing results in ultra-processed foods and has been widely criticized for promoting overnutrition and obesity, containing too much sugar and salt, too little fiber, and otherwise being unhealthful in respect to dietary needs of humans and farm animals.
Processing levels
Primary food processing
Primary food processing turns agricultural products, such as raw wheat kernels or livestock, into something that can eventually be eaten. This category includes ingredients that are produced by ancient processes such as drying, threshing, winnowing and milling grain, shelling nuts, and butchering animals for meat. It also includes deboning and cutting meat, freezing and smoking fish and meat, extracting and filtering oils, canning food, preserving food through food irradiation, and candling eggs, as well as homogenizing and pasteurizing milk.
Contamination and spoilage problems in primary food processing can lead to significant public health threats, as the resulting foods are used so widely. However, many forms of processing contribute to improved food safety and longer shelf life before the food spoils. Commercial food processing uses control systems such as hazard analysis and critical control points (HACCP) and failure mode and effects analysis (FMEA) to |
https://en.wikipedia.org/wiki/Cable%20bacteria | Cable bacteria are filamentous bacteria that conduct electricity across distances over 1 cm in sediment and groundwater aquifers. Cable bacteria allow for long-distance electron transport, which connects electron donors to electron acceptors, connecting previously separated oxidation and reduction reactions. Cable bacteria couple the reduction of oxygen or nitrate at the sediment's surface to the oxidation of sulfide in the deeper, anoxic, sediment layers.
Discovery
Long-distance electrical conductance in sediment was first observed in 2010 as a spatial separation of sulfide oxidation and oxygen reduction in marine sediment that was interrupted and re-established at a rate faster than could be explained by chemical diffusion. It was later found that this electrical conductance could be observed across a non-conductive layer of glass microspheres, where the only possible conductive structures were filamentous bacteria belonging to the family Desulfobulbaceae. The conductivity of single, live filaments was later demonstrated by observing the oxidation state of cytochromes using Raman microscopy. The same phenomenon was later observed in freshwater sediments and groundwater aquifers. Within a 15 cm thick top layer of sediment, cable bacteria densities providing total length of up to 2 km per square centimeter of surface have been observed.
Morphology
Cable bacteria filaments are 0.4—1.7 µm in diameter and up to 15 mm long. Filaments consist of rod-shaped cells with an average length of 3 µm. Filaments are long strings composed of cells stacked together, and can be as long as 30-70mm. Some filaments are composed of upwards of 10,000 cells. Each cell has between 15 and 54 ridges, and ridges span the entire length of the filament. These ridges are hypothesized to contain the cells' conductive structures.
Junctions
Cells in a filament are connected by junctions. The diameter of junctions between cells in the filament varies from being smaller than the cell diameter, th |
https://en.wikipedia.org/wiki/NDST1 | Bifunctional heparan sulfate N-deacetylase/N-sulfotransferase 1 is an enzyme. In humans, it is encoded by the NDST1 gene. |
https://en.wikipedia.org/wiki/Logo%20Software | Logo Software is a Turkish business software company based in Gebze, Kocaeli. The company was founded in 1984, and was officially established in 1986. Logo Yazılım has more than 1,300 employees and over 800 business partners in 7 different locations in 4 different countries. It started developing Logo Paas in 2015 as part of its servitization strategy and offers cloud applications through this platform. Logo Yazılım also provides digital transformation consultancy and special project management services.
In 2000, Logo Yazılım completed its initial public offering (IPO) and was the first IT company to go public in Turkey. The company is traded on Borsa Istanbul under the code LOGO and currently has a free float percentage rate of 66.37%. The company was the first software company to be included in "Turquality®", the Turkish Ministry of Trade's branding program. Logo ranked 67th on the list of "Turkey's Most Valuable Brands" according to the 2022 report by Brand Finance, which is a brand evaluation company.
History
Foundation and early years (1984 - 2000)
Launched on July 1, 1984, by eight young engineers, Logo Yazılım was one of Turkey's first start-ups and was officially founded and registered on March 5, 1986. In the same year, Logo introduced Logo Commercial System (LTS) to users and released many DOS and Windows-based products in the following years. The first version of Logo accounting programs, LKS - Logo Classic Series, is the predecessor of today's GO application. Following the LKS program running on the MS-DOS operating system, Logo Classic Series 2 marked the beginning of the current versions, in which Microsoft-based data is kept on Microsoft SQL Server. Logo Gold was Logo's first ERP application developed to be optimized with the MS-DOS operating system using the Btrieve database, and was launched in 1992. Following this application, Logo Gold Open, the predecessor of Tiger which is used today, was presented using Windows operating system and MS SQL.
|
https://en.wikipedia.org/wiki/Multilateral%20Interoperability%20Programme | The Multilateral Interoperability Programme (MIP) is an effort to deliver an assured capability for interoperability of information to support multinational, combined and joint operations. The MIP goal is to support all levels from corps to battalion. MIP's focus is on command and control systems. MIP is a consortium of 29 NATO and Non-NATO nations that meet quarterly to define interoperability specifications for the exchange information between their national Command and Control systems.
Overview
The Multilateral Interoperability Programme referred to as MIP, is an interoperability organisation established by national Command and Control Information Systems (C2IS) developers with a requirement to share relevant Command and Control information in a multinational or coalition environment.
As a result of collaboration within the programme, MIP produces a set of specifications which when implemented by the nations, provide the required interoperability capability.
MIP provides a venue for system level interoperability testing of national MIP implementations as well as providing a forum for exchanging information relevant to national implementation and fielding plans to enable synchronization.
MIP is NOT empowered to direct how nations develop their own C2IS.
The MIP meetings are held in Greding, Bavaria, Germany.
The NATO Data Administration Group (NDAG) cooperates with the MIP in building the Joint Consultation, Command and Control Information Exchange Data Model (JC3IEDM).
C2IEDM
The C2IEDM (the predecessor to the JC3IEDM), or Command and Control Information Exchange Data Model, is a data model that is managed by the Multilateral Interoperability Programme (MIP). It originated with experts from various NATO partners and from the Partnership for Peace nations. This data model is in the process of being submitted to Object Management Group (OMG) for consideration as the standard for Information exchange. It falls under the shared operational picture exchange s |
https://en.wikipedia.org/wiki/List%20of%20national%20animals | This is a list of countries that have officially designated one or more animals as their national animals.
National animal
See also
State animal
List of national birds
Animals as heraldic charges
Floral emblem
National personification |
https://en.wikipedia.org/wiki/Nuclear%20Fusion%20%28journal%29 | Nuclear Fusion is a peer reviewed international scientific journal that publishes articles, letters and review articles, special issue articles, conferences summaries and book reviews on the theoretical and practical research based on controlled thermonuclear fusion. The journal was first published in September, 1960 by IAEA and its head office was housed at the headquarter of IAEA in Vienna, Austria. Since 2002, the journal has been jointly published by IAEA and IOP Publishing.
The Nuclear Fusion Award
Since 2006, this award has been given each year to a particular paper of highest standard. The editorial board selects the recipient from all the research papers published in the Nuclear Fusion journal two years prior to the award year. The list of recipients are -
Tim Luce (2006)
Clemente Angioni, Max Planck Institute for Plasma Physics, Germany (2007)
Todd Evans (2008)
Steve Sabbagh, Columbia University, USA (2009)
John Rice, Massachusetts Institute of Technology (2010)
Hajime Urano, JAEA, Japan (2011)
Pat Diamond, University of California at San Diego (2012)
Dennis Whyte, Massachusetts Institute of Technology, USA (2013)
Phil Snyder, General Atomics, USA (2014)
Robert Goldston, Princeton Plasma Physics Laboratory, USA (2015)
Sebastijan Brezinsek, EUROfusion Consortium and Forschungszentrum Jülich, Germany (2016)
Francois Ryter, Max Planck Institute for Plasma Physics, Germany (2017)
A. Kallenbach, Max Planck Institute for Plasma Physics, Germany (2018)
N.T. Howard, Massachusetts Institute of Technology, USA (2019)
C. Theiler, Swiss Plasma Center, Switzerland (2020) |
https://en.wikipedia.org/wiki/Rayleigh%20dissipation%20function | In physics, the Rayleigh dissipation function, named after Lord Rayleigh, is a function used to handle the effects of velocity-proportional frictional forces in Lagrangian mechanics. If the frictional force on a particle with velocity can be written as , the Rayleigh dissipation function can be defined for a system of particles as
This function represents half of the rate of energy dissipation of the system through friction. The force of friction is negative the velocity gradient of the dissipation function, , analogous to a force being equal to the negative position gradient of a potential. This relationship is represented in terms of the set of generalized coordinates as
.
As friction is not conservative, it is included in the term of Lagrange's equations,
.
Applying of the value of the frictional force described by generalized coordinates into the Euler-Lagrange equations gives
. |
https://en.wikipedia.org/wiki/Amorphous%20computing | Amorphous computing refers to computational systems that use very large numbers of identical, parallel processors each having limited computational ability and local interactions. The term Amorphous Computing was coined at MIT in 1996 in a paper entitled "Amorphous Computing Manifesto" by Abelson, Knight, Sussman, et al.
Examples of naturally occurring amorphous computations can be found in many fields, such as: developmental biology (the development of multicellular organisms from a single cell), molecular biology (the organization of sub-cellular compartments and intra-cell signaling), neural networks, and chemical engineering (non-equilibrium systems) to name a few. The study of amorphous computation is hardware agnostic—it is not concerned with the physical substrate (biological, electronic, nanotech, etc.) but rather with the characterization of amorphous algorithms as abstractions with the goal of both understanding existing natural examples and engineering novel systems.
Amorphous computers tend to have many of the following properties:
Implemented by redundant, potentially faulty, massively parallel devices.
Devices having limited memory and computational abilities.
Devices being asynchronous.
Devices having no a priori knowledge of their location.
Devices communicating only locally.
Exhibit emergent or self-organizational behavior (patterns or states larger than an individual device).
Fault-tolerant, especially to the occasional malformed device or state perturbation.
Algorithms, tools, and patterns
(Some of these algorithms have no known names. Where a name is not known, a descriptive one is given.)
"Fickian communication". Devices communicate by generating messages which diffuse through the medium in which the devices dwell. Message strength will follow the inverse square law as described by Fick's law of diffusion. Examples of such communication are common in biological and chemical systems.
"Link diffusive communication". Devices com |
https://en.wikipedia.org/wiki/Pencil%20sharpener | A pencil sharpener (or pencil pointer, or in Ireland a parer or topper) is a tool for sharpening a pencil's writing point by shaving away its worn surface. Pencil sharpeners may be operated manually or by an electric motor. It is common for many sharpeners to have a casing around them, which can be removed for emptying the pencil shavings debris into a bin.
History
Before the development of dedicated pencil sharpeners, a pencil was sharpened by whittling with a knife.
The development of pencil sharpeners began in France when a French book from 1822 reported in detail about an invention of Mr. C. A. Boucher (Paris) for the construction of a pencil sharpener. He was working with pantographs and apparently needed a device to precisely sharpen the pencils. The device of Mr. Boucher was technically sensible and functional. His idea was also internationally known and recognized, as shown by corresponding reports in German literature at this time. But Mr. Boucher had not applied a patent for his pencil sharpener. Commercial use of his inventions is unlikely.
French mathematician Bernard Lassimonne (Limoges) applied for the world's first patent (French patent #2444) on a pencil sharpener in 1828. Pencil sharpener devices using his patent were actually produced and sold by Binant, a shop for painting accessories in Paris. In 1833 in England, Cooper & Eckstein patented the so-called Styloxynon, a simple device consisting of two sharp files set together at right angle in a small block of rosewood. This is the oldest pencil sharpener that has surviving examples.
In the 1830s and 1840s, some French people, all based in Paris, were engaged in construction of simple pencil sharpening tools, like François Joseph Lahausse. These devices were partially sold, but without supra-regional significance. In 1847 the French nobleman Thierry des Estivaux invented a simple hand-held pencil sharpener in its recognizable modern form. The first American pencil sharpener was patented by Wa |
https://en.wikipedia.org/wiki/While%20loop | In most computer programming languages, a while loop is a control flow statement that allows code to be executed repeatedly based on a given Boolean condition. The while loop can be thought of as a repeating if statement.
Overview
The while construct consists of a block of code and a condition/expression. The condition/expression is evaluated, and if the condition/expression is true, the code within all of their following in the block is executed. This repeats until the condition/expression becomes false. Because the while loop checks the condition/expression before the block is executed, the control structure is often also known as a pre-test loop. Compare this with the do while loop, which tests the condition/expression after the loop has executed.
For example, in the C programming language (as well as Java, C#, Objective-C, and C++, which use the same syntax in this case), the code fragment
int x = 0;
while (x < 5) {
printf ("x = %d\n", x);
x++;
}
first checks whether x is less than 5, which it is, so then the {loop body} is entered, where the printf function is run and x is incremented by 1. After completing all the statements in the loop body, the condition, (x < 5), is checked again, and the loop is executed again, this process repeating until the variable x has the value 5.
Note that it is possible, and in some cases desirable, for the condition to always evaluate to true, creating an infinite loop. When such a loop is created intentionally, there is usually another control structure (such as a break statement) that controls termination of the loop.
For example:
while (true) {
// do complicated stuff
if (someCondition)
break;
// more stuff
}
Demonstrating while loops
These while loops will calculate the factorial of the number 5:
ActionScript 3
var counter: int = 5;
var factorial: int = 1;
while (counter > 1) {
factorial *= counter;
counter--;
}
Printf("Factorial = %d", factorial);
Ada
with Ada.Integer_Text_IO; |
https://en.wikipedia.org/wiki/Mixminion | Mixminion is the standard implementation of the Type III anonymous remailer protocol. Mixminion can send and receive anonymous e-mail.
Mixminion uses a mix network architecture to provide strong anonymity, and prevent eavesdroppers and other attackers from linking senders and recipients. Volunteers run servers (called "mixes") that receive messages, decrypt them, re-order them, and re-transmit them toward their eventual destination. Every e-mail passes through several mixes so that no single mix can link message senders with recipients.
To send an anonymous message, mixminion breaks it into uniform-sized chunks (also called "packets"), pads the packets to a uniform size, and chooses a path through the mix network for each packet. The software encrypts every packet with the public keys for each server in its path, one by one. When it is time to transmit a packet, mixminion sends it to the first mix in the path. The first mix decrypts the packet, learns which mix will receive the packet, and relays it. Eventually, the packet arrives at a final (or "exit") mix, which sends it to the chosen recipient. Because no mix sees any more of the path besides the immediately adjacent mixes, they cannot link senders to recipients.
Mixminion supports Single-Use Reply Blocks (or SURBs) to allow anonymous recipients. A SURB encodes a half-path to a recipient, so that each mix in the sequence can unwrap a single layer of the path, and encrypt the message for the recipient. When the message reaches the recipient, the recipient can decode the message and learn which SURB was used to send it; the sender does not know which recipient has received the anonymous message.
The most current version of Mixminion Message Sender is 1.2.7 and was released on 11 February 2009.
On 2 September 2011, a news announcement was made that stated the source was uploaded to GitHub
See also
Anonymity
Anonymous P2P
Anonymous remailer
Cypherpunk anonymous remailer (Type I)
Mixmaster anonymous remail |
https://en.wikipedia.org/wiki/Image-based%20flow%20visualization | In scientific visualization, image-based flow visualization (or visualisation) is a computer modelling technique developed by Jarke van Wijk to visualize two dimensional flows of liquids such as water and air, like the wind movement of a tornado. Compared with integration techniques it has the advantage of producing a whole image at every step, as the technique relies upon graphical computing methods for frame-by-frame capture of the model of advective transport of a decaying dye. It is a method from the texture advection family.
Principle
The core idea is to create a noise texture on a regular grid and then bend this grid according to the flow (the vector field). The bent grid is then sampled at the original grid locations. Thus, the output is a version of the noise, that is displaced according to the flow.
The advantage of this approach is that it can be accelerated on modern graphics hardware, thus allowing for real-time or almost real-time simulation of 2D flow data. This is particularly handy if one wants to visualise multiple scaled versions of the vector field to first gain an overview and then concentrate on the details. |
https://en.wikipedia.org/wiki/Ragon%20Institute | Phillip T. and Susan M. Ragon Institute is a medical institute founded in 2009 at the Massachusetts General Hospital (MGH) by the funding from founder and CEO of InterSystems Phillip Ragon and his wife Susan Ragon to find vaccines for diseases of the immune system, particularly HIV/AIDS. The institute includes scientists from Harvard University, the Massachusetts Institute of Technology (MIT) and the MGH, a level I trauma center which is the largest teaching hospital affiliated with Harvard Medical School.
The Phillip T. and Susan M. Ragon Institute, or the Ragon Institute of MGH, MIT and Harvard, was officially established in February 2009 with a dual mission: to contribute to the accelerated discovery of an HIV/AIDS vaccine and subsequently to establish itself as a world leader in the collaborative study of immunology.
Aims
Founded with a commitment of $100 million from the Ragons (Phillip T. aka Terry, and Susan), the institute is structured and positioned to significantly contribute to a global effort to successfully develop an HIV/AIDS vaccine through:
creating non-traditional partnerships among experts with different but complementary backgrounds (e.g., engineers, basic immunologists, computational biologists, immunogeneticists, clinicians);
providing a means for rapidly funding promising studies (e.g., elite controllers, innovative viral vectors) and emerging concepts in the field (e.g., innate immune system memory);
integrating key facets of current vaccine development efforts that have tended to follow separate tracks (e.g., seeking a combined antibody and T-cell solution);
providing a substantial pool of accessible, flexible funding that will help lower the threshold for scientists to pursue risky, unconventional avenues of study that are unlikely to attract funding from traditional sources. Such funding will encourage innovation, compress the time it takes to conduct bench-to-bedside research and attract new minds to the field.
The Ragon Institute' |
https://en.wikipedia.org/wiki/Littlewood%20subordination%20theorem | In mathematics, the Littlewood subordination theorem, proved by J. E. Littlewood in 1925, is a theorem in operator theory and complex analysis. It states that any holomorphic univalent self-mapping of the unit disk in the complex numbers that fixes 0 induces a contractive composition operator on various function spaces of holomorphic functions on the disk. These spaces include the Hardy spaces, the Bergman spaces and Dirichlet space.
Subordination theorem
Let h be a holomorphic univalent mapping of the unit disk D into itself such that h(0) = 0. Then the composition operator Ch defined on holomorphic functions f on D by
defines a linear operator with operator norm less than 1 on the Hardy spaces , the Bergman spaces .
(1 ≤ p < ∞) and the Dirichlet space .
The norms on these spaces are defined by:
Littlewood's inequalities
Let f be a holomorphic function on the unit disk D and let h be a holomorphic univalent mapping of D into itself with h(0) = 0. Then
if 0 < r < 1 and 1 ≤ p < ∞
This inequality also holds for 0 < p < 1, although in this case there is no operator interpretation.
Proofs
Case p = 2
To prove the result for H2 it suffices to show that for f a polynomial
Let U be the unilateral shift defined by
This has adjoint U* given by
Since f(0) = a0, this gives
and hence
Thus
Since U*f has degree less than f, it follows by induction that
and hence
The same method of proof works for A2 and
General Hardy spaces
If f is in Hardy space Hp, then it has a factorization
with fi an inner function and fo an outer function.
Then
Inequalities
Taking 0 < r < 1, Littlewood's inequalities follow by applying the Hardy space inequalities to the function
The inequalities can also be deduced, following , using subharmonic functions. The inequaties in turn immediately imply the subordination theorem for general Bergman spaces.
Notes |
https://en.wikipedia.org/wiki/Agaricus%20lanatoniger | Agaricus lanatoniger is an agaric fungus in the family Agaricaceae, endemic to New Zealand.
Taxonomy
A. lanatoniger was first described in 1974 by Belgian mycologist Paul Heinemann and collected by Egon Horak in December 1967. The holotype specimen was collected in the Westland Province, of New Zealand by Lake Haupiri, underneath red beech (Nothofagus fusca) and rimu (Dacrydium cupressinum) trees. The original paper reference number was incorrect, but is correctly listed as PDD 27107 in a report on New Zealand Agaricus species in 1999.
Description
The pileus of Agarcius lanatoniger can vary from a spherical to a convex shape. Smaller specimens tend to have more spherical pileus, while larger are more flattened convex shape, although both have round shape when viewed from above. The dark brown, felt-like pileus or cap can be up to wide in diameter.
The gills consist of thin pink filaments, stemming from the underside of the pileus without touching the stem. This forms a small ring around the stem less than long. About a third of the length of the stem is a thick skirt. This extends out from the stem. Above the skirt, the stem is tan or pale colour. While underneath, the colour transitions from light brown to dark brown or black like the cap's colour.
The stem ranges from in length and with a diameter of , generally thicker toward the base. Inside the stem is a white, hollow column beginning at gill level but sealed at the bottom. Beneath the ground, the bulbous shape has many small root-like filaments. The stem's insides are white with a hollow center.
Similarity to Agaricus purpureoniger
The sequence of A. lanatoniger, when compared to A. purpureoniger differed only by one nucleotide, suggesting that they be the same species. When physically compared, the A. purpureoniger is more purple. However, over the last three decades, all samples of A. purpureoniger have been found in similar locations as A. lanatoniger, specifically in the northwestern regions o |
https://en.wikipedia.org/wiki/Coded%20exposure%20photography | Coded exposure photography, also known as a flutter shutter, is the name given to any mathematical algorithm that reduces the effects of motion blur in photography. The key element of the coded exposure process is the mathematical formula that affects the shutter frequency. This involves the calculation of the relationship between the photon exposure of the light sensor and the randomized code. The camera is made to take a series of snapshots with random time intervals using a simple computer, this creates a blurred image that can be reconciled into a clear image using the algorithm.
Motion de-blurring technology grew due to increasing demand for clearer images in sporting events and other digital media. The relative inexpensiveness of the coded exposure technology makes it a viable alternative to expensive cameras and equipment that are built to take millions of images per second.
History
Photography was developed to enable imaging of the visible world. Early cameras utilized film made of plastic coated with compounds of silver. The film is highly sensitive to light. When photons (light) hit the film a reaction occurs which semi-permanently stores the data on its surface. This film is then developed by exposing it to several chemicals to create the image. The film is highly sensitive and the process is complicated. It must be stored away from light to prevent spoilage.
Digital cameras utilize digital technologies to create images. This process involves exposing light-sensitive material to photons, creating electrical signals that are recorded in computer files. This process is simple and has improved the availability of photography. One problem that digital cameras have faced is motion blur. Motion blur occurs when the camera or the subject are in motion. When motion blur happens, the resulting image is blurry, fuzzy edges and indistinct features. One solution to remove motion blur in photography is to increase the shutter speed of the camera. Unlike the coded |
https://en.wikipedia.org/wiki/Capsela | Capsela is a construction toy brand consisting primarily of gears and motors in spherical plastic capsules that can be connected to form various static or dynamic toys suitable for land or water. The capsules typically have six hollow octagonal connectors protruding, where an octagonal sleeve piece bridges between two capsules. The hollow connection pegs on a capsule can have electrical or rotary adapters inside, reaching into the next capsule. There are electric motor capsules, gear capsules, switch capsules, pontoon capsules, and others; there are also battery holders, wheels, propellers, impellers, lights, wires, and miscellaneous supporting pieces.
Capsela products were originally manufactured by the Mitsubishi Pencil Company.
Product line
Capsela products were originally sold in at least four ranges:
A series of increasingly complex generic construction sets, similar in style to Meccano or Lego Technic. (Sets 135, 150, 200, 250, 400, 450, 500, 700 and 1000, ranging from 30 to 108 parts.)
Capsela Computer, the flagship range, based around a multi-function computer capsule capable of controlling motors and lights. Sets CRC2000 and ICR5000, the latter featuring an infrared remote control.
Capsela Voice Command systems (3000 and 6000) have a small computer capable of responding to a number of voice commands. The 6000 system has a rudimentary wireless infrared remote control.
A sister product branded as SpaceLink, has no motorized parts (with the exception of spring powered wheel units) and instead focuses on science fiction-themed accessories (cockpits, small action figures, rockets, etc.). This range appears to have been designed to be more appealing to children than the more educational original products. This includes sets 330, 345, 360, 610, 650, 670, 690, 805, 820, 835, 850, 860 and 895, ranging from 8 to 53 parts.
Capsela Powertram, a series designed to bridge Capsela and Spacelink, containing motor-driven construction components and Spacelink accesso |
https://en.wikipedia.org/wiki/8chan | 8kun, previously called 8chan, Infinitechan or Infinitychan (stylized as ∞chan), is an imageboard website composed of user-created message boards. An owner moderates each board, with minimal interaction from site administration. The site has been linked to white supremacism, neo-Nazism, the alt-right, racism and antisemitism, hate crimes, and multiple mass shootings. The site has been known to host child pornography; as a result, it was filtered out from Google Search in 2015. Several of the site's boards played an active role in the Gamergate harassment campaign, encouraging Gamergate affiliates to frequent 8chan after 4chan banned the topic. 8chan is the home of the discredited QAnon conspiracy theory.
Shortly before the 2019 El Paso shooting, a four-page message justifying the attack was posted to the site, and police have stated that they are "reasonably confident" it was posted by the perpetrator. In the aftermath of the back-to-back mass shootings on August 3 in El Paso and August 4 in Dayton, Ohio, respectively, the site was taken off clearnet on August 5, 2019, when network infrastructure provider Cloudflare stopped providing their content delivery network (CDN) service. Voxility, a web services company that had been renting servers to Epik, the site's new domain registrar, as well as Epik's CDN provider subsidiary BitMitigate, also terminated service. After several attempts to return to clearnet were ultimately stymied by providers denying service to 8chan, the site returned to the clearnet as 8kun in November 2019 through a Russian hosting provider.
History
8chan was created in October 2013 by computer programmer Fredrick Brennan. Brennan created the website after observing what he perceived to be rapidly escalating surveillance and a loss of free speech on the Internet. Brennan, who considered the imageboard 4chan to have grown into authoritarianism, described 8chan as a "free-speech-friendly" alternative, and originally conceptualized the site while e |
https://en.wikipedia.org/wiki/Quantum%20point%20contact | A quantum point contact (QPC) is a narrow constriction between two wide electrically conducting regions, of a width comparable to the electronic wavelength (nano- to micrometer).
The importance of QPC lies in the fact that they prove quantisation of ballistic conductance in mesoscopic systems. The conductance of a QPC is quantized in units of , the so-called conductance quantum.
Quantum point contacts were first reported in 1988 by a Dutch team from Delft University of Technology and Philips Research and, independently, by a British team from the Cavendish Laboratory. They are based on earlier work by the British group which showed how split gates could be used to convert a two-dimensional electron gas into one-dimension, first in silicon and then in gallium arsenide.
This quantisation is reminiscent of the quantisation of the Hall conductance, but is measured in the absence of a magnetic field. The zero-field conductance quantisation and the smooth transition to the quantum Hall effect on applying a magnetic field are essentially consequences of the equipartition of current among an integer number of propagating modes in the constriction.
Fabrication
There are several different ways of fabricating a quantum point contact. It can be realized in a break-junction by pulling apart a piece of conductor until it breaks. The breaking point forms the point contact. In a more controlled way, quantum point contacts are formed in a two-dimensional electron gas (2DEG), e.g. in GaAs/AlGaAs heterostructures. By applying a voltage to suitably shaped gate electrodes, the electron gas can be locally depleted and many different types of conducting regions can be created in the plane of the 2DEG, among them quantum dots and quantum point contacts. Another means of creating a QPC is by positioning the tip of a scanning tunneling microscope close to the surface of a conductor.
Properties
Geometrically, a quantum point contact is a constriction in the transverse direction which pr |
https://en.wikipedia.org/wiki/REPAIRtoire | REPAIRtoire is a database of resources for systems biology of DNA damage and repair.
See also
DNA repair |
https://en.wikipedia.org/wiki/Avaya%20VSP%207000%20Series | Avaya Virtual Services Platform 7000 Series or VSP 7000 is a set standalone/stackable switches, used in enterprise data networks, and data centers, manufactured by Avaya. This product is primarily offered to satisfy the Top-of-Rack (ToR) role for server farms and virtualized data centers. It supports Avaya's extended Shortest Path Bridging (SPB) implementation "Fabric Connect", and is future-ready for Edge Virtual Bridging (EVB) – IEEE 802.1Qbg, and Fiber Channel over Ethernet (FCoE). The system incorporates fifth generation application-specific integrated circuit (ASIC) chips with redundant and hot-swappable power supplies, fans, and expansion modules. The VSP 7000's unique architecture allows it to be meshedfully or partiallywith like devices, creating a high-capacity, low-latency network of up to 500 units, supporting up to 16,000 ports of 10GbE supported by a virtual backplane of up to 280 Tbit/s
History
In November 2010, Avaya introduced its Virtual Enterprise Network Architecture (VENA) and in May 2011 this Switch, featuring 24 ports of fixed 10 Gigabit Ethernet was released as part of the VENA strategy. The VSP 7000 is future-ready for 40GbE and 100GbE by virtue of the Media Dependent Adapter (MDA) slot on the front panel, enabling in-service deployment of high-speed connections. The Switch is also future-ready for storage area networking. The company gained the foundational technology for the VENA strategy through its $915 million acquisition of Nortel Enterprise Solutions business unit in December 2009. The Switch supports 24 fixed ports 1 or 10 Gigabit via SFP+ sockets, either front-to-back or back-to-front cooling, field replaceable fan trays, and field replaceable redundant AC or DC power supplies. The Application-specific integrated circuit (ASIC) technology allows the MDA to support additional 10GbE ports (8-port SFP+ and 10GBASE-T currently available, increasing port density to 32 ports), with future 40GbE (2-port), and 100GbE (1-port) planned. |
https://en.wikipedia.org/wiki/Physalaemin | Physalaemin is a tachykinin peptide obtained from the Physalaemus frog, closely related to substance P. Its structure was first elucidated in 1964.
Like all tachykinins, physalaemin is a sialagogue (increases salivation) and a potent vasodilator with hypotensive effects.
Structure
Physalaemin (PHY) is known to take on both a linear and helical three dimensional structure. Grace et al. (2010) have shown that in aqueous environments, PHY preferentially takes on the linear conformation whereas in an environment that simulates a cellular membrane, PHY takes on a helical confirmation from the Pro4 residue to the C-Terminus. This helical conformation is essential to allow the binding of PHY to neurokinin-1 (NK1) receptors. Consensus sequences between Substance P (a mammalian tachykinin and agonist of NK1) and PHY have been used to confirm that the helical confirmation is necessary for PHY to bind to NK1.
Use In Research
Not only is PHY closely related to Substance P (SP), but it also has a higher affinity for the mammalian neurokinin receptors that Substance P can bind to. Researchers can make use of this behavior of PHY to study the behavior of smooth muscle - a tissue where NK1 can be found. Shiina et al. (2010) used PHY to show that tachykinins as a whole can cause the longitudinal contraction of smooth muscle tissue in esophageal tissue.
Singh et Maji made use of PHY's similarity to SP along with its sequence similarity to Amyloid B-peptide 25-35 [AB(25-35)]. Despite its sequence similarity to SP, Singh et Maji showed that PHY had distinct amyloid forming capabilities . Under artificially elevated concentrations of tetrafluoroethylene (TFE) and a short incubation time, PHY was able to form amyloid fibrils. These fibrils originating from tackynins like PHY were also shown to reduce the neurotoxicity of other Amyloid fibers associated with amyloid induced diseases such as Alzheimer's disease. |
https://en.wikipedia.org/wiki/Uncompressed%20video | Uncompressed video is digital video that either has never been compressed or was generated by decompressing previously compressed digital video. It is commonly used by video cameras, video monitors, video recording devices (including general-purpose computers), and in video processors that perform functions such as image resizing, image rotation, deinterlacing, and text and graphics overlay. It is conveyed over various types of baseband digital video interfaces, such as HDMI, DVI, DisplayPort and SDI. Standards also exist for the carriage of uncompressed video over computer networks.
Some HD video cameras output uncompressed video, whereas others compress the video using a lossy compression method such as MPEG or H.264. In any lossy compression process, some of the video information is removed, which creates compression artifacts and reduces the quality of the resulting decompressed video. When editing video, it is preferred to work with video that has never been compressed (or was losslessly compressed) as this maintains the best possible quality, with compression performed after completion of editing.
Uncompressed video should not be confused with raw video. Raw video represents largely unprocessed data (e.g. without demosaicing) captured by an imaging device.
Recording
A standalone video recorder is a device that receives uncompressed video and stores it in either uncompressed or compressed form. These devices typically have a video output that can be used to monitor or playback recorded video. When playing back compressed video, the compressed video is uncompressed by the device before being output. Such devices may also have a communication interface, such as Ethernet or USB, which can used to exchange video files with an external computer, and in some cases control the recorder from an external computer as well.
Recording to a computer is a relatively inexpensive alternative to implementing a digital video recorder, but the computer and its video storage d |
https://en.wikipedia.org/wiki/Null%20dust%20solution | In mathematical physics, a null dust solution (sometimes called a null fluid) is a Lorentzian manifold in which the Einstein tensor is null. Such a spacetime can be interpreted as an exact solution of Einstein's field equation, in which the only mass–energy present in the spacetime is due to some kind of massless radiation.
Mathematical definition
By definition, the Einstein tensor of a null dust solution has the form
where is a null vector field. This definition makes sense purely geometrically, but if we place a stress–energy tensor on our spacetime of the form
,
then Einstein's field equation is satisfied, and such a stress–energy tensor has a clear physical interpretation in terms of massless radiation. The vector field specifies the direction in which the radiation is moving; the scalar multiplier specifies its intensity.
Physical interpretation
Physically speaking, a null dust describes either gravitational radiation, or some kind of nongravitational radiation which is described by a relativistic classical field theory (such as electromagnetic radiation), or a combination of these two. Null dusts include vacuum solutions as a special case.
Phenomena which can be modeled by null dust solutions include:
a beam of neutrinos assumed for simplicity to be massless (treated according to classical physics),
a very high-frequency electromagnetic wave,
a beam of incoherent electromagnetic radiation.
In particular, a plane wave of incoherent electromagnetic radiation is a linear superposition of plane waves, all moving in the same direction but having randomly chosen phases and frequencies. (Even though the Einstein field equation is nonlinear, a linear superposition of comoving plane waves is possible.)
Here, each electromagnetic plane wave has a well defined frequency and phase, but the superposition does not. Individual electromagnetic plane waves are modeled by null electrovacuum solutions, while an incoherent mixture can be modeled by a null dust.
|
https://en.wikipedia.org/wiki/Committee%20to%20End%20Pay%20Toilets%20in%20America | The Committee to End Pay Toilets in America, or CEPTIA, was a 1970s grass-roots political organization which was one of the main forces behind the elimination of pay toilets in many American cities and states.
History
Founded in 1970 by then-nineteen-year-old Ira Gessel, the Committee's purpose was to "eliminate pay toilets in the U.S. through legislation and public pressure."
Starting a national crusade to cast away coin-operated commodes, Gessel told newsmen, "You can have a fifty-dollar bill, but if you don't have a dime, that metal box is between you and relief." Membership in the organization cost only $0.25, and members received the Committee's newsletter, the Free Toilet Paper. Headquartered in Dayton, Ohio, USA, the group had as many as 1,500 members, in seven chapters.
The group also sponsored the Thomas Crapper Memorial Award, which was given to "the person who has made an outstanding contribution to the cause of CEPTIA and free toilets."
In 1973, Chicago became the first American city to act when the city council voted 37–8 in support of a ban on pay toilets in that city. According to at least one source, this was "... a direct response, evidently," to CEPTIA.
Achievements
According to The Wall Street Journal, there were, in 1974, at least 50,000 pay toilets in America, mostly made by the Nik-O-Lok Company. Despite this flourishing commerce, CEPTIA was successful over the next few years in obtaining bans in New York, New Jersey, Minnesota, California, Florida, and Ohio. Lobbying was so successful that by June 1976, twelve states had enacted bans and the group announced that it was disbanding, declaring its mission mostly achieved.
Criticism
While CEPTIA's campaign was successful in largely eliminating pay toilets in the United States, critics charge that the result was not a flourishing of free public toilets, but rather many fewer public toilets of any sort than in other countries that did not see a movement against pay toilets. In recent y |
https://en.wikipedia.org/wiki/Society%20for%20Evidence-Based%20Gender%20Medicine | The Society For Evidence-Based Gender Medicine (SEGM) is a non-profit organization that is known for opposing standards of care for transgender youth and engaging in political lobbying. The group routinely cites the theory of rapid-onset gender dysphoria and has claimed that conversion therapy can only be practiced on the basis of sexual orientation rather than gender identity. SEGM is often cited in anti-transgender legislation and court cases, sometimes providing evidence briefs themselves. It is not officially recognized as a scientific organization by the international medical community.
Researchers at the Yale School of Medicine issued a report which described SEGM as a small group of anti-trans activists. Joshua Safer, a spokesperson for the Endocrine Society, described them as outside the medical mainstream.
SEGM is closely affiliated with Genspect: seven advisors to SEGM are on Genspect's team of advisors, including Stella O'Malley, Genspect's founder.
Activities and positions
William Malone, a founder of SEGM, has opposed the informed consent model for transgender healthcare, where adults older than 18 can start hormones after signing an informed consent document without requiring an evaluation by a mental health professional. He told Medscape that "cognitive maturity doesn’t occur until the age of 25."
SEGM made a submission in defense of the state of Arizona's ban on Medicaid coverage for transgender healthcare. In it, they advanced the controversial idea of rapid-onset gender dysphoria (ROGD), which suggests a subtype of gender dysphoria caused by peer influence and social contagion. ROGD has been condemned as unevidenced and nonscientific by the majority of the worlds' major psychological bodies. Lambda Legal and Cooley LLP filed an amicus brief opposing the ban on behalf of LGBT advocacy organizations such as PFLAG, the Southern Arizona Gender Alliance, and the TransActive Gender Project. The Pediatric Endocrine Society and the World Professional |
https://en.wikipedia.org/wiki/Comparison%20of%20mobile%20operating%20systems | This is a comparison on mobile operating systems. Only the latest versions are shown in the table below, even though older versions may still be marketed.
About OS
Advanced controls
Accessibility features
App ecosystem
Browser
Basic features
Communication and connectivity
Language and inputs
Maps and navigation
Media playback and controls
Peripheral support
Photo and video
Productivity
Ringtones and alerts
Security and privacy
Sound and voice
Other features
See also
Comparison of open-source mobile phones
List of custom Android distributions
Comparison of satellite navigation software |
https://en.wikipedia.org/wiki/Lifeyo | LIFEYO is a website hosting company founded in Newport Beach, CA. Lifeyo allows users to make websites without using HTML or managing a hosting account.
Lifeyo competes with other website builders such as Yola, Jimdo, Webs, Weebly, and Wix.
Features
Lifeyo offers templates that allow users to get started building their website quickly. Users can choose from different website types (business, personal, portfolio, and blogging) that provide a starting point with placeholder content. The WYSIWYG website builder allows users to customize website pages in a visual manner that does not require any HTML coding expertise. Users of Lifeyo can build websites using the blogging interface and photo gallery features. A feature that was unique to Lifeyo is the ability to collaborate on a website with multiple editors. The service is free and does not apply ads on websites created by users. For a fee the company offers an upgrade to a domain name.
History
Lifeyo was released to the public in 2010 at the South By Southwest Interactive conference in Austin, Texas. It was a startup launched by Mike Kai and Wiwat Ruengmee. Lifeyo has offices based in Newport Beach, California and Bangkok, Thailand.
Reception
In 2010, Erez Zukerman of AOL News gave a positive review to Lifeyo. Zukerman described Lifeyo as an "Incredible content-management system for someone just trying to put together their first website.” |
https://en.wikipedia.org/wiki/History%20of%20Linux | Linux began in 1991 as a personal project by Finnish student Linus Torvalds to create a new free operating system kernel. The resulting Linux kernel has been marked by constant growth throughout its history. Since the initial release of its source code in 1991, it has grown from a small number of C files under a license prohibiting commercial distribution to the 4.15 version in 2018 with more than 23.3 million lines of source code, not counting comments, under the GNU General Public License v2.
Events leading to creation
After AT&T had dropped out of the Multics project, the Unix operating system was conceived and implemented by Ken Thompson and Dennis Ritchie (both of AT&T Bell Laboratories) in 1969 and first released in 1970. Later they rewrote it in a new programming language, C, to make it portable. The availability and portability of Unix caused it to be widely adopted, copied and modified by academic institutions and businesses.
In 1977, the Berkeley Software Distribution (BSD) was developed by the Computer Systems Research Group (CSRG) from UC Berkeley, based on the 6th edition of Unix from AT&T. Since BSD contained Unix code that AT&T owned, AT&T filed a lawsuit (USL v. BSDi) in the early 1990s against the University of California. This strongly limited the development and adoption of BSD.
Onyx Systems began selling early microcomputer-based Unix workstations in 1980. Later, Sun Microsystems, founded as a spin-off of a student project at Stanford University, also began selling Unix-based desktop workstations in 1982. While Sun workstations didn't utilize commodity PC hardware like Linux was later developed for, it represented the first successful commercial attempt at distributing a primarily single-user microcomputer that ran a Unix operating system.
In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s, there |
https://en.wikipedia.org/wiki/Coordinative%20definition | A coordinative definition is a postulate which assigns a partial meaning to the theoretical terms of a scientific theory by correlating the mathematical objects of the pure or formal/syntactical aspects of a theory with physical objects in the world. The idea was formulated by the logical positivists and arises out of a formalist vision of mathematics as pure symbol manipulation.
Formalism
In order to get a grasp on the motivations which inspired the development of the idea of coordinative definitions, it is important to understand the doctrine of formalism as it is conceived in the philosophy of mathematics. For the formalists, mathematics, and particularly geometry, is divided into two parts: the pure and the applied. The first part consists in an uninterpreted axiomatic system, or syntactic calculus, in which terms such as point, straight line and between (the so-called primitive terms) have their meanings assigned to them implicitly by the axioms in which they appear. On the basis of deductive rules eternally specified in advance, pure geometry provides a set of theorems derived in a purely logical manner from the axioms. This part of mathematics is therefore a priori but devoid of any empirical meaning, not synthetic in the sense of Kant.
It is only by connecting these primitive terms and theorems with physical objects such as rulers or rays of light that, according to the formalist, pure mathematics becomes applied mathematics and assumes an empirical meaning. The method of correlating the abstract mathematical objects of the pure part of theories with physical objects consists in coordinative definitions.
It was characteristic of logical positivism to consider a scientific theory to be nothing more than a set of sentences, subdivided into the class of theoretical sentences, the class of observational sentences, and the class of mixed sentences. The first class contains terms which refer to theoretical entities, that is to entities not directly observabl |
https://en.wikipedia.org/wiki/National%20Longitudinal%20Surveys | The National Longitudinal Surveys (NLS) are a set of surveys sponsored by the Bureau of Labor Statistics (BLS) of the U.S. Department of Labor. These surveys have gathered information at multiple points in time on the labor market experiences and other significant life events of several groups of men and women. Each of the NLS samples consists of several thousand individuals, many of whom have been surveyed over several decades.
Surveys
The National Longitudinal Survey of Youth 1997 (NLSY97) began in 1997 with 8,984 men and women born in 1980-84 (ages 12–17 in 1997). Sample members were interviewed annually from 1997 to 2011 and biennially thereafter. The 2015 interview was conducted with 7,103 men and women ages 30–36. Data are available from Round 1 (1997–98) to Round 17 (2015–16).
The National Longitudinal Survey of Youth 1979 (NLSY79) began in 1979 with 12,686 men and women born in 1957-64 (ages 14–22 in 1979). Sample members were interviewed annually from 1979-1994 and biennially thereafter. Oversamples of military and economically disadvantaged, nonblack/non-Hispanic respondents were dropped in 1985 and 1991, leaving a sample size of 9,964. The 2014 interview (Round 26) was conducted with 7,071 men and women ages 49–58.
The NLSY79 Children and Young Adults (NLSCYA) began in 1986 with children born to female NLSY79 respondents. Biennial data collection consists of interviews with the mothers and interviews with the children themselves; from 1994 onward, children turning age 15 and older during the survey year have been administered a Young Adult questionnaire that is similar to the NLSY79 questionnaire. In 2014, 276 children (ages 0–14) and 5,735 young adults (ages 15–42) were interviewed. To date, about 10,500 children have been interviewed in at least one survey round.
The National Longitudinal Surveys of Young Women and Mature Women (NLSW) comprised two separate surveys. The Young Women's survey began in 1968 with 5,159 women born in 1943-53 (ages 14–2 |
https://en.wikipedia.org/wiki/Biaugmented%20pentagonal%20prism | In geometry, the biaugmented pentagonal prism is one of the Johnson solids (). As the name suggests, it can be constructed by doubly augmenting a pentagonal prism by attaching square pyramids () to two of its nonadjacent equatorial faces. (The solid obtained by attaching pyramids to adjacent equatorial faces is not convex, and thus not a Johnson solid.)
External links
Johnson solids |
https://en.wikipedia.org/wiki/Network%20intelligence | Network intelligence (NI) is a technology that builds on the concepts and capabilities of deep packet inspection (DPI), packet capture and business intelligence (BI). It examines, in real time, IP data packets that cross communications networks by identifying the protocols used and extracting packet content and metadata for rapid analysis of data relationships and communications patterns. Also, sometimes referred to as Network Acceleration or piracy.
NI is used as a middleware to capture and feed information to network operator applications for bandwidth management, traffic shaping, policy management, charging and billing (including usage-based and content billing), service assurance, revenue assurance, market research mega panel analytics, lawful interception and cyber security. It is currently being incorporated into a wide range of applications by vendors who provide technology solutions to Communications Service Providers (CSPs), governments and large enterprises. NI extends network controls, business capabilities, security functions and data mining for new products and services needed since the emergence of Web 2.0 and wireless 3G and 4G technologies.
Background
The evolution and growth of Internet and wireless technologies offer possibilities for new types of products and services, as well as opportunities for hackers and criminal organizations to exploit weaknesses and perpetrate cyber crime. Network optimization and security solutions therefore need to address the exponential increases in IP traffic, methods of access, types of activity and volume of content generated. Traditional DPI tools from established vendors have historically addressed specific network infrastructure applications such as bandwidth management, performance optimization and quality of service (QoS).
DPI focuses on recognizing different types of IP traffic as part of a CSP's infrastructure. NI provides more granular analysis. It enables vendors to create an information layer wi |
https://en.wikipedia.org/wiki/Retronasal%20smell | Retronasal smell, retronasal olfaction, is the ability to perceive flavor dimensions of foods and drinks. Retronasal smell is a sensory modality that produces flavor. It is best described as a combination of traditional smell (orthonasal smell) and taste modalities. Retronasal smell creates flavor from smell molecules in foods or drinks shunting up through the nasal passages as one is chewing. When people use the term "smell", they are usually referring to "orthonasal smell", or the perception of smell molecules that enter directly through the nose and up the nasal passages. Retronasal smell is critical for experiencing the flavor of foods and drinks. Flavor should be contrasted with taste, which refers to five specific dimensions: (1) sweet, (2) salty, (3) bitter, (4) sour, and (5) umami. Perceiving anything beyond these five dimensions, such as distinguishing the flavor of an apple from a pear for example, requires the sense of retronasal smell.
History
Evolutionarily, smell has long been presumed to be a less-important sense for humans, especially compared to vision. Vision appears to dominate human stimuli perception, but researchers now argue that smell cues are highly informative to humans despite being less obviously so. Before his death in 1826, French gastronome Brillat-Savarin published his book, The Physiology of Taste; Or, Meditations on Transcendental Gastronomy: Theoretical, Historical, and Practical Work, in which he makes the first mention of the importance of smell in the “combined sense” of taste. He defines taste in terms of the five taste dimensions in addition to flavor created with the nasal apparatus. Avery Gilbert, in his book The Nose Knows, reviews the work of Henry T. Finck, an American philosopher from the late 1800s who published a groundbreaking essay titled “The Gastronomic Value of Odours.” Flink called flavor a “second way of smelling,” and much subsequent scientific investigation in the early 1900s focused on attempting to break |
https://en.wikipedia.org/wiki/Kirchhoff%20equations | In fluid dynamics, the Kirchhoff equations, named after Gustav Kirchhoff, describe the motion of a rigid body in an ideal fluid.
where and are the angular and linear velocity vectors at the point , respectively; is the moment of inertia tensor, is the body's mass; is
a unit normal to the surface of the body at the point ;
is a pressure at this point; and are the hydrodynamic
torque and force acting on the body, respectively;
and likewise denote all other torques and forces acting on the
body. The integration is performed over the fluid-exposed portion of the
body's surface.
If the body is completely submerged body in an infinitely large volume of irrotational, incompressible, inviscid fluid, that is at rest at infinity, then the vectors and can be found via explicit integration, and the dynamics of the body is described by the Kirchhoff – Clebsch equations:
Their first integrals read
Further integration produces explicit expressions for position and velocities. |
https://en.wikipedia.org/wiki/SharkWire%20Online | SharkWire Online is a specialized GameShark device with a serial port and modem added, accompanied by a now-defunct dialup Internet portal service. Launched in January 2000, it was sold only in the US, by InterAct which is most famous for its GameShark and Dexdrive. This unlicensed platform was the only Nintendo 64 online service to have been released other than Nintendo's official Randnet service which had already been released only in Japan in December 1999.
History
The SharkWire Online's Nintendo 64 accessories were developed by Datel in the UK, for InterAct to sell in the US. The now-defunct dialup portal system was developed between InterAct and its communications partners: Spyglass, Inc. for its Mosaic web browser application; D3 Networks, an Internet Service Provider (ISP) for game devices, which built and operated the SharkWire Online dialup network and content portal; and GTE Internetworking for its local dial-up access via its DiaLinx network and Global Network Infrastructure (GNI) backbone.
The SharkWire Online was presented to the public at the Electronic Entertainment Expo (E3) in Los Angeles on May 13–15, 1999. A PlayStation version was preannounced but canceled in development. Later in 1999, it was test marketed in Atlanta, Georgia, Dallas, Texas, Minneapolis, Minnesota, but not released to the rest of the US until January 1, 2000. The company considered the possibility of eventually supporting online multiplayer gaming, and opening up access to the wider Internet beyond their proprietary portal.
The company ran a $5–10 million advertising campaign created by advertising agency J. Walter Thompson across TV, print, radio, direct, and interactive media. It portrayed an aggressive image of teenaged "hacking" versus the FBI, which "gives kids a feeling of control and power over the establishment".
In 2003, SharkWire Online and all other trademarks of GameShark were sold to Mad Catz, and InterAct ceased operations.
Usage
The SharkWire Online produ |
https://en.wikipedia.org/wiki/Lateral%20superior%20genicular%20artery | The lateral superior genicular artery is a branch of the popliteal artery that supplies a portion of the knee joint.
Anatomy
Course and relations
It passes above the lateral condyle of the femur. It runs deep to the tendon of the biceps femoris.
Branches
It divides into a superficial and a deep branch; the superficial branch supplies the vastus lateralis, and anastomoses with the descending branch of the lateral femoral circumflex and the lateral inferior genicular arteries; the deep branch supplies the lower part of the femur and knee-joint, and forms an anastomotic arch across the front of the bone with the highest genicular and the medial inferior genicular arteries.
Additional images
See also
Patellar anastomosis |
https://en.wikipedia.org/wiki/Aerospike%20%28database%29 | Aerospike Database is a real-time, high performance NoSQL database. Designed for applications that cannot experience any downtime and require high read & write throughput. Aerospike is optimized to run on NVMe SSDs capable of efficiently storing large datasets (Gigabytes to Petabytes). Aerospike can also be deployed as a fully in-memory cache database. Aerospike offers Key-Value, JSON Document, and Graph data models. Aerospike is open source distributed NoSQL database management system, marketed by the company also named Aerospike.
History
Aerospike was first known as Citrusleaf. In August 2012, the company - which had been providing its database since 2010 - rebranded both the company and software name to Aerospike. The name "Aerospike" is derived from the aerospike engine, a type of rocket nozzle that is able to maintain its output efficiency over a large range of altitudes, and is intended to refer to the software's ability to scale up. In 2012, Aerospike acquired AlchemyDB, and integrated the two databases' functions, including the addition of a relational data management system. On June 24, 2014, Aerospike was opensourced under the AGPL 3.0 license for the Aerospike database server and the Apache License Version 2.0 for its Aerospike client software development kit.
Release history
Features
Aerospike Database is modeled under the shared-nothing architecture and written in C. It operates in three layers: a data storage layer, a self-managed distribution layer, and a cluster-aware client layer.
Aerospike uses hybrid memory architecture: the database indices are stored fully in main random-access memory, while the data is stored on a persistent device using the data layer. The data layer stores the data in solid-state drive, NVMe or Persistent memory. Reading the data is done using a direct access to the record position on disk using a direct pointer from the primary index, and data writes are optimized through large block writes to reduce latency. This archi |
https://en.wikipedia.org/wiki/Aurora%20Pulsed%20Radiation%20Simulator | The Aurora Pulsed Radiation Simulator (also known as the Aurora flash x-ray simulator) is a 14 TW flash gamma-ray simulator that was designed to simulate the effects of a nuclear weapon's bremsstrahlung, or gamma radiation, pulses on military electronic systems. It was built in 1971 by the U.S. Defense Atomic Support Agency (DASA), which eventually became the Defense Threat Reduction Agency, and the U.S. Department of Energy (DOE).
More than long and weighing at 1,450 tons, the Aurora Simulator was the first gamma radiation simulator of its size in the world at the time. It was also one of only four large machines in the United States that were built specifically to test complete nuclear weapons packages, with the other three being the Hermes I to III simulators at Sandia Base, New Mexico. Situated at the Harry Diamond Laboratories (which later became a part of the Army Research Laboratory) in Adelphi, Maryland, it was used to test complete weapons electronics packages from the warheads of intercontinental ballistic missiles (ICBMs) to satellites. After more than 20 years of use during the Cold War, the Aurora Simulator was officially decommissioned and disassembled in 1996.
In 1986, the Aurora facility set the world record for the largest amount of high-power microwave power generated from a virtual cathode oscillator. As a result, HDL was recognized by the American Defense Preparedness Association (ADPA) in 1987.
History
Following the use of the atomic bomb in World War II, studies on flash radiography found that field emission flash x-rays, which were previously used to analyze explosions, could simulate radiation from a nuclear bomb. Given this realization, the U.S. military began to prioritize the development of flash x-ray machines to test parts of missile packages during the 1960s.
After the Soviet Union demonstrated the use of the world's first anti-ballistic missiles (ABM) in 1964, DASA launched a series of projects in response that aimed to hasten th |
https://en.wikipedia.org/wiki/Cray%20X2 | The Cray X2 is a vector processing node for the Cray XT5h supercomputer, developed and sold by Cray Inc. and launched in 2007.
The X2, developed under the code name Black Widow, was originally expected to be a standalone supercomputer system, superseding the Cray X1 parallel vector supercomputer. However, the X2 was eventually launched as one of the four processor "blade" options for the XT5h system.
An X2 blade comprises two nodes, each with four symmetric multiprocessing vector processors and 32 or 64 GB of shared memory. Each node has a peak performance of more than 100 gigaflops. X2 processors are connected using a radix-64 "fat-tree" interconnect implemented by the YARC router ASIC. X2 blades also link into the XT5h system via its SeaStar2+ processor interconnect.
Up to 256 X2 blades can be installed in an XT5h system. The X2 processor nodes integrate with the Cray XT5h's UNICOS/lc OS, user environment, and storage subsystem, as part of the Rainier project.
External links
Cray XT5h Supercomputer
Cray Introduces Next-Generation Supercomputers
Thinking Ahead: Future Architectures from Cray
The BlackWidow High-Radix Clos Network
Cray X2 Vector Processing Blade
X2
Vector supercomputers |
https://en.wikipedia.org/wiki/Valve%20actuator | A valve actuator is the mechanism for opening and closing a valve. Manually operated valves require someone in attendance to adjust them using a direct or geared mechanism attached to the valve stem. Power-operated actuators, using gas pressure, hydraulic pressure or electricity, allow a valve to be adjusted remotely, or allow rapid operation of large valves. Power-operated valve actuators may be the final elements of an automatic control loop which automatically regulates some flow, level or other process. Actuators may be only to open and close the valve, or may allow intermediate positioning; some valve actuators include switches or other ways to remotely indicate the position of the valve.
Used for the automation of industrial valves, actuators can be found in all kinds of process plants. They are used in waste water treatment plants, power plants, refineries, mining and nuclear processes, food factories, and pipelines. Valve actuators play a major part in automating process control. The valves to be automated vary both in design and dimension. The diameters of the valves range from one-tenth of an inch to several feet.
Types
The common types of actuators are: manual, pneumatic, hydraulic, electric and spring.
Manual
A manual actuator employs levers, gears, or wheels to move the valve stem with a certain action. Manual actuators are powered by hand. Manual actuators are inexpensive, typically self-contained, and easy to operate by humans. However, some large valves are impossible to operate manually and some valves may be located in remote, toxic, or hostile environments that prevent manual operations in some conditions. As a safety feature, certain types of situations may require quicker operation than manual actuators can provide to close the valve.
Pneumatic
Air (or other gas) pressure is the power source for pneumatic valve actuators. They are used on linear or quarter-turn valves. Air pressure acts on a piston or bellows diaphragm creating linear fo |
https://en.wikipedia.org/wiki/IMA%20Journal%20of%20Mathematical%20Control%20and%20Information | The IMA Journal of Mathematical Control and Information is published by Oxford University Press on behalf of the Institute of Mathematics and its Applications. The Journal publishes articles in control and information theory which aim to develop solutions for unsolved problems in the field.
External links
Journal homepage
Submission website
Institute of Mathematics and its Applications |
https://en.wikipedia.org/wiki/Psychological%20warfare | Psychological warfare (PSYWAR), or the basic aspects of modern psychological operations (PsyOp), have been known by many other names or terms, including Military Information Support Operations (MISO), Psy Ops, political warfare, "Hearts and Minds", and propaganda. The term is used "to denote any action which is practiced mainly by psychological methods with the aim of evoking a planned psychological reaction in other people".
Various techniques are used, and are aimed at influencing a target audience's value system, belief system, emotions, motives, reasoning, or behavior. It is used to induce confessions or reinforce attitudes and behaviors favorable to the originator's objectives, and are sometimes combined with black operations or false flag tactics. It is also used to destroy the morale of enemies through tactics that aim to depress troops' psychological states.
Target audiences can be governments, organizations, groups, and individuals, and is not just limited to soldiers. Civilians of foreign territories can also be targeted by technology and media so as to cause an effect on the government of their country.
Mass communication such as radio allows for direct communication with an enemy populace, and therefore has been used in many efforts. Social media channels and the internet allow for campaigns of disinformation and misinformation performed by agents anywhere in the world.
History
Early
Since prehistoric times, warlords and chiefs have recognized the importance of weakening the morale of their opponents. In the Battle of Pelusium (525 BC) between the Persian Empire and ancient Egypt, the Persian forces used cats and other animals as a psychological tactic against the Egyptians, who avoided harming cats due to religious belief and superstitions.
Currying favor with supporters was the other side of psychological warfare, and an early practitioner of this was Alexander the Great, who successfully conquered large parts of Europe and the Middle East and h |
https://en.wikipedia.org/wiki/Aerolysin | In molecular biology, aerolysin is a cytolytic pore-forming toxin exported by Aeromonas hydrophila, a Gram-negative bacterium associated with diarrhoeal diseases and deep wound infections. It is also produced by the caterpillar of the moth Megalopyge opercularis, sometimes called the Tree Asp. The mature toxin binds to eukaryotic cells and aggregates to form holes (approximately 3 nm in diameter) leading to the destruction of the membrane permeability barrier and osmotic lysis. The structure of proaerolysin has been determined to 2.8A resolution and shows the protoxin to adopt a novel fold. Images of an aerolysin oligomer derived from electron microscopy have helped to construct a model of the protein in its heptameric conformation, and to outline a mechanism by which this assembly might insert into lipid bilayers to form ion channels. |
https://en.wikipedia.org/wiki/Source%20counts | The source counts distribution of radio-sources from a radio-astronomical survey is the cumulative distribution of the number of sources (N) brighter than a given flux density (S). As it is usually plotted on a log-log scale its distribution is known as the log N – log S plot. It is one of several cosmological tests that were conceived in the 1930s to check the viability of and compare new cosmological models.
Early work to catalogue radio sources aimed to determine the source count distribution as a discriminating test of different cosmological models. For example, a uniform distribution of radio sources at low redshift, such as might be found in a 'steady-state Euclidean universe,' would produce a slope of −1.5 in the cumulative distribution of log(N) versus log(S).
Data from the early Cambridge 2C survey (published 1955) apparently implied a (log(N), log(S)) slope of nearly −3.0. This appeared to invalidate the steady state theory of Fred Hoyle, Hermann Bondi and Thomas Gold. Unfortunately many of these weaker sources were subsequently found to be due to 'confusion' (the blending of several weak sources in the side-lobes of the interferometer, producing a stronger response).
By contrast, analysis from the contemporaneous Mills Cross data (by Slee and Mills) were consistent with an index of −1.5.
Later and more accurate surveys from Cambridge, 3C, 3CR, and 4C, also showed source count slopes steeper than −1.5, though by a smaller margin than 2C. This convinced some cosmologists that the steady state theory was wrong, although residual problems with confusion provided some defense for Hoyle and his colleagues.
The immediate interest in testing the steady-state theory through source-counts was reduced by the discovery of the 3K microwave background radiation in the mid 1960s, which essentially confirmed the Big-Bang model.
Later radio survey data have shown a complex picture — the 3C and 4C claims appear to hold up, while at fainter levels the source counts |
https://en.wikipedia.org/wiki/Patent%20Blue%20V | Patent Blue V, also called Food Blue 5, Sulphan Blue, Acid Blue 3, L-Blau 3, C-Blau 20, Patentblau V, Sky Blue, or C.I. 42051, is a sky blue synthetic triphenylmethane dye used as a food coloring. As a food additive, it has E number E131. It is a sodium or calcium salt of [4-(α-(4-diethylaminophenyl)-5-hydroxy- 2,4-disulfophenylmethylidene)-2,5-cyclohexadien-1-ylidene] diethylammonium hydroxide inner salt.
Use as dye
It is not widely used, but in Europe it can be found in Scotch eggs, certain jelly sweets, blue Curaçao, certain gelatin desserts, among others. An important advantage is the very deep color it produces even at low concentration, a disadvantage is that it fades fairly quickly when exposed to light.
In medicine, Patent Blue V is used in lymphangiography and sentinel node biopsy as a dye to color lymph vessels. It is also used in dental disclosing tablets as a stain to show dental plaque on teeth.
The color of the dye is pH-dependent. In aqueous solution, its color will vary from a deep blue in alkaline or weakly acidic medium to a yellow–orange in stronger acidic
conditions. It is useful as a pH indicator for the range 0.8–3.0. The structure is also redox-sensitive, changing from a reduced yellow form to an oxidized red form in solution. The reduction potential of around 0.77 volts is similar to that of other triphenylmethane dyes. It is usable as a reversible redox indicator in some analytical methods.
Because of its pH-dependent color, Patent Blue V was included in chemistry sets from Salter Science in the 1970s and 80s under the name Sky Blue.
Regulation
Patent Blue V is banned as a food dye in Australia and US, because health officials in these countries suspect that it may cause allergic reactions, with symptoms ranging from itching and nettle rash to nausea, hypotension, and in rare cases anaphylactic shock; it is therefore not recommended in those countries for children. |
https://en.wikipedia.org/wiki/U-Key | A U-Key is an implementation of the MIFARE RFID chip, encased in a plastic key style housing. It is used as a prepayment system on vending machines and for some self-service diving air compressors in Switzerland, and they will be most likely made by Selecta (company) |
https://en.wikipedia.org/wiki/Bathometer | A bathometer (also bathymeter) is an instrument for measuring water depth. It was previously used mainly in oceanographical studies, but is rarely employed nowadays. The term originates from Greek βαθύς (bathys), "deep" and μέτρον (métron), "measure".
History
The earliest idea for a bathometer is due to Leon Battista Alberti (1404–1472) who sunk a hollow sphere attached to some ballast with a hook. When the ball reached the bottom it detached from the ballast and resurfaced. The depth was determined (rather inaccurately) by the time it took to surface. Jacob Perkins (1766–1849) proposed a bathometer based on the compressibility of water. In this instrument the movement of a piston compressing a body of water enclosed in its cylinder is dependent on the pressure of the water outside the cylinder, and hence its depth. The amount the piston moved can be measured when it is returned to the surface.
A bathometer that did not need to be submerged was invented in 1876 by William Siemens, stimulated by the needs of the telegraph industry. Siemens' instrument was the first to come into widespread use and is so different and so much more practical than anything that had gone before that he is often credited as the inventor of the bathometer. His instrument consisted of a tube of mercury and worked similar to a barometer. The pressure of the mercury acting under the force of gravity pushed down on, and deformed, a thin steel sheet. The height of the mercury in the column was thus proportional to the strength of the Earth's gravity field. The theory of the instrument was that the greater the depth of water under the ship, the lower the gravitational force would be. This is because water has a much lower density than the rocks of the Earth's crust. Starting in the mid-nineteenth century, submarine telegraph cables were being laid around the world. Accurate knowledge of the depth of the ocean bed was important for this work. Previously, depth was determined by ta |
https://en.wikipedia.org/wiki/List%20of%20scorewriters | This is a list of music notation programs (excluding discontinued products) which have articles on Wikipedia.
For programs specifically for writing guitar tablature, see the list of guitar tablature software. For discontinued products, see list of discontinued scorewriters.
Free software
Denemo, a scorewriter primarily providing a front-end for LilyPond
Frescobaldi, a GUI front-end for LilyPond. [Linux, FreeBSD, OS X and Microsoft Windows]
Impro-Visor, a GUI- and text-based scorewriter for constructing lead sheets and jazz solos on Linux, OS X, and Windows
LilyPond, a text-based scorewriter with several backends including PS, PDF and SVG
MuseScore, a WYSIWYG scorewriter for Linux, Windows, and OS X
MusiXTeX, a set of macros and fonts that allow music typesetting in TeX
NoteEdit, a KDE scorewriter
Rosegarden, a scorewriter for Linux
Philip's Music Writer, a text-based scorewriter originally written for Acorn RISC OS (released as a commercial program in the 1990s), later ported to POSIX and licensed under the GNU GPL
Proprietary
Microsoft Windows
Capella
Cubase Score V1-5 (first run on version numbers)
Cubase SX
Cubase V4-9.5 (second run on version numbers)
Dorico
Encore
Finale plus the following lite versions: Allegro, PrintMusic, NotePad, Songwriter
Forte (notation program)
Guitar Pro (primarily for guitars and bands, but also notates other instruments including drums)
Igor Engraver
MagicScore, plus Music Notation for MS Word and lite version MagicScore School and free versions MagicScore onLine and MagicScore Note
Mozart
Mus2
MusEdit
MusiCAD
MusicEase, notates standard music, shaped notes and tablature; transposes and imports abc music.
Music Construction Set (obsolete; was also for Apple II, Atari 400, and Commodore 64)
Music Write
MusicTime Deluxe
Musink Lite, a WYSIWYM scorewriter and publication tool for Windows
Notion
Notation Composer
NoteWorthy Composer
Overture, plus lite version Score Writer.
SCORE, one of the earli |
https://en.wikipedia.org/wiki/GNU%20Free%20Documentation%20License | The GNU Free Documentation License (GNU FDL or simply GFDL) is a copyleft license for free documentation, designed by the Free Software Foundation (FSF) for the GNU Project. It is similar to the GNU General Public License, giving readers the rights to copy, redistribute, and modify (except for "invariant sections") a work and requires all copies and derivatives to be available under the same license. Copies may also be sold commercially, but, if produced in larger quantities (greater than 100), the original document or source code must be made available to the work's recipient.
The GFDL was designed for manuals, textbooks, other reference and instructional materials, and documentation which often accompanies GNU software. However, it can be used for any text-based work, regardless of subject matter. For example, the free online encyclopedia Wikipedia uses the GFDL (coupled with the Creative Commons Attribution Share-Alike License) for much of its text, excluding text that was imported from other sources after the 2009 licensing update that is only available under the Creative Commons license.
History
The GFDL was released in draft form for feedback in September 1999. After revisions, version 1.1 was issued in March 2000, version 1.2 in November 2002, and version 1.3 in November 2008. The current state of the license is version 1.3.
The first discussion draft of the GNU Free Documentation License version 2 was released on September 26, 2006, along with a draft of the new GNU Simpler Free Documentation License.
On December 1, 2007, Wikipedia founder Jimmy Wales announced that a long period of discussion and negotiation between and amongst the Free Software Foundation, Creative Commons, the Wikimedia Foundation and others had produced a proposal supported by both the FSF and Creative Commons to modify the Free Documentation License in such a fashion as to allow the possibility for the Wikimedia Foundation to migrate the projects to the similar Creative Commons Att |
https://en.wikipedia.org/wiki/20%CE%B1%2C22R-Dihydroxycholesterol | 20α,22R-Dihydroxycholesterol, or (3β)-cholest-5-ene-3,20,22-triol is an endogenous, metabolic intermediate in the biosynthesis of the steroid hormones from cholesterol. Cholesterol ((3β)-cholest-5-en-3-ol) is hydroxylated by cholesterol side-chain cleavage enzyme (P450scc) to form 22R-hydroxycholesterol, which is subsequently hydroxylated again by P450scc to form 20α,22R-dihydroxycholesterol, and finally the bond between carbons 20 and 22 is cleaved by P450scc to form pregnenolone ((3β)-3-hydroxypregn-5-en-20-one), the precursor to the steroid hormones.
See also
22R-Hydroxycholesterol
27-Hydroxycholesterol |
https://en.wikipedia.org/wiki/Quick%20Response%20Engine | Quick Response Engine was a planning and scheduling program developed for the OS/400 platform. The program was developed by the Acacia Technologies division of Computer Associates in 1996. In 2002 the group was sold to SSA Global Technologies. |
https://en.wikipedia.org/wiki/MESI%20protocol | The MESI protocol is an Invalidate-based cache coherence protocol, and is one of the most common protocols that support write-back caches. It is also known as the Illinois protocol due to its development at the University of Illinois at Urbana-Champaign. Write back caches can save considerable bandwidth generally wasted on a write through cache. There is always a dirty state present in write back caches that indicates that the data in the cache is different from that in main memory. The Illinois Protocol requires a cache to cache transfer on a miss if the block resides in another cache. This protocol reduces the number of main memory transactions with respect to the MSI protocol. This marks a significant improvement in performance.
States
The letters in the acronym MESI represent four exclusive states that a cache line can be marked with (encoded using two additional bits):
Modified (M) The cache line is present only in the current cache, and is dirty - it has been modified (M state) from the value in main memory. The cache is required to write the data back to main memory at some time in the future, before permitting any other read of the (no longer valid) main memory state. The write-back changes the line to the Shared state(S).
Exclusive (E) The cache line is present only in the current cache, but is clean - it matches main memory. It may be changed to the Shared state at any time, in response to a read request. Alternatively, it may be changed to the Modified state when writing to it.
Shared (S) Indicates that this cache line may be stored in other caches of the machine and is clean - it matches the main memory. The line may be discarded (changed to the Invalid state) at any time.
Invalid (I) Indicates that this cache line is invalid (unused).
For any given pair of caches, the permitted states of a given cache line are as follows:
When the block is marked M (modified) or E (exclusive), the copies of the block in other Caches are marked as I (Invalid).
|
https://en.wikipedia.org/wiki/Goodyear%20MPP | The Goodyear Massively Parallel Processor (MPP) was a
massively parallel processing supercomputer built by Goodyear Aerospace
for the NASA Goddard Space Flight Center. It was designed to deliver enormous computational power at lower cost than other existing supercomputer architectures, by using thousands of simple processing elements, rather than one or a few highly complex CPUs. Development of the MPP began circa 1979; it was delivered in May 1983, and was in general use from 1985 until 1991.
It was based on Goodyear's earlier STARAN array processor, a 4x256 1-bit processing element (PE) computer. The MPP was a 128x128 2-dimensional array of 1-bit wide PEs. In actuality 132x128 PEs were configured with a 4x128 configuration added for fault tolerance to substitute for up to 4 rows (or columns) of processors in the presence of problems. The PEs operated in a single instruction, multiple data (SIMD) fashioneach PE performed the same operation simultaneously, on different data elements, under the control of a microprogrammed control unit.
After the MPP was retired in 1991, it was donated to the Smithsonian Institution, and is now in the collection of the National Air and Space Museum's Steven F. Udvar-Hazy Center. It was succeeded at Goddard by the MasPar MP-1 and Cray T3D massively parallel computers.
Applications
The MPP was initially developed for high-speed analysis of satellite images. In early tests, it was able to extract and separate different land-use areas on Landsat imagery in 18 seconds, as compared with 7 hours on a DEC VAX-11/780.
Once the system was put into production use, NASA's Office of Space Science and Applications solicited proposals from scientists across the country to test and implement a wide range of computational algorithms on the MPP. 40 projects were accepted, to form the "MPP Working Group"; results of most of them were presented at the First Symposium on the Frontiers of Massively Parallel Computation, in 1986.
Some examples of app |
https://en.wikipedia.org/wiki/Zooflagellate | In some older systems of classification, Zoomastigophora is a phylum (more commonly known as zooflagellates) within the kingdom Protista. Organisms within this group have a spherical, elongated body with a single central nucleus. They are single-celled, heterotrophic eukaryotes and may form symbiotic relationships with other organisms, including Trichomonas. Some species are parasitic, causing diseases such as the African Sleeping Sickness, caused by the zooflagellate Trypanosoma brucei. Zooflagellates have one or more flagella but do not have plastids or cell walls.
A few are mutualistic, such as those that live in the guts of termites and aid the bacteria present in breaking down wood. |
https://en.wikipedia.org/wiki/Matei%20Machedon | Matei Machedon (born 10 February 1960 in Romania) is a Romanian–American mathematician, specializing in partial differential equations and mathematical physics.
Machedon graduated from the University of Chicago with B.A./M.S. in 1982. He received his Ph.D. from Princeton University in 1986 with thesis advisor Charles Fefferman. Machedon was a C.L.E. Moore Instructor at Massachusetts Institute of Technology from 1986 to 1988. At Princeton University he was an assistant professor from 1988 to 1994. He was at the Institute for Advanced Study for the academic year 1994–1995. At the University of Maryland he was an associate professor from 1994 to 1998 and is since 1998 a full professor.
Machedon held a Sloan Fellowship for the academic year 1985–1986 and for the two academic years 1989–1991. He has frequently collaborated with Sergiu Klainerman and Manoussos Grillakis. In 1998 Machedon was an invited speaker at the International Congress of Mathematicians in Berlin.
Selected publications |
https://en.wikipedia.org/wiki/Laser%20ablation%20electrospray%20ionization | Laser ablation electrospray ionization (LAESI) is an ambient ionization method for mass spectrometry that combines laser ablation from a mid-infrared (mid-IR) laser with a secondary electrospray ionization (ESI) process. The mid-IR laser is used to generate gas phase particles which are then ionized through interactions with charged droplets from the ESI source. LAESI was developed in Professor Akos Vertes lab by Dr. Peter Nemes in 2007 and it was marketed commercially by Protea Biosciences, Inc until 2017. Fiber-LAESI for single-cell analysis approach was developed by Dr. Bindesh Shrestha in Professor Vertes lab in 2009. LAESI is a novel ionization source for mass spectrometry (MS) that has been used to perform MS imaging of plants, tissues, cell pellets, and even single cells. In addition, LAESI has been used to analyze historic documents and untreated biofluids such as urine and blood. The technique of LAESI is performed at atmospheric pressure and therefore overcomes many of the obstacles of traditional MS techniques, including extensive and invasive sample preparation steps and the use of high vacuum. Because molecules and aerosols are ionized by interacting with an electrospray plume, LAESI's ionization mechanism is similar to SESI and EESI techniques.
LAESI can be used to perform MS analysis of many different classes of compounds ranging from small molecules, such as pharmaceuticals, saccharides, lipids, and metabolites to larger biomolecules like peptides and proteins. LAESI has also been shown to have a quantitative dynamic range of 4 decades and a limit of detection (LOD) of 8 fmol with verapamil, a small pharmaceutical molecule. The technique has a lateral resolution of <200 μm for imaging applications and has been used for 3D imaging of plant tissues. Additionally, in cell-by-cell LAESI imaging experiments single cells can be used as the pixels of the molecular image. This LAESI application uses etched optical fibers to produce laser spot sizes of |
https://en.wikipedia.org/wiki/Threema | Threema is a paid cross-platform encrypted instant messaging app developed by Threema GmbH in Switzerland and launched in 2012. The service operates on a decentralized architecture and offers end-to-end encryption. Users can make voice and video calls, send photos, files, and voice notes, share locations, and make groups. Unlike many other popular secure messaging apps, Threema does not require phone numbers or email address for registration, only a one-time purchase. Threema is available on iOS and Android and has clients for Windows, macOS, Linux, and can be accessed via web browser but requires a mobile app to function.
Features
The service claims to be based on the privacy by design principles by not requiring a phone number or other personally identifiable information. This helps anonymize the users to a degree.
Threema uses a user ID, created after the initial app launch by a random generator, instead of requiring a linked email address or phone number to send messages. It is possible to find other users by phone number or email address if the user allows the app to synchronize their address book. Linking a phone number or email address to a Threema ID is optional. Hence, the service can be used anonymously. Users can verify the identity of their Threema contacts by scanning their QR code when they meet physically. The QR code contains the public key of the user, which is cryptographically tied to the ID and will not change during the lifetime of the identity. Using this strong authentication feature, users can make sure they have the correct public key from their chat partners, which provides additional security against a Man-in-the-middle attack. Threema knows three levels of authentication (trust levels of the contact's identity). The verification level of each contact is displayed in the Threema application as dots next to the corresponding contact.
In addition to text messaging, users can make voice and video calls, send multimedia, locations, voice |
https://en.wikipedia.org/wiki/%CE%A9-logic | In set theory, Ω-logic is an infinitary logic and deductive system proposed by as part of an attempt to generalize the theory of determinacy of pointclasses to cover the structure . Just as the axiom of projective determinacy yields a canonical theory of , he sought to find axioms that would give a canonical theory for the larger structure. The theory he developed involves a controversial argument that the continuum hypothesis is false.
Analysis
Woodin's Ω-conjecture asserts that if there is a proper class of Woodin cardinals (for technical reasons, most results in the theory are most easily stated under this assumption), then Ω-logic satisfies an analogue of the completeness theorem. From this conjecture, it can be shown that, if there is any single axiom which is comprehensive over (in Ω-logic), it must imply that the continuum is not . Woodin also isolated a specific axiom, a variation of Martin's maximum, which states that any Ω-consistent (over ) sentence is true; this axiom implies that the continuum is .
Woodin also related his Ω-conjecture to a proposed abstract definition of large cardinals: he took a "large cardinal property" to be a property of ordinals which implies that α is a strong inaccessible, and which is invariant under forcing by sets of cardinal less than α. Then the Ω-conjecture implies that if there are arbitrarily large models containing a large cardinal, this fact will be provable in Ω-logic.
The theory involves a definition of Ω-validity: a statement is an Ω-valid consequence of a set theory T if it holds in every model of T having the form for some ordinal and some forcing notion . This notion is clearly preserved under forcing, and in the presence of a proper class of Woodin cardinals it will also be invariant under forcing (in other words, Ω-satisfiability is preserved under forcing as well). There is also a notion of Ω-provability; here the "proofs" consist of universally Baire sets and are checked by verifying that for every |
https://en.wikipedia.org/wiki/Harlaniella | Harlaniella podolica is a problematic Ediacaran species of elongate, striated or crudely 'segmented' tubes, once thought to represent a trace fossil, but now believed to represent internal casts of the body of an unknown organism. |
https://en.wikipedia.org/wiki/Methemalbumin | Methemalbumin (MHA) is an albumin complex consisting of albumin and heme.
This complex gives brown color to plasma and occurs in hemolytic and hemorrhagic disorders.
Its presence in plasma is used to differentiate between hemorrhagic and edematous pancreatitis.
The Schumm test is used to differentiate intravascular haemolysis from extravascular haemolysis, as in haemolytic anaemias. A positive result is indicative of intravascular haemolysis. |
https://en.wikipedia.org/wiki/Geodesics%20on%20an%20ellipsoid | The study of geodesics on an ellipsoid arose in connection with geodesy specifically with the solution of triangulation networks. The figure of the Earth is well approximated by an oblate ellipsoid, a slightly flattened sphere. A geodesic is the shortest path between two points on a curved surface, analogous to a straight line on a plane surface. The solution of a triangulation network on an ellipsoid is therefore a set of exercises in spheroidal trigonometry .
If the Earth is treated as a sphere, the geodesics are great circles (all of which are closed) and the problems reduce to ones in spherical trigonometry. However, showed that the effect of the rotation of the Earth results in its resembling a slightly oblate ellipsoid: in this case, the equator and the meridians are the only simple closed geodesics. Furthermore, the shortest path between two points on the equator does not necessarily run along the equator. Finally, if the ellipsoid is further perturbed to become a triaxial ellipsoid (with three distinct semi-axes), only three geodesics are closed.
Geodesics on an ellipsoid of revolution
There are several ways of defining geodesics . A simple definition is as the shortest path between two points on a surface. However, it is frequently more useful to define them as paths with zero geodesic curvature—i.e., the analogue of straight lines on a curved surface. This definition encompasses geodesics traveling so far across the ellipsoid's surface that they start to return toward the starting point, so that other routes are more direct, and includes paths that intersect or re-trace themselves. Short enough segments of a geodesics are still the shortest route between their endpoints, but geodesics are not necessarily globally minimal (i.e. shortest among all possible paths). Every globally-shortest path is a geodesic, but not vice versa.
By the end of the 18th century, an ellipsoid of revolution (the term spheroid is also used) was a well-accepted appro |
https://en.wikipedia.org/wiki/Disk-based%20backup | Disk-based backup refers to technology that allows one to back up large amounts of data to a disk storage unit. It is the technology which is often supplemented by tape drives for data archival or replication to another facility for disaster recovery. Additionally, backup-to-disk has several advantages over traditional tape backup for both technical and business reasons. With continued improvements in storage devices to provide faster access and higher storage capacity, a prime consideration for backup and restore operations, backup-to-disk will become more prominent in organizations.
Technical Advantages
There are technical advantages to backup-to-disk technology. One of the main advantages is the speed at which backups can be performed to the disk appliance. Backing up data to a backup-to-disk technology can be up to four times faster than traditional SCSI tape devices. While the new Serial Attached SCSI (SAS) connected tape drives are faster than the original tape drives, the disk appliance is still faster than most tape technologies. These faster backup times lead to shorter backup windows allowing the technology to backup the data while in a smaller amount of time thus increasing the window for processing which is also a benefit for the business.
Another advantage that backup-to-disk offers is data deduplication and compression. The disk appliances offer either de-duplication at the source or at the destination. The deduplication at the destination is faster and requires less performance overhead on the source host. The deduplication requires less disk space on the disk appliance as it stores only one copy of the possible multiple copies of one file on the network.
Many of the backup-to-disk technologies advertise up to 15 to 1 compression ratios. This also allows the information technology department to store more data on less disk space. With deduplication a disk appliance with 5 terabytes of raw disk space can store as much as 30 terabytes of compressed |
https://en.wikipedia.org/wiki/Drumhead%20%28sign%29 | The term drumhead refers to a type of removable sign that was prevalent on North American railroads of the first half of the 20th century. The sign was mounted at the rear of named passenger trains, and consisted of a box with internal illumination that shone through a tinted panel bearing the logo of the railroad or specific train. Since the box and the sign were usually circular in shape and resembled small drums, they came to be known as drumheads.
Railroad drumheads were removable so that they could be mounted on different passenger cars (usually on the rear of observations), as needed for specific trains.
See also
Headboard |
https://en.wikipedia.org/wiki/TLS-PSK | Transport Layer Security pre-shared key ciphersuites (TLS-PSK) is a set of cryptographic protocols that provide secure communication based on pre-shared keys (PSKs). These pre-shared keys are symmetric keys shared in advance among the communicating parties.
There are several cipher suites: The first set of ciphersuites use only symmetric key operations for authentication. The second set use a Diffie–Hellman key exchange authenticated with a pre-shared key. The third set combine public key authentication of the server with pre-shared key authentication of the client.
Usually, Transport Layer Security (TLS) uses public key certificates or Kerberos for authentication. TLS-PSK uses symmetric keys, shared in advance among the communicating parties, to establish a TLS connection. There are several reasons to use PSKs:
Using pre-shared keys can, depending on the ciphersuite, avoid the need for public key operations. This is useful if TLS is used in performance-constrained environments with limited CPU power.
Pre-shared keys may be more convenient from a key management point of view. For instance, in closed environments where the connections are mostly configured manually in advance, it may be easier to configure a PSK than to use certificates. Another case is when the parties already have a mechanism for setting up a shared secret key, and that mechanism could be used to “bootstrap” a key for authenticating a TLS connection.
Standards
: "Pre-Shared Key Ciphersuites for Transport Layer Security (TLS)".
: "Pre-Shared Key (PSK) Ciphersuites with NULL Encryption for Transport Layer Security (TLS)".
: "Pre-Shared Key Cipher Suites for TLS with SHA-256/384 and AES Galois Counter Mode".
: "ECDHE_PSK Cipher Suites for Transport Layer Security (TLS)".
See also
Transport layer security Secure Remote Password (TLS-SRP)
AES Galois Counter Mode (GCM)
Elliptic curve Diffie–Hellman (ECDHE)
Null encryption
SHA-256 |
https://en.wikipedia.org/wiki/Obstacle | An obstacle (also called a barrier, impediment, or stumbling block) is an object, thing, action or situation that causes an obstruction. Different types of obstacles include physical, economic, biopsychosocial, cultural, political, technological and military.
Types
Physical
As physical obstacles, we can enumerate all those physical barriers that block the action and prevent the progress or the achievement of a concrete goal. Examples:
architectural barriers that hinder access to people with reduced mobility;
doors, gates, and access control systems, designed to keep intruders or attackers out;
large objects, fallen trees or collapses through passageways, paths, roads, railroads, waterways or airfields, preventing mobility;
sandbanks, rocks or coral reefs, preventing free navigation;
hills, mountains and weather phenomena preventing the free traffic of aircraft;
meteors, meteorites, micrometeorites, cosmic dust, comets, space debris, strong electromagnetic radiation or gravitational field, preventing a spacecraft from navigating freely in space.
Sports
In sports, a variety of physical barriers or obstacles were introduced in the competition rules to make them even more difficult and competitive:
in the athletics, there are barriers in obstacle running contests of 110 meters and 3000 meters, as well as in high jump and in pole vault;
in equestrian competitions, there are also jumps over obstacles;
in tennis and volleyball, a net stands as an obstacle that divides the court;
in the cycling, motorcycle and motor racing, circuit designs are interposed with difficult paths to obstruct and render more difficult the competition;
in team sports, like soccer, football, basketball and volleyball, attack players are hampered by defensive players, that make it difficult to move or throw the ball towards the goal;
in other sports, such as Parkour, the competitor aims to move from one point to another in the most fluid and fast as possible, jumping obstacles of |
https://en.wikipedia.org/wiki/Cure | A cure is a substance or procedure that ends a medical condition, such as a medication, a surgical operation, a change in lifestyle or even a philosophical mindset that helps end a person's sufferings; or the state of being healed, or cured. The medical condition could be a disease, mental illness, genetic disorder, or simply a condition a person considers socially undesirable, such as baldness or lack of breast tissue.
An incurable disease may or may not be a terminal illness; conversely, a curable illness can still result in the patient's death.
The proportion of people with a disease that are cured by a given treatment, called the cure fraction or cure rate, is determined by comparing disease-free survival of treated people against a matched control group that never had the disease.
Another way of determining the cure fraction and/or "cure time" is by measuring when the hazard rate in a diseased group of individuals returns to the hazard rate measured in the general population.
Inherent in the idea of a cure is the permanent end to the specific instance of the disease. When a person has the common cold, and then recovers from it, the person is said to be cured, even though the person might someday catch another cold. Conversely, a person that has successfully managed a disease, such as diabetes mellitus, so that it produces no undesirable symptoms for the moment, but without actually permanently ending it, is not cured.
Related concepts, whose meaning can differ, include response, remission and recovery.
Statistical model
In complex diseases, such as cancer, researchers rely on statistical comparisons of disease-free survival (DFS) of patients against matched, healthy control groups. This logically rigorous approach essentially equates indefinite remission with cure. The comparison is usually made through the Kaplan-Meier estimator approach.
The simplest cure rate model was published by Joseph Berkson and Robert P. Gage in 1952. In this model, the surviva |
https://en.wikipedia.org/wiki/RLC%20circuit | An RLC circuit is an electrical circuit consisting of a resistor (R), an inductor (L), and a capacitor (C), connected in series or in parallel. The name of the circuit is derived from the letters that are used to denote the constituent components of this circuit, where the sequence of the components may vary from RLC.
The circuit forms a harmonic oscillator for current, and resonates in a manner similar to an LC circuit. Introducing the resistor increases the decay of these oscillations, which is also known as damping. The resistor also reduces the peak resonant frequency. Some resistance is unavoidable even if a resistor is not specifically included as a component.
RLC circuits have many applications as oscillator circuits. Radio receivers and television sets use them for tuning to select a narrow frequency range from ambient radio waves. In this role, the circuit is often referred to as a tuned circuit. An RLC circuit can be used as a band-pass filter, band-stop filter, low-pass filter or high-pass filter. The tuning application, for instance, is an example of band-pass filtering. The RLC filter is described as a second-order circuit, meaning that any voltage or current in the circuit can be described by a second-order differential equation in circuit analysis.
The three circuit elements, R, L and C, can be combined in a number of different topologies. All three elements in series or all three elements in parallel are the simplest in concept and the most straightforward to analyse. There are, however, other arrangements, some with practical importance in real circuits. One issue often encountered is the need to take into account inductor resistance. Inductors are typically constructed from coils of wire, the resistance of which is not usually desirable, but it often has a significant effect on the circuit.
Basic concepts
Resonance
An important property of this circuit is its ability to resonate at a specific frequency, the resonance frequency, . Frequencies a |
https://en.wikipedia.org/wiki/Incomplete%20Fermi%E2%80%93Dirac%20integral | In mathematics, the incomplete Fermi–Dirac integral for an index j is given by
This is an alternate definition of the incomplete polylogarithm.
See also
Complete Fermi–Dirac integral
External links
GNU Scientific Library - Reference Manual
Special functions |
https://en.wikipedia.org/wiki/Endolithic%20lichen | An endolithic lichen is a crustose lichen that grows inside solid rock, growing between the grains, with only the fruiting bodies exposed to the air. An example is Caloplaca luteominea subspecies bolandri. |
https://en.wikipedia.org/wiki/Boletus%20separans | Boletus separans is a species of bolete fungus in the family Boletaceae. It was described as new to science in 1873 by American mycologist Charles Horton Peck. In 1998, Roy Halling and Ernst Both transferred the bolete to the genus Xanthoconium. Molecular phylogenetic analysis published in 2013 shows that it is more closely related to Boletus sensu stricto than to Xanthoconium.
The species is a choice edible mushroom.
See also
List of Boletus species
List of North American boletes |
https://en.wikipedia.org/wiki/Travelport | Travelport Worldwide Ltd provides distribution, technology, payment solutions for the travel and tourism industry. It is the smallest, by revenue, of the top three global distribution systems (GDS) after Amadeus IT Group and Sabre Corporation.
The company also provides IT services to airlines, such as shopping, ticketing, and departure control.
History
The company was formed by Cendant in 2001 following its acquisitions of Galileo GDS for $2.9 billion and CheapTickets for $425 million.
In 2004, the company acquired Orbitz for $1.25 billion and Flairview Travel for $88 million.
In 2005, the company acquired eBookers for $350 million and Gullivers Travel Associates for $1.1 billion.
In August 2006, Cendant sold Orbitz and Galileo to The Blackstone Group for $4.3 billion, forming Travelport.
In August 2007, Travelport acquired Worldspan for $1.4 billion.
In July 2007, the company completed the partial corporate spin-off of Orbitz via an initial public offering.
In May 2010, the company acquired Sprice.com.
In 2011, the company sold Gullivers Travel Associates to Kuoni Travel for $720 million.
On September 25, 2014, the company became a public company via an initial public offering on the New York Stock Exchange.
In 2015, Travelport acquired Mobile Travel Technologies for €55 million.
On March 10, 2023, the company acquired Deem from Enterprise Holdings for an undisclosed amount.
Awards
In 2017, Travelport was the first GDS to be awarded the International Air Transport Association NDC (New Distribution Capability) Level 3 certification as an aggregator of travel content. In 2018, it became the first GDS operator to manage the live booking of flights using the NDC standard.
Acquisition
On May 30, 2019, the company was acquired by affiliates of Siris Capital Group and Evergreen Coast Capital, an affiliate of Elliott Management Corporation, for $4.4 billion. |
https://en.wikipedia.org/wiki/Paramormyrops%20hopkinsi | Paramormyrops hopkinsi is a species of freshwater electric fish. It was discovered in the Ivindo River in Gabon, in west-Central Africa by Dr. Carl D. Hopkins of Cornell University. It is distributed throughout the Ivindo River basin of Gabon and the Ntem River basin of Cameroon. Described originally as a Brienomyrus in 1985 it was transferred to Paramormyrops in 2007. The electric discharge has two phases: a head-positive phase followed by a head-negative phase. The mean duration of the EOD is 2.8 ms for females, 2.96 for males. The Fourier transform of the EOD peaks at 536 Hz for females, 468 for males. Both male and female EODs have a head-negative voltage bump about 5 to 6 ms after the main head positive phase (arrows). |
https://en.wikipedia.org/wiki/La%20Abad%C3%ADa%20del%20Crimen | La abadía del crimen (The Abbey of Crime) is a video game written by Paco Menéndez with graphics made by Juan Delcán and published in 1987 by Opera Soft. It was conceived as a version of Umberto Eco's 1980 book The Name of the Rose. Paco Menéndez and Opera Soft were unable to secure the rights for the name, so the game was released as La abadía del crimen. "The Abbey of the Crime" was the working title of the novel The Name of the Rose.
This game is an adventure with isometric graphics. A Franciscan friar, William of Occam (William of Baskerville in the book) and his young novice Adso have to discover the perpetrator of a series of murders in a medieval Italian abbey.
Gameplay
The player controls the movement of the friar Fra William (mistakenly described as a monk in the user manual). The player also has the possibility of controlling the movement of the novice Adso within the same screen in which Fra William is. If the key for controlling the novice is not pressed, he follows Fra William most of the time. The game features other characters representing the monks of the abbey who behave according to programmed artificial intelligence to move throughout the mapping of the abbey and show a series of dialogs shown by written text which is moved along the lower part of the screen.
An extensive mapping of the abbey is represented in a large series of screens with 3D isometric graphics. A series of objects has to be collected in order to successfully complete the game. The action occurs in seven days subdivided in different Canonical hours. The time (day + current hour) is indicated at the bottom left of the screen.
The game starts with the abbot welcoming Fra William and explaining that a monk has disappeared. He also explains to Fra William that he is obligated to obey the orders of the abbot and the rules of the monastery, attend religious services and meals and stay in his cell at night while the research of the crimes is pursued. During the game, the novice Ad |
https://en.wikipedia.org/wiki/Bound%20entanglement | Bound entanglement is a weak form of quantum entanglement, from which no singlets can be distilled with local operations and classical communication (LOCC).
Bound entanglement was discovered by M. Horodecki, P. Horodecki, and R. Horodecki. Bipartite entangled states that have a non-negative partial transpose are all bound-entangled. Moreover, a particular quantum state for 2x4 systems has been presented. Such states are not detected by the Peres-Horodecki criterion as entangled, thus other entanglement criteria are needed for their detection. There are a number of examples for such states.
There are also multipartite entangled states that have a negative partial transpose with respect to some bipartitions, while they have a positive partial transpose to the other partitions, nevertheless, they are undistillable.
The possible existence of bipartite bound entangled states with a negative partial transpose is still under intensive study.
Properties of bound entangled states with a positive partial transpose
Bipartite bound entangled states do not exist in 2x2 or 2x3 systems, only in larger ones.
Rank-2 bound entangled states do not exist.
Bipartite bound entangled states with a positive partial transpose are useless for teleportation, as they cannot lead to a larger fidelity than the classical limit.
Bound entangled states with a positive partial transpose in 3x3 systems have a Schmidt number 2.
It has been shown that bipartite bound entangled states with a positive partial transpose exist in symmetric systems. It has also been shown that in symmetric systems multipartite bound entangled states exists for which all partial transposes are non-negative.
Asher Peres conjectured that bipartite bound entangled states with positive partial transpose cannot violate a Bell inequality. After a long search for counterexamples, the conjecture turned out to be false.
While no singlets can be distilled from bound entangled state, they can be still useful for some quant |
https://en.wikipedia.org/wiki/Boris%20Beizer | Boris Beizer (1934-2018) was an American software engineer and author. He received his B.S. degree in physics from the City College of New York in 1956, an MS in Electrical Engineering (1963) and a PhD in computer science from the University of Pennsylvania in 1966. He wrote many books and articles on topics such as system architecture and software testing. His books Software Testing Techniques and Software System Testing and Quality Assurance are frequently consulted references on the subject. He directed testing for the FAA's Weather Message Switching Center and several other large communications systems. He was a speaker at many testing conferences and was also known for his seminars on testing. He consulted on software testing and quality assurance with many organizations throughout the world.
Personal life
Boris Beizer was born in Brussels, Belgium. He emigrated to the United States in May, 1941.
List of publications (partial)
Software Quality Reflections Essays, dialogues and poems, 2000
Black-box testing: techniques for functional testing of software and systems (1995) ; Japanese edition, Nikei
The frozen keyboard: living with bad software (1988)
Personal computer quality: a guide for victims and vendors (1986)
Software system testing and quality assurance (1984)
Software testing techniques (1983) ; expanded Second edition 1990; Japanese Edition Nikei
Micro-analysis of computer system performance (1978) ; Russian Edition
Communications Processor System (1977) with Kenneth Hagstrom
The architecture and engineering of digital computer complexes, Volume 1 (1971), Volume 2; Polish and Russian Editions
Engineering applications of Boolean algebra (1958) with Stephen W. Leibholz
Novels under pseudonym Ethan I. Shedley |
https://en.wikipedia.org/wiki/Relief%20Therapeutics | Relief Therapeutics is a Swiss biopharmaceutical company based in Geneva. The company focuses on developing drugs for serious diseases with few or no existing treatment options. Its lead compound, RLF-100, is a synthetic form of a natural peptide that protects the lung. The company was incorporated as Relief Therapeutics Holdings AG (RFLB.S) and listed on the SIX Swiss Exchange in 2016.
History
Relief Therapeutics was founded in 2013 by Gael Hédou with the aim of developing new treatments for diseases with high unmet needs. The company today considers itself the successor to Mondobiotech, which was founded in 2000 by Fabio Cavalli and Dorian Bevec. Mondobiotech began research into Vasoactive intestinal peptide (VIP), a naturally occurring substance in humans that was first identified in the 1970s. They were granted US and European patents for a synthetic version of VIP known as aviptadil in 2006.
On June 23, 2013, Mondobiotech merged with Italian pharmaceutical company Pierrel Research International to form a new Contract research organization known as Therametrics. On July 14, 2016, Therametrics merged with Relief Therapeutics to form Relief Therapeutics Holdings AG, which inherited all patents related to aviptadil.
COVID-19 research
Aviptadil
In the wake of the COVID-19 pandemic, scientists at Relief conducted initial studies into the efficacy of RLF-100 in treating severe COVID-19 patients. In June 2020, the U.S. Food and Drug Administration granted fast-track designation to RLF-100 for treatment of respiratory distress in COVID-19. In September 2020, Relief partnered with US-Israeli firm NRX Pharmaceuticals (formerly NeuroRx Inc) for the co-development of the drug and the co-ordination of US trials. In April 2021, a reformulated version of aviptadil, known as Zyesami, was included in a National Institutes of Health (NIH) sponsored Phase 3 trial with the aim of testing aviptadil against remdesivir. In May 2021, NRX submitted a request for an Emergency Use Aut |
https://en.wikipedia.org/wiki/Simplicial%20presheaf | In mathematics, more specifically in homotopy theory, a simplicial presheaf is a presheaf on a site (e.g., the category of topological spaces) taking values in simplicial sets (i.e., a contravariant functor from the site to the category of simplicial sets). Equivalently, a simplicial presheaf is a simplicial object in the category of presheaves on a site. The notion was introduced by A. Joyal in the 1970s. Similarly, a simplicial sheaf on a site is a simplicial object in the category of sheaves on the site.
Example: Consider the étale site of a scheme S. Each U in the site represents the presheaf . Thus, a simplicial scheme, a simplicial object in the site, represents a simplicial presheaf (in fact, often a simplicial sheaf).
Example: Let G be a presheaf of groupoids. Then taking nerves section-wise, one obtains a simplicial presheaf . For example, one might set . These types of examples appear in K-theory.
If is a local weak equivalence of simplicial presheaves, then the induced map is also a local weak equivalence.
Homotopy sheaves of a simplicial presheaf
Let F be a simplicial presheaf on a site. The homotopy sheaves of F is defined as follows. For any in the site and a 0-simplex s in F(X), set and . We then set to be the sheaf associated with the pre-sheaf .
Model structures
The category of simplicial presheaves on a site admits many different model structures.
Some of them are obtained by viewing simplicial presheaves as functors
The category of such functors is endowed with (at least) three model structures, namely the projective, the Reedy, and the injective model structure. The weak equivalences / fibrations in the first are maps
such that
is a weak equivalence / fibration of simplicial sets, for all U in the site S. The injective model structure is similar, but with weak equivalences and cofibrations instead.
Stack
A simplicial presheaf F on a site is called a stack if, for any X and any hypercovering H →X, the canonical map
is a weak |
https://en.wikipedia.org/wiki/Visual%20Instruction%20Set | Visual Instruction Set, or VIS, is a SIMD instruction set extension for SPARC V9 microprocessors developed by Sun Microsystems. There are five versions of VIS: VIS 1, VIS 2, VIS 2+, VIS 3 and VIS 4.
History
VIS 1 was introduced in 1994 and was first implemented by Sun in their UltraSPARC microprocessor (1995) and by Fujitsu in their SPARC64 GP microprocessors (2000).
VIS 2 was first implemented by the UltraSPARC III. All subsequent UltraSPARC and SPARC64 microprocessors implement the instruction set.
VIS 3 was first implemented in the SPARC T4 microprocessor.
VIS 4 was first implemented in the SPARC M7 microprocessor.
Differences vs x86
VIS is not an instruction toolkit like Intel's MMX and SSE. MMX has only 8 registers shared with the FPU stack, while SPARC processors have 32 registers, also aliased to the double-precision (64-bit) floating point registers.
As with the SIMD instruction set extensions on other RISC processors, VIS strictly conforms to the main principle of RISC: keep the instruction set concise and efficient.
This design is very different from comparable extensions on CISC processors, such as MMX, SSE, SSE2, SSE3, SSE4, 3DNow!.
Sometimes, programmers must use several VIS instructions to accomplish an operation that can be done with only one MMX or SSE instruction, but it should be kept in mind that fewer instructions do not automatically result in better performance.
Functionality
VIS re-uses existing SPARC V9 64-bit floating point registers to hold multiple 8, 16, or 32-bit integer values. In this respect, VIS is more similar to the design of MMX than other SIMD architectures such as SSE/SSE2/AltiVec.
VIS includes a number of operations primarily for graphics support, so most of them are only for integers. These include 3D to 2D conversion, edge processing and pixel distance.
There are four ways to use VIS in code:
The GCC - option
Use inline assembly
Use inline template in VSDK, similar to compiler intrinsics, which have C functio |
https://en.wikipedia.org/wiki/Frederik%20Zeuthen | Frederik Ludvig Bang Zeuthen (9 September 1888 – 24 February 1959) was a Danish economist. He became an internationally recognized economist in the 1930s and published his research in English, French and German, as well as Danish. He was especially known for his theoretical microeconomics work in general equilibrium theory and the theories of market influences and pricing. He was one of the pioneers of the mathematical theory of monopolistic competition. At the same time, he was interested in social policy and distribution of income.
Background and career
Frederik Zeuthen's father was the mathematician Hieronymus Georg Zeuthen. Frederik attended Østre Borgerdyd Gymnasium and then studied economics at the University of Copenhagen. His teachers Lauritz Vilhelm Birck and Harald Ludvig Westergaard inspired him to do research in economics. After working for some years outside of academe, Zeuthen earned in 1928 a doctorate with a dissertation Den Økonomiske Fordeling (The Economic Distribution) on price formation. He was appointed a professor in economics in 1930—a position he held until retiring in 1958. He died the following year at age 70.
Economic research
Zeuthen's book Problems of Monopoly and Economic Warfare, published in London in 1930 with a foreword by the famous economist Joseph Schumpeter, established for Zeuthen an international reputation among economists. Internationally, the term Zeuthen strategy is still used for a particular negotiation strategy studied in game theory. Zeuthen's article Das Prinzip der Knappheit, technische Kombination und Ökonomische Qualität (The principle of scarcity, technical combination and economic quality), published in the Austrian Zeitschrift für Nationalökonomie in 1933, was ground-breaking, as the article presented the first formulation using mathematical inequalities for Léon Walras's general theory of economic equilibrium.
However, Frederik Zeuthen's main theoretical work was his book Økonomisk Teori og Metode, publishe |
https://en.wikipedia.org/wiki/Ajilim%C3%B3jili | Ajilimójili is a hot or hot and sweet chili sauce from Puerto Rico, traditionally served over grilled seafood, vegetables, pasteles, boiled tuber vegetables and especially grilled meats.
Description
Ajilimójili is a combination of olive oil, garlic, cilantro, chilies, green bell pepper, cumin, oregano, vinegar, citrus (lemon, lime or sour orange) chopped or blended, simmered and cooled to serve. A variant, sweet ajilimójili, adds ingredients such as sweet red peppers, ajicitos, honey, tomato sauce and butter.
The sauce is one of the essential elements of Puerto Rican cooking.
See also
Puerto Rican cuisine
Salsa (sauce)
Mojo (sauce) |
https://en.wikipedia.org/wiki/American%20Society%20for%20Mass%20Spectrometry | The American Society for Mass Spectrometry (ASMS) is a professional association based in the United States that supports the scientific field of mass spectrometry. As of 2018, the society had approximately 10,000 members primarily from the US, but also from around the world. The society holds a large annual meeting, typically in late May or early June as well as other topical conferences and workshops. The society publishes the Journal of the American Society for Mass Spectrometry.
Awards
The Society recognizes achievements and promotes academic research through four annual awards. The Biemann Medal and the John B. Fenn Award for a Distinguished Contribution in Mass Spectrometry both are awarded in recognition of singular achievements or contributions in fundamental or applied mass spectrometry, with the Biemann Medal being focused on individuals who are early in their careers. The Ronald A. Hites Award is awarded for outstanding original research demonstrated in papers published in the Journal of the American Society for Mass Spectrometry. The Research Awards are given to young scientists in mass spectrometry, based on the evaluation of their proposed research.
Publications
Journal of the American Society for Mass Spectrometry
Measuring Mass: From Positive Rays to Proteins
Past presidents
The past presidents of ASMS are:
Conferences
The Society holds an annual conference in late May or early June as well as topical conferences (at Asilomar State Beach in California and Sanibel Island, Florida) and a fall workshop, which is also focused on a single topic. Conferences on Mass Spectrometry and Allied Topics have been held yearly since 1953.
See also
International Mass Spectrometry Foundation
List of female mass spectrometrists |
https://en.wikipedia.org/wiki/Satellite%20navigation%20software | Satellite navigation software or GNSS navigation software usually falls into one of the following two categories:
Navigation with route calculation and directions from the software to the user of the route to take, based on a vector-based map, normally for motorized vehicles with some motorized forms added on as an afterthought.
Navigation tracking, often with a map "picture" in the background, but showing where you have been, and allowing "routes" to be preprogrammed, giving a line you can follow on the screen. This type can also be used for geocaching.
Terminology
Track
A track is a trace of somewhere that you have actually been (often called a "breadcrumb trail"). The GNSS unit (external or internal) periodically sends details of the location which are recorded by the software, either by taking a reading based on a set time interval, based on a set distance, based on a change in direction by more than a certain angle, or a combination of these. Each point is stored together with its date and time. The resulting track can be displayed as a series of the recorded points or a line connecting them.
Retracing your steps is a simple matter of following the track back to the source.
Route
A route is a preset series of points that make up a set route to follow for your destination. Most software allows the route and the track to be displayed at the same time.
Waypoint
Waypoints are used to mark particular locations, typically used as markers along the "way" to somewhere. They are either key entered by users or downloaded from other sources, depending upon the sophistication of the device. Although not linked to tracks or routes, they can be used to simplify the construction of routes, by being able to be re-used. Frequently, waypoints serve a "safety" purpose, enabling a route to be taken around obstacles such as shallow water (marine navigation) or streams/cliffs/other hazards which may prevent a safe passage directly from point "A" to point "B".
Platforms
|
https://en.wikipedia.org/wiki/Interfacial%20thermal%20resistance | Interfacial thermal resistance, also known as thermal boundary resistance, or Kapitza resistance, is a measure of resistance to thermal flow at the interface between two materials. While these terms may be used interchangeably, Kapitza resistance technically refers to an atomically perfect, flat interface whereas thermal boundary resistance is a more broad term. This thermal resistance differs from contact resistance (not to be confused with electrical contact resistance) because it exists even at atomically perfect interfaces. Owing to differences in electronic and vibrational properties in different materials, when an energy carrier (phonon or electron, depending on the material) attempts to traverse the interface, it will scatter at the interface. The probability of transmission after scattering will depend on the available energy states on side 1 and side 2 of the interface.
Assuming a constant thermal flux is applied across an interface, this interfacial thermal resistance will lead to a finite temperature discontinuity at the interface. From an extension of Fourier's law, we can write
where is the applied flux, is the observed temperature drop, is the thermal boundary resistance, and is its inverse, or thermal boundary conductance.
Understanding the thermal resistance at the interface between two materials is of primary significance in the study of its thermal properties. Interfaces often contribute significantly to the observed properties of the materials. This is even more critical for nanoscale systems where interfaces could significantly affect the properties relative to bulk materials.
Low thermal resistance at interfaces is technologically important for applications where very high heat dissipation is necessary. This is of particular concern to the development of microelectronic semiconductor devices as defined by the International Technology Roadmap for Semiconductors in 2004 where an 8 nm feature size device is projected to generate up to |
https://en.wikipedia.org/wiki/Buffer%20strip | A buffer strip is an area of land maintained in permanent vegetation that helps to control air quality, soil quality, and water quality, along with other environmental problems, dealing primarily on land that is used in agriculture. Buffer strips trap sediment, and enhance filtration of nutrients and pesticides by slowing down surface runoff that could enter the local surface waters. The root systems of the planted vegetation in these buffers hold soil particles together which alleviate the soil of wind erosion and stabilize stream banks providing protection against substantial erosion and landslides. Farmers can also use buffer strips to square up existing crop fields to provide safety for equipment while also farming more efficiently.
Buffer strips can have several different configurations of vegetation found on them varying from simply grass to combinations of grass, trees, and shrubs. Areas with diverse vegetation provide more protection from nutrient and pesticide flow and at the same time provide better biodiversity amongst plants and animals.
Many country, state, and local governments provide financial incentives for conservation programs such as buffer strips because they help stabilize the environment, help reduce nitrogen emissions to water and soil loss by wind erosion, while simultaneously providing substantial environmental co-benefits, even when the land is being used. Buffer strips not only stabilize the land but can also provide a visual demonstration that land is under stewardship.
Types
Buffers within fields
A grassed waterway reduces soil erosion and captures most nutrients and pesticides that would normally wash out of crop fields and into major waters. These waterways help to carry surface water at a non-erosive velocity to an area where it will have a stable outlet. Outlets must be adequate enough to allow water to drain without ponding or flooding the area being protected, while also preventing erosion of the water into the outlet which ca |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.