source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Switched%20fabric
Switched fabric or switching fabric is a network topology in which network nodes interconnect via one or more network switches (particularly crossbar switches). Because a switched fabric network spreads network traffic across multiple physical links, it yields higher total throughput than broadcast networks, such as the early 10BASE5 version of Ethernet and most wireless networks such as Wi-Fi. The generation of high-speed serial data interconnects that appeared in 2001–2004 which provided point-to-point connectivity between processor and peripheral devices are sometimes referred to as fabrics; however, they lack features such as a message-passing protocol. For example, HyperTransport, the computer processor interconnect technology, continues to maintain a processor bus focus even after adopting a higher speed physical layer. Similarly, PCI Express is just a serial version of PCI; it adheres to PCI's host/peripheral load/store direct memory access (DMA)-based architecture on top of a serial physical and link layer. Fibre Channel In the Fibre Channel Switched Fabric (FC-SW-6) topology, devices are connected to each other through one or more Fibre Channel switches. While this topology has the best scalability of the three FC topologies (the other two are Arbitrated Loop and point-to-point), it is the only one requiring switches, which are costly hardware devices. Visibility among devices (called nodes) in a fabric is typically controlled with Fibre Channel zoning. Multiple switches in a fabric usually form a mesh network, with devices being on the "edges" ("leaves") of the mesh. Most Fibre Channel network designs employ two separate fabrics for redundancy. The two fabrics share the edge nodes (devices), but are otherwise unconnected. One of the advantages of such setup is capability of failover, meaning that in case one link breaks or a fabric goes out of order, datagrams can be sent via the second fabric. The fabric topology allows the connection of up to the
https://en.wikipedia.org/wiki/Diffusing%20capacity
Diffusing capacity of the lung (DL) (also known as Transfer factor is another expression for the formerly used diffusing capacity.) measures the transfer of gas from air in the lung, to the red blood cells in lung blood vessels. It is part of a comprehensive series of pulmonary function tests to determine the overall ability of the lung to transport gas into and out of the blood. DL, especially DLCO, is reduced in certain diseases of the lung and heart. DLCO measurement has been standardized according to a position paper by a task force of the European Respiratory and American Thoracic Societies. In respiratory physiology, the diffusing capacity has a long history of great utility, representing conductance of gas across the alveolar-capillary membrane and also takes into account factors affecting the behaviour of a given gas with hemoglobin. The term may be considered a misnomer as it represents neither diffusion nor a capacity (as it is typically measured under submaximal conditions) nor capacitance. In addition, gas transport is only diffusion limited in extreme cases, such as for oxygen uptake at very low ambient oxygen or very high pulmonary blood flow. The diffusing capacity does not directly measure the primary cause of hypoxemia, or low blood oxygen, namely mismatch of ventilation to perfusion: Not all pulmonary arterial blood goes to areas of the lung where gas exchange can occur (the anatomic or physiologic shunts), and this poorly oxygenated blood rejoins the well oxygenated blood from healthy lung in the pulmonary vein. Together, the mixture has less oxygen than that blood from the healthy lung alone, and so is hypoxemic. Similarly, not all inspired air goes to areas of the lung where gas exchange can occur (the anatomic and the physiological dead spaces), and so is wasted. Testing The single-breath diffusing capacity test is the most common way to determine . The test is performed by having the subject blow out all of the air that they can, le
https://en.wikipedia.org/wiki/Texas%20Instruments%20signing%20key%20controversy
The Texas Instruments signing key controversy resulted from Texas Instruments' (TI) response to a project to factorize the 512-bit RSA cryptographic keys needed to write custom firmware to TI devices. Project In July 2009, Benjamin Moody, a United-TI forum user, published the factors of a 512-bit RSA key used to sign the TI-83+ series graphing calculator. The discovery of the private key would allow end users to flash their own operating systems onto the device without having to use any special software. Moody used two free implementations of the general number field sieve, msieve and ggnfs; the computation took 73 days on a 1.9 GHz dual-core processor. This demonstrates the progress of hardware development: the factorization of the similar 512-bit RSA-155 in 1999 using the same algorithm required a large dedicated research group, 8000 MIPS-years of computing time, and a Cray C916 supercomputer. In response, members of the wider TI graphing calculators community (at yAronet) set up a BOINC-based distributed computing project, RSA Lattice Siever (RSALS for short), that quickly factored the other keys. RSA Lattice Siever remained active for nearly three years after outliving its initial purpose, by factoring other integers for the mathematical community. After factoring over 400 integers, RSALS moved to RSALS-inspired NFS@home at the end of August 2012. Legal response Texas Instruments began by sending out two initial Digital Millennium Copyright Act (DMCA) take-down requests to the hackers, referring to sites or forum posts that they controlled. The hackers responded by removing the keys, without consulting an attorney. TI then sent further DMCA notices to a variety of websites displaying the keys, including United-TI, Reddit, and Wikipedia. Texas Instruments' efforts then became subject to the Streisand effect, and the keys were mirrored on a number of sites, including WikiLeaks. In September 2009, Dan Goodin from The Register alerted the Electronic Frontier Foun
https://en.wikipedia.org/wiki/Unix%20architecture
A Unix architecture is a computer operating system system architecture that embodies the Unix philosophy. It may adhere to standards such as the Single UNIX Specification (SUS) or similar POSIX IEEE standard. No single published standard describes all Unix architecture computer operating systems — this is in part a legacy of the Unix wars. Description There are many systems which are Unix-like in their architecture. Notable among these are the Linux distributions. The distinctions between Unix and Unix-like systems have been the subject of heated legal battles, and the holders of the UNIX brand, The Open Group, object to "Unix-like" and similar terms. For distinctions between SUS branded UNIX architectures and other similar architectures, see Unix-like. Kernel A Unix kernel — the core or key components of the operating system — consists of many kernel subsystems like process management, scheduling, file management, device management, network management, memory management, and dealing with interrupts from hardware devices. Each of the subsystems has some features: Concurrency: As Unix is a multiprocessing OS, many processes run concurrently to improve the performance of the system. Virtual memory (VM): Memory management subsystem implements the virtual memory concept and users need not worry about the executable program size and the RAM size. Paging: It is a technique to minimize the internal as well as the external fragmentation in the physical memory. Virtual file system (VFS): A VFS is a file system used to help the user to hide the different file systems complexities. A user can use the same standard file system related calls to access different file systems. The kernel provides these and other basic services: interrupt and trap handling, separation between user and system space, system calls, scheduling, timer and clock handling, file descriptor management. Features Some key features of the Unix architecture concept are: Unix systems use a central
https://en.wikipedia.org/wiki/Proximal%20gradient%20methods%20for%20learning
Proximal gradient (forward backward splitting) methods for learning is an area of research in optimization and statistical learning theory which studies algorithms for a general class of convex regularization problems where the regularization penalty may not be differentiable. One such example is regularization (also known as Lasso) of the form Proximal gradient methods offer a general framework for solving regularization problems from statistical learning theory with penalties that are tailored to a specific problem application. Such customized penalties can help to induce certain structure in problem solutions, such as sparsity (in the case of lasso) or group structure (in the case of group lasso). Relevant background Proximal gradient methods are applicable in a wide variety of scenarios for solving convex optimization problems of the form where is convex and differentiable with Lipschitz continuous gradient, is a convex, lower semicontinuous function which is possibly nondifferentiable, and is some set, typically a Hilbert space. The usual criterion of minimizes if and only if in the convex, differentiable setting is now replaced by where denotes the subdifferential of a real-valued, convex function . Given a convex function an important operator to consider is its proximal operator defined by which is well-defined because of the strict convexity of the norm. The proximal operator can be seen as a generalization of a projection. We see that the proximity operator is important because is a minimizer to the problem if and only if where is any positive real number. Moreau decomposition One important technique related to proximal gradient methods is the Moreau decomposition, which decomposes the identity operator as the sum of two proximity operators. Namely, let be a lower semicontinuous, convex function on a vector space . We define its Fenchel conjugate to be the function The general form of Moreau's decomposition states that for any
https://en.wikipedia.org/wiki/Davies%20attack
In cryptography, the Davies attack is a dedicated statistical cryptanalysis method for attacking the Data Encryption Standard (DES). The attack was originally created in 1987 by Donald Davies. In 1994, Eli Biham and Alex Biryukov made significant improvements to the technique. It is a known-plaintext attack based on the non-uniform distribution of the outputs of pairs of adjacent S-boxes. It works by collecting many known plaintext/ciphertext pairs and calculating the empirical distribution of certain characteristics. Bits of the key can be deduced given sufficiently many known plaintexts, leaving the remaining bits to be found through brute force. There are tradeoffs between the number of required plaintexts, the number of key bits found, and the probability of success; the attack can find 24 bits of the key with 252 known plaintexts and 53% success rate. The Davies attack can be adapted to other Feistel ciphers besides DES. In 1998, Pornin developed techniques for analyzing and maximizing a cipher's resistance to this kind of cryptanalysis.
https://en.wikipedia.org/wiki/Teenage%20Mutant%20Ninja%20Turtles%20%28NES%20video%20game%29
Teenage Mutant Ninja Turtles, known as in Japan and Teenage Mutant Hero Turtles in Europe, is a 1989 side-scrolling action-platform game for the Nintendo Entertainment System released by Konami. In North America it was published under Konami's Ultra Games imprint in the US and the equivalent PALCOM brand in Europe and Australia. Alongside the arcade game (also developed by Konami), it was one of the first video games based on the 1987 Teenage Mutant Ninja Turtles animated series, being released after the show's second season. The game sold more than cartridges worldwide. Plot Shredder kidnaps April and gains the Life Transformer Gun, a weapon capable of returning Splinter to his human form. In order to save April, the turtles (Leo, Mikey, Donny and Raph) embark on the streets of New York to confront the Foot Clan. While traversing the sewers, the turtles encounter Bebop, a mutated pig, and Rocksteady, a mutant rhino. Though the turtles defeat Bebop, Rocksteady escapes with April O’Neil. The turtles then chase Rocksteady to an abandoned warehouse, fight him, and rescue April. After disabling bombs in the Hudson River dam, Shredder captures Splinter, so the turtles give chase in the Party Wagon. Hot in pursuit, the turtles scour the city and eventually find that Splinter is held captive by the robotic Mecaturtle on a skyscraper rooftop. After the turtles save Splinter, Shredder escapes in a helicopter. The turtles give chase, tracking him to JFK airport, where they encounter Big Mouser. After defeating Big Mouser, the turtles head to Shredder's secret Foot Clan base in the South Bronx via the Turtle Blimp. Once there, they locate and battle the Technodrome underground. The turtles descend into the Technodrome’s reactor and ultimately defeat Shredder. With the Life Transformer Gun, the turtles help Splinter return to his human form. With a tough mission accomplished, the turtles and April celebrate with a pizza. Gameplay Teenage Mutant Ninja Turtles is a s
https://en.wikipedia.org/wiki/Pea
The pea is most commonly the small spherical seed or the seed-pod of the flowering plant species Lathyrus oleraceus. Each pod contains several peas, which can be green or yellow. Botanically, pea pods are fruit, since they contain seeds and develop from the ovary of a (pea) flower. The name is also used to describe other edible seeds from the Fabaceae such as the pigeon pea (Cajanus cajan), the cowpea (Vigna unguiculata), and the seeds from several species of Lathyrus. Peas are annual plants, with a life cycle of one year. They are a cool-season crop grown in many parts of the world; planting can take place from winter to early summer depending on location. The average pea weighs between 0.1 and 0.36 gram. The immature peas (and in snow peas the tender pod as well) are used as a vegetable, fresh, frozen or canned; varieties of the species typically called field peas are grown to produce dry peas like the split pea shelled from a matured pod. These are the basis of pease porridge and pea soup, staples of medieval cuisine; in Europe, consuming fresh immature green peas was an innovation of early modern cuisine. Description A pea is a most commonly green, occasionally golden yellow, or infrequently purple pod-shaped vegetable, widely grown as a cool-season vegetable crop. The seeds may be planted as soon as the soil temperature reaches , with the plants growing best at temperatures of . They do not thrive in the summer heat of warmer temperate and lowland tropical climates, but do grow well in cooler, high-altitude, tropical areas. Many cultivars reach maturity about 60 days after planting. Peas have both low-growing and vining cultivars. The vining cultivars grow thin tendrils from leaves that coil around any available support and can climb to be high. A traditional approach to supporting climbing peas is to thrust branches pruned from trees or other woody plants upright into the soil, providing a lattice for the peas to climb. Branches used in this fashion are
https://en.wikipedia.org/wiki/Mass%20spectrometric%20immunoassay
Mass spectrometric immunoassay (MSIA) is a rapid method is used to detect and/ or quantify antigens and or antibody analytes. This method uses an analyte affinity (either through antigens or antibodies) isolation to extract targeted molecules and internal standards from biological fluid in preparation for matrix assisted laser desorption ionization-time of flight mass spectrometry (MALDI-TOF-MS). This method allows for "top down" and "bottom up" analysis. This sensitive method allows for a new and improved process for detecting multiple antigens and antibodies in a single assay. This assay is also capable of distinguishing mass shifted forms of the same molecule via a panantibody, as well as distinguish point mutations in proteins. Each specific form is detected uniquely based on their characteristic molecular mass. MSIA has dual specificity because of the antibody-antigen reaction coupled with the power of a mass spectrometer. There are various other immunoassy techniques that have been used previously such as radioimmunoassay (RIA) and enzyme immunoassay (EIA and ELISA). These techniques are extremely sensitive however, there are many limitations to these methods. For example, quantification for ELISA and EIA require several hours because the binding has to reach equilibrium. RIA's disadvantage is that you need radioactive particles which are universally known to be carcinogens. The creation of MSIA fulfilled the need to determine the presence of one or more antigens in a specimen as well as the quantification of those said species. History This assay was patented in 2006 by Randall Nelson, Peter Williams and Jennifer Reeve Krone. The idea first came about with the development of ELISA and RIA. An earlier patent method suggested tagging antigens or antibodies with stable isotopes or long-lived radioactive elements. But limitations to both methods called for a better detection methods of a protein or proteins. The invention combines antigen-antibody binding
https://en.wikipedia.org/wiki/Synchrotron%20light%20source
A synchrotron light source is a source of electromagnetic radiation (EM) usually produced by a storage ring, for scientific and technical purposes. First observed in synchrotrons, synchrotron light is now produced by storage rings and other specialized particle accelerators, typically accelerating electrons. Once the high-energy electron beam has been generated, it is directed into auxiliary components such as bending magnets and insertion devices (undulators or wigglers) in storage rings and free electron lasers. These supply the strong magnetic fields perpendicular to the beam that are needed to stimulate the high energy electrons to emit photons. The major applications of synchrotron light are in condensed matter physics, materials science, biology and medicine. A large fraction of experiments using synchrotron light involve probing the structure of matter from the sub-nanometer level of electronic structure to the micrometer and millimeter levels important in medical imaging. An example of a practical industrial application is the manufacturing of microstructures by the LIGA process. Synchrotron is one of the most expensive kinds of light source known, but it is practically the only viable luminous source of wide-band radiation in far infrared wavelength range for some applications, such as far-infrared absorption spectrometry. Spectral brightness The primary figure of merit used to compare different sources of synchrotron radiation has been referred to as the "brightness", the "brilliance", and the "spectral brightness," with the latter term being recommended as the best choice by the Working Group on Synchrotron Nomenclature. Regardless of the name chosen, the term is a measure of the total flux of photons in a given six-dimensional phase space per unit bandwidth (BW). The spectral brightness is given by: where is the photons per second of the beam, and are the root mean square values for the size of the beam in the axes perpendicular to the beam
https://en.wikipedia.org/wiki/Pierre%20Boutroux
Pierre Léon Boutroux (; 6 December 1880 – 15 August 1922) was a French mathematician and historian of science. Boutroux is chiefly known for his work in the history and philosophy of mathematics. Biography He was born in Paris on 6 December 1880 into a well connected family of the French intelligentsia. His father was the philosopher Émile Boutroux. His mother was Aline Catherine Eugénie Poincaré, sister of the scientist and mathematician Henri Poincaré. A cousin of Aline, Raymond Poincaré was to be President of France. He occupied the mathematics chair at Princeton University from 1913 until 1914. He occupied the History of sciences chair from 1920 to 1922. Boutroux published his major work Les principes de l'analyse mathématique in two volumes; Volume 1 in 1914 and Volume 2 in 1919. This is a comprehensive view of the whole field of mathematics at the time. He was an Invited Speaker of the ICM in 1904 at Heidelberg, in 1908 at Rome, and in 1920 at Strasbourg. He died on 15 August 1922, aged 41 years. Works L'Imagination et les mathématiques selon Descartes (1900) Sur quelques propriétés des fonctions entières (1903) Œuvres de Blaise Pascal, publiées suivant l'ordre chronologique, avec documents complémentaires, introductions et notes, par Léon Brunschvicg et Pierre Boutroux (1908) Leçons sur les fonctions définies par les équations différentielles du premier ordre, professées au Collège de France (1908) Les Principes de l'analyse mathématique, exposé historique et critique (2 volumes, 1914-1919) Texte en ligne 1 2 Contient : (I) Les nombres, les grandeurs, les figures, le calcul combinatoire, le calcul algébrique, calcul des fonctions, l'algèbre géométrique. (2) La géométrie algébrique. Extensions de l'algèbre et constructions logiques. Extensions de l'algèbre ; les développements en séries. La méthode analytique en mathématiques. Analyse infinitésimale. Analyse des principes mathématiques. Analyse de la notion de fonction. L'Idéal scientifique des ma
https://en.wikipedia.org/wiki/PeopleFinders.com
PeopleFinders.com is a public records business located in Sacramento, California. The company offers a variety of public records including address histories, phone numbers and background checks. History In 1999 Rob Miller founded PeopleFinders. Products and services PeopleFinders is largely a "People Search" company, providing individuals with various types of public records that will allow them to obtain contact information for most private citizens in the United States. In addition, PeopleFinders.com offers background checks, criminal records and a variety of other public records related to marriage, divorce, birth, death, bankruptcy, and property ownership. This data is collected from various sources, including government agencies, property records and other public records sources. Also in the interest of finding people, PeopleFinders offers a reverse phone lookup service that allows users to find people or businesses behind unknown phone numbers. Blog The PeopleFinders Blog was launched in May 2009. This blog covered topics that are related to people searches, current events, how to obtain background checks and other public record reports, and other subjects of interest to their customers and the general public. New blog posts were typically published three times per week; every Monday, Wednesday, and Friday. Growth In 2008 PeopleFinders.com was noted for being one of the fastest growing internet based companies. Deloitte, an independent firm that provides audit and consulting services, listed the company as number 24 in the Technology Fast 50 for the Silicon Valley. This ranking was based on the amount of revenue growth the company experienced between 2003 and 2007. Mark Jensen, Managing Partner, National Venture Capital Services, Deloitte & Touche LLP, stated that PeopleFinders was one of a very few companies to "accomplish such a fast growth rate over the past five years". PeopleFinders was also listed as a Deloitte Fast 500 company in both 2008 and
https://en.wikipedia.org/wiki/Type-1.5%20superconductor
Type-1.5 superconductors are multicomponent superconductors characterized by two or more coherence lengths, at least one of which is shorter than the magnetic field penetration length , and at least one of which is longer. This is in contrast to single-component superconductors, where there is only one coherence length and the superconductor is necessarily either type 1 () or type 2 () (often a coherence length is defined with extra factor, with such a definition the corresponding inequalities are and ). When placed in magnetic field, type-1.5 superconductors should form quantum vortices: magnetic-flux-carrying excitations. They allow magnetic field to pass through superconductors due to a vortex-like circulation of superconducting particles (electronic pairs). In type-1.5 superconductors these vortices have long-range attractive, short-range repulsive interaction. As a consequence a type-1.5 superconductor in a magnetic field can form a phase separation into domains with expelled magnetic field and clusters of quantum vortices which are bound together by attractive intervortex forces. The domains of the Meissner state retain the two-component superconductivity, while in the vortex clusters one of the superconducting components is suppressed. Thus such materials should allow coexistence of various properties of type-I and type-II superconductors. Description Type-I superconductors completely expel external magnetic fields if the strength of the applied field is sufficiently low. Also the supercurrent can flow only on the surface of such a superconductor but not in its interior. This state is called the Meissner state. However at elevated magnetic field, when the magnetic field energy becomes comparable with the superconducting condensation energy, the superconductivity is destroyed by the formation of macroscopically large inclusions of non-superconducting phase. Type-II superconductors, besides the Meissner state, possess another state: a sufficiently strong
https://en.wikipedia.org/wiki/Flap%20endonuclease
Flap endonucleases (FENs, also known as 5' durgs in older references) are a class of nucleolytic enzymes that act as both 5'-3' exonucleases and structure-specific endonucleases on specialised DNA structures that occur during the biological processes of DNA replication, DNA repair, and DNA recombination. Flap endonucleases have been identified in eukaryotes, prokaryotes, archaea, and some viruses. Organisms can have more than one FEN homologue; this redundancy may give an indication of the importance of these enzymes. In prokaryotes, the FEN enzyme is found as an N-terminal domain of DNA polymerase I, but some prokaryotes appear to encode a second homologue. The endonuclease activity of FENs was initially identified as acting on a DNA duplex which has a single-stranded 5' overhang on one of the strands (termed a "5' flap", hence the name flap endonuclease). FENs catalyse hydrolytic cleavage of the phosphodiester bond at the junction of single- and double-stranded DNA. Some FENs can also act as 5'-3' exonucleases on the 5' terminus of the flap strand and on 'nicked' DNA substrates. Protein structure models based on X-ray crystallography data suggest that FENs have a flexible arch created by two α-helices through which the single 5' strand of the 5' flap structure can thread. Flap endonucleases have been used in biotechnology, for example the Taqman PCR assay and the Invader Assay for mutation and single nucleotide polymorphism (SNP) detection. See also Endonucleases
https://en.wikipedia.org/wiki/Black%20cross
Black cross or Black Cross may refer to: Black Cross (Teutonic Order), heraldic insignia of the Teutonic order (since 1205) Black Cross (Germany), military emblem of Prussia and Germany, derived from the cross used by the Teutonic order Anarchist Black Cross, an anarchist support organization Black Cross Navigation and Trading Company, the successor of the Black Star Line "Black Cross (Hezekiah Jones)", a 1948 poem by Joseph Simon Newman, recorded by Lord Buckley and by Bob Dylan Black Cross, also known as Knights of the Teutonic Order, a 1960 film from Poland "Black Cross" (song), debut single of the band 45 Grave Kroaz Du, a Breton flag See also Blue Cross (disambiguation) Bronze Cross Green Cross Red Cross (disambiguation) Silver Cross White Cross (disambiguation) Yellow cross Cross, black Cross symbols
https://en.wikipedia.org/wiki/Generic%20Eclipse%20Modeling%20System
Generic Eclipse Modeling System (GEMS) is a configurable toolkit for creating domain-specific modeling and program synthesis environments for Eclipse. The project aims to bridge the gap between the communities experienced with visual metamodeling tools like those built around the Eclipse modeling technologies, such as the Eclipse Modeling Framework (EMF) and Graphical Modeling Framework (GMF). GEMS helps developers rapidly create a graphical modeling tool from a visual language description or metamodel without any coding in third-generation languages. Graphical modeling tools created with GEMS automatically support complex capabilities, such as remote updating and querying, template creation, styling with Cascading Style Sheets (CSS), and model linking. The configuration is accomplished through metamodels specifying the modeling paradigm of the application domain, i.e. a domain-specific modeling language (DSML). The modeling paradigm contains all the syntactic, semantic, and presentation information regarding the domain; which concepts will be used to construct models, what relationships may exist among those concepts, how the concepts may be organized and viewed by the modeler, and rules governing the construction of models. The modeling paradigm defines the family of models that can be created using the resultant modeling environment. The built-in metamodeling language is based on the UML class diagram notation. Metamodels in other eCore readable formats can be used as well. Metamodel constraints can be specified in declarative languages (e.g. OCL, Prolog) or, alternatively, in Java. Once a metamodel has been created, GEMS plug-in generator can be invoked to create the modeling tool. The generated plug-in uses Eclipse's Graphical Editing Framework (GEF) and Draw2D plug-in to visualize the DSML as a diagram. GEMS extension points can be used to create an interpreter which traverses the domain-specific model and generates code. Interpreters can also interpret the
https://en.wikipedia.org/wiki/Amplitude%20gate
An amplitude gate (also, slicer or slice amplifier) is a circuit whose output is only the part of the input signal that lies between two amplitude boundary level values.
https://en.wikipedia.org/wiki/Ferdinand%20Georg%20Frobenius
Ferdinand Georg Frobenius (26 October 1849 – 3 August 1917) was a German mathematician, best known for his contributions to the theory of elliptic functions, differential equations, number theory, and to group theory. He is known for the famous determinantal identities, known as Frobenius–Stickelberger formulae, governing elliptic functions, and for developing the theory of biquadratic forms. He was also the first to introduce the notion of rational approximations of functions (nowadays known as Padé approximants), and gave the first full proof for the Cayley–Hamilton theorem. He also lent his name to certain differential-geometric objects in modern mathematical physics, known as Frobenius manifolds. Biography Ferdinand Georg Frobenius was born on 26 October 1849 in Charlottenburg, a suburb of Berlin, from parents Christian Ferdinand Frobenius, a Protestant parson, and Christine Elizabeth Friedrich. He entered the Joachimsthal Gymnasium in 1860 when he was nearly eleven. In 1867, after graduating, he went to the University of Göttingen where he began his university studies but he only studied there for one semester before returning to Berlin, where he attended lectures by Kronecker, Kummer and Karl Weierstrass. He received his doctorate (awarded with distinction) in 1870 supervised by Weierstrass. His thesis was on the solution of differential equations. In 1874, after having taught at secondary school level first at the Joachimsthal Gymnasium then at the Sophienrealschule, he was appointed to the University of Berlin as an extraordinary professor of mathematics. Frobenius was only in Berlin a year before he went to Zürich to take up an appointment as an ordinary professor at the Eidgenössische Polytechnikum. For seventeen years, between 1875 and 1892, Frobenius worked in Zürich. It was there that he married, brought up his family, and did much important work in widely differing areas of mathematics. In the last days of December 1891 Kronecker died and, therefore,
https://en.wikipedia.org/wiki/Alex%20Wilkie
Alex James Wilkie FRS (born 1948 in Northampton) is a British mathematician known for his contributions to model theory and logic. Previously Reader in Mathematical Logic at the University of Oxford, he was appointed to the Fielden Chair of Pure Mathematics at the University of Manchester in 2007. Education Alex Wilkie attended Aylesbury Grammar School and went on to gain his BSc in mathematics with first class honours from University College London in 1969, his MSc (in mathematical logic) from the University of London in 1970, and his PhD from the Bedford College, University of London in 1973 under the supervision of Wilfrid Hodges with a dissertation titled Models of Number Theory. Career and research After his PhD he went on to an appointment as a lecturer in mathematics at Leicester University from 1972 to 1973, then a research fellow at the Open University from 1973 until 1978. He spent two periods as a junior lecturer in mathematics at Oxford University (1978–80 and 1981-2) with (1980–1) as a visiting assistant professor at Yale University. In 1980 Wilkie solved Tarski's high school algebra problem. In October 1982 Wilkie was appointed as a research fellow in the department of mathematics at the University of Paris VII, then returned to England the following year to take up a three-year SERC (now EPSRC) advanced research fellowship at the University of Manchester. After two years he was appointed lecturer in the Department of Mathematics. In 1986 he went on to Oxford where he was appointed to the readership in mathematical logic there which had become vacant upon the retirement of Robin Gandy. He remained in this post until appointment to the Fielden Chair at Manchester. Awards and honours Wilkie was elected a Fellow of the Royal Society in 2001. To quote the citation Wilkie has combined logical techniques and differential-geometric techniques to establish fundamental Finiteness Theorems for sets definable using the exponential function, and more general
https://en.wikipedia.org/wiki/Molecular%20spring
A Molecular spring is a device or part of a biological system based on molecular mechanics and is associated with molecular vibration. Any molecule can be deformed in several ways - A-A bond length, A-A-A angle, A-A-A-A torsion angle. Deformed molecules store energy, which can be released and cause mechanical work as the molecules return into their optimal geometrical conformation. The term molecular string is usually used in nano-science and molecular biology, however theoretically also macroscopic molecular springs can be considered, if it is manufactured. Such a device composed for example of arranged ultra-high molecular mass polymer fibres (Helicene, Polyacetylene) could store extraordinary (0.1-10MJ/kg in comparison to 0.0003MJ/kg of clockwork spring) amount of energy which can be stored and released almost instantly, with high energy conversion efficiency. The amount of energy storable in molecular spring is limited by the value of deformation the molecule can withstand until it undergoes chemical change. Manufacturing of such macroscopic device is however out of reach of contemporary technology, because of difficulties of synthesis and molecular arrangement of such long polymer molecules. In addition, the force needed to draw molecular string to its maximum length could be impractically high - comparable to the tensile strength of particular polymer molecule (~100GPa for some carbon compounds) See also Ultra high molecular weight polyethylene Carbon nanotube Carbon nanotube springs
https://en.wikipedia.org/wiki/Tessellation%20%28computer%20graphics%29
In computer graphics, tessellation is the dividing of datasets of polygons (sometimes called vertex sets) presenting objects in a scene into suitable structures for rendering. Especially for real-time rendering, data is tessellated into triangles, for example in OpenGL 4.0 and Direct3D 11. In graphics rendering A key advantage of tessellation for realtime graphics is that it allows detail to be dynamically added and subtracted from a 3D polygon mesh and its silhouette edges based on control parameters (often camera distance). In previously leading realtime techniques such as parallax mapping and bump mapping, surface details could be simulated at the pixel level, but silhouette edge detail was fundamentally limited by the quality of the original dataset. In Direct3D 11 pipeline (a part of DirectX 11), the graphics primitive is the patch. The tessellator generates a triangle-based tessellation of the patch according to tessellation parameters such as the TessFactor, which controls the degree of fineness of the mesh. The tessellation, along with shaders such as a Phong shader, allows for producing smoother surfaces than would be generated by the original mesh. By offloading the tessellation process onto the GPU hardware, smoothing can be performed in real time. Tessellation can also be used for implementing subdivision surfaces, level of detail scaling and fine displacement mapping. OpenGL 4.0 uses a similar pipeline, where tessellation into triangles is controlled by the Tessellation Control Shader and a set of four tessellation parameters. In computer-aided design In computer-aided design the constructed design is represented by a boundary representation topological model, where analytical 3D surfaces and curves, limited to faces, edges, and vertices, constitute a continuous boundary of a 3D body. Arbitrary 3D bodies are often too complicated to analyze directly. So they are approximated (tessellated) with a mesh of small, easy-to-analyze pieces of 3D volume—us
https://en.wikipedia.org/wiki/Astacin
Astacins are a family of multidomain metalloendopeptidases which are either secreted or membrane-anchored. These metallopeptidases belong to the MEROPS peptidase family M12, subfamily M12A (astacin family, clan MA(M)). The protein fold of the peptidase domain for members of this family resembles that of thermolysin, the type example for clan MA and the predicted active site residues for members of this family and thermolysin occur in the motif HEXXH. The astacin family of metalloendopeptidases (EC 3.4.24.21) encompasses a range of proteins found in hydra to humans, in mature and developmental systems. Their functions include activation of growth factors, degradation of polypeptides, and processing of extracellular proteins. The proteins are synthesised with N-terminal signal and pro-enzyme sequences, and many contain multiple domains C-terminal to the protease domain. They are either secreted from cells, or are associated with the plasma membrane. The astacin molecule adopts a kidney shape, with a deep active-site cleft between its N- and C-terminal domains. The zinc ion, which lies at the bottom of the cleft, exhibits a unique penta-coordinated mode of binding, involving 3 histidine residues, a tyrosine and a water molecule (which is also bound to the carboxylate side chain of Glu93). The N-terminal domain comprises 2 alpha-helices and a 5-stranded beta-sheet. The overall topology of this domain is shared by the archetypal zinc-endopeptidase thermolysin. Astacin protease domains also share common features with serralysins, matrix metalloendopeptidases, and snake venom proteases; they cleave peptide bonds in polypeptides such as insulin B chain and bradykinin, and in proteins such as casein and gelatin; and they have arylamidase activity. History In 1965 R. Zwilling observed during his doctorate work that the oncosphaere of the mouse tapeworm Hymenolepis diminuta easily hatched in vitro in the presence of the digestive fluid of its intermediate host Tenebrio moli
https://en.wikipedia.org/wiki/Lifting%20scheme
The lifting scheme is a technique for both designing wavelets and performing the discrete wavelet transform (DWT). In an implementation, it is often worthwhile to merge these steps and design the wavelet filters while performing the wavelet transform. This is then called the second-generation wavelet transform. The technique was introduced by Wim Sweldens. The lifting scheme factorizes any discrete wavelet transform with finite filters into a series of elementary convolution operators, so-called lifting steps, which reduces the number of arithmetic operations by nearly a factor two. Treatment of signal boundaries is also simplified. The discrete wavelet transform applies several filters separately to the same signal. In contrast to that, for the lifting scheme, the signal is divided like a zipper. Then a series of convolution–accumulate operations across the divided signals is applied. Basics The simplest version of a forward wavelet transform expressed in the lifting scheme is shown in the figure above. means predict step, which will be considered in isolation. The predict step calculates the wavelet function in the wavelet transform. This is a high-pass filter. The update step calculates the scaling function, which results in a smoother version of the data. As mentioned above, the lifting scheme is an alternative technique for performing the DWT using biorthogonal wavelets. In order to perform the DWT using the lifting scheme, the corresponding lifting and scaling steps must be derived from the biorthogonal wavelets. The analysis filters () of the particular wavelet are first written in polyphase matrix where . The polyphase matrix is a 2 × 2 matrix containing the analysis low-pass and high-pass filters, each split up into their even and odd polynomial coefficients and normalized. From here the matrix is factored into a series of 2 × 2 upper- and lower-triangular matrices, each with diagonal entries equal to 1. The upper-triangular matrices contain the co
https://en.wikipedia.org/wiki/Project%20X%20%28accelerator%29
Project-X is a proposed high-intensity proton accelerator complex which is to be built at the Fermi National Accelerator Laboratory. It is planned to produce beams of different energies up to 8 GeV for precision experiments involving kaons and muons. The complex can also be used to create a high-intensity neutrino beam for neutrino oscillation experiments such as NOνA and the Long Baseline Neutrino Experiment. Project-X will be based on superconducting RF components such as those developed for the International Linear Collider. Immediate plans are for cost-effective upgrades in proton luminosity referred to as the Proton Improvement Plan-II. More future-thinking proposals see Project X as laying the groundwork for a possible Muon collider at the Fermilab site. Background Project X is a long range plan to bring accelerators at Fermilab campus to new frontiers. The plan for accelerators focuses on two of the three frontiers that are long-term plans of Fermilab. In the intensity frontier, the new high-intensity accelerators will support experiments that require intense particle beams to understand particles such as neutrinos, muons, kaons, and nuclei. In the energy frontier, the accelerators will support the detection of new particles and forces with potential future projects such as a multi-TeV Muon Collider. Stages The immediate plan of Project X is to focus on the intensity frontier. The project is broken down into 3 stages. Stage one includes upgrades to existing facilities to support immediate experiments. This stage has translated into work done in the Proton Improvement Plan. Stage two includes delivery of three concurrent beam levels: 2.9 MW at 3 GeV; 50–200 kW at 8 GeV and 2.3 MW at 60–120 GeV. Stage three is to build next generation accelerators as the front end to the energy frontier based on international collaboration in projects such as the Neutrino Factory and Muon Collider.
https://en.wikipedia.org/wiki/Null%20allele
A null allele is a nonfunctional allele (a variant of a gene) caused by a genetic mutation. Such mutations can cause a complete lack of production of the associated gene product or a product that does not function properly; in either case, the allele may be considered nonfunctional. A null allele cannot be distinguished from deletion of the entire locus solely from phenotypic observation. A mutant allele that produces no RNA transcript is called an RNA null (shown by Northern blotting or by DNA sequencing of a deletion allele), and one that produces no protein is called a protein null (shown by Western blotting). A genetic null or amorphic allele has the same phenotype when homozygous as when heterozygous with a deficiency that disrupts the locus in question. A genetic null allele may be both a protein null and an RNA null, but may also express normal levels of a gene product that is nonfunctional due to mutation. Null alleles can have lethal effects depending on the importance of the mutated gene. For example, mice homozygous for a null allele for insulin die 48 to 72 hours after birth. Null alleles can also have beneficial effects, such as the elevated harvest index of semi-dwarf rice of the green revolution caused by null alleles in GA20ox-2. Evidence Polymerase chain reaction (PCR) A microsatellite null allele is an allele at a microsatellite locus that does not amplify to detectable levels in a polymerase chain reaction test. Microsatellite regions are usually characterized by short, repeated sequences of nucleotides. Primers that are specific to a particular locus are used in PCR amplification to bind to these nucleotide sequence repeats and are used as genetic markers. The primers anneal to either end of the locus and are derived from source organisms in a genomic library. Divergence from the reference sequences (from genetic mutations) results in poor annealing of the primers so that the marker cannot be used, representative of a null allele. Parentage
https://en.wikipedia.org/wiki/Martin%20Bower
Martin Bower (born 1952) is a model maker and designer of special effects miniatures for both film and television. His credits include the television series Space: 1999 (1975-8), Blake's 7 (1978–81) and The Tripods (1984–85), and the films Alien (1979), Flash Gordon (1980) and Outland (1981). He has a long-standing professional relationship with special effects director Brian Johnson. Notable designs Blake's 7 "The Liberator" models, and the teleport bracelets, as seen in multiple episodes of the series Space:1999 Mark IX Hawk model as seen in the episode "War Games" Ultra Probe model as seen in the episode "Dragon's Domain" Laser Tank model as seen in the episode "The Infernal Machine" Super Swift model as seen in the episodes "The Bringers of Wonder, Part One" and "The Bringers of Wonder, Part Two" Voyager One model as seen in the episode "Voyager's Return" The Tripods Models of the eponymous tripods used throughout the two series
https://en.wikipedia.org/wiki/Staphylococcus%20virus%20G1
Staphylococcus virus G1 is a virus of the family Herelleviridae, genus Kayvirus. As a member of the group I of the Baltimore classification, Staphylococcus virus G1 is a dsDNA virus. All the family Herelleviridae members share a nonenveloped morphology consisting of a head and a tail separated by a neck. Its genome is linear. The propagation of the virions includes the attaching to a host cell (a bacterium, as Staphylococcus virus G1 is a bacteriophage) and the injection of the double stranded DNA; the host transcribes and translates it to manufacture new particles. To replicate its genetic content requires host cell DNA polymerases and, hence, the process is highly dependent on the cell cycle. The Gp67 protein of G1 has been found to interact with its host's RNA polymerase though an interaction with a sigma factor. The phage contains a genome of 138,715 base pairs with a 30.4% of GC content and 214 predicted genes; this means that the 88.5% of the DNA is coding open reading frames, and therefore the gene density (the number of genes per kilobase) is 1.54.
https://en.wikipedia.org/wiki/Penetration%20depth
Penetration depth is a measure of how deep light or any electromagnetic radiation can penetrate into a material. It is defined as the depth at which the intensity of the radiation inside the material falls to 1/e (about 37%) of its original value at (or more properly, just beneath) the surface. When electromagnetic radiation is incident on the surface of a material, it may be (partly) reflected from that surface and there will be a field containing energy transmitted into the material. This electromagnetic field interacts with the atoms and electrons inside the material. Depending on the nature of the material, the electromagnetic field might travel very far into the material, or may die out very quickly. For a given material, penetration depth will generally be a function of wavelength. Beer–Lambert law According to Beer–Lambert law, the intensity of an electromagnetic wave inside a material falls off exponentially from the surface as If denotes the penetration depth, we have Penetration depth is one term that describes the decay of electromagnetic waves inside of a material. The above definition refers to the depth at which the intensity or power of the field decays to 1/e of its surface value. In many contexts one is concentrating on the field quantities themselves: the electric and magnetic fields in the case of electromagnetic waves. Since the power of a wave in a particular medium is proportional to the square of a field quantity, one may speak of a penetration depth at which the magnitude of the electric (or magnetic) field has decayed to 1/e of its surface value, and at which point the power of the wave has thereby decreased to or about 13% of its surface value: Note that is identical to the skin depth, the latter term usually applying to metals in reference to the decay of electrical currents (which follow the decay in the electric or magnetic field due to a plane wave incident on a bulk conductor). The attenuation constant is also identical
https://en.wikipedia.org/wiki/Roseburia
{{Automatic taxobox | image = | image_caption = | taxon = Roseburia | authority = Duncan et al. 2006 | type_species = Roseburia cecicola | subdivision_ranks = Species | subdivision = Roseburia cecicolaRoseburia faecis Roseburia hominisRoseburia intestinalis Roseburia inulinivorans }}Roseburia'' is a genus of butyrate-producing, Gram-positive anaerobic bacteria that inhabit the human colon. Named in honor of Theodor Rosebury, they are members of the phylum Bacillota (formerly known as Firmicutes). Increased abundance of Roseburia is associated with weight loss and reduced glucose intolerance in mice.
https://en.wikipedia.org/wiki/Helper%20virus
A helper virus is a virus that allows an otherwise-deficient coinfecting virus to replicate. These can be naturally occurring as with Hepatitis D virus, which requires Hepatitis B virus to coinfect cells in order to replicate. Helper viruses are also commonly used to replicate and spread viral vectors for gene expression and gene therapy. See also Helper dependent virus Virophage
https://en.wikipedia.org/wiki/List%20of%20university%20statistical%20consulting%20centers
This list of university statistical consulting centers (or centres) is a simple list of universities in which there is a specifically designated team providing statistical consultancy services. Often this service will be available only to enquirers from within the same university. Australia Belgium Canada Germany Hungary Indonesia Spain United States The following is a list of university statistical consulting centers in the United States of America. Vietnam See also Statistical consultant Statistician
https://en.wikipedia.org/wiki/Fluxus%20%28programming%20environment%29
Fluxus is a live coding environment for 3D graphics, music and games. It uses the programming language Racket (a dialect of Scheme/Lisp) to work with a games engine with built-in 3D graphics, physics simulation and sound synthesis. All programming is done on-the-fly, where the code editor appears on top of the graphics that the code is generating. It is an important reference for research and practice in exploratory programming, pedagogy, live performance and games programming.
https://en.wikipedia.org/wiki/Ristra
A ristra (), also known as a sarta, is an arrangement of drying chile pepper pods, garlic bulbs, or other vegetables for later consumption. In addition to its practical use, the ristra has come to be a trademark of decorative design in the state of New Mexico as well as southern Arizona. Typically, large chiles such as New Mexico chiles and Anaheim peppers are used, although any kind of chile may be used. Garlic can also be arranged into a ristra for drying and curing after the bulbs have matured and the leaves have died away. Ristras are commonly used for decoration and "are said to bring health and good luck." See also List of dried foods
https://en.wikipedia.org/wiki/Standard%20normal%20table
In statistics, a standard normal table, also called the unit normal table or Z table, is a mathematical table for the values of , the cumulative distribution function of the normal distribution. It is used to find the probability that a statistic is observed below, above, or between values on the standard normal distribution, and by extension, any normal distribution. Since probability tables cannot be printed for every normal distribution, as there are an infinite variety of normal distributions, it is common practice to convert a normal to a standard normal (known as a z-score) and then use the standard normal table to find probabilities. Normal and standard normal distribution Normal distributions are symmetrical, bell-shaped distributions that are useful in describing real-world data. The standard normal distribution, represented by , is the normal distribution having a mean of 0 and a standard deviation of 1. Conversion If is a random variable from a normal distribution with mean and standard deviation , its Z-score may be calculated from by subtracting and dividing by the standard deviation: If is the mean of a sample of size from some population in which the mean is and the standard deviation is , the standard error is If is the total of a sample of size from some population in which the mean is and the standard deviation is , the expected total is and the standard error is Reading a Z table Formatting / layout tables are typically composed as follows: The label for rows contains the integer part and the first decimal place of . The label for columns contains the second decimal place of . The values within the table are the probabilities corresponding to the table type. These probabilities are calculations of the area under the normal curve from the starting point (0 for cumulative from mean, negative infinity for cumulative and positive infinity for complementary cumulative) to . Example: To find 0.69, one would look down the r
https://en.wikipedia.org/wiki/Richard%20Lary
Richard F. "Richie" Lary (born 1948, Brooklyn) is the RL of the PDP-8 RL Monitor System, which subsequently became MS/8. Years later, while working for Digital Equipment Corporation, he was also involved with other DEC hardware and software, including "principal architect for OS/8" and "working on the VAX architecture." Biography He graduated from Stuyvesant High School in 1965, along with ; they both were on the school's Math Team and "later wound up working on the VAX architecture." They were $2/hour summertime Fortran programmers in 1965, using an IBM 1130. Lary left DEC in 2000, forming a company he and his wife Ellen Lary, also a former DEC employee, named TuteLary.
https://en.wikipedia.org/wiki/When%20Einstein%20Walked%20with%20G%C3%B6del
When Einstein Walked With Gödel: Excursions to the Edge of Thought is the third nonfiction book authored by American philosopher and essayist Jim Holt. The book was initially released by Farrar, Straus and Giroux on 15 May 2018. Background The book is a collection of Holt's previous 25 essays on such diverse themes as science, philosophy, the nature of time, eugenics, quantum physics, string theory, relativity, the future of the universe, and the foundation of mathematics; Holt also explores the works of such scientists as Alan Turing, Benoit Mandelbrot, Emmy Noether, and others. Reception Parul Sehgal of The New York Times stated "In these pieces, plucked from the last 20 years, Holt takes on infinity and the infinitesimal, the illusion of time, the birth of eugenics, the so-called new atheism, smartphones and distraction. It is an elegant history of recent ideas. There are a few historical correctives — he dismantles the notion that Ada Lovelace, the daughter of Lord Byron, was the first computer programmer. But he generally prefers to perch in the middle of a muddle — say, the string theory wars — and hear evidence from both sides without rushing to adjudication. The essays orbit around three chief concerns: How do we conceive of the world (metaphysics), how do we know what we know (epistemology) and how do we conduct ourselves (ethics)". Steven Poole of The Wall Street Journal commented "...this collection of previously published essays by Jim Holt, who is one of the very best modern science writers".
https://en.wikipedia.org/wiki/Mir-52%20microRNA%20precursor%20family
In molecular biology mir-52 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms. See also MicroRNA
https://en.wikipedia.org/wiki/Comparison%20of%20MQTT%20implementations
MQTT is an ISO standard (ISO/IEC PRF 20922) publish–subscribe-based messaging protocol. It works on top of the Internet protocol suite TCP/IP. It is designed for connections with remote locations where a "small code footprint" is required or the network bandwidth is limited. The publish-subscribe messaging pattern requires a message broker. All comparison categories use the stable version of each implementation listed in the overview section. The comparison is limited to features that relate to the MQTT protocol. Overview The following table lists MQTT both libraries and implementations, along with general information about each. A more complete list of MQTT implementations can be found on GitHub. Protocol support There are several versions of the MQTT protocol currently standardized. Below is a list containing the more recent versions of the MQTT protocol, with the organization that standardized them. MQTT-SN v1.2, standardized by IBM. MQTT v3.1, standardized by Eurotech and IBM. MQTT v3.1.1, standardized by OASIS. MQTT v5.0, standardized by OASIS. The following table lists the versions of MQTT that each implementation supports, and also lists their support for SSL/TLS and TCP. The security provided by SSL/TLS may be desirable depending on the type traffic being sent between devices, as MQTT transmits messages in the clear. Quality of service levels offered From the MQTT page, quality of service (QoS) is described as,Quality of service refers to traffic prioritization and resource reservation control mechanisms rather than the achieved service quality. Quality of service is the ability to provide different priority to different applications, users, or data flows, or to guarantee a certain level of performance to a data flow.A description of each QoS level is found below. At most once delivery (fire and forget) At least once delivery (acknowledged delivery) Exactly once delivery (assured delivery) The following table lists each implementation's s
https://en.wikipedia.org/wiki/Passive%20acoustics
Passive acoustics is the action of listening for sounds, often at specific frequencies or for purposes of specific analyses. It is often used for passive acoustic monitoring (PAM), the act of recording animal and environmental sounds through the use of acoustic sensors for the purpose of tracking animals and answering other ecological questions. PAM has risen as a valuable method for a variety of queries relating to conservation, applied and behavioral ecology, and biodiversity. There are a variety of tools and machinery available for monitoring in both terrestrial and aquatic environments. This method of ecological monitoring is especially valuable for species that vocalize and are hard to observe visually. Passive acoustic monitoring also allows for observation across spatial and temporal scales previously unseen due to the limitations of data collection. As applied to underwater acoustics, also termed hydroacoustics, passive acoustics can be used to listen for underwater explosions, earthquakes, volcanic eruptions, sounds produced by fish and other marine animals, vessel activity or aquatic detecting equipment (as in hydroacoustics to track fish).
https://en.wikipedia.org/wiki/Free-orbit%20experiment%20with%20laser%20interferometry%20X-rays
The Free-orbit Experiment with Laser Interferometry X-Rays (FELIX) belongs to a category of experiments exploring whether macroscopic systems can be in superposition states. It was originally proposed by the physicist Roger Penrose in his 2004 book The Road to Reality specifically to prove whether unconventional decoherence processes such as gravitationally induced decoherence or spontaneous wave-function collapse of a quantum system occur. Later revised to take place as a tabletop experiment, if successful, it is estimated that a mass of roughly 1014 atoms would have been superposed, approximately nine orders of magnitude more massive than any superposition observed to that date (2003). Configuration The proposed experimental setup is basically a variation of the Michelson interferometer but for a single photon. Additionally, one of the mirrors has to be very tiny and fixed on an isolated micromechanical-oscillator. This allows it to move when the photon is reflected on it, so that it may become superposed with the photon. The purpose is to vary the size of the mirror to investigate the effect of the mass on the time it takes for the quantum system to collapse. Originally the arms of the interferometer had to stretch into the hundreds of thousands of kilometers to achieve a photon roundtrip-time comparable to the oscillator's period, but that meant that the experiment had to take place in-orbit, reducing its viability. The revised proposal requires that the mirrors be placed into high-finesse optical cavities that will trap the photons long enough to achieve the desired delay. There are various technological challenges, but all are within high-end laboratory capabilities. The primary requirement is that the mass of the cavity remains as small as possible. To avoid noise on the interferometer and have a low probability of emitting more than one photon each time, a very low absolute temperature for the experiment is needed, on the order of 60 μK. For similar re
https://en.wikipedia.org/wiki/Kernel%20principal%20component%20analysis
In the field of multivariate statistics, kernel principal component analysis (kernel PCA) is an extension of principal component analysis (PCA) using techniques of kernel methods. Using a kernel, the originally linear operations of PCA are performed in a reproducing kernel Hilbert space. Background: Linear PCA Recall that conventional PCA operates on zero-centered data; that is, , where is one of the multivariate observations. It operates by diagonalizing the covariance matrix, in other words, it gives an eigendecomposition of the covariance matrix: which can be rewritten as . (See also: Covariance matrix as a linear operator) Introduction of the Kernel to PCA To understand the utility of kernel PCA, particularly for clustering, observe that, while N points cannot, in general, be linearly separated in dimensions, they can almost always be linearly separated in dimensions. That is, given N points, , if we map them to an N-dimensional space with where , it is easy to construct a hyperplane that divides the points into arbitrary clusters. Of course, this creates linearly independent vectors, so there is no covariance on which to perform eigendecomposition explicitly as we would in linear PCA. Instead, in kernel PCA, a non-trivial, arbitrary function is 'chosen' that is never calculated explicitly, allowing the possibility to use very-high-dimensional 's if we never have to actually evaluate the data in that space. Since we generally try to avoid working in the -space, which we will call the 'feature space', we can create the N-by-N kernel which represents the inner product space (see Gramian matrix) of the otherwise intractable feature space. The dual form that arises in the creation of a kernel allows us to mathematically formulate a version of PCA in which we never actually solve the eigenvectors and eigenvalues of the covariance matrix in the -space (see Kernel trick). The N-elements in each column of K represent the dot product of one point of the tr
https://en.wikipedia.org/wiki/Kathryn%20Peddrew
Kathryn Peddrew (June 14, 1922 - March 4, 2012) was an African-American mathematician, engineer, and scientist who played a crucial role in the National Advisory Committee for Aeronautics (NACA) and the National Aeronautics and Space Administration (NASA). She was one of the African-American women who worked as a "human computer" at NACA's Langley Research Center in the 1940s and 1950s. Early life and education Peddrew was born on June 14, 1922 in Martinsburg, Virginia. She attended Storer College in her home state of West Virginia. She focused her studies on chemistry and graduated with a chemistry degree in 1943. After college, she began looking for research opportunities. Her first choice was to travel with one of her former professors to New Guinea to study quinine deafness. Unfortunately, these plans fell through as the research program had made no plans for female housing. Career at NACA and NASA Peddrew saw an advertisement from NACA (eventually known as NASA) saying that they were hiring chemists. At the time, there was a large increase in women being hired by NACA due to men going overseas to fight in WWII. She decided to apply for this position and was hired. However, when she arrived at the job, she was relocated to the West Area Computing Unit after it was discovered that she was African American. There she worked in the all-black West Building at NACA. She and her colleagues were referred to as the “West Computers”, a group that consisted of Dorothy Vaughn, Mary Jackson, Miriam Daniel Mann, and Peddrew herself. Here she conducted aeronautical and aerospace research, doing the majority of her work in the Instrument Research Division. The unit was responsible for performing complex calculations that were critical to various aeronautical research projects. Despite facing racial segregation and discrimination, Peddrew and her colleagues persisted in their work and contributed to the development of supersonic flight, as well as the early stages of the
https://en.wikipedia.org/wiki/Seshadri%20constant
In algebraic geometry, a Seshadri constant is an invariant of an ample line bundle L at a point P on an algebraic variety. It was introduced by Demailly to measure a certain rate of growth, of the tensor powers of L, in terms of the jets of the sections of the Lk. The object was the study of the Fujita conjecture. The name is in honour of the Indian mathematician C. S. Seshadri. It is known that Nagata's conjecture on algebraic curves is equivalent to the assertion that for more than nine general points, the Seshadri constants of the projective plane are maximal. There is a general conjecture for algebraic surfaces, the Nagata–Biran conjecture. Definition Let be a smooth projective variety, an ample line bundle on it, a point of , = { all irreducible curves passing through }. . Here, denotes the intersection number of and , measures how many times passing through . Definition: One says that is the Seshadri constant of at the point , a real number. When is an abelian variety, it can be shown that is independent of the point chosen, and it is written simply .
https://en.wikipedia.org/wiki/Manasollasa
The also known as Abhilashitartha Chintamani, is an early 12th-century Sanskrit text composed by the Kalyani Chalukya king Someshvara III, who ruled in present-day Karnataka . It is an encyclopedic work covering topics such as polity, governance, ethics, economics, astronomy, astrology, rhetoric, veterinary medicine, horticulture, perfumes, food, architecture, games, painting, poetry, dance and music. The text is a valuable source of socio-cultural information on 11th- and 12th-century India. The encyclopedic treatise is structured as five sub-books with a cumulative total of 100 chapters. It is notable for its extensive discussion of arts, particularly music and dance. It is also notable for including chapters on food recipes and festivals, many of which are a part of modern Indian culture. Another medieval era Sanskrit text with the title Mānasollāsa also exists, consisting of devotional praise hymns (stotra), and it is different from the encyclopedic treatise. Nomenclature The title Manasollasa (मानसोल्लास) is a compound Sanskrit word, consisting of manas ( मनस्) or "mind" and ullasa (उल्लास) or "rejoicing, delighting". It means "the joy, delighter or entertainer of the mind". Alternatively, the compound word can be broken as manasa and ullasa, which mean "happiness of mind". The work is divided into five sub-books suffixed as Viṁśati (विंशति) which means 20 and refers to the 20 chapters in the sub-book. In modern scholarship, it is referred by IAST spellings "Manasollāsa", and "Mānasollāsa". It is also known as the Abhilashitartha Chintamani () (literally, the precious gem that fulfills wish). This text title is spelled Manasollāsa, and there is another medieval era Indian text with the title, Mānasollāsa (), written in Stotra (poetry of praise) style related to Dakshinamurti, very different in scope and attributed to Adi Shankara or Suresvara. Date and author The text was completed in 1129 CE by Someshvara III. He became the king in 1127, was part of
https://en.wikipedia.org/wiki/The%20Conservation%20Foundation%20%28Illinois%29
The Conservation Foundation is a non-governmental, not-for-profit 501(c)(3) organization with a mission to preserve open space and natural lands, protect rivers and watersheds, and promote stewardship of the environment. It serves northern Illinois counties of DuPage, Kane, Kendall and Will. McDonald Farm In 1992, Lenore Clow McDonald donated her 60-acre farm to The Conservation Foundation, subject to conservation easement. The farm now serves as headquarters for the organization. Located on the farm is a partner organization, The Resiliency Institute, which uses permaculture education and design to foster resilient lifestyles, landscapes and communities. Green Earth Harvest A major program at the Foundation is Green Earth Harvest. This is a certified organic farm that provides local, organic vegetables through environmentally sustainable methods.
https://en.wikipedia.org/wiki/Lucien%20LaCoste
Lucien LaCoste (1908 – 1995) was a physicist and metrologist. He coinvented the modern gravimeter and invented the zero-length spring and vehicle-mounted gravimeters. He was also co-founder of LaCoste Romberg, a prominent company selling gravimetric instruments. Life LaCoste discovered the zero-length spring in 1932 while performing an assignment in Arnold Romberg's undergraduate physics course. A zero-length spring is a spring supported in such a way that its exerted force is proportional to its length, rather than the distance it is compressed. That is, over at least part of its travel, it does not conform to Hooke's Law of spring compression. The zero-length spring is extremely important to seismometers and gravimeters because it permits the design of vertical pendulums with (theoretically) infinite periods. In practice, periods of a thousand seconds are possible, a hundredfold increase from other forms of pendulum. Over a short period starting in 1932, the design of these instruments was revolutionized, obsoleting all previous designs. During this period, LaCoste and his physics teacher Arnold Romberg invented the first modern seismographs and gravimeters, using steel and quartz (respectively) zero-length springs. While a graduate student, LaCoste decided to go into business together with Romberg, selling advanced gravimeters to oil-exploration companies. LaCoste's most famous invention is the ship, and aircraft-mounted gravimeter. These revolutionized exploration for minerals by allowing wide-ranging geological surveys. The chief problem that Lacoste defeated was to distinguish the accelerations of the vehicles from the accelerations due to gravity, and measure the minute changes in gravity. Since the accelerations from the vehicle typically are hundreds to thousands of times more forceful than the measured changes, this invention was considered impossible until LaCoste demonstrated it.
https://en.wikipedia.org/wiki/Malettinin
Malettinins are polyketide-derived, antimicrobials made by Hypoxylon. External links Malettinins B-D: new polyketide metabolites from an unidentified fungal colonist of Hypoxylon Stromata (NRRL 29110). Antimicrobials
https://en.wikipedia.org/wiki/Copy-and-paste%20programming
Copy-and-paste programming, sometimes referred to as just pasting, is the production of highly repetitive computer programming code, as produced by copy and paste operations. It is primarily a pejorative term; those who use the term are often implying a lack of programming competence. It may also be the result of technology limitations (e.g., an insufficiently expressive development environment) as subroutines or libraries would normally be used instead. However, there are occasions when copy-and-paste programming is considered acceptable or necessary, such as for boilerplate, loop unrolling (when not supported automatically by the compiler), or certain programming idioms, and it is supported by some source code editors in the form of snippets. Origins Copy-and-paste programming is often done by inexperienced or student programmers, who find the act of writing code from scratch difficult or irritating and prefer to search for a pre-written solution or partial solution they can use as a basis for their own problem solving. (See also Cargo cult programming) Inexperienced programmers who copy code often do not fully understand the pre-written code they are taking. As such, the problem arises more from their inexperience and lack of courage in programming than from the act of copying and pasting, per se. The code often comes from disparate sources such as friends' or co-workers' code, Internet forums, code provided by the student's professors/TAs, or computer science textbooks. The result risks being a disjointed clash of styles, and may have superfluous code that tackles problems for which new solutions are no longer required. A further problem is that bugs can easily be introduced by assumptions and design choices made in the separate sources that no longer apply when placed in a new environment. Such code may also, in effect, be unintentionally obfuscated, as the names of variables, classes, functions and the like are typically left unchanged, even though their
https://en.wikipedia.org/wiki/Introduction%20of%20species%20to%20Mana%20Island
The New Zealand Department of Conservation has introduced a number of animal and plant species to Mana Island, near Porirua, New Zealand, as part of an ecological restoration programme since taking over conservation management of the island in 1987. Some were reintroductions of species wiped out during years of agricultural use or the subsequent explosion in the number of house mice on the island. A pest control programme eliminated the mice by 1990 and many species have been introduced since. Notable successes include the spotted skink, Duvaucel's gecko and the flax weevil, which are now regarded as well established, and the yellow-crowned parakeet, which has been described as abundant and widespread. Attempts have been made to introduce several seabird species in what the Department of Conservation describes as "the world's most complex seabird translocation project"; results have been mixed. Background Mana Island is a small (approx ) island off the southwest coast of the North Island, New Zealand. Since 1987 the island has been owned and managed by New Zealand's Department of Conservation and is administered as a scientific reserve. The island was previously used for agriculture, particularly the raising of sheep and cattle, and after grazing ceased, the growth of grass allowed a rapid increase of the house mouse population. The mice preyed on several native species of lizard and birds and damaged newly planted trees. An eradication programme successfully reduced the mouse population from 5 million in August 1989 to zero by February 1990. The Department of Conservation carries out works to revegetate the island (which was probably originally forested) and restore it to an example of a Cook Strait island ecosystem. The Friends of Mana Island volunteer group was formed in 1998 to carry out work at the island and to raise funds for its conservation. A formal ecological restoration plan was established in 1999 and measures taken to reduce the growth of
https://en.wikipedia.org/wiki/Sphenopetrosal%20fissure
The sphenopetrosal fissure (or sphenopetrosal suture) is the cranial suture between the sphenoid bone and the petrous portion of the temporal bone. It is in the middle cranial fossa. External links Skull
https://en.wikipedia.org/wiki/Moment%20of%20inertia
The moment of inertia, otherwise known as the mass moment of inertia, angular mass, second moment of mass, or most accurately, rotational inertia, of a rigid body is a quantity that determines the torque needed for a desired angular acceleration about a rotational axis, akin to how mass determines the force needed for a desired acceleration. It depends on the body's mass distribution and the axis chosen, with larger moments requiring more torque to change the body's rate of rotation by a given amount. It is an extensive (additive) property: for a point mass the moment of inertia is simply the mass times the square of the perpendicular distance to the axis of rotation. The moment of inertia of a rigid composite system is the sum of the moments of inertia of its component subsystems (all taken about the same axis). Its simplest definition is the second moment of mass with respect to distance from an axis. For bodies constrained to rotate in a plane, only their moment of inertia about an axis perpendicular to the plane, a scalar value, matters. For bodies free to rotate in three dimensions, their moments can be described by a symmetric 3-by-3 matrix, with a set of mutually perpendicular principal axes for which this matrix is diagonal and torques around the axes act independently of each other. In mechanical engineering, simply "inertia" is often used to refer to "inertial mass" or "moment of inertia". Introduction When a body is free to rotate around an axis, torque must be applied to change its angular momentum. The amount of torque needed to cause any given angular acceleration (the rate of change in angular velocity) is proportional to the moment of inertia of the body. Moments of inertia may be expressed in units of kilogram metre squared (kg·m2) in SI units and pound-foot-second squared (lbf·ft·s2) in imperial or US units. The moment of inertia plays the role in rotational kinetics that mass (inertia) plays in linear kinetics—both characterize the resistance
https://en.wikipedia.org/wiki/Equation%20solving
In mathematics, to solve an equation is to find its solutions, which are the values (numbers, functions, sets, etc.) that fulfill the condition stated by the equation, consisting generally of two expressions related by an equals sign. When seeking a solution, one or more variables are designated as unknowns. A solution is an assignment of values to the unknown variables that makes the equality in the equation true. In other words, a solution is a value or a collection of values (one for each unknown) such that, when substituted for the unknowns, the equation becomes an equality. A solution of an equation is often called a root of the equation, particularly but not only for polynomial equations. The set of all solutions of an equation is its solution set. An equation may be solved either numerically or symbolically. Solving an equation numerically means that only numbers are admitted as solutions. Solving an equation symbolically means that expressions can be used for representing the solutions. For example, the equation is solved for the unknown by the expression , because substituting for in the equation results in , a true statement. It is also possible to take the variable to be the unknown, and then the equation is solved by . Or and can both be treated as unknowns, and then there are many solutions to the equation; a symbolic solution is , where the variable may take any value. Instantiating a symbolic solution with specific numbers gives a numerical solution; for example, gives (that is, ), and gives . The distinction between known variables and unknown variables is generally made in the statement of the problem, by phrases such as "an equation in and ", or "solve for and ", which indicate the unknowns, here and . However, it is common to reserve , , , ... to denote the unknowns, and to use , , , ... to denote the known variables, which are often called parameters. This is typically the case when considering polynomial equations, such as quad
https://en.wikipedia.org/wiki/Armed%20Forces%20Institute%20of%20Pathology%2C%20Bangladesh
Armed Forces Institute of Pathology is a medical research laboratory of the Bangladesh Armed Forces. It is the premier diagnostic institution of Bangladesh Armed Forces. It is situated in Dhaka Cantonment. Major General Nishat Jubaida is the current Commandant of this institution. History Armed Forces Institute of Pathology was established as a small laboratory designated as Army Pathological Laboratory (APL) in 1951 as part of the Pakistan Army. After the Independence of Bangladesh it became part of Bangladesh Army. In 1974 APL was restructured and re-equipped with modern equipment. In 1987, the institute was awarded the Independence Day Award. In 1991 Army Pathological Laboratory was renamed to Armed Forces Institute of Pathology. This Institute is recognized by the Bangladesh College of Physicians and Surgeons (BCPS), University of Dhaka, Bangabandhu Sheikh Mujib Medical University (BSMMU) and also by Bangladesh University of Professionals (BUP) for postgraduate studies and training.
https://en.wikipedia.org/wiki/Intrascleral%20plexus
The intrascleral plexus is the network of blood vessels within the substance of the sclera. Blood enters from the small connecting channels from Schlemm's canal; it exits through superficial vessels, then passing it on to the anterior ciliary veins.
https://en.wikipedia.org/wiki/SuEXEC
Apache suEXEC is a feature of the Apache web server. It allows users to run Common Gateway Interface (CGI) and Server Side Includes (SSI) applications as a different user. Normally, all web server processes run as the default web server user (often wwwrun, www-data, apache or nobody). The suEXEC feature consists of a module for the web server and a binary executable which acts as a wrapper. suEXEC was introduced in Apache 1.2 and is often included in the default Apache package provided by most Linux distributions. If a client requests a CGI and suEXEC is activated, it will call the suEXEC binary which then wraps the CGI scripts and executes it under the user account of the server process (virtual host) defined in the virtual host directive. Additionally, suEXEC perform a multi-step check on the executed CGI to ensure security for the server (including path-checks, a limit of permitted commands, etc.) Example User "Alice" has a website including some Common Gateway Interface script files in her own public_html (a common web root directory name) folder, which can be accessed by https://example.com/~alice. User "Bob" now views Alice's webpage, which requires Apache to run one of these CGI scripts. Instead of running all scripts as "wwwrun" (which results in the need that all scripts have to be readable and executable for the "wwwrun" group if the file is owned by that group or for all users otherwise), the scripts in /home/alice/public_html will be wrapped using suEXEC and run with Alice's user ID resulting in higher security and eliminating the need to make the scripts readable and executable for all users or everyone in the "wwwrun" group (instead only Alice herself needs to be able to run the script).
https://en.wikipedia.org/wiki/Johann%20Georg%20Schr%C3%B6pfer
Johann Georg Schrepfer, or Johann Georg Schröpfer (1738? – 8 October 1774 in Leipzig), was a German charlatan, independent Freemason and necromancer. He performed ghost-raising séances for which he secretly used special effects, possibly including magic lantern projections of ghosts on smoke, which inspired the phantasmagoria shows popular in Europe from the 1790s to the 1830s. Little is certain about Schrepfer's life. Many accounts stem from either adherent or opposing Freemasons and Schrepfer himself was an impostor who told many lies about himself. Biography Schrepfer's year of birth is often stated to be 1730 and sometimes as 1739, but he was baptized in Nürnberg on March 26, 1738. He was listed as the eighth child of the host of "Zum Rothen Roße". Father Schrepfer later ran the "Goldenen Lamm" and seems to have gone bankrupt in 1753. Johann Georg reportedly served for some time as a hussar in the Prussian army at the start of the Seven Years' War. However, he also claimed to have been a cavalry commander with the imperial troops and to have received many wounds in battle (although his autopsy report clearly stated that no scars or signs of old injuries were found on his body). He came to Leipzig in 1759 and started as a "Küper" (controller of goods) in Hotel de Saxe. He became a citizen of the city in August 1761 and was registered as a waiter. September 20, 1761 he married Johanna Katharina Herr, the daughter of the quartermaster of the local tailors. She was already highly pregnant and soon gave birth to their first daughter. In 1769 they purchased the "Weisslederische Coffeé-Hauß" with Johanna's money. It was a café with a billiard room in the very center of town (at the corner of Klostergasse and the Barfußgäßchen, the location of restaurant "Zill's Tunnel" since 1841). Saxony went through some hunger years in 1770–1771. Schrepfer had debts and his family seemed to live in relative poverty, but he reportedly maintained a frivolous lifestyle and often dr
https://en.wikipedia.org/wiki/Cash-flow%20return%20on%20investment
Cash-flow return on investment (CFROI) is a valuation model that assumes the stock market sets prices based on cash flow, not on corporate performance and earnings. For the corporation, it is essentially internal rate of return (IRR). CFROI is compared to a hurdle rate to determine if investment/product is performing adequately. The hurdle rate is the total cost of capital for the corporation calculated by a mix of cost of debt financing plus investors' expected return on equity investments. The CFROI must exceed the hurdle rate to satisfy both the debt financing and the investors expected return. Michael J. Mauboussin in his 2006 book More Than You Know: Finding Financial Wisdom in Unconventional Places, quoted an analysis by Credit Suisse First Boston, that, measured by CFROI, performance of companies tend to converge after five years in terms of their survival rates. The CFROI for a firm or a division can then be written as follows: This annuity is called the economic depreciation: where n is the expected life of the asset and Kc is the replacement cost in current dollars. See also Return of capital
https://en.wikipedia.org/wiki/Cold%20chill
A cold chill (also known as chills, the chills or simply thrills) is described by David Huron as, "a pleasant tingling feeling, associated with the flexing of hair follicles resulting in goose bumps (technically called piloerection), accompanied by a cold sensation, and sometimes producing a shudder or shiver." Dimpled skin is often visible due to cold chills especially on the back of the neck or upper spine. Unlike shivering, however, it is not caused by temperature, but rather is an emotionally triggered response when one is deeply affected by things such as music, speech, or recollection. It is similar to autonomous sensory meridian response; both sensations consist of a pleasant tingling feeling that affects the skin on the back of the neck and spine. See also Chills Autonomous sensory meridian response Frisson
https://en.wikipedia.org/wiki/Riesz%20mean
In mathematics, the Riesz mean is a certain mean of the terms in a series. They were introduced by Marcel Riesz in 1911 as an improvement over the Cesàro mean. The Riesz mean should not be confused with the Bochner–Riesz mean or the Strong–Riesz mean. Definition Given a series , the Riesz mean of the series is defined by Sometimes, a generalized Riesz mean is defined as Here, the are a sequence with and with as . Other than this, the are taken as arbitrary. Riesz means are often used to explore the summability of sequences; typical summability theorems discuss the case of for some sequence . Typically, a sequence is summable when the limit exists, or the limit exists, although the precise summability theorems in question often impose additional conditions. Special cases Let for all . Then Here, one must take ; is the Gamma function and is the Riemann zeta function. The power series can be shown to be convergent for . Note that the integral is of the form of an inverse Mellin transform. Another interesting case connected with number theory arises by taking where is the Von Mangoldt function. Then Again, one must take c > 1. The sum over ρ is the sum over the zeroes of the Riemann zeta function, and is convergent for λ > 1. The integrals that occur here are similar to the Nörlund–Rice integral; very roughly, they can be connected to that integral via Perron's formula.
https://en.wikipedia.org/wiki/Omental%20foramen
In human anatomy, the omental foramen (epiploic foramen, foramen of Winslow after the anatomist Jacob B. Winslow, or uncommonly aditus; ), is the passage of communication, or foramen, between the greater sac (general cavity (of the abdomen)), and the lesser sac. Borders It has the following borders: anterior: the free border of the lesser omentum, known as the hepatoduodenal ligament. This has two layers and within these layers are the common bile duct, hepatic artery, and hepatic portal vein. posterior: the peritoneum covering the inferior vena cava superior: the peritoneum covering the caudate lobe of the liver inferior: the peritoneum covering the commencement of the duodenum and the hepatic artery, the latter passing forward below the foramen before ascending between the two layers of the lesser omentum. left lateral: gastrosplenic ligament and splenorenal ligament As the portal vein is the most posterior structure in the hepatoduodenal ligament, and the inferior vena cava lies under the posterior wall, the epiploic foramen can be remembered as lying between the two great veins of the abdomen. Additional images See also Terms for anatomical location Omental bursa (Lesser sac) Greater sac Lesser omentum Greater omentum Peritoneum
https://en.wikipedia.org/wiki/Staatsarchiv%20Ludwigsburg
The Staatsarchiv Ludwigsburg (Ludwigsburg State Archives), located in Ludwigsburg, Germany, is a public institutional repository for roughly 680 state authorities within the District of Stuttgart, Germany. Holdings The historical holdings come from the North Württemberg (de) area. Notable holdings, unveiled through the process of denazification, include documents related to the Holocaust (more than 500,000). The archives also house the files of the Police Headquarters Stuttgart (de), the decorative folders and role books of the Staatstheater Stuttgart and all birth records of the Landesfrauenklinik Stuttgart (Stuttgart Women's State Hospital); the documents of the Teutonic Order, official books of the convent Ellwangen and documents of Ulm, Esslingen am Neckar, Heilbronn, and other former free imperial cities. The collections also include personnel files of the Deutsche Reichsbahn (the German National Railway) and the Deutsche Bundesbahn (the German Federal Railway) from the Reichsbahndirektion/Bundesbahndirektion Stuttgart (de) (Federal Railway Directorate). Facilities The Ludwigsburg State Archives relocated, in 1995, from a castle to the Arsenal Barracks and Armory located in the center of Ludwigsburg. The Arsenal Barracks (Arsenalkaserne) houses the public areas – including a reading room, an auditorium, an exhibition room, library, and an administration office The Armory (Zeughaus) houses the stacks, which comprise more than of archives Both buildings were mechanically and structurally re-purposed to meet the safekeeping requirements of modern archives. The buildings are connected by an underground corridor, through which a transport system delivers documents to researchers in the reading room. Governmental oversight As part of the Baden-Württemberg administrative reorganization of 2005, Staatsarchiv Ludwigsburg has been a department of the newly created Landesarchiv Baden-Württemberg (de) (Baden-Württemberg State Archives), which was inaugurated
https://en.wikipedia.org/wiki/Positivity%20effect
The positivity effect is the ability to constructively analyze a situation where the desired results are not achieved, but still obtain positive feedback that assists one's future progression. The findings of a study show that the optimism bias in future-oriented thinking fulfils a self-improvement purpose while also suggesting this bias probably reflects a common underpinning motivational process across various future-thinking domains, either episodic or semantic. In attribution The positivity effect as an attribution phenomenon relates to the habits and characteristics of people when evaluating the causes of their behaviors. To positively attribute is to be open to attributing a person’s inherent disposition as the cause of their positive behaviors, and the situations surrounding them as the potential cause of their negative behaviors. In perception Two studies by Emilio Ferrara have shown that, on online social networks like Twitter and Instagram, users prefer to share positive news, and are emotionally affected by positive news more than twice as much as they are by negative news. According to the research recorded by Dan Zarella, the more positive a person is on social media, the more followers they will get because "users become less engaged when content on their feed becomes more negative" (Lee 1). So, when someone posts a lot of positive things, it makes people want to be a part of their social media presence. People on social media seek out positivity. See also List of biases in judgment and decision making List of memory biases Optimism bias Pollyanna principle Positivity offset Rosy retrospection Self-serving bias Toxic positivity Wishful thinking Notes
https://en.wikipedia.org/wiki/Comparison%20of%20American%20and%20British%20English
The English language was introduced to the Americas by British colonisation, beginning in the late 16th and early 17th centuries. The language also spread to numerous other parts of the world as a result of British trade and colonisation and the spread of the former British Empire, which, by 1921, included 470–570 million people, about a quarter of the world's population. Note that in England, Wales, Ireland and especially parts of Scotland there are differing varieties of the English language, so the term 'British English' is an oversimplification. Written forms of 'British' and American English as found in newspapers and textbooks vary little in their essential features, with only occasional noticeable differences. Over the past 400 years, the forms of the language used in the Americas—especially in the United States—and that used in the United Kingdom have diverged in a few minor ways, leading to the versions now often referred to as American English and British English. Differences between the two include pronunciation, grammar, vocabulary (lexis), spelling, punctuation, idioms, and formatting of dates and numbers. However, the differences in written and most spoken grammar structure tend to be much fewer than in other aspects of the language in terms of mutual intelligibility. A few words have completely different meanings in the two versions or are even unknown or not used in one of the versions. One particular contribution towards formalising these differences came from Noah Webster, who wrote the first American dictionary (published 1828) with the intention of showing that people in the United States spoke a different dialect from those spoken in the UK, much like a regional accent. This divergence between American English and British English has provided opportunities for humorous comment: e.g. in fiction George Bernard Shaw says that the United States and United Kingdom are "two countries divided by a common language"; and Oscar Wilde says that "We have
https://en.wikipedia.org/wiki/Rosser%27s%20trick
In mathematical logic, Rosser's trick is a method for proving a variant of Gödel's incompleteness theorems not relying on the assumption that the theory being considered is ω-consistent (Smorynski 1977, p. 840; Mendelson 1977, p. 160). This method was introduced by J. Barkley Rosser in 1936, as an improvement of Gödel's original proof of the incompleteness theorems that was published in 1931. While Gödel's original proof uses a sentence that says (informally) "This sentence is not provable", Rosser's trick uses a formula that says "If this sentence is provable, there is a shorter proof of its negation". Background Rosser's trick begins with the assumptions of Gödel's incompleteness theorem. A theory is selected which is effective, consistent, and includes a sufficient fragment of elementary arithmetic. Gödel's proof shows that for any such theory there is a formula which has the intended meaning that is a natural number code (a Gödel number) for a formula and is the Gödel number for a proof, from the axioms of , of the formula encoded by . (In the remainder of this article, no distinction is made between the number and the formula encoded by , and the number coding a formula is denoted .) Furthermore, the formula is defined as . It is intended to define the set of formulas provable from . The assumptions on also show that it is able to define a negation function , with the property that if is a code for a formula then is a code for the formula . The negation function may take any value whatsoever for inputs that are not codes of formulas. The Gödel sentence of the theory is a formula , sometimes denoted , such that proves  ↔. Gödel's proof shows that if is consistent then it cannot prove its Gödel sentence; but in order to show that the negation of the Gödel sentence is also not provable, it is necessary to add a stronger assumption that the theory is ω-consistent, not merely consistent. For example, the theory , in which PA is Peano axioms, pr
https://en.wikipedia.org/wiki/PB286LP
The PB286LP, released in 1989, was Packard Bell's first laptop computer. The laptop featured an 80C286 processor clocked at 12 MHz and 1 MB of RAM, along with a single ISA expansion slot. Packard Bell released the PB286LP in 1989 among a slew of products aimed at the corporate market. Technology writers gave it mostly positive reviews, although some noted its 16-lb weight as hefty and its monochrome LCD as somewhat flawed. Originally only capable of CGA-mode graphics, the laptop was updated in 1990 to support VGA. Packard Bell discontinued the PB286LP in 1991, in favor of more-compact, notebook-sized computers. Development and specifications The PB286LP was the first laptop of Packard Bell, a defunct trademark of Teledyne Technologies revived as a computer manufacturer by Beny Alagem, Alex Sandel and Jason Barzilay in 1986. The laptop was announced in November 1988 amid a barrage of computers aimed at the corporate market. The PB286LP, which features an 80C286 processor clocked at 12 MHz—downclockable to 6 MHz—was marketed to existing buyers of 80286-based personal computers. It arrived stock with 1 MB of RAM SIMMs, expandable up to 5 MB. Packard Bell included a single full-length ISA expansion slot to offer users a means of expanding and upgrading the laptop. The laptop carries a 1.44 MB floppy disk drive and was optioned with either a 20 MB or a 40 MB hard disk drive. Conner and TEAC manufactured the hard and floppy drives respectively. In addition the PB286LP had an optional 2400-bit/s modem. The systems chipset of the PB286LP was manufactured by Intel, while the BIOS was provided by Award. The PB286LP's video chipset was designed in-house by Packard Bell and has an interface eight bits wide and support for both double-scan CGA and MDA. The laptop's monochrome LCD measures 9.7 inches diagonally with an aspect ratio of 1.6:1. Its removable nickel–cadmium battery had a claimed battery life of three hours in 1990. Packard Bell allowed users to enable a battery-con
https://en.wikipedia.org/wiki/Gallai%E2%80%93Edmonds%20decomposition
In graph theory, the Gallai–Edmonds decomposition is a partition of the vertices of a graph into three subsets which provides information on the structure of maximum matchings in the graph. Tibor Gallai and Jack Edmonds independently discovered it and proved its key properties. The Gallai–Edmonds decomposition of a graph can be found using the blossom algorithm. Properties Given a graph , its Gallai–Edmonds decomposition consists of three disjoint sets of vertices, , , and , whose union is : the set of all vertices of . First, the vertices of are divided into essential vertices (vertices which are covered by every maximum matching in ) and inessential vertices (vertices which are left uncovered by at least one maximum matching in ). The set is defined to contain all the inessential vertices. Essential vertices are split into and : the set is defined to contain all essential vertices adjacent to at least one vertex of , and is defined to contain all essential vertices not adjacent to any vertices of . It is common to identify the sets , , and with the subgraphs induced by those sets. For example, we say "the components of " to mean the connected components of the subgraph induced by . The Gallai–Edmonds decomposition has the following properties: The components of are factor-critical graphs: each component has an odd number of vertices, and when any one of these vertices is left out, there is a perfect matching of the remaining vertices. In particular, each component has a near-perfect matching: a matching that covers all but one of the vertices. The subgraph induced by has a perfect matching. Every subset has neighbors in at least components of . Every maximum matching in has the following structure: it consists of a near-perfect matching of each component of , a perfect matching of , and edges from all vertices in to distinct components of . If has components, then the number of edges in any maximum matching in is . Construction The Gal
https://en.wikipedia.org/wiki/Gabriela%20Araujo-Pardo
Martha Gabriela Araujo-Pardo is a Mexican mathematician specializing in graph theory, including work on graph coloring, Kneser graphs, cages, and finite geometry. She is a researcher at the National Autonomous University of Mexico in the Mathematics Institute, Juriquilla Campus, and the 2022–2024 president of the Mexican Mathematical Society. Education and career Araujo studied mathematics at the National Autonomous University of Mexico (UNAM), where she completed her Ph.D. in 2000. Her dissertation, Daisy Structure in Desarguesian Projective Planes, was supervised by Luis Montejano Peimbert. She has worked for the UNAM Mathematics Institute since 2000, with a postdoctoral research visit to the Polytechnic University of Catalonia in Spain. She is president of the Mexican Mathematical Society for the 2022–2024 term. Recognition In 2013, Araujo won UNAM's Sor Juana Inés de la Cruz 2013 award, and was elected to the Mexican Academy of Sciences.
https://en.wikipedia.org/wiki/Network%20Security%20Services
Network Security Services (NSS) is a collection of cryptographic computer libraries designed to support cross-platform development of security-enabled client and server applications with optional support for hardware TLS/SSL acceleration on the server side and hardware smart cards on the client side. NSS provides a complete open-source implementation of cryptographic libraries supporting Transport Layer Security (TLS) / Secure Sockets Layer (SSL) and S/MIME. NSS releases prior to version 3.14 are tri-licensed under the Mozilla Public License 1.1, the GNU General Public License, and the GNU Lesser General Public License. Since release 3.14, NSS releases are licensed under GPL-compatible Mozilla Public License 2.0. History NSS originated from the libraries developed when Netscape invented the SSL security protocol. FIPS 140 validation and NISCC testing The NSS software crypto module has been validated five times (in 1997, 1999, 2002, 2007, and 2010) for conformance to FIPS 140 at Security Levels 1 and 2. NSS was the first open source cryptographic library to receive FIPS 140 validation. The NSS libraries passed the NISCC TLS/SSL and S/MIME test suites (1.6 million test cases of invalid input data). Applications that use NSS AOL, Red Hat, Sun Microsystems/Oracle Corporation, Google and other companies and individual contributors have co-developed NSS. Mozilla provides the source code repository, bug tracking system, and infrastructure for mailing lists and discussion groups. They and others named below use NSS in a variety of products, including the following: Mozilla client products, including Firefox, Thunderbird, SeaMonkey, and Firefox for mobile. AOL Communicator and AOL Instant Messenger (AIM) Open source client applications such as Evolution, Pidgin, and OpenOffice.org 2.0 onward (and its descendants). Server products from Red Hat: Red Hat Directory Server, Red Hat Certificate System, and the mod nss SSL module for the Apache web server. Sun server produ
https://en.wikipedia.org/wiki/Laccaria%20proxima
Laccaria proxima is a species of edible mushroom in the genus Laccaria from the conifer forest of California, as well as eastern and northern North America.
https://en.wikipedia.org/wiki/Ethernut
Ethernut is an open source hardware and software project for use as an embedded-Ethernet-system. Hardware Currently four different reference designs exist (Ethernut 1, Ethernut 2, Ethernut 3 and Ethernut 5) which are available as ready to use boards. All reference designs share the same set of interfaces: Ethernet Connector RS-232 ports Power connector Nearly all I/O pins of the microcontroller are available on an expansion port The designs mainly differ in the used microcontroller (Ethernut 1 & 2: 8-bit AVR, Ethernut 3 & 5: 32-bit ARM) and the available memory size. Since the Ethernut 3 board an MMC slot is available. Ethernut 5 also includes USB host and device interfaces. History With the rise of the idea of the Internet of Things, different groups of people with different aims started around 2000 developing embedded systems which would be able to integrate in existing TCP/IP networks. The limited resources of this mostly 8-bit embedded systems, which rarely reached or exceeded 64 kB of main memory, imposed a great challenge. In the year 2000 the British developer Dave Hudson released the source code of a real-time operating system with integrated TCP/IP stack named Liquorice. Another challenge was that suitable hardware for this kind of development was nearly non-existent. Nearly all network controllers of that time used the PCI-Bus interface which is too complex for small microcontrollers. But one of the few still-in-production controllers with ISA-Bus interface, which also was not recommended for new designs, was used together with ATmega103 microcontroller to create a single board computer. The layout of this board was released under BSD license and was one of the early open hardware projects. After one year Hudson ceased work on the project. But the source code of Liquorice was used as software foundation for the new Ethernut project which aimed to integrate software and hardware development. After the first few versions had been released to
https://en.wikipedia.org/wiki/Jaro%E2%80%93Winkler%20distance
In computer science and statistics, the Jaro–Winkler similarity is a string metric measuring an edit distance between two sequences. It is a variant of the Jaro distance metric metric (1989, Matthew A. Jaro) proposed in 1990 by William E. Winkler. The Jaro–Winkler distance uses a prefix scale which gives more favourable ratings to strings that match from the beginning for a set prefix length . The higher the Jaro–Winkler distance for two strings is, the less similar the strings are. The score is normalized such that 0 means an exact match and 1 means there is no similarity. The original paper actually defined the metric in terms of similarity, so the distance is defined as the inversion of that value (distance = 1 − similarity). Although often referred to as a distance metric, the Jaro–Winkler distance is not a metric in the mathematical sense of that term because it does not obey the triangle inequality. Definition Jaro similarity The Jaro similarity of two given strings and is Where: is the length of the string ; is the number of matching characters (see below); is the number of transpositions (see below). Jaro similarity score is 0 if the strings do not match at all, and 1 if they are an exact match. In the first step, each character of is compared with all its matching characters in . Two characters from and respectively, are considered matching only if they are the same and not farther than characters apart. For example, the following two nine character long strings, FAREMVIEL and FARMVILLE, have 8 matching characters. 'F', 'A' and 'R' are in the same position in both strings. Also 'M', 'V', 'I', 'E' and 'L' are within three (result of ) characters away. If no matching characters are found then the strings are not similar and the algorithm terminates by returning Jaro similarity score 0. If non-zero matching characters are found, the next step is to find the number of transpositions. Transposition is the number of matching characters
https://en.wikipedia.org/wiki/TI%20Advanced%20Scientific%20Computer
The Advanced Scientific Computer (ASC) is a supercomputer designed and manufactured by Texas Instruments (TI) between 1966 and 1973. The ASC's central processing unit (CPU) supported vector processing, a performance-enhancing technique which was key to its high-performance. The ASC, along with the Control Data Corporation STAR-100 supercomputer (which was introduced in the same year), were the first computers to feature vector processing. However, this technique's potential was not fully realized by either the ASC or STAR-100 due to an insufficient understanding of the technique; it was the Cray Research Cray-1 supercomputer, announced in 1975 that would fully realize and popularize vector processing. The more successful implementation of vector processing in the Cray-1 would demarcate the ASC (and STAR-100) as first-generation vector processors, with the Cray-1 belonging in the second. History TI began as a division of Geophysical Service Incorporated (GSI), a company that performed seismic surveys for oil exploration companies. GSI was now a subsidiary of TI, and TI wanted to apply the latest computer technology to the processing and analysis of seismic datasets. The ASC project started as the Advanced Seismic Computer. As the project developed, TI decided to expand its scope. "Seismic" was replaced by "Scientific" in the name, allowing the project to retain the designation ASC. Originally the software, including an operating system and a FORTRAN compiler, were done under contract by Computer Usage Company, under direction of George R. Trimble, Jr. but later taken over by TI itself. Southern Methodist University in Dallas developed an ALGOL compiler for the ASC. Architecture The ASC was based around a single high-speed shared memory, which was accessed by the CPU and eight I/O channel controllers, in an organization similar to Seymour Cray's groundbreaking CDC 6600. Memory was accessed solely under the control of the memory control unit (MCU). The MCU was a tw
https://en.wikipedia.org/wiki/Quantum%20cryptography
Quantum cryptography is the science of exploiting quantum mechanical properties to perform cryptographic tasks. The best known example of quantum cryptography is quantum key distribution, which offers an information-theoretically secure solution to the key exchange problem. The advantage of quantum cryptography lies in the fact that it allows the completion of various cryptographic tasks that are proven or conjectured to be impossible using only classical (i.e. non-quantum) communication. For example, it is impossible to copy data encoded in a quantum state. If one attempts to read the encoded data, the quantum state will be changed due to wave function collapse (no-cloning theorem). This could be used to detect eavesdropping in quantum key distribution (QKD). History In the early 1970s, Stephen Wiesner, then at Columbia University in New York, introduced the concept of quantum conjugate coding. His seminal paper titled "Conjugate Coding" was rejected by the IEEE Information Theory Society but was eventually published in 1983 in SIGACT News. In this paper he showed how to store or transmit two messages by encoding them in two "conjugate observables", such as linear and circular polarization of photons, so that either, but not both, properties may be received and decoded. It was not until Charles H. Bennett, of the IBM's Thomas J. Watson Research Center, and Gilles Brassard met in 1979 at the 20th IEEE Symposium on the Foundations of Computer Science, held in Puerto Rico, that they discovered how to incorporate Wiesner's findings. "The main breakthrough came when we realized that photons were never meant to store information, but rather to transmit it." In 1984, building upon this work, Bennett and Brassard proposed a method for secure communication, which is now called BB84. Independently, in 1991 Artur Ekert proposed to use Bell's inequalities to achieve secure key distribution. Ekert's protocol for the key distribution, as it was subsequently shown by Dominic Ma
https://en.wikipedia.org/wiki/Discrete-time%20proportional%20hazards
Hazard rate models are widely used to model duration data in a wide range of disciplines, from bio-statistics to economics. Grouped duration data are widespread in many applications. Unemployment durations are typically measured over weeks or months and these time intervals may be considered too large for continuous approximations to hold. In this case, we will typically have grouping points , where . Models allow for time-invariant and time-variant covariates, but the latter require stronger assumptions in terms of exogeneity. The discrete-time hazard function can be written as: where is the survivor function. It can be shown that this can be rewritten as: These probabilities provide the building blocks for setting up the Likelihood function, which ends up being: This maximum likelihood maximization depends on the specification of the baseline hazard functions. These specifications include fully parametric models, piece-wise-constant proportional hazard models, or partial likelihood approaches that estimate the baseline hazard as a nuisance function. Alternatively, one can be more flexible for the baseline hazard and impose more structure for This approach performs well for certain measures and can approximate arbitrary hazard functions relatively well, while not imposing stringent computational requirements. When the covariates are omitted from the analysis, the maximum likelihood boils down to the Kaplan-Meier estimator of the survivor function. Another way to model discrete duration data is to model transitions using binary choice models.
https://en.wikipedia.org/wiki/Ecotoxicology%20%28journal%29
Ecotoxicology is a peer-reviewed scientific journal covering ecotoxicology. It was established in 1992 and is published ten times per year by Springer Science+Business Media. The editor-in-chief is Lee R. Shugart. According to the Journal Citation Reports, the journal has a 2015 impact factor of 2.329.
https://en.wikipedia.org/wiki/Berkeley%20Madonna
Berkeley Madonna is a mathematical modelling software package, developed at the University of California at Berkeley by Robert Macey and George Oster. It numerically solves ordinary differential equations and difference equations, originally developed to execute STELLA programs. Its strength lies in a relatively simple syntax to define differential equations coupled with a simple yet powerful user interface. In particular, Berkeley Madonna provides the facility of putting parameters onto a slider that can in turn be moved by a user to change the value. Such visualizations enable quick assessments of whether or not a particular model class is suitable to describe the data to be analyzed and modeled, and, later, communicating models easily to other disciplines such as medical decision makers. Uses It has become a standard in the development and communication of pharmacometric models describing drug concentration and its effects in drug development as well as modeling of physiological processes. A user community exists in the form of a LinkedIn user group with more than 750 members (February 2023). The use of system dynamics modeling has expanded into other areas such as system physics, epidemiology, environmental health, and population ecology. Versions There are two versions of Berkeley Madonna: a free version with slightly limited functionality and a licensed version that is registered to individuals.
https://en.wikipedia.org/wiki/Rosenbrock%20function
In mathematical optimization, the Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is used as a performance test problem for optimization algorithms. It is also known as Rosenbrock's valley or Rosenbrock's banana function. The global minimum is inside a long, narrow, parabolic shaped flat valley. To find the valley is trivial. To converge to the global minimum, however, is difficult. The function is defined by It has a global minimum at , where . Usually, these parameters are set such that and . Only in the trivial case where the function is symmetric and the minimum is at the origin. Multidimensional generalizations Two variants are commonly encountered. One is the sum of uncoupled 2D Rosenbrock problems, and is defined only for even s: This variant has predictably simple solutions. A second, more involved variant is has exactly one minimum for (at ) and exactly two minima for —the global minimum at and a local minimum near . This result is obtained by setting the gradient of the function equal to zero, noticing that the resulting equation is a rational function of . For small the polynomials can be determined exactly and Sturm's theorem can be used to determine the number of real roots, while the roots can be bounded in the region of . For larger this method breaks down due to the size of the coefficients involved. Stationary points Many of the stationary points of the function exhibit a regular pattern when plotted. This structure can be exploited to locate them. Optimization examples The Rosenbrock function can be efficiently optimized by adapting appropriate coordinate system without using any gradient information and without building local approximation models (in contrast to many derivate-free optimizers). The following figure illustrates an example of 2-dimensional Rosenbrock function optimization by adaptive coordinate descent from starting point . The solution with the function va
https://en.wikipedia.org/wiki/Olive%20oil%20regulation%20and%20adulteration
Olive oil regulation and adulteration are complex issues overseen and studied by various governmental bodies, non-governmental organizations, and private researchers across the world. Background The International Olive Council (IOC) is an intergovernmental organization based in Madrid, Spain, with 16 member states plus the European Union. It promotes olive oil around the world by tracking production, defining quality standards, and monitoring authenticity. More than 98 percent of the world's olives are grown in IOC member nations. The United States is not a member of the IOC, and the US Department of Agriculture does not legally recognize its classifications (such as extra-virgin olive oil). On October 25, 2011, the United States adopted new olive oil standards, a revision of those that have been in place since 1948, which affect importers and domestic growers and producers by ensuring conformity with the benchmarks commonly accepted in the U.S. and abroad. The IOC officially governs 95 per cent of international production and holds great influence over the rest. IOC terminology is precise, but it can lead to confusion between the words that describe production and the words used on retail labels. Olive oil is classified by how it was produced, by its chemistry, and by its flavor. All production begins by transforming the olive fruit into olive paste. This paste is then malaxed to allow the microscopic oil droplets to concentrate. The oil is extracted by means of pressure (traditional method) or centrifugation (modern method). After extraction the remnant solid substance, called pomace, still contains a small quantity of oil. The EU regulates the use of different protected designation of origin labels for olive oils. US Customs regulations on "country of origin" state that if a non-origin nation is shown on the label, then the real origin must be shown on the same side of the label and in comparable size letters so as not to mislead the consumer. Yet most major
https://en.wikipedia.org/wiki/Biosatellite
A bio satellite is an artificial satellite designed to carry plants or animals in outer space. They are used to research the effects of space (cosmic radiation, weightlessness, etc.) on biological matter while in orbit around a celestial body. The first satellite carrying an animal (a dog, "Laika") was Soviet Sputnik 2 on November 3, 1957. On August 20, 1960 Soviet Sputnik 5 launched and recovered dogs from Earth orbit. NASA launched 3 satellites between 1966 and 1969 for the Biosatellite program. The most famous biosatellites include: Biosatellite program launched by NASA between 1966 and 1969. Bion space program by Soviet Union The Mars Gravity Biosatellite Orbiting Frog Otolith (OFO-A) See also Animals in space Biosatellite (NASA)
https://en.wikipedia.org/wiki/CD4%2B%20T%20cells%20and%20antitumor%20immunity
Understanding of the antitumor immunity role of CD4+ T cells has grown substantially since the late 1990s. CD4+ T cells (mature T-helper cells) play an important role in modulating immune responses to pathogens and tumor cells, and are important in orchestrating overall immune responses. Immunosurveillance and immunoediting This discovery furthered the development of a previously hypothesized theory, the immunosurveillance theory. The immunosurveillance theory suggests that the immune system routinely patrols the cells of the body, and, upon recognition of a cell, or group of cells, that has become cancerous, it will attempt to destroy them, thus preventing the growth of some tumors. (Burnet, 1970) More recent evidence has suggested that immunosurveillance is only part of a larger role the immune system plays in fighting cancer. Remodeling of this theory has led to the progression of the immunoediting theory, in which there are 3 phases, Elimination, Equilibrium and Escape. Elimination phase As mentioned, the elimination phase is synonymous with the classic immunosurveillance theory. In 2001, it was shown that mice deficient in RAG-2 (Recombinase Activator Gene 2) were far less capable of preventing MCA induced tumours than were wild type mice. (Shankaran et al., 2001, Bui and Schreiber, 2007) RAG proteins are necessary for the recombination events necessary to produce TCRs and Igs, and as such RAG-2 deficient mice are incapable of producing functional T, B or NK cells. RAG-2 deficient mice were chosen over other methods of inducting immunodeficiency (such as SCID mice) as an absence of these proteins does not affect DNA repair mechanisms, which becomes important when dealing with cancer, as DNA repair problems can lead to cancers themselves. This experiment provides clear evidence that the immune system does, in fact, play a role in eradication of tumor cells. Further knock out experiments showed important roles of αβ T cells, γδ T cells and NK cells in tumour
https://en.wikipedia.org/wiki/Order%20%28mathematics%29
Order in mathematics may refer to: Set theory Total order and partial order, a binary relation generalizing the usual ordering of numbers and of words in a dictionary Ordered set Order in Ramsey theory, uniform structures in consequence to critical set cardinality Algebra Order (group theory), the cardinality of a group or period of an element Order of a polynomial (disambiguation) Order of a square matrix, its dimension Order (ring theory), an algebraic structure Ordered group Ordered field Analysis Order (differential equation) or order of highest derivative, of a differential equation Leading-order terms NURBS order, a number one greater than the degree of the polynomial representation of a non-uniform rational B-spline Order of convergence, a measurement of convergence Order of derivation Order of an entire function Order of a power series, the lowest degree of its terms Ordered list, a sequence or tuple Orders of approximation in Big O notation Z-order (curve), a space-filling curve Arithmetic Multiplicative order in modular arithmetic Order of operations Orders of magnitude, a class of scale or magnitude of any amount Combinatorics Order in the Josephus permutation Ordered selections and partitions of the twelvefold way in combinatorics Ordered set, a bijection, cyclic order, or permutation Unordered subset or combination Weak order of permutations Fractals Complexor, or complex order in fractals Order of extension in Lakes of Wada Order of fractal dimension (Rényi dimensions) Orders of construction in the Pythagoras tree Geometry Long-range aperiodic order, in pinwheel tiling, for instance Graphs Graph order, the number of nodes in a graph First order and second order logic of graphs Topological ordering of directed acyclic graphs Degeneracy ordering of undirected graphs Elimination ordering of chordal graphs Order, the complexity of a structure within a graph: see haven (graph theory) and bramble (graph theory) Logic In logic, model theory and
https://en.wikipedia.org/wiki/The%20Most%20Good%20You%20Can%20Do
The Most Good You Can Do: How Effective Altruism Is Changing Ideas About Living Ethically is a 2015 Yale University Press book by moral philosopher and bioethicist Peter Singer describing and arguing for the ideas of effective altruism. As a follow-up to The Life You Can Save, which makes the moral argument for donating money to improve the lives of people in extreme poverty, the new book focuses on the broader question of how to do the most good. Reception Interviews Oliver Milman interviewed Peter Singer about the book for The Guardian shortly before the book's release. Hamilton Nolan interviewed Singer for Gawker a week after the release. Singer was also interviewed on ABC Online (an Australian media network) about his book. He also did a longer interview with the Melbourne radio channel of the network. Singer also participated in an Ask Me Anything on Reddit, fielding questions about his book, on April 14, 2015 (a week after the book's release). Book reviews Nicholas Kristof reviewed the book for The New York Times, beginning with a discussion of the earning to give strategy. Kristof had three reservations about the book: (1) it is not clear where to draw the line with respect to altruism, (2) in addition to humanitarian motives, loyalty is also important and many give to universities or the arts out of loyalty, (3) the idea of taking a job solely because it is well-paying made him flinch. Kristof concluded on a positive note: "Singer's argument is powerful, provocative and, I think, basically right. The world would be a better place if we were as tough-minded in how we donate money as in how we make it." University of Chicago Law School professor Eric Posner reviewed the book for Slate Magazine, concluding: "So what's an effective altruist to do? The utilitarian imperative to search out and help the people with the highest marginal utility of money around the world is in conflict with our limited knowledge about foreign cultures, which makes it difficult
https://en.wikipedia.org/wiki/Wingertshellicus
Wingertshellicus is an extinct genus of arthropod that has been found in Hunsrück Slate, that is located in the Rhenish Massif in Germany, and lived about 405 million years ago, during the Lower Emsian (part of the Lower Devonian). Etymology Wingertshellicus has been named after the Wingertshell Member, part of the Hunsrück Slate near the German town of Bundenbach, in which it was found. Morphology Wingertshellicus measured about long. The body of consists of just two main parts (tagmata): a short head and an elongated trunk. The head possess a pair of large stalked compound eyes and 7 pairs of appendages. The first appendages are long, annulated antennae. Within the remaining 6 pairs of leg-like appendages, the anterior 2 pairs are short and stout while the posterior 3 pairs were significantly elongated. The segmentation of the post-antennular appendages obscured by their limited preservation. A ventral triangular projection with terminal mouth opening located between the posteroir appendages. There is no complex mouthparts (e.g. mandible; labium) nor subdivision of a thorax for the 6 elongated appendages like those of an insect. The trunk comprises about 40 similar segments (somites), each carried a pair of biramous appendages with enlarged basipod (basal segment), flap-like exopods (outer branch) and slender endopods (inner branch). The appendages were subequal, lacking any evidence of specializations. the trunk terminated by a pair of fluke-like appendages with flagellate structures. Taxonomy Wingertshellicus backesi was originally described in 2001 from specimens that were exposed on their belly (or ventrally), and was tentatively identified as a crustaceanomorph at that time. Two years later, Devonohexapodus bocksbergensis was erected based on a specimen on its side (or laterally) from the same stratum. At first sight, it seems that the ventrally exposed specimens only have small and slender appendages, while the laterally exposed has three pairs of l
https://en.wikipedia.org/wiki/God%20Created%20the%20Integers
God Created the Integers: The Mathematical Breakthroughs That Changed History is a 2005 anthology, edited by Stephen Hawking, of "excerpts from thirty-one of the most important works in the history of mathematics." The title of the book is a reference to a quotation attributed to mathematician Leopold Kronecker, who once wrote that "God made the integers; all else is the work of man." Content The works are grouped by author and ordered chronologically. Each section is prefaced by notes on the mathematician's life and work. The anthology includes works by the following mathematicians: Selections from the works of Euler, Bolyai, Lobachevsky and Galois, which are included in the second edition of the book (published in 2007), were not included in the first edition. Editions
https://en.wikipedia.org/wiki/Vegeta%20%28software%29
Vegeta is an HTTP load testing tool written in Go that can be used as a command in a command-line interface or as a library. The program tests how an HTTP-based application behaves when multiple users access it at the same time by generating a background load of GET requests. Vegeta is used to generate a sustained, constant number of requests per second in order to discover how long a service can sustain a peak load before dropping in performance. In addition to preemptive load testing, the program can also be used for shadow testing, where traffic from a live version of an application is mirrored onto a test version to determine how it handles the same traffic load, without causing potential disruption to the live version of the application. Shadow testing is done in this way in order to analyze anticipated server performance. Vegeta is provided for use by web hosting services such as Scaleway to use varied and multiple requests to stress test client HTTP services. It is also used with dedicated load-testing platform services such as BlazeMeter. Usage The command-line usage is in the format of . The three global flags are which specifies the number of CPUs to use, which enables profiling, and which prints the software version and then terminates the program. The commands available are , , , and , each with its own various command flag options, and both attack input and report output can be done in an optional JSON format when specified with the appropriate flag. Vegeta can specify targets as URLs in a separate file with optional custom headers and requests, which can then be used as an input option on the command line. Example An example usage would be to issue from the command-line. This example uses the echo command to output , and then executes the attack command for that output for five seconds. After that, it uses the tee command to write results to a file called results.bin, and runs the report command to display the output of the attack results.
https://en.wikipedia.org/wiki/Hurler%20syndrome
Hurler syndrome, also known as mucopolysaccharidosis Type IH (MPS-IH), Hurler's disease, and formerly gargoylism, is a genetic disorder that results in the buildup of large sugar molecules called glycosaminoglycans (GAGs) in lysosomes. The inability to break down these molecules results in a wide variety of symptoms caused by damage to several different organ systems, including but not limited to the nervous system, skeletal system, eyes, and heart. The underlying mechanism is a deficiency of alpha-L iduronidase, an enzyme responsible for breaking down GAGs. Without this enzyme, a buildup of dermatan sulfate and heparan sulfate occurs in the body. Symptoms appear during childhood, and early death usually occurs. Other, less severe forms of MPS Type I include Hurler–Scheie syndrome (MPS-IHS) and Scheie syndrome (MPS-IS). Hurler syndrome is classified as a lysosomal storage disease. It is clinically related to Hunter syndrome (MPS II); however, Hunter syndrome is X-linked, while Hurler syndrome is autosomal recessive. Signs and symptoms Children with Hurler syndrome may appear normal at birth and develop symptoms over the first years of life. Symptoms vary between patients. One of the first abnormalities that may be detected is coarsening of the facial features; these symptoms can begin at 3–6 months of age. The head can be large with prominent frontal bones. The skull can be elongated. The nose can have a flattened nasal bridge with continuous nasal discharge. The eye sockets may be widely spaced, and the eyes may protrude from the skull. The lips can be large, and affected children may hold their jaws open constantly. Skeletal abnormalities occur by about age 6 months, but may not be clinically obvious until 10–14 months. Patients may experience debilitating spine and hip deformities, carpal tunnel syndrome, and joint stiffness. Patients may be normal height in infancy, but stop growing by the age of two years. They may not reach a height of greater than . Oth
https://en.wikipedia.org/wiki/Chaguar
Chaguar is the common name of several related species of South American plants of the family Bromeliaceae, among them Bromelia serra, Bromelia hieronymi, Deinacanthon urbanianum and Pseudananas sagenarius, which are non-woody forest plants with sword-shaped evergreen leaves, resembling yucca. The different varieties grow in the semi-arid parts of the Gran Chaco ecoregion. The term chaguar is of Quechua origin; in areas where the Guaraní have had an influence, it is also known as caraguatá. This plant is mainly employed by the Wichí tribal groups in the provinces of Salta and Formosa, Argentina, to provide a resistant fiber that can be woven to make bags and purses, ponchos, skirts, fishing nets, string, ropes, hammocks, mats, covers and clothing. Along with those made from hardwoods such as quebracho, chaguar crafts make up an important part of the economy of some Wichí groups, though the profits are scarce. Chaguar is not cultivated. It grows in the semi-shade of a middle layer of the Chaco forest, and reproduces by stolons. Desertification of the Chaco has decreased its presence, but the plant is neither endangered nor of primary ecological importance. Farmers consider it a pest and, since its spines scare cattle, they sometimes burn the chaguarales (plant colonies) during the dry season. There are no restrictions on the collection and use of chaguar among the Wichí, but the task is time-consuming and labor-intensive. This fact and the environmental ethics of the tribe discourage overexploitation.
https://en.wikipedia.org/wiki/Shrimp%20paste
Shrimp paste or prawn sauce is a fermented condiment commonly used in Southeast Asian and Southern Chinese cuisines. It is primarily made from finely crushed shrimp or krill mixed with salt, and then fermented for several weeks. They are either sold in their wet form or are sun-dried and either cut into rectangular blocks or sold in bulk. It is an essential ingredient in many curries, sauces and sambal. Shrimp paste can be found in many meals in Cambodia, Indonesia, Laos, Malaysia, Myanmar, the Philippines, Singapore, Thailand, and Vietnam. It is often an ingredient in dip for fish or vegetables. History Shrimp paste originated in continental Southeast Asia, probably among the Cham and Mon people, from where it spread southwards to insular Southeast Asia. In Java, fermented shrimp paste (trasi or terasi), as mentioned in two ancient Sundanese scriptures, Carita Purwaka Caruban Nagari and Mertasinga, had been around before sixth century. According to Carita Purwaka Caruban Nagari, Cirebon had angered the King of Galuh Kingdom after they stopped paying a tribute (in the forms of shrimp paste and salt, their regional products) to him. In Mertasinga, it was mentioned that Cirebon was attacked by Galuh Kingdom because they stopped sending trasi to the king. Shrimp paste was one of Java's most popular exports bought by traders from neighboring islands and abroad. According to Purwaka Caruban Nagari, Chinese Muslim explorer, Zheng He of Yunnan, used to buy trasi from Cirebon and brought it back to his homeland. He was the one who introduced trasi to China, a foreign condiment which later became popular and inspired locals to make their own version. In 1707, William Dampier described trasi in his book "A New Voyage Round the World"; "A composition of a strong odor, but it became a very tasty meal for the indigenous people." Dampier described it further as a mixture of shrimp and small fish made into a kind of soft pickle with salt and water, and then the dough was pack
https://en.wikipedia.org/wiki/Natamycin
Natamycin, also known as pimaricin, is an antifungal medication used to treat fungal infections around the eye. This includes infections of the eyelids, conjunctiva, and cornea. It is used as eyedrops. Natamycin is also used in the food industry as a preservative. Allergic reactions may occur. It is unclear if medical use during pregnancy or breastfeeding is safe. It is in the macrolide and polyene families of medications. It results in fungal death by altering the cell membrane. Natamycin was discovered in 1955 and approved for medical use in the United States in 1978. It is on the World Health Organization's List of Essential Medicines. It is produced by fermentation of certain types of the bacterium Streptomyces. Uses Medical Natamycin is used to treat fungal infections, including Candida, Aspergillus, Cephalosporium, Fusarium, and Penicillium. It is applied topically as a cream, in eye drops, or (for oral infections) in a lozenge. Natamycin shows negligible absorption into the body when administered in these ways. When taken orally, little or none is absorbed from the gastrointestinal tract, making it inappropriate for systemic infections. Natamycin lozenges are also prescribed to treat yeast infections and oral thrush. Food Natamycin has been used for decades in the food industry as a hurdle to fungal outgrowth in dairy products and other foods. Potential advantages for the usage of natamycin might include the replacement of traditional chemical preservatives, a neutral flavor impact, and less dependence on pH for efficacy, as is common with chemical preservatives. It can be applied in a variety of ways: as an aqueous suspension (such as mixed into a brine) sprayed on the product or into which the product is dipped, or in powdered form (along with an anticaking agent such as cellulose) sprinkled on or mixed into the product. While not currently approved for use on meats in the United States, some countries allow natamycin to be applied to the surfac
https://en.wikipedia.org/wiki/List%20of%20assigned%20/8%20IPv4%20address%20blocks
Some large blocks of IPv4 addresses, the former Class A network blocks, are assigned in whole to single organizations or related groups of organizations, either by the Internet Corporation for Assigned Names and Numbers (ICANN), through the Internet Assigned Numbers Authority (IANA), or a regional Internet registry. Each block contains 256 = 2 = 16,777,216 addresses, which covers the whole range of the last three delimited segments of an IP address. As IPv4 address exhaustion has advanced to its final stages, some organizations, such as Stanford University, formerly using , have returned their allocated blocks (in this case to APNIC) to assist in the delay of the exhaustion date. List of reserved /8 blocks List of assigned /8 blocks to commercial organisations List of assigned /8 blocks to the United States Department of Defense List of assigned /8 blocks to the regional Internet registries The regional Internet registries (RIR) allocate IPs within a particular region of the world. Note that this list may not include current assignments of /8 blocks to all regional or national Internet registries. Original list of IPv4 assigned address blocks The original list of IPv4 address blocks was published in September 1981. In previous versions of the document, network numbers were 8-bit numbers rather than the 32-bit numbers used in IPv4. At that time, three networks were added that were not listed earlier: 42.rrr.rrr.rrr, 43.rrr.rrr.rrr, and 44.rrr.rrr.rrr. The relevant portion of RFC 790 is reproduced here with minor changes: 000.rrr.rrr.rrr Reserved [JBP] 001.rrr.rrr.rrr BBN-PR BBN Packet Radio Network [DCA2] 002.rrr.rrr.rrr SF-PR-1 SF Packet Radio Network [JEM] 003.rrr.rrr.rrr BBN-RCC BBN RCC Network [SGC] 004.rrr.rrr.rrr SATNET Atlantic Satellite Network [DM11] 005.rrr.rrr.rrr SILL-PR Ft. Sill Packet Radio Network[JEM] 006.rrr.rrr.rrr SF-PR-2 SF Packet Radio Network [JEM] 007.rrr.rrr.rrr CHAOS MIT CHAOS Network [MOON] 008.rrr.rrr.rrr CLARKNET SA
https://en.wikipedia.org/wiki/Daffodil%20Polytechnic%20Institute
Daffodil Polytechnic Institute is a private polytechnic Institute located in Dhaka, Bangladesh.The campus is located at Dhanmondi. Daffodil Polytechnic Institute which has been functioning since 2006 to develop professionals in different fields of education and training under Bangladesh Technical Education Board (BTEB). It is the first and only polytechnic institute of the country which has been awarded internationally. Daffodil Polytechnic is one of the top ranking polytechnics in Bangladesh. https://dpi.ac Departments Currently there are eight departments: Civil Department Electrical Department Computer Science Department Textile Department Apparel Manufacturing Department Telecommunication Department Architecture and Interior Design Department Graphic Design History The polytechnic was established in 2006 with the approval of Bangladesh Technical Education Board and the Government of Bangladesh's Ministry of Education. Campuses The institute has multiple campuses within Dhaka. The main campus & the academic building 1 is located in Dhanmondi and the other campus is in Kalabagan with library and hostel facilities for both male and female students. Academics Departments Computer Engineering Technology Electrical Engineering Technology Civil Engineering Technology Architecture & Interior Design Technology Textile Engineering Technology Garments Design & Pattern Making Technology Telecommunication Engineering Technology Graphic Design Engineering Technology Principals Mohammad Nuruzzaman (31 July 2006 - 30 April 2013) K M Hasan Ripon (1 May 2013 - 31 May 2016) Wiz khalifa( 1 June 2016 - 30 April 2019) K M Hasan Ripon (1 May 2019 – Present) Online admission The Polytechnic facilitates online admission for applicants from distant areas. Clubs Kolorob Cultural Club Computer club Language club DPI Alumni Association Blood donating club Tourism club International Activities A gorup of Students from Daffodil Polytech Instit
https://en.wikipedia.org/wiki/Coronoid%20process%20of%20the%20ulna
The coronoid process of the ulna is a triangular process projecting forward from the anterior proximal portion of the ulna. Structure Its base is continuous with the body of the bone, and of considerable strength. Anatomy Its apex is pointed, slightly curved upward, and in flexion of the forearm is received into the coronoid fossa of the humerus. Its upper surface is smooth, convex, and forms the lower part of the semilunar notch. Its antero-inferior surface is concave, and marked by a rough impression for the insertion of the brachialis muscle. At the junction of this surface with the front of the body is a rough eminence, the tuberosity of the ulna, which gives insertion to a part of the brachialis; to the lateral border of this tuberosity the oblique cord is attached. Its lateral surface presents a narrow, oblong, articular depression, the radial notch. Its medial surface, by its prominent, free margin, serves for the attachment of part of the ulnar collateral ligament. At the front part of this surface is a small rounded eminence for the origin of one head of the flexor digitorum superficialis muscle; behind the eminence is a depression for part of the origin of the flexor digitorum profundus muscle; descending from the eminence is a ridge which gives origin to one head of the pronator teres muscle. Frequently, the flexor pollicis longus muscle arises from the lower part of the coronoid process by a rounded bundle of muscular fibers. Function The coronoid process stabilises the elbow joint and prevents hyperflexion. Clinical significance The coronoid process can be fractured from its anteromedial facet. Additional images
https://en.wikipedia.org/wiki/PayScale
Payscale is an American compensation software and data company which helps employers manage employee compensation and employees understand their worth in the job market. History The website was launched on January 1, 2002. It was founded by Joe Giordano and John Gaffney. Mike Metzger served as CEO from 2004 to 2019. Scott Torrey, a 20-year veteran of SAP Concur, started as CEO on August 26, 2019 and stepped down on November 16, 2021. The current CEO of PayScale is Alex Hart. On April 24, 2014, Warburg Pincus acquired Payscale in a deal worth up to $100 million. On April 25, 2019, Francisco Partners announced a majority investment in Payscale at an enterprise value of $325 million. Overview Payscale was developed to help people and businesses obtain accurate, real-time information on job market compensation. While Payscale started by crowdsourcing compensation data from employees to power its products for employers, its Software as a Service offerings have evolved to allow businesses to utilize multiple compensation data sources, including Payscale's Crowdsourced and Company Sourced offerings as well as data from other providers. Customers can also manage their employee compensation strategy and structure within the platform and perform robust compensation analytics. For employees, the service works via the Internet by enabling individuals to submit their job profile and salary data, which is then compared to others like them. They receive a free report on their market worth. The company generates revenue by selling SaaS subscriptions, compensation data and services to employers, to aid in determining correct market rates for hiring, benchmarking and budgeting, and by targeted advertising to employees that visit its website. Payscale surveys its users' income and background, and since 2007, it has published an annual ranking of American colleges and universities by their estimated return on investment. The rankings have been popular with the public but cont
https://en.wikipedia.org/wiki/Anaphylaxis
Anaphylaxis is a serious, potentially fatal allergic reaction and medical emergency that is rapid in onset and requires immediate medical attention regardless of use of emergency medication on site. It typically causes more than one of the following: an itchy rash, throat closing due to swelling that can obstruct or stop breathing; severe tongue swelling that can also interfere with or stop breathing; shortness of breath, vomiting, lightheadedness, loss of consciousness, low blood pressure, and medical shock. These symptoms typically start in minutes to hours and then increase very rapidly to life-threatening levels. Urgent medical treatment is required to prevent serious harm and death, even if the patient has used an epipen or has taken other medications in response, and even if symptoms appear to be improving. Common causes include allergies to insect bites and stings, allergies to foods – including nuts, milk, fish, shellfish, eggs and some fresh fruits or dried fruits; allergies to sulfites – a class of food preservatives and a byproduct in some fermented foods like vinegar; allergies to medications – including some antibiotics and non-steroidal anti-inflammatory drugs (NSAIDs) like aspirin; allergy to general anaesthetic (used to make people sleep during surgery); allergy to contrast agents – dyes used in some medical tests to help certain areas of the body show up better on scans; allergy to latex – a type of rubber found in some rubber gloves and condoms. Other causes can include physical exercise, and cases may also occur in some people due to escalating reactions to simple throat irritation or may also occur without an obvious reason. The mechanism involves the release of inflammatory mediators in a rapidly escalating cascade from certain types of white blood cells triggered by either immunologic or non-immunologic mechanisms. Diagnosis is based on the presenting symptoms and signs after exposure to a potential allergen or irritant and in some cases, reac
https://en.wikipedia.org/wiki/WHEP-TRS%20protein%20domain
In molecular biology, the protein domain WHEP-TRS refers to helix-turn-helix domains. They are found in variable numbers in glutamyl-prolyl tRNA synthetase (EPRS). This protein domain has an important function in protein–protein interactions between synthetases. WHEP domains exhibit high-affinity interactions with tRNA, indicating a putative evolutionary relationship to facilitate tRNA binding to fused synthetases, thereby enhancing catalytic efficiency. Protein interactions EPRS is a component of the interferon-gamma-activated inhibitor of translation (GAIT) complex, which interacts with stem-loop elements (GAIT elements) in mRNAs encoding proinflammatory proteins, for example, vascular endothelial growth factor-A (VEGFA). WHEP domains interact with the GAIT element in the 3′UTR of target mRNAs and with the regulatory protein NS1-associated protein-1 (NSAP1). Structure A conserved domain of 46 amino acids, called WHEP-TRS has been shown to exist in a number of higher eukaryote aminoacyl-transfer RNA synthetases. This domain is present one to six times in several enzymes. There are three copies in mammalian aminoacyl-tRNA synthetase in a region that separates the N-terminal glutamyl-tRNA synthetase domain from the C-terminal prolyl-tRNA synthetase domain, and six copies in the intercatalytic region of the Drosophila enzyme. The domain is found at the N-terminal extremity of the mammalian tryptophanyl- tRNA synthetase and histidyl-tRNA synthetase, and the mammalian, insect, nematode and plant glycyl- tRNA synthetases.
https://en.wikipedia.org/wiki/Diffusion%20map
Diffusion maps is a dimensionality reduction or feature extraction algorithm introduced by Coifman and Lafon which computes a family of embeddings of a data set into Euclidean space (often low-dimensional) whose coordinates can be computed from the eigenvectors and eigenvalues of a diffusion operator on the data. The Euclidean distance between points in the embedded space is equal to the "diffusion distance" between probability distributions centered at those points. Different from linear dimensionality reduction methods such as principal component analysis (PCA), diffusion maps are part of the family of nonlinear dimensionality reduction methods which focus on discovering the underlying manifold that the data has been sampled from. By integrating local similarities at different scales, diffusion maps give a global description of the data-set. Compared with other methods, the diffusion map algorithm is robust to noise perturbation and computationally inexpensive. Definition of diffusion maps Following and, diffusion maps can be defined in four steps. Connectivity Diffusion maps exploit the relationship between heat diffusion and random walk Markov chain. The basic observation is that if we take a random walk on the data, walking to a nearby data-point is more likely than walking to another that is far away. Let be a measure space, where is the data set and represents the distribution of the points on . Based on this, the connectivity between two data points, and , can be defined as the probability of walking from to in one step of the random walk. Usually, this probability is specified in terms of a kernel function of the two points: . For example, the popular Gaussian kernel: More generally, the kernel function has the following properties ( is symmetric) ( is positivity preserving). The kernel constitutes the prior definition of the local geometry of the data-set. Since a given kernel will capture a specific feature of the data set, its c
https://en.wikipedia.org/wiki/Laser%20rapid%20manufacturing
Laser Rapid Manufacturing (LRM) is one of the advanced additive manufacturing processes that is capable of fabricating engineering components directly from a solid model. Technique In this technique, a solid model of the component to be fabricated is made either by 3D imaging system or by designer using computer-aided design (CAD) software or by math data as an output of numerical analysis. Thus obtained model is sliced into thin layers along the vertical axis. The thin layers are converted into corresponding numerical controlled (NC) code and are sent to LRM station in suitable format (e.g. G&M code). LRM station employs a laser beam as a heat source to melt a thin layer on the surface of the substrate/deposited material and fed material to deposit a new layer as per shape and dimensions defined in NC code. A number of such layers deposited one over another and it results in three-dimensional (3D) components directly from the solid model. Benefits LRM eliminates many manufacturing steps such as materials-machine planning, man-machine interaction, intermittent quality checks, assembly and related human errors etc. Therefore, LRM offers many advantages over conventional subtractive techniques, such as reduced production time, better process control and capability to form functionally graded parts. It is also an attractive candidate for refurbishing applications because of low heat input, limited dilution with minimal distortion and capability of adding finer near-net shaped features to the components. Similar techniques Manufacturing techniques, similar to LRM, are being developed with different names at various laboratories, such as Laser Engineered Net Shaping (LENSTM) at Sandia National Laboratories (USA), Freeform Laser Consolidation at National Research Council (Canada), Selective Laser Powder Remelting (SLPR) at Fraunhofer Society (Germany), Selective Laser Cladding (SLC) at the University of Liverpool (UK), Shape deposition Manufacturing (SDM) at Stanford U
https://en.wikipedia.org/wiki/Rainbow%20table
A rainbow table is a precomputed table for caching the outputs of a cryptographic hash function, usually for cracking password hashes. Passwords are typically stored not in plain text form, but as hash values. If such a database of hashed passwords falls into the hands of an attacker, they can use a precomputed rainbow table to recover the plaintext passwords. A common defense against this attack is to compute the hashes using a key derivation function that adds a "salt" to each password before hashing it, with different passwords receiving different salts, which are stored in plain text along with the hash. Rainbow tables are a practical example of a space–time tradeoff: they use less computer processing time and more storage than a brute-force attack which calculates a hash on every attempt, but more processing time and less storage than a simple table that stores the hash of every possible password. Rainbow tables were invented by Philippe Oechslin as an application of an earlier, simpler algorithm by Martin Hellman. Background For user authentication, passwords are stored either as plaintext or hashes. Since passwords stored as plaintext are easily stolen if database access is compromised, databases typically store hashes instead. Thus, no one – including the authentication system – can learn a password merely by looking at the value stored in the database. When a user enters a password for authentication, a hash is computed for it and then compared to the stored hash for that user. Authentication fails if the two hashes do not match; moreover, authentication would equally fail if a hashed value were entered as a password, since the authentication system would hash it a second time. To learn a password from a hash is to find a string which, when input into the hash function, creates that same hash. This is the same as inverting the hash function. Though brute-force attacks (e.g. dictionary attacks) may be used to try to invert a hash function, they can bec
https://en.wikipedia.org/wiki/Fiber-optic%20display
A fiber-optic display is a light-emitting display that uses fiber optics to display images or text. Fiber-optic displays can either be static or dynamic, with the typical lighting source being halogen light bulbs. Static fiber optic displays Static fiber optic displays have been commonly used for some types of traffic signals. One common use for static fiber optic displays are lane control lights, which display either a green downward-pointing arrow or a red X to indicate the open/closed status of road lanes. Dynamic fiber optic displays Dynamic fiber optic displays typically display alphanumeric text, and utilize electromechanical shutters to open or close the ends of the fiber strands to display an alphanumeric pixel. These type of displays were commonly used as variable-message signs on highways. Compared to eggcrate displays, dynamic fiber optic displays offered lower energy consumption due to requiring fewer bulbs, and offered improved nighttime legibility. For daytime legibility, they were sometimes combined with flip-disc displays to be reflective in daylight and emissive at night. Display technology
https://en.wikipedia.org/wiki/Context%20mixing
Context mixing is a type of data compression algorithm in which the next-symbol predictions of two or more statistical models are combined to yield a prediction that is often more accurate than any of the individual predictions. For example, one simple method (not necessarily the best) is to average the probabilities assigned by each model. The random forest is another method: it outputs the prediction that is the mode of the predictions output by individual models. Combining models is an active area of research in machine learning. The PAQ series of data compression programs use context mixing to assign probabilities to individual bits of the input. Application to Data Compression Suppose that we are given two conditional probabilities, and , and we wish to estimate , the probability of event X given both conditions and . There is insufficient information for probability theory to give a result. In fact, it is possible to construct scenarios in which the result could be anything at all. But intuitively, we would expect the result to be some kind of average of the two. The problem is important for data compression. In this application, and are contexts, is the event that the next bit or symbol of the data to be compressed has a particular value, and and are the probability estimates by two independent models. The compression ratio depends on how closely the estimated probability approaches the true but unknown probability of event . It is often the case that contexts and have occurred often enough to accurately estimate and by counting occurrences of in each context, but the two contexts either have not occurred together frequently, or there are insufficient computing resources (time and memory) to collect statistics for the combined case. For example, suppose that we are compressing a text file. We wish to predict whether the next character will be a linefeed, given that the previous character was a period (context ) and that the last linefeed o