source
stringlengths
31
203
text
stringlengths
28
2k
https://en.wikipedia.org/wiki/Branch%20Queue
In Computer Architecture, While Branch predictions Branch queue takes place. When Branch Predictor predicts if the branch is taken or not, Branch queue stores the predictions that to be used later. Branch queue consists 2 values only. Taken or Not Taken. Branch queue helps other algorithms to increase parallelism and optimization. It is not software implemented or Hardware one, It falls under hardware software co-design. References Computer architecture fa:صف پرش
https://en.wikipedia.org/wiki/Ford%20Instrument%20Company
The Ford Instrument Company was a U.S. corporation known for being the primary supplier of fire control Rangekeepers and analog computers for the United States Navy before and during World War II. A personal blog, Doug Coward's Analog History Museum, includes a page with details for the Ford Instrument Company Computer Mark I that was used after 1939 on WW II naval guns up to 5 inch and anti-aircraft guns. This page has a background stating that the Ford Instrument Company is a subsidiary of Sperry Rand, indicating that the displayed page was supplied by Sperry while operating as Sperry Rand, 1955 and 1978. References Defunct computer companies of the United States Military computers
https://en.wikipedia.org/wiki/ISCB%20Fellow
ISCB Fellowship is an award granted to scientists that the International Society for Computational Biology (ISCB) judges to have made “outstanding contributions to the fields of computational biology and bioinformatics”. , there are 76 Fellows of the ISCB including Michael Ashburner, Alex Bateman, Bonnie Berger, Steven E. Brenner, Janet Kelso, Daphne Koller, Michael Levitt, Sarah Teichmann and Shoshana Wodak. See List of Fellows of the International Society for Computational Biology for a comprehensive listing. Fellows of the International Society for Computational Biology The first seven fellows of the ISCB were laureates of the ISCB Senior Scientist Award from 2003 to 2009: Webb Miller David Haussler Temple F. Smith Michael Waterman Janet Thornton David J. Lipman David Sankoff Since 2009, new fellows have been nominated from the community of ISCB members and voted on annually by a selection committee. New fellows are traditionally inaugurated at the annual Intelligent Systems for Molecular Biology (ISMB) conference. References Bioinformatics Computational biology Academic awards
https://en.wikipedia.org/wiki/Korean%20Mathematical%20Olympiad
The Korean Mathematical Olympiad is a mathematical olympiad held by the Korean Mathematical Society (KMS) in Republic of Korea. History In 1988, only high school students were tested and middle school students were supposed to take the 'high school exam'. From the 11th exam, the middle school students' test was introduced (JKMO, Junior Korean Mathematical Olympiad). In the 53rd International Mathematical Olympiad (IMO), the Korean delegation won 209 out of 252 out of the total score and 6 gold medals and ranked the first place for the first time in history. From then, Korean mathematicians have made outstanding achievements in advanced math research and International Mathematical Olympiad. Business Background In order to obtain excellent grades in the International Mathematics Olympiad, the Korean Mathematical Society holds the Korean Mathematical Olympiad, and through the operation of the seasonal school, KMS will discover gifted students and educate them to contribute the development of mathematics, science, and engineering in Korea. Students who have won higher than a bronze prize in a High School Level Secondary Test or those who have completed a Middle School Level Winter School are entitled to take a Final Test. This consists of 6 questions in a narrative form. It is taken for two days at the end of March, with students required to solve 3 problems in 4 hours and 30 minutes each day. Awards are divided into Grand Prize, Excellence Prize, and Encouragement Prize, and 13 people are selected as representative candidates. Six students are selected from the candidates to represent Korea in the IMO. Gauss Part and Euler Part Since 2017, KMO has been divided into two parts: Gauss Part and Euler Part. The Gauss Part is open to all students under the age of 20 and the Euler Part is available for high school students other than science high school students. Final representative selection process About 12 to 13 students (twice the number of final candidates) are
https://en.wikipedia.org/wiki/Octopussy%20%28software%29
Octopussy, also known as 8Pussy, is a free and open-source computer-software which monitors systems, by constantly analyzing the syslog data they generate and transmit to such a central Octopussy server (thus often called a SIEM solution). Therefore, software like Octopussy plays an important role in maintaining an information security management system within ISO/IEC 27001-compliant environments. Octopussy has the ability to monitor any device that supports the syslog protocol, such as servers, routers, switches, firewalls, load balancers, and its important applications and services. The main purpose of the software is to alert its administrators and users to different kinds of events, like system outages, attacks on systems or errors in applications. However, unlike Nagios or Icinga, Octopussy is not a state-checker and therefore problems cannot be resolved within the application. The software also makes no prescription whatsoever on which messages must be/must not be analyzed. As such, Octopussy can be seen as less powerful than other popular commercial software in the same category (event monitoring and log analysis). Octopussy is compatible with many Linux system distributions like Debian, Ubuntu, OpenSUSE, CentOS, RHEL and even meta-distributions as Gentoo or Arch Linux. Although Octopussy was originally designed to run on Linux, it could be ported to other Unix variants like FreeBSD with minimal effort. Octopussy has extensive report generating features and also various interfaces to other software, like e.g. NSCA (Nagios), Jabber/XMPP and Zabbix. With the help of software like Snare even Windows EventLogs can be processed. Octopussy is licensed under the terms of the GNU General Public License. Characteristics Although Octopussy is free and open-source software it has a variety of characteristics also found in some professional enterprise applications like Splunk, SAWMILL or Kiwi Syslog. Octopussy features At the time of writing, Octopussy comes wit
https://en.wikipedia.org/wiki/Self-verifying%20finite%20automaton
In automata theory, a self-verifying finite automaton (SVFA) is a special kind of a nondeterministic finite automaton (NFA) with a symmetric kind of nondeterminism introduced by Hromkovič and Schnitger. Generally, in self-verifying nondeterminism, each computation path is concluded with any of the three possible answers: yes, no, and I do not know. For each input string, no two paths may give contradictory answers, namely both answers yes and no on the same input are not possible. At least one path must give answer yes or no, and if it is yes then the string is considered accepted. SVFA accept the same class of languages as deterministic finite automata (DFA) and NFA but have different state complexity. Formal definition An SVFA is represented formally by a 6-tuple, A=(Q, Σ, Δ, q0, Fa, Fr) such that (Q, Σ, Δ, q0, Fa) is an NFA, and Fa, Fr are disjoint subsets of Q. For each word w = a1a2 … an, a computation is a sequence of states r0,r1, …, rn, in Q with the following conditions: r0 = q0 ri+1 ∈ Δ(ri, ai+1), for i = 0, …, n−1. If rn ∈ Fa then the computation is accepting, and if rn ∈ Fr then the computation is rejecting. There is a requirement that for each w there is at least one accepting computation or at least one rejecting computation but not both. Results Each DFA is a SVFA, but not vice versa. Jirásková and Pighizzini proved that for every SVFA of n states, there exists an equivalent DFA of states. Furthermore, for each positive integer n, there exists an n-state SVFA such that the minimal equivalent DFA has exactly states. Other results on the state complexity of SVFA were obtained by Jirásková and her colleagues. References Finite automata
https://en.wikipedia.org/wiki/OpenIO
OpenIO offered object storage for a wide range of high-performance applications. OpenIO was founded in 2015 by Laurent Denel (CEO), Jean-François Smigielski (CTO) and five other co-founders; it leveraged open source software, developed since 2006, based on a grid technology that enabled dynamic behaviour and supported heterogenous hardware. In October 2017 OpenIO completed a $5 million funding round. In July 2020 OpenIO had been acquired by OVH and withdrawn from the market to become the core technology of OVHcloud object storage offering. Software OpenIO is a software-defined object store that supports S3 and can be deployed on-premises, cloud-hosted or at the edge, on any hardware mix. It has been designed from the beginning for performance and cost-efficiency at any scale, and it has been optimized for Big Data, HPC and AI. OpenIO stores objects within a flat structure within a massively distributed directory with indirections, which allows the data query path to be independent of the number of nodes and the performance not to be affected by the growth of capacity. Servers are organized as a grid of nodes massively distributed, where each node takes part in directory and storage services, which ensures that there is no single point of failure and that new nodes are automatically discovered and immediately available without the need to rebalance data. The software is built on top of a technology that ensures optimal data placement based on real-time metrics and allows the addition or removal of storage devices with automatic performance and load impact optimization. For data protection OpenIO has synchronous and asynchronous replication with multiple copies, and an erasure coding implementation based on Reed-Solomon that can be deployed in one data center or geo-distributed or stretched clusters. The software has a feature that catches all events that occur in the cluster and can pass them up in the stack or to applications running on OpenIO nodes. This enable
https://en.wikipedia.org/wiki/ISO/IEC%2090003
ISO/IEC 90003 Software engineering -- Guidelines for the application of ISO 9001:2008 to computer software is a guidelines developed for organizations in the application of ISO 9001 to the acquisition, supply, development, operation and maintenance of computer software and related support services. This standard was developed by technical committee ISO/IEC JTC 1/SC 7 Software and systems engineering. ISO/IEC 90003 originally published as ISO 9000-3 for the first time in December 1997, was issued for the first time as an ISO/IEC 90003 in February 2004. The review cycle of ISO 90003 is every 5 years. Main requirements of the standard The ISO/IEC 90003:2014 adopts the ISO structure in 8 chapters in the following breakdown: Scope Normative references Terms and definitions Quality management system Management responsibility Resource management Product realization Measurement, analysis and improvement See also List of ISO standards List of IEC standards International Organization for Standardization References External links ISO/IEC 90003—Software engineering -- Guidelines for the application of ISO 9001 to computer software ISO/IEC JTC 1/SC 7—Software and systems engineering 90003 Software quality
https://en.wikipedia.org/wiki/SKEW
SKEW is the ticker symbol for the CBOE Skew Index, a measure of the perceived tail risk of the distribution of S&P 500 investment returns over a 30-day horizon. The index values are calculated and published by the Chicago Board Options Exchange (CBOE) based on current S&P 500 options market data. SKEW is similar to the VIX index, but instead of measuring implied volatility based on a normal distribution, it measures an implied risk of future returns realizing outlier behavior. The index model defines such an outlier as two or more standard deviations below the mean, which would characterize a black swan event or market crash. The index value typically reflects trading activity of portfolio managers hedging tail risk with options, to protect portfolios from a large, sudden decline in the market. A SKEW value of 100 indicates the options market perceives a low risk of outlier returns; values increasing above 100 reflect an increased perception of risk for future outlier event(s). See also Skewness risk Taleb distribution Volatility skew References External links Google Finance page for current SKEW values, and historical charts American stock market indices S&P Global Options (finance) Derivatives (finance) Mathematical finance Technical analysis
https://en.wikipedia.org/wiki/Hazard%20substitution
Hazard substitution is a hazard control strategy in which a material or process is replaced with another that is less hazardous. Substitution is the second most effective of the five members of the hierarchy of hazard controls in protecting workers, after elimination. Substitution and elimination are most effective early in the design process, when they may be inexpensive and simple to implement, while for an existing process they may require major changes in equipment and procedures. The concept of prevention through design emphasizes integrating the more effective control methods such as elimination and substitution early in the design phase. Hazard substitutions can involve not only changing one chemical for another, but also using the same chemical in a less hazardous form. Substitutions can also be made to processes and equipment. In making a substitution, the hazards of the new material should be considered and monitored, so that a new hazard is not unwittingly introduced, causing "regrettable substitutions". Substitution can also fail as a strategy if the hazardous process or material is reintroduced at a later stage in the design or production phases, or if cost or quality concerns cause a substitution to not be adopted. Examples Chemicals A common substitution is to replace a toxic chemical with a less toxic one. Some examples include replacing the solvent benzene, a carcinogen, with toluene; switching from organic solvents to water-based detergents; and replacing paints containing lead with those containing non-leaded pigments. Dry cleaning can avoid the use of toxic perchloroethylene by using petroleum-based solvents, supercritical carbon dioxide, or wet cleaning techniques. Chemical substitutions are an example of green chemistry. Chemicals can also be substituted with a different form of the same chemical. In general, inhalation exposure to dusty powders can be reduced by using a slurry or suspension of particles in a liquid solvent inste
https://en.wikipedia.org/wiki/Bioclogging
Bioclogging or biological clogging is the clogging of pore space in soil by microbial biomass; their body and their byproducts such as extracellular polymeric substance (EPS). The microbial biomass blocks the pathway of water in the pore space, forming a certain thickness of the impermeable layer in the soil, and it reduces the rate of infiltration of water remarkably. Bioclogging is observed under continuous ponded infiltration at various field conditions such as artificial recharge ponds, percolation trenches, irrigation channels, sewage treatment systems, constructed wetlands, and landfill liners. It also affects groundwater flow in the aquifer, such as ground source heat pumps, permeable reactive barriers, and microbial enhanced oil recovery. In the situation where infiltration of water at an appropriate rate is needed, bioclogging can be problematic and countermeasures such as regular drying of the system are taken. In some cases, bioclogging can be utilized to make an impermeable layer to minimize the rate of infiltration. General description Change in permeability with time Bioclogging is observed as the decrease in the infiltration rate. A decrease in the infiltration rate under ponded infiltration was observed in the 1940s for studying the infiltration of artificial recharge ponds and the water-spreading on agricultural soils. When soils are continuously submerged, permeability or saturated hydraulic conductivity changes in 3 stages which was explained as follows. Permeability decreases for 10 to 20 days possibly due to physical changes in the structure of the soil. Permeability increases due to dissolving the entrapped air in soil into the percolating water. Permeability decreases for 2 to 4 weeks due to the disintegration of aggregates and biological clogging of soil pores with microbial cells and their synthesized products, slimes, or polysaccharides. The 3 stages are not necessarily distinct in every field condition of bioclogging; when the seco
https://en.wikipedia.org/wiki/University%20Professor%20of%20Natural%20Philosophy%20%28Dublin%29
The University Chair of Natural Philosophy is a professorship in the School of Mathematics at Trinity College Dublin. It was established in 1847. From 1724 to 1847 the Erasmus Smith's Professorship of Natural and Experimental Philosophy had a mathematical and theoretical orientation, with many holders being also mathematicians. Several, such as Bartholomew Lloyd (1822) and James MacCullagh (1843), previously held the Erasmus Smith's Professor of Mathematics position. In 1847 the University Chair of Natural Philosophy was founded and took on the applied mathematics and theoretical physics role, while Erasmus Smith's Professor of Natural and Experimental Philosophy (1724) effectively became the chair of experimental physics. List of the professors 1847–1870: John Jellett (1817–1888) 1870–1884: Richard Townsend (1821–1884) 1884–1890: Benjamin Williamson (1827–1916) 1890–1902: Francis Tarleton (1841–1920) 1902–1910: Frederick Purser (1839–1910) 1910–1925: Matthew Fry (1863–1943) 1925–1930: J. L. Synge (1897–1995) 1930–1957: Albert McConnell (1903–1993) 1957–1962: vacant 1962–1966: John Chisholm (born 1926) 1966–1997: David Spearman (born 1937) 1997–2002: vacant 2002–present: Samson Shatashvili See also List of professorships at the University of Dublin Natural philosophy References 1847 establishments in Ireland Natural Philosophy, University Natural Philosophy, University, Dublin Natural Philosophy, University, Dublin
https://en.wikipedia.org/wiki/Vegetable%20Production%20System
The Vegetable Production System (Veggie) is a plant growth system developed and used by NASA in outer space environments. The purpose of Veggie is to provide a self-sufficient and sustainable food source for astronauts as well as a means of recreation and relaxation through therapeutic gardening. Veggie was designed in conjunction with ORBITEC and is currently being used aboard the International Space Station, with another Veggie module planned to be delivered to the ISS in 2017. Overview Veggie is part of an overarching project concerning research on growing crops in zero gravity. Among the goals of this project are to learn about how plants grow in a weightless environment and to learn about how plants can efficiently be grown for crew use in space. Veggie was designed to be low maintenance, using low power and having a low launch mass. Thus, Veggie provides a minorly regulated environment with minimal control over the atmosphere and temperature of the module. The successor to the Veggie project is the Advanced Plant Habitat (APH), components of which will be delivered to the International Space Station during the Cygnus CRS OA-7 and SpaceX CRS-11 missions in 2017. In 2018 the Veggie-3 experiment was tested with plant pillows and root mats. One of the goals is to grow food for crew consumption. Crops tested at this time include cabbage, lettuce, and mizuna. Design A Veggie module weighs less than and uses 90 watts. It consists of three parts: a lighting system, a bellows enclosure, and a reservoir. The lighting system regulates the amount and intensity of light plants receive, the bellows enclosure keeps the environment inside the unit separate from its surroundings, and the reservoir connects to plant pillows where the seeds grow. Lighting system Veggie's lighting system consists of three different types of coloreds LEDs: red, blue, and green. Each color corresponds to a different light intensity that the plants will receive. Although the lighting syst
https://en.wikipedia.org/wiki/County%20Herb%20Committee
The County Herb Committees were a nationwide medicinal plant collecting scheme, established by the British Ministry of Health during the Second World War. History The County Herb Committees were set up at a time when the German occupation of France and the disruption of shipping lanes interfered with drug supplies. As in the First World War, the British found that the Germans still largely dominated the pharmaceutical industry and consequently by the early 1940s there were critical shortages of essential medicines in hospitals and homes across Britain. First the Vegetable Drugs Committee (VDC) of the Ministry of Supply was established in March, 1941, and the involvement of the Royal Botanic Gardens, Kew, was led by Dr. Ronald Melville, an economic botanist and pharmacist. A total of 70 Committees were set up across England, Scotland and Wales, and grants of a total of £1,191 were made available to set up drying centres that could deal directly with trading companies. The first drying centre in Oxfordshire was in the home of Dr. W. O. James and his wife Gladys in the village of Islip, as well as in the Oxford Botanic Garden where they set up the Oxford Medicinal Plants Scheme. By the end of the war, there were 250 drying centres across country. Guides for herb collectors The committee found that a number of imported drugs were derived from plants that were also native to Britain. By 1941 they were publishing guides for herb collectors in the rural British communities. Various groups such as Boy Scouts, Girl Guides, Women's Institutes and the elderly, were enlisted as collectors by Sir Arthur William Hill, Director of Kew. In 1941 the wholesale pharmaceutical company Brome and Schimmer published a booklet called Herb Gathering, describing the many roots, flowers and herbs needed by the Ministry of Health, and how to collect and dry them. The National Federation of Women's Institutes cooperated with the Ministry of Supply and the County Herb Committees by colle
https://en.wikipedia.org/wiki/State%20complexity
State complexity is an area of theoretical computer science dealing with the size of abstract automata, such as different kinds of finite automata. The classical result in the area is that simulating an -state nondeterministic finite automaton by a deterministic finite automaton requires exactly states in the worst case. Transformation between variants of finite automata Finite automata can be deterministic and nondeterministic, one-way (DFA, NFA) and two-way (2DFA, 2NFA). Other related classes are unambiguous (UFA), self-verifying (SVFA) and alternating (AFA) finite automata. These automata can also be two-way (2UFA, 2SVFA, 2AFA). All these machines can accept exactly the regular languages. However, the size of different types of automata necessary to accept the same language (measured in the number of their states) may be different. For any two types of finite automata, the state complexity tradeoff between them is an integer function where is the least number of states in automata of the second type sufficient to recognize every language recognized by an -state automaton of the first type. The following results are known. NFA to DFA: states. This is the subset construction by Rabin and Scott, proved optimal by Lupanov. UFA to DFA: states, see Leung, An earlier lower bound by Schmidt was smaller. NFA to UFA: states, see Leung. There was an earlier smaller lower bound by Schmidt. SVFA to DFA: states, see Jirásková and Pighizzini 2DFA to DFA: states, see Kapoutsis. Earlier construction by Shepherdson used more states, and an earlier lower bound by Moore was smaller. 2DFA to NFA: , see Kapoutsis. Earlier construction by Birget used more states. 2NFA to NFA: , see Kapoutsis. 2NFA to NFA accepting the complement: states, see Vardi. AFA to DFA: states, see Chandra, Kozen and Stockmeyer. AFA to NFA: states, see Fellah, Jürgensen and Yu. 2AFA to DFA: , see Ladner, Lipton and Stockmeyer. 2AFA to NFA: , see Geffert and Okhotin. The 2DFA vs. 2NFA
https://en.wikipedia.org/wiki/Nature%20Energy
Nature Energy is a monthly peer-reviewed scientific journal published by Nature Portfolio. The editor-in-chief is Nicky Dean. The 2017 efficiency record (26.6%) in solar cell technology was published in the journal. According to the Journal Citation Reports, the journal has a 2021 impact factor of 67.439, ranking it 1st out of 119 journals in the category "Energy & Fuels" and 2nd out of 345 journals in the category "Materials Science, Multidisciplinary". References External links Nature Research academic journals English-language journals Monthly journals Academic journals established in 2016
https://en.wikipedia.org/wiki/Tasmanian%20coniferous%20shrubbery
The vegetation in Tasmania's alpine environments is predominately woody and shrub-like. One vegetation type is coniferous shrubbery, characterised by the gymnosperm species Microcachrys tetragona, Pherosphaera hookeriana, Podocarpus lawrencei, and Diselma archeri. Distribution of these species is relevant with abiotic factors including edaphic conditions and fire frequency, and increasingly, the threat of climate change towards species survival exists. Conservation and management of coniferous shrubbery are necessary considering that the paleoendemic species, Microcachrys, Pherosphaera and Diselma, have persisted in western Tasmanian environments for millions of years. Distribution These coniferous shrub species are restricted to subalpine and alpine heathlands in western Tasmania, with the exception of Podocarpus lawrencei which lives on the mainland. The alpine environments where these conifers occur have high levels of conifer endemism, which is an ecologically habitat for coniferous shrub species. Coniferous shrub species can be observed in Mount Field National Park in Tasmania's south west along the Tarn Shelf. All species can be observed in rocky environments with shallow soil above . Ecology Both the alpine environment and the harsh maritime climate have the pressures and limitations of wind exposure and ice abrasion for the woody and shrub-like habit of coniferous shrubbery. The lack of protective snow cover on Tasmanian mountains means that vegetation must be mechanically resistant to these elements, hence an ecologically habitat for coniferous shrub species. This is contrasted to alps of mainland Australia or New Zealand, where the presence of prolonged snow lie lead to the development of a grassland-herbland vegetation community. Low productivity of the environment is indicated through the slow growth habit of the conifers, and the effects of fire are detrimental to the species. As well as this, physiological drought intolerance in conifers could in
https://en.wikipedia.org/wiki/Single-cell%20transcriptomics
Single-cell transcriptomics examines the gene expression level of individual cells in a given population by simultaneously measuring the RNA concentration (conventionally only messenger RNA (mRNA)) of hundreds to thousands of genes. Single-cell transcriptomics makes it possible to unravel heterogeneous cell populations, reconstruct cellular developmental pathways, and model transcriptional dynamics — all previously masked in bulk RNA sequencing. Background The development of high-throughput RNA sequencing (RNA-seq) and microarrays has made gene expression analysis a routine. RNA analysis was previously limited to tracing individual transcripts by Northern blots or quantitative PCR. Higher throughput and speed allow researchers to frequently characterize the expression profiles of populations of thousands of cells. The data from bulk assays has led to identifying genes differentially expressed in distinct cell populations, and biomarker discovery. These studies are limited as they provide measurements for whole tissues and, as a result, show an average expression profile for all the constituent cells. This has a couple of drawbacks. Firstly, different cell types within the same tissue can have distinct roles in multicellular organisms. They often form subpopulations with unique transcriptional profiles. Correlations in the gene expression of the subpopulations can often be missed due to the lack of subpopulation identification. Secondly, bulk assays fail to recognize whether a change in the expression profile is due to a change in regulation or composition — for example if one cell type arises to dominate the population. Lastly, when your goal is to study cellular progression through differentiation, average expression profiles can only order cells by time rather than by developmental stage. Consequently, they cannot show trends in gene expression levels specific to certain stages. Recent advances in biotechnology allow the measurement of gene expression in hund
https://en.wikipedia.org/wiki/Photopolymerization-based%20signal%20amplification
Photopolymerization-based signal amplification (PBA) is a method of amplifying detection signals from molecular recognition events in an immunoassay by utilizing a radical polymerization initiated through illumination by light. To contrast between a negative and a positive result, PBA is linked to a colorimetric method, thereby resulting in a change in color when a targeted analyte is detected, i.e., a positive signal. PBA is also used to quantify the concentration of the analyte by measuring intensity of the color. Method PBA is achieved by sequentially adding three kinds of solutions to a test strip and illuminating it with green light. First, a droplet of a patient’s sample is loaded on a test strip whose surface is covered with immobilized antibodies. If the sample contains the target antigens, they bind to the immobilized antibodies. (Figure 1a) Second, eosin-conjugated antibodies are added to the patient’s sample. This second antibody specifically binds with the bound antigens, thereby causing each bound antigen to be sandwiched between the first and eosin-conjugated antibodies. (Figure 1b) After ten minutes, the droplet on the surface is rinsed away in order to make sure that only the sandwiched binding complexes are left on the surface before adding the third solution. Lastly, a droplet of mixture of monomers (e.g., PEGDA and N-vinyl pyrrolidone) and phenolphthalein is added to the test strip, and the droplet is illuminated with green visible light, by which the eosin molecules become excited and produce radicals. (Figure 1c) As a result, propagation is caused and polymers are formed. Since phenolphthalein molecules are surrounded by the polymers and thus left on the surface even after another rinse, the test strip turns red when a base is added. (Figure 1d) On the other hand, if the patient’s sample does not include any targeted antigens, the sandwiched binding complexes on the surface will not be formed, which leads to no red color. Principle Regen
https://en.wikipedia.org/wiki/Parasitic%20chromosome
Parasitic chromosomes are "selfish" chromosomes that propagate throughout cell divisions, even if they confer no benefit to the overall organism's survival. Parasitic chromosomes can persist even if slightly detrimental to survival, as is characteristic of some selfish genetic elements. Parasitic chromosomes are often B chromosomes, such that they are not necessarily present in the majority of the species population and are not needed for basic life functions, in contrast to A chromosomes. Parasitic chromosomes are classified as selfish genetic elements. Parasitic chromosomes, if detrimental to an organism's survival, often are selected against by natural selection over time, but if the chromosome is able to act like a selfish DNA element, it can spread throughout a population. An example of a parasitic chromosome is the b24 chromosome in grasshoppers. References Genetics
https://en.wikipedia.org/wiki/Post-tech
Post-tech, (or post-technology, post-digital-technology) is type of technology that is more concerned about being human than about technology. It advocates a design that is not merely focused on efficiency and exploiting users by increasing their time spent with digital devices and technology itself but to support the user's focus and intent, well-being, and independence (from technology). With this interstitial spaces could also be created, similar to what Michel Foucault describes as Heterotopia (space). See also Human-centered computing (discipline) Human-computer interaction Attention economy References Human–computer interaction Ubiquitous computing Postmodernism Product design
https://en.wikipedia.org/wiki/Quantum%20coin%20flipping
Consider two remote players, connected by a channel, that don't trust each other. The problem of them agreeing on a random bit by exchanging messages over this channel, without relying on any trusted third party, is called the coin flipping problem in cryptography. Quantum coin flipping uses the principles of quantum mechanics to encrypt messages for secure communication. It is a cryptographic primitive which can be used to construct more complex and useful cryptographic protocols, e.g. Quantum Byzantine agreement. Unlike other types of quantum cryptography (in particular, quantum key distribution), quantum coin flipping is a protocol used between two users who do not trust each other. Consequently, both users (or players) want to win the coin toss and will attempt to cheat in various ways. It is known that if the communication between the players is over a classical channel, i.e. a channel over which quantum information cannot be communicated, then one player can (in principle) always cheat regardless of which protocol is used. We say in principle because it might be that cheating requires an unfeasible amount of computational resource. Under standard computational assumptions, coin flipping can be achieved with classical communication. The most basic figure of merit for a coin-flipping protocol is given by its bias, a number between and . The bias of a protocol captures the success probability of an all-powerful cheating player who uses the best conceivable strategy. A protocol with bias means that no player can cheat. A protocol with bias means that at least one player can always succeed at cheating. Obviously, the smaller the bias better the protocol. When the communication is over a quantum channel, it has been shown that even the best conceivable protocol can not have a bias less than . Consider the case where each player knows the preferred bit of the other. A coin flipping problem which makes this additional assumption constitutes the weaker variant
https://en.wikipedia.org/wiki/Outline%20of%20machine%20learning
The following outline is provided as an overview of and topical guide to machine learning: Machine learning – subfield of soft computing within computer science that evolved from the study of pattern recognition and computational learning theory in artificial intelligence. In 1959, Arthur Samuel defined machine learning as a "field of study that gives computers the ability to learn without being explicitly programmed". Machine learning explores the study and construction of algorithms that can learn from and make predictions on data. Such algorithms operate by building a model from an example training set of input observations in order to make data-driven predictions or decisions expressed as outputs, rather than following strictly static program instructions. What type of thing is machine learning? An academic discipline A branch of science An applied science A subfield of computer science A branch of artificial intelligence A subfield of soft computing Application of statistics Branches of machine learning Subfields of machine learning Computational learning theory – studying the design and analysis of machine learning algorithms. Grammar induction Meta-learning Cross-disciplinary fields involving machine learning Adversarial machine learning Predictive analytics Quantum machine learning Robot learning Developmental robotics Applications of machine learning Applications of machine learning Bioinformatics Biomedical informatics Computer vision Customer relationship management – Data mining Earth sciences Email filtering Inverted pendulum – balance and equilibrium system. Natural language processing (NLP) Named Entity Recognition Automatic summarization Automatic taxonomy construction Dialog system Grammar checker Language recognition Handwriting recognition Optical character recognition Speech recognition Text to Speech Synthesis (TTS) Speech Emotion Recognition (SER) Machine translation Question answering Speech
https://en.wikipedia.org/wiki/7825
7825 (seven thousand, eight hundred [and] twenty-five) is the natural number following 7824 and preceding 7826. In mathematics 7825 is the smallest number n when it is impossible to assign two colors to natural numbers 1 through n such that every Pythagorean triple is multicolored, i.e. where the Boolean Pythagorean triples problem becomes false. The 200-terabyte proof to verify this is the largest ever made. 7825 is a magic constant of n × n normal magic square and n-Queens Problem for n = 25. References Integers
https://en.wikipedia.org/wiki/LiquidSky
LiquidSky was a New York Citybased provider of cloud visualization, acquired by Walmart in 2018. The company's flagship product was a cloud gaming service of the same name, launched on March 24, 2017, and shut down in 2018. Announced at Consumer Electronics Show 2017, the service aimed to tackle issues that other providers struggle with, including latency and input lag. Its major competitors include Sony's PlayStation Now and NVIDIA's GeForce Now for PC and Mac. History The company was founded in 2014 by Ian McLoughlin and Scott Johnston as LiquidSky Software Inc. Prior to founding LiquidSky, McLoughlin had expressed displeasure of how existing cloud gaming solutions handled latency and input lag, particularly OnLive. Ian stated that his goal when founding LiquidSky was to offer a service with minimal latency and input lag, thus offering a cleaner experience for gamers. Testing began sometime around late 2014, with the company quietly bringing in waves of testers to try out the service. The company found success, raising almost $12 million in funding for the service. At the 2017 Consumer Electronics Show, after roughly two years of testing, McLoughlin and the company announced that the service would be launching in Q1 2017, with an ad-supported plan, upgraded datacenter hardware, and a redesigned client. Other shown off included concepts for standalone hardware and demo footage of Battlefield 1 running at 60 FPS. Originally slated for release on March 14, 2017, it was delayed to March 24, 2017, due to server issues and launched with support for Microsoft Windows. On July 11, 2017, LiquidSky updated their Android client to reflect the launch of the service. The company stated that support for macOS was in development. LiquidSky shut down their service on December 17, 2018, while they focus their efforts on building a new streaming platform. As of 2018, LiquidSky was acquired by Walmart and now focusing on other projects. Features Among the features stated fo
https://en.wikipedia.org/wiki/The%20Computer%20Language%20Benchmarks%20Game
The Computer Language Benchmarks Game (formerly called The Great Computer Language Shootout) is a free software project for comparing how a given subset of simple algorithms can be implemented in various popular programming languages. The project consists of: A set of very simple algorithmic problems Various implementations to the above problems in various programming languages A set of unit tests to verify that the submitted implementations solve the problem statement A framework for running and timing the implementations A website to facilitate the interactive comparison of the results Supported languages Due to resource constraints, only a small subset of common programming languages are supported, up to the discretion of the game's operator. Metrics The following aspects of each given implementation are measured: overall user runtime peak memory allocation gzip'ped size of the solution's source code sum of total CPU time over all threads individual CPU utilization It is common to see multiple solutions in the same programming language for the same problem. This highlights that within the constraints of a given language, a solution can be given which is either of high abstraction, is memory efficient, is fast, or can be parallelized better. Benchmark programs It was a design choice from the start to only include very simple toy problems, each providing a different kind of programming challenge. This provides users of the Benchmark Game the opportunity to scrutinize the various implementations. binary-trees chameneos-redux fannkuch-redux fasta k-nucleotide mandelbrot meteor-contest n-body pidigits regex-redux reverse-complement spectral-norm thread-ring History The project was known as The Great Computer Language Shootout until 2007. A port for Windows was maintained separately between 2002 and 2003. The sources have been archived on GitLab. There are also older forks on GitHub. The project is continuously evolving. The list of su
https://en.wikipedia.org/wiki/FGLM%20algorithm
FGLM is one of the main algorithms in computer algebra, named after its designers, Faugère, Gianni, Lazard and Mora. They introduced their algorithm in 1993. The input of the algorithm is a Gröbner basis of a zero-dimensional ideal in the ring of polynomials over a field with respect to a monomial order and a second monomial order. As its output, it returns a Gröbner basis of the ideal with respect to the second ordering. The algorithm is a fundamental tool in computer algebra and has been implemented in most of the computer algebra systems. The complexity of FGLM is O(nD3), where n is the number of variables of the polynomials and D is the degree of the ideal. There are several generalization and various applications for FGLM. References Computer algebra Commutative algebra Polynomials
https://en.wikipedia.org/wiki/Teo%20Mora
Ferdinando 'Teo' Mora is an Italian mathematician, and since 1990 until 2019 a professor of algebra at the University of Genoa. Life and work Mora's degree is in mathematics from the University of Genoa in 1974. Mora's publications span forty years; his notable contributions in computer algebra are the tangent cone algorithm and its extension of Buchberger theory of Gröbner bases and related algorithm earlier to non-commutative polynomial rings and more recently to effective rings; less significant the notion of Gröbner fan; marginal, with respect to the other authors, his contribution to the FGLM algorithm. Mora is on the managing-editorial-board of the journal AAECC published by Springer, and was also formerly an editor of the Bulletin of the Iranian Mathematical Society. He is the author of the tetralogy Solving Polynomial Equation Systems: Solving Polynomial Equation Systems I: The Kronecker-Duval Philosophy, on equations in one variable Solving Polynomial Equation Systems II: Macaulay's paradigm and Gröbner technology, on equations in several variables Solving Polynomial Equation Systems III: Algebraic Solving, Solving Polynomial Equation Systems IV: Buchberger Theory and Beyond, on the Buchberger algorithm Personal life Mora lives in Genoa. Mora published a book trilogy in 1977-1978 (reprinted 2001-2003) called on the history of horror films. Italian television said in 2014 that the books are an "authoritative guide with in-depth detailed descriptions and analysis." See also FGLM algorithm, Buchberger's algorithm Gröbner fan, Gröbner basis Algebraic geometry#Computational algebraic geometry, System of polynomial equations References Notes Further reading . and volumes: , . Reprinted 2001. also in: External links Official page Teo Mora and Michela Ceria, Do It Yourself: Buchberger and Janet bases over effective rings, Part 1: Buchberger Algorithm via Spear’s Theorem, Zacharias’ Representation, Weisspfenning
https://en.wikipedia.org/wiki/List%20of%20resurrected%20species
This is a list of resurrected species. So far, there is no scientifically verified method of DNA resurrection of a previously extinct species. Cloning Pyrenean Ibex The Pyrenean ibex (Capra pyrenaica pyrenaica) is an Iberian ibex subspecies with the unfortunate moniker of the first animal to go extinct twice. Endemic to the Pyrenees and Cantabrian Mountains, this ibex was driven to extinction due to competition with livestock and introduced wild ungulates. Several attempts were made to clone the Pyrenean ibex, and one individual was born to a domestic goat mother in 2003. However, this newborn died within minutes due to a lung defect. Seed Germination Judean Date Palm The Judean date palm is a cultivar of the date palm (Phoenix dactylifera) that is historically endemic to ancient Judea - modern Israel and Palestine. It is genetically unique, and closely related to modern Iraqi and Moroccan varieties. Between 1963 and 1991, archaeologists discovered Judean date seeds in excavation sites. Through radiocarbon dating, they were determined to be between 1,900 and 2,300 years old. In 2008, researchers at the Arava Institute for Environmental Studies began to germinate the seeds. As of 2023, 7 Judean date palms have successfully germinated. In 2020, researchers began to harvest dates from these trees. Experiments to revive this cultivar are ongoing. Thawing Pithovirus sibericum: 30,000 year old giant virus Panagrolaimus kolymaensis: This novel species of nematode was resurrected from cryptobiosis in 2023. The nematodes had been frozen in the Siberian permafrost since the Pleistocene, approximately 46,000 years ago. See also De-extinction Breeding back Lazarus taxon Rewilding References Extinction Conservation biology Emerging technologies
https://en.wikipedia.org/wiki/Jean%20Gallier
Jean Henri Gallier (born 1949) is a researcher in computational logic at the University of Pennsylvania, where he holds appointments in the Computer and Information Science Department and the Department of Mathematics. Biography Gallier was born January 5, 1949, in Nancy, France, and holds dual French and American citizenship. He earned his baccalauréat at the Lycée de Sèvres in 1966, and a degree in civil engineering at the École Nationale des Ponts et Chaussées in 1972. He then moved to the University of California, Los Angeles for his graduate studies, earning a Ph.D. in computer science in 1978 under the joint supervision of Sheila Greibach and Emily Perlinski Friedman. His dissertation was entitled Semantics and Correctness of Classes of Deterministic and Nondeterministic Recursive Programs. After postdoctoral study at the University of California, Santa Barbara, he joined the University of Pennsylvania Department of Computer and Information Science in 1978. At Pennsylvania, he was promoted to full professor in 1990, gained a secondary appointment to the Department of Mathematics in 1994, and directed the French Institute of Culture and Technology from 2001 to 2004. Contributions Gallier's most heavily cited research paper, with his student William F. Dowling, gives a linear time algorithm for Horn-satisfiability. This is a variant of the Boolean satisfiability problem: its input is a Boolean formula in conjunctive normal form with at most one positive literal per clause, and the goal is to assign truth values to the variables of the formula to make the whole formula true. Solving Horn-satisfiability problems is the central computational paradigm in the Prolog programming language. Gallier is also the author of five books in computational logic, computational geometry, low-dimensional topology, and discrete mathematics. Selected publications Research papers Books References External links Home page Living people American computer scientists French comp
https://en.wikipedia.org/wiki/Basis%20theorem%20%28computability%29
In computability theory, there are a number of basis theorems. These theorems show that particular kinds of sets always must have some members that are, in terms of Turing degree, not too complicated. One family of basis theorems concern nonempty effectively closed sets (that is, nonempty sets in the arithmetical hierarchy); these theorems are studied as part of classical computability theory. Another family of basis theorems concern nonempty lightface analytic sets (that is, in the analytical hierarchy); these theorems are studied as part of hyperarithmetical theory. Effectively closed sets Effectively closed sets are a topic of study in classical computability theory. An effectively closed set is the set of all paths through some computable subtree of the binary tree . These sets are closed, in the topological sense, as subsets of the Cantor space , and the complement of an effective closed set is an effective open set in the sense of effective Polish spaces. Kleene proved in 1952 that there is a nonempty, effectively closed set with no computable point (Cooper 1999, p. 134). Basis theorems show that there must be points that are not "too far" from being computable, in an informal sense. A class is a basis for effectively closed sets if every nonempty effectively closed set includes a member of  (Cooper 2003, p. 329). Basis theorems show that particular classes are bases in this sense. These theorems include (Cooper 1999, p. 134): The low basis theorem: each nonempty set has a member that is of low degree. The hyperimmune-free basis theorem: each nonempty set has a member that is of hyperimmune-free degree. The r.e. basis theorem: each nonempty set has a member that is of recursively enumerable (r.e.) degree. Here, a set is low if its Turing jump , the degree of the halting problem. has hyperimmune-free degree if every total -computable function is dominated by a total computable function (meaning for all ). No two of the above three theorem
https://en.wikipedia.org/wiki/Vacuum%20metallurgy
Vacuum metallurgy is the field of materials technology that deals with making, shaping, or treating metals in a controlled atmosphere, at pressures significantly less than normal atmospheric pressure. The purpose of vacuum metallurgy is to prevent contamination of metal by gases in the atmosphere. Alternatively, in some processes, a reactive gas may be introduced into the process to become part of the resultant product. Examples of vacuum metallurgy include vacuum degassing of molten steel in steelmaking operations, vacuum deposition of thin metal layers in manufacture of optics and semiconductors, vacuum casting, vacuum arc remelting of alloys, and vacuum induction melting. See also Electron-beam welding References Metallurgy
https://en.wikipedia.org/wiki/Simon%20Birrell
Simon Birrell (born 26 July 1966) is a British entrepreneur, technologist and film maker. He was part of the team that invented ambient intelligence and who, with Eli Zelkha, coined the term. Biography Early life, education and career Born in 1966 in Bristol, UK. He graduated from Cambridge University in 1988 with a degree in Natural Sciences. He has been a founder or co-founder of three companies. Euro-Profile/i-Profile – a business intelligence company based out of Silicon Valley which was acquired by Virgo Capital (2008), Vemm Brazil, a publisher of consumer advice websites in Brazil which was acquired by QuinStreet (2015) and Silicon Artists, a Madrid-based entertainment technology company funded by Silicon Valley-based Tandem Computers. Ambient intelligence In 1998, Birrell was part of the team at Palo Alto Ventures that invented and developed the ambient intelligence concept and who, with Eli Zelkha, coined the term. It was presented by Roel Pieper of Philips at The Digital Living Room Conference on 22 June 1998. Since its invention in 1998, Ambient Intelligence labs have been formed at leading universities and ambient intelligence has become part of the core strategies of many of the world's leading technology companies, including Microsoft, Google, Amazon and IBM. Robotics and deep learning Birrell is researching deep learning and robotics at Cambridge University. He is the author of the blog Artificial Human Companions. Video games, virtual reality and other activities He developed some of the very first video games for Richard Branson's Virgin Interactive in 1983. These included Bug Bomb – BBC Micro (1983), Microbe – BBC Micro (1983), High-Rise Horror – Commodore 64 (1984), Strangeloop – Commodore 64 (1985), Shogun – Commodore 64 / Amstrad (co-design). From 1993 to 1995, Birrell was the CTO of an early virtual reality company in Spain called Realidad Virtual S.L. At Realidad Virtual, he developed Pandora – the first Spanish online virtual reality p
https://en.wikipedia.org/wiki/Chronicles%20of%20Terror
Chronicles of Terror is a digital internet archive established by the in August 2016. Initially, it provided access to the depositions of Polish citizens who after World War II were interviewed as witnesses before the Main Commission for the Investigation of German Crimes in Poland. From 17 September 2017, the database also presents the accounts of Poles who fell victim to repressions perpetrated by Soviet totalitarianism. History The Main Commission for the Investigation of German Crimes in Poland, set up in 1945, documented German crimes committed during World War II, conducted investigations and published the results of its activities. In 1949, its name was changed to the Main Commission for the Investigation of Hitlerite Crimes in Poland. The commission was active throughout the country. The commission's tasks, their scope expanded to include Communist terror, were taken over by the investigative division of the Institute of National Remembrance (IPN) in 1998. The commission's documentation, including witness testimonies, was also passed over to the archives of the institute. Acting on the basis of an agreement dated 11 February 2016, the IPN started submitting digital copies of these materials with the objective of making them available over the internet. The Chronicles of Terror website premiered on 4 August 2016, initially providing access to a database containing more than 500 testimonies given before the commission. On 25 January 2017 the Witold Pilecki Center for Totalitarian Studies signed an agreement with the Hoover Institution, pursuant to which Chronicles of Terror would be supplemented by depositions relating to Communist crimes, taken from the archives of Anders' Army. After World War II, documents containing the accounts of Poles, both soldiers and civilians who left the Soviet Union together with the Polish Army, were deposited with the Hoover Library, for it was feared that they could have been seized by the Communist authorities. Testimony
https://en.wikipedia.org/wiki/Naira%20Hovakimyan
Naira Hovakimyan (born September 21, 1966) is an Armenian control theorist who holds the W. Grafton and Lillian B. Wilkins professorship of the Mechanical Science and Engineering at the University of Illinois at Urbana-Champaign. She is the director of AVIATE Center of flying cars at UIUC, funded through a NASA University Leadership Initiative. She was the inaugural director of the Intelligent Robotics Laboratory during 2015–2017, associated with the Coordinated Science Laboratory at the University of Illinois at Urbana-Champaign. Education Naira Hovakimyan received her MS degree in Theoretical Mechanics and Applied Mathematics in 1988 from Yerevan State University in Armenia. She got her Ph.D. in Physics and Mathematics in 1992, in Moscow, from the Institute of Applied Mathematics of Russian Academy of Sciences, majoring in optimal control and differential games. Academic life Before joining the faculty of the University of Illinois at Urbana–Champaign in 2008, Naira Hovakimyan has spent time as a research scientist at Stuttgart University in Germany, at INRIA in France, at Georgia Institute of Technology, and she was on faculty of Aerospace and Ocean engineering of Virginia Tech during 2003–2008. She is currently W. Grafton and Lillian B. Wilkins Professor of Mechanical Science and Engineering at UIUC. In 2015, she was named as inaugural director for Intelligent Robotics Laboratory of CSL at UIUC. Currently she is the director of AVIATE Center of flying cars at UIUC, funded through a NASA University Leadership Initiative. She has co-authored two books, ten book chapters, eleven patents, and more than 450 journal and conference papers. Research areas Her research interests are in control and optimization, autonomous systems, machine learning, cybersecurity, neural networks, game theory and their applications in aerospace, robotics, mechanical, agricultural, electrical, petroleum, biomedical engineering and elderly care. Honors She is the 2011 recipient of
https://en.wikipedia.org/wiki/Gr%C3%B6bner%20fan
In computer algebra, the Gröbner fan of an ideal in the ring of polynomials is a concept in the theory of Gröbner bases. It is defined to be a fan consisting of cones that correspond to different monomial orders on that ideal. The concept was introduced by Mora and Robbiano in 1988. The result is a weaker version of the result presented in the same issue of the journal by Bayer and Morrison. Gröbner fan is a base for the nowadays active field of tropical geometry. One implementation of the Gröbner fan is called Gfan, based on an article of Fukuda, et al. which is included in some computer algebra systems such as Singular, Macaulay2, and CoCoA. See also Gröbner basis Tropical geometry References Computer algebra Algebraic geometry Commutative algebra
https://en.wikipedia.org/wiki/Feller%E2%80%93Tornier%20constant
In mathematics, the Feller–Tornier constant CFT is the density of the set of all positive integers that have an even number of distinct prime factors raised to a power larger than one (ignoring any prime factors which appear only to the first power). It is named after William Feller (1906–1970) and Erhard Tornier (1894–1982) Omega function The Big Omega function is given by See also: Prime omega function. The Iverson bracket is With these notations, we have Prime zeta function The prime zeta function P is give by The Feller–Tornier constant satisfies See also Riemann zeta function L-function Euler product Twin prime References Mathematical constants Zeta and L-functions Infinite products
https://en.wikipedia.org/wiki/Eight-segment%20display
An eight-segment display is a type of display based on eight segments that can be turned on or off according to the font pattern to be produced. It is similar to a nine-segment display whose middle vertical bars are vertical, except that on an eight-segment display, the bars F and G are merged. Applications One application was in the Sharp EL-8, an early electronic calculator. The eight-segment display produces more rounded digits than a seven-segment display, yielding a more "script-like" output, with the trade-off that fewer possible alphabetic characters can be displayed because the bars F and G are merged (see table below). Displaying An eight segment display can sometimes display alphabetic characters with less readability because the segments F and G are combined and the corners are rounded. The asymmetrical layout of the elements produced a distinctive "handwritten" digit style, with a half-height "0". Because of graphical confusion, it is unable to display the following characters: Examples See also Seven-segment display Nine-segment display Fourteen-segment display Sixteen-segment display Dot matrix display Vacuum fluorescent display Display technology
https://en.wikipedia.org/wiki/Porter%27s%20constant
In mathematics, Porter's constant C arises in the study of the efficiency of the Euclidean algorithm. It is named after J. W. Porter of University College, Cardiff. Euclid's algorithm finds the greatest common divisor of two positive integers and . Hans Heilbronn proved that the average number of iterations of Euclid's algorithm, for fixed and averaged over all choices of relatively prime integers , is Porter showed that the error term in this estimate is a constant, plus a polynomially-small correction, and Donald Knuth evaluated this constant to high accuracy. It is: where is the Euler–Mascheroni constant is the Riemann zeta function is the Glaisher–Kinkelin constant See also Lochs' theorem Lévy's constant References Mathematical constants Analytic number theory
https://en.wikipedia.org/wiki/DNA%20spiking
DNA spiking, also known as custom spiking, is the differing ratio of bases at a single degenerate position when synthesizing oligonucleotides. DNA spiking is a unequal proportions of bases at a given position (for example, 10% Adenine, 75% Guanine, 5% Cytosine & 10% Thymine). As an example, with the degenerate code R = A + G, 50% of the time that R position is adenine and the other 50% of the time it is guanine. However, with DNA Spiking, the R position could be adenine 70% of the time and guanine 30% of the time. The proportions do not need to be 70:30, the ratios can be anything else such as 12:82 and 64:36. DNA spiking can also refer to a spike control in PCR, which is when DNA is added to a sample that will provide some signal (e.g. a plasmid or some synthetic DNA with a specific known sequence) to a reaction, and seeing if the reaction will amplify. This method is used to discover if the PCR method is working correctly, as in a PCR machine it may not amplify DNA properly, so by adding spiked DNA it can be observed how much DNA is produced. This is then compared to the amount of DNA that would be theoretically predicted if the machine was working properly so that any malfunctions can be discovered. See also RNA spike-in References DNA Biochemistry
https://en.wikipedia.org/wiki/Caffe%20%28software%29
Caffe (Convolutional Architecture for Fast Feature Embedding) is a deep learning framework, originally developed at University of California, Berkeley. It is open source, under a BSD license. It is written in C++, with a Python interface. History Yangqing Jia created the Caffe project during his PhD at UC Berkeley. It is currently hosted on GitHub. Features Caffe supports many different types of deep learning architectures geared towards image classification and image segmentation. It supports CNN, RCNN, LSTM and fully-connected neural network designs. Caffe supports GPU- and CPU-based acceleration computational kernel libraries such as Nvidia cuDNN and Intel MKL. Applications Caffe is being used in academic research projects, startup prototypes, and even large-scale industrial applications in vision, speech, and multimedia. Yahoo! has also integrated Caffe with Apache Spark to create CaffeOnSpark, a distributed deep learning framework. Caffe2 In April 2017, Facebook announced Caffe2, which included new features such as recurrent neural network (RNN). At the end of March 2018, Caffe2 was merged into PyTorch. See also Comparison of deep learning software References External links Deep learning software Free science software Free statistical software Image processing Information technology companies of the United States Software using the BSD license
https://en.wikipedia.org/wiki/Divisare
Divisare is the largest library of contemporary architecture online, founded in 1998 in Rome, Italy. Published projects are classified in curated albums, each dedicated to a specific topic or typology. History Divisare is one of the first architecture websites in the history of the web. Online since 1998 as Europaconcorsi, it was rebranded in August 2015, launching three new different platforms: Divisare, Europaconcorsi and Architettura Italiana. At first, Europaconcorsi was a service to inform architects about design competitions in Europe (in Italian, "concorsi" means competitions), on recently issued tenders as well as the results of competitions, with rankings of the winners and the participating projects. From the outset Europaconcorsi was conceived as a structure of related but distinct databases: informations on competitions, projects and designers were gathered in autonomous data structures. With almost 20 years of history behind it, and over 140,000 projects, Divisare is the largest online archive of contemporary architecture. Curated by the team based in Rome, Italy, Divisare is structured as a library of thematic albums (organized per elements, cities, houses, ideas, materiality, plans & details, private interiors, public interiors, topics and types) featuring the most relevant pieces of architecture ever designed. Complementary with the daily showcase on the online platform, registered subscribers received Journal, a weekly review curated by the editorial staff. Divisare means imagining, designing with the mind, literally "to devise". The term was used by Leon Battista Alberti to define the work of the architect: "Him I call an Architect, who, by sure and wonderful Art and Method, is able, both with Thought and Invention, to devise." (Leon Battista Alberti — De Re Aedificatoria, 1450 - original Italian: "Architettore chiamerò io colui, il quale saprà con certa, e maravigliosa ragione, e regola, sì con la mente, e con lo animo divisare"). During Nove
https://en.wikipedia.org/wiki/Materials%20with%20memory
In continuum physics, materials with memory, also referred as materials with hereditary effects are a class of materials whose constitutive equations contains a dependence upon the past history of thermodynamic, kinetic, electromagnetic or other kind of state variables. Historical notes The study of these materials arises from the pioneering articles of Ludwig Boltzmann and Vito Volterra, in which they sought an extension of the concept of an elastic material. The key assumption of their theory was that the local stress value at a time depends upon the history of the local deformation up to . In general, in materials with memory the local value of some constitutive quantity (stress, heat flux, electric current, polarization and magnetization, etc.) at a time depends upon the history of the state variables (deformation, temperature, electric and magnetic fields, etc.). The hypothesis that the remote history of a variable has less influence than its values in the recent past, was stated in modern continuum mechanics as the fading memory principle by Bernard Coleman and Walter Noll. This assumption was implicit in the pioneer works: when restricted to cyclic hystories, it traces back to the closed cycle principle stated by Volterra, which leads to a constitutive relation of integral convolution type. In the linear case, this relation takes the form of a Volterra equation Constitutive relations of materials with memory In the linear case, this relation takes the form of a Volterra equation See also Biomaterial Biomechanics Dielectric relaxation Hysteresis Rheology Viscosity Viscoelasticity Viscoplasticity Notes References Articles , published also as . . (online version ). . . In this paper, written only few years after the discovery of the effect itself, Dario Graffi proposes a theory of the Luxemburg effect based on Volterra's theory of hereditary phenomena. (online version ). (online version ). In this paper, Graffi introduces the free energy now called G
https://en.wikipedia.org/wiki/Product%20analysis
Product analysis involves examining product features, costs, availability, quality, appearance and other aspects. Product analysis is conducted by potential buyers, by product managers attempting to understand competitors and by third party reviewers. Product analysis can also be used as part of product design to convert a high-level product description into project deliverables and requirements. It involves all facts of the product, its purpose, its operation, and its characteristics. Techniques Related techniques include product breakdown, systems analysis, systems engineering, value engineering, value analysis and functional analysis. Product breakdown: Recursively divide the product into components and subcomponents. Technology analysis Technological analysis is sometimes applied in decision-making often related to investments, policy-decisions and public spending. They can be done by a variety of organization-types such as for-profit companies, non-profit think tanks, research institutes, public platforms and government agencies and evaluate established, emerging and potential future technologies on a variety of measures and metrics – all of which are related to ideals and goals such as minimal global greenhouse gas emissions – such as life-cycle-sustainability, openness, performance, control, financial costs, resource costs, health impacts and more. Results are sometimes published as public reports or as scientific peer-reviewed studies. Based on such reports standardization can enable interventions or efforts which balance competition and cooperation and improve sustainability, reduce waste and redundancy, or accelerate innovation. They can also be used for the creation of standardized system designs that integrate a variety of technologies as their components. Other applications include risk assessment and research of defense applications. They can also be used or created for determining the hypothetical or existing optimal solution/s and to identify
https://en.wikipedia.org/wiki/GlobalVision
GlobalVision is a SaaS software company specializing in packaging and labeling. History Global Vision Inc. (GlobalVision) waReuben Malz established Global Vision Inc. (GlobalVision) in the year 1990.s founded in 1990 by Reuben Malz. It is headquartered in Montreal, Canada, and has regional offices in the UK and Germany. In 2013, Pfizer contracted with GlobalVision for packaging quality control. GlobalVision introduced its first web-based quality control platform, QCanywhere, in 2013. This was followed by the introduction of a more advanced version of the platform, Proofware, in 2014. In 2017, GlobalVision released a desktop quality control platform. In 2018, GlobalVision partnered with Pantone and X-Rite to launch the world’s first digital color inspection system for print and packaging. References Canadian companies established in 1990 Software companies established in 1990 Companies based in Montreal Automation software Braille technology Packaging industry Packaging Machine vision Print production Platform virtualization software
https://en.wikipedia.org/wiki/IBM%20Quantum%20Platform
IBM Quantum Platform (previously as IBM Quantum Experience) is an online platform allowing public and premium access to cloud-based quantum computing services provided by IBM. This includes access to a set of IBM's prototype quantum processors, a set of tutorials on quantum computation, and access to an interactive textbook. As of February 2021, there are over 20 devices on the service, six of which are freely available for the public. This service can be used to run algorithms and experiments, and explore tutorials and simulations around what might be possible with quantum computing. IBM's quantum processors are made up of superconducting transmon qubits, located in dilution refrigerators at the IBM Research headquarters at the Thomas J. Watson Research Center. Users interact with a quantum processor through the quantum circuit model of computation. Circuits can be created either using graphically with the Quantum Composer, or programmatically within the Jupyter notebooks of the Quantum Lab. Circuits are created using Qiskit and can be compiled down to OpenQASM for execution on real quantum systems. History The service was launched in May 2016 as the IBM Quantum Experience with a five-qubit quantum processor and matching simulator connected in a star shaped pattern. At this time, users could only interact with the hardware through the quantum composer GUI. Quantum circuits were also limited to the specific two-qubit gates available on the hardware. In July 2016, IBM launched the IBM Quantum Experience community forum. This was subsequently replaced by a Slack workspace. In January 2017, IBM made a number of additions to the IBM Quantum Experience, including increasing the set of two-qubit interactions available on the five-qubit quantum processor, expanding the simulator to custom topologies up to twenty qubits, and allowing users to interact with the device and simulator using quantum assembly language code. In March 2017, IBM released Qiskit to enable u
https://en.wikipedia.org/wiki/Drug%20disposal
Drug disposal is the discarding of drugs. Individuals commonly dispose of unused drugs that remain after the end of medical treatment. Health care organizations dispose of drugs on a larger scale for a range of reasons, including having leftover drugs after treating patients and discarding of expired drugs. Failure to properly dispose of drugs creates opportunities for others (of whom the drug is unintended) to take them inappropriately. Inappropriate disposal of drugs can also cause drug pollution. People dispose of drugs in various ways; even organizations with expertise on drugs may give inconsistent information to consumers about drug disposal. Proper waste management system including distribution, control, and disposal not only helps the Healthcare Centres but also promote environmental health. Sources of drugs Medication waste includes both hazardous and non-hazardous waste, controlled substances, and expired pharmaceuticals. Medication waste can come from multiple levels in the drug's lifespan. First, it can come from production factories from where they were created. This includes unwanted pharmaceutical ingredients and materials that can no longer be used in the drug manufacturing process. Second, medication waste can be generated from healthcare facilities including hospitals, clinics, and pharmacies. Medication waste from this source can be from over prescribing of drugs from healthcare providers, hospital labs, expired drugs, opened drug containers and partially used medications. Furthermore, these wastes can include materials, such as syringes, vials, IV bags, and tubing that contain excess drugs or contaminated in the process of handling hazardous pharmaceuticals, such as chemotherapy drugs. Some states have regulations that require healthcare facilities to destroy unused medications. Lastly, pharmaceutical waste can come from excessive consumption of over-the-counter medications from patients. Sometimes normal use of a drug can result in waste. F
https://en.wikipedia.org/wiki/Impinging%20mixer
Impinging mixers combine and disperse resins within each other, and are often used in reaction injection molding (RIM). Mixing occurs as two high velocity streams collide in a mixing chamber. High velocity results in a turbulent rather than a laminar flow. Impingement mixing is most effective when it occurs at the center of the mixing chamber. Thermosetting plastics cure by a chemical reaction between two resins. The resins must be mixed immediately before they are injected into a mold. The mixing can be done by impingement mixing, where two streams to collide at high velocity in a mixing chamber. As soon as the mixing chamber is full, a piston immediately pushes the mixed resin into the mold, leaving very little mixed resin curing outside the mold. References Injection molding
https://en.wikipedia.org/wiki/Zellballen
A zellballen is a small nest of chromaffin cells or chief cells with pale eosinophilic staining. Zellballen are separated into groups by segmenting bands of fibrovascular stroma, and are surrounded by supporting sustentacular cells. A zellballen pattern is diagnostic for paraganglioma or pheochromocytoma. Zellballen is German for "ball of cells". References Cell biology Human cells Neuroendocrine cells Adrenal gland
https://en.wikipedia.org/wiki/Elective%20genetic%20and%20genomic%20testing
Elective genetic and genomic testing are DNA tests performed for an individual who does not have an indication for testing. An elective genetic test analyzes selected sites in the human genome while an elective genomic test analyzes the entire human genome. Some elective genetic and genomic tests require a physician to order the test to ensure that individuals understand the risks and benefits of testing as well as the results. Other DNA-based tests, such as a genealogical DNA test do not require a physician's order. Elective testing is generally not paid for by health insurance companies. With the advent of personalized medicine, also called precision medicine, an increasing number of individuals are undertaking elective genetic and genomic testing. History Genetic testing for a variety of disorders has seen many advances starting with cytogenetics to evaluate human chromosomes for aneuploidy and other chromosome abnormalities. The development of molecular cytogenetics involving techniques such as fluorescence in situ hybridization (FISH) followed, permitting the detection of more subtle changes in the karyotype. Techniques to determine the precise sequence of nucleotides in DNA by DNA sequencing, notably Sanger sequencing was developed in the 1970s. In the 1980s the DNA microarray appeared, permitting laboratories to find copy number variants associated with disease that are below the level of detection of cytogenetics but too large to be detected by DNA sequencing. In recent years the development of high-throughput or next-generation sequencing has dramatically lowered the cost of DNA sequencing permitting laboratories to evaluate all 20,000 genes of the human genome at once through exome sequencing and whole genome sequencing. A catalogue of the many uses of these techniques can be found in the section: genetic testing. Most elective genetic and genomic testing employs either a DNA microarray or next-generation sequencing. Historically, all laboratory tes
https://en.wikipedia.org/wiki/ShEx
Shape Expressions (ShEx) is a data modelling language for validating and describing a Resource Description Framework (RDF). It was proposed at the 2012 RDF Validation Workshop as a high-level, concise language for RDF validation. The shapes can be defined in a human-friendly compact syntax called ShExC or using any RDF serialization formats like JSON-LD or Turtle. ShEx expressions can be used both to describe RDF and to automatically check the conformance of RDF data. The syntax of ShEx is similar to Turtle and SPARQL while the semantics is inspired by regular expression languages like RelaxNG. Example PREFIX : <http://example.org/> PREFIX schema: <http://schema.org/> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> :Person { schema:name xsd:string ; schema:knows @:Person * ; } The previous example declares that nodes conforming to shape Person must have one property schema:name with a string value and zero or more properties schema:knows whose values must conform with shape Person. Implementations Online playgrounds and demos ShExSimple: Online demo based on shex.js rdfshape: online demo based on shaclex References Further reading Specification Other See also SHACL Wikidata XML schemas Resource Description Framework Data modeling languages Declarative programming languages RDF data access Semantic Web World Wide Web Consortium standards SPARQL
https://en.wikipedia.org/wiki/Infinite%20chess
Infinite chess is any variation of the game of chess played on an unbounded chessboard. Versions of infinite chess have been introduced independently by multiple players, chess theorists, and mathematicians, both as a playable game and as a model for theoretical study. It has been found that even though the board is unbounded, there are ways in which a player can win the game in a finite number of moves. Background Classical (FIDE) chess is played on an 8×8 board (64 squares). However, the history of chess includes variants of the game played on boards of various sizes. A predecessor game called courier chess was played on a slightly larger 12×8 board (96 squares) in the 12th century, and continued to be played for at least six hundred years. Japanese chess (shogi) has been played historically on boards of various sizes; the largest is taikyoku shōgi ("ultimate chess"). This chess-like game, which dates to the mid 16th century, was played on a 36×36 board (1296 squares). Each player starts with 402 pieces of 209 different types, and a well-played game would require several days of play, possibly requiring each player to make over a thousand moves. Chess player Jianying Ji was one of many to propose infinite chess, suggesting a setup with the chess pieces in the same relative positions as in classical chess, with knights replaced by nightriders and a rule preventing pieces from travelling too far from opposing pieces. Numerous other chess players, chess theorists, and mathematicians who study game theory have conceived of variations of infinite chess, often with different objectives in mind. Chess players sometimes use the scheme simply to alter the strategy; since chess pieces, and in particular the king, cannot be trapped in corners on an infinite board, new patterns are required to form a checkmate. Theorists conceive of infinite chess variations to expand the theory of chess in general, or as a model to study other mathematical, economic, or game-playing st
https://en.wikipedia.org/wiki/Lifting%20property
In mathematics, in particular in category theory, the lifting property is a property of a pair of morphisms in a category. It is used in homotopy theory within algebraic topology to define properties of morphisms starting from an explicitly given class of morphisms. It appears in a prominent way in the theory of model categories, an axiomatic framework for homotopy theory introduced by Daniel Quillen. It is also used in the definition of a factorization system, and of a weak factorization system, notions related to but less restrictive than the notion of a model category. Several elementary notions may also be expressed using the lifting property starting from a list of (counter)examples. Formal definition A morphism in a category has the left lifting property with respect to a morphism , and also has the right lifting property with respect to , sometimes denoted or , iff the following implication holds for each morphism and in the category: if the outer square of the following diagram commutes, then there exists completing the diagram, i.e. for each and such that there exists such that and . This is sometimes also known as the morphism being orthogonal to the morphism ; however, this can also refer to the stronger property that whenever and are as above, the diagonal morphism exists and is also required to be unique. For a class of morphisms in a category, its left orthogonal or with respect to the lifting property, respectively its right orthogonal or , is the class of all morphisms which have the left, respectively right, lifting property with respect to each morphism in the class . In notation, Taking the orthogonal of a class is a simple way to define a class of morphisms excluding non-isomorphisms from , in a way which is useful in a diagram chasing computation. Thus, in the category Set of sets, the right orthogonal of the simplest non-surjection is the class of surjections. The left and right orthogonals of the simplest non-inje
https://en.wikipedia.org/wiki/Gene%20amplification%20in%20Paramecium%20tetraurelia
Gene amplification in Paramecium tetraurelia is an example of gene amplification that has occurred in the unicellular organism Paramecium tetraurelia. Gene duplication occurs in a large number of organisms as part of evolution or as the cause or result of disease (as in the case of the amylase genes in humans, and genes in cancer cells respectively). Gene duplication often leads to amplification of their gene products due to transcription and translation of all gene duplicates. Evidence of gene duplication has been observed the inheritance patterns of Paramecium tetraurelia, a common model organism. In one strain of P. tetraurelia, d4-95, a recessive mutant allele of a gene known as pawn-B found in this strain is inherited through gene duplication and amplification between generations, and even self-fertilizations. The inheritance of this allele is the first description of gene duplication and amplification in the micronucleus of ciliates. Additionally, it appears that the duplication of the mutant allele occurred after mutagenesis due to the similarity in nucleotide sequences of different copies of the mutant allele, especially in the coding region. When the d4-95 strain was crossed with a wild-type P. tetraurelia, F2 and later progeny often expressed the phenotype of the pawn-B mutant, despite carrying a wild-type gene at the pawn-B locus. This phenotype was maintained in progeny even after the self-fertilization of theoretical wild-type homozygotes that had been recovered from the cross. As is the case of other Paramecium, P. tetraurelia exhibits a number of non-Mendelian modes of inheritance, partially due to the existence of both macro- and micronuclei. In both the macro- and micronucleus of the d4-95 strain of P. tetraurelia contained many more copies of the mutant gene than in the wild type strain. This occurs due to the ability of most of the extra pawn-B gene copies to be heritable independently from the original pawn-B locus. Additionally, there is evi
https://en.wikipedia.org/wiki/Notch%20tensile%20strength
The notch tensile strength (NTS) of a material is the value given by performing a standard tensile strength test on a notched specimen of the material. The ratio between the NTS and the tensile strength is called the notch strength ratio (NSR). See also Charpy impact test References Physical quantities Fracture mechanics Materials testing Elasticity (physics)
https://en.wikipedia.org/wiki/Bull%20Gamma%2060
The Bull Gamma 60 was a large transistorized mainframe computer designed by Compagnie des Machines Bull. Initially announced in 1957, the first unit shipped in 1960. It holds the distinction of being the world's first multi-threaded computer, and the first to feature an architecture specially designed for parallelism. The Gamma 60 spearheaded numerous groundbreaking technologies during the early 1960s, notably in multi-programming, utilizing tools that were still in their nascent stages. Upon its release, its architecture garnered significant attention among machine designers, becoming a subject of study alongside contemporary supercomputers and being cited as an example for progress in computer design. Despite its innovations, the Gamma 60's large footprint (close to 4000 sq.ft), high cost, energy consumption, and complexity ultimately resulted in limited commercial success, with about only twenty units sold worldwide. Its main competitors included the IBM 7070, 7090, and 7030 "Stretch". The last Gamma 60 remained in service until 1974. Design The Gamma 60 marked Bull's entry into core memory, solid-state logic and magnetic tape capabilities. Its architectural core was based on a large, high-speed central memory, with an arbitrator (known as the Program Distributor) responsible for distributing data and instructions to the various units within the computer. The processor was segmented into a central unit and a series of discrete, specialized processing units. This design allowed for the concurrent operation of up to five clusters, each containing five processing units. Each unit in the computer, whether a processing unit or a peripheral device, operated autonomously and would request data and instructions from the central unit when they became available. Data transmission to and from the processing units occurred through two independent buses—one for transmission and another for retrieval. Processor The processor operated in a 24-bit parallel configuration,
https://en.wikipedia.org/wiki/Gaussian%20correlation%20inequality
The Gaussian correlation inequality (GCI), formerly known as the Gaussian correlation conjecture (GCC), is a mathematical theorem in the fields of mathematical statistics and convex geometry. The statement The Gaussian correlation inequality states: Let be an n-dimensional Gaussian probability measure on , i.e. a multivariate normal distribution, centered at the origin. Then for all convex sets that are symmetric about the origin, As a simple example for n=2, one can think of darts being thrown at a board, with their landing spots in the plane distributed according to a 2-variable normal distribution centered at the origin. (This is a reasonable assumption for any given darts player, with different players being described by different normal distributions.) If we now consider a circle and a rectangle in the plane, both centered at the origin, then the proportion of the darts landing in the intersection of both shapes is no less than the product of the proportions of the darts landing in each shape. This can also be formulated in terms of conditional probabilities: if you're informed that your last dart hit the rectangle, then this information will increase your estimate of the probability that the dart hit the circle. History A special case of the inequality was conjectured in 1955; further development was given by Olive Jean Dunn in 1958. The general case was stated in 1972, also as a conjecture. The case of dimension n=2 was proved in 1977 and certain special cases of higher dimension have also been proven in subsequent years. The general case of the inequality remained open until 2014, when Thomas Royen, a retired German statistician, proved it using relatively elementary tools. In fact, Royen generalized the conjecture and proved it for multivariate gamma distributions. The proof did not gain attention when it was published in 2014, due to Royen's relative anonymity and the fact that the proof was published in a predatory journal. Another reason was
https://en.wikipedia.org/wiki/Noise%20%28spectral%20phenomenon%29
Noise refers to many types of random, troublesome, problematic, or unwanted signals. Acoustic noise may mar aesthetic experience, such as attending a concert hall. It may also be a medical issue inherent in the biology of hearing. In technology, noise is unwanted signals in a device or apparatus, commonly of an electrical nature. The nature of noise is much studied in mathematics and is a prominent topic in statistics. This article provides a survey of specific topics linked to their primary articles. Acoustic noise In transportation Aircraft noise Jet noise, caused by high-velocity jets and turbulent eddies Noise and vibration on maritime vessels Noise, vibration, and harshness, quality criteria for vehicles Traffic noise, including roadway noise and train noise Other acoustic noise Acoustic noise, technical term for any sound, even deliberate Artificial noise, in spectator sports Background noise, in acoustics, any sound other than the monitored one Comfort noise, used in telecommunications to fill silent gaps Grey noise, random noise with a psychoacoustic adjusted spectrum Industrial noise, relevant to hearing damage and industrial hygiene Noise pollution, that affects negatively the quality of life Noise in biology Cellular noise, in biology, random variability between cells Developmental noise, variations among living beings with the same genome Neuronal noise, in neuroscience Synaptic noise, in neuroscience Transcriptional noise, in biochemistry, errors in genetic transcription Noise in computer graphics Noise in computer graphics refers to various pseudo-random functions used to create textures, including: Gradient noise, created by interpolation of a lattice of pseudorandom gradients Perlin noise, a type of gradient noise developed in 1983 Simplex noise, a method for constructing an n-dimensional noise function comparable to Perlin noise Simulation noise, a function that creates a divergence-free field Value noise, created by interpolation of a lat
https://en.wikipedia.org/wiki/Larry%20D.%20Nichols
Larry D. Nichols, born 1939 in the United States, is a puzzle designer. He grew up in Xenia, Ohio, and studied chemistry at DePauw University in Greencastle, Indiana, before moving to Massachusetts to attend Harvard Graduate School. He is best known for the invention of mechanical puzzles including 'The Nichols Cube Puzzle' (1974), patent US365520. He has lived with his wife Karen in Arlington, Massachusetts since 1959. The Nichols Cube Puzzle (Twizzle) In 1957, 17 years before Dr. Rubik’s invention popularly known as the “Rubik's Cube”, Dr. Nichols conceived of a twist cube puzzle with six colored faces. It was a 2×2×2 cube assembled from eight unit cubes with magnets on their inside faces, allowing the cubes to rotate in groups of four around three axes. The object of the puzzle was to mix the colors on the faces of the cube and then restore them. After making many preliminary models, in 1968 a working prototype was constructed, and on April 11, 1972, U.S. patent 3,655,201 was issued covering the Nichols' Cube. The patent focused on the 2×2×2 puzzle but mentioned the possibility of larger versions. Nichols' patents became the subject of a court hearing, between his employer, and the Ideal Toy Company. In 1985, a U.S. District Court ruled that Rubik’s Cube infringed the Nichols patent, but in 1986 the Court of Appeals ruled that only the smaller 2×2×2 Rubik’s Pocket Cube was guilty of infringement, and not the popular 3×3×3 Rubik's Cube. Analogues of the invention Manipulative toy William O Gustafson 1960 Rubik's Cube 1974 Frank Fox got a British patent for a spherical sliding puzzle in 1974 Terutoshi Ishige received a Japanese patent for a 3x3x3 in 1976 Career Nichols received a degree in chemistry in 1958 from DePauw University, where he was a Rector Scholar, and went on to earn a doctorate at Harvard. As a Harvard student, he invented and produced games and puzzles. He became chief scientist for the Moleculon Research Corporation of Cambridge,
https://en.wikipedia.org/wiki/Physiological%20functional%20capacity
Physiological functional capacity (PFC) is the ability to perform the physical tasks of daily life and the ease with which these tasks can be performed. PFC declines at some point with advancing age even in healthy adults, resulting in a reduced capacity to perform certain physical tasks. This can eventually result in increased incidence of functional disability, increased use of health care services, loss of independence, and reduced quality of life. See also Human body Frailty syndrome Frailty index Functional residual capacity, where it pertains to the lungs Physiology References Ageing Physiology
https://en.wikipedia.org/wiki/Genetics%20of%20synesthesia
The genetic mechanism of synesthesia has long been debated, with researchers previously claiming it was a single X-linked trait due to seemingly higher prevalence in women and no evidence of male-male transmission This is where the only synesthetic parent is male and the male child has synesthesia, meaning that the trait cannot be solely linked to the X chromosome. The Mendelian nature of the trait was further disproven when case studies showed that the Phenotype of synesthesia could be differentially expressed in monozygotic (genotypically identical) twins While both twins had the same genome with the potential for phenotypic expression of synesthesia, only one had documented synesthesia. Therefore, the condition is now thought to be oligogenic, with Locus heterogeneity and multiple forms of inheritance, and expression, implying that synesthesia is determined by more than one gene, more than one location in those genes, and a complex mode of inheritance. Several full genome linkage scans have shown particular areas of the genome whose inheritance seem to correlate with the inheritance of synesthesia. Using the LOD score which describes the likelihood that two genes are near each other on a chromosome, and thus will be inherited together, areas of strong or suggestive linkage with inheritance of synesthesia were found. The area with the highest LOD score in the genome of an individual with auditory-visual synesthesia has been shown to be linked with autism as well, another disorder with sensory and perceptual abnormalities. Other regions of linkage include genes that are related to the development of the cerebral cortex (TBR1), dyslexia, and apoptosis (EFHC1), the last of which could be potentially related to the retention of the neonatal synesthetic pathways in the universal synesthesia/pruning hypothesis. This hypothesis posits that every person is born a synesthete and the ‘extra’ connections are pruned during normal neurodevelopment in non-synesthetes, and n
https://en.wikipedia.org/wiki/R/place
r/place is a recurring collaborative project and social experiment hosted on the content aggregator site Reddit. Originally launched on April Fools' Day 2017, it has since been repeated again on April Fools' Day 2022 and on July 20, 2023. The 2017 experiment involved an online canvas located at a subreddit called r/place. Registered users could edit the canvas by changing the color of a single pixel with a replacement from a 16-color palette. After each pixel was placed, a timer prevented the user from placing any more pixels for a period of time varying from 5 minutes (depending on whether the user had verified their email address). The idea of the experiment was conceived by Josh Wardle. It was ended by Reddit administrators about 72 hours after its creation, on 3 April 2017. Over a million users edited the canvas, placing a total of approximately 16 million pixels, and, at the time the experiment was ended, over 90,000 users were actively viewing or editing the canvas. The experiment was commended for its representation of the culture of Reddit's online communities, and of Internet culture as a whole. Overview The experiment, during the 2017 edition, was based in a subreddit called r/place, in which individual registered users could place a single colored pixel (or "tile") on an online canvas of one million (1000 x 1000) pixel squares, and wait a certain amount of time before placing another. In 2017, the waiting time varied from 5 to 20 minutes throughout the experiment, and the user could choose their pixel's color from a palette of sixteen colors. The 2022 edition started with the same size and colors as 2017, but the canvas was later expanded to four million (2000 x 2000) pixel squares, and the palette gradually gained sixteen more colors for a total of 32. The 2023 edition also started with the same size as the 2022 and 2017 editions (1000 x 1000), and started with 8 colors. It was later expanded to 2 million (2000 x 1000) pixel squares, with 16 colors, t
https://en.wikipedia.org/wiki/Seatooth
Seatooth is a wireless technology standard for exchanging data through water and the water-air boundary using low frequency radio waves (from 1 Hz to 2.485 GHz). WFS Technologies Ltd launched Seatooth, the world's first commercially available underwater radio modem, to the subsea market in 2006. 2007 saw the launch of the first underwater wireless broadband data link, followed by the first hybrid radio/acoustic modem. In comparing wireless technologies subsea radio waves prefer shallow water and can cross the air/water/seabed boundaries easily. Subsea radio communication is generally limited to under through seawater. Subsea radio waves are unaffected by turbidity, salinity and pressure gradients and also has a notable difference between acoustic and optical technologies, in that radio waves can pass through the water-air and water-seabed boundaries easily. Subsea acoustics are efficient at long-range of up to and have relatively low power consumption for their range. Acoustic communication systems generally perform poorly in shallow water and complex environments and has a limited bandwidth. Subsea optical has an ultra-high bandwidth and a very short range. Subsea optical communication does not cross the water/air boundary and is susceptible to turbidity. Most underwater sensor networks choose acoustics as the medium for wireless transmission. Electromagnetic waves offer great merits for transmission in special underwater environments. Applications for subsea wireless sensor technologies can include subsea wireless sensor networks (WSN) for production monitoring, or oil and gas pipeline monitoring within a wireless linear sensor network (LSN). References External links Networking standards Wireless
https://en.wikipedia.org/wiki/Deformers
Deformers is a multiplayer brawler game developed by Ready at Dawn and published by GameTrust. Gameplay Deformers is a multiplayer brawler game played from a third-person perspective. Players take control of squishy ball-shaped creatures called a Forms that are capable of rolling, jumping, dashing, and absorbing items. In the deathmatch and team deathmatch modes, the goal is to roll around and knock opponents off the arena platform. Form-ball is a mode that plays like association football, a ball must be knocked into a net to score goals. The game supports online 2, 3, and 4-player splitscreen as well as online multiplayer with up to eight players. Development Deformers was developed by video game studio Ready at Dawn. The project began as a physics technology demonstration. A duo at the studio began working on prototype for Deformers in August 2014 while most employees were occupied with The Order: 1886. It was announced that Deformers would shut down on 9 August 2018. Release Deformers was initially scheduled to launch for Windows, PlayStation 4, and Xbox One on February 14, 2017, however it was delayed so that the development team could address issues from feedback they had received. An open beta of the Windows version ran from April 1–4 on digital distribution service Steam. Deformers was the second game published by GameStop's publishing program GameTrust. The game was released on April 21, 2017. Reception Deformers received a mixed reception from critics. It was nominated for "Original Score - Video Game" at the 2017 Hollywood Music in Media Awards. References External links Multiplayer video games Multiplayer online games Inactive multiplayer online games Online-only games PlayStation 4 games Indie games 2017 video games Video games developed in the United States Video games scored by Austin Wintory Video games with cross-platform play Windows games Xbox One games Products and services discontinued in 2018 Ready at Dawn games
https://en.wikipedia.org/wiki/Worksoft
Worksoft, Inc. is a software testing company founded in 1998 and headquartered in Addison, Texas. The company provides an automation platform for test automation, business process discovery, and documentation supporting enterprise applications, including packaged and web apps. In addition to its headquarters in Addison, Texas, the company has offices in London and Munich. History Worksoft was founded in 1998 by Linda Hayes, a co-founder of AutoTester, Inc, and was initially funded by a contract with Fidelity Investments for Y2K testing. Worksoft Certify was the first script-less automation tool designed for business analysts and is now a leader in the ERP automation industry. Texas-based Austin Ventures and California-based Crecendo Ventures are major investors. In 2010, Worksoft acquired TestFactory, a software testing company specializing in SAP. In 2019, Worksoft was acquired by Marlin Equity Partners for an undisclosed sum. Products Worksoft Certify is a test automation platform focused on business process testing. Worksoft Certify can be used to test ERP applications, web apps, mobile apps, and more. The software is SAP certified for integration with SAP applications. Other products include Worksoft Analyze, Worksoft Business Reporting Tool (BPP), Worksoft Execution Suite, and Process Capture 2.0. References Software companies established in 1998 1998 establishments in Texas Software testing Companies based in Addison, Texas 2019 mergers and acquisitions
https://en.wikipedia.org/wiki/Toukiden%202
is a 2016 action role-playing game developed by Omega Force and published by Koei Tecmo for Microsoft Windows, PlayStation 3, PlayStation 4, and PlayStation Vita. The game is the sequel to Toukiden: The Age of Demons. Gameplay New to the Toukiden series, the element of Control is also empowered by Mitama, which specialises in summoning Oni to fight for the player. Mitama grow and develop during battles. In order to increase the levels of your Boosts, you must meet the upgrade conditions of each Boost and have sufficient proficiency with them. The upgrade conditions differ for each Mitama and Boost. Each Mitama can learn up to nine Boosts, but you can only equip three of them at a time. World exploration has shifted to be open world. Hunting grounds from previous games return with alterations. Certain sections of the world map will be locked depending on the player's progression with the main story; destinations may also be marked when fulfilling requests. Mahoroba Village serves as the game's new primary hub,day and night system. The protagonist's spare inventory and setups can be accessed from chests scattered across the world map. Players can raise multiple Tenko and give them names. If well-cared for, they will occasionally find items and follow the protagonist. Combat has been altered to be seamless encounters that happen anywhere in the field. "Demon Hand" is a new secondary feature for traveling and fighting giant demons. The invention can hook onto objects or minor demons to quickly close distances or destroy obstacles. Against giant demons it can do the following: stun them, lock onto a specific body part or completely obliterate a demon's body part(s). This feature is controlled via touchscreen in the Vita version. Slayers have a new faster running ability Automatons are deployed to scavenge for items in different locations, though they may receive damage upon returning. Damaged automatons can be repaired or made stronger by augmenting them with var
https://en.wikipedia.org/wiki/Bathymodiolus%20platifrons
Bathymodiolus platifrons, described by Hashimoto and Okutani in 1994, is a deep-sea mussel that is common in hydrothermal vents and methane seeps in the Western Pacific Ocean. Symbiosis Bathymodiolus platifrons harbours methane-oxidizing bacteria in its gill, which help to transfer methane into material and energy to help it to thrive in such environments. References platifrons Molluscs described in 1994 Chemosynthetic symbiosis
https://en.wikipedia.org/wiki/Anriette%20Esterhuysen
Anriette Esterhuysen is a human rights defender and computer networking pioneer from South Africa. She has pioneered the use of Internet and Communications Technologies (ICTs) to promote social justice in South Africa and throughout the world, focusing on affordable Internet access. She was the Executive Director of the Association for Progressive Communications from 2000 until April 2017, when she became APC's Director of Policy and Strategy. In November 2019 United Nations Secretary-General António Guterres appointed Esterhuysen to Chair the Internet Governance Forum’s Multistakeholder Advisory Group. Education and work Esterhuysen holds a BA in Social Science, a Post Graduate Diploma in Library and Information Science and a BA in Musicology from the University of the Witwatersrand. During her university years, she was campus coordinator for the South African Student Press Union. In 1994, Anriette became the inaugural Executive Director of SANGONeT, a South African communications NGO, which itself was a merger of Worknet (a communications NGO and member of the Association for Progressive Communications) and the Development Resources Center's Handsnet initiative. SANGONeT provided Internet connectivity, technical training, and website hosting to civil society organisations, trade unions, and others engaged in the mass democratic movement. In 1992-1993 she was Director of Information Services at South Africa's Development Resources Centre, where she set up a library and online news service for the NGO sector in South and Southern Africa, while mobilising information technologies to facilitate information and communication exchange in the broader development sector. Prior to that, she was Chief Librarian and consultant for the South African Council of Churches (SACC), where she facilitated training in documentation techniques and information management. From 1980 onwards, Esterhuysen was active in the struggle against Apartheid. Through her work at the SACC, an
https://en.wikipedia.org/wiki/Head%20of%20Franz%20Kafka
The Head of Franz Kafka (), also known as the Statue of Kafka, is an outdoor sculpture by David Černý depicting Bohemian German-language writer Franz Kafka, installed outside the Quadrio shopping centre in Prague, Czech Republic. The kinetic sculpture is 11 metres tall and made of 42 rotating panels. Each layer is mechanized and rotates individually. References External links 2014 sculptures Busts of writers Cultural depictions of Franz Kafka Kinetic sculptures in the Czech Republic Monuments and memorials in Prague Outdoor sculptures in Prague Sculptures of men in the Czech Republic Works about Franz Kafka New Town, Prague Heads in the arts
https://en.wikipedia.org/wiki/Sex-linked%20barring
Sex-linked barring is a plumage pattern on individual feathers in chickens, which is characterized by alternating pigmented and apigmented bars. The pigmented bar can either contain red pigment (phaeomelanin) or black pigment (eumelanin) whereas the apigmented bar is always white. The locus is therefore often referred to as an ‘eumelanin diluter’ or ‘melanin disruptor’. Typical sex-linked barred breeds include the Barred Plymouth Rock, Delaware, Old English Crele Games as well as Coucou de Renne. Plumage appearance The presence of a white bar on a dark background is distinguishing sex-linked barring from Autosomal barring, another plumage pattern in chickens which is created by a black bar on a light color background (white/ beige or brown) as exemplified by the breed Egyptian Fayoumi. The absence of pigment in the white bar has been attributed to a lack of pigment producing cells (melanocytes) in the feather follicle during feather growth. Initially it was proposed that this lack was the result of cell death as a consequence of the B locus mutation but later research demonstrated that the lack is the result of premature cell differentiation rather than apoptosis. Male chicken of traditional sex-linked barred breeds like the Barred Plymouth Rock usually show much wider and clearer white bands than females of the same breed. Further characteristics of sex-linked barred chickens are the dilution of skin pigment in the legs as well as a white dot at the top of the head of freshly hatched chicks which can be used for autosexing: homozygous males have a much bigger spot than hemizygous females. Genetics Sex-linked barring has been established as the dominant locus B by traditional mendelian genetics in the beginning of the 20th century. The responsible gene was predicted to be located on the Z chromosome and since male birds are homogametic (ZZ), they can be either hetero- or homozygous for sex-linked barring. Females are always hemizygous at this locus (ZW). In 2
https://en.wikipedia.org/wiki/Criticism%20of%20Windows%2010
Windows 10, a proprietary operating system released by Microsoft in July 2015, has been criticized by reviewers and users. Due to issues mostly about privacy, it has been the subject of a number of negative assessments by various groups. General criticism Critics have noted that Windows10 heavily emphasizes freemium services and contains various advertising facilities. Some outlets have considered these to be a hidden "cost" of the free upgrade offer. Examples include media storefronts, Office 365, paid functionality in bundled games such as Microsoft Solitaire Collection, default settings that display promotions of "suggested" apps in Start menu and "tips" on the lock screen that may contain advertising, ads displayed in File Explorer for Office 365 subscriptions on Redstone 2 builds, and notifications promoting the Microsoft Edge web browser when a different browser is set as default. Update system Windows 10 Home is permanently set to download all updates automatically, including cumulative updates, security patches, and drivers, and users cannot individually select updates to install or not. Microsoft offers a diagnostic tool that can be used to hide updates and prevent them from being reinstalled, but only after they had been already installed, then uninstalled without rebooting the system. However, the software agreement states, specifically for users of Windows10 in Canada, that they may pause updates by disconnecting their device from the Internet. Tom Warren of The Verge felt that, given web browsers such as Google Chrome had already adopted such an automatic update system, such a requirement would help to keep all Windows10 devices secure, and felt that "if you're used to family members calling you for technical support because they've failed to upgrade to the latest Windows service pack or some malware disabled Windows Update then those days will hopefully be over." Concerns were raised that due to these changes, users would be unable to skip the au
https://en.wikipedia.org/wiki/Event%20storming
Event storming is a workshop-based method to quickly find out what is happening in the domain of a software program. Compared to other methods it is extremely lightweight and intentionally requires no support by a computer. The result is expressed in sticky notes on a wide wall. The business process is "stormed out" as a series of domain events which are denoted as orange stickies. It was invented by Alberto Brandolini in the context of domain-driven design (DDD). Event storming can be used as a means for business process modeling and requirements engineering. The idea is to bring together software developers and domain experts and learn from each other. The name was chosen to show that the focus should be on the domain events and the method works similar to brainstorming or agile modeling's model storming. Requirements It is important for an event storming workshop to have the right people present. This includes people who know the questions to ask (typically developers) and those who know the answers (domain experts, product owners). The modeling will be placed on a wide wall with a roll of paper rolled out on it. The sticky notes will be placed on this paper. You will require at least 5 distinct colors for the sticky notes. Steps The first step is to find the domain events and write them on orange sticky notes. When all domain events are found the second step is to find the command that caused each of the domain events. Commands are written on blue notes and placed directly before the corresponding domain event. In the third step the aggregates within which commands are executed and where events happen are identified. The aggregates are written in yellow stickies. The concepts gathered during an event storming session fall into several categories, each with its own colour of sticky note: An event that occurs in the business process. Written in past tense. A person who executes a command through a view. Processes a command according to business rule
https://en.wikipedia.org/wiki/Mongoose%20OS
Mongoose OS is an Internet of Things (IoT) Firmware Development Framework available under Apache License Version 2.0. It supports low power, connected microcontrollers such as: ESP32, ESP8266, TI CC3200, TI CC3220, STM32 (STM32L4, STM32F4, STM32F7 series). Its purpose is to be a complete environment for prototyping, development and managing connected devices. It is designed to reduce the time and costs associated with IoT projects. Mongoose OS serves as the gap between Arduino firmware suitable for prototyping and bare-metal microcontrollers' native SDKs. It is developed by Cesanta Software Ltd., a company based in Dublin (Ireland), and dual licensed. Features User friendly Over the Air (OTA) updating of embedded ICs. Secure connectivity and crypto support Integrated Mongoose Web Server Programming in either JavaScript (integrated mJS engine) or C. Integration with private and public clouds: AWS IoT, Microsoft Azure IoT, Google IoT Core, IBM Watson IoT, Mosquitto, HiveMQ, etc. License Mongoose OS is Open Source and dual-licensed: Mongoose OS Community Edition - Apache License Version 2.0 Mongoose OS Enterprise Edition - Commercial License References External links Embedded operating systems Free software operating systems Internet of things
https://en.wikipedia.org/wiki/P-variation
In mathematical analysis, p-variation is a collection of seminorms on functions from an ordered set to a metric space, indexed by a real number . p-variation is a measure of the regularity or smoothness of a function. Specifically, if , where is a metric space and I a totally ordered set, its p-variation is where D ranges over all finite partitions of the interval I. The p variation of a function decreases with p. If f has finite p-variation and g is an α-Hölder continuous function, then has finite -variation. The case when p is one is called total variation, and functions with a finite 1-variation are called bounded variation functions. Link with Hölder norm One can interpret the p-variation as a parameter-independent version of the Hölder norm, which also extends to discontinuous functions. If f is α–Hölder continuous (i.e. its α–Hölder norm is finite) then its -variation is finite. Specifically, on an interval [a,b], . Conversely, if f is continuous and has finite p-variation, there exists a reparameterisation, , such that is Hölder continuous. If p is less than q then the space of functions of finite p-variation on a compact set is continuously embedded with norm 1 into those of finite q-variation. I.e. . However unlike the analogous situation with Hölder spaces the embedding is not compact. For example, consider the real functions on [0,1] given by . They are uniformly bounded in 1-variation and converge pointwise to a discontinuous function f but this not only is not a convergence in p-variation for any p but also is not uniform convergence. Application to Riemann–Stieltjes integration If f and g are functions from  [a, b] to ℝ with no common discontinuities and with f having finite p-variation and g having finite q-variation, with then the Riemann–Stieltjes Integral is well-defined. This integral is known as the Young integral because it comes from . The value of this definite integral is bounded by the Young-Loève estimate as follows w
https://en.wikipedia.org/wiki/Computational%20thermodynamics
Computational thermodynamics is the use of computers to simulate thermodynamic problems specific to materials science, particularly used in the construction of phase diagrams. Several open and commercial programs exist to perform these operations. The concept of the technique is minimization of Gibbs free energy of the system; the success of this method is due not only to properly measuring thermodynamic properties, such as those in the list of thermodynamic properties, but also due to the extrapolation of the properties of metastable allotropes of the chemical elements. History The computational modeling of metal-based phase diagrams, which dates back to the beginning of the previous century mainly by Johannes van Laar and to the modeling of regular solutions, has evolved in more recent years to the CALPHAD (CALculation of PHAse Diagrams). This has been pioneered by American metallurgist Larry Kaufman since the 1970s. Current state Computational thermodynamics may be considered a part of materials informatics and is a cornerstone of the concepts behind the materials genome project. While crystallographic databases are used mainly as a reference source, thermodynamic databases represent one of the earliest examples of informatics, as these databases were integrated into thermochemical computations to map phase stability in binary and ternary alloys. Many concepts and software used in computational thermodynamics are credited to the SGTE Group, a consortium devoted to the development of thermodynamic databases; the open elements database is freely available based on the paper by Dinsdale. This so-called "unary" system proves to be a common basis for the development of binary and multiple systems and is used by both commercial and open software in this field. However, as stated in recent CALPHAD papers and meetings, such a Dinsdale/SGTE database will likely need to be corrected over time despite the utility in keeping a common base. In this case, most published as
https://en.wikipedia.org/wiki/Sunbury%20Research%20Centre
The Sunbury Research Centre -- also known as ICBT Sunbury—is a main research institute of BP in north-east Surrey. History It began in 1917 as the Sunbury Research Station. Research began with the employment of two chemists to look into the viscosity of fuel oil for the Navy in the First World War, and the production of toluene. In the 1920s research took place into cracking, at the plant at Uphall in Scotland (West Lothian). The first new building opened in July 1931. 76 staff were there in 1929, 99 in 1934 and 197 in 1939. By the 1950s, BP Research was in a 39-acre site in Sunbury. Geophysical research had also taken place at Kirklington Hall Research Station in Nottinghamshire, until 1957. Around 1958, the site was expanded with a new Physics laboratory and five other buildings. A Linear electron accelerator was installed. By early 1958, Kirklington Hall had been sold. Products that the British Petroleum Company made in the 1950s were BP Motor Spirit and BP Energol (visco-static motor oil). But Britain would not produce much oil of its own until the mid-1970s when North Sea oil arrived at the Forties Oil Field. Construction Three new buildings were built from 1998 as part of Phase 1. Since 2001, four new buildings were built as part of Phase 2. Structure It is situated off the A244 (via the A308) in the north of Sunbury-on-Thames, and Surrey, on the Surrey boundary with London. To the east nearby is Sunbury Common. The retail division of BP UK is at Witan Gate House. BP employs around 15,000 people in the UK. It has an enhanced oil recovery laboratory. See also Castrol Technology Centre Oil fields operated by BP Peter Mather (businessman), Head of BP UK National Physical Laboratory (United Kingdom) References The History of the British Petroleum Company, October 1982, External links ICBT Sunbury New building 1917 establishments in England 1917 in technology BP buildings and structures Chemical industry in the United Kingdom Chemical research in
https://en.wikipedia.org/wiki/Tricentis%20Tosca
Tricentis Tosca is a software testing tool that is used to automate end-to-end testing for software applications. It is developed by Tricentis. Tricentis Tosca combines multiple aspects of software testing (test case design, test automation, test data design and generation, and analytics) to test GUIs and APIs from a business perspective. Two of the most frequently-noted technologies used in Tricentis Tosca are related to Model-based testing and Risk-based testing. Model-based testing Instead of using scripting for test automation, Tricentis Tosca applies a model-based testing approach and creates a model of the application under test. Technical details about the application under test, test case logic, and test data are saved separately and joined together at test execution time. When an element in the application under test changes, the technical details are updated once in the central model. Since the test cases inherit from this model, the various test cases that test the updated element do not need to be modified manually to reflect the change. Risk-based testing Based on a risk assessment of the application under test’s requirements, Tricentis Tosca uses risk-based test design to suggest the most effective test cases and identify the risk contribution of each test cases. It also uses a variety of methodologies (such as equivalence partitioning, boundary testing, and combinatorial methods such as linear expansion) to try to minimize the number of test cases while increasing risk coverage. After the tests are executed, the tool aggregates risk coverage from business, technical, and compliance perspectives. Additional technologies Tricentis Tosca features technologies for GUI testing, API testing, Mobile testing, Service virtualization, Test data design and generation, Business intelligence and data warehouse testing, and Exploratory testing. It has SAP-certified integration with SAP solutions, and is used to automate testing for SAP technologies such as S
https://en.wikipedia.org/wiki/Instabug
Instabug is a software company that provides bug reporting, app performance monitoring, crash reporting, in-app chats, and user surveys for mobile apps. The company was founded in 2014. As of September 2019, Instabug has reached over 25,000 companies, 400 millions reported issues and feedbacks received, and 2 billion devices running their SDK worldwide. Instabug's software development kit (SDK) used by Android, iOS, Cordova, Ionic, Xamarin, and web developers during beta testing as well as in live production versions of their apps. It is known for customizable “Shake to Send” feature on the mobile user side and detailed reports on the developer side. Users can attach annotated screenshots, video recordings, and voice notes to supplement their bug reports, which automatically includes network and device logs and repro steps. The SDK also integrates with a range of third-party tools used by developers, including Slack, Zapier, JIRA, Trello, Zendesk, and more. History Instabug started in 2012 in Cairo, Egypt by two co-founders: Omar Gabr and Moataz Soliman. During their last semester at Cairo University, Gabr and Soliman created AStarApps, which offered a range of services to mobile developers, including an extremely early iOS version of Instabug's bug reporting tool with its intuitive “Shake to Send” feature. Shortly after, Gabr and Soliman pivoted from AStarApps to focus on the bug reporter and what is now known as Instabug, Inc. They were incubated by Flat6Labs. In 2013, Instabug secured early seed investments from Georges Harik, Cairo Angels, Middle East Venture Partners, and angel investor Hala Fadel. They also won first place and $50,000 in the MIT Enterprise Forum Arab Startup Competition that year. In 2014, Instabug introduced an Android SDK and crash reporting. In 2016, Instabug was incubated by Y Combinator, “the world’s most powerful startup incubator”, as part of its Winter batch. In June of that year, Instabug raised $1.7 million in seed funding led
https://en.wikipedia.org/wiki/Model-driven%20application
A model-driven application is a software application that the functions or behaviors are based on, or in control of, some evolutionary applied models of the target things to the application. The applied models are served as a part of the application system in which it can be changed at runtime. The target things are what the application deals with, such as the objects and affairs in business for a business application. Follows the definition of application in TOGAF, a model-driven business application could be described as an IT system that supports business functions and services running on the models of the (things in) business. History The ideal of the architecture for a model-driven application was first put forward by Tong-Ying Yu on the Enterprise Engineering Forum in 1999, which have been studied and spread through some internet media for a long time. It had influence on the field of enterprise application development in China; there were successful cases of commercial development of enterprise/business applications in the architectural style of a model-driven application. Gartner Group carried out some studies into the subject in 2008; they defined the model-driven packaged applications as "enterprise applications that have explicit metadata-driven models of the supported processes, data and relationships, and that generate runtime components through metadata models, either dynamically interpreted or compiled, rather than hardcoded." The model-driven application architecture is one of few technology trends to driven the next generation of application modernization, that claimed by some industrial researchers in 2012. Instance Business process management (BPM) is the significant practice to the model-driven application. According to the definition, a BPM system is model-driven if the functions are operated based on the business process models which are built and changed at the operational time but not the design or implementation time; the biggest advantage
https://en.wikipedia.org/wiki/H%C3%A9ctor%20Altuve%20Ferrer
Héctor J. Altuve Ferrer (born October 5, 1947) is a Dean at the Schweitzer Engineering Laboratories, Monterrey, Mexico. He was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2015 for contributions to power line and transformer protection. Altuve Ferrer attended Central University of Las Villas in Santa Clara from which he got his BSEE. After obtaining Ph.D. from the Igor Sikorsky Kyiv Polytechnic Institute, he served on the faculty of the School of Electrical Engineering at Central University from 1969 to 1993. Between 1993 and 2000, Altuve served as professor of the Graduate Doctoral Program in the Mechanical and Electrical Engineering School at the Universidad Autónoma de Nuevo León in Monterrey, Mexico and from 1999 to 2000 was the Schweitzer Visiting Professor in the Department of Electrical Engineering at Washington State University. References 20th-century births Living people Academic staff of the Autonomous University of Nuevo León Washington State University faculty Fellow Members of the IEEE Year of birth missing (living people) Place of birth missing (living people)
https://en.wikipedia.org/wiki/Natalino%20Camilleri
Natalino Camilleri from the Nitero, Inc., Austin, TX was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2015 for leadership in radio frequency integrated circuits and systems. Early life Camilleri received the B.S.E.E. degree from the University of Malta and the M.S.E.E. and Ph.D. degrees from the University of Texas at Austin. Career He possesses over 20 years of experience in the development and creation of RF systems and integrated circuits. He is currently the President of RF & Wireless Design Services Inc., Cupertino, CA, a consulting company performing RFIC design. Earlier in his career, he started up the Radio Product Line at Advanced Micro Devices, and was able to take some early product ideas and develop the award-winning chip set for digital spread-spectrum (DSS) cordless phones. He led the group through an acquisition by DSPG and was the President of RF Integrated Systems Inc. He was instrumental in the development of the 900-MHz giga-range phones by Panasonic that use DSS chip sets. At AMD, he also developed integrated circuits (ICs) for wireless local-area network (WLAN) and GSM applications. Prior to his involvement with AMD, he was the Silicon RF Technology Manager at Motorola, where he developed leading-edge circuits for new technologies such as SiGe, BiCMOS, LDMOS, and CMOS. He was instrumental in developing the LDMOS discrete technology and was the founding father for the LDMOS IC technology for Motorola. These products are now the premiere technology for Motorola’s RF business. He originally joined Motorola to start up the GaAs operation as the Design and Application Manager. He was instrumental in kicking off the early technology and IC products that later fueled GaAs RFIC business for Motorola. Prior to Motorola, he was with HP-Avantek and Texas Instruments, where he was a Senior Member of the Technical Staff and developed ICs for microwave and millimeter-wave applications. He has authored or co-authored
https://en.wikipedia.org/wiki/Youngky%20Kim
Youngky Kim from the Networks Business, Samsung Electronics, Suwon, South Korea was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2015 for leadership in mobile communication systems. References Fellow Members of the IEEE Living people Year of birth missing (living people) Place of birth missing (living people) Samsung people
https://en.wikipedia.org/wiki/Risk%20score
Risk score (or risk scoring) is the name given to a general practice in applied statistics, bio-statistics, econometrics and other related disciplines, of creating an easily calculated number (the score) that reflects the level of risk in the presence of some risk factors (e.g. risk of mortality or disease in the presence of symptoms or genetic profile, risk financial loss considering credit and financial history, etc.). Risk scores are designed to be: Simple to calculate: In many cases all you need to calculate a score is a pen and a piece of paper (although some scores use rely on more sophisticated or less transparent calculations that require a computer program). Easily interpreted: The result of the calculation is a single number, and higher score usually means higher risk. Furthermore, many scoring methods enforce some form of monotonicity along the measured risk factors to allow a straight forward interpretation of the score (e.g. risk of mortality only increases with age, risk of payment default only increase with the amount of total debt the customer has, etc.). Actionable: Scores are designed around a set of possible actions that should be taken as a result of the calculated score. Effective score-based policies can be designed and executed by setting thresholds on the value of the score and associating them with escalating actions. Formal definition A typical scoring method is composed of 3 components: A set of consistent rules (or weights) that assign a numerical value ("points") to each risk factor that reflect our estimation of underlying risk. A formula (typically a simple sum of all accumulated points) that calculates the score. A set of thresholds that helps to translate the calculated score into a level of risk, or an equivalent formula or set of rules to translate the calculated score back into probabilities (leaving the nominal evaluation of severity to the practitioner). Items 1 & 2 can be achieved by using some form of regression, th
https://en.wikipedia.org/wiki/Dan%20Keun%20Sung
Dan Keun Sung (born 19 July 1952) is a South Korean professor of electronic engineering at the Korea Advanced Institute of Science and Technology (KAIST) in Daejeon. Sung was born in South Gyeongsang Province. He received his B.Sc. in electronic engineering from Seoul National University in 1975 before going on to the University of Texas at Austin, where he received his M.Sc. (1982) and Ph.D. (1986) in electrical and computer engineering. He was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2015 for contributions to network resource management. He is also a member of the Korean Academy of Science and Technology. References 1952 births Living people People from South Gyeongsang Province Seoul National University alumni Cockrell School of Engineering alumni Academic staff of KAIST Fellow Members of the IEEE
https://en.wikipedia.org/wiki/Wonyong%20Sung
Wonyong Sung (born 1955) is a South Korean professor of electronic and information engineering at Seoul National University (SNU). Sung received his B.S. in engineering from SNU in 1978 and his M.S. in the same field from KAIST in 1980. After working for GoldStar for three years, he went on to the University of California, Santa Barbara for his Ph.D., which he completed in 1987. He was named a Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2015 for "contributions to real-time signal processing systems". References 1955 births Living people Seoul National University alumni KAIST alumni University of California, Santa Barbara alumni Academic staff of Seoul National University Fellow Members of the IEEE
https://en.wikipedia.org/wiki/Gyu%20Hyeong%20Cho
Gyu Hyeong Cho (; Hanja: ; born 1953) is an electrical engineer at the Korea Advanced Institute of Science and Technology (KAIST) in Daejeon, South Korea. Cho was named a Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for his contributions to power management circuit design. References 1953 births Living people Fellow Members of the IEEE South Korean engineers Place of birth missing (living people)
https://en.wikipedia.org/wiki/Wan%20Kyun%20Chung
Wan Kyun Chung from the Pohang University of Science and Technology, Pohang, Korea was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for developments in robust control theory for mechanical systems. References Fellow Members of the IEEE Living people Academic staff of Pohang University of Science and Technology Year of birth missing (living people) Place of birth missing (living people)
https://en.wikipedia.org/wiki/Robert%20Deng
Robert Deng from the Singapore Management University, Singapore was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for contributions to security algorithms, protocols and systems. References Fellow Members of the IEEE Living people Year of birth missing (living people) Place of birth missing (living people)
https://en.wikipedia.org/wiki/Family%20symmetries
In particle physics, the family symmetries or horizontal symmetries are various discrete, global, or local symmetries between quark-lepton families or generations. In contrast to the intrafamily or vertical symmetries (collected in the conventional Standard Model and Grand Unified Theories) which operate inside each family, these symmetries presumably underlie physics of the family flavors. They may be treated as a new set of quantum charges assigned to different families of quarks and leptons. Spontaneous symmetry breaking of these symmetries is believed to lead to an adequate description of the flavor mixing of quarks and leptons of different families.  This is certainly one of the major problems that presently confront particle physics. Despite its great success in explaining the basic interactions of nature, the Standard Model still suffers from an absence of such a unique ability to explain the flavor mixing angles or weak mixing angles (as they are conventionally referred to) whose observed values are collected in the corresponding Cabibbo–Kobayashi–Maskawa matrices. While being conceptually useful and leading in some cases to the physically valuable patterns of the flavor mixing, the family symmetries are not yet observationally confirmed. Introduction The Standard Model is based on the internal symmetries of the unitary product group  the members of which have a quite different nature. The color symmetry  has the vectorlike structure due to which the lefthanded and righthanded quarks are transformed identically as its fundamental triplets. At the same time, the electroweak symmetry consisting of the weak isospin and hypercharge is chiral. So, the lefthanded components of all quarks and leptons are the doublets,     whereas  their righthanded components are its singlets    Here, the quark-lepton families are numbered by the index both for the quark and lepton ones. The up and down righthanded quarks and leptons are written separately and f
https://en.wikipedia.org/wiki/Society%20for%20Research%20on%20Biological%20Rhythms
The Society for Research on Biological Rhythms (SRBR) is an international chronobiological research society with three key goals: to promote the advancement and dissemination of basic and applied research in all aspects of biological rhythms. to enhance the education and training of students and researchers in the field. to foster interdisciplinary communication and an international exchange of ideas. The society holds biennial meetings and informal gatherings, and participates in peer-reviewed science and evidence-based policy making. It is one of four prominent existing Chronology Research Societies and one of the 14 societies that make up The World Federation of Societies for Chronobiology. The organization is currently composed of 1,000 scientists and clinicians dedicated to studying biological rhythms and their impact. The society has its own official journal, the Journal of Biological Rhythms. Through its journal and meetings the society engages scientists of all backgrounds and nationalities. It advocates the need for funding in research areas in biological rhythms such as sleep and supports other research efforts such as the National Institutes of Health and National Science Foundation. Founding and Early History In 1986, Benjamin Rusak founded the Journal of Biological Rhythms. Rusak wanted an accompanying society that would hold meetings concerning research on biological rhythms, so he asked Fred W. Turek to organize the first meeting. The society was officially founded on November 12, 1986 by Fred Turek, Dave Hudson, Joe Takahashi, and Gene Block. The society is sometimes cited as being founded in 1988, as this was when the first meeting occurred. Colin Pittendrigh, Turek's Ph.D. adviser, and one of the fathers of the field, was initially opposed to the society, worrying that it would have an isolating effect on the field rather than connecting it to related disciplines. Pittendrigh came around to the idea after the successful first meeting. Ture
https://en.wikipedia.org/wiki/Minerva%20Cordero
Minerva Cordero Braña is a Puerto Rican mathematician and a professor of mathematics at the University of Texas at Arlington. She is also the university's Senior Associate Dean for the College of Science, where she is responsible for the advancement of the research mission of the college. President Biden awarded her the Presidential Award for Excellence in Science, Mathematics, and Engineering Mentoring (PAESMEM) on February 8, 2022. Early life and education Cordero was born in Bayamón, Puerto Rico. Her mother, whose schooling stopped after the fifth grade, made education a top priority in the family home. She told her children "the best thing I can give you is an education." Cordero and her siblings would do their homework together and discussed what they learned in school each day. Cordero said, "We learned each other's subjects." Wanting to go to college, Cordero bought herself a college exam preparation book in high school and studied for the college-entrance exam. She states that her exam scores were the highest scores for her high school, Miguel Melendez Munoz High School. Cordero attended the Universidad de Puerto Rico in Rio Piedras and received her B.S. in Mathematics in 1981. She was granted a National Science Foundation Minority Graduate Fellowship which she used to attend the University of California at Berkeley to obtain her masters in mathematics in 1983. She continued her studies at the University of Iowa, and obtained her Ph.D. in mathematics in 1989 under Norman Johnson. Career and research Cordero's research is in the area of finite semifields (non-associative algebras) and their associated planes (viewed affinely or projectively) in the general area of finite geometry. After earning her Ph.D., Cordero worked as an associate and an assistant professor at Texas Tech University until 2001, when she joined the faculty at the University of Texas at Arlington. Cordero served as the Mathematical Association of America's Governor-at-Large for Minor
https://en.wikipedia.org/wiki/ChatScript
ChatScript is a combination Natural Language engine and dialog management system designed initially for creating chatbots, but is currently also used for various forms of NL processing. It is written in C++. The engine is an open source project at SourceForge. and GitHub. ChatScript was written by Bruce Wilcox and originally released in 2011, after Suzette (written in ChatScript) won the 2010 Loebner Prize, fooling one of four human judges. Features In general ChatScript aims to author extremely concisely, since the limiting scalability of hand-authored chatbots is how much/fast one can write the script. Because ChatScript is designed for interactive conversation, it automatically maintains user state across volleys. A volley is any number of sentences the user inputs at once and the chatbots response. The basic element of scripting is the rule. A rule consists of a type, a label (optional), a pattern, and an output. There are three types of rules. Gambits are something a chatbot might say when it has control of the conversation. Rejoinders are rules that respond to a user remark tied to what the chatbot just said. Responders are rules that respond to arbitrary user input which is not necessarily tied to what the chatbot just said. Patterns describe conditions under which a rule may fire. Patterns range from extremely simplistic to deeply complex (analogous to Regex but aimed for NL). Heavy use is typically made of concept sets, which are lists of words sharing a meaning. ChatScript contains some 2000 predefined concepts and scripters can easily write their own. Output of a rule intermixes literal words to be sent to the user along with common C-style programming code. Rules are bundled into collections called topics. Topics can have keywords, which allows the engine to automatically search the topic for relevant rules based on user input. Example code Topic: ~food( ~fruit fruit food eat) t: What is your favorite food? a: (~fruit) I like fruit
https://en.wikipedia.org/wiki/Rabih%20Jabr
Rabih Jabr from the American University of Beirut, Lebanon was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for application of robust optimization to power systems. References Fellow Members of the IEEE Living people Year of birth missing (living people) Place of birth missing (living people) Academic staff of the American University of Beirut
https://en.wikipedia.org/wiki/Joungho%20Kim
Joungho Kim is an electrical engineer with the Korea Advanced Institute of Science and Technology (KAIST) in Daejeon, South Korea. Kim was named a Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for his contributions to modeling signal and power integrity in 3D integrated circuits. References Fellow Members of the IEEE Living people South Korean engineers Year of birth missing (living people) Academic staff of KAIST Place of birth missing (living people)
https://en.wikipedia.org/wiki/Lee%20Jong-ho%20%28engineer%29
Lee Jong-ho () is a South Korean electronic engineer and professor of electrical and computer engineering at Seoul National University. He serves as Minister of Science and ICT in the Yoon Suk-yeol government since May 2022. He was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for contributions to development and characterization of bulk multiple-gate field effect transistors. The following year he received the Kyung-Ahm Prize in Engineering. References Living people 1966 births People from South Gyeongsang Province Kyungpook National University alumni Seoul National University alumni Academic staff of Kyungpook National University Academic staff of Seoul National University South Korean engineers Electronics engineers Science ministers Government ministers of South Korea Fellow Members of the IEEE
https://en.wikipedia.org/wiki/Inkyu%20Lee
Inkyu Lee from the Korea University, Seoul, Korea was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for contributions to multiple antenna systems for wireless communications. References Fellow Members of the IEEE Living people Year of birth missing (living people) Place of birth missing (living people)
https://en.wikipedia.org/wiki/Weisi%20Lin
Weisi Lin is a professor at the School of Computer Science and Engineering of Nanyang Technological University (NTU), Singapore. He was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 "for contributions to perceptual modeling and processing of visual signals." He is also a Fellow of IET. He has been a Highly Cited Researcher (2019, 2020, 2021, and 2022, awarded by Clarivate Analytics), and a Distinguished Lecturer for both IEEE Circuits and Systems Society (2016-2017) and Asia-Pacific Signal and Information Processing Association (APSIPA, 2012-2013). He has been elected for the COE Research Award 2023, NTU. References Fellow Members of the IEEE Living people Academic staff of Nanyang Technological University Singaporean engineers Year of birth missing (living people)
https://en.wikipedia.org/wiki/Andrea%20Neto
Andrea Neto from the TU Delft- Delft University of Technology, Delft, Netherlands was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for contributions to dielectric lens antennas and wideband arrays. References Fellow Members of the IEEE Living people Year of birth missing (living people) Place of birth missing (living people) Dutch electrical engineers Academic staff of the Delft University of Technology
https://en.wikipedia.org/wiki/Branislav%20Notaros
Branislav Notaros is an electrical engineer from Colorado State University in Fort Collins. He was named a Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for his contributions to higher order methods in computational electromagnetics. He serves as the General Chair for the 2022 IEEE International Symposium on Antennas and Propagation and USNC-URSI Radio Science Meeting and a Track Editor for the IEEE Transactions on Antennas and Propagation. He serves as the President for the Applied Computational Electromagnetics Society (ACES), the Chair for USNC-URSI Commission B, and the Meetings Committee Chair for the IEEE Antennas and Propagation Society Personal life Dr. Notaros received Dipl.Ing. (B.S.) in Electrical Engineering, University of Belgrade, Sch. of Elec. Eng., Yugoslavia in 1988, M.S. in Electrical Engineering, University of Belgrade, School of Electrical Eng., Yugoslavia in 1992, and Ph.D. in Electrical Engineering, University of Belgrade, School of Electrical Eng., Yugoslavia in 1995. Dr. Notaros teaches many classes at Colorado State University in the Electrical and Computer Engineering Department, with a concentration on electromagnetics, antennas, and radar. He is married to Olivera Notaros, who also teaches in the department. References Fellow Members of the IEEE Serbian engineers Living people Year of birth missing (living people) Colorado State University faculty Place of birth missing (living people) American electrical engineers
https://en.wikipedia.org/wiki/Pablo%20Parrilo
Pablo A. Parrilo from MIT (Massachusetts Institute of Technology) was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2016 for contributions to semidefinite and sum-of-squares optimization. He was named a SIAM Fellow in 2018. References Fellow Members of the IEEE Fellows of the Society for Industrial and Applied Mathematics Living people Massachusetts Institute of Technology faculty 21st-century American engineers Year of birth missing (living people)